Rev 1702 | Go to most recent revision | Details | Compare with Previous | Last modification | View Log | RSS feed
| Rev | Author | Line No. | Line |
|---|---|---|---|
| 1281 | palkovsky | 1 | /* |
| 2 | * Copyright (C) 2006 Ondrej Palkovsky |
||
| 3 | * All rights reserved. |
||
| 4 | * |
||
| 5 | * Redistribution and use in source and binary forms, with or without |
||
| 6 | * modification, are permitted provided that the following conditions |
||
| 7 | * are met: |
||
| 8 | * |
||
| 9 | * - Redistributions of source code must retain the above copyright |
||
| 10 | * notice, this list of conditions and the following disclaimer. |
||
| 11 | * - Redistributions in binary form must reproduce the above copyright |
||
| 12 | * notice, this list of conditions and the following disclaimer in the |
||
| 13 | * documentation and/or other materials provided with the distribution. |
||
| 14 | * - The name of the author may not be used to endorse or promote products |
||
| 15 | * derived from this software without specific prior written permission. |
||
| 16 | * |
||
| 17 | * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR |
||
| 18 | * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES |
||
| 19 | * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. |
||
| 20 | * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, |
||
| 21 | * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT |
||
| 22 | * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
||
| 23 | * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
||
| 24 | * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
||
| 25 | * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF |
||
| 26 | * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
||
| 27 | */ |
||
| 28 | |||
| 1757 | jermar | 29 | /** @addtogroup genericipc |
| 1702 | cejka | 30 | * @{ |
| 31 | */ |
||
| 1757 | jermar | 32 | /** |
| 33 | * @file |
||
| 34 | * @brief IRQ notification framework. |
||
| 1284 | palkovsky | 35 | * |
| 36 | * This framework allows applications to register to receive a notification |
||
| 37 | * when interrupt is detected. The application may provide a simple 'top-half' |
||
| 38 | * handler as part of its registration, which can perform simple operations |
||
| 39 | * (read/write port/memory, add information to notification ipc message). |
||
| 40 | * |
||
| 41 | * The structure of a notification message is as follows: |
||
| 1693 | palkovsky | 42 | * - METHOD: interrupt number |
| 43 | * - ARG1: payload modified by a 'top-half' handler |
||
| 44 | * - ARG2: payload |
||
| 45 | * - ARG3: payload |
||
| 46 | * - in_phone_hash: interrupt counter (may be needed to assure correct order |
||
| 1284 | palkovsky | 47 | * in multithreaded drivers) |
| 48 | */ |
||
| 49 | |||
| 1281 | palkovsky | 50 | #include <arch.h> |
| 51 | #include <mm/slab.h> |
||
| 52 | #include <errno.h> |
||
| 53 | #include <ipc/ipc.h> |
||
| 54 | #include <ipc/irq.h> |
||
| 1284 | palkovsky | 55 | #include <atomic.h> |
| 1288 | jermar | 56 | #include <syscall/copy.h> |
| 1507 | vana | 57 | #include <console/console.h> |
| 1281 | palkovsky | 58 | |
| 59 | typedef struct { |
||
| 60 | SPINLOCK_DECLARE(lock); |
||
| 61 | answerbox_t *box; |
||
| 62 | irq_code_t *code; |
||
| 1284 | palkovsky | 63 | atomic_t counter; |
| 1281 | palkovsky | 64 | } ipc_irq_t; |
| 65 | |||
| 66 | |||
| 67 | static ipc_irq_t *irq_conns = NULL; |
||
| 68 | static int irq_conns_size; |
||
| 69 | |||
| 70 | #include <print.h> |
||
| 71 | /* Execute code associated with IRQ notification */ |
||
| 72 | static void code_execute(call_t *call, irq_code_t *code) |
||
| 73 | { |
||
| 74 | int i; |
||
| 1693 | palkovsky | 75 | __native dstval = 0; |
| 1628 | decky | 76 | |
| 1281 | palkovsky | 77 | if (!code) |
| 78 | return; |
||
| 79 | |||
| 80 | for (i=0; i < code->cmdcount;i++) { |
||
| 81 | switch (code->cmds[i].cmd) { |
||
| 82 | case CMD_MEM_READ_1: |
||
| 1693 | palkovsky | 83 | dstval = *((__u8 *)code->cmds[i].addr); |
| 1281 | palkovsky | 84 | break; |
| 85 | case CMD_MEM_READ_2: |
||
| 1693 | palkovsky | 86 | dstval = *((__u16 *)code->cmds[i].addr); |
| 1281 | palkovsky | 87 | break; |
| 88 | case CMD_MEM_READ_4: |
||
| 1693 | palkovsky | 89 | dstval = *((__u32 *)code->cmds[i].addr); |
| 1281 | palkovsky | 90 | break; |
| 91 | case CMD_MEM_READ_8: |
||
| 1693 | palkovsky | 92 | dstval = *((__u64 *)code->cmds[i].addr); |
| 1281 | palkovsky | 93 | break; |
| 94 | case CMD_MEM_WRITE_1: |
||
| 95 | *((__u8 *)code->cmds[i].addr) = code->cmds[i].value; |
||
| 96 | break; |
||
| 97 | case CMD_MEM_WRITE_2: |
||
| 98 | *((__u16 *)code->cmds[i].addr) = code->cmds[i].value; |
||
| 99 | break; |
||
| 100 | case CMD_MEM_WRITE_4: |
||
| 101 | *((__u32 *)code->cmds[i].addr) = code->cmds[i].value; |
||
| 102 | break; |
||
| 103 | case CMD_MEM_WRITE_8: |
||
| 104 | *((__u64 *)code->cmds[i].addr) = code->cmds[i].value; |
||
| 105 | break; |
||
| 1284 | palkovsky | 106 | #if defined(ia32) || defined(amd64) |
| 107 | case CMD_PORT_READ_1: |
||
| 1693 | palkovsky | 108 | dstval = inb((long)code->cmds[i].addr); |
| 1284 | palkovsky | 109 | break; |
| 110 | case CMD_PORT_WRITE_1: |
||
| 111 | outb((long)code->cmds[i].addr, code->cmds[i].value); |
||
| 112 | break; |
||
| 113 | #endif |
||
| 1507 | vana | 114 | #if defined(ia64) |
| 115 | case CMD_IA64_GETCHAR: |
||
| 1693 | palkovsky | 116 | dstval = _getc(&ski_uconsole); |
| 1507 | vana | 117 | break; |
| 118 | #endif |
||
| 1628 | decky | 119 | #if defined(ppc32) |
| 1625 | decky | 120 | case CMD_PPC32_GETCHAR: |
| 1693 | palkovsky | 121 | dstval = cuda_get_scancode(); |
| 1625 | decky | 122 | break; |
| 123 | #endif |
||
| 1281 | palkovsky | 124 | default: |
| 125 | break; |
||
| 126 | } |
||
| 1693 | palkovsky | 127 | if (code->cmds[i].dstarg && code->cmds[i].dstarg < 4) { |
| 128 | call->data.args[code->cmds[i].dstarg] = dstval; |
||
| 129 | } |
||
| 1281 | palkovsky | 130 | } |
| 131 | } |
||
| 132 | |||
| 133 | static void code_free(irq_code_t *code) |
||
| 134 | { |
||
| 135 | if (code) { |
||
| 136 | free(code->cmds); |
||
| 137 | free(code); |
||
| 138 | } |
||
| 139 | } |
||
| 140 | |||
| 141 | static irq_code_t * code_from_uspace(irq_code_t *ucode) |
||
| 142 | { |
||
| 143 | irq_code_t *code; |
||
| 144 | irq_cmd_t *ucmds; |
||
| 1288 | jermar | 145 | int rc; |
| 1281 | palkovsky | 146 | |
| 147 | code = malloc(sizeof(*code), 0); |
||
| 1288 | jermar | 148 | rc = copy_from_uspace(code, ucode, sizeof(*code)); |
| 149 | if (rc != 0) { |
||
| 150 | free(code); |
||
| 151 | return NULL; |
||
| 152 | } |
||
| 1281 | palkovsky | 153 | |
| 154 | if (code->cmdcount > IRQ_MAX_PROG_SIZE) { |
||
| 155 | free(code); |
||
| 156 | return NULL; |
||
| 157 | } |
||
| 158 | ucmds = code->cmds; |
||
| 159 | code->cmds = malloc(sizeof(code->cmds[0]) * (code->cmdcount), 0); |
||
| 1288 | jermar | 160 | rc = copy_from_uspace(code->cmds, ucmds, sizeof(code->cmds[0]) * (code->cmdcount)); |
| 161 | if (rc != 0) { |
||
| 162 | free(code->cmds); |
||
| 163 | free(code); |
||
| 164 | return NULL; |
||
| 165 | } |
||
| 1281 | palkovsky | 166 | |
| 167 | return code; |
||
| 168 | } |
||
| 169 | |||
| 170 | /** Unregister task from irq */ |
||
| 171 | void ipc_irq_unregister(answerbox_t *box, int irq) |
||
| 172 | { |
||
| 173 | ipl_t ipl; |
||
| 1595 | palkovsky | 174 | int mq = irq + IPC_IRQ_RESERVED_VIRTUAL; |
| 1281 | palkovsky | 175 | |
| 176 | ipl = interrupts_disable(); |
||
| 1595 | palkovsky | 177 | spinlock_lock(&irq_conns[mq].lock); |
| 178 | if (irq_conns[mq].box == box) { |
||
| 179 | irq_conns[mq].box = NULL; |
||
| 180 | code_free(irq_conns[mq].code); |
||
| 181 | irq_conns[mq].code = NULL; |
||
| 1281 | palkovsky | 182 | } |
| 183 | |||
| 1595 | palkovsky | 184 | spinlock_unlock(&irq_conns[mq].lock); |
| 1281 | palkovsky | 185 | interrupts_restore(ipl); |
| 186 | } |
||
| 187 | |||
| 188 | /** Register an answerbox as a receiving end of interrupts notifications */ |
||
| 189 | int ipc_irq_register(answerbox_t *box, int irq, irq_code_t *ucode) |
||
| 190 | { |
||
| 191 | ipl_t ipl; |
||
| 192 | irq_code_t *code; |
||
| 1595 | palkovsky | 193 | int mq = irq + IPC_IRQ_RESERVED_VIRTUAL; |
| 1281 | palkovsky | 194 | |
| 195 | ASSERT(irq_conns); |
||
| 196 | |||
| 197 | if (ucode) { |
||
| 198 | code = code_from_uspace(ucode); |
||
| 199 | if (!code) |
||
| 200 | return EBADMEM; |
||
| 201 | } else |
||
| 202 | code = NULL; |
||
| 203 | |||
| 204 | ipl = interrupts_disable(); |
||
| 1595 | palkovsky | 205 | spinlock_lock(&irq_conns[mq].lock); |
| 1281 | palkovsky | 206 | |
| 1595 | palkovsky | 207 | if (irq_conns[mq].box) { |
| 208 | spinlock_unlock(&irq_conns[mq].lock); |
||
| 1281 | palkovsky | 209 | interrupts_restore(ipl); |
| 210 | code_free(code); |
||
| 211 | return EEXISTS; |
||
| 212 | } |
||
| 1595 | palkovsky | 213 | irq_conns[mq].box = box; |
| 214 | irq_conns[mq].code = code; |
||
| 215 | atomic_set(&irq_conns[mq].counter, 0); |
||
| 216 | spinlock_unlock(&irq_conns[mq].lock); |
||
| 1281 | palkovsky | 217 | interrupts_restore(ipl); |
| 218 | |||
| 219 | return 0; |
||
| 220 | } |
||
| 221 | |||
| 1595 | palkovsky | 222 | /** Add call to proper answerbox queue |
| 223 | * |
||
| 224 | * Assume irq_conns[mq].lock is locked */ |
||
| 225 | static void send_call(int mq, call_t *call) |
||
| 226 | { |
||
| 227 | spinlock_lock(&irq_conns[mq].box->irq_lock); |
||
| 228 | list_append(&call->link, &irq_conns[mq].box->irq_notifs); |
||
| 229 | spinlock_unlock(&irq_conns[mq].box->irq_lock); |
||
| 230 | |||
| 231 | waitq_wakeup(&irq_conns[mq].box->wq, 0); |
||
| 232 | } |
||
| 233 | |||
| 234 | /** Send notification message |
||
| 235 | * |
||
| 236 | */ |
||
| 1693 | palkovsky | 237 | void ipc_irq_send_msg(int irq, __native a1, __native a2, __native a3) |
| 1595 | palkovsky | 238 | { |
| 239 | call_t *call; |
||
| 240 | int mq = irq + IPC_IRQ_RESERVED_VIRTUAL; |
||
| 241 | |||
| 242 | spinlock_lock(&irq_conns[mq].lock); |
||
| 243 | |||
| 244 | if (irq_conns[mq].box) { |
||
| 245 | call = ipc_call_alloc(FRAME_ATOMIC); |
||
| 246 | if (!call) { |
||
| 247 | spinlock_unlock(&irq_conns[mq].lock); |
||
| 248 | return; |
||
| 249 | } |
||
| 250 | call->flags |= IPC_CALL_NOTIF; |
||
| 1693 | palkovsky | 251 | IPC_SET_METHOD(call->data, irq); |
| 252 | IPC_SET_ARG1(call->data, a1); |
||
| 1595 | palkovsky | 253 | IPC_SET_ARG2(call->data, a2); |
| 254 | IPC_SET_ARG3(call->data, a3); |
||
| 1693 | palkovsky | 255 | /* Put a counter to the message */ |
| 256 | call->private = atomic_preinc(&irq_conns[mq].counter); |
||
| 1595 | palkovsky | 257 | |
| 258 | send_call(mq, call); |
||
| 259 | } |
||
| 260 | spinlock_unlock(&irq_conns[mq].lock); |
||
| 261 | } |
||
| 262 | |||
| 1698 | jermar | 263 | /** Notify task that an irq had occurred. |
| 1281 | palkovsky | 264 | * |
| 265 | * We expect interrupts to be disabled |
||
| 266 | */ |
||
| 267 | void ipc_irq_send_notif(int irq) |
||
| 268 | { |
||
| 269 | call_t *call; |
||
| 1595 | palkovsky | 270 | int mq = irq + IPC_IRQ_RESERVED_VIRTUAL; |
| 1281 | palkovsky | 271 | |
| 272 | ASSERT(irq_conns); |
||
| 1595 | palkovsky | 273 | spinlock_lock(&irq_conns[mq].lock); |
| 1281 | palkovsky | 274 | |
| 1595 | palkovsky | 275 | if (irq_conns[mq].box) { |
| 1281 | palkovsky | 276 | call = ipc_call_alloc(FRAME_ATOMIC); |
| 1591 | palkovsky | 277 | if (!call) { |
| 1595 | palkovsky | 278 | spinlock_unlock(&irq_conns[mq].lock); |
| 1591 | palkovsky | 279 | return; |
| 280 | } |
||
| 1281 | palkovsky | 281 | call->flags |= IPC_CALL_NOTIF; |
| 1693 | palkovsky | 282 | /* Put a counter to the message */ |
| 283 | call->private = atomic_preinc(&irq_conns[mq].counter); |
||
| 284 | /* Set up args */ |
||
| 285 | IPC_SET_METHOD(call->data, irq); |
||
| 1281 | palkovsky | 286 | |
| 287 | /* Execute code to handle irq */ |
||
| 1595 | palkovsky | 288 | code_execute(call, irq_conns[mq].code); |
| 289 | |||
| 290 | send_call(mq, call); |
||
| 1281 | palkovsky | 291 | } |
| 292 | |||
| 1595 | palkovsky | 293 | spinlock_unlock(&irq_conns[mq].lock); |
| 1281 | palkovsky | 294 | } |
| 295 | |||
| 296 | |||
| 1595 | palkovsky | 297 | /** Initialize table of interrupt handlers |
| 298 | * |
||
| 299 | * @param irqcount Count of required hardware IRQs to be supported |
||
| 300 | */ |
||
| 1281 | palkovsky | 301 | void ipc_irq_make_table(int irqcount) |
| 302 | { |
||
| 303 | int i; |
||
| 304 | |||
| 1595 | palkovsky | 305 | irqcount += IPC_IRQ_RESERVED_VIRTUAL; |
| 306 | |||
| 1281 | palkovsky | 307 | irq_conns_size = irqcount; |
| 308 | irq_conns = malloc(irqcount * (sizeof(*irq_conns)), 0); |
||
| 309 | for (i=0; i < irqcount; i++) { |
||
| 310 | spinlock_initialize(&irq_conns[i].lock, "irq_ipc_lock"); |
||
| 311 | irq_conns[i].box = NULL; |
||
| 312 | irq_conns[i].code = NULL; |
||
| 313 | } |
||
| 314 | } |
||
| 315 | |||
| 316 | /** Disconnect all irq's notifications |
||
| 317 | * |
||
| 1757 | jermar | 318 | * @todo It may be better to do some linked list, so that |
| 1281 | palkovsky | 319 | * we wouldn't need to go through whole array every cleanup |
| 320 | */ |
||
| 321 | void ipc_irq_cleanup(answerbox_t *box) |
||
| 322 | { |
||
| 323 | int i; |
||
| 324 | ipl_t ipl; |
||
| 325 | |||
| 326 | for (i=0; i < irq_conns_size; i++) { |
||
| 327 | ipl = interrupts_disable(); |
||
| 328 | spinlock_lock(&irq_conns[i].lock); |
||
| 329 | if (irq_conns[i].box == box) |
||
| 330 | irq_conns[i].box = NULL; |
||
| 331 | spinlock_unlock(&irq_conns[i].lock); |
||
| 332 | interrupts_restore(ipl); |
||
| 333 | } |
||
| 334 | } |
||
| 1702 | cejka | 335 | |
| 1757 | jermar | 336 | /** @} |
| 1702 | cejka | 337 | */ |