Rev 2848 | Rev 2851 | Go to most recent revision | Show entire file | Ignore whitespace | Details | Blame | Last modification | View Log | RSS feed
Rev 2848 | Rev 2849 | ||
---|---|---|---|
Line 16... | Line 16... | ||
16 | #include <syscall/copy.h> |
16 | #include <syscall/copy.h> |
17 | #include <udebug/udebug.h> |
17 | #include <udebug/udebug.h> |
18 | #include <udebug/udebug_ipc.h> |
18 | #include <udebug/udebug_ipc.h> |
19 | 19 | ||
20 | /** |
20 | /** |
21 | * Get and lock a phone's callee task. |
21 | * Get a phone's callee task id. |
22 | * |
22 | * |
23 | * This will return a pointer to the task to which the phone |
23 | * This will return the id of the task to which the phone |
24 | * is connected. It will lock the task, making sure it exists. |
24 | * is connected. |
25 | * |
25 | * |
26 | * Interrupts must be already disabled. |
26 | * Interrupts must be already disabled. |
27 | * |
- | |
28 | * (TODO: make sure the udebug-cleanup of the task hasn't |
- | |
29 | * started yet) |
- | |
30 | */ |
27 | */ |
31 | static task_t *get_lock_callee_task(phone_t *phone) |
28 | static task_id_t get_callee_task_id(phone_t *phone) |
32 | { |
29 | { |
33 | answerbox_t *box; |
30 | answerbox_t *box; |
34 | task_t *ta; |
- | |
35 | task_id_t taskid; |
31 | task_id_t taskid; |
36 | 32 | ||
37 | spinlock_lock(&phone->lock); |
33 | spinlock_lock(&phone->lock); |
38 | if (phone->state != IPC_PHONE_CONNECTED) { |
34 | if (phone->state != IPC_PHONE_CONNECTED) { |
39 | spinlock_unlock(&phone->lock); |
35 | spinlock_unlock(&phone->lock); |
Line 41... | Line 37... | ||
41 | } |
37 | } |
42 | 38 | ||
43 | box = phone->callee; |
39 | box = phone->callee; |
44 | 40 | ||
45 | spinlock_lock(&box->lock); |
41 | spinlock_lock(&box->lock); |
46 | ta = box->task; |
- | |
47 | taskid = ta->taskid; |
42 | taskid = box->task->taskid; |
48 | spinlock_unlock(&box->lock); |
43 | spinlock_unlock(&box->lock); |
49 | spinlock_unlock(&phone->lock); |
44 | spinlock_unlock(&phone->lock); |
50 | 45 | ||
- | 46 | return taskid; |
|
- | 47 | } |
|
- | 48 | ||
- | 49 | /** |
|
51 | /* Locking decoupled using taskid */ |
50 | * Get and lock a phone's callee task. |
- | 51 | * |
|
- | 52 | * This will return a pointer to the task to which the phone |
|
- | 53 | * is connected. It will lock the task, making sure it exists. |
|
- | 54 | * |
|
- | 55 | * Interrupts must be already disabled. |
|
- | 56 | * |
|
- | 57 | * (TODO: make sure the udebug-cleanup of the task hasn't |
|
- | 58 | * started yet) |
|
- | 59 | */ |
|
- | 60 | static task_t *get_lock_callee_task(phone_t *phone) |
|
52 | 61 | { |
|
- | 62 | task_t *ta; |
|
- | 63 | task_id_t taskid; |
|
- | 64 | ||
- | 65 | taskid = get_callee_task_id(phone); |
|
- | 66 | ||
53 | spinlock_lock(&tasks_lock); |
67 | spinlock_lock(&tasks_lock); |
54 | ta = task_find_by_id(taskid); |
68 | ta = task_find_by_id(taskid); |
55 | if (ta == NULL) { |
69 | if (ta == NULL) { |
56 | spinlock_unlock(&tasks_lock); |
70 | spinlock_unlock(&tasks_lock); |
57 | return NULL; |
71 | return NULL; |
Line 61... | Line 75... | ||
61 | spinlock_unlock(&tasks_lock); |
75 | spinlock_unlock(&tasks_lock); |
62 | 76 | ||
63 | return ta; |
77 | return ta; |
64 | } |
78 | } |
65 | 79 | ||
66 | /** |
80 | /** |
67 | * Verify that thread t is valid for debugging ops. |
81 | * Prepare a thread for a debugging operation. |
68 | * |
82 | * |
69 | * Verifies that t belongs to task ta and that debugging operations |
83 | * Simply put, return thread t with t->debug_lock held, |
70 | * may be used on it. |
84 | * but only if it verifies all conditions. |
71 | * |
85 | * |
- | 86 | * Specifically, verifies that thread t exists, is a userspace thread, |
|
- | 87 | * belongs to the callee of 'phone'. It also locks t->debug_lock, |
|
- | 88 | * making sure that t->debug_active is true - that the thread is |
|
- | 89 | * in a valid debugging session. |
|
- | 90 | * |
|
- | 91 | * Returns EOK if all went well, or an error code otherwise. |
|
72 | * Thread t's lock must already be held and interrupts must be disabled. |
92 | * Interrupts must be already disabled when calling this function. |
- | 93 | * |
|
- | 94 | * Note: This function sports complicated locking. |
|
73 | */ |
95 | */ |
74 | static int verify_thread(thread_t *t, task_t *ta) |
96 | static int _thread_op_begin(phone_t *phone, thread_t *t) |
75 | { |
97 | { |
- | 98 | int rc; |
|
- | 99 | task_id_t taskid; |
|
- | 100 | int task_match; |
|
- | 101 | DEADLOCK_PROBE_INIT(p_tasklock); |
|
- | 102 | ||
- | 103 | taskid = get_callee_task_id(phone); |
|
- | 104 | ||
76 | /* Verify that 't' exists and belongs to task 'ta' */ |
105 | /* Need to lock down the thread and than it's owner task */ |
- | 106 | grab_locks: |
|
- | 107 | spinlock_lock(&threads_lock); |
|
- | 108 | ||
77 | if (!thread_exists(t) || (t->task != ta)) { |
109 | if (!thread_exists(t)) { |
- | 110 | spinlock_unlock(&threads_lock); |
|
78 | return ENOENT; |
111 | return ENOENT; |
79 | } |
112 | } |
80 | 113 | ||
- | 114 | spinlock_lock(&t->debug_lock); |
|
- | 115 | spinlock_lock(&t->lock); |
|
- | 116 | ||
- | 117 | if (!spinlock_trylock(&t->task->lock)) { |
|
- | 118 | spinlock_unlock(&t->lock); |
|
- | 119 | spinlock_unlock(&t->debug_lock); |
|
- | 120 | DEADLOCK_PROBE(p_tasklock, DEADLOCK_THRESHOLD); |
|
- | 121 | goto grab_locks; /* avoid deadlock */ |
|
- | 122 | } |
|
- | 123 | ||
- | 124 | /* Now verify that it's the callee */ |
|
- | 125 | task_match = (t->task->taskid == taskid); |
|
- | 126 | ||
- | 127 | spinlock_unlock(&t->task->lock); |
|
- | 128 | ||
- | 129 | if (!task_match) { |
|
- | 130 | /* No such thread belonging to callee */ |
|
- | 131 | rc = ENOENT; |
|
- | 132 | goto error_exit; |
|
- | 133 | } |
|
- | 134 | ||
81 | /* Verify that 't' is a userspace thread */ |
135 | /* Verify that 't' is a userspace thread */ |
82 | if ((t->flags & THREAD_FLAG_USPACE) == 0) { |
136 | if ((t->flags & THREAD_FLAG_USPACE) == 0) { |
83 | /* It's not, deny its existence */ |
137 | /* It's not, deny its existence */ |
84 | return ENOENT; |
138 | rc = ENOENT; |
- | 139 | goto error_exit; |
|
85 | } |
140 | } |
86 | 141 | ||
87 | if ((t->debug_active != true) || (t->debug_stop != true)) { |
142 | if ((t->debug_active != true) || (t->debug_stop != true)) { |
88 | /* Not in debugging session or already has GO */ |
143 | /* Not in debugging session or already has GO */ |
89 | return EBUSY; |
144 | rc = ENOENT; |
- | 145 | goto error_exit; |
|
90 | } |
146 | } |
91 | 147 | ||
- | 148 | spinlock_unlock(&threads_lock); |
|
- | 149 | spinlock_unlock(&t->lock); |
|
- | 150 | ||
- | 151 | /* Only t->debug_lock left */ |
|
- | 152 | ||
- | 153 | return EOK; /* All went well */ |
|
- | 154 | ||
- | 155 | ||
- | 156 | /* Executed when a check on the thread fails */ |
|
92 | return EOK; |
157 | error_exit: |
- | 158 | spinlock_unlock(&t->lock); |
|
- | 159 | spinlock_unlock(&t->debug_lock); |
|
- | 160 | spinlock_unlock(&threads_lock); |
|
- | 161 | ||
- | 162 | /* No locks left here */ |
|
- | 163 | return rc; /* Some errors occured */ |
|
- | 164 | } |
|
- | 165 | ||
- | 166 | static void _thread_op_end(thread_t *t) |
|
- | 167 | { |
|
- | 168 | spinlock_unlock(&t->debug_lock); |
|
93 | } |
169 | } |
94 | 170 | ||
95 | static int udebug_rp_begin(call_t *call, phone_t *phone) |
171 | static int udebug_rp_begin(call_t *call, phone_t *phone) |
96 | { |
172 | { |
97 | task_t *ta; |
173 | task_t *ta; |
Line 227... | Line 303... | ||
227 | 303 | ||
228 | 304 | ||
229 | static int udebug_rp_go(call_t *call, phone_t *phone) |
305 | static int udebug_rp_go(call_t *call, phone_t *phone) |
230 | { |
306 | { |
231 | thread_t *t; |
307 | thread_t *t; |
232 | task_t *ta; |
- | |
233 | ipl_t ipl; |
308 | ipl_t ipl; |
234 | int rc; |
309 | int rc; |
235 | 310 | ||
236 | klog_printf("debug_go()"); |
311 | klog_printf("debug_go()"); |
237 | 312 | ||
238 | ipl = interrupts_disable(); |
313 | t = (thread_t *)IPC_GET_ARG2(call->data); |
239 | 314 | ||
240 | ta = get_lock_callee_task(phone); |
- | |
241 | spinlock_unlock(&ta->lock); |
- | |
242 | // TODO: don't lock ta |
- | |
243 | - | ||
244 | t = (thread_t *) IPC_GET_ARG2(call->data); |
- | |
245 | - | ||
246 | spinlock_lock(&threads_lock); |
- | |
247 | if (!thread_exists(t)) { |
- | |
248 | spinlock_unlock(&threads_lock); |
- | |
249 | interrupts_restore(ipl); |
315 | ipl = interrupts_disable(); |
250 | return ENOENT; |
- | |
251 | } |
- | |
252 | - | ||
253 | spinlock_lock(&t->debug_lock); |
- | |
254 | 316 | ||
255 | /* Verify that thread t may be operated on */ |
317 | /* On success, this will lock t->debug_lock */ |
256 | rc = verify_thread(t, ta); |
318 | rc = _thread_op_begin(phone, t); |
257 | if (rc != EOK) { |
319 | if (rc != EOK) { |
258 | spinlock_unlock(&t->debug_lock); |
- | |
259 | spinlock_unlock(&threads_lock); |
- | |
260 | interrupts_restore(ipl); |
320 | interrupts_restore(ipl); |
261 | return rc; |
321 | return rc; |
262 | } |
322 | } |
263 | 323 | ||
264 | /* |
- | |
265 | * Since t->debug_active == true and t->debug_lock is held, |
- | |
266 | * we can safely release threads_lock and t will continue |
- | |
267 | * to exist (and will stay in debug_active state) |
- | |
268 | */ |
- | |
269 | spinlock_unlock(&threads_lock); |
- | |
270 | - | ||
271 | t->debug_go_call = call; |
324 | t->debug_go_call = call; |
272 | t->debug_stop = false; |
325 | t->debug_stop = false; |
273 | 326 | ||
274 | /* |
327 | /* |
275 | * Neither t's lock nor threads_lock may be held during wakeup |
328 | * Neither t's lock nor threads_lock may be held during wakeup |
276 | */ |
329 | */ |
277 | waitq_wakeup(&t->go_wq, WAKEUP_FIRST); |
330 | waitq_wakeup(&t->go_wq, WAKEUP_FIRST); |
278 | 331 | ||
279 | spinlock_unlock(&t->debug_lock); |
332 | _thread_op_end(t); |
280 | interrupts_restore(ipl); |
333 | interrupts_restore(ipl); |
281 | 334 | ||
282 | return 0; /* no backsend */ |
335 | return 0; /* no backsend */ |
283 | } |
336 | } |
284 | 337 | ||
285 | static int udebug_rp_args_read(call_t *call, phone_t *phone) |
338 | static int udebug_rp_args_read(call_t *call, phone_t *phone) |
286 | { |
339 | { |
287 | thread_t *t; |
340 | thread_t *t; |
288 | task_t *ta; |
- | |
289 | void *uspace_buffer; |
341 | void *uspace_buffer; |
290 | int rc; |
342 | int rc; |
291 | ipl_t ipl; |
343 | ipl_t ipl; |
292 | unative_t buffer[6]; |
344 | unative_t buffer[6]; |
293 | 345 | ||
294 | klog_printf("debug_args_read()"); |
346 | klog_printf("debug_args_read()"); |
295 | 347 | ||
296 | ipl = interrupts_disable(); |
- | |
297 | ta = get_lock_callee_task(phone); |
- | |
298 | klog_printf("task %llu", ta->taskid); |
- | |
299 | spinlock_unlock(&ta->lock); |
- | |
300 | - | ||
301 | t = (thread_t *) IPC_GET_ARG2(call->data); |
348 | t = (thread_t *)IPC_GET_ARG2(call->data); |
302 | - | ||
303 | spinlock_lock(&threads_lock); |
- | |
304 | - | ||
305 | if (!thread_exists(t)) { |
- | |
306 | spinlock_unlock(&threads_lock); |
- | |
307 | interrupts_restore(ipl); |
- | |
308 | return ENOENT; |
- | |
309 | } |
- | |
310 | 349 | ||
311 | spinlock_lock(&t->debug_lock); |
350 | ipl = interrupts_disable(); |
312 | 351 | ||
313 | /* Verify that thread t may be operated on */ |
352 | /* On success, this will lock t->debug_lock */ |
314 | rc = verify_thread(t, ta); |
353 | rc = _thread_op_begin(phone, t); |
315 | if (rc != EOK) { |
354 | if (rc != EOK) { |
316 | spinlock_unlock(&t->debug_lock); |
- | |
317 | spinlock_unlock(&threads_lock); |
- | |
318 | interrupts_restore(ipl); |
355 | interrupts_restore(ipl); |
319 | return rc; |
356 | return rc; |
320 | } |
357 | } |
321 | 358 | ||
322 | /* |
- | |
323 | * We can now safely release threads_lock as debug_active == true |
- | |
324 | * and t->debug_lock is held. |
- | |
325 | */ |
- | |
326 | spinlock_unlock(&threads_lock); |
- | |
327 | - | ||
328 | //FIXME: additionally we need to verify that we are inside a syscall |
359 | //FIXME: additionally we need to verify that we are inside a syscall |
329 | 360 | ||
330 | /* Copy to a local buffer before releasing the lock */ |
361 | /* Copy to a local buffer before releasing the lock */ |
331 | memcpy(buffer, t->syscall_args, 6 * sizeof(unative_t)); |
362 | memcpy(buffer, t->syscall_args, 6 * sizeof(unative_t)); |
332 | 363 | ||
333 | spinlock_unlock(&t->debug_lock); |
364 | _thread_op_end(t); |
334 | interrupts_restore(ipl); |
365 | interrupts_restore(ipl); |
335 | 366 | ||
336 | /* Now copy to userspace */ |
367 | /* Now copy to userspace */ |
337 | 368 | ||
338 | uspace_buffer = (void *)IPC_GET_ARG3(call->data); |
369 | uspace_buffer = (void *)IPC_GET_ARG3(call->data); |
339 | 370 | ||
340 | rc = copy_to_uspace(uspace_buffer, buffer, 6 * sizeof(unative_t)); |
371 | rc = copy_to_uspace(uspace_buffer, buffer, 6 * sizeof(unative_t)); |
341 | if (rc != 0) { |
372 | if (rc != 0) { |
342 | spinlock_unlock(&ta->lock); |
- | |
343 | klog_printf("debug_args_read() - copy failed"); |
373 | klog_printf("debug_args_read() - copy failed"); |
344 | return rc; |
374 | return rc; |
345 | } |
375 | } |
346 | 376 | ||
347 | klog_printf("debug_args_read() done"); |
377 | klog_printf("debug_args_read() done"); |
Line 349... | Line 379... | ||
349 | } |
379 | } |
350 | 380 | ||
351 | static int udebug_rp_regs_read(call_t *call, phone_t *phone) |
381 | static int udebug_rp_regs_read(call_t *call, phone_t *phone) |
352 | { |
382 | { |
353 | thread_t *t; |
383 | thread_t *t; |
354 | task_t *ta; |
- | |
355 | void *uspace_buffer; |
384 | void *uspace_buffer; |
356 | unative_t to_copy; |
385 | unative_t to_copy; |
357 | int rc; |
386 | int rc; |
358 | istate_t *state; |
387 | istate_t *state; |
359 | istate_t state_copy; |
388 | istate_t state_copy; |
360 | ipl_t ipl; |
389 | ipl_t ipl; |
361 | 390 | ||
362 | klog_printf("debug_regs_read()"); |
391 | klog_printf("debug_regs_read()"); |
363 | 392 | ||
364 | ta = get_lock_callee_task(phone); |
- | |
365 | spinlock_unlock(&ta->lock); |
- | |
366 | //FIXME: don't lock ta |
- | |
367 | - | ||
368 | ipl = interrupts_disable(); |
- | |
369 | spinlock_lock(&threads_lock); |
- | |
370 | - | ||
371 | t = (thread_t *) IPC_GET_ARG2(call->data); |
393 | t = (thread_t *) IPC_GET_ARG2(call->data); |
372 | 394 | ||
373 | if (!thread_exists(t)) { |
- | |
374 | spinlock_unlock(&threads_lock); |
- | |
375 | interrupts_restore(ipl); |
395 | ipl = interrupts_disable(); |
376 | return ENOENT; |
- | |
377 | } |
- | |
378 | - | ||
379 | spinlock_lock(&t->debug_lock); |
- | |
380 | 396 | ||
381 | /* Verify that thread t may be operated on */ |
397 | /* On success, this will lock t->debug_lock */ |
382 | rc = verify_thread(t, ta); |
398 | rc = _thread_op_begin(phone, t); |
383 | if (rc != EOK) { |
399 | if (rc != EOK) { |
384 | spinlock_unlock(&t->debug_lock); |
- | |
385 | spinlock_unlock(&threads_lock); |
- | |
386 | interrupts_restore(ipl); |
400 | interrupts_restore(ipl); |
387 | return rc; |
401 | return rc; |
388 | } |
402 | } |
389 | 403 | ||
390 | /* |
- | |
391 | * We can now safely release threads_lock as debug_active == true |
- | |
392 | * and t->debug_lock is held. |
- | |
393 | */ |
- | |
394 | spinlock_unlock(&threads_lock); |
- | |
395 | - | ||
396 | state = t->uspace_state; |
404 | state = t->uspace_state; |
397 | if (state == NULL) { |
405 | if (state == NULL) { |
398 | spinlock_unlock(&threads_lock); |
406 | _thread_op_end(t); |
399 | interrupts_restore(ipl); |
407 | interrupts_restore(ipl); |
400 | klog_printf("debug_regs_read() - istate not available"); |
408 | klog_printf("debug_regs_read() - istate not available"); |
401 | return EBUSY; |
409 | return EBUSY; |
402 | } |
410 | } |
403 | 411 | ||
404 | /* Copy to a local buffer so that we can release the lock */ |
412 | /* Copy to a local buffer so that we can release the lock */ |
405 | memcpy(&state_copy, state, sizeof(state_copy)); |
413 | memcpy(&state_copy, state, sizeof(state_copy)); |
406 | spinlock_unlock(&t->debug_lock); |
414 | _thread_op_end(t); |
407 | interrupts_restore(ipl); |
415 | interrupts_restore(ipl); |
408 | 416 | ||
409 | uspace_buffer = (void *)IPC_GET_ARG3(call->data); |
417 | uspace_buffer = (void *)IPC_GET_ARG3(call->data); |
410 | to_copy = IPC_GET_ARG4(call->data); |
418 | to_copy = IPC_GET_ARG4(call->data); |
411 | if (to_copy > sizeof(istate_t)) to_copy = sizeof(istate_t); |
419 | if (to_copy > sizeof(istate_t)) to_copy = sizeof(istate_t); |
Line 421... | Line 429... | ||
421 | 429 | ||
422 | klog_printf("debug_regs_read() done"); |
430 | klog_printf("debug_regs_read() done"); |
423 | return 1; /* actually need becksend with retval 0 */ |
431 | return 1; /* actually need becksend with retval 0 */ |
424 | } |
432 | } |
425 | 433 | ||
- | 434 | ||
- | 435 | ||
426 | static int udebug_rp_regs_write(call_t *call, phone_t *phone) |
436 | static int udebug_rp_regs_write(call_t *call, phone_t *phone) |
427 | { |
437 | { |
428 | thread_t *t; |
438 | thread_t *t; |
429 | task_t *ta; |
- | |
430 | void *uspace_data; |
439 | void *uspace_data; |
431 | unative_t to_copy; |
440 | unative_t to_copy; |
432 | int rc; |
441 | int rc; |
433 | istate_t *state; |
442 | istate_t *state; |
434 | istate_t data_copy; |
443 | istate_t data_copy; |
Line 446... | Line 455... | ||
446 | if (rc != 0) { |
455 | if (rc != 0) { |
447 | klog_printf("debug_regs_write() - copy failed"); |
456 | klog_printf("debug_regs_write() - copy failed"); |
448 | return rc; |
457 | return rc; |
449 | } |
458 | } |
450 | 459 | ||
451 | ta = get_lock_callee_task(phone); |
- | |
452 | spinlock_unlock(&ta->lock); |
- | |
453 | //FIXME: don't lock ta |
- | |
454 | - | ||
455 | /* Now try to change the thread's uspace_state */ |
460 | /* Now try to change the thread's uspace_state */ |
456 | 461 | ||
457 | ipl = interrupts_disable(); |
462 | ipl = interrupts_disable(); |
458 | spinlock_lock(&threads_lock); |
- | |
459 | - | ||
460 | t = (thread_t *) IPC_GET_ARG2(call->data); |
463 | t = (thread_t *) IPC_GET_ARG2(call->data); |
461 | 464 | ||
462 | if (!thread_exists(t)) { |
- | |
463 | spinlock_unlock(&threads_lock); |
- | |
464 | interrupts_restore(ipl); |
- | |
465 | return ENOENT; |
- | |
466 | } |
- | |
467 | - | ||
468 | spinlock_lock(&t->debug_lock); |
465 | /* On success, this will lock t->debug_lock */ |
469 | - | ||
470 | /* Verify that thread t may be operated on */ |
- | |
471 | rc = verify_thread(t, ta); |
466 | rc = _thread_op_begin(phone, t); |
472 | if (rc != EOK) { |
467 | if (rc != EOK) { |
473 | spinlock_unlock(&t->debug_lock); |
- | |
474 | spinlock_unlock(&threads_lock); |
- | |
475 | interrupts_restore(ipl); |
468 | interrupts_restore(ipl); |
476 | return rc; |
469 | return rc; |
477 | } |
470 | } |
478 | 471 | ||
479 | state = t->uspace_state; |
472 | state = t->uspace_state; |
480 | if (state == NULL) { |
473 | if (state == NULL) { |
481 | spinlock_unlock(&t->debug_lock); |
474 | _thread_op_end(t); |
482 | interrupts_restore(ipl); |
475 | interrupts_restore(ipl); |
483 | klog_printf("debug_regs_write() - istate not available"); |
476 | klog_printf("debug_regs_write() - istate not available"); |
484 | return EBUSY; |
477 | return EBUSY; |
485 | } |
478 | } |
486 | 479 | ||
487 | memcpy(t->uspace_state, &data_copy, sizeof(t->uspace_state)); |
480 | memcpy(t->uspace_state, &data_copy, sizeof(t->uspace_state)); |
488 | 481 | ||
489 | spinlock_unlock(&t->debug_lock); |
482 | _thread_op_end(t); |
490 | interrupts_restore(ipl); |
483 | interrupts_restore(ipl); |
491 | 484 | ||
492 | /* Set answer values */ |
485 | /* Set answer values */ |
493 | 486 | ||
494 | IPC_SET_ARG1(call->data, to_copy); |
487 | IPC_SET_ARG1(call->data, to_copy); |