Rev 2090 | Rev 2093 | Go to most recent revision | Only display areas with differences | Regard whitespace | Details | Blame | Last modification | View Log | RSS feed
Rev 2090 | Rev 2092 | ||
---|---|---|---|
1 | /* |
1 | /* |
2 | * Copyright (c) 2001-2004 Jakub Jermar |
2 | * Copyright (c) 2001-2004 Jakub Jermar |
3 | * All rights reserved. |
3 | * All rights reserved. |
4 | * |
4 | * |
5 | * Redistribution and use in source and binary forms, with or without |
5 | * Redistribution and use in source and binary forms, with or without |
6 | * modification, are permitted provided that the following conditions |
6 | * modification, are permitted provided that the following conditions |
7 | * are met: |
7 | * are met: |
8 | * |
8 | * |
9 | * - Redistributions of source code must retain the above copyright |
9 | * - Redistributions of source code must retain the above copyright |
10 | * notice, this list of conditions and the following disclaimer. |
10 | * notice, this list of conditions and the following disclaimer. |
11 | * - Redistributions in binary form must reproduce the above copyright |
11 | * - Redistributions in binary form must reproduce the above copyright |
12 | * notice, this list of conditions and the following disclaimer in the |
12 | * notice, this list of conditions and the following disclaimer in the |
13 | * documentation and/or other materials provided with the distribution. |
13 | * documentation and/or other materials provided with the distribution. |
14 | * - The name of the author may not be used to endorse or promote products |
14 | * - The name of the author may not be used to endorse or promote products |
15 | * derived from this software without specific prior written permission. |
15 | * derived from this software without specific prior written permission. |
16 | * |
16 | * |
17 | * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR |
17 | * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR |
18 | * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES |
18 | * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES |
19 | * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. |
19 | * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. |
20 | * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, |
20 | * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, |
21 | * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT |
21 | * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT |
22 | * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
22 | * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
23 | * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
23 | * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
24 | * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
24 | * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
25 | * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF |
25 | * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF |
26 | * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
26 | * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
27 | */ |
27 | */ |
28 | 28 | ||
29 | /** @addtogroup genericproc |
29 | /** @addtogroup genericproc |
30 | * @{ |
30 | * @{ |
31 | */ |
31 | */ |
32 | /** @file |
32 | /** @file |
33 | */ |
33 | */ |
34 | 34 | ||
35 | #ifndef KERN_TASK_H_ |
35 | #ifndef KERN_TASK_H_ |
36 | #define KERN_TASK_H_ |
36 | #define KERN_TASK_H_ |
37 | 37 | ||
- | 38 | #include <cpu.h> |
|
38 | #include <synch/spinlock.h> |
39 | #include <synch/spinlock.h> |
39 | #include <synch/mutex.h> |
40 | #include <synch/mutex.h> |
40 | #include <synch/rwlock.h> |
41 | #include <synch/rwlock.h> |
41 | #include <synch/futex.h> |
42 | #include <synch/futex.h> |
42 | #include <adt/btree.h> |
43 | #include <adt/btree.h> |
43 | #include <adt/list.h> |
44 | #include <adt/list.h> |
44 | #include <security/cap.h> |
45 | #include <security/cap.h> |
45 | #include <arch/proc/task.h> |
46 | #include <arch/proc/task.h> |
46 | #include <arch/proc/thread.h> |
47 | #include <arch/proc/thread.h> |
47 | #include <arch/context.h> |
48 | #include <arch/context.h> |
48 | #include <arch/fpu_context.h> |
49 | #include <arch/fpu_context.h> |
49 | #include <arch/cpu.h> |
50 | #include <arch/cpu.h> |
50 | #include <mm/tlb.h> |
51 | #include <mm/tlb.h> |
51 | #include <proc/scheduler.h> |
52 | #include <proc/scheduler.h> |
52 | 53 | ||
53 | #define IPC_MAX_PHONES 16 |
54 | #define IPC_MAX_PHONES 16 |
54 | #define THREAD_NAME_BUFLEN 20 |
55 | #define THREAD_NAME_BUFLEN 20 |
55 | 56 | ||
56 | struct answerbox; |
57 | struct answerbox; |
57 | struct task; |
58 | struct task; |
58 | struct thread; |
59 | struct thread; |
59 | 60 | ||
60 | typedef enum { |
61 | typedef enum { |
61 | IPC_PHONE_FREE = 0, /**< Phone is free and can be allocated */ |
62 | IPC_PHONE_FREE = 0, /**< Phone is free and can be allocated */ |
62 | IPC_PHONE_CONNECTING, /**< Phone is connecting somewhere */ |
63 | IPC_PHONE_CONNECTING, /**< Phone is connecting somewhere */ |
63 | IPC_PHONE_CONNECTED, /**< Phone is connected */ |
64 | IPC_PHONE_CONNECTED, /**< Phone is connected */ |
64 | IPC_PHONE_HUNGUP, /**< Phone is hung up, waiting for answers to come */ |
65 | IPC_PHONE_HUNGUP, /**< Phone is hung up, waiting for answers to come */ |
65 | IPC_PHONE_SLAMMED /**< Phone was hungup from server */ |
66 | IPC_PHONE_SLAMMED /**< Phone was hungup from server */ |
66 | } ipc_phone_state_t; |
67 | } ipc_phone_state_t; |
67 | 68 | ||
68 | /** Structure identifying phone (in TASK structure) */ |
69 | /** Structure identifying phone (in TASK structure) */ |
69 | typedef struct { |
70 | typedef struct { |
70 | SPINLOCK_DECLARE(lock); |
71 | SPINLOCK_DECLARE(lock); |
71 | link_t link; |
72 | link_t link; |
72 | struct answerbox *callee; |
73 | struct answerbox *callee; |
73 | ipc_phone_state_t state; |
74 | ipc_phone_state_t state; |
74 | atomic_t active_calls; |
75 | atomic_t active_calls; |
75 | } phone_t; |
76 | } phone_t; |
76 | 77 | ||
77 | typedef struct answerbox { |
78 | typedef struct answerbox { |
78 | SPINLOCK_DECLARE(lock); |
79 | SPINLOCK_DECLARE(lock); |
79 | 80 | ||
80 | struct task *task; |
81 | struct task *task; |
81 | 82 | ||
82 | waitq_t wq; |
83 | waitq_t wq; |
83 | 84 | ||
84 | link_t connected_phones; /**< Phones connected to this answerbox */ |
85 | link_t connected_phones; /**< Phones connected to this answerbox */ |
85 | link_t calls; /**< Received calls */ |
86 | link_t calls; /**< Received calls */ |
86 | link_t dispatched_calls; /* Should be hash table in the future */ |
87 | link_t dispatched_calls; /* Should be hash table in the future */ |
87 | 88 | ||
88 | link_t answers; /**< Answered calls */ |
89 | link_t answers; /**< Answered calls */ |
89 | 90 | ||
90 | SPINLOCK_DECLARE(irq_lock); |
91 | SPINLOCK_DECLARE(irq_lock); |
91 | link_t irq_notifs; /**< Notifications from IRQ handlers */ |
92 | link_t irq_notifs; /**< Notifications from IRQ handlers */ |
92 | link_t irq_head; /**< IRQs with notifications to this answerbox. */ |
93 | link_t irq_head; /**< IRQs with notifications to this answerbox. */ |
93 | } answerbox_t; |
94 | } answerbox_t; |
94 | 95 | ||
95 | /** Task structure. */ |
96 | /** Task structure. */ |
96 | typedef struct task { |
97 | typedef struct task { |
97 | /** Task lock. |
98 | /** Task lock. |
98 | * |
99 | * |
99 | * Must be acquired before threads_lock and thread lock of any of its threads. |
100 | * Must be acquired before threads_lock and thread lock of any of its threads. |
100 | */ |
101 | */ |
101 | SPINLOCK_DECLARE(lock); |
102 | SPINLOCK_DECLARE(lock); |
102 | 103 | ||
103 | char *name; |
104 | char *name; |
104 | struct thread *main_thread; /**< Pointer to the main thread. */ |
105 | struct thread *main_thread; /**< Pointer to the main thread. */ |
105 | link_t th_head; /**< List of threads contained in this task. */ |
106 | link_t th_head; /**< List of threads contained in this task. */ |
106 | as_t *as; /**< Address space. */ |
107 | as_t *as; /**< Address space. */ |
107 | task_id_t taskid; /**< Unique identity of task */ |
108 | task_id_t taskid; /**< Unique identity of task */ |
108 | context_id_t context; /**< Task security context */ |
109 | context_id_t context; /**< Task security context */ |
109 | 110 | ||
110 | /** If this is true, new threads can become part of the task. */ |
111 | /** If this is true, new threads can become part of the task. */ |
111 | bool accept_new_threads; |
112 | bool accept_new_threads; |
112 | 113 | ||
113 | count_t refcount; /**< Number of references (i.e. threads). */ |
114 | count_t refcount; /**< Number of references (i.e. threads). */ |
114 | 115 | ||
115 | cap_t capabilities; /**< Task capabilities. */ |
116 | cap_t capabilities; /**< Task capabilities. */ |
116 | 117 | ||
117 | /* IPC stuff */ |
118 | /* IPC stuff */ |
118 | answerbox_t answerbox; /**< Communication endpoint */ |
119 | answerbox_t answerbox; /**< Communication endpoint */ |
119 | phone_t phones[IPC_MAX_PHONES]; |
120 | phone_t phones[IPC_MAX_PHONES]; |
120 | atomic_t active_calls; /**< Active asynchronous messages. |
121 | atomic_t active_calls; /**< Active asynchronous messages. |
121 | * It is used for limiting uspace to |
122 | * It is used for limiting uspace to |
122 | * certain extent. */ |
123 | * certain extent. */ |
123 | 124 | ||
124 | task_arch_t arch; /**< Architecture specific task data. */ |
125 | task_arch_t arch; /**< Architecture specific task data. */ |
125 | 126 | ||
126 | /** |
127 | /** |
127 | * Serializes access to the B+tree of task's futexes. This mutex is |
128 | * Serializes access to the B+tree of task's futexes. This mutex is |
128 | * independent on the task spinlock. |
129 | * independent on the task spinlock. |
129 | */ |
130 | */ |
130 | mutex_t futexes_lock; |
131 | mutex_t futexes_lock; |
131 | btree_t futexes; /**< B+tree of futexes referenced by this task. */ |
132 | btree_t futexes; /**< B+tree of futexes referenced by this task. */ |
132 | 133 | ||
133 | uint64_t cycles; /**< Accumulated accounting. */ |
134 | uint64_t cycles; /**< Accumulated accounting. */ |
134 | } task_t; |
135 | } task_t; |
135 | 136 | ||
136 | /** CPU structure. |
- | |
137 | * |
- | |
138 | * There is one structure like this for every processor. |
- | |
139 | */ |
- | |
140 | typedef struct { |
- | |
141 | SPINLOCK_DECLARE(lock); |
- | |
142 | - | ||
143 | tlb_shootdown_msg_t tlb_messages[TLB_MESSAGE_QUEUE_LEN]; |
- | |
144 | count_t tlb_messages_count; |
- | |
145 | - | ||
146 | context_t saved_context; |
- | |
147 | - | ||
148 | atomic_t nrdy; |
- | |
149 | runq_t rq[RQ_COUNT]; |
- | |
150 | volatile count_t needs_relink; |
- | |
151 | - | ||
152 | SPINLOCK_DECLARE(timeoutlock); |
- | |
153 | link_t timeout_active_head; |
- | |
154 | - | ||
155 | count_t missed_clock_ticks; /**< When system clock loses a tick, it is recorded here |
- | |
156 | so that clock() can react. This variable is |
- | |
157 | CPU-local and can be only accessed when interrupts |
- | |
158 | are disabled. */ |
- | |
159 | - | ||
160 | /** |
- | |
161 | * Processor ID assigned by kernel. |
- | |
162 | */ |
- | |
163 | unsigned int id; |
- | |
164 | - | ||
165 | int active; |
- | |
166 | int tlb_active; |
- | |
167 | - | ||
168 | uint16_t frequency_mhz; |
- | |
169 | uint32_t delay_loop_const; |
- | |
170 | - | ||
171 | cpu_arch_t arch; |
- | |
172 | - | ||
173 | struct thread *fpu_owner; |
- | |
174 | - | ||
175 | /** |
- | |
176 | * Stack used by scheduler when there is no running thread. |
- | |
177 | */ |
- | |
178 | uint8_t *stack; |
- | |
179 | } cpu_t; |
- | |
180 | - | ||
181 | typedef void (* timeout_handler_t)(void *arg); |
137 | typedef void (* timeout_handler_t)(void *arg); |
182 | 138 | ||
183 | typedef struct { |
139 | typedef struct { |
184 | SPINLOCK_DECLARE(lock); |
140 | SPINLOCK_DECLARE(lock); |
185 | 141 | ||
186 | link_t link; /**< Link to the list of active timeouts on THE->cpu */ |
142 | link_t link; /**< Link to the list of active timeouts on THE->cpu */ |
187 | 143 | ||
188 | uint64_t ticks; /**< Timeout will be activated in this amount of clock() ticks. */ |
144 | uint64_t ticks; /**< Timeout will be activated in this amount of clock() ticks. */ |
189 | 145 | ||
190 | timeout_handler_t handler; /**< Function that will be called on timeout activation. */ |
146 | timeout_handler_t handler; /**< Function that will be called on timeout activation. */ |
191 | void *arg; /**< Argument to be passed to handler() function. */ |
147 | void *arg; /**< Argument to be passed to handler() function. */ |
192 | 148 | ||
193 | cpu_t *cpu; /**< On which processor is this timeout registered. */ |
149 | cpu_t *cpu; /**< On which processor is this timeout registered. */ |
194 | } timeout_t; |
150 | } timeout_t; |
195 | 151 | ||
196 | /** Thread states. */ |
152 | /** Thread states. */ |
197 | typedef enum { |
153 | typedef enum { |
198 | Invalid, /**< It is an error, if thread is found in this state. */ |
154 | Invalid, /**< It is an error, if thread is found in this state. */ |
199 | Running, /**< State of a thread that is currently executing on some CPU. */ |
155 | Running, /**< State of a thread that is currently executing on some CPU. */ |
200 | Sleeping, /**< Thread in this state is waiting for an event. */ |
156 | Sleeping, /**< Thread in this state is waiting for an event. */ |
201 | Ready, /**< State of threads in a run queue. */ |
157 | Ready, /**< State of threads in a run queue. */ |
202 | Entering, /**< Threads are in this state before they are first readied. */ |
158 | Entering, /**< Threads are in this state before they are first readied. */ |
203 | Exiting, /**< After a thread calls thread_exit(), it is put into Exiting state. */ |
159 | Exiting, /**< After a thread calls thread_exit(), it is put into Exiting state. */ |
204 | Undead /**< Threads that were not detached but exited are in the Undead state. */ |
160 | Undead /**< Threads that were not detached but exited are in the Undead state. */ |
205 | } state_t; |
161 | } state_t; |
206 | 162 | ||
207 | /** Join types. */ |
163 | /** Join types. */ |
208 | typedef enum { |
164 | typedef enum { |
209 | None, |
165 | None, |
210 | TaskClnp, /**< The thread will be joined by ktaskclnp thread. */ |
166 | TaskClnp, /**< The thread will be joined by ktaskclnp thread. */ |
211 | TaskGC /**< The thread will be joined by ktaskgc thread. */ |
167 | TaskGC /**< The thread will be joined by ktaskgc thread. */ |
212 | } thread_join_type_t; |
168 | } thread_join_type_t; |
213 | 169 | ||
214 | /** Thread structure. There is one per thread. */ |
170 | /** Thread structure. There is one per thread. */ |
215 | typedef struct thread { |
171 | typedef struct thread { |
216 | link_t rq_link; /**< Run queue link. */ |
172 | link_t rq_link; /**< Run queue link. */ |
217 | link_t wq_link; /**< Wait queue link. */ |
173 | link_t wq_link; /**< Wait queue link. */ |
218 | link_t th_link; /**< Links to threads within containing task. */ |
174 | link_t th_link; /**< Links to threads within containing task. */ |
219 | 175 | ||
220 | /** Lock protecting thread structure. |
176 | /** Lock protecting thread structure. |
221 | * |
177 | * |
222 | * Protects the whole thread structure except list links above. |
178 | * Protects the whole thread structure except list links above. |
223 | */ |
179 | */ |
224 | SPINLOCK_DECLARE(lock); |
180 | SPINLOCK_DECLARE(lock); |
225 | 181 | ||
226 | char name[THREAD_NAME_BUFLEN]; |
182 | char name[THREAD_NAME_BUFLEN]; |
227 | 183 | ||
228 | void (* thread_code)(void *); /**< Function implementing the thread. */ |
184 | void (* thread_code)(void *); /**< Function implementing the thread. */ |
229 | void *thread_arg; /**< Argument passed to thread_code() function. */ |
185 | void *thread_arg; /**< Argument passed to thread_code() function. */ |
230 | 186 | ||
231 | /** From here, the stored context is restored when the thread is scheduled. */ |
187 | /** From here, the stored context is restored when the thread is scheduled. */ |
232 | context_t saved_context; |
188 | context_t saved_context; |
233 | /** From here, the stored timeout context is restored when sleep times out. */ |
189 | /** From here, the stored timeout context is restored when sleep times out. */ |
234 | context_t sleep_timeout_context; |
190 | context_t sleep_timeout_context; |
235 | /** From here, the stored interruption context is restored when sleep is interrupted. */ |
191 | /** From here, the stored interruption context is restored when sleep is interrupted. */ |
236 | context_t sleep_interruption_context; |
192 | context_t sleep_interruption_context; |
237 | 193 | ||
238 | bool sleep_interruptible; /**< If true, the thread can be interrupted from sleep. */ |
194 | bool sleep_interruptible; /**< If true, the thread can be interrupted from sleep. */ |
239 | waitq_t *sleep_queue; /**< Wait queue in which this thread sleeps. */ |
195 | waitq_t *sleep_queue; /**< Wait queue in which this thread sleeps. */ |
240 | timeout_t sleep_timeout; /**< Timeout used for timeoutable sleeping. */ |
196 | timeout_t sleep_timeout; /**< Timeout used for timeoutable sleeping. */ |
241 | volatile int timeout_pending; /**< Flag signalling sleep timeout in progress. */ |
197 | volatile int timeout_pending; /**< Flag signalling sleep timeout in progress. */ |
242 | 198 | ||
243 | /** True if this thread is executing copy_from_uspace(). False otherwise. */ |
199 | /** True if this thread is executing copy_from_uspace(). False otherwise. */ |
244 | bool in_copy_from_uspace; |
200 | bool in_copy_from_uspace; |
245 | /** True if this thread is executing copy_to_uspace(). False otherwise. */ |
201 | /** True if this thread is executing copy_to_uspace(). False otherwise. */ |
246 | bool in_copy_to_uspace; |
202 | bool in_copy_to_uspace; |
247 | 203 | ||
248 | /** |
204 | /** |
249 | * If true, the thread will not go to sleep at all and will |
205 | * If true, the thread will not go to sleep at all and will |
250 | * call thread_exit() before returning to userspace. |
206 | * call thread_exit() before returning to userspace. |
251 | */ |
207 | */ |
252 | bool interrupted; |
208 | bool interrupted; |
253 | 209 | ||
254 | thread_join_type_t join_type; /**< Who joinins the thread. */ |
210 | thread_join_type_t join_type; /**< Who joinins the thread. */ |
255 | bool detached; /**< If true, thread_join_timeout() cannot be used on this thread. */ |
211 | bool detached; /**< If true, thread_join_timeout() cannot be used on this thread. */ |
256 | waitq_t join_wq; /**< Waitq for thread_join_timeout(). */ |
212 | waitq_t join_wq; /**< Waitq for thread_join_timeout(). */ |
257 | 213 | ||
258 | fpu_context_t *saved_fpu_context; |
214 | fpu_context_t *saved_fpu_context; |
259 | int fpu_context_exists; |
215 | int fpu_context_exists; |
260 | 216 | ||
261 | /* |
217 | /* |
262 | * Defined only if thread doesn't run. |
218 | * Defined only if thread doesn't run. |
263 | * It means that fpu context is in CPU that last time executes this thread. |
219 | * It means that fpu context is in CPU that last time executes this thread. |
264 | * This disables migration. |
220 | * This disables migration. |
265 | */ |
221 | */ |
266 | int fpu_context_engaged; |
222 | int fpu_context_engaged; |
267 | 223 | ||
268 | rwlock_type_t rwlock_holder_type; |
224 | rwlock_type_t rwlock_holder_type; |
269 | 225 | ||
270 | void (* call_me)(void *); /**< Funtion to be called in scheduler before the thread is put asleep. */ |
226 | void (* call_me)(void *); /**< Funtion to be called in scheduler before the thread is put asleep. */ |
271 | void *call_me_with; /**< Argument passed to call_me(). */ |
227 | void *call_me_with; /**< Argument passed to call_me(). */ |
272 | 228 | ||
273 | state_t state; /**< Thread's state. */ |
229 | state_t state; /**< Thread's state. */ |
274 | int flags; /**< Thread's flags. */ |
230 | int flags; /**< Thread's flags. */ |
275 | 231 | ||
276 | cpu_t *cpu; /**< Thread's CPU. */ |
232 | cpu_t *cpu; /**< Thread's CPU. */ |
277 | task_t *task; /**< Containing task. */ |
233 | task_t *task; /**< Containing task. */ |
278 | 234 | ||
279 | uint64_t ticks; /**< Ticks before preemption. */ |
235 | uint64_t ticks; /**< Ticks before preemption. */ |
280 | 236 | ||
281 | uint64_t cycles; /**< Thread accounting. */ |
237 | uint64_t cycles; /**< Thread accounting. */ |
282 | uint64_t last_cycle; /**< Last sampled cycle. */ |
238 | uint64_t last_cycle; /**< Last sampled cycle. */ |
283 | bool uncounted; /**< Thread doesn't affect accumulated accounting. */ |
239 | bool uncounted; /**< Thread doesn't affect accumulated accounting. */ |
284 | 240 | ||
285 | int priority; /**< Thread's priority. Implemented as index to CPU->rq */ |
241 | int priority; /**< Thread's priority. Implemented as index to CPU->rq */ |
286 | uint32_t tid; /**< Thread ID. */ |
242 | uint32_t tid; /**< Thread ID. */ |
287 | 243 | ||
288 | thread_arch_t arch; /**< Architecture-specific data. */ |
244 | thread_arch_t arch; /**< Architecture-specific data. */ |
289 | 245 | ||
290 | uint8_t *kstack; /**< Thread's kernel stack. */ |
246 | uint8_t *kstack; /**< Thread's kernel stack. */ |
291 | } thread_t; |
247 | } thread_t; |
292 | 248 | ||
293 | extern spinlock_t tasks_lock; |
249 | extern spinlock_t tasks_lock; |
294 | extern btree_t tasks_btree; |
250 | extern btree_t tasks_btree; |
295 | 251 | ||
296 | extern void task_init(void); |
252 | extern void task_init(void); |
297 | extern task_t *task_create(as_t *as, char *name); |
253 | extern task_t *task_create(as_t *as, char *name); |
298 | extern void task_destroy(task_t *t); |
254 | extern void task_destroy(task_t *t); |
299 | extern task_t *task_run_program(void *program_addr, char *name); |
255 | extern task_t *task_run_program(void *program_addr, char *name); |
300 | extern task_t *task_find_by_id(task_id_t id); |
256 | extern task_t *task_find_by_id(task_id_t id); |
301 | extern int task_kill(task_id_t id); |
257 | extern int task_kill(task_id_t id); |
302 | extern uint64_t task_get_accounting(task_t *t); |
258 | extern uint64_t task_get_accounting(task_t *t); |
303 | 259 | ||
304 | extern void cap_set(task_t *t, cap_t caps); |
260 | extern void cap_set(task_t *t, cap_t caps); |
305 | extern cap_t cap_get(task_t *t); |
261 | extern cap_t cap_get(task_t *t); |
306 | 262 | ||
307 | 263 | ||
308 | #ifndef task_create_arch |
264 | #ifndef task_create_arch |
309 | extern void task_create_arch(task_t *t); |
265 | extern void task_create_arch(task_t *t); |
310 | #endif |
266 | #endif |
311 | 267 | ||
312 | #ifndef task_destroy_arch |
268 | #ifndef task_destroy_arch |
313 | extern void task_destroy_arch(task_t *t); |
269 | extern void task_destroy_arch(task_t *t); |
314 | #endif |
270 | #endif |
315 | 271 | ||
316 | extern unative_t sys_task_get_id(task_id_t *uspace_task_id); |
272 | extern unative_t sys_task_get_id(task_id_t *uspace_task_id); |
317 | 273 | ||
318 | #endif |
274 | #endif |
319 | 275 | ||
320 | /** @} |
276 | /** @} |
321 | */ |
277 | */ |
322 | 278 |