Subversion Repositories HelenOS

Rev

Rev 2071 | Rev 2090 | Go to most recent revision | Details | Compare with Previous | Last modification | View Log | RSS feed

Rev Author Line No. Line
2089 decky 1
/*3D
2071 jermar 2
 * Copyright (c) 2001-2004 Jakub Jermar
1 jermar 3
 * All rights reserved.
4
 *
5
 * Redistribution and use in source and binary forms, with or without
6
 * modification, are permitted provided that the following conditions
7
 * are met:
8
 *
9
 * - Redistributions of source code must retain the above copyright
10
 *   notice, this list of conditions and the following disclaimer.
11
 * - Redistributions in binary form must reproduce the above copyright
12
 *   notice, this list of conditions and the following disclaimer in the
13
 *   documentation and/or other materials provided with the distribution.
14
 * - The name of the author may not be used to endorse or promote products
15
 *   derived from this software without specific prior written permission.
16
 *
17
 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
18
 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
19
 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
20
 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
21
 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
22
 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
23
 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
24
 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
25
 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
26
 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
27
 */
28
 
1820 decky 29
/** @addtogroup genericproc
1702 cejka 30
 * @{
31
 */
32
/** @file
33
 */
34
 
1888 jermar 35
#ifndef KERN_TASK_H_
36
#define KERN_TASK_H_
1 jermar 37
 
38
#include <synch/spinlock.h>
1460 jermar 39
#include <synch/mutex.h>
2089 decky 40
#include <synch/rwlock.h>
1583 jermar 41
#include <synch/futex.h>
1159 jermar 42
#include <adt/btree.h>
788 jermar 43
#include <adt/list.h>
1174 jermar 44
#include <security/cap.h>
1184 jermar 45
#include <arch/proc/task.h>
2089 decky 46
#include <arch/proc/thread.h>
47
#include <arch/context.h>
48
#include <arch/fpu_context.h>
49
#include <arch/cpu.h>
50
#include <mm/tlb.h>
51
#include <proc/scheduler.h>
1 jermar 52
 
2089 decky 53
#define IPC_MAX_PHONES  16
54
#define THREAD_NAME_BUFLEN	20
55
 
56
struct answerbox;
57
struct task;
58
struct thread;
59
 
60
typedef enum {
61
	IPC_PHONE_FREE = 0,     /**< Phone is free and can be allocated */
62
	IPC_PHONE_CONNECTING,   /**< Phone is connecting somewhere */
63
	IPC_PHONE_CONNECTED,    /**< Phone is connected */
64
	IPC_PHONE_HUNGUP,  	/**< Phone is hung up, waiting for answers to come */
65
	IPC_PHONE_SLAMMED       /**< Phone was hungup from server */
66
} ipc_phone_state_t;
67
 
68
/** Structure identifying phone (in TASK structure) */
69
typedef struct {
70
	SPINLOCK_DECLARE(lock);
71
	link_t link;
72
	struct answerbox *callee;
73
	ipc_phone_state_t state;
74
	atomic_t active_calls;
75
} phone_t;
76
 
77
typedef struct answerbox {
78
	SPINLOCK_DECLARE(lock);
79
 
80
	struct task *task;
81
 
82
	waitq_t wq;
83
 
84
	link_t connected_phones;	/**< Phones connected to this answerbox */
85
	link_t calls;			/**< Received calls */
86
	link_t dispatched_calls;	/* Should be hash table in the future */
87
 
88
	link_t answers;			/**< Answered calls */
89
 
90
	SPINLOCK_DECLARE(irq_lock);
91
	link_t irq_notifs;       	/**< Notifications from IRQ handlers */
92
	link_t irq_head;		/**< IRQs with notifications to this answerbox. */
93
} answerbox_t;
94
 
703 jermar 95
/** Task structure. */
2089 decky 96
typedef struct task {
1579 jermar 97
	/** Task lock.
98
	 *
99
	 * Must be acquired before threads_lock and thread lock of any of its threads.
100
	 */
623 jermar 101
	SPINLOCK_DECLARE(lock);
1579 jermar 102
 
1062 jermar 103
	char *name;
2089 decky 104
	struct thread *main_thread;	/**< Pointer to the main thread. */
410 jermar 105
	link_t th_head;		/**< List of threads contained in this task. */
703 jermar 106
	as_t *as;		/**< Address space. */
1159 jermar 107
	task_id_t taskid;	/**< Unique identity of task */
1820 decky 108
	context_id_t context;	/**< Task security context */
1005 palkovsky 109
 
1579 jermar 110
	/** If this is true, new threads can become part of the task. */
111
	bool accept_new_threads;
112
 
113
	count_t refcount;	/**< Number of references (i.e. threads). */
114
 
1174 jermar 115
	cap_t capabilities;	/**< Task capabilities. */
116
 
998 palkovsky 117
	/* IPC stuff */
955 palkovsky 118
	answerbox_t answerbox;  /**< Communication endpoint */
119
	phone_t phones[IPC_MAX_PHONES];
1568 palkovsky 120
	atomic_t active_calls;  /**< Active asynchronous messages.
121
				 *   It is used for limiting uspace to
122
				 *   certain extent. */
1184 jermar 123
 
1187 jermar 124
	task_arch_t arch;	/**< Architecture specific task data. */
1460 jermar 125
 
126
	/**
1468 jermar 127
	 * Serializes access to the B+tree of task's futexes. This mutex is
128
	 * independent on the task spinlock.
129
	 */
1460 jermar 130
	mutex_t futexes_lock;
131
	btree_t futexes;	/**< B+tree of futexes referenced by this task. */
2039 decky 132
 
133
	uint64_t cycles;	/**< Accumulated accounting. */
2089 decky 134
} task_t;
1 jermar 135
 
2089 decky 136
/** CPU structure.
137
 *
138
 * There is one structure like this for every processor.
139
 */
140
typedef struct {
141
	SPINLOCK_DECLARE(lock);
142
 
143
	tlb_shootdown_msg_t tlb_messages[TLB_MESSAGE_QUEUE_LEN];
144
	count_t tlb_messages_count;
145
 
146
	context_t saved_context;
147
 
148
	atomic_t nrdy;
149
	runq_t rq[RQ_COUNT];
150
	volatile count_t needs_relink;
151
 
152
	SPINLOCK_DECLARE(timeoutlock);
153
	link_t timeout_active_head;
154
 
155
	count_t missed_clock_ticks;	/**< When system clock loses a tick, it is recorded here
156
					     so that clock() can react. This variable is
157
					     CPU-local and can be only accessed when interrupts
158
					     are disabled. */
159
 
160
	/**
161
	 * Processor ID assigned by kernel.
162
	 */
163
	int id;
164
 
165
	int active;
166
	int tlb_active;
167
 
168
	uint16_t frequency_mhz;
169
	uint32_t delay_loop_const;
170
 
171
	cpu_arch_t arch;
172
 
173
	struct thread *fpu_owner;
174
 
175
	/**
176
	 * Stack used by scheduler when there is no running thread.
177
	 */
178
	uint8_t *stack;
179
} cpu_t;
180
 
181
typedef void (* timeout_handler_t)(void *arg);
182
 
183
typedef struct {
184
	SPINLOCK_DECLARE(lock);
185
 
186
	link_t link;			/**< Link to the list of active timeouts on THE->cpu */
187
 
188
	uint64_t ticks;			/**< Timeout will be activated in this amount of clock() ticks. */
189
 
190
	timeout_handler_t handler;	/**< Function that will be called on timeout activation. */
191
	void *arg;			/**< Argument to be passed to handler() function. */
192
 
193
	cpu_t *cpu;			/**< On which processor is this timeout registered. */
194
} timeout_t;
195
 
196
/** Thread states. */
197
typedef enum {
198
	Invalid,	/**< It is an error, if thread is found in this state. */
199
	Running,	/**< State of a thread that is currently executing on some CPU. */
200
	Sleeping,	/**< Thread in this state is waiting for an event. */
201
	Ready,		/**< State of threads in a run queue. */
202
	Entering,	/**< Threads are in this state before they are first readied. */
203
	Exiting,	/**< After a thread calls thread_exit(), it is put into Exiting state. */
204
	Undead		/**< Threads that were not detached but exited are in the Undead state. */
205
} state_t;
206
 
207
/** Join types. */
208
typedef enum {
209
	None,
210
	TaskClnp,	/**< The thread will be joined by ktaskclnp thread. */
211
	TaskGC		/**< The thread will be joined by ktaskgc thread. */
212
} thread_join_type_t;
213
 
214
/** Thread structure. There is one per thread. */
215
typedef struct thread {
216
	link_t rq_link;				/**< Run queue link. */
217
	link_t wq_link;				/**< Wait queue link. */
218
	link_t th_link;				/**< Links to threads within containing task. */
219
 
220
	/** Lock protecting thread structure.
221
	 *
222
	 * Protects the whole thread structure except list links above.
223
	 */
224
	SPINLOCK_DECLARE(lock);
225
 
226
	char name[THREAD_NAME_BUFLEN];
227
 
228
	void (* thread_code)(void *);		/**< Function implementing the thread. */
229
	void *thread_arg;			/**< Argument passed to thread_code() function. */
230
 
231
	/** From here, the stored context is restored when the thread is scheduled. */
232
	context_t saved_context;
233
	/** From here, the stored timeout context is restored when sleep times out. */
234
	context_t sleep_timeout_context;
235
	/** From here, the stored interruption context is restored when sleep is interrupted. */
236
	context_t sleep_interruption_context;
237
 
238
	bool sleep_interruptible;		/**< If true, the thread can be interrupted from sleep. */
239
	waitq_t *sleep_queue;			/**< Wait queue in which this thread sleeps. */
240
	timeout_t sleep_timeout;		/**< Timeout used for timeoutable sleeping.  */
241
	volatile int timeout_pending;		/**< Flag signalling sleep timeout in progress. */
242
 
243
	/** True if this thread is executing copy_from_uspace(). False otherwise. */
244
	bool in_copy_from_uspace;
245
	/** True if this thread is executing copy_to_uspace(). False otherwise. */
246
	bool in_copy_to_uspace;
247
 
248
	/**
249
	 * If true, the thread will not go to sleep at all and will
250
	 * call thread_exit() before returning to userspace.
251
	 */
252
	bool interrupted;			
253
 
254
	thread_join_type_t	join_type;	/**< Who joinins the thread. */
255
	bool detached;				/**< If true, thread_join_timeout() cannot be used on this thread. */
256
	waitq_t join_wq;			/**< Waitq for thread_join_timeout(). */
257
 
258
	fpu_context_t *saved_fpu_context;
259
	int fpu_context_exists;
260
 
261
	/*
262
	 * Defined only if thread doesn't run.
263
	 * It means that fpu context is in CPU that last time executes this thread.
264
	 * This disables migration.
265
	 */
266
	int fpu_context_engaged;
267
 
268
	rwlock_type_t rwlock_holder_type;
269
 
270
	void (* call_me)(void *);		/**< Funtion to be called in scheduler before the thread is put asleep. */
271
	void *call_me_with;			/**< Argument passed to call_me(). */
272
 
273
	state_t state;				/**< Thread's state. */
274
	int flags;				/**< Thread's flags. */
275
 
276
	cpu_t *cpu;				/**< Thread's CPU. */
277
	task_t *task;				/**< Containing task. */
278
 
279
	uint64_t ticks;				/**< Ticks before preemption. */
280
 
281
	uint64_t cycles;			/**< Thread accounting. */
282
	uint64_t last_cycle;		/**< Last sampled cycle. */
283
	bool uncounted;				/**< Thread doesn't affect accumulated accounting. */
284
 
285
	int priority;				/**< Thread's priority. Implemented as index to CPU->rq */
286
	uint32_t tid;				/**< Thread ID. */
287
 
288
	thread_arch_t arch;			/**< Architecture-specific data. */
289
 
290
	uint8_t *kstack;			/**< Thread's kernel stack. */
291
} thread_t;
292
 
1 jermar 293
extern spinlock_t tasks_lock;
1159 jermar 294
extern btree_t tasks_btree;
1 jermar 295
 
296
extern void task_init(void);
1062 jermar 297
extern task_t *task_create(as_t *as, char *name);
1579 jermar 298
extern void task_destroy(task_t *t);
1062 jermar 299
extern task_t *task_run_program(void *program_addr, char *name);
1178 jermar 300
extern task_t *task_find_by_id(task_id_t id);
1579 jermar 301
extern int task_kill(task_id_t id);
2039 decky 302
extern uint64_t task_get_accounting(task_t *t);
1 jermar 303
 
2089 decky 304
extern void cap_set(task_t *t, cap_t caps);
305
extern cap_t cap_get(task_t *t);
1579 jermar 306
 
2089 decky 307
 
1185 jermar 308
#ifndef task_create_arch
309
extern void task_create_arch(task_t *t);
310
#endif
311
 
1587 jermar 312
#ifndef task_destroy_arch
313
extern void task_destroy_arch(task_t *t);
314
#endif
315
 
1780 jermar 316
extern unative_t sys_task_get_id(task_id_t *uspace_task_id);
1176 jermar 317
 
1 jermar 318
#endif
1702 cejka 319
 
1888 jermar 320
/** @}
1702 cejka 321
 */