Subversion Repositories HelenOS

Rev

Rev 2071 | Rev 2090 | Go to most recent revision | Only display areas with differences | Ignore whitespace | Details | Blame | Last modification | View Log | RSS feed

Rev 2071 Rev 2089
1
/*
1
/*3D
2
 * Copyright (c) 2001-2004 Jakub Jermar
2
 * Copyright (c) 2001-2004 Jakub Jermar
3
 * All rights reserved.
3
 * All rights reserved.
4
 *
4
 *
5
 * Redistribution and use in source and binary forms, with or without
5
 * Redistribution and use in source and binary forms, with or without
6
 * modification, are permitted provided that the following conditions
6
 * modification, are permitted provided that the following conditions
7
 * are met:
7
 * are met:
8
 *
8
 *
9
 * - Redistributions of source code must retain the above copyright
9
 * - Redistributions of source code must retain the above copyright
10
 *   notice, this list of conditions and the following disclaimer.
10
 *   notice, this list of conditions and the following disclaimer.
11
 * - Redistributions in binary form must reproduce the above copyright
11
 * - Redistributions in binary form must reproduce the above copyright
12
 *   notice, this list of conditions and the following disclaimer in the
12
 *   notice, this list of conditions and the following disclaimer in the
13
 *   documentation and/or other materials provided with the distribution.
13
 *   documentation and/or other materials provided with the distribution.
14
 * - The name of the author may not be used to endorse or promote products
14
 * - The name of the author may not be used to endorse or promote products
15
 *   derived from this software without specific prior written permission.
15
 *   derived from this software without specific prior written permission.
16
 *
16
 *
17
 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
17
 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
18
 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
18
 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
19
 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
19
 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
20
 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
20
 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
21
 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
21
 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
22
 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
22
 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
23
 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
23
 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
24
 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
24
 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
25
 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
25
 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
26
 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
26
 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
27
 */
27
 */
28
 
28
 
29
/** @addtogroup genericproc
29
/** @addtogroup genericproc
30
 * @{
30
 * @{
31
 */
31
 */
32
/** @file
32
/** @file
33
 */
33
 */
34
 
34
 
35
#ifndef KERN_TASK_H_
35
#ifndef KERN_TASK_H_
36
#define KERN_TASK_H_
36
#define KERN_TASK_H_
37
 
37
 
38
#include <typedefs.h>
-
 
39
#include <synch/spinlock.h>
38
#include <synch/spinlock.h>
40
#include <synch/mutex.h>
39
#include <synch/mutex.h>
-
 
40
#include <synch/rwlock.h>
41
#include <synch/futex.h>
41
#include <synch/futex.h>
42
#include <adt/btree.h>
42
#include <adt/btree.h>
43
#include <adt/list.h>
43
#include <adt/list.h>
44
#include <ipc/ipc.h>
-
 
45
#include <security/cap.h>
44
#include <security/cap.h>
46
#include <arch/proc/task.h>
45
#include <arch/proc/task.h>
-
 
46
#include <arch/proc/thread.h>
-
 
47
#include <arch/context.h>
-
 
48
#include <arch/fpu_context.h>
-
 
49
#include <arch/cpu.h>
-
 
50
#include <mm/tlb.h>
-
 
51
#include <proc/scheduler.h>
-
 
52
 
-
 
53
#define IPC_MAX_PHONES  16
-
 
54
#define THREAD_NAME_BUFLEN  20
-
 
55
 
-
 
56
struct answerbox;
-
 
57
struct task;
-
 
58
struct thread;
-
 
59
 
-
 
60
typedef enum {
-
 
61
    IPC_PHONE_FREE = 0,     /**< Phone is free and can be allocated */
-
 
62
    IPC_PHONE_CONNECTING,   /**< Phone is connecting somewhere */
-
 
63
    IPC_PHONE_CONNECTED,    /**< Phone is connected */
-
 
64
    IPC_PHONE_HUNGUP,   /**< Phone is hung up, waiting for answers to come */
-
 
65
    IPC_PHONE_SLAMMED       /**< Phone was hungup from server */
-
 
66
} ipc_phone_state_t;
-
 
67
 
-
 
68
/** Structure identifying phone (in TASK structure) */
-
 
69
typedef struct {
-
 
70
    SPINLOCK_DECLARE(lock);
-
 
71
    link_t link;
-
 
72
    struct answerbox *callee;
-
 
73
    ipc_phone_state_t state;
-
 
74
    atomic_t active_calls;
-
 
75
} phone_t;
-
 
76
 
-
 
77
typedef struct answerbox {
-
 
78
    SPINLOCK_DECLARE(lock);
-
 
79
 
-
 
80
    struct task *task;
-
 
81
 
-
 
82
    waitq_t wq;
-
 
83
 
-
 
84
    link_t connected_phones;    /**< Phones connected to this answerbox */
-
 
85
    link_t calls;           /**< Received calls */
-
 
86
    link_t dispatched_calls;    /* Should be hash table in the future */
-
 
87
 
-
 
88
    link_t answers;         /**< Answered calls */
-
 
89
 
-
 
90
    SPINLOCK_DECLARE(irq_lock);
-
 
91
    link_t irq_notifs;          /**< Notifications from IRQ handlers */
-
 
92
    link_t irq_head;        /**< IRQs with notifications to this answerbox. */
-
 
93
} answerbox_t;
47
 
94
 
48
/** Task structure. */
95
/** Task structure. */
49
struct task {
96
typedef struct task {
50
    /** Task lock.
97
    /** Task lock.
51
     *
98
     *
52
     * Must be acquired before threads_lock and thread lock of any of its threads.
99
     * Must be acquired before threads_lock and thread lock of any of its threads.
53
     */
100
     */
54
    SPINLOCK_DECLARE(lock);
101
    SPINLOCK_DECLARE(lock);
55
   
102
   
56
    char *name;
103
    char *name;
57
    thread_t *main_thread;  /**< Pointer to the main thread. */
104
    struct thread *main_thread; /**< Pointer to the main thread. */
58
    link_t th_head;     /**< List of threads contained in this task. */
105
    link_t th_head;     /**< List of threads contained in this task. */
59
    as_t *as;       /**< Address space. */
106
    as_t *as;       /**< Address space. */
60
    task_id_t taskid;   /**< Unique identity of task */
107
    task_id_t taskid;   /**< Unique identity of task */
61
    context_id_t context;   /**< Task security context */
108
    context_id_t context;   /**< Task security context */
62
 
109
 
63
    /** If this is true, new threads can become part of the task. */
110
    /** If this is true, new threads can become part of the task. */
64
    bool accept_new_threads;
111
    bool accept_new_threads;
65
 
112
 
66
    count_t refcount;   /**< Number of references (i.e. threads). */
113
    count_t refcount;   /**< Number of references (i.e. threads). */
67
 
114
 
68
    cap_t capabilities; /**< Task capabilities. */
115
    cap_t capabilities; /**< Task capabilities. */
69
 
116
 
70
    /* IPC stuff */
117
    /* IPC stuff */
71
    answerbox_t answerbox;  /**< Communication endpoint */
118
    answerbox_t answerbox;  /**< Communication endpoint */
72
    phone_t phones[IPC_MAX_PHONES];
119
    phone_t phones[IPC_MAX_PHONES];
73
    atomic_t active_calls;  /**< Active asynchronous messages.
120
    atomic_t active_calls;  /**< Active asynchronous messages.
74
                 *   It is used for limiting uspace to
121
                 *   It is used for limiting uspace to
75
                 *   certain extent. */
122
                 *   certain extent. */
76
   
123
   
77
    task_arch_t arch;   /**< Architecture specific task data. */
124
    task_arch_t arch;   /**< Architecture specific task data. */
78
   
125
   
79
    /**
126
    /**
80
     * Serializes access to the B+tree of task's futexes. This mutex is
127
     * Serializes access to the B+tree of task's futexes. This mutex is
81
     * independent on the task spinlock.
128
     * independent on the task spinlock.
82
     */
129
     */
83
    mutex_t futexes_lock;
130
    mutex_t futexes_lock;
84
    btree_t futexes;    /**< B+tree of futexes referenced by this task. */
131
    btree_t futexes;    /**< B+tree of futexes referenced by this task. */
85
   
132
   
86
    uint64_t cycles;    /**< Accumulated accounting. */
133
    uint64_t cycles;    /**< Accumulated accounting. */
-
 
134
} task_t;
-
 
135
 
-
 
136
/** CPU structure.
87
};
137
 *
-
 
138
 * There is one structure like this for every processor.
-
 
139
 */
-
 
140
typedef struct {
-
 
141
    SPINLOCK_DECLARE(lock);
-
 
142
 
-
 
143
    tlb_shootdown_msg_t tlb_messages[TLB_MESSAGE_QUEUE_LEN];
-
 
144
    count_t tlb_messages_count;
-
 
145
   
-
 
146
    context_t saved_context;
-
 
147
 
-
 
148
    atomic_t nrdy;
-
 
149
    runq_t rq[RQ_COUNT];
-
 
150
    volatile count_t needs_relink;
-
 
151
 
-
 
152
    SPINLOCK_DECLARE(timeoutlock);
-
 
153
    link_t timeout_active_head;
-
 
154
 
-
 
155
    count_t missed_clock_ticks; /**< When system clock loses a tick, it is recorded here
-
 
156
                         so that clock() can react. This variable is
-
 
157
                         CPU-local and can be only accessed when interrupts
-
 
158
                         are disabled. */
-
 
159
 
-
 
160
    /**
-
 
161
     * Processor ID assigned by kernel.
-
 
162
     */
-
 
163
    int id;
-
 
164
   
-
 
165
    int active;
-
 
166
    int tlb_active;
-
 
167
 
-
 
168
    uint16_t frequency_mhz;
-
 
169
    uint32_t delay_loop_const;
-
 
170
 
-
 
171
    cpu_arch_t arch;
-
 
172
 
-
 
173
    struct thread *fpu_owner;
-
 
174
   
-
 
175
    /**
-
 
176
     * Stack used by scheduler when there is no running thread.
-
 
177
     */
-
 
178
    uint8_t *stack;
-
 
179
} cpu_t;
-
 
180
 
-
 
181
typedef void (* timeout_handler_t)(void *arg);
-
 
182
 
-
 
183
typedef struct {
-
 
184
    SPINLOCK_DECLARE(lock);
-
 
185
 
-
 
186
    link_t link;            /**< Link to the list of active timeouts on THE->cpu */
-
 
187
   
-
 
188
    uint64_t ticks;         /**< Timeout will be activated in this amount of clock() ticks. */
-
 
189
 
-
 
190
    timeout_handler_t handler;  /**< Function that will be called on timeout activation. */
-
 
191
    void *arg;          /**< Argument to be passed to handler() function. */
-
 
192
   
-
 
193
    cpu_t *cpu;         /**< On which processor is this timeout registered. */
-
 
194
} timeout_t;
-
 
195
 
-
 
196
/** Thread states. */
-
 
197
typedef enum {
-
 
198
    Invalid,    /**< It is an error, if thread is found in this state. */
-
 
199
    Running,    /**< State of a thread that is currently executing on some CPU. */
-
 
200
    Sleeping,   /**< Thread in this state is waiting for an event. */
-
 
201
    Ready,      /**< State of threads in a run queue. */
-
 
202
    Entering,   /**< Threads are in this state before they are first readied. */
-
 
203
    Exiting,    /**< After a thread calls thread_exit(), it is put into Exiting state. */
-
 
204
    Undead      /**< Threads that were not detached but exited are in the Undead state. */
-
 
205
} state_t;
-
 
206
 
-
 
207
/** Join types. */
-
 
208
typedef enum {
-
 
209
    None,
-
 
210
    TaskClnp,   /**< The thread will be joined by ktaskclnp thread. */
-
 
211
    TaskGC      /**< The thread will be joined by ktaskgc thread. */
-
 
212
} thread_join_type_t;
-
 
213
 
-
 
214
/** Thread structure. There is one per thread. */
-
 
215
typedef struct thread {
-
 
216
    link_t rq_link;             /**< Run queue link. */
-
 
217
    link_t wq_link;             /**< Wait queue link. */
-
 
218
    link_t th_link;             /**< Links to threads within containing task. */
-
 
219
   
-
 
220
    /** Lock protecting thread structure.
-
 
221
     *
-
 
222
     * Protects the whole thread structure except list links above.
-
 
223
     */
-
 
224
    SPINLOCK_DECLARE(lock);
-
 
225
 
-
 
226
    char name[THREAD_NAME_BUFLEN];
-
 
227
 
-
 
228
    void (* thread_code)(void *);       /**< Function implementing the thread. */
-
 
229
    void *thread_arg;           /**< Argument passed to thread_code() function. */
-
 
230
 
-
 
231
    /** From here, the stored context is restored when the thread is scheduled. */
-
 
232
    context_t saved_context;
-
 
233
    /** From here, the stored timeout context is restored when sleep times out. */
-
 
234
    context_t sleep_timeout_context;
-
 
235
    /** From here, the stored interruption context is restored when sleep is interrupted. */
-
 
236
    context_t sleep_interruption_context;
-
 
237
 
-
 
238
    bool sleep_interruptible;       /**< If true, the thread can be interrupted from sleep. */
-
 
239
    waitq_t *sleep_queue;           /**< Wait queue in which this thread sleeps. */
-
 
240
    timeout_t sleep_timeout;        /**< Timeout used for timeoutable sleeping.  */
-
 
241
    volatile int timeout_pending;       /**< Flag signalling sleep timeout in progress. */
-
 
242
 
-
 
243
    /** True if this thread is executing copy_from_uspace(). False otherwise. */
-
 
244
    bool in_copy_from_uspace;
-
 
245
    /** True if this thread is executing copy_to_uspace(). False otherwise. */
-
 
246
    bool in_copy_to_uspace;
-
 
247
   
-
 
248
    /**
-
 
249
     * If true, the thread will not go to sleep at all and will
-
 
250
     * call thread_exit() before returning to userspace.
-
 
251
     */
-
 
252
    bool interrupted;          
-
 
253
   
-
 
254
    thread_join_type_t  join_type;  /**< Who joinins the thread. */
-
 
255
    bool detached;              /**< If true, thread_join_timeout() cannot be used on this thread. */
-
 
256
    waitq_t join_wq;            /**< Waitq for thread_join_timeout(). */
-
 
257
 
-
 
258
    fpu_context_t *saved_fpu_context;
-
 
259
    int fpu_context_exists;
-
 
260
 
-
 
261
    /*
-
 
262
     * Defined only if thread doesn't run.
-
 
263
     * It means that fpu context is in CPU that last time executes this thread.
-
 
264
     * This disables migration.
-
 
265
     */
-
 
266
    int fpu_context_engaged;
-
 
267
 
-
 
268
    rwlock_type_t rwlock_holder_type;
-
 
269
 
-
 
270
    void (* call_me)(void *);       /**< Funtion to be called in scheduler before the thread is put asleep. */
-
 
271
    void *call_me_with;         /**< Argument passed to call_me(). */
-
 
272
 
-
 
273
    state_t state;              /**< Thread's state. */
-
 
274
    int flags;              /**< Thread's flags. */
-
 
275
   
-
 
276
    cpu_t *cpu;             /**< Thread's CPU. */
-
 
277
    task_t *task;               /**< Containing task. */
-
 
278
 
-
 
279
    uint64_t ticks;             /**< Ticks before preemption. */
-
 
280
   
-
 
281
    uint64_t cycles;            /**< Thread accounting. */
-
 
282
    uint64_t last_cycle;        /**< Last sampled cycle. */
-
 
283
    bool uncounted;             /**< Thread doesn't affect accumulated accounting. */
-
 
284
 
-
 
285
    int priority;               /**< Thread's priority. Implemented as index to CPU->rq */
-
 
286
    uint32_t tid;               /**< Thread ID. */
-
 
287
   
-
 
288
    thread_arch_t arch;         /**< Architecture-specific data. */
-
 
289
 
-
 
290
    uint8_t *kstack;            /**< Thread's kernel stack. */
-
 
291
} thread_t;
88
 
292
 
89
extern spinlock_t tasks_lock;
293
extern spinlock_t tasks_lock;
90
extern btree_t tasks_btree;
294
extern btree_t tasks_btree;
91
 
295
 
92
extern void task_init(void);
296
extern void task_init(void);
93
extern task_t *task_create(as_t *as, char *name);
297
extern task_t *task_create(as_t *as, char *name);
94
extern void task_destroy(task_t *t);
298
extern void task_destroy(task_t *t);
95
extern task_t *task_run_program(void *program_addr, char *name);
299
extern task_t *task_run_program(void *program_addr, char *name);
96
extern task_t *task_find_by_id(task_id_t id);
300
extern task_t *task_find_by_id(task_id_t id);
97
extern int task_kill(task_id_t id);
301
extern int task_kill(task_id_t id);
98
extern uint64_t task_get_accounting(task_t *t);
302
extern uint64_t task_get_accounting(task_t *t);
99
 
303
 
-
 
304
extern void cap_set(task_t *t, cap_t caps);
-
 
305
extern cap_t cap_get(task_t *t);
-
 
306
 
100
 
307
 
101
#ifndef task_create_arch
308
#ifndef task_create_arch
102
extern void task_create_arch(task_t *t);
309
extern void task_create_arch(task_t *t);
103
#endif
310
#endif
104
 
311
 
105
#ifndef task_destroy_arch
312
#ifndef task_destroy_arch
106
extern void task_destroy_arch(task_t *t);
313
extern void task_destroy_arch(task_t *t);
107
#endif
314
#endif
108
 
315
 
109
extern unative_t sys_task_get_id(task_id_t *uspace_task_id);
316
extern unative_t sys_task_get_id(task_id_t *uspace_task_id);
110
 
317
 
111
#endif
318
#endif
112
 
319
 
113
/** @}
320
/** @}
114
 */
321
 */
115
 
322