Subversion Repositories HelenOS

Rev

Rev 2089 | Rev 2092 | Go to most recent revision | Blame | Compare with Previous | Last modification | View Log | Download | RSS feed

  1. /*
  2.  * Copyright (c) 2001-2004 Jakub Jermar
  3.  * All rights reserved.
  4.  *
  5.  * Redistribution and use in source and binary forms, with or without
  6.  * modification, are permitted provided that the following conditions
  7.  * are met:
  8.  *
  9.  * - Redistributions of source code must retain the above copyright
  10.  *   notice, this list of conditions and the following disclaimer.
  11.  * - Redistributions in binary form must reproduce the above copyright
  12.  *   notice, this list of conditions and the following disclaimer in the
  13.  *   documentation and/or other materials provided with the distribution.
  14.  * - The name of the author may not be used to endorse or promote products
  15.  *   derived from this software without specific prior written permission.
  16.  *
  17.  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
  18.  * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
  19.  * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
  20.  * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
  21.  * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
  22.  * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
  23.  * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
  24.  * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
  25.  * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
  26.  * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
  27.  */
  28.  
  29. /** @addtogroup genericproc
  30.  * @{
  31.  */
  32. /** @file
  33.  */
  34.  
  35. #ifndef KERN_TASK_H_
  36. #define KERN_TASK_H_
  37.  
  38. #include <synch/spinlock.h>
  39. #include <synch/mutex.h>
  40. #include <synch/rwlock.h>
  41. #include <synch/futex.h>
  42. #include <adt/btree.h>
  43. #include <adt/list.h>
  44. #include <security/cap.h>
  45. #include <arch/proc/task.h>
  46. #include <arch/proc/thread.h>
  47. #include <arch/context.h>
  48. #include <arch/fpu_context.h>
  49. #include <arch/cpu.h>
  50. #include <mm/tlb.h>
  51. #include <proc/scheduler.h>
  52.  
  53. #define IPC_MAX_PHONES  16
  54. #define THREAD_NAME_BUFLEN  20
  55.  
  56. struct answerbox;
  57. struct task;
  58. struct thread;
  59.  
  60. typedef enum {
  61.     IPC_PHONE_FREE = 0,     /**< Phone is free and can be allocated */
  62.     IPC_PHONE_CONNECTING,   /**< Phone is connecting somewhere */
  63.     IPC_PHONE_CONNECTED,    /**< Phone is connected */
  64.     IPC_PHONE_HUNGUP,   /**< Phone is hung up, waiting for answers to come */
  65.     IPC_PHONE_SLAMMED       /**< Phone was hungup from server */
  66. } ipc_phone_state_t;
  67.  
  68. /** Structure identifying phone (in TASK structure) */
  69. typedef struct {
  70.     SPINLOCK_DECLARE(lock);
  71.     link_t link;
  72.     struct answerbox *callee;
  73.     ipc_phone_state_t state;
  74.     atomic_t active_calls;
  75. } phone_t;
  76.  
  77. typedef struct answerbox {
  78.     SPINLOCK_DECLARE(lock);
  79.  
  80.     struct task *task;
  81.  
  82.     waitq_t wq;
  83.  
  84.     link_t connected_phones;    /**< Phones connected to this answerbox */
  85.     link_t calls;           /**< Received calls */
  86.     link_t dispatched_calls;    /* Should be hash table in the future */
  87.  
  88.     link_t answers;         /**< Answered calls */
  89.  
  90.     SPINLOCK_DECLARE(irq_lock);
  91.     link_t irq_notifs;          /**< Notifications from IRQ handlers */
  92.     link_t irq_head;        /**< IRQs with notifications to this answerbox. */
  93. } answerbox_t;
  94.  
  95. /** Task structure. */
  96. typedef struct task {
  97.     /** Task lock.
  98.      *
  99.      * Must be acquired before threads_lock and thread lock of any of its threads.
  100.      */
  101.     SPINLOCK_DECLARE(lock);
  102.    
  103.     char *name;
  104.     struct thread *main_thread; /**< Pointer to the main thread. */
  105.     link_t th_head;     /**< List of threads contained in this task. */
  106.     as_t *as;       /**< Address space. */
  107.     task_id_t taskid;   /**< Unique identity of task */
  108.     context_id_t context;   /**< Task security context */
  109.  
  110.     /** If this is true, new threads can become part of the task. */
  111.     bool accept_new_threads;
  112.  
  113.     count_t refcount;   /**< Number of references (i.e. threads). */
  114.  
  115.     cap_t capabilities; /**< Task capabilities. */
  116.  
  117.     /* IPC stuff */
  118.     answerbox_t answerbox;  /**< Communication endpoint */
  119.     phone_t phones[IPC_MAX_PHONES];
  120.     atomic_t active_calls;  /**< Active asynchronous messages.
  121.                  *   It is used for limiting uspace to
  122.                  *   certain extent. */
  123.    
  124.     task_arch_t arch;   /**< Architecture specific task data. */
  125.    
  126.     /**
  127.      * Serializes access to the B+tree of task's futexes. This mutex is
  128.      * independent on the task spinlock.
  129.      */
  130.     mutex_t futexes_lock;
  131.     btree_t futexes;    /**< B+tree of futexes referenced by this task. */
  132.    
  133.     uint64_t cycles;    /**< Accumulated accounting. */
  134. } task_t;
  135.  
  136. /** CPU structure.
  137.  *
  138.  * There is one structure like this for every processor.
  139.  */
  140. typedef struct {
  141.     SPINLOCK_DECLARE(lock);
  142.  
  143.     tlb_shootdown_msg_t tlb_messages[TLB_MESSAGE_QUEUE_LEN];
  144.     count_t tlb_messages_count;
  145.    
  146.     context_t saved_context;
  147.  
  148.     atomic_t nrdy;
  149.     runq_t rq[RQ_COUNT];
  150.     volatile count_t needs_relink;
  151.  
  152.     SPINLOCK_DECLARE(timeoutlock);
  153.     link_t timeout_active_head;
  154.  
  155.     count_t missed_clock_ticks; /**< When system clock loses a tick, it is recorded here
  156.                          so that clock() can react. This variable is
  157.                          CPU-local and can be only accessed when interrupts
  158.                          are disabled. */
  159.  
  160.     /**
  161.      * Processor ID assigned by kernel.
  162.      */
  163.     unsigned int id;
  164.    
  165.     int active;
  166.     int tlb_active;
  167.  
  168.     uint16_t frequency_mhz;
  169.     uint32_t delay_loop_const;
  170.  
  171.     cpu_arch_t arch;
  172.  
  173.     struct thread *fpu_owner;
  174.    
  175.     /**
  176.      * Stack used by scheduler when there is no running thread.
  177.      */
  178.     uint8_t *stack;
  179. } cpu_t;
  180.  
  181. typedef void (* timeout_handler_t)(void *arg);
  182.  
  183. typedef struct {
  184.     SPINLOCK_DECLARE(lock);
  185.  
  186.     link_t link;            /**< Link to the list of active timeouts on THE->cpu */
  187.    
  188.     uint64_t ticks;         /**< Timeout will be activated in this amount of clock() ticks. */
  189.  
  190.     timeout_handler_t handler;  /**< Function that will be called on timeout activation. */
  191.     void *arg;          /**< Argument to be passed to handler() function. */
  192.    
  193.     cpu_t *cpu;         /**< On which processor is this timeout registered. */
  194. } timeout_t;
  195.  
  196. /** Thread states. */
  197. typedef enum {
  198.     Invalid,    /**< It is an error, if thread is found in this state. */
  199.     Running,    /**< State of a thread that is currently executing on some CPU. */
  200.     Sleeping,   /**< Thread in this state is waiting for an event. */
  201.     Ready,      /**< State of threads in a run queue. */
  202.     Entering,   /**< Threads are in this state before they are first readied. */
  203.     Exiting,    /**< After a thread calls thread_exit(), it is put into Exiting state. */
  204.     Undead      /**< Threads that were not detached but exited are in the Undead state. */
  205. } state_t;
  206.  
  207. /** Join types. */
  208. typedef enum {
  209.     None,
  210.     TaskClnp,   /**< The thread will be joined by ktaskclnp thread. */
  211.     TaskGC      /**< The thread will be joined by ktaskgc thread. */
  212. } thread_join_type_t;
  213.  
  214. /** Thread structure. There is one per thread. */
  215. typedef struct thread {
  216.     link_t rq_link;             /**< Run queue link. */
  217.     link_t wq_link;             /**< Wait queue link. */
  218.     link_t th_link;             /**< Links to threads within containing task. */
  219.    
  220.     /** Lock protecting thread structure.
  221.      *
  222.      * Protects the whole thread structure except list links above.
  223.      */
  224.     SPINLOCK_DECLARE(lock);
  225.  
  226.     char name[THREAD_NAME_BUFLEN];
  227.  
  228.     void (* thread_code)(void *);       /**< Function implementing the thread. */
  229.     void *thread_arg;           /**< Argument passed to thread_code() function. */
  230.  
  231.     /** From here, the stored context is restored when the thread is scheduled. */
  232.     context_t saved_context;
  233.     /** From here, the stored timeout context is restored when sleep times out. */
  234.     context_t sleep_timeout_context;
  235.     /** From here, the stored interruption context is restored when sleep is interrupted. */
  236.     context_t sleep_interruption_context;
  237.  
  238.     bool sleep_interruptible;       /**< If true, the thread can be interrupted from sleep. */
  239.     waitq_t *sleep_queue;           /**< Wait queue in which this thread sleeps. */
  240.     timeout_t sleep_timeout;        /**< Timeout used for timeoutable sleeping.  */
  241.     volatile int timeout_pending;       /**< Flag signalling sleep timeout in progress. */
  242.  
  243.     /** True if this thread is executing copy_from_uspace(). False otherwise. */
  244.     bool in_copy_from_uspace;
  245.     /** True if this thread is executing copy_to_uspace(). False otherwise. */
  246.     bool in_copy_to_uspace;
  247.    
  248.     /**
  249.      * If true, the thread will not go to sleep at all and will
  250.      * call thread_exit() before returning to userspace.
  251.      */
  252.     bool interrupted;          
  253.    
  254.     thread_join_type_t  join_type;  /**< Who joinins the thread. */
  255.     bool detached;              /**< If true, thread_join_timeout() cannot be used on this thread. */
  256.     waitq_t join_wq;            /**< Waitq for thread_join_timeout(). */
  257.  
  258.     fpu_context_t *saved_fpu_context;
  259.     int fpu_context_exists;
  260.  
  261.     /*
  262.      * Defined only if thread doesn't run.
  263.      * It means that fpu context is in CPU that last time executes this thread.
  264.      * This disables migration.
  265.      */
  266.     int fpu_context_engaged;
  267.  
  268.     rwlock_type_t rwlock_holder_type;
  269.  
  270.     void (* call_me)(void *);       /**< Funtion to be called in scheduler before the thread is put asleep. */
  271.     void *call_me_with;         /**< Argument passed to call_me(). */
  272.  
  273.     state_t state;              /**< Thread's state. */
  274.     int flags;              /**< Thread's flags. */
  275.    
  276.     cpu_t *cpu;             /**< Thread's CPU. */
  277.     task_t *task;               /**< Containing task. */
  278.  
  279.     uint64_t ticks;             /**< Ticks before preemption. */
  280.    
  281.     uint64_t cycles;            /**< Thread accounting. */
  282.     uint64_t last_cycle;        /**< Last sampled cycle. */
  283.     bool uncounted;             /**< Thread doesn't affect accumulated accounting. */
  284.  
  285.     int priority;               /**< Thread's priority. Implemented as index to CPU->rq */
  286.     uint32_t tid;               /**< Thread ID. */
  287.    
  288.     thread_arch_t arch;         /**< Architecture-specific data. */
  289.  
  290.     uint8_t *kstack;            /**< Thread's kernel stack. */
  291. } thread_t;
  292.  
  293. extern spinlock_t tasks_lock;
  294. extern btree_t tasks_btree;
  295.  
  296. extern void task_init(void);
  297. extern task_t *task_create(as_t *as, char *name);
  298. extern void task_destroy(task_t *t);
  299. extern task_t *task_run_program(void *program_addr, char *name);
  300. extern task_t *task_find_by_id(task_id_t id);
  301. extern int task_kill(task_id_t id);
  302. extern uint64_t task_get_accounting(task_t *t);
  303.  
  304. extern void cap_set(task_t *t, cap_t caps);
  305. extern cap_t cap_get(task_t *t);
  306.  
  307.  
  308. #ifndef task_create_arch
  309. extern void task_create_arch(task_t *t);
  310. #endif
  311.  
  312. #ifndef task_destroy_arch
  313. extern void task_destroy_arch(task_t *t);
  314. #endif
  315.  
  316. extern unative_t sys_task_get_id(task_id_t *uspace_task_id);
  317.  
  318. #endif
  319.  
  320. /** @}
  321.  */
  322.