Rev 4084 | Rev 4384 | Go to most recent revision | Only display areas with differences | Ignore whitespace | Details | Blame | Last modification | View Log | RSS feed
Rev 4084 | Rev 4173 | ||
---|---|---|---|
1 | /* |
1 | /* |
2 | * Copyright (c) 2001-2004 Jakub Jermar |
2 | * Copyright (c) 2001-2004 Jakub Jermar |
3 | * All rights reserved. |
3 | * All rights reserved. |
4 | * |
4 | * |
5 | * Redistribution and use in source and binary forms, with or without |
5 | * Redistribution and use in source and binary forms, with or without |
6 | * modification, are permitted provided that the following conditions |
6 | * modification, are permitted provided that the following conditions |
7 | * are met: |
7 | * are met: |
8 | * |
8 | * |
9 | * - Redistributions of source code must retain the above copyright |
9 | * - Redistributions of source code must retain the above copyright |
10 | * notice, this list of conditions and the following disclaimer. |
10 | * notice, this list of conditions and the following disclaimer. |
11 | * - Redistributions in binary form must reproduce the above copyright |
11 | * - Redistributions in binary form must reproduce the above copyright |
12 | * notice, this list of conditions and the following disclaimer in the |
12 | * notice, this list of conditions and the following disclaimer in the |
13 | * documentation and/or other materials provided with the distribution. |
13 | * documentation and/or other materials provided with the distribution. |
14 | * - The name of the author may not be used to endorse or promote products |
14 | * - The name of the author may not be used to endorse or promote products |
15 | * derived from this software without specific prior written permission. |
15 | * derived from this software without specific prior written permission. |
16 | * |
16 | * |
17 | * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR |
17 | * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR |
18 | * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES |
18 | * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES |
19 | * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. |
19 | * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. |
20 | * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, |
20 | * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, |
21 | * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT |
21 | * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT |
22 | * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
22 | * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
23 | * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
23 | * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
24 | * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
24 | * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
25 | * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF |
25 | * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF |
26 | * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
26 | * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
27 | */ |
27 | */ |
28 | 28 | ||
29 | /** @addtogroup main |
29 | /** @addtogroup main |
30 | * @{ |
30 | * @{ |
31 | */ |
31 | */ |
32 | 32 | ||
33 | /** |
33 | /** |
34 | * @file |
34 | * @file |
35 | * @brief Main initialization kernel function for all processors. |
35 | * @brief Main initialization kernel function for all processors. |
36 | * |
36 | * |
37 | * During kernel boot, all processors, after architecture dependent |
37 | * During kernel boot, all processors, after architecture dependent |
38 | * initialization, start executing code found in this file. After |
38 | * initialization, start executing code found in this file. After |
39 | * bringing up all subsystems, control is passed to scheduler(). |
39 | * bringing up all subsystems, control is passed to scheduler(). |
40 | * |
40 | * |
41 | * The bootstrap processor starts executing main_bsp() while |
41 | * The bootstrap processor starts executing main_bsp() while |
42 | * the application processors start executing main_ap(). |
42 | * the application processors start executing main_ap(). |
43 | * |
43 | * |
44 | * @see scheduler() |
44 | * @see scheduler() |
45 | * @see main_bsp() |
45 | * @see main_bsp() |
46 | * @see main_ap() |
46 | * @see main_ap() |
47 | */ |
47 | */ |
48 | 48 | ||
49 | #include <arch/asm.h> |
49 | #include <arch/asm.h> |
50 | #include <context.h> |
50 | #include <context.h> |
51 | #include <print.h> |
51 | #include <print.h> |
52 | #include <panic.h> |
52 | #include <panic.h> |
53 | #include <debug.h> |
53 | #include <debug.h> |
54 | #include <config.h> |
54 | #include <config.h> |
55 | #include <time/clock.h> |
55 | #include <time/clock.h> |
56 | #include <time/timeout.h> |
56 | #include <time/timeout.h> |
57 | #include <proc/scheduler.h> |
57 | #include <proc/scheduler.h> |
58 | #include <proc/thread.h> |
58 | #include <proc/thread.h> |
59 | #include <proc/task.h> |
59 | #include <proc/task.h> |
60 | #include <proc/tasklet.h> |
60 | #include <proc/tasklet.h> |
61 | #include <main/kinit.h> |
61 | #include <main/kinit.h> |
62 | #include <main/version.h> |
62 | #include <main/version.h> |
63 | #include <console/kconsole.h> |
63 | #include <console/kconsole.h> |
64 | #include <console/console.h> |
64 | #include <console/console.h> |
65 | #include <cpu.h> |
65 | #include <cpu.h> |
66 | #include <align.h> |
66 | #include <align.h> |
67 | #include <interrupt.h> |
67 | #include <interrupt.h> |
68 | #include <mm/frame.h> |
68 | #include <mm/frame.h> |
69 | #include <mm/page.h> |
69 | #include <mm/page.h> |
70 | #include <genarch/mm/page_pt.h> |
70 | #include <genarch/mm/page_pt.h> |
71 | #include <mm/tlb.h> |
71 | #include <mm/tlb.h> |
72 | #include <mm/as.h> |
72 | #include <mm/as.h> |
73 | #include <mm/slab.h> |
73 | #include <mm/slab.h> |
74 | #include <synch/waitq.h> |
74 | #include <synch/waitq.h> |
75 | #include <synch/futex.h> |
75 | #include <synch/futex.h> |
76 | #include <arch/arch.h> |
76 | #include <arch/arch.h> |
77 | #include <arch.h> |
77 | #include <arch.h> |
78 | #include <arch/faddr.h> |
78 | #include <arch/faddr.h> |
79 | #include <ipc/ipc.h> |
79 | #include <ipc/ipc.h> |
80 | #include <macros.h> |
80 | #include <macros.h> |
81 | #include <adt/btree.h> |
81 | #include <adt/btree.h> |
82 | #include <smp/smp.h> |
82 | #include <smp/smp.h> |
83 | #include <ddi/ddi.h> |
83 | #include <ddi/ddi.h> |
84 | #include <main/main.h> |
84 | #include <main/main.h> |
- | 85 | #include <event/event.h> |
|
85 | 86 | ||
86 | /** Global configuration structure. */ |
87 | /** Global configuration structure. */ |
87 | config_t config; |
88 | config_t config; |
88 | 89 | ||
89 | /** Initial user-space tasks */ |
90 | /** Initial user-space tasks */ |
90 | init_t init = { |
91 | init_t init = { |
91 | .cnt = 0 |
92 | .cnt = 0 |
92 | }; |
93 | }; |
93 | 94 | ||
94 | /** Boot allocations. */ |
95 | /** Boot allocations. */ |
95 | ballocs_t ballocs = { |
96 | ballocs_t ballocs = { |
96 | .base = NULL, |
97 | .base = NULL, |
97 | .size = 0 |
98 | .size = 0 |
98 | }; |
99 | }; |
99 | 100 | ||
100 | context_t ctx; |
101 | context_t ctx; |
101 | 102 | ||
102 | /* |
103 | /* |
103 | * These 'hardcoded' variables will be intialized by |
104 | * These 'hardcoded' variables will be intialized by |
104 | * the linker or the low level assembler code with |
105 | * the linker or the low level assembler code with |
105 | * appropriate sizes and addresses. |
106 | * appropriate sizes and addresses. |
106 | */ |
107 | */ |
107 | 108 | ||
108 | /** Virtual address of where the kernel is loaded. */ |
109 | /** Virtual address of where the kernel is loaded. */ |
109 | uintptr_t hardcoded_load_address = 0; |
110 | uintptr_t hardcoded_load_address = 0; |
110 | /** Size of the kernel code in bytes. */ |
111 | /** Size of the kernel code in bytes. */ |
111 | size_t hardcoded_ktext_size = 0; |
112 | size_t hardcoded_ktext_size = 0; |
112 | /** Size of the kernel data in bytes. */ |
113 | /** Size of the kernel data in bytes. */ |
113 | size_t hardcoded_kdata_size = 0; |
114 | size_t hardcoded_kdata_size = 0; |
114 | /** Lowest safe stack virtual address. */ |
115 | /** Lowest safe stack virtual address. */ |
115 | uintptr_t stack_safe = 0; |
116 | uintptr_t stack_safe = 0; |
116 | 117 | ||
117 | /* |
118 | /* |
118 | * These two functions prevent stack from underflowing during the |
119 | * These two functions prevent stack from underflowing during the |
119 | * kernel boot phase when SP is set to the very top of the reserved |
120 | * kernel boot phase when SP is set to the very top of the reserved |
120 | * space. The stack could get corrupted by a fooled compiler-generated |
121 | * space. The stack could get corrupted by a fooled compiler-generated |
121 | * pop sequence otherwise. |
122 | * pop sequence otherwise. |
122 | */ |
123 | */ |
123 | static void main_bsp_separated_stack(void); |
124 | static void main_bsp_separated_stack(void); |
124 | #ifdef CONFIG_SMP |
125 | #ifdef CONFIG_SMP |
125 | static void main_ap_separated_stack(void); |
126 | static void main_ap_separated_stack(void); |
126 | #endif |
127 | #endif |
127 | 128 | ||
128 | #define CONFIG_STACK_SIZE ((1 << STACK_FRAMES) * STACK_SIZE) |
129 | #define CONFIG_STACK_SIZE ((1 << STACK_FRAMES) * STACK_SIZE) |
129 | 130 | ||
130 | /** Main kernel routine for bootstrap CPU. |
131 | /** Main kernel routine for bootstrap CPU. |
131 | * |
132 | * |
132 | * The code here still runs on the boot stack, which knows nothing about |
133 | * The code here still runs on the boot stack, which knows nothing about |
133 | * preemption counts. Because of that, this function cannot directly call |
134 | * preemption counts. Because of that, this function cannot directly call |
134 | * functions that disable or enable preemption (e.g. spinlock_lock()). The |
135 | * functions that disable or enable preemption (e.g. spinlock_lock()). The |
135 | * primary task of this function is to calculate address of a new stack and |
136 | * primary task of this function is to calculate address of a new stack and |
136 | * switch to it. |
137 | * switch to it. |
137 | * |
138 | * |
138 | * Assuming interrupts_disable(). |
139 | * Assuming interrupts_disable(). |
139 | * |
140 | * |
140 | */ |
141 | */ |
141 | void main_bsp(void) |
142 | void main_bsp(void) |
142 | { |
143 | { |
143 | config.cpu_count = 1; |
144 | config.cpu_count = 1; |
144 | config.cpu_active = 1; |
145 | config.cpu_active = 1; |
145 | 146 | ||
146 | config.base = hardcoded_load_address; |
147 | config.base = hardcoded_load_address; |
147 | config.kernel_size = ALIGN_UP(hardcoded_ktext_size + |
148 | config.kernel_size = ALIGN_UP(hardcoded_ktext_size + |
148 | hardcoded_kdata_size, PAGE_SIZE); |
149 | hardcoded_kdata_size, PAGE_SIZE); |
149 | config.stack_size = CONFIG_STACK_SIZE; |
150 | config.stack_size = CONFIG_STACK_SIZE; |
150 | 151 | ||
151 | /* Initialy the stack is placed just after the kernel */ |
152 | /* Initialy the stack is placed just after the kernel */ |
152 | config.stack_base = config.base + config.kernel_size; |
153 | config.stack_base = config.base + config.kernel_size; |
153 | 154 | ||
154 | /* Avoid placing stack on top of init */ |
155 | /* Avoid placing stack on top of init */ |
155 | count_t i; |
156 | count_t i; |
156 | for (i = 0; i < init.cnt; i++) { |
157 | for (i = 0; i < init.cnt; i++) { |
157 | if (PA_overlaps(config.stack_base, config.stack_size, |
158 | if (PA_overlaps(config.stack_base, config.stack_size, |
158 | init.tasks[i].addr, init.tasks[i].size)) |
159 | init.tasks[i].addr, init.tasks[i].size)) |
159 | config.stack_base = ALIGN_UP(init.tasks[i].addr + |
160 | config.stack_base = ALIGN_UP(init.tasks[i].addr + |
160 | init.tasks[i].size, config.stack_size); |
161 | init.tasks[i].size, config.stack_size); |
161 | } |
162 | } |
162 | 163 | ||
163 | /* Avoid placing stack on top of boot allocations. */ |
164 | /* Avoid placing stack on top of boot allocations. */ |
164 | if (ballocs.size) { |
165 | if (ballocs.size) { |
165 | if (PA_overlaps(config.stack_base, config.stack_size, |
166 | if (PA_overlaps(config.stack_base, config.stack_size, |
166 | ballocs.base, ballocs.size)) |
167 | ballocs.base, ballocs.size)) |
167 | config.stack_base = ALIGN_UP(ballocs.base + |
168 | config.stack_base = ALIGN_UP(ballocs.base + |
168 | ballocs.size, PAGE_SIZE); |
169 | ballocs.size, PAGE_SIZE); |
169 | } |
170 | } |
170 | 171 | ||
171 | if (config.stack_base < stack_safe) |
172 | if (config.stack_base < stack_safe) |
172 | config.stack_base = ALIGN_UP(stack_safe, PAGE_SIZE); |
173 | config.stack_base = ALIGN_UP(stack_safe, PAGE_SIZE); |
173 | 174 | ||
174 | context_save(&ctx); |
175 | context_save(&ctx); |
175 | context_set(&ctx, FADDR(main_bsp_separated_stack), config.stack_base, |
176 | context_set(&ctx, FADDR(main_bsp_separated_stack), config.stack_base, |
176 | THREAD_STACK_SIZE); |
177 | THREAD_STACK_SIZE); |
177 | context_restore(&ctx); |
178 | context_restore(&ctx); |
178 | /* not reached */ |
179 | /* not reached */ |
179 | } |
180 | } |
180 | 181 | ||
181 | 182 | ||
182 | /** Main kernel routine for bootstrap CPU using new stack. |
183 | /** Main kernel routine for bootstrap CPU using new stack. |
183 | * |
184 | * |
184 | * Second part of main_bsp(). |
185 | * Second part of main_bsp(). |
185 | * |
186 | * |
186 | */ |
187 | */ |
187 | void main_bsp_separated_stack(void) |
188 | void main_bsp_separated_stack(void) |
188 | { |
189 | { |
189 | /* Keep this the first thing. */ |
190 | /* Keep this the first thing. */ |
190 | the_initialize(THE); |
191 | the_initialize(THE); |
191 | 192 | ||
192 | version_print(); |
193 | version_print(); |
193 | 194 | ||
194 | LOG("\nconfig.base=%#" PRIp " config.kernel_size=%" PRIs |
195 | LOG("\nconfig.base=%#" PRIp " config.kernel_size=%" PRIs |
195 | "\nconfig.stack_base=%#" PRIp " config.stack_size=%" PRIs, |
196 | "\nconfig.stack_base=%#" PRIp " config.stack_size=%" PRIs, |
196 | config.base, config.kernel_size, config.stack_base, |
197 | config.base, config.kernel_size, config.stack_base, |
197 | config.stack_size); |
198 | config.stack_size); |
198 | 199 | ||
199 | #ifdef CONFIG_KCONSOLE |
200 | #ifdef CONFIG_KCONSOLE |
200 | /* |
201 | /* |
201 | * kconsole data structures must be initialized very early |
202 | * kconsole data structures must be initialized very early |
202 | * because other subsystems will register their respective |
203 | * because other subsystems will register their respective |
203 | * commands. |
204 | * commands. |
204 | */ |
205 | */ |
205 | LOG_EXEC(kconsole_init()); |
206 | LOG_EXEC(kconsole_init()); |
206 | #endif |
207 | #endif |
207 | 208 | ||
208 | /* |
209 | /* |
209 | * Exception handler initialization, before architecture |
210 | * Exception handler initialization, before architecture |
210 | * starts adding its own handlers |
211 | * starts adding its own handlers |
211 | */ |
212 | */ |
212 | LOG_EXEC(exc_init()); |
213 | LOG_EXEC(exc_init()); |
213 | 214 | ||
214 | /* |
215 | /* |
215 | * Memory management subsystems initialization. |
216 | * Memory management subsystems initialization. |
216 | */ |
217 | */ |
217 | LOG_EXEC(arch_pre_mm_init()); |
218 | LOG_EXEC(arch_pre_mm_init()); |
218 | LOG_EXEC(frame_init()); |
219 | LOG_EXEC(frame_init()); |
219 | 220 | ||
220 | /* Initialize at least 1 memory segment big enough for slab to work. */ |
221 | /* Initialize at least 1 memory segment big enough for slab to work. */ |
221 | LOG_EXEC(slab_cache_init()); |
222 | LOG_EXEC(slab_cache_init()); |
222 | LOG_EXEC(btree_init()); |
223 | LOG_EXEC(btree_init()); |
223 | LOG_EXEC(as_init()); |
224 | LOG_EXEC(as_init()); |
224 | LOG_EXEC(page_init()); |
225 | LOG_EXEC(page_init()); |
225 | LOG_EXEC(tlb_init()); |
226 | LOG_EXEC(tlb_init()); |
226 | LOG_EXEC(ddi_init()); |
227 | LOG_EXEC(ddi_init()); |
227 | LOG_EXEC(tasklet_init()); |
228 | LOG_EXEC(tasklet_init()); |
228 | LOG_EXEC(arch_post_mm_init()); |
229 | LOG_EXEC(arch_post_mm_init()); |
229 | LOG_EXEC(arch_pre_smp_init()); |
230 | LOG_EXEC(arch_pre_smp_init()); |
230 | LOG_EXEC(smp_init()); |
231 | LOG_EXEC(smp_init()); |
231 | 232 | ||
232 | /* Slab must be initialized after we know the number of processors. */ |
233 | /* Slab must be initialized after we know the number of processors. */ |
233 | LOG_EXEC(slab_enable_cpucache()); |
234 | LOG_EXEC(slab_enable_cpucache()); |
234 | 235 | ||
235 | printf("Detected %" PRIc " CPU(s), %" PRIu64" MiB free memory\n", |
236 | printf("Detected %" PRIc " CPU(s), %" PRIu64" MiB free memory\n", |
236 | config.cpu_count, SIZE2MB(zone_total_size())); |
237 | config.cpu_count, SIZE2MB(zone_total_size())); |
237 | 238 | ||
238 | LOG_EXEC(cpu_init()); |
239 | LOG_EXEC(cpu_init()); |
239 | 240 | ||
240 | LOG_EXEC(calibrate_delay_loop()); |
241 | LOG_EXEC(calibrate_delay_loop()); |
241 | LOG_EXEC(clock_counter_init()); |
242 | LOG_EXEC(clock_counter_init()); |
242 | LOG_EXEC(timeout_init()); |
243 | LOG_EXEC(timeout_init()); |
243 | LOG_EXEC(scheduler_init()); |
244 | LOG_EXEC(scheduler_init()); |
244 | LOG_EXEC(task_init()); |
245 | LOG_EXEC(task_init()); |
245 | LOG_EXEC(thread_init()); |
246 | LOG_EXEC(thread_init()); |
246 | LOG_EXEC(futex_init()); |
247 | LOG_EXEC(futex_init()); |
247 | 248 | ||
248 | if (init.cnt > 0) { |
249 | if (init.cnt > 0) { |
249 | count_t i; |
250 | count_t i; |
250 | for (i = 0; i < init.cnt; i++) |
251 | for (i = 0; i < init.cnt; i++) |
251 | LOG("init[%" PRIc "].addr=%#" PRIp ", init[%" PRIc |
252 | LOG("init[%" PRIc "].addr=%#" PRIp ", init[%" PRIc |
252 | "].size=%#" PRIs "\n", i, init.tasks[i].addr, i, |
253 | "].size=%#" PRIs "\n", i, init.tasks[i].addr, i, |
253 | init.tasks[i].size); |
254 | init.tasks[i].size); |
254 | } else |
255 | } else |
255 | printf("No init binaries found\n"); |
256 | printf("No init binaries found\n"); |
256 | 257 | ||
257 | LOG_EXEC(ipc_init()); |
258 | LOG_EXEC(ipc_init()); |
- | 259 | LOG_EXEC(event_init()); |
|
258 | LOG_EXEC(klog_init()); |
260 | LOG_EXEC(klog_init()); |
259 | 261 | ||
260 | #ifdef CONFIG_KCONSOLE |
- | |
261 | LOG_EXEC(kconsole_notify_init()); |
- | |
262 | #endif |
- | |
263 | - | ||
264 | /* |
262 | /* |
265 | * Create kernel task. |
263 | * Create kernel task. |
266 | */ |
264 | */ |
267 | task_t *kernel = task_create(AS_KERNEL, "kernel"); |
265 | task_t *kernel = task_create(AS_KERNEL, "kernel"); |
268 | if (!kernel) |
266 | if (!kernel) |
269 | panic("Cannot create kernel task."); |
267 | panic("Cannot create kernel task."); |
270 | 268 | ||
271 | /* |
269 | /* |
272 | * Create the first thread. |
270 | * Create the first thread. |
273 | */ |
271 | */ |
274 | thread_t *kinit_thread |
272 | thread_t *kinit_thread |
275 | = thread_create(kinit, NULL, kernel, 0, "kinit", true); |
273 | = thread_create(kinit, NULL, kernel, 0, "kinit", true); |
276 | if (!kinit_thread) |
274 | if (!kinit_thread) |
277 | panic("Cannot create kinit thread."); |
275 | panic("Cannot create kinit thread."); |
278 | LOG_EXEC(thread_ready(kinit_thread)); |
276 | LOG_EXEC(thread_ready(kinit_thread)); |
279 | 277 | ||
280 | /* |
278 | /* |
281 | * This call to scheduler() will return to kinit, |
279 | * This call to scheduler() will return to kinit, |
282 | * starting the thread of kernel threads. |
280 | * starting the thread of kernel threads. |
283 | */ |
281 | */ |
284 | scheduler(); |
282 | scheduler(); |
285 | /* not reached */ |
283 | /* not reached */ |
286 | } |
284 | } |
287 | 285 | ||
288 | 286 | ||
289 | #ifdef CONFIG_SMP |
287 | #ifdef CONFIG_SMP |
290 | /** Main kernel routine for application CPUs. |
288 | /** Main kernel routine for application CPUs. |
291 | * |
289 | * |
292 | * Executed by application processors, temporary stack |
290 | * Executed by application processors, temporary stack |
293 | * is at ctx.sp which was set during BSP boot. |
291 | * is at ctx.sp which was set during BSP boot. |
294 | * This function passes control directly to |
292 | * This function passes control directly to |
295 | * main_ap_separated_stack(). |
293 | * main_ap_separated_stack(). |
296 | * |
294 | * |
297 | * Assuming interrupts_disable()'d. |
295 | * Assuming interrupts_disable()'d. |
298 | * |
296 | * |
299 | */ |
297 | */ |
300 | void main_ap(void) |
298 | void main_ap(void) |
301 | { |
299 | { |
302 | /* |
300 | /* |
303 | * Incrementing the active CPU counter will guarantee that the |
301 | * Incrementing the active CPU counter will guarantee that the |
304 | * *_init() functions can find out that they need to |
302 | * *_init() functions can find out that they need to |
305 | * do initialization for AP only. |
303 | * do initialization for AP only. |
306 | */ |
304 | */ |
307 | config.cpu_active++; |
305 | config.cpu_active++; |
308 | 306 | ||
309 | /* |
307 | /* |
310 | * The THE structure is well defined because ctx.sp is used as stack. |
308 | * The THE structure is well defined because ctx.sp is used as stack. |
311 | */ |
309 | */ |
312 | the_initialize(THE); |
310 | the_initialize(THE); |
313 | 311 | ||
314 | arch_pre_mm_init(); |
312 | arch_pre_mm_init(); |
315 | frame_init(); |
313 | frame_init(); |
316 | page_init(); |
314 | page_init(); |
317 | tlb_init(); |
315 | tlb_init(); |
318 | arch_post_mm_init(); |
316 | arch_post_mm_init(); |
319 | 317 | ||
320 | cpu_init(); |
318 | cpu_init(); |
321 | calibrate_delay_loop(); |
319 | calibrate_delay_loop(); |
322 | arch_post_cpu_init(); |
320 | arch_post_cpu_init(); |
323 | 321 | ||
324 | the_copy(THE, (the_t *) CPU->stack); |
322 | the_copy(THE, (the_t *) CPU->stack); |
325 | 323 | ||
326 | /* |
324 | /* |
327 | * If we woke kmp up before we left the kernel stack, we could |
325 | * If we woke kmp up before we left the kernel stack, we could |
328 | * collide with another CPU coming up. To prevent this, we |
326 | * collide with another CPU coming up. To prevent this, we |
329 | * switch to this cpu's private stack prior to waking kmp up. |
327 | * switch to this cpu's private stack prior to waking kmp up. |
330 | */ |
328 | */ |
331 | context_save(&CPU->saved_context); |
329 | context_save(&CPU->saved_context); |
332 | context_set(&CPU->saved_context, FADDR(main_ap_separated_stack), |
330 | context_set(&CPU->saved_context, FADDR(main_ap_separated_stack), |
333 | (uintptr_t) CPU->stack, CPU_STACK_SIZE); |
331 | (uintptr_t) CPU->stack, CPU_STACK_SIZE); |
334 | context_restore(&CPU->saved_context); |
332 | context_restore(&CPU->saved_context); |
335 | /* not reached */ |
333 | /* not reached */ |
336 | } |
334 | } |
337 | 335 | ||
338 | 336 | ||
339 | /** Main kernel routine for application CPUs using new stack. |
337 | /** Main kernel routine for application CPUs using new stack. |
340 | * |
338 | * |
341 | * Second part of main_ap(). |
339 | * Second part of main_ap(). |
342 | * |
340 | * |
343 | */ |
341 | */ |
344 | void main_ap_separated_stack(void) |
342 | void main_ap_separated_stack(void) |
345 | { |
343 | { |
346 | /* |
344 | /* |
347 | * Configure timeouts for this cpu. |
345 | * Configure timeouts for this cpu. |
348 | */ |
346 | */ |
349 | timeout_init(); |
347 | timeout_init(); |
350 | 348 | ||
351 | waitq_wakeup(&ap_completion_wq, WAKEUP_FIRST); |
349 | waitq_wakeup(&ap_completion_wq, WAKEUP_FIRST); |
352 | scheduler(); |
350 | scheduler(); |
353 | /* not reached */ |
351 | /* not reached */ |
354 | } |
352 | } |
355 | #endif /* CONFIG_SMP */ |
353 | #endif /* CONFIG_SMP */ |
356 | 354 | ||
357 | /** @} |
355 | /** @} |
358 | */ |
356 | */ |
359 | 357 |