Subversion Repositories HelenOS-historic

Rev

Rev 1315 | Rev 1434 | Go to most recent revision | Only display areas with differences | Ignore whitespace | Details | Blame | Last modification | View Log | RSS feed

Rev 1315 Rev 1389
1
/*
1
/*
2
 * Copyright (C) 2001-2004 Jakub Jermar
2
 * Copyright (C) 2001-2004 Jakub Jermar
3
 * All rights reserved.
3
 * All rights reserved.
4
 *
4
 *
5
 * Redistribution and use in source and binary forms, with or without
5
 * Redistribution and use in source and binary forms, with or without
6
 * modification, are permitted provided that the following conditions
6
 * modification, are permitted provided that the following conditions
7
 * are met:
7
 * are met:
8
 *
8
 *
9
 * - Redistributions of source code must retain the above copyright
9
 * - Redistributions of source code must retain the above copyright
10
 *   notice, this list of conditions and the following disclaimer.
10
 *   notice, this list of conditions and the following disclaimer.
11
 * - Redistributions in binary form must reproduce the above copyright
11
 * - Redistributions in binary form must reproduce the above copyright
12
 *   notice, this list of conditions and the following disclaimer in the
12
 *   notice, this list of conditions and the following disclaimer in the
13
 *   documentation and/or other materials provided with the distribution.
13
 *   documentation and/or other materials provided with the distribution.
14
 * - The name of the author may not be used to endorse or promote products
14
 * - The name of the author may not be used to endorse or promote products
15
 *   derived from this software without specific prior written permission.
15
 *   derived from this software without specific prior written permission.
16
 *
16
 *
17
 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
17
 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
18
 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
18
 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
19
 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
19
 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
20
 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
20
 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
21
 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
21
 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
22
 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
22
 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
23
 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
23
 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
24
 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
24
 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
25
 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
25
 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
26
 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
26
 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
27
 */
27
 */
28
 
28
 
29
/**
29
/**
30
 * @file    main.c
30
 * @file    main.c
31
 * @brief   Main initialization kernel function for all processors.
31
 * @brief   Main initialization kernel function for all processors.
32
 *
32
 *
33
 * During kernel boot, all processors, after architecture dependent
33
 * During kernel boot, all processors, after architecture dependent
34
 * initialization, start executing code found in this file. After
34
 * initialization, start executing code found in this file. After
35
 * bringing up all subsystems, control is passed to scheduler().
35
 * bringing up all subsystems, control is passed to scheduler().
36
 *
36
 *
37
 * The bootstrap processor starts executing main_bsp() while
37
 * The bootstrap processor starts executing main_bsp() while
38
 * the application processors start executing main_ap().
38
 * the application processors start executing main_ap().
39
 *
39
 *
40
 * @see scheduler()
40
 * @see scheduler()
41
 * @see main_bsp()
41
 * @see main_bsp()
42
 * @see main_ap()
42
 * @see main_ap()
43
 */
43
 */
44
 
44
 
45
#include <arch/asm.h>
45
#include <arch/asm.h>
46
#include <context.h>
46
#include <context.h>
47
#include <print.h>
47
#include <print.h>
48
#include <panic.h>
48
#include <panic.h>
49
#include <debug.h>
49
#include <debug.h>
50
#include <config.h>
50
#include <config.h>
51
#include <time/clock.h>
51
#include <time/clock.h>
52
#include <proc/scheduler.h>
52
#include <proc/scheduler.h>
53
#include <proc/thread.h>
53
#include <proc/thread.h>
54
#include <proc/task.h>
54
#include <proc/task.h>
55
#include <main/kinit.h>
55
#include <main/kinit.h>
56
#include <main/version.h>
56
#include <main/version.h>
57
#include <console/kconsole.h>
57
#include <console/kconsole.h>
58
#include <cpu.h>
58
#include <cpu.h>
59
#include <align.h>
59
#include <align.h>
60
#include <interrupt.h>
60
#include <interrupt.h>
61
#include <arch/mm/memory_init.h>
61
#include <arch/mm/memory_init.h>
62
#include <mm/frame.h>
62
#include <mm/frame.h>
63
#include <mm/page.h>
63
#include <mm/page.h>
64
#include <genarch/mm/page_pt.h>
64
#include <genarch/mm/page_pt.h>
65
#include <mm/tlb.h>
65
#include <mm/tlb.h>
66
#include <mm/as.h>
66
#include <mm/as.h>
67
#include <mm/slab.h>
67
#include <mm/slab.h>
68
#include <synch/waitq.h>
68
#include <synch/waitq.h>
69
#include <synch/futex.h>
69
#include <synch/futex.h>
70
#include <arch/arch.h>
70
#include <arch/arch.h>
71
#include <arch.h>
71
#include <arch.h>
72
#include <arch/faddr.h>
72
#include <arch/faddr.h>
73
#include <typedefs.h>
73
#include <typedefs.h>
74
#include <ipc/ipc.h>
74
#include <ipc/ipc.h>
75
#include <macros.h>
75
#include <macros.h>
76
#include <adt/btree.h>
76
#include <adt/btree.h>
77
 
77
 
78
#ifdef CONFIG_SMP
78
#ifdef CONFIG_SMP
79
#include <arch/smp/apic.h>
79
#include <arch/smp/apic.h>
80
#include <arch/smp/mps.h>
80
#include <arch/smp/mps.h>
81
#endif /* CONFIG_SMP */
81
#endif /* CONFIG_SMP */
82
#include <smp/smp.h>
82
#include <smp/smp.h>
83
 
83
 
84
/** Global configuration structure. */
84
/** Global configuration structure. */
85
config_t config = {
85
config_t config = {
86
    .mm_initialized = false
86
    .mm_initialized = false
87
};
87
};
88
 
88
 
89
/** Initial user-space tasks */
89
/** Initial user-space tasks */
90
init_t init = {
90
init_t init = {
91
    0
91
    0
92
};
92
};
93
 
93
 
94
context_t ctx;
94
context_t ctx;
95
 
95
 
96
/**
96
/**
97
 * These 'hardcoded' variables will be intialized by
97
 * These 'hardcoded' variables will be intialized by
98
 * the linker or the low level assembler code with
98
 * the linker or the low level assembler code with
99
 * appropriate sizes and addresses.
99
 * appropriate sizes and addresses.
100
 */
100
 */
101
__address hardcoded_load_address = 0;
101
__address hardcoded_load_address = 0;
102
size_t hardcoded_ktext_size = 0;
102
size_t hardcoded_ktext_size = 0;
103
size_t hardcoded_kdata_size = 0;
103
size_t hardcoded_kdata_size = 0;
104
 
104
 
105
void main_bsp(void);
105
void main_bsp(void);
106
void main_ap(void);
106
void main_ap(void);
107
 
107
 
108
/*
108
/*
109
 * These two functions prevent stack from underflowing during the
109
 * These two functions prevent stack from underflowing during the
110
 * kernel boot phase when SP is set to the very top of the reserved
110
 * kernel boot phase when SP is set to the very top of the reserved
111
 * space. The stack could get corrupted by a fooled compiler-generated
111
 * space. The stack could get corrupted by a fooled compiler-generated
112
 * pop sequence otherwise.
112
 * pop sequence otherwise.
113
 */
113
 */
114
static void main_bsp_separated_stack(void);
114
static void main_bsp_separated_stack(void);
115
#ifdef CONFIG_SMP
115
#ifdef CONFIG_SMP
116
static void main_ap_separated_stack(void);
116
static void main_ap_separated_stack(void);
117
#endif
117
#endif
118
 
118
 
119
#define CONFIG_STACK_SIZE   ((1<<STACK_FRAMES)*STACK_SIZE)
119
#define CONFIG_STACK_SIZE   ((1<<STACK_FRAMES)*STACK_SIZE)
120
 
120
 
121
/** Main kernel routine for bootstrap CPU.
121
/** Main kernel routine for bootstrap CPU.
122
 *
122
 *
123
 * Initializes the kernel by bootstrap CPU.
123
 * Initializes the kernel by bootstrap CPU.
124
 * This function passes control directly to
124
 * This function passes control directly to
125
 * main_bsp_separated_stack().
125
 * main_bsp_separated_stack().
126
 *
126
 *
127
 * Assuming interrupts_disable().
127
 * Assuming interrupts_disable().
128
 *
128
 *
129
 */
129
 */
130
void main_bsp(void)
130
void main_bsp(void)
131
{
131
{
132
    __address stackaddr;
132
    __address stackaddr;
133
 
133
 
134
    config.cpu_count = 1;
134
    config.cpu_count = 1;
135
    config.cpu_active = 1;
135
    config.cpu_active = 1;
136
   
136
   
137
    config.base = hardcoded_load_address;
137
    config.base = hardcoded_load_address;
138
    config.memory_size = get_memory_size();
138
    config.memory_size = get_memory_size();
139
   
139
   
140
    config.kernel_size = ALIGN_UP(hardcoded_ktext_size + hardcoded_kdata_size, PAGE_SIZE);
140
    config.kernel_size = ALIGN_UP(hardcoded_ktext_size + hardcoded_kdata_size, PAGE_SIZE);
141
    stackaddr = config.base + config.kernel_size;
141
    stackaddr = config.base + config.kernel_size;
142
   
142
   
143
    /* Avoid placing kernel on top of init */
143
    /* Avoid placing kernel on top of init */
144
    count_t i;
144
    count_t i;
145
    bool overlap = false;
145
    bool overlap = false;
146
    for (i = 0; i < init.cnt; i++)
146
    for (i = 0; i < init.cnt; i++)
147
        if (PA_overlaps(stackaddr, CONFIG_STACK_SIZE, init.tasks[i].addr, init.tasks[i].size)) {
147
        if (PA_overlaps(stackaddr, CONFIG_STACK_SIZE, init.tasks[i].addr, init.tasks[i].size)) {
148
            stackaddr = ALIGN_UP(init.tasks[i].addr + init.tasks[i].size, CONFIG_STACK_SIZE);
148
            stackaddr = ALIGN_UP(init.tasks[i].addr + init.tasks[i].size, CONFIG_STACK_SIZE);
149
            init.tasks[i].size = ALIGN_UP(init.tasks[i].size, CONFIG_STACK_SIZE) + CONFIG_STACK_SIZE;
149
            init.tasks[i].size = ALIGN_UP(init.tasks[i].size, CONFIG_STACK_SIZE) + CONFIG_STACK_SIZE;
150
            overlap = true;
150
            overlap = true;
151
        }
151
        }
152
   
152
   
153
    if (!overlap)
153
    if (!overlap)
154
        config.kernel_size += CONFIG_STACK_SIZE;
154
        config.kernel_size += CONFIG_STACK_SIZE;
155
   
155
   
156
    context_save(&ctx);
156
    context_save(&ctx);
157
    context_set(&ctx, FADDR(main_bsp_separated_stack), stackaddr, THREAD_STACK_SIZE);
157
    context_set(&ctx, FADDR(main_bsp_separated_stack), stackaddr, THREAD_STACK_SIZE);
158
    context_restore(&ctx);
158
    context_restore(&ctx);
159
    /* not reached */
159
    /* not reached */
160
}
160
}
161
 
161
 
162
 
162
 
163
/** Main kernel routine for bootstrap CPU using new stack.
163
/** Main kernel routine for bootstrap CPU using new stack.
164
 *
164
 *
165
 * Second part of main_bsp().
165
 * Second part of main_bsp().
166
 *
166
 *
167
 */
167
 */
168
void main_bsp_separated_stack(void)
168
void main_bsp_separated_stack(void)
169
{
169
{
170
    task_t *k;
170
    task_t *k;
171
    thread_t *t;
171
    thread_t *t;
172
    count_t i;
172
    count_t i;
173
   
173
   
174
    the_initialize(THE);
174
    the_initialize(THE);
175
 
175
 
176
    /*
176
    /*
177
     * kconsole data structures must be initialized very early
177
     * kconsole data structures must be initialized very early
178
     * because other subsystems will register their respective
178
     * because other subsystems will register their respective
179
     * commands.
179
     * commands.
180
     */
180
     */
181
    kconsole_init();
181
    kconsole_init();
182
   
182
   
183
    /*
183
    /*
184
     * Exception handler initialization, before architecture
184
     * Exception handler initialization, before architecture
185
     * starts adding its own handlers
185
     * starts adding its own handlers
186
     */
186
     */
187
    exc_init();
187
    exc_init();
188
 
188
 
189
    /*
189
    /*
190
     * Memory management subsystems initialization.
190
     * Memory management subsystems initialization.
191
     */
191
     */
192
    arch_pre_mm_init();
192
    arch_pre_mm_init();
193
    frame_init();       /* Initialize at least 1 memory segment big enough for slab to work */
193
    frame_init();       /* Initialize at least 1 memory segment big enough for slab to work */
194
    slab_cache_init();
194
    slab_cache_init();
195
    btree_init();
195
    btree_init();
196
    as_init();
196
    as_init();
197
    page_init();
197
    page_init();
198
    tlb_init();
198
    tlb_init();
199
    config.mm_initialized = true;
199
    config.mm_initialized = true;
200
    arch_post_mm_init();   
200
    arch_post_mm_init();   
201
 
201
 
202
    version_print();
202
    version_print();
203
    printf("%.*p: hardcoded_ktext_size=%zdK, hardcoded_kdata_size=%zdK\n", sizeof(__address) * 2, config.base, hardcoded_ktext_size / 1024, hardcoded_kdata_size / 1024);
203
    printf("%.*p: hardcoded_ktext_size=%zdK, hardcoded_kdata_size=%zdK\n", sizeof(__address) * 2, config.base, hardcoded_ktext_size >> 10, hardcoded_kdata_size >> 10);
204
 
204
 
205
    arch_pre_smp_init();
205
    arch_pre_smp_init();
206
    smp_init();
206
    smp_init();
207
   
207
   
208
    slab_enable_cpucache(); /* Slab must be initialized AFTER we know the number of processors */
208
    slab_enable_cpucache(); /* Slab must be initialized AFTER we know the number of processors */
209
 
209
 
210
    printf("config.memory_size=%zdM\n", config.memory_size/(1024*1024));
210
    printf("config.memory_size=%zdM\n", config.memory_size >> 20);
211
    printf("config.cpu_count=%zd\n", config.cpu_count);
211
    printf("config.cpu_count=%zd\n", config.cpu_count);
212
    cpu_init();
212
    cpu_init();
213
   
213
   
214
    calibrate_delay_loop();
214
    calibrate_delay_loop();
215
    timeout_init();
215
    timeout_init();
216
    scheduler_init();
216
    scheduler_init();
217
    task_init();
217
    task_init();
218
    thread_init();
218
    thread_init();
219
    futex_init();
219
    futex_init();
220
   
220
   
221
    for (i = 0; i < init.cnt; i++)
221
    for (i = 0; i < init.cnt; i++)
222
        printf("init[%zd].addr=%.*p, init[%zd].size=%zd\n", i, sizeof(__address)*2, init.tasks[i].addr, i, init.tasks[i].size);
222
        printf("init[%zd].addr=%.*p, init[%zd].size=%zd\n", i, sizeof(__address) * 2, init.tasks[i].addr, i, init.tasks[i].size);
223
   
223
   
224
    ipc_init();
224
    ipc_init();
225
 
225
 
226
    /*
226
    /*
227
     * Create kernel task.
227
     * Create kernel task.
228
     */
228
     */
229
    k = task_create(AS_KERNEL, "KERNEL");
229
    k = task_create(AS_KERNEL, "KERNEL");
230
    if (!k)
230
    if (!k)
231
        panic("can't create kernel task\n");
231
        panic("can't create kernel task\n");
232
   
232
   
233
    /*
233
    /*
234
     * Create the first thread.
234
     * Create the first thread.
235
     */
235
     */
236
    t = thread_create(kinit, NULL, k, 0, "kinit");
236
    t = thread_create(kinit, NULL, k, 0, "kinit");
237
    if (!t)
237
    if (!t)
238
        panic("can't create kinit thread\n");
238
        panic("can't create kinit thread\n");
239
    thread_ready(t);
239
    thread_ready(t);
240
   
240
   
241
    /*
241
    /*
242
     * This call to scheduler() will return to kinit,
242
     * This call to scheduler() will return to kinit,
243
     * starting the thread of kernel threads.
243
     * starting the thread of kernel threads.
244
     */
244
     */
245
    scheduler();
245
    scheduler();
246
    /* not reached */
246
    /* not reached */
247
}
247
}
248
 
248
 
249
 
249
 
250
#ifdef CONFIG_SMP
250
#ifdef CONFIG_SMP
251
/** Main kernel routine for application CPUs.
251
/** Main kernel routine for application CPUs.
252
 *
252
 *
253
 * Executed by application processors, temporary stack
253
 * Executed by application processors, temporary stack
254
 * is at ctx.sp which was set during BP boot.
254
 * is at ctx.sp which was set during BP boot.
255
 * This function passes control directly to
255
 * This function passes control directly to
256
 * main_ap_separated_stack().
256
 * main_ap_separated_stack().
257
 *
257
 *
258
 * Assuming interrupts_disable()'d.
258
 * Assuming interrupts_disable()'d.
259
 *
259
 *
260
 */
260
 */
261
void main_ap(void)
261
void main_ap(void)
262
{
262
{
263
    /*
263
    /*
264
     * Incrementing the active CPU counter will guarantee that the
264
     * Incrementing the active CPU counter will guarantee that the
265
     * pm_init() will not attempt to build GDT and IDT tables again.
265
     * pm_init() will not attempt to build GDT and IDT tables again.
266
     * Neither frame_init() will do the complete thing. Neither cpu_init()
266
     * Neither frame_init() will do the complete thing. Neither cpu_init()
267
     * will do.
267
     * will do.
268
     */
268
     */
269
    config.cpu_active++;
269
    config.cpu_active++;
270
 
270
 
271
    /*
271
    /*
272
     * The THE structure is well defined because ctx.sp is used as stack.
272
     * The THE structure is well defined because ctx.sp is used as stack.
273
     */
273
     */
274
    the_initialize(THE);
274
    the_initialize(THE);
275
   
275
   
276
    arch_pre_mm_init();
276
    arch_pre_mm_init();
277
    frame_init();
277
    frame_init();
278
    page_init();
278
    page_init();
279
    tlb_init();
279
    tlb_init();
280
    arch_post_mm_init();
280
    arch_post_mm_init();
281
   
281
   
282
    cpu_init();
282
    cpu_init();
283
   
283
   
284
    calibrate_delay_loop();
284
    calibrate_delay_loop();
285
 
285
 
286
    l_apic_init();
286
    l_apic_init();
287
    l_apic_debug();
287
    l_apic_debug();
288
 
288
 
289
    the_copy(THE, (the_t *) CPU->stack);
289
    the_copy(THE, (the_t *) CPU->stack);
290
 
290
 
291
    /*
291
    /*
292
     * If we woke kmp up before we left the kernel stack, we could
292
     * If we woke kmp up before we left the kernel stack, we could
293
     * collide with another CPU coming up. To prevent this, we
293
     * collide with another CPU coming up. To prevent this, we
294
     * switch to this cpu's private stack prior to waking kmp up.
294
     * switch to this cpu's private stack prior to waking kmp up.
295
     */
295
     */
296
    context_set(&CPU->saved_context, FADDR(main_ap_separated_stack), (__address) CPU->stack, CPU_STACK_SIZE);
296
    context_set(&CPU->saved_context, FADDR(main_ap_separated_stack), (__address) CPU->stack, CPU_STACK_SIZE);
297
    context_restore(&CPU->saved_context);
297
    context_restore(&CPU->saved_context);
298
    /* not reached */
298
    /* not reached */
299
}
299
}
300
 
300
 
301
 
301
 
302
/** Main kernel routine for application CPUs using new stack.
302
/** Main kernel routine for application CPUs using new stack.
303
 *
303
 *
304
 * Second part of main_ap().
304
 * Second part of main_ap().
305
 *
305
 *
306
 */
306
 */
307
void main_ap_separated_stack(void)
307
void main_ap_separated_stack(void)
308
{
308
{
309
    /*
309
    /*
310
     * Configure timeouts for this cpu.
310
     * Configure timeouts for this cpu.
311
     */
311
     */
312
    timeout_init();
312
    timeout_init();
313
 
313
 
314
    waitq_wakeup(&ap_completion_wq, WAKEUP_FIRST);
314
    waitq_wakeup(&ap_completion_wq, WAKEUP_FIRST);
315
    scheduler();
315
    scheduler();
316
    /* not reached */
316
    /* not reached */
317
}
317
}
318
#endif /* CONFIG_SMP */
318
#endif /* CONFIG_SMP */
319
 
319