Subversion Repositories HelenOS-historic

Compare Revisions

Ignore whitespace Rev 694 → Rev 691

/kernel/trunk/arch/amd64/src/mm/memory_init.c
31,9 → 31,8
#include <arch/mm/page.h>
#include <print.h>
 
__u8 e820counter = 0xff;
struct e820memmap_ e820table[MEMMAP_E820_MAX_RECORDS];
__u32 e801memorysize;
__u8 e820counter __attribute__ ((section ("BOOT_DATA"))) = 0xff;
struct e820memmap_ e820table[MEMMAP_E820_MAX_RECORDS] __attribute__ ((section ("BOOT_DATA"))) ;
 
size_t get_memory_size(void)
{
/kernel/trunk/arch/amd64/src/boot/boot.S
33,45 → 33,55
#include <arch/cpuid.h>
#include <arch/boot/boot.h>
 
#define START_STACK (BOOT_OFFSET - BOOT_STACK_SIZE)
#define START_STACK_64 0xffffffff80007c00
#
# This is where we require any SPARTAN-kernel-compatible boot loader
# to pass control in real mode.
#
# Protected mode tables are statically initialised during compile
# time. So we can just load the respective table registers and
# switch to protected mode.
#
 
#define START_STACK (BOOTSTRAP_OFFSET-0x400)
.section K_TEXT_START, "ax"
# .code16
# .global kernel_image_start
# .global multiboot_image_start
# kernel_image_start:
# cli
# xorw %ax,%ax
# movw %ax,%ds
# movw %ax,%es
# movw %ax,%ss # initialize stack segment register
# movl $(START_STACK), %esp # initialize stack pointer
#
# call memmap_arch_init
#
# movl $0x80000000, %eax
# cpuid
# cmp $0x80000000, %eax # any function > 80000000h?
# jbe no_long_mode
# movl $(AMD_CPUID_EXTENDED), %eax # Extended function code 80000001
# cpuid
# bt $29, %edx # Test if long mode is supported.
# jnc no_long_mode
#
# # Load gdtr, idtr
# lgdt real_bootstrap_gdtr_boot
#
# movl %cr0,%eax
# orl $0x1,%eax
# movl %eax,%cr0 # switch to protected mode
#
# jmpl $gdtselector(KTEXT32_DES), $now_in_prot
#
# no_long_mode:
# 1:
# jmp 1b
#
.code16
.global kernel_image_start
.global multiboot_image_start
kernel_image_start:
cli
xorw %ax,%ax
movw %ax,%ds
movw %ax,%es
movw %ax,%ss # initialize stack segment register
movl $(START_STACK), %esp # initialize stack pointer
call memmap_arch_init
movl $0x80000000, %eax
cpuid
cmp $0x80000000, %eax # any function > 80000000h?
jbe no_long_mode
movl $(AMD_CPUID_EXTENDED), %eax # Extended function code 80000001
cpuid
bt $29, %edx # Test if long mode is supported.
jnc no_long_mode
 
# Load gdtr, idtr
lgdt real_bootstrap_gdtr_boot
movl %cr0,%eax
orl $0x1,%eax
movl %eax,%cr0 # switch to protected mode
 
jmpl $gdtselector(KTEXT32_DES), $now_in_prot
 
no_long_mode:
1:
jmp 1b
.code32
.align 4
multiboot_header:
78,11 → 88,11
.long MULTIBOOT_HEADER_MAGIC
.long MULTIBOOT_HEADER_FLAGS
.long -(MULTIBOOT_HEADER_MAGIC + MULTIBOOT_HEADER_FLAGS) # checksum
.long multiboot_header
.long unmapped_ktext_start
.long multiboot_header + BOOT_OFFSET
.long unmapped_ktext_start + BOOT_OFFSET
.long 0
.long 0
.long multiboot_image_start
.long multiboot_image_start + BOOT_OFFSET
 
multiboot_image_start:
movl $START_STACK, %esp # initialize stack pointer
99,39 → 109,44
jmpl $gdtselector(KTEXT_DES), $multiboot_meeting_point + BOOT_OFFSET
multiboot_meeting_point:
# Protected 32-bit. We want to reuse the code-seg descriptor,
# the Default operand size must not be 1 when entering long mode
hlt
 
# Protected 32-bit. We want to reuse the code-seg descriptor,
# the Default operand size must not be 1 when entering long mode
now_in_prot:
# Set up stack & data descriptors
movw $gdtselector(KDATA_DES), %ax
movw %ax, %ds
movw %ax, %ss
 
movb $0xd1, %al # enable A20 using the keyboard controller
outb %al, $0x64
movb $0xdf, %al
outb %al, $0x60
 
pushl %ebx # save parameters from GRUB
pushl %eax
# Enable 64-bit page transaltion entries - CR4.PAE = 1.
# Paging is not enabled until after long mode is enabled
movl %cr4, %eax
btsl $5, %eax
movl %eax, %cr4
 
# Set up paging tables
leal ptl_0, %eax
movl %eax, %cr3
# Enable long mode
movl $EFER_MSR_NUM, %ecx # EFER MSR number
rdmsr # Read EFER
btsl $AMD_LME_FLAG, %eax # Set LME=1
wrmsr # Write EFER
movl $EFER_MSR_NUM, %ecx # EFER MSR number
rdmsr # Read EFER
btsl $AMD_LME_FLAG, %eax # Set LME=1
wrmsr # Write EFER
# Enable paging to activate long mode (set CR0.PG=1)
movl %cr0, %eax
btsl $31, %eax
movl %eax, %cr0
# At this point we are in compatibility mode
jmpl $gdtselector(KTEXT_DES), $start64
 
.code64
138,10 → 153,16
start64:
movq $(PA2KA(START_STACK)), %rsp
 
# Copy kernel to higher physical memory
movq $BOOTSTRAP_OFFSET, %rsi
movq $BOOTSTRAP_OFFSET + BOOT_OFFSET, %rdi
movq $_hardcoded_kernel_size, %rcx
cld
rep movsb
call main_bsp # never returns
cli
hlt
1:
jmp 1b
.section K_DATA_START, "aw", @progbits
.align 4096
197,6 → 218,11
.fill 510,8,0
.quad ptl_1 + (PTL_WRITABLE | PTL_PRESENT)
 
.global real_bootstrap_gdtr_boot
real_bootstrap_gdtr_boot:
.word gdtselector(GDT_ITEMS)
.long KA2PA(gdt)-BOOT_OFFSET
 
.global protected_bootstrap_gdtr
protected_bootstrap_gdtr:
.word gdtselector(GDT_ITEMS)
/kernel/trunk/arch/amd64/src/boot/memmap.S
0,0 → 1,115
/*
* Copyright (C) 2005 Josef Cejka
* All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
*
* - Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* - Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
* - The name of the author may not be used to endorse or promote products
* derived from this software without specific prior written permission.
*
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
* IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
* OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
* IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
* INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
* NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
* DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
* THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
* THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
 
 
#include <arch/boot/memmap.h>
 
E820_SMAP = 0x534d4150
 
.global memmap_arch_init
.global e801memorysize
 
.code16
.section K_TEXT_START_2, "ax"
 
memmap_arch_init:
e820begin:
xorl %ebx,%ebx # during first call, ebx must be 0
movw $e820table_boot,%di
movb $MEMMAP_E820_MAX_RECORDS,e820counter_boot
e820loop:
movl $E820_SMAP,%edx # control sequence "SMAP"
 
movl $0x0000e820,%eax # service
movl $MEMMAP_E820_RECORD_SIZE,%ecx
int $0x15
jc e820err
cmpl $E820_SMAP,%eax # verifying BIOS
jne e820err
 
cmpl $MEMMAP_E820_RECORD_SIZE,%ecx
jne e820err # bad record size - bug in bios
movw %di,%ax # next record
addw $MEMMAP_E820_RECORD_SIZE,%ax
movw %ax,%di
decb e820counter_boot # buffer is full
jz e820end
testl %ebx,%ebx
jnz e820loop
e820end:
movb $MEMMAP_E820_MAX_RECORDS,%al
subb e820counter_boot,%al
movb %al,e820counter_boot # store # of valid entries in e820counter
 
jmp e801begin
 
e820err:
movb $0,e820counter_boot
 
# method e801 - get size of memory
 
e801begin:
xorw %dx,%dx
xorw %cx,%cx
xorw %bx,%bx
movw $0xe801,%ax
stc
int $0x15
jc e801end
# fix problem with some BIOSes which use ax:bx rather than cx:dx
testw %cx,%cx
jnz e801cxdx
testw %dx,%dx
jnz e801cxdx
 
movw %ax,%cx
movw %bx,%dx
e801cxdx:
andl $0xffff,%edx
shll $6,%edx
andl $0xffff,%ecx
addl %ecx,%edx
addl $0x0400,%edx # add lower 1 MB - it's not included by e801 method
movl %edx,e801memorysize
e801end:
ret
 
 
.section K_DATA_START, "aw", @progbits
 
#memory size in 1 kb chunks
e801memorysize:
.long 0