Rev 4344 | Go to most recent revision | Show entire file | Ignore whitespace | Details | Blame | Last modification | View Log | RSS feed
| Rev 4344 | Rev 4346 | ||
|---|---|---|---|
| Line 1... | Line 1... | ||
| 1 | /* |
1 | /* |
| 2 | * Copyright (c) 2005 Jakub Jermar |
2 | * Copyright (c) 2009 Jakub Jermar |
| 3 | * All rights reserved. |
3 | * All rights reserved. |
| 4 | * |
4 | * |
| 5 | * Redistribution and use in source and binary forms, with or without |
5 | * Redistribution and use in source and binary forms, with or without |
| 6 | * modification, are permitted provided that the following conditions |
6 | * modification, are permitted provided that the following conditions |
| 7 | * are met: |
7 | * are met: |
| Line 40... | Line 40... | ||
| 40 | #include <bitops.h> |
40 | #include <bitops.h> |
| 41 | #include <debug.h> |
41 | #include <debug.h> |
| 42 | #include <align.h> |
42 | #include <align.h> |
| 43 | #include <config.h> |
43 | #include <config.h> |
| 44 | 44 | ||
| 45 | #ifdef CONFIG_SMP |
- | |
| 46 | /** Entries locked in DTLB of BSP. |
- | |
| 47 | * |
- | |
| 48 | * Application processors need to have the same locked entries in their DTLBs as |
- | |
| 49 | * the bootstrap processor. |
- | |
| 50 | */ |
- | |
| 51 | static struct { |
- | |
| 52 | uintptr_t virt_page; |
- | |
| 53 | uintptr_t phys_page; |
- | |
| 54 | int pagesize_code; |
- | |
| 55 | } bsp_locked_dtlb_entry[DTLB_MAX_LOCKED_ENTRIES]; |
- | |
| 56 | - | ||
| 57 | /** Number of entries in bsp_locked_dtlb_entry array. */ |
- | |
| 58 | static count_t bsp_locked_dtlb_entries = 0; |
- | |
| 59 | #endif /* CONFIG_SMP */ |
- | |
| 60 | - | ||
| 61 | /** Perform sparc64 specific initialization of paging. */ |
45 | /** Perform sparc64 specific initialization of paging. */ |
| 62 | void page_arch_init(void) |
46 | void page_arch_init(void) |
| 63 | { |
47 | { |
| 64 | if (config.cpu_active == 1) { |
48 | if (config.cpu_active == 1) |
| 65 | page_mapping_operations = &ht_mapping_operations; |
49 | page_mapping_operations = &ht_mapping_operations; |
| 66 | } else { |
- | |
| 67 | - | ||
| 68 | #ifdef CONFIG_SMP |
- | |
| 69 | unsigned int i; |
- | |
| 70 | - | ||
| 71 | /* |
- | |
| 72 | * Copy locked DTLB entries from the BSP. |
- | |
| 73 | */ |
- | |
| 74 | for (i = 0; i < bsp_locked_dtlb_entries; i++) { |
- | |
| 75 | dtlb_insert_mapping(bsp_locked_dtlb_entry[i].virt_page, |
- | |
| 76 | bsp_locked_dtlb_entry[i].phys_page, |
- | |
| 77 | bsp_locked_dtlb_entry[i].pagesize_code, true, |
- | |
| 78 | false); |
- | |
| 79 | } |
- | |
| 80 | #endif |
- | |
| 81 | - | ||
| 82 | } |
- | |
| 83 | } |
50 | } |
| 84 | 51 | ||
| 85 | /** Map memory-mapped device into virtual memory. |
52 | /** Map memory-mapped device into virtual memory. |
| 86 | * |
53 | * |
| 87 | * So far, only DTLB is used to map devices into memory. Chances are that there |
54 | * We are currently using identity mapping for mapping device registers. |
| 88 | * will be only a limited amount of devices that the kernel itself needs to |
- | |
| 89 | * lock in DTLB. |
- | |
| 90 | * |
55 | * |
| 91 | * @param physaddr Physical address of the page where the device is located. |
56 | * @param physaddr Physical address of the page where the device is |
| 92 | * Must be at least page-aligned. |
57 | * located. |
| 93 | * @param size Size of the device's registers. Must not exceed 4M and must |
58 | * @param size Size of the device's registers. This argument is |
| 94 | * include extra space caused by the alignment. |
59 | * ignored. |
| 95 | * |
60 | * |
| 96 | * @return Virtual address of the page where the device is mapped. |
61 | * @return Virtual address of the page where the device is mapped. |
| 97 | */ |
62 | */ |
| 98 | uintptr_t hw_map(uintptr_t physaddr, size_t size) |
63 | uintptr_t hw_map(uintptr_t physaddr, size_t size) |
| 99 | { |
64 | { |
| 100 | unsigned int order; |
- | |
| 101 | unsigned int i; |
- | |
| 102 | - | ||
| 103 | ASSERT(config.cpu_active == 1); |
- | |
| 104 | - | ||
| 105 | struct { |
- | |
| 106 | int pagesize_code; |
- | |
| 107 | size_t increment; |
- | |
| 108 | count_t count; |
- | |
| 109 | } sizemap[] = { |
- | |
| 110 | { PAGESIZE_8K, 0, 1 }, /* 8K */ |
- | |
| 111 | { PAGESIZE_8K, MMU_PAGE_SIZE, 2 }, /* 16K */ |
- | |
| 112 | { PAGESIZE_8K, MMU_PAGE_SIZE, 4 }, /* 32K */ |
- | |
| 113 | { PAGESIZE_64K, 0, 1}, /* 64K */ |
- | |
| 114 | { PAGESIZE_64K, 8 * MMU_PAGE_SIZE, 2 }, /* 128K */ |
- | |
| 115 | { PAGESIZE_64K, 8 * MMU_PAGE_SIZE, 4 }, /* 256K */ |
- | |
| 116 | { PAGESIZE_512K, 0, 1 }, /* 512K */ |
- | |
| 117 | { PAGESIZE_512K, 64 * MMU_PAGE_SIZE, 2 }, /* 1M */ |
- | |
| 118 | { PAGESIZE_512K, 64 * MMU_PAGE_SIZE, 4 }, /* 2M */ |
- | |
| 119 | { PAGESIZE_4M, 0, 1 }, /* 4M */ |
- | |
| 120 | { PAGESIZE_4M, 512 * MMU_PAGE_SIZE, 2 } /* 8M */ |
- | |
| 121 | }; |
- | |
| 122 | - | ||
| 123 | ASSERT(ALIGN_UP(physaddr, MMU_PAGE_SIZE) == physaddr); |
- | |
| 124 | ASSERT(size <= 8 * 1024 * 1024); |
- | |
| 125 | - | ||
| 126 | if (size <= MMU_FRAME_SIZE) |
- | |
| 127 | order = 0; |
- | |
| 128 | else |
- | |
| 129 | order = (fnzb64(size - 1) + 1) - MMU_FRAME_WIDTH; |
- | |
| 130 | - | ||
| 131 | /* |
- | |
| 132 | * Use virtual addresses that are beyond the limit of physical memory. |
- | |
| 133 | * Thus, the physical address space will not be wasted by holes created |
- | |
| 134 | * by frame_alloc(). |
- | |
| 135 | */ |
- | |
| 136 | ASSERT(PA2KA(last_frame)); |
- | |
| 137 | uintptr_t virtaddr = ALIGN_UP(PA2KA(last_frame), |
- | |
| 138 | 1 << (order + FRAME_WIDTH)); |
- | |
| 139 | last_frame = ALIGN_UP(KA2PA(virtaddr) + size, |
- | |
| 140 | 1 << (order + FRAME_WIDTH)); |
- | |
| 141 | - | ||
| 142 | for (i = 0; i < sizemap[order].count; i++) { |
- | |
| 143 | /* |
- | |
| 144 | * First, insert the mapping into DTLB. |
- | |
| 145 | */ |
- | |
| 146 | dtlb_insert_mapping(virtaddr + i * sizemap[order].increment, |
- | |
| 147 | physaddr + i * sizemap[order].increment, |
- | |
| 148 | sizemap[order].pagesize_code, true, false); |
- | |
| 149 | - | ||
| 150 | #ifdef CONFIG_SMP |
- | |
| 151 | /* |
- | |
| 152 | * Second, save the information about the mapping for APs. |
- | |
| 153 | */ |
- | |
| 154 | bsp_locked_dtlb_entry[bsp_locked_dtlb_entries].virt_page = |
- | |
| 155 | virtaddr + i * sizemap[order].increment; |
- | |
| 156 | bsp_locked_dtlb_entry[bsp_locked_dtlb_entries].phys_page = |
- | |
| 157 | physaddr + i * sizemap[order].increment; |
- | |
| 158 | bsp_locked_dtlb_entry[bsp_locked_dtlb_entries].pagesize_code = |
- | |
| 159 | sizemap[order].pagesize_code; |
- | |
| 160 | bsp_locked_dtlb_entries++; |
- | |
| 161 | #endif |
- | |
| 162 | } |
- | |
| 163 | - | ||
| 164 | return virtaddr; |
65 | return PA2KA(physaddr); |
| 165 | } |
66 | } |
| 166 | 67 | ||
| 167 | /** @} |
68 | /** @} |
| 168 | */ |
69 | */ |