Rev 1890 | Rev 1892 | Go to most recent revision | Show entire file | Ignore whitespace | Details | Blame | Last modification | View Log | RSS feed
| Rev 1890 | Rev 1891 | ||
|---|---|---|---|
| Line 38... | Line 38... | ||
| 38 | #include <genarch/mm/asid_fifo.h> |
38 | #include <genarch/mm/asid_fifo.h> |
| 39 | #include <debug.h> |
39 | #include <debug.h> |
| 40 | 40 | ||
| 41 | #ifdef CONFIG_TSB |
41 | #ifdef CONFIG_TSB |
| 42 | #include <arch/mm/tsb.h> |
42 | #include <arch/mm/tsb.h> |
| - | 43 | #include <arch/memstr.h> |
|
| - | 44 | #include <synch/mutex.h> |
|
| - | 45 | #include <arch/asm.h> |
|
| - | 46 | #include <mm/frame.h> |
|
| - | 47 | #include <bitops.h> |
|
| - | 48 | #include <macros.h> |
|
| 43 | #endif |
49 | #endif |
| 44 | 50 | ||
| 45 | /** Architecture dependent address space init. */ |
51 | /** Architecture dependent address space init. */ |
| 46 | void as_arch_init(void) |
52 | void as_arch_init(void) |
| 47 | { |
53 | { |
| 48 | as_operations = &as_ht_operations; |
54 | as_operations = &as_ht_operations; |
| 49 | asid_fifo_init(); |
55 | asid_fifo_init(); |
| 50 | } |
56 | } |
| 51 | 57 | ||
| - | 58 | int as_constructor_arch(as_t *as, int flags) |
|
| - | 59 | { |
|
| - | 60 | #ifdef CONFIG_TSB |
|
| - | 61 | int order = fnzb32(((ITSB_ENTRY_COUNT+DTSB_ENTRY_COUNT)*sizeof(tsb_entry_t))>>FRAME_WIDTH); |
|
| - | 62 | uintptr_t tsb = (uintptr_t) frame_alloc(order, flags); |
|
| - | 63 | ||
| - | 64 | if (!tsb) |
|
| - | 65 | return -1; |
|
| - | 66 | ||
| - | 67 | as->arch.itsb = (tsb_entry_t *) tsb; |
|
| - | 68 | as->arch.dtsb = (tsb_entry_t *) (tsb + ITSB_ENTRY_COUNT * sizeof(tsb_entry_t)); |
|
| - | 69 | #endif |
|
| - | 70 | return 0; |
|
| - | 71 | } |
|
| - | 72 | ||
| - | 73 | int as_destructor_arch(as_t *as) |
|
| - | 74 | { |
|
| - | 75 | #ifdef CONFIG_TSB |
|
| - | 76 | count_t cnt = ((ITSB_ENTRY_COUNT+DTSB_ENTRY_COUNT)*sizeof(tsb_entry_t))>>FRAME_WIDTH; |
|
| - | 77 | frame_free((uintptr_t) as->arch.itsb); |
|
| - | 78 | return cnt; |
|
| - | 79 | #else |
|
| - | 80 | return 0; |
|
| - | 81 | #endif |
|
| - | 82 | } |
|
| - | 83 | ||
| - | 84 | int as_create_arch(as_t *as, int flags) |
|
| - | 85 | { |
|
| - | 86 | #ifdef CONFIG_TSB |
|
| - | 87 | ipl_t ipl; |
|
| - | 88 | ||
| - | 89 | memsetb((uintptr_t) as->arch.itsb, (ITSB_ENTRY_COUNT+DTSB_ENTRY_COUNT)*sizeof(tsb_entry_t), 0); |
|
| - | 90 | ipl = interrupts_disable(); |
|
| - | 91 | mutex_lock_active(&as->lock); /* completely unnecessary, but polite */ |
|
| - | 92 | tsb_invalidate(as, 0, (count_t) -1); |
|
| - | 93 | mutex_unlock(&as->lock); |
|
| - | 94 | interrupts_restore(ipl); |
|
| - | 95 | #endif |
|
| - | 96 | return 0; |
|
| - | 97 | } |
|
| - | 98 | ||
| 52 | /** Perform sparc64-specific tasks when an address space becomes active on the processor. |
99 | /** Perform sparc64-specific tasks when an address space becomes active on the processor. |
| 53 | * |
100 | * |
| 54 | * Install ASID and map TSBs. |
101 | * Install ASID and map TSBs. |
| 55 | * |
102 | * |
| 56 | * @param as Address space. |
103 | * @param as Address space. |
| Line 76... | Line 123... | ||
| 76 | ctx.v = 0; |
123 | ctx.v = 0; |
| 77 | ctx.context = as->asid; |
124 | ctx.context = as->asid; |
| 78 | mmu_secondary_context_write(ctx.v); |
125 | mmu_secondary_context_write(ctx.v); |
| 79 | 126 | ||
| 80 | #ifdef CONFIG_TSB |
127 | #ifdef CONFIG_TSB |
| 81 | if (as != AS_KERNEL) { |
- | |
| 82 | uintptr_t base = ALIGN_DOWN(config.base, 1 << KERNEL_PAGE_WIDTH); |
128 | uintptr_t base = ALIGN_DOWN(config.base, 1 << KERNEL_PAGE_WIDTH); |
| 83 | 129 | ||
| 84 | ASSERT(as->arch.itsb && as->arch.dtsb); |
130 | ASSERT(as->arch.itsb && as->arch.dtsb); |
| 85 | 131 | ||
| 86 | uintptr_t tsb = as->arch.itsb; |
132 | uintptr_t tsb = (uintptr_t) as->arch.itsb; |
| 87 | - | ||
| 88 | if (!overlaps(tsb, 8*PAGE_SIZE, base, 1 << KERNEL_PAGE_WIDTH)) { |
- | |
| 89 | /* |
- | |
| 90 | * TSBs were allocated from memory not covered |
- | |
| 91 | * by the locked 4M kernel DTLB entry. We need |
- | |
| 92 | * to map both TSBs explicitly. |
- | |
| 93 | */ |
- | |
| 94 | dtlb_demap(TLB_DEMAP_PAGE, TLB_DEMAP_NUCLEUS, tsb); |
- | |
| 95 | dtlb_insert_mapping(tsb, KA2PA(tsb), PAGESIZE_64K, true, true); |
- | |
| 96 | } |
- | |
| 97 | 133 | ||
| - | 134 | if (!overlaps(tsb, 8*PAGE_SIZE, base, 1 << KERNEL_PAGE_WIDTH)) { |
|
| 98 | /* |
135 | /* |
| - | 136 | * TSBs were allocated from memory not covered |
|
| - | 137 | * by the locked 4M kernel DTLB entry. We need |
|
| 99 | * Setup TSB Base registers. |
138 | * to map both TSBs explicitly. |
| 100 | */ |
139 | */ |
| 101 | tsb_base_reg_t tsb_base; |
- | |
| 102 | - | ||
| 103 | tsb_base.value = 0; |
- | |
| 104 | tsb_base.size = TSB_SIZE; |
- | |
| 105 | tsb_base.split = 0; |
- | |
| 106 | - | ||
| 107 | tsb_base.base = as->arch.itsb >> PAGE_WIDTH; |
140 | dtlb_demap(TLB_DEMAP_PAGE, TLB_DEMAP_NUCLEUS, tsb); |
| 108 | itsb_base_write(tsb_base.value); |
- | |
| 109 | tsb_base.base = as->arch.dtsb >> PAGE_WIDTH; |
141 | dtlb_insert_mapping(tsb, KA2PA(tsb), PAGESIZE_64K, true, true); |
| 110 | dtsb_base_write(tsb_base.value); |
- | |
| 111 | } |
142 | } |
| - | 143 | ||
| - | 144 | /* |
|
| - | 145 | * Setup TSB Base registers. |
|
| - | 146 | */ |
|
| - | 147 | tsb_base_reg_t tsb_base; |
|
| - | 148 | ||
| - | 149 | tsb_base.value = 0; |
|
| - | 150 | tsb_base.size = TSB_SIZE; |
|
| - | 151 | tsb_base.split = 0; |
|
| - | 152 | ||
| - | 153 | tsb_base.base = ((uintptr_t) as->arch.itsb) >> PAGE_WIDTH; |
|
| - | 154 | itsb_base_write(tsb_base.value); |
|
| - | 155 | tsb_base.base = ((uintptr_t) as->arch.dtsb) >> PAGE_WIDTH; |
|
| - | 156 | dtsb_base_write(tsb_base.value); |
|
| 112 | #endif |
157 | #endif |
| 113 | } |
158 | } |
| 114 | 159 | ||
| 115 | /** Perform sparc64-specific tasks when an address space is removed from the processor. |
160 | /** Perform sparc64-specific tasks when an address space is removed from the processor. |
| 116 | * |
161 | * |
| Line 127... | Line 172... | ||
| 127 | * because we only read members that are |
172 | * because we only read members that are |
| 128 | * currently read-only. |
173 | * currently read-only. |
| 129 | */ |
174 | */ |
| 130 | 175 | ||
| 131 | #ifdef CONFIG_TSB |
176 | #ifdef CONFIG_TSB |
| 132 | if (as != AS_KERNEL) { |
- | |
| 133 | uintptr_t base = ALIGN_DOWN(config.base, 1 << KERNEL_PAGE_WIDTH); |
177 | uintptr_t base = ALIGN_DOWN(config.base, 1 << KERNEL_PAGE_WIDTH); |
| 134 | 178 | ||
| 135 | ASSERT(as->arch.itsb && as->arch.dtsb); |
179 | ASSERT(as->arch.itsb && as->arch.dtsb); |
| 136 | 180 | ||
| 137 | uintptr_t tsb = as->arch.itsb; |
181 | uintptr_t tsb = (uintptr_t) as->arch.itsb; |
| 138 | - | ||
| 139 | if (!overlaps(tsb, 8*PAGE_SIZE, base, 1 << KERNEL_PAGE_WIDTH)) { |
- | |
| 140 | /* |
- | |
| 141 | * TSBs were allocated from memory not covered |
- | |
| 142 | * by the locked 4M kernel DTLB entry. We need |
- | |
| 143 | * to demap the entry installed by as_install_arch(). |
- | |
| 144 | */ |
- | |
| 145 | dtlb_demap(TLB_DEMAP_PAGE, TLB_DEMAP_NUCLEUS, tsb); |
- | |
| 146 | } |
- | |
| 147 | 182 | ||
| - | 183 | if (!overlaps(tsb, 8*PAGE_SIZE, base, 1 << KERNEL_PAGE_WIDTH)) { |
|
| - | 184 | /* |
|
| - | 185 | * TSBs were allocated from memory not covered |
|
| - | 186 | * by the locked 4M kernel DTLB entry. We need |
|
| - | 187 | * to demap the entry installed by as_install_arch(). |
|
| - | 188 | */ |
|
| - | 189 | dtlb_demap(TLB_DEMAP_PAGE, TLB_DEMAP_NUCLEUS, tsb); |
|
| 148 | } |
190 | } |
| 149 | #endif |
191 | #endif |
| 150 | } |
192 | } |
| 151 | 193 | ||
| 152 | /** @} |
194 | /** @} |