Rev 3779 | Rev 3902 | Go to most recent revision | Only display areas with differences | Regard whitespace | Details | Blame | Last modification | View Log | RSS feed
Rev 3779 | Rev 3880 | ||
---|---|---|---|
1 | /* |
1 | /* |
2 | * Copyright (c) 2005 Jakub Jermar |
2 | * Copyright (c) 2005 Jakub Jermar |
3 | * All rights reserved. |
3 | * All rights reserved. |
4 | * |
4 | * |
5 | * Redistribution and use in source and binary forms, with or without |
5 | * Redistribution and use in source and binary forms, with or without |
6 | * modification, are permitted provided that the following conditions |
6 | * modification, are permitted provided that the following conditions |
7 | * are met: |
7 | * are met: |
8 | * |
8 | * |
9 | * - Redistributions of source code must retain the above copyright |
9 | * - Redistributions of source code must retain the above copyright |
10 | * notice, this list of conditions and the following disclaimer. |
10 | * notice, this list of conditions and the following disclaimer. |
11 | * - Redistributions in binary form must reproduce the above copyright |
11 | * - Redistributions in binary form must reproduce the above copyright |
12 | * notice, this list of conditions and the following disclaimer in the |
12 | * notice, this list of conditions and the following disclaimer in the |
13 | * documentation and/or other materials provided with the distribution. |
13 | * documentation and/or other materials provided with the distribution. |
14 | * - The name of the author may not be used to endorse or promote products |
14 | * - The name of the author may not be used to endorse or promote products |
15 | * derived from this software without specific prior written permission. |
15 | * derived from this software without specific prior written permission. |
16 | * |
16 | * |
17 | * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR |
17 | * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR |
18 | * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES |
18 | * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES |
19 | * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. |
19 | * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. |
20 | * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, |
20 | * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, |
21 | * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT |
21 | * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT |
22 | * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
22 | * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
23 | * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
23 | * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
24 | * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
24 | * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
25 | * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF |
25 | * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF |
26 | * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
26 | * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
27 | */ |
27 | */ |
28 | 28 | ||
29 | /** @addtogroup ia64 |
29 | /** @addtogroup ia64 |
30 | * @{ |
30 | * @{ |
31 | */ |
31 | */ |
32 | /** @file |
32 | /** @file |
33 | */ |
33 | */ |
34 | 34 | ||
35 | #ifndef KERN_ia64_ASM_H_ |
35 | #ifndef KERN_ia64_ASM_H_ |
36 | #define KERN_ia64_ASM_H_ |
36 | #define KERN_ia64_ASM_H_ |
37 | 37 | ||
38 | #include <config.h> |
38 | #include <config.h> |
39 | #include <arch/types.h> |
39 | #include <arch/types.h> |
40 | #include <arch/register.h> |
40 | #include <arch/register.h> |
41 | 41 | ||
42 | #define IA64_IOSPACE_ADDRESS 0xE001000000000000ULL |
42 | #define IA64_IOSPACE_ADDRESS 0xE001000000000000ULL |
43 | 43 | ||
44 | static inline void outb(ioport_t port, uint8_t v) |
44 | static inline void outb(ioport_t port, uint8_t v) |
45 | { |
45 | { |
46 | *((uint8_t *)(IA64_IOSPACE_ADDRESS + |
46 | *((uint8_t *)(IA64_IOSPACE_ADDRESS + |
47 | ((port & 0xfff) | ((port >> 2) << 12)))) = v; |
47 | ((port & 0xfff) | ((port >> 2) << 12)))) = v; |
48 | 48 | ||
49 | asm volatile ("mf\n" ::: "memory"); |
49 | asm volatile ("mf\n" ::: "memory"); |
50 | } |
50 | } |
51 | 51 | ||
52 | static inline void outw(ioport_t port, uint16_t v) |
52 | static inline void outw(ioport_t port, uint16_t v) |
53 | { |
53 | { |
54 | *((uint16_t *)(IA64_IOSPACE_ADDRESS + |
54 | *((uint16_t *)(IA64_IOSPACE_ADDRESS + |
55 | ((port & 0xfff) | ((port >> 2) << 12)))) = v; |
55 | ((port & 0xfff) | ((port >> 2) << 12)))) = v; |
56 | 56 | ||
57 | asm volatile ("mf\n" ::: "memory"); |
57 | asm volatile ("mf\n" ::: "memory"); |
58 | } |
58 | } |
59 | 59 | ||
60 | static inline void outl(ioport_t port, uint32_t v) |
60 | static inline void outl(ioport_t port, uint32_t v) |
61 | { |
61 | { |
62 | *((uint32_t *)(IA64_IOSPACE_ADDRESS + |
62 | *((uint32_t *)(IA64_IOSPACE_ADDRESS + |
63 | ((port & 0xfff) | ((port >> 2) << 12)))) = v; |
63 | ((port & 0xfff) | ((port >> 2) << 12)))) = v; |
64 | 64 | ||
65 | asm volatile ("mf\n" ::: "memory"); |
65 | asm volatile ("mf\n" ::: "memory"); |
66 | } |
66 | } |
67 | 67 | ||
68 | static inline uint8_t inb(ioport_t port) |
68 | static inline uint8_t inb(ioport_t port) |
69 | { |
69 | { |
70 | asm volatile ("mf\n" ::: "memory"); |
70 | asm volatile ("mf\n" ::: "memory"); |
71 | 71 | ||
72 | return *((uint8_t *)(IA64_IOSPACE_ADDRESS + |
72 | return *((uint8_t *)(IA64_IOSPACE_ADDRESS + |
73 | ((port & 0xfff) | ((port >> 2) << 12)))); |
73 | ((port & 0xfff) | ((port >> 2) << 12)))); |
74 | } |
74 | } |
75 | 75 | ||
76 | static inline uint16_t inw(ioport_t port) |
76 | static inline uint16_t inw(ioport_t port) |
77 | { |
77 | { |
78 | asm volatile ("mf\n" ::: "memory"); |
78 | asm volatile ("mf\n" ::: "memory"); |
79 | 79 | ||
80 | return *((uint16_t *)(IA64_IOSPACE_ADDRESS + |
80 | return *((uint16_t *)(IA64_IOSPACE_ADDRESS + |
81 | ((port & 0xffE) | ((port >> 2) << 12)))); |
81 | ((port & 0xffE) | ((port >> 2) << 12)))); |
82 | } |
82 | } |
83 | 83 | ||
84 | static inline uint32_t inl(ioport_t port) |
84 | static inline uint32_t inl(ioport_t port) |
85 | { |
85 | { |
86 | asm volatile ("mf\n" ::: "memory"); |
86 | asm volatile ("mf\n" ::: "memory"); |
87 | 87 | ||
88 | return *((uint32_t *)(IA64_IOSPACE_ADDRESS + |
88 | return *((uint32_t *)(IA64_IOSPACE_ADDRESS + |
89 | ((port & 0xfff) | ((port >> 2) << 12)))); |
89 | ((port & 0xfff) | ((port >> 2) << 12)))); |
90 | } |
90 | } |
91 | 91 | ||
92 | /** Return base address of current stack |
92 | /** Return base address of current stack |
93 | * |
93 | * |
94 | * Return the base address of the current stack. |
94 | * Return the base address of the current stack. |
95 | * The stack is assumed to be STACK_SIZE long. |
95 | * The stack is assumed to be STACK_SIZE long. |
96 | * The stack must start on page boundary. |
96 | * The stack must start on page boundary. |
97 | */ |
97 | */ |
98 | static inline uintptr_t get_stack_base(void) |
98 | static inline uintptr_t get_stack_base(void) |
99 | { |
99 | { |
100 | uint64_t v; |
100 | uint64_t v; |
101 | 101 | ||
102 | //I'm not sure why but this code bad inlines in scheduler, |
102 | //I'm not sure why but this code bad inlines in scheduler, |
103 | //so THE shifts about 16B and causes kernel panic |
103 | //so THE shifts about 16B and causes kernel panic |
104 | //asm volatile ("and %0 = %1, r12" : "=r" (v) : "r" (~(STACK_SIZE-1))); |
104 | //asm volatile ("and %0 = %1, r12" : "=r" (v) : "r" (~(STACK_SIZE-1))); |
105 | //return v; |
105 | //return v; |
106 | 106 | ||
107 | //this code have the same meaning but inlines well |
107 | //this code have the same meaning but inlines well |
108 | asm volatile ("mov %0 = r12" : "=r" (v) ); |
108 | asm volatile ("mov %0 = r12" : "=r" (v) ); |
109 | return v & (~(STACK_SIZE-1)); |
109 | return v & (~(STACK_SIZE-1)); |
110 | } |
110 | } |
111 | 111 | ||
112 | /** Return Processor State Register. |
112 | /** Return Processor State Register. |
113 | * |
113 | * |
114 | * @return PSR. |
114 | * @return PSR. |
115 | */ |
115 | */ |
116 | static inline uint64_t psr_read(void) |
116 | static inline uint64_t psr_read(void) |
117 | { |
117 | { |
118 | uint64_t v; |
118 | uint64_t v; |
119 | 119 | ||
120 | asm volatile ("mov %0 = psr\n" : "=r" (v)); |
120 | asm volatile ("mov %0 = psr\n" : "=r" (v)); |
121 | 121 | ||
122 | return v; |
122 | return v; |
123 | } |
123 | } |
124 | 124 | ||
125 | /** Read IVA (Interruption Vector Address). |
125 | /** Read IVA (Interruption Vector Address). |
126 | * |
126 | * |
127 | * @return Return location of interruption vector table. |
127 | * @return Return location of interruption vector table. |
128 | */ |
128 | */ |
129 | static inline uint64_t iva_read(void) |
129 | static inline uint64_t iva_read(void) |
130 | { |
130 | { |
131 | uint64_t v; |
131 | uint64_t v; |
132 | 132 | ||
133 | asm volatile ("mov %0 = cr.iva\n" : "=r" (v)); |
133 | asm volatile ("mov %0 = cr.iva\n" : "=r" (v)); |
134 | 134 | ||
135 | return v; |
135 | return v; |
136 | } |
136 | } |
137 | 137 | ||
138 | /** Write IVA (Interruption Vector Address) register. |
138 | /** Write IVA (Interruption Vector Address) register. |
139 | * |
139 | * |
140 | * @param v New location of interruption vector table. |
140 | * @param v New location of interruption vector table. |
141 | */ |
141 | */ |
142 | static inline void iva_write(uint64_t v) |
142 | static inline void iva_write(uint64_t v) |
143 | { |
143 | { |
144 | asm volatile ("mov cr.iva = %0\n" : : "r" (v)); |
144 | asm volatile ("mov cr.iva = %0\n" : : "r" (v)); |
145 | } |
145 | } |
146 | 146 | ||
147 | 147 | ||
148 | /** Read IVR (External Interrupt Vector Register). |
148 | /** Read IVR (External Interrupt Vector Register). |
149 | * |
149 | * |
150 | * @return Highest priority, pending, unmasked external interrupt vector. |
150 | * @return Highest priority, pending, unmasked external interrupt vector. |
151 | */ |
151 | */ |
152 | static inline uint64_t ivr_read(void) |
152 | static inline uint64_t ivr_read(void) |
153 | { |
153 | { |
154 | uint64_t v; |
154 | uint64_t v; |
155 | 155 | ||
156 | asm volatile ("mov %0 = cr.ivr\n" : "=r" (v)); |
156 | asm volatile ("mov %0 = cr.ivr\n" : "=r" (v)); |
157 | 157 | ||
158 | return v; |
158 | return v; |
159 | } |
159 | } |
160 | 160 | ||
161 | static inline uint64_t cr64_read(void) |
161 | static inline uint64_t cr64_read(void) |
162 | { |
162 | { |
163 | uint64_t v; |
163 | uint64_t v; |
164 | 164 | ||
165 | asm volatile ("mov %0 = cr64\n" : "=r" (v)); |
165 | asm volatile ("mov %0 = cr64\n" : "=r" (v)); |
166 | 166 | ||
167 | return v; |
167 | return v; |
168 | } |
168 | } |
169 | 169 | ||
170 | 170 | ||
171 | /** Write ITC (Interval Timer Counter) register. |
171 | /** Write ITC (Interval Timer Counter) register. |
172 | * |
172 | * |
173 | * @param v New counter value. |
173 | * @param v New counter value. |
174 | */ |
174 | */ |
175 | static inline void itc_write(uint64_t v) |
175 | static inline void itc_write(uint64_t v) |
176 | { |
176 | { |
177 | asm volatile ("mov ar.itc = %0\n" : : "r" (v)); |
177 | asm volatile ("mov ar.itc = %0\n" : : "r" (v)); |
178 | } |
178 | } |
179 | 179 | ||
180 | /** Read ITC (Interval Timer Counter) register. |
180 | /** Read ITC (Interval Timer Counter) register. |
181 | * |
181 | * |
182 | * @return Current counter value. |
182 | * @return Current counter value. |
183 | */ |
183 | */ |
184 | static inline uint64_t itc_read(void) |
184 | static inline uint64_t itc_read(void) |
185 | { |
185 | { |
186 | uint64_t v; |
186 | uint64_t v; |
187 | 187 | ||
188 | asm volatile ("mov %0 = ar.itc\n" : "=r" (v)); |
188 | asm volatile ("mov %0 = ar.itc\n" : "=r" (v)); |
189 | 189 | ||
190 | return v; |
190 | return v; |
191 | } |
191 | } |
192 | 192 | ||
193 | /** Write ITM (Interval Timer Match) register. |
193 | /** Write ITM (Interval Timer Match) register. |
194 | * |
194 | * |
195 | * @param v New match value. |
195 | * @param v New match value. |
196 | */ |
196 | */ |
197 | static inline void itm_write(uint64_t v) |
197 | static inline void itm_write(uint64_t v) |
198 | { |
198 | { |
199 | asm volatile ("mov cr.itm = %0\n" : : "r" (v)); |
199 | asm volatile ("mov cr.itm = %0\n" : : "r" (v)); |
200 | } |
200 | } |
201 | 201 | ||
202 | /** Read ITM (Interval Timer Match) register. |
202 | /** Read ITM (Interval Timer Match) register. |
203 | * |
203 | * |
204 | * @return Match value. |
204 | * @return Match value. |
205 | */ |
205 | */ |
206 | static inline uint64_t itm_read(void) |
206 | static inline uint64_t itm_read(void) |
207 | { |
207 | { |
208 | uint64_t v; |
208 | uint64_t v; |
209 | 209 | ||
210 | asm volatile ("mov %0 = cr.itm\n" : "=r" (v)); |
210 | asm volatile ("mov %0 = cr.itm\n" : "=r" (v)); |
211 | 211 | ||
212 | return v; |
212 | return v; |
213 | } |
213 | } |
214 | 214 | ||
215 | /** Read ITV (Interval Timer Vector) register. |
215 | /** Read ITV (Interval Timer Vector) register. |
216 | * |
216 | * |
217 | * @return Current vector and mask bit. |
217 | * @return Current vector and mask bit. |
218 | */ |
218 | */ |
219 | static inline uint64_t itv_read(void) |
219 | static inline uint64_t itv_read(void) |
220 | { |
220 | { |
221 | uint64_t v; |
221 | uint64_t v; |
222 | 222 | ||
223 | asm volatile ("mov %0 = cr.itv\n" : "=r" (v)); |
223 | asm volatile ("mov %0 = cr.itv\n" : "=r" (v)); |
224 | 224 | ||
225 | return v; |
225 | return v; |
226 | } |
226 | } |
227 | 227 | ||
228 | /** Write ITV (Interval Timer Vector) register. |
228 | /** Write ITV (Interval Timer Vector) register. |
229 | * |
229 | * |
230 | * @param v New vector and mask bit. |
230 | * @param v New vector and mask bit. |
231 | */ |
231 | */ |
232 | static inline void itv_write(uint64_t v) |
232 | static inline void itv_write(uint64_t v) |
233 | { |
233 | { |
234 | asm volatile ("mov cr.itv = %0\n" : : "r" (v)); |
234 | asm volatile ("mov cr.itv = %0\n" : : "r" (v)); |
235 | } |
235 | } |
236 | 236 | ||
237 | /** Write EOI (End Of Interrupt) register. |
237 | /** Write EOI (End Of Interrupt) register. |
238 | * |
238 | * |
239 | * @param v This value is ignored. |
239 | * @param v This value is ignored. |
240 | */ |
240 | */ |
241 | static inline void eoi_write(uint64_t v) |
241 | static inline void eoi_write(uint64_t v) |
242 | { |
242 | { |
243 | asm volatile ("mov cr.eoi = %0\n" : : "r" (v)); |
243 | asm volatile ("mov cr.eoi = %0\n" : : "r" (v)); |
244 | } |
244 | } |
245 | 245 | ||
246 | /** Read TPR (Task Priority Register). |
246 | /** Read TPR (Task Priority Register). |
247 | * |
247 | * |
248 | * @return Current value of TPR. |
248 | * @return Current value of TPR. |
249 | */ |
249 | */ |
250 | static inline uint64_t tpr_read(void) |
250 | static inline uint64_t tpr_read(void) |
251 | { |
251 | { |
252 | uint64_t v; |
252 | uint64_t v; |
253 | 253 | ||
254 | asm volatile ("mov %0 = cr.tpr\n" : "=r" (v)); |
254 | asm volatile ("mov %0 = cr.tpr\n" : "=r" (v)); |
255 | 255 | ||
256 | return v; |
256 | return v; |
257 | } |
257 | } |
258 | 258 | ||
259 | /** Write TPR (Task Priority Register). |
259 | /** Write TPR (Task Priority Register). |
260 | * |
260 | * |
261 | * @param v New value of TPR. |
261 | * @param v New value of TPR. |
262 | */ |
262 | */ |
263 | static inline void tpr_write(uint64_t v) |
263 | static inline void tpr_write(uint64_t v) |
264 | { |
264 | { |
265 | asm volatile ("mov cr.tpr = %0\n" : : "r" (v)); |
265 | asm volatile ("mov cr.tpr = %0\n" : : "r" (v)); |
266 | } |
266 | } |
267 | 267 | ||
268 | /** Disable interrupts. |
268 | /** Disable interrupts. |
269 | * |
269 | * |
270 | * Disable interrupts and return previous |
270 | * Disable interrupts and return previous |
271 | * value of PSR. |
271 | * value of PSR. |
272 | * |
272 | * |
273 | * @return Old interrupt priority level. |
273 | * @return Old interrupt priority level. |
274 | */ |
274 | */ |
275 | static ipl_t interrupts_disable(void) |
275 | static ipl_t interrupts_disable(void) |
276 | { |
276 | { |
277 | uint64_t v; |
277 | uint64_t v; |
278 | 278 | ||
279 | asm volatile ( |
279 | asm volatile ( |
280 | "mov %0 = psr\n" |
280 | "mov %0 = psr\n" |
281 | "rsm %1\n" |
281 | "rsm %1\n" |
282 | : "=r" (v) |
282 | : "=r" (v) |
283 | : "i" (PSR_I_MASK) |
283 | : "i" (PSR_I_MASK) |
284 | ); |
284 | ); |
285 | 285 | ||
286 | return (ipl_t) v; |
286 | return (ipl_t) v; |
287 | } |
287 | } |
288 | 288 | ||
289 | /** Enable interrupts. |
289 | /** Enable interrupts. |
290 | * |
290 | * |
291 | * Enable interrupts and return previous |
291 | * Enable interrupts and return previous |
292 | * value of PSR. |
292 | * value of PSR. |
293 | * |
293 | * |
294 | * @return Old interrupt priority level. |
294 | * @return Old interrupt priority level. |
295 | */ |
295 | */ |
296 | static ipl_t interrupts_enable(void) |
296 | static ipl_t interrupts_enable(void) |
297 | { |
297 | { |
298 | uint64_t v; |
298 | uint64_t v; |
299 | 299 | ||
300 | asm volatile ( |
300 | asm volatile ( |
301 | "mov %0 = psr\n" |
301 | "mov %0 = psr\n" |
302 | "ssm %1\n" |
302 | "ssm %1\n" |
303 | ";;\n" |
303 | ";;\n" |
304 | "srlz.d\n" |
304 | "srlz.d\n" |
305 | : "=r" (v) |
305 | : "=r" (v) |
306 | : "i" (PSR_I_MASK) |
306 | : "i" (PSR_I_MASK) |
307 | ); |
307 | ); |
308 | 308 | ||
309 | return (ipl_t) v; |
309 | return (ipl_t) v; |
310 | } |
310 | } |
311 | 311 | ||
312 | /** Restore interrupt priority level. |
312 | /** Restore interrupt priority level. |
313 | * |
313 | * |
314 | * Restore PSR. |
314 | * Restore PSR. |
315 | * |
315 | * |
316 | * @param ipl Saved interrupt priority level. |
316 | * @param ipl Saved interrupt priority level. |
317 | */ |
317 | */ |
318 | static inline void interrupts_restore(ipl_t ipl) |
318 | static inline void interrupts_restore(ipl_t ipl) |
319 | { |
319 | { |
320 | if (ipl & PSR_I_MASK) |
320 | if (ipl & PSR_I_MASK) |
321 | (void) interrupts_enable(); |
321 | (void) interrupts_enable(); |
322 | else |
322 | else |
323 | (void) interrupts_disable(); |
323 | (void) interrupts_disable(); |
324 | } |
324 | } |
325 | 325 | ||
326 | /** Return interrupt priority level. |
326 | /** Return interrupt priority level. |
327 | * |
327 | * |
328 | * @return PSR. |
328 | * @return PSR. |
329 | */ |
329 | */ |
330 | static inline ipl_t interrupts_read(void) |
330 | static inline ipl_t interrupts_read(void) |
331 | { |
331 | { |
332 | return (ipl_t) psr_read(); |
332 | return (ipl_t) psr_read(); |
333 | } |
333 | } |
334 | 334 | ||
335 | /** Disable protection key checking. */ |
335 | /** Disable protection key checking. */ |
336 | static inline void pk_disable(void) |
336 | static inline void pk_disable(void) |
337 | { |
337 | { |
338 | asm volatile ("rsm %0\n" : : "i" (PSR_PK_MASK)); |
338 | asm volatile ("rsm %0\n" : : "i" (PSR_PK_MASK)); |
339 | } |
339 | } |
340 | 340 | ||
341 | extern void cpu_halt(void); |
341 | extern void cpu_halt(void); |
342 | extern void cpu_sleep(void); |
342 | extern void cpu_sleep(void); |
343 | extern void asm_delay_loop(uint32_t t); |
343 | extern void asm_delay_loop(uint32_t t); |
344 | 344 | ||
345 | extern void switch_to_userspace(uintptr_t, uintptr_t, uintptr_t, uintptr_t, |
345 | extern void switch_to_userspace(uintptr_t, uintptr_t, uintptr_t, uintptr_t, |
346 | uint64_t, uint64_t); |
346 | uint64_t, uint64_t); |
347 | 347 | ||
348 | #endif |
348 | #endif |
349 | 349 | ||
350 | /** @} |
350 | /** @} |
351 | */ |
351 | */ |
352 | 352 |