Rev 2018 | Rev 2784 | Go to most recent revision | Details | Compare with Previous | Last modification | View Log | RSS feed
Rev | Author | Line No. | Line |
---|---|---|---|
224 | palkovsky | 1 | # |
2071 | jermar | 2 | # Copyright (c) 2005 Ondrej Palkovsky |
224 | palkovsky | 3 | # All rights reserved. |
4 | # |
||
5 | # Redistribution and use in source and binary forms, with or without |
||
6 | # modification, are permitted provided that the following conditions |
||
7 | # are met: |
||
8 | # |
||
9 | # - Redistributions of source code must retain the above copyright |
||
10 | # notice, this list of conditions and the following disclaimer. |
||
11 | # - Redistributions in binary form must reproduce the above copyright |
||
12 | # notice, this list of conditions and the following disclaimer in the |
||
13 | # documentation and/or other materials provided with the distribution. |
||
14 | # - The name of the author may not be used to endorse or promote products |
||
15 | # derived from this software without specific prior written permission. |
||
16 | # |
||
17 | # THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR |
||
18 | # IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES |
||
19 | # OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. |
||
20 | # IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, |
||
21 | # INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT |
||
22 | # NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
||
23 | # DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
||
24 | # THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
||
25 | # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF |
||
26 | # THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
||
27 | # |
||
28 | |||
1021 | jermar | 29 | #define IREGISTER_SPACE 120 |
224 | palkovsky | 30 | |
1021 | jermar | 31 | #define IOFFSET_RAX 0x0 |
32 | #define IOFFSET_RBX 0x8 |
||
33 | #define IOFFSET_RCX 0x10 |
||
34 | #define IOFFSET_RDX 0x18 |
||
35 | #define IOFFSET_RSI 0x20 |
||
36 | #define IOFFSET_RDI 0x28 |
||
37 | #define IOFFSET_R8 0x30 |
||
38 | #define IOFFSET_R9 0x38 |
||
39 | #define IOFFSET_R10 0x40 |
||
40 | #define IOFFSET_R11 0x48 |
||
41 | #define IOFFSET_R12 0x50 |
||
42 | #define IOFFSET_R13 0x58 |
||
43 | #define IOFFSET_R14 0x60 |
||
44 | #define IOFFSET_R15 0x68 |
||
45 | #define IOFFSET_RBP 0x70 |
||
46 | |||
224 | palkovsky | 47 | # Mask for interrupts 0 - 31 (bits 0 - 31) where 0 means that int has no error word |
48 | # and 1 means interrupt with error word |
||
49 | #define ERROR_WORD_INTERRUPT_LIST 0x00027D00 |
||
50 | |||
51 | #include <arch/pm.h> |
||
808 | palkovsky | 52 | #include <arch/mm/page.h> |
224 | palkovsky | 53 | |
54 | .text |
||
55 | .global interrupt_handlers |
||
803 | palkovsky | 56 | .global syscall_entry |
224 | palkovsky | 57 | .global panic_printf |
58 | |||
59 | panic_printf: |
||
60 | movq $halt, (%rsp) |
||
61 | jmp printf |
||
62 | |||
252 | palkovsky | 63 | .global cpuid |
242 | palkovsky | 64 | .global has_cpuid |
2018 | decky | 65 | .global get_cycle |
251 | palkovsky | 66 | .global read_efer_flag |
67 | .global set_efer_flag |
||
1288 | jermar | 68 | .global memcpy |
69 | .global memcpy_from_uspace |
||
70 | .global memcpy_to_uspace |
||
71 | .global memcpy_from_uspace_failover_address |
||
72 | .global memcpy_to_uspace_failover_address |
||
73 | |||
74 | #define MEMCPY_DST %rdi |
||
75 | #define MEMCPY_SRC %rsi |
||
76 | #define MEMCPY_SIZE %rdx |
||
77 | |||
78 | /** |
||
79 | * Copy memory from/to userspace. |
||
80 | * |
||
81 | * This is almost conventional memcpy(). |
||
82 | * The difference is that there is a failover part |
||
83 | * to where control is returned from a page fault if |
||
84 | * the page fault occurs during copy_from_uspace() |
||
85 | * or copy_to_uspace(). |
||
86 | * |
||
87 | * @param MEMCPY_DST Destination address. |
||
88 | * @param MEMCPY_SRC Source address. |
||
89 | * @param MEMCPY_SIZE Number of bytes to copy. |
||
90 | * |
||
91 | * @retrun MEMCPY_SRC on success, 0 on failure. |
||
92 | */ |
||
93 | memcpy: |
||
94 | memcpy_from_uspace: |
||
95 | memcpy_to_uspace: |
||
96 | movq MEMCPY_SRC, %rax |
||
97 | |||
98 | movq MEMCPY_SIZE, %rcx |
||
99 | shrq $3, %rcx /* size / 8 */ |
||
251 | palkovsky | 100 | |
1288 | jermar | 101 | rep movsq /* copy as much as possible word by word */ |
102 | |||
103 | movq MEMCPY_SIZE, %rcx |
||
104 | andq $7, %rcx /* size % 8 */ |
||
105 | jz 0f |
||
106 | |||
107 | rep movsb /* copy the rest byte by byte */ |
||
108 | |||
109 | 0: |
||
110 | ret /* return MEMCPY_SRC, success */ |
||
111 | |||
112 | memcpy_from_uspace_failover_address: |
||
113 | memcpy_to_uspace_failover_address: |
||
114 | xorq %rax, %rax /* return 0, failure */ |
||
115 | ret |
||
116 | |||
242 | palkovsky | 117 | ## Determine CPUID support |
118 | # |
||
119 | # Return 0 in EAX if CPUID is not support, 1 if supported. |
||
120 | # |
||
121 | has_cpuid: |
||
122 | pushfq # store flags |
||
123 | popq %rax # read flags |
||
348 | jermar | 124 | movq %rax,%rdx # copy flags |
125 | btcl $21,%edx # swap the ID bit |
||
126 | pushq %rdx |
||
242 | palkovsky | 127 | popfq # propagate the change into flags |
128 | pushfq |
||
348 | jermar | 129 | popq %rdx # read flags |
242 | palkovsky | 130 | andl $(1<<21),%eax # interested only in ID bit |
348 | jermar | 131 | andl $(1<<21),%edx |
132 | xorl %edx,%eax # 0 if not supported, 1 if supported |
||
242 | palkovsky | 133 | ret |
134 | |||
251 | palkovsky | 135 | cpuid: |
136 | movq %rbx, %r10 # we have to preserve rbx across function calls |
||
242 | palkovsky | 137 | |
251 | palkovsky | 138 | movl %edi,%eax # load the command into %eax |
139 | |||
140 | cpuid |
||
141 | movl %eax,0(%rsi) |
||
142 | movl %ebx,4(%rsi) |
||
143 | movl %ecx,8(%rsi) |
||
144 | movl %edx,12(%rsi) |
||
145 | |||
146 | movq %r10, %rbx |
||
147 | ret |
||
148 | |||
2018 | decky | 149 | get_cycle: |
242 | palkovsky | 150 | xorq %rax,%rax |
151 | rdtsc |
||
152 | ret |
||
251 | palkovsky | 153 | |
154 | set_efer_flag: |
||
155 | movq $0xc0000080, %rcx |
||
156 | rdmsr |
||
157 | btsl %edi, %eax |
||
158 | wrmsr |
||
159 | ret |
||
242 | palkovsky | 160 | |
251 | palkovsky | 161 | read_efer_flag: |
162 | movq $0xc0000080, %rcx |
||
163 | rdmsr |
||
164 | ret |
||
242 | palkovsky | 165 | |
224 | palkovsky | 166 | # Push all general purpose registers on stack except %rbp, %rsp |
799 | palkovsky | 167 | .macro save_all_gpr |
168 | movq %rax, IOFFSET_RAX(%rsp) |
||
169 | movq %rcx, IOFFSET_RCX(%rsp) |
||
170 | movq %rdx, IOFFSET_RDX(%rsp) |
||
171 | movq %rsi, IOFFSET_RSI(%rsp) |
||
172 | movq %rdi, IOFFSET_RDI(%rsp) |
||
173 | movq %r8, IOFFSET_R8(%rsp) |
||
174 | movq %r9, IOFFSET_R9(%rsp) |
||
175 | movq %r10, IOFFSET_R10(%rsp) |
||
176 | movq %r11, IOFFSET_R11(%rsp) |
||
1094 | palkovsky | 177 | #ifdef CONFIG_DEBUG_ALLREGS |
178 | movq %rbx, IOFFSET_RBX(%rsp) |
||
179 | movq %rbp, IOFFSET_RBP(%rsp) |
||
799 | palkovsky | 180 | movq %r12, IOFFSET_R12(%rsp) |
181 | movq %r13, IOFFSET_R13(%rsp) |
||
182 | movq %r14, IOFFSET_R14(%rsp) |
||
183 | movq %r15, IOFFSET_R15(%rsp) |
||
1094 | palkovsky | 184 | #endif |
224 | palkovsky | 185 | .endm |
186 | |||
799 | palkovsky | 187 | .macro restore_all_gpr |
188 | movq IOFFSET_RAX(%rsp), %rax |
||
189 | movq IOFFSET_RCX(%rsp), %rcx |
||
190 | movq IOFFSET_RDX(%rsp), %rdx |
||
191 | movq IOFFSET_RSI(%rsp), %rsi |
||
192 | movq IOFFSET_RDI(%rsp), %rdi |
||
193 | movq IOFFSET_R8(%rsp), %r8 |
||
194 | movq IOFFSET_R9(%rsp), %r9 |
||
195 | movq IOFFSET_R10(%rsp), %r10 |
||
196 | movq IOFFSET_R11(%rsp), %r11 |
||
1094 | palkovsky | 197 | #ifdef CONFIG_DEBUG_ALLREGS |
198 | movq IOFFSET_RBX(%rsp), %rbx |
||
199 | movq IOFFSET_RBP(%rsp), %rbp |
||
799 | palkovsky | 200 | movq IOFFSET_R12(%rsp), %r12 |
201 | movq IOFFSET_R13(%rsp), %r13 |
||
202 | movq IOFFSET_R14(%rsp), %r14 |
||
203 | movq IOFFSET_R15(%rsp), %r15 |
||
1094 | palkovsky | 204 | #endif |
224 | palkovsky | 205 | .endm |
1021 | jermar | 206 | |
1094 | palkovsky | 207 | #ifdef CONFIG_DEBUG_ALLREGS |
208 | # define INTERRUPT_ALIGN 256 |
||
209 | #else |
||
210 | # define INTERRUPT_ALIGN 128 |
||
211 | #endif |
||
212 | |||
224 | palkovsky | 213 | ## Declare interrupt handlers |
214 | # |
||
215 | # Declare interrupt handlers for n interrupt |
||
216 | # vectors starting at vector i. |
||
217 | # |
||
1021 | jermar | 218 | # The handlers call exc_dispatch(). |
224 | palkovsky | 219 | # |
220 | .macro handler i n |
||
221 | |||
1021 | jermar | 222 | /* |
223 | * Choose between version with error code and version without error code. |
||
224 | * Both versions have to be of the same size. amd64 assembly is, however, |
||
225 | * a little bit tricky. For instance, subq $0x80, %rsp and subq $0x78, %rsp |
||
226 | * can result in two instructions with different op-code lengths. |
||
1121 | jermar | 227 | * Therefore we align the interrupt handlers. |
1021 | jermar | 228 | */ |
224 | palkovsky | 229 | |
1021 | jermar | 230 | .iflt \i-32 |
231 | .if (1 << \i) & ERROR_WORD_INTERRUPT_LIST |
||
232 | /* |
||
233 | * Version with error word. |
||
234 | */ |
||
235 | subq $IREGISTER_SPACE, %rsp |
||
236 | .else |
||
237 | /* |
||
238 | * Version without error word, |
||
239 | */ |
||
240 | subq $(IREGISTER_SPACE+8), %rsp |
||
241 | .endif |
||
242 | .else |
||
243 | /* |
||
244 | * Version without error word, |
||
245 | */ |
||
246 | subq $(IREGISTER_SPACE+8), %rsp |
||
247 | .endif |
||
224 | palkovsky | 248 | |
1021 | jermar | 249 | save_all_gpr |
224 | palkovsky | 250 | |
1021 | jermar | 251 | movq $(\i), %rdi # %rdi - first parameter |
252 | movq %rsp, %rsi # %rsi - pointer to istate |
||
253 | call exc_dispatch # exc_dispatch(i, istate) |
||
254 | |||
799 | palkovsky | 255 | restore_all_gpr |
256 | # $8 = Skip error word |
||
1021 | jermar | 257 | addq $(IREGISTER_SPACE+8), %rsp |
224 | palkovsky | 258 | iretq |
259 | |||
1094 | palkovsky | 260 | .align INTERRUPT_ALIGN |
224 | palkovsky | 261 | .if (\n-\i)-1 |
262 | handler "(\i+1)",\n |
||
263 | .endif |
||
264 | .endm |
||
1094 | palkovsky | 265 | |
266 | .align INTERRUPT_ALIGN |
||
224 | palkovsky | 267 | interrupt_handlers: |
268 | h_start: |
||
269 | handler 0 IDT_ITEMS |
||
270 | h_end: |
||
803 | palkovsky | 271 | |
224 | palkovsky | 272 | |
803 | palkovsky | 273 | syscall_entry: |
806 | palkovsky | 274 | # Switch to hidden gs |
275 | swapgs |
||
808 | palkovsky | 276 | # %gs:0 now points to pointer to stack page |
277 | mov %gs:0, %r10 # We have a ptr to stack page in r10 |
||
278 | addq $PAGE_SIZE-16, %r10 # We need some space to store old %sp |
||
806 | palkovsky | 279 | |
280 | movq %rsp, 0(%r10) # Save old stack pointer to stack |
||
281 | movq %r10, %rsp # Change to new stack |
||
282 | pushq %rcx # Return address |
||
283 | pushq %r11 # Save flags |
||
284 | |||
285 | # Switch back to remain consistent |
||
286 | swapgs |
||
287 | |||
955 | palkovsky | 288 | sti |
806 | palkovsky | 289 | movq %r9, %rcx # Exchange last parameter as a third |
1212 | palkovsky | 290 | |
803 | palkovsky | 291 | call syscall_handler |
955 | palkovsky | 292 | cli # We will be touching stack pointer |
293 | |||
806 | palkovsky | 294 | popq %r11 |
295 | popq %rcx |
||
296 | movq 0(%rsp), %rsp |
||
297 | sysretq |
||
803 | palkovsky | 298 | |
1278 | palkovsky | 299 | |
224 | palkovsky | 300 | .data |
301 | .global interrupt_handler_size |
||
302 | |||
820 | jermar | 303 | interrupt_handler_size: .quad (h_end-h_start)/IDT_ITEMS |