Subversion Repositories HelenOS-historic

Rev

Rev 1295 | Only display areas with differences | Ignore whitespace | Details | Blame | Last modification | View Log | RSS feed

Rev 1295 Rev 1611
1
#
1
#
2
# Copyright (C) 2005 Jakub Jermar
2
# Copyright (C) 2005 Jakub Jermar
3
# All rights reserved.
3
# All rights reserved.
4
#
4
#
5
# Redistribution and use in source and binary forms, with or without
5
# Redistribution and use in source and binary forms, with or without
6
# modification, are permitted provided that the following conditions
6
# modification, are permitted provided that the following conditions
7
# are met:
7
# are met:
8
#
8
#
9
# - Redistributions of source code must retain the above copyright
9
# - Redistributions of source code must retain the above copyright
10
#   notice, this list of conditions and the following disclaimer.
10
#   notice, this list of conditions and the following disclaimer.
11
# - Redistributions in binary form must reproduce the above copyright
11
# - Redistributions in binary form must reproduce the above copyright
12
#   notice, this list of conditions and the following disclaimer in the
12
#   notice, this list of conditions and the following disclaimer in the
13
#   documentation and/or other materials provided with the distribution.
13
#   documentation and/or other materials provided with the distribution.
14
# - The name of the author may not be used to endorse or promote products
14
# - The name of the author may not be used to endorse or promote products
15
#   derived from this software without specific prior written permission.
15
#   derived from this software without specific prior written permission.
16
#
16
#
17
# THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
17
# THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
18
# IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
18
# IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
19
# OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
19
# OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
20
# IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
20
# IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
21
# INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
21
# INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
22
# NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
22
# NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
23
# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
23
# DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
24
# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
24
# THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
25
# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
25
# (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
26
# THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
26
# THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
27
#
27
#
28
 
28
 
29
#include <arch/register.h>
29
#include <arch/register.h>
30
 
30
 
31
.text
31
.text
32
 
32
 
33
/** Copy memory from/to userspace.
33
/** Copy memory from/to userspace.
34
 *
34
 *
35
 * This memcpy() has been taken from the assembler output of
35
 * This memcpy() has been taken from the assembler output of
36
 * the generic _memcpy() and modified to have the failover part.
36
 * the generic _memcpy() and modified to have the failover part.
37
 *
37
 *
38
 * @param in0 Destination address.
38
 * @param in0 Destination address.
39
 * @param in1 Source address.
39
 * @param in1 Source address.
40
 * @param in2 Number of byte to copy.
40
 * @param in2 Number of byte to copy.
41
 */
41
 */
42
.global memcpy
42
.global memcpy
43
.global memcpy_from_uspace
43
.global memcpy_from_uspace
44
.global memcpy_to_uspace
44
.global memcpy_to_uspace
45
.global memcpy_from_uspace_failover_address
45
.global memcpy_from_uspace_failover_address
46
.global memcpy_to_uspace_failover_address
46
.global memcpy_to_uspace_failover_address
47
memcpy:
47
memcpy:
48
memcpy_from_uspace:
48
memcpy_from_uspace:
49
memcpy_to_uspace:
49
memcpy_to_uspace:
50
	alloc loc0 = ar.pfs, 3, 1, 0, 0
50
	alloc loc0 = ar.pfs, 3, 1, 0, 0
51
 
51
 
52
	shr.u r18 = in2, 3
52
	adds r14 = 7, in1
53
	mov r8 = in1 ;;			/* prepare to return in1 on success */
-
 
54
	cmp.ne p6, p7 = 0, r18	;;
53
	mov r2 = ar.lc
55
(p7)	mov r17 = r0
54
	mov r8 = in1 ;;
56
(p7)	br.cond.dptk 1f ;;
55
	and r14 = -8, r14 ;;
57
	mov r16 = r0
56
	cmp.ne p6, p7 = r14, in1
58
	mov r17 = r0 ;;
57
(p7)	br.cond.dpnt 3f	;;
59
 
-
 
60
0:
58
0:
61
	shladd r14 = r17, 3, r0
59
	cmp.ne p6, p7 = 0, in2
62
	adds r16 = 1, r16 ;;
60
(p7)	br.cond.dpnt 2f	;;
63
	add r15 = r14, r8
61
(p6)	adds r14 = -1, in2
64
	sxt4 r17 = r16
62
(p6)	mov r16 = r0
65
	add r14 = r14, in0 ;;
63
(p6)	mov r17 = r0 ;;
66
	ld8 r15 = [r15]
-
 
67
	cmp.gtu p6, p7 = r18, r17 ;;
-
 
68
	st8 [r14] = r15
64
(p6)	mov ar.lc = r14
69
(p6)	br.cond.dptk 0b
-
 
70
 
-
 
71
1:
65
1:
72
	and in2 = 7, in2 ;;
66
	add r14 = r16, r8
73
	cmp.eq p6, p7 = 0, in2 ;;
67
	add r15 = r16, in0
74
(p6)	mov ar.pfs = loc0
-
 
75
(p6)	br.ret.dptk.many rp
-
 
76
	shladd r14 = r17, 3, r0
68
	adds r17 = 1, r17 ;;
77
	mov r16 = r0
69
	ld1 r14 = [r14]
78
	mov r17 = r0 ;;
70
	mov r16 = r17 ;;
79
	add in0 = in0, r14
71
	st1 [r15] = r14
80
	add r18 = r8, r14 ;;
72
	br.cloop.sptk.few 1b ;;
81
 
-
 
82
2:
73
2:
-
 
74
	mov ar.lc = r2
-
 
75
 
-
 
76
	mov ar.pfs = loc0
-
 
77
	br.ret.sptk.many rp
-
 
78
3:
-
 
79
	adds r14 = 7, in0 ;;
-
 
80
	and r14 = -8, r14 ;;
-
 
81
	cmp.eq p6, p7 = r14, in0
-
 
82
(p7)	br.cond.dptk 0b
-
 
83
	shr.u r18 = in2, 3 ;;
-
 
84
	cmp.ne p6, p7 = 0, r18
-
 
85
(p7)	br.cond.dpnt 5f	;;
-
 
86
(p6)	adds r14 = -1, r18
-
 
87
(p6)	mov r16 = r0
-
 
88
(p6)	mov r17 = r0 ;;
-
 
89
(p6)	mov ar.lc = r14
-
 
90
4:
-
 
91
	shladd r14 = r16, 3, r0
-
 
92
	adds r16 = 1, r17 ;;
-
 
93
	add r15 = r8, r14
83
	add r14 = r16, r18
94
	add r14 = in0, r14
-
 
95
	mov r17 = r16 ;;
-
 
96
	ld8 r15 = [r15] ;;
-
 
97
	st8 [r14] = r15
-
 
98
	br.cloop.sptk.few 4b
-
 
99
5:
-
 
100
	and r15 = 7, in2
-
 
101
	shladd r14 = r18, 3, r0
-
 
102
	mov r16 = r0
-
 
103
	mov r18 = r0 ;;
-
 
104
	cmp.eq p6, p7 = 0, r15
-
 
105
	add in0 = r14, in0
84
	adds r17 = 1, r17
106
	adds r15 = -1, r15
-
 
107
	add r17 = r14, r8
-
 
108
(p6)	br.cond.dpnt 2b	;;
-
 
109
	mov ar.lc = r15
-
 
110
6:
-
 
111
	add r14 = r16, r17
-
 
112
	add r15 = r16, in0
85
	add r15 = in0, r16 ;;
113
	adds r16 = 1, r18 ;;
86
	ld1 r14 = [r14]
114
	ld1 r14 = [r14]
87
	sxt4 r16 = r17	;;
115
	mov r18 = r16 ;;
88
	st1 [r15] = r14
116
	st1 [r15] = r14
89
	cmp.gtu p6, p7 = in2, r16
117
	br.cloop.sptk.few 6b ;;
90
(p6)	br.cond.dptk 2b
118
	mov ar.lc = r2
91
 
119
 
92
	mov ar.pfs = loc0
120
	mov ar.pfs = loc0
93
	br.ret.sptk.many rp
121
	br.ret.sptk.many rp
94
	
122
	
95
memcpy_from_uspace_failover_address:
123
memcpy_from_uspace_failover_address:
96
memcpy_to_uspace_failover_address:
124
memcpy_to_uspace_failover_address:
97
	mov r8 = r0			/* return 0 on failure */
125
	mov r8 = r0			/* return 0 on failure */
98
	mov ar.pfs = loc0
126
	mov ar.pfs = loc0
99
	br.ret.sptk.many rp
127
	br.ret.sptk.many rp
100
 
128
 
101
.global memsetb
129
.global memsetb
102
memsetb:
130
memsetb:
103
	br _memsetb
131
	br _memsetb
104
 
132
 
105
.global cpu_halt
133
.global cpu_halt
106
cpu_halt:
134
cpu_halt:
107
	br cpu_halt
135
	br cpu_halt
108
 
136
 
109
.global panic_printf
137
.global panic_printf
110
panic_printf:
138
panic_printf:
111
	{
139
	{
112
		br.call.sptk.many b0=printf
140
		br.call.sptk.many b0=printf
113
	}
141
	}
114
	br halt
142
	br halt
115
 
143
 
116
/** Switch to userspace - low level code.
144
/** Switch to userspace - low level code.
117
 *
145
 *
118
 * @param in0 Userspace entry point address.
146
 * @param in0 Userspace entry point address.
119
 * @param in1 Userspace stack pointer address.
147
 * @param in1 Userspace stack pointer address.
120
 * @param in2 Userspace register stack pointer address.
148
 * @param in2 Userspace register stack pointer address.
121
 * @param in3 Userspace address of thread uspace_arg_t structure.
149
 * @param in3 Userspace address of thread uspace_arg_t structure.
122
 * @param in4 Value to be stored in IPSR.
150
 * @param in4 Value to be stored in IPSR.
123
 * @param in5 Value to be stored in RSC.
151
 * @param in5 Value to be stored in RSC.
124
 */
152
 */
125
.global switch_to_userspace
153
.global switch_to_userspace
126
switch_to_userspace:
154
switch_to_userspace:
127
	alloc loc0 = ar.pfs, 6, 3, 0, 0
155
	alloc loc0 = ar.pfs, 6, 3, 0, 0
128
	rsm (PSR_IC_MASK | PSR_I_MASK)		/* disable interruption collection and interrupts */
156
	rsm (PSR_IC_MASK | PSR_I_MASK)		/* disable interruption collection and interrupts */
129
	srlz.d ;;
157
	srlz.d ;;
130
	srlz.i ;;
158
	srlz.i ;;
131
	
159
	
132
	mov cr.ipsr = in4
160
	mov cr.ipsr = in4
133
	mov cr.iip = in0
161
	mov cr.iip = in0
134
	mov r12 = in1
162
	mov r12 = in1
135
 
163
 
136
	xor r1 = r1, r1
164
	xor r1 = r1, r1
137
	
165
	
138
	mov loc1 = cr.ifs
166
	mov loc1 = cr.ifs
139
	movl loc2 = PFM_MASK ;;
167
	movl loc2 = PFM_MASK ;;
140
	and loc1 = loc2, loc1 ;;
168
	and loc1 = loc2, loc1 ;;
141
	mov cr.ifs = loc1 ;;			/* prevent decrementing BSP by rfi */
169
	mov cr.ifs = loc1 ;;			/* prevent decrementing BSP by rfi */
142
 
170
 
143
	invala
171
	invala
144
	
172
	
145
	mov loc1 = ar.rsc ;;
173
	mov loc1 = ar.rsc ;;
146
	and loc1 = ~3, loc1 ;;			
174
	and loc1 = ~3, loc1 ;;			
147
	mov ar.rsc = loc1 ;;			/* put RSE into enforced lazy mode */
175
	mov ar.rsc = loc1 ;;			/* put RSE into enforced lazy mode */
148
 
176
 
149
	flushrs ;;
177
	flushrs ;;
150
	
178
	
151
	mov ar.bspstore = in2 ;;
179
	mov ar.bspstore = in2 ;;
152
	mov ar.rsc = in5 ;;
180
	mov ar.rsc = in5 ;;
153
	
181
	
154
	mov r8 = in3
182
	mov r8 = in3
155
	
183
	
156
	rfi ;;
184
	rfi ;;
157
 
185