Subversion Repositories HelenOS

Rev

Rev 2116 | Rev 3940 | Go to most recent revision | Only display areas with differences | Ignore whitespace | Details | Blame | Last modification | View Log | RSS feed

Rev 2116 Rev 2141
1
/*
1
/*
2
 * Copyright (c) 2006 Jakub Jermar
2
 * Copyright (c) 2006 Jakub Jermar
3
 * All rights reserved.
3
 * All rights reserved.
4
 *
4
 *
5
 * Redistribution and use in source and binary forms, with or without
5
 * Redistribution and use in source and binary forms, with or without
6
 * modification, are permitted provided that the following conditions
6
 * modification, are permitted provided that the following conditions
7
 * are met:
7
 * are met:
8
 *
8
 *
9
 * - Redistributions of source code must retain the above copyright
9
 * - Redistributions of source code must retain the above copyright
10
 *   notice, this list of conditions and the following disclaimer.
10
 *   notice, this list of conditions and the following disclaimer.
11
 * - Redistributions in binary form must reproduce the above copyright
11
 * - Redistributions in binary form must reproduce the above copyright
12
 *   notice, this list of conditions and the following disclaimer in the
12
 *   notice, this list of conditions and the following disclaimer in the
13
 *   documentation and/or other materials provided with the distribution.
13
 *   documentation and/or other materials provided with the distribution.
14
 * - The name of the author may not be used to endorse or promote products
14
 * - The name of the author may not be used to endorse or promote products
15
 *   derived from this software without specific prior written permission.
15
 *   derived from this software without specific prior written permission.
16
 *
16
 *
17
 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
17
 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
18
 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
18
 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
19
 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
19
 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
20
 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
20
 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
21
 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
21
 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
22
 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
22
 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
23
 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
23
 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
24
 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
24
 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
25
 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
25
 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
26
 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
26
 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
27
 */
27
 */
28
 
28
 
29
/** @addtogroup genericddi
29
/** @addtogroup genericddi
30
 * @{
30
 * @{
31
 */
31
 */
32
 
32
 
33
/**
33
/**
34
 * @file
34
 * @file
35
 * @brief   Device Driver Interface functions.
35
 * @brief   Device Driver Interface functions.
36
 *
36
 *
37
 * This file contains functions that comprise the Device Driver Interface.
37
 * This file contains functions that comprise the Device Driver Interface.
38
 * These are the functions for mapping physical memory and enabling I/O
38
 * These are the functions for mapping physical memory and enabling I/O
39
 * space to tasks.
39
 * space to tasks.
40
 */
40
 */
41
 
41
 
42
#include <ddi/ddi.h>
42
#include <ddi/ddi.h>
43
#include <ddi/ddi_arg.h>
43
#include <ddi/ddi_arg.h>
44
#include <proc/task.h>
44
#include <proc/task.h>
45
#include <security/cap.h>
45
#include <security/cap.h>
46
#include <mm/frame.h>
46
#include <mm/frame.h>
47
#include <mm/as.h>
47
#include <mm/as.h>
48
#include <synch/spinlock.h>
48
#include <synch/spinlock.h>
49
#include <syscall/copy.h>
49
#include <syscall/copy.h>
50
#include <adt/btree.h>
50
#include <adt/btree.h>
51
#include <arch.h>
51
#include <arch.h>
52
#include <align.h>
52
#include <align.h>
53
#include <errno.h>
53
#include <errno.h>
54
 
54
 
55
/** This lock protects the parea_btree. */
55
/** This lock protects the parea_btree. */
56
SPINLOCK_INITIALIZE(parea_lock);
56
SPINLOCK_INITIALIZE(parea_lock);
57
 
57
 
58
/** B+tree with enabled physical memory areas. */
58
/** B+tree with enabled physical memory areas. */
59
static btree_t parea_btree;
59
static btree_t parea_btree;
60
 
60
 
61
/** Initialize DDI. */
61
/** Initialize DDI. */
62
void ddi_init(void)
62
void ddi_init(void)
63
{
63
{
64
    btree_create(&parea_btree);
64
    btree_create(&parea_btree);
65
}
65
}
66
 
66
 
67
/** Enable piece of physical memory for mapping by physmem_map().
67
/** Enable piece of physical memory for mapping by physmem_map().
68
 *
68
 *
69
 * @param parea Pointer to physical area structure.
69
 * @param parea Pointer to physical area structure.
70
 *
70
 *
71
 * @todo This function doesn't check for overlaps. It depends on the kernel to
71
 * @todo This function doesn't check for overlaps. It depends on the kernel to
72
 * create disjunct physical memory areas.
72
 * create disjunct physical memory areas.
73
 */
73
 */
74
void ddi_parea_register(parea_t *parea)
74
void ddi_parea_register(parea_t *parea)
75
{
75
{
76
    ipl_t ipl;
76
    ipl_t ipl;
77
 
77
 
78
    ipl = interrupts_disable();
78
    ipl = interrupts_disable();
79
    spinlock_lock(&parea_lock);
79
    spinlock_lock(&parea_lock);
80
   
80
   
81
    /*
81
    /*
82
     * TODO: we should really check for overlaps here.
82
     * TODO: we should really check for overlaps here.
83
     * However, we should be safe because the kernel is pretty sane and
83
     * However, we should be safe because the kernel is pretty sane and
84
     * memory of different devices doesn't overlap.
84
     * memory of different devices doesn't overlap.
85
     */
85
     */
86
    btree_insert(&parea_btree, (btree_key_t) parea->pbase, parea, NULL);
86
    btree_insert(&parea_btree, (btree_key_t) parea->pbase, parea, NULL);
87
 
87
 
88
    spinlock_unlock(&parea_lock);
88
    spinlock_unlock(&parea_lock);
89
    interrupts_restore(ipl);   
89
    interrupts_restore(ipl);   
90
}
90
}
91
 
91
 
92
/** Map piece of physical memory into virtual address space of current task.
92
/** Map piece of physical memory into virtual address space of current task.
93
 *
93
 *
94
 * @param pf Physical address of the starting frame.
94
 * @param pf Physical address of the starting frame.
95
 * @param vp Virtual address of the starting page.
95
 * @param vp Virtual address of the starting page.
96
 * @param pages Number of pages to map.
96
 * @param pages Number of pages to map.
97
 * @param flags Address space area flags for the mapping.
97
 * @param flags Address space area flags for the mapping.
98
 *
98
 *
99
 * @return 0 on success, EPERM if the caller lacks capabilities to use this
99
 * @return 0 on success, EPERM if the caller lacks capabilities to use this
100
 *  syscall, ENOENT if there is no task matching the specified ID or the
100
 *  syscall, ENOENT if there is no task matching the specified ID or the
101
 *  physical address space is not enabled for mapping and ENOMEM if there
101
 *  physical address space is not enabled for mapping and ENOMEM if there
102
 *  was a problem in creating address space area. ENOTSUP is returned when
102
 *  was a problem in creating address space area.
103
 *  an attempt to create an illegal address alias is detected.
-
 
104
 */
103
 */
105
static int ddi_physmem_map(uintptr_t pf, uintptr_t vp, count_t pages, int flags)
104
static int ddi_physmem_map(uintptr_t pf, uintptr_t vp, count_t pages, int flags)
106
{
105
{
107
    ipl_t ipl;
106
    ipl_t ipl;
108
    cap_t caps;
107
    cap_t caps;
109
    mem_backend_data_t backend_data;
108
    mem_backend_data_t backend_data;
110
 
109
 
111
    backend_data.base = pf;
110
    backend_data.base = pf;
112
    backend_data.frames = pages;
111
    backend_data.frames = pages;
113
   
112
   
114
    /*
113
    /*
115
     * Make sure the caller is authorised to make this syscall.
114
     * Make sure the caller is authorised to make this syscall.
116
     */
115
     */
117
    caps = cap_get(TASK);
116
    caps = cap_get(TASK);
118
    if (!(caps & CAP_MEM_MANAGER))
117
    if (!(caps & CAP_MEM_MANAGER))
119
        return EPERM;
118
        return EPERM;
120
 
119
 
121
    ipl = interrupts_disable();
120
    ipl = interrupts_disable();
122
 
121
 
123
    /*
122
    /*
124
     * Check if the physical memory area is enabled for mapping.
123
     * Check if the physical memory area is enabled for mapping.
125
     * If the architecture supports virtually indexed caches, intercept
124
     * If the architecture supports virtually indexed caches, intercept
126
     * attempts to create an illegal address alias.
125
     * attempts to create an illegal address alias.
127
     */
126
     */
128
    spinlock_lock(&parea_lock);
127
    spinlock_lock(&parea_lock);
129
    parea_t *parea;
128
    parea_t *parea;
130
    btree_node_t *nodep;
129
    btree_node_t *nodep;
131
    parea = (parea_t *) btree_search(&parea_btree, (btree_key_t) pf, &nodep);
130
    parea = (parea_t *) btree_search(&parea_btree, (btree_key_t) pf, &nodep);
132
    if (!parea || parea->frames < pages || ((flags & AS_AREA_CACHEABLE) &&
131
    if (!parea || parea->frames < pages || ((flags & AS_AREA_CACHEABLE) &&
133
        !parea->cacheable) || (!(flags & AS_AREA_CACHEABLE) &&
132
        !parea->cacheable) || (!(flags & AS_AREA_CACHEABLE) &&
134
        parea->cacheable)) {
133
        parea->cacheable)) {
135
        /*
134
        /*
136
         * This physical memory area cannot be mapped.
135
         * This physical memory area cannot be mapped.
137
         */
136
         */
138
        spinlock_unlock(&parea_lock);
137
        spinlock_unlock(&parea_lock);
139
        interrupts_restore(ipl);
138
        interrupts_restore(ipl);
140
        return ENOENT;
139
        return ENOENT;
141
    }
140
    }
142
 
-
 
143
#ifdef CONFIG_VIRT_IDX_DCACHE
-
 
144
    if (PAGE_COLOR(parea->vbase) != PAGE_COLOR(vp)) {
-
 
145
        /*
-
 
146
         * Refuse to create an illegal address alias.
-
 
147
         */
-
 
148
        spinlock_unlock(&parea_lock);
-
 
149
        interrupts_restore(ipl);
-
 
150
        return ENOTSUP;
-
 
151
    }
-
 
152
#endif /* CONFIG_VIRT_IDX_DCACHE */
-
 
153
 
-
 
154
    spinlock_unlock(&parea_lock);
141
    spinlock_unlock(&parea_lock);
155
 
142
 
156
    spinlock_lock(&TASK->lock);
143
    spinlock_lock(&TASK->lock);
157
   
144
   
158
    if (!as_area_create(TASK->as, flags, pages * PAGE_SIZE, vp, AS_AREA_ATTR_NONE,
145
    if (!as_area_create(TASK->as, flags, pages * PAGE_SIZE, vp, AS_AREA_ATTR_NONE,
159
        &phys_backend, &backend_data)) {
146
        &phys_backend, &backend_data)) {
160
        /*
147
        /*
161
         * The address space area could not have been created.
148
         * The address space area could not have been created.
162
         * We report it using ENOMEM.
149
         * We report it using ENOMEM.
163
         */
150
         */
164
        spinlock_unlock(&TASK->lock);
151
        spinlock_unlock(&TASK->lock);
165
        interrupts_restore(ipl);
152
        interrupts_restore(ipl);
166
        return ENOMEM;
153
        return ENOMEM;
167
    }
154
    }
168
   
155
   
169
    /*
156
    /*
170
     * Mapping is created on-demand during page fault.
157
     * Mapping is created on-demand during page fault.
171
     */
158
     */
172
   
159
   
173
    spinlock_unlock(&TASK->lock);
160
    spinlock_unlock(&TASK->lock);
174
    interrupts_restore(ipl);
161
    interrupts_restore(ipl);
175
    return 0;
162
    return 0;
176
}
163
}
177
 
164
 
178
/** Enable range of I/O space for task.
165
/** Enable range of I/O space for task.
179
 *
166
 *
180
 * @param id Task ID of the destination task.
167
 * @param id Task ID of the destination task.
181
 * @param ioaddr Starting I/O address.
168
 * @param ioaddr Starting I/O address.
182
 * @param size Size of the enabled I/O space..
169
 * @param size Size of the enabled I/O space..
183
 *
170
 *
184
 * @return 0 on success, EPERM if the caller lacks capabilities to use this
171
 * @return 0 on success, EPERM if the caller lacks capabilities to use this
185
 *  syscall, ENOENT if there is no task matching the specified ID.
172
 *  syscall, ENOENT if there is no task matching the specified ID.
186
 */
173
 */
187
static int ddi_iospace_enable(task_id_t id, uintptr_t ioaddr, size_t size)
174
static int ddi_iospace_enable(task_id_t id, uintptr_t ioaddr, size_t size)
188
{
175
{
189
    ipl_t ipl;
176
    ipl_t ipl;
190
    cap_t caps;
177
    cap_t caps;
191
    task_t *t;
178
    task_t *t;
192
    int rc;
179
    int rc;
193
   
180
   
194
    /*
181
    /*
195
     * Make sure the caller is authorised to make this syscall.
182
     * Make sure the caller is authorised to make this syscall.
196
     */
183
     */
197
    caps = cap_get(TASK);
184
    caps = cap_get(TASK);
198
    if (!(caps & CAP_IO_MANAGER))
185
    if (!(caps & CAP_IO_MANAGER))
199
        return EPERM;
186
        return EPERM;
200
   
187
   
201
    ipl = interrupts_disable();
188
    ipl = interrupts_disable();
202
    spinlock_lock(&tasks_lock);
189
    spinlock_lock(&tasks_lock);
203
   
190
   
204
    t = task_find_by_id(id);
191
    t = task_find_by_id(id);
205
   
192
   
206
    if ((!t) || (!context_check(CONTEXT, t->context))) {
193
    if ((!t) || (!context_check(CONTEXT, t->context))) {
207
        /*
194
        /*
208
         * There is no task with the specified ID
195
         * There is no task with the specified ID
209
         * or the task belongs to a different security
196
         * or the task belongs to a different security
210
         * context.
197
         * context.
211
         */
198
         */
212
        spinlock_unlock(&tasks_lock);
199
        spinlock_unlock(&tasks_lock);
213
        interrupts_restore(ipl);
200
        interrupts_restore(ipl);
214
        return ENOENT;
201
        return ENOENT;
215
    }
202
    }
216
 
203
 
217
    /* Lock the task and release the lock protecting tasks_btree. */
204
    /* Lock the task and release the lock protecting tasks_btree. */
218
    spinlock_lock(&t->lock);
205
    spinlock_lock(&t->lock);
219
    spinlock_unlock(&tasks_lock);
206
    spinlock_unlock(&tasks_lock);
220
 
207
 
221
    rc = ddi_iospace_enable_arch(t, ioaddr, size);
208
    rc = ddi_iospace_enable_arch(t, ioaddr, size);
222
   
209
   
223
    spinlock_unlock(&t->lock);
210
    spinlock_unlock(&t->lock);
224
    interrupts_restore(ipl);
211
    interrupts_restore(ipl);
225
    return rc;
212
    return rc;
226
}
213
}
227
 
214
 
228
/** Wrapper for SYS_PHYSMEM_MAP syscall.
215
/** Wrapper for SYS_PHYSMEM_MAP syscall.
229
 *
216
 *
230
 * @param phys_base Physical base address to map
217
 * @param phys_base Physical base address to map
231
 * @param virt_base Destination virtual address
218
 * @param virt_base Destination virtual address
232
 * @param pages Number of pages
219
 * @param pages Number of pages
233
 * @param flags Flags of newly mapped pages
220
 * @param flags Flags of newly mapped pages
234
 *
221
 *
235
 * @return 0 on success, otherwise it returns error code found in errno.h
222
 * @return 0 on success, otherwise it returns error code found in errno.h
236
 */
223
 */
237
unative_t sys_physmem_map(unative_t phys_base, unative_t virt_base,
224
unative_t sys_physmem_map(unative_t phys_base, unative_t virt_base,
238
    unative_t pages, unative_t flags)
225
    unative_t pages, unative_t flags)
239
{
226
{
240
    return (unative_t) ddi_physmem_map(ALIGN_DOWN((uintptr_t) phys_base,
227
    return (unative_t) ddi_physmem_map(ALIGN_DOWN((uintptr_t) phys_base,
241
        FRAME_SIZE), ALIGN_DOWN((uintptr_t) virt_base, PAGE_SIZE),
228
        FRAME_SIZE), ALIGN_DOWN((uintptr_t) virt_base, PAGE_SIZE),
242
        (count_t) pages, (int) flags);
229
        (count_t) pages, (int) flags);
243
}
230
}
244
 
231
 
245
/** Wrapper for SYS_ENABLE_IOSPACE syscall.
232
/** Wrapper for SYS_ENABLE_IOSPACE syscall.
246
 *
233
 *
247
 * @param uspace_io_arg User space address of DDI argument structure.
234
 * @param uspace_io_arg User space address of DDI argument structure.
248
 *
235
 *
249
 * @return 0 on success, otherwise it returns error code found in errno.h
236
 * @return 0 on success, otherwise it returns error code found in errno.h
250
 */
237
 */
251
unative_t sys_iospace_enable(ddi_ioarg_t *uspace_io_arg)
238
unative_t sys_iospace_enable(ddi_ioarg_t *uspace_io_arg)
252
{
239
{
253
    ddi_ioarg_t arg;
240
    ddi_ioarg_t arg;
254
    int rc;
241
    int rc;
255
   
242
   
256
    rc = copy_from_uspace(&arg, uspace_io_arg, sizeof(ddi_ioarg_t));
243
    rc = copy_from_uspace(&arg, uspace_io_arg, sizeof(ddi_ioarg_t));
257
    if (rc != 0)
244
    if (rc != 0)
258
        return (unative_t) rc;
245
        return (unative_t) rc;
259
       
246
       
260
    return (unative_t) ddi_iospace_enable((task_id_t) arg.task_id,
247
    return (unative_t) ddi_iospace_enable((task_id_t) arg.task_id,
261
        (uintptr_t) arg.ioaddr, (size_t) arg.size);
248
        (uintptr_t) arg.ioaddr, (size_t) arg.size);
262
}
249
}
263
 
250
 
264
/** Disable or enable preemption.
251
/** Disable or enable preemption.
265
 *
252
 *
266
 * @param enable If non-zero, the preemption counter will be decremented,
253
 * @param enable If non-zero, the preemption counter will be decremented,
267
 *  leading to potential enabling of preemption. Otherwise the preemption
254
 *  leading to potential enabling of preemption. Otherwise the preemption
268
 *  counter will be incremented, preventing preemption from occurring.
255
 *  counter will be incremented, preventing preemption from occurring.
269
 *
256
 *
270
 * @return Zero on success or EPERM if callers capabilities are not sufficient.
257
 * @return Zero on success or EPERM if callers capabilities are not sufficient.
271
 */
258
 */
272
unative_t sys_preempt_control(int enable)
259
unative_t sys_preempt_control(int enable)
273
{
260
{
274
        if (!cap_get(TASK) & CAP_PREEMPT_CONTROL)
261
        if (!cap_get(TASK) & CAP_PREEMPT_CONTROL)
275
                return EPERM;
262
                return EPERM;
276
        if (enable)
263
        if (enable)
277
                preemption_enable();
264
                preemption_enable();
278
        else
265
        else
279
                preemption_disable();
266
                preemption_disable();
280
        return 0;
267
        return 0;
281
}
268
}
282
 
269
 
283
/** @}
270
/** @}
284
 */
271
 */
285
 
272