Rev 3380 | Go to most recent revision | Details | Compare with Previous | Last modification | View Log | RSS feed
| Rev | Author | Line No. | Line |
|---|---|---|---|
| 633 | palkovsky | 1 | /* |
| 2071 | jermar | 2 | * Copyright (c) 2001-2004 Jakub Jermar |
| 633 | palkovsky | 3 | * All rights reserved. |
| 4 | * |
||
| 5 | * Redistribution and use in source and binary forms, with or without |
||
| 6 | * modification, are permitted provided that the following conditions |
||
| 7 | * are met: |
||
| 8 | * |
||
| 9 | * - Redistributions of source code must retain the above copyright |
||
| 10 | * notice, this list of conditions and the following disclaimer. |
||
| 11 | * - Redistributions in binary form must reproduce the above copyright |
||
| 12 | * notice, this list of conditions and the following disclaimer in the |
||
| 13 | * documentation and/or other materials provided with the distribution. |
||
| 14 | * - The name of the author may not be used to endorse or promote products |
||
| 15 | * derived from this software without specific prior written permission. |
||
| 16 | * |
||
| 17 | * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR |
||
| 18 | * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES |
||
| 19 | * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. |
||
| 20 | * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, |
||
| 21 | * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT |
||
| 22 | * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
||
| 23 | * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
||
| 24 | * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
||
| 25 | * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF |
||
| 26 | * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
||
| 27 | */ |
||
| 28 | |||
| 4017 | decky | 29 | /** @addtogroup amd64 |
| 1702 | cejka | 30 | * @{ |
| 31 | */ |
||
| 32 | /** @file |
||
| 33 | */ |
||
| 34 | |||
| 1888 | jermar | 35 | #ifndef KERN_amd64_ATOMIC_H_ |
| 36 | #define KERN_amd64_ATOMIC_H_ |
||
| 633 | palkovsky | 37 | |
| 38 | #include <arch/types.h> |
||
| 1100 | palkovsky | 39 | #include <arch/barrier.h> |
| 40 | #include <preemption.h> |
||
| 633 | palkovsky | 41 | |
| 42 | static inline void atomic_inc(atomic_t *val) { |
||
| 43 | #ifdef CONFIG_SMP |
||
| 4017 | decky | 44 | asm volatile ( |
| 45 | "lock incq %[count]\n" |
||
| 46 | : [count] "+m" (val->count) |
||
| 47 | ); |
||
| 633 | palkovsky | 48 | #else |
| 4017 | decky | 49 | asm volatile ( |
| 50 | "incq %[count]\n" |
||
| 51 | : [count] "+m" (val->count) |
||
| 52 | ); |
||
| 633 | palkovsky | 53 | #endif /* CONFIG_SMP */ |
| 54 | } |
||
| 55 | |||
| 56 | static inline void atomic_dec(atomic_t *val) { |
||
| 57 | #ifdef CONFIG_SMP |
||
| 4017 | decky | 58 | asm volatile ( |
| 59 | "lock decq %[count]\n" |
||
| 60 | : [count] "+m" (val->count) |
||
| 61 | ); |
||
| 633 | palkovsky | 62 | #else |
| 4017 | decky | 63 | asm volatile ( |
| 64 | "decq %[count]\n" |
||
| 65 | : [count] "+m" (val->count) |
||
| 66 | ); |
||
| 633 | palkovsky | 67 | #endif /* CONFIG_SMP */ |
| 68 | } |
||
| 69 | |||
| 1104 | jermar | 70 | static inline long atomic_postinc(atomic_t *val) |
| 633 | palkovsky | 71 | { |
| 1692 | palkovsky | 72 | long r = 1; |
| 4017 | decky | 73 | |
| 2082 | decky | 74 | asm volatile ( |
| 4017 | decky | 75 | "lock xaddq %[r], %[count]\n" |
| 76 | : [count] "+m" (val->count), [r] "+r" (r) |
||
| 633 | palkovsky | 77 | ); |
| 4017 | decky | 78 | |
| 633 | palkovsky | 79 | return r; |
| 80 | } |
||
| 81 | |||
| 1104 | jermar | 82 | static inline long atomic_postdec(atomic_t *val) |
| 633 | palkovsky | 83 | { |
| 1692 | palkovsky | 84 | long r = -1; |
| 633 | palkovsky | 85 | |
| 2082 | decky | 86 | asm volatile ( |
| 4017 | decky | 87 | "lock xaddq %[r], %[count]\n" |
| 88 | : [count] "+m" (val->count), [r] "+r" (r) |
||
| 633 | palkovsky | 89 | ); |
| 90 | |||
| 91 | return r; |
||
| 92 | } |
||
| 93 | |||
| 4017 | decky | 94 | #define atomic_preinc(val) (atomic_postinc(val) + 1) |
| 95 | #define atomic_predec(val) (atomic_postdec(val) - 1) |
||
| 633 | palkovsky | 96 | |
| 1780 | jermar | 97 | static inline uint64_t test_and_set(atomic_t *val) { |
| 98 | uint64_t v; |
||
| 633 | palkovsky | 99 | |
| 2082 | decky | 100 | asm volatile ( |
| 4017 | decky | 101 | "movq $1, %[v]\n" |
| 102 | "xchgq %[v], %[count]\n" |
||
| 103 | : [v] "=r" (v), [count] "+m" (val->count) |
||
| 633 | palkovsky | 104 | ); |
| 105 | |||
| 106 | return v; |
||
| 107 | } |
||
| 108 | |||
| 109 | |||
| 1104 | jermar | 110 | /** amd64 specific fast spinlock */ |
| 1100 | palkovsky | 111 | static inline void atomic_lock_arch(atomic_t *val) |
| 112 | { |
||
| 1780 | jermar | 113 | uint64_t tmp; |
| 4017 | decky | 114 | |
| 1100 | palkovsky | 115 | preemption_disable(); |
| 2082 | decky | 116 | asm volatile ( |
| 3163 | jermar | 117 | "0:\n" |
| 1100 | palkovsky | 118 | #ifdef CONFIG_HT |
| 3163 | jermar | 119 | "pause\n" |
| 633 | palkovsky | 120 | #endif |
| 4017 | decky | 121 | "mov %[count], %[tmp]\n" |
| 122 | "testq %[tmp], %[tmp]\n" |
||
| 3380 | jermar | 123 | "jnz 0b\n" /* lightweight looping on locked spinlock */ |
| 1100 | palkovsky | 124 | |
| 4017 | decky | 125 | "incq %[tmp]\n" /* now use the atomic operation */ |
| 126 | "xchgq %[count], %[tmp]\n" |
||
| 127 | "testq %[tmp], %[tmp]\n" |
||
| 3163 | jermar | 128 | "jnz 0b\n" |
| 4017 | decky | 129 | : [count] "+m" (val->count), [tmp] "=&r" (tmp) |
| 3163 | jermar | 130 | ); |
| 1100 | palkovsky | 131 | /* |
| 132 | * Prevent critical section code from bleeding out this way up. |
||
| 133 | */ |
||
| 134 | CS_ENTER_BARRIER(); |
||
| 135 | } |
||
| 136 | |||
| 137 | #endif |
||
| 1702 | cejka | 138 | |
| 1888 | jermar | 139 | /** @} |
| 1702 | cejka | 140 | */ |