/SPARTAN/trunk/src/proc/scheduler.c |
---|
149,7 → 149,7 |
CPU->nrdy--; |
spinlock_unlock(&CPU->lock); |
atomic_dec(&nrdy); |
atomic_dec((int *) &nrdy); |
r->n--; |
/* |
/SPARTAN/trunk/src/proc/thread.c |
---|
136,7 → 136,7 |
r->n++; |
spinlock_unlock(&r->lock); |
atomic_inc(&nrdy); |
atomic_inc((int *) &nrdy); |
avg = nrdy / config.cpu_active; |
spinlock_lock(&cpu->lock); |
/SPARTAN/trunk/arch/amd64/include/atomic.h |
---|
30,8 → 30,7 |
#define __amd64_ATOMIC_H__ |
/* Count_t is 32-bits on AMD-64 */ |
static inline void atomic_inc(volatile count_t *val) { |
static inline void atomic_inc(volatile int *val) { |
#ifdef __SMP__ |
__asm__ volatile ("lock incl (%0)\n" : : "r" (val)); |
#else |
39,7 → 38,7 |
#endif /* __SMP__ */ |
} |
static inline void atomic_dec(volatile count_t *val) { |
static inline void atomic_dec(volatile int *val) { |
#ifdef __SMP__ |
__asm__ volatile ("lock decl (%0)\n" : : "r" (val)); |
#else |
/SPARTAN/trunk/arch/ia32/include/atomic.h |
---|
31,7 → 31,7 |
#include <arch/types.h> |
static inline void atomic_inc(volatile count_t *val) { |
static inline void atomic_inc(volatile int *val) { |
#ifdef __SMP__ |
__asm__ volatile ("lock incl (%0)\n" : : "r" (val)); |
#else |
39,7 → 39,7 |
#endif /* __SMP__ */ |
} |
static inline void atomic_dec(volatile count_t *val) { |
static inline void atomic_dec(volatile int *val) { |
#ifdef __SMP__ |
__asm__ volatile ("lock decl (%0)\n" : : "r" (val)); |
#else |