/kernel/trunk/generic/src/console/cmd.c |
---|
310,7 → 310,6 |
int cmd_help(cmd_arg_t *argv) |
{ |
link_t *cur; |
ipl_t ipl; |
spinlock_lock(&cmd_lock); |
339,7 → 338,6 |
int cmd_desc(cmd_arg_t *argv) |
{ |
link_t *cur; |
ipl_t ipl; |
spinlock_lock(&cmd_lock); |
368,9 → 366,6 |
/** Search symbol table */ |
int cmd_symaddr(cmd_arg_t *argv) |
{ |
__address symaddr; |
char *symbol; |
symtab_print_search(argv->buffer); |
return 1; |
510,7 → 505,6 |
/** Write 4 byte value to address */ |
int cmd_set4(cmd_arg_t *argv) |
{ |
char *symbol; |
__u32 *addr ; |
__u32 arg1 = argv[1].intval; |
bool pointer = false; |
/kernel/trunk/generic/src/console/kconsole.c |
---|
91,7 → 91,6 |
*/ |
int cmd_register(cmd_info_t *cmd) |
{ |
ipl_t ipl; |
link_t *cur; |
spinlock_lock(&cmd_lock); |
160,8 → 159,6 |
{ |
int namelen = strlen(name); |
const char *curname; |
char *foundsym = NULL; |
int foundpos = 0; |
spinlock_lock(&cmd_lock); |
469,7 → 466,6 |
index_t start = 0, end = 0; |
cmd_info_t *cmd = NULL; |
link_t *cur; |
ipl_t ipl; |
int i; |
if (!parse_argument(cmdline, len, &start, &end)) { |
/kernel/trunk/generic/src/proc/scheduler.c |
---|
490,7 → 490,7 |
*/ |
ipl = interrupts_disable(); |
spinlock_lock(&CPU->lock); |
count = nrdy / config.cpu_active; |
count = atomic_get(&nrdy) / config.cpu_active; |
count -= CPU->nrdy; |
spinlock_unlock(&CPU->lock); |
interrupts_restore(ipl); |
618,7 → 618,7 |
/* |
* Tell find_best_thread() to wake us up later again. |
*/ |
CPU->kcpulbstarted = 0; |
atomic_set(&CPU->kcpulbstarted,0); |
goto loop; |
} |
/kernel/trunk/generic/src/proc/thread.c |
---|
95,7 → 95,7 |
void thread_init(void) |
{ |
THREAD = NULL; |
nrdy = 0; |
atomic_set(&nrdy,0); |
} |
111,7 → 111,7 |
cpu_t *cpu; |
runq_t *r; |
ipl_t ipl; |
int i, avg, send_ipi = 0; |
int i, avg; |
ipl = interrupts_disable(); |
135,7 → 135,7 |
spinlock_unlock(&r->lock); |
atomic_inc(&nrdy); |
avg = nrdy / config.cpu_active; |
avg = atomic_get(&nrdy) / config.cpu_active; |
spinlock_lock(&cpu->lock); |
if ((++cpu->nrdy) > avg) { |
/kernel/trunk/generic/src/main/kinit.c |
---|
68,11 → 68,13 |
*/ |
void kinit(void *arg) |
{ |
thread_t *t; |
int i; |
#ifdef CONFIG_USERSPACE |
vm_t *m; |
vm_area_t *a; |
task_t *u; |
thread_t *t; |
int i; |
#endif |
interrupts_disable(); |
/kernel/trunk/generic/src/main/main.c |
---|
125,7 → 125,9 |
* pop sequence otherwise. |
*/ |
static void main_bsp_separated_stack(void); |
#ifdef CONFIG_SMP |
static void main_ap_separated_stack(void); |
#endif |
/** Bootstrap CPU main kernel routine |
* |
/kernel/trunk/generic/src/synch/rwlock.c |
---|
225,7 → 225,7 |
case ESYNCH_OK_ATOMIC: |
panic("_mutex_lock_timeout()==ESYNCH_OK_ATOMIC\n"); |
break; |
dafault: |
default: |
panic("invalid ESYNCH\n"); |
break; |
} |
/kernel/trunk/generic/src/synch/spinlock.c |
---|
45,7 → 45,7 |
*/ |
void spinlock_initialize(spinlock_t *sl, char *name) |
{ |
sl->val = 0; |
atomic_set(&sl->val, 0); |
#ifdef CONFIG_DEBUG_SPINLOCK |
sl->name = name; |
#endif |
151,7 → 151,7 |
*/ |
void spinlock_unlock(spinlock_t *sl) |
{ |
ASSERT(sl->val != 0); |
ASSERT(atomic_get(&sl->val) != 0); |
/* |
* Prevent critical section code from bleeding out this way down. |
158,7 → 158,7 |
*/ |
CS_LEAVE_BARRIER(); |
sl->val = 0; |
atomic_set(&sl->val,0); |
preemption_enable(); |
} |
/kernel/trunk/generic/src/debug/symtab.c |
---|
66,8 → 66,6 |
int namelen = strlen(name); |
char *curname; |
int i,j; |
char *foundsym = NULL; |
int foundpos = 0; |
int colonoffset = -1; |
for (i=0;name[i];i++) |
/kernel/trunk/generic/src/mm/vm.c |
---|
188,7 → 188,6 |
void vm_install(vm_t *m) |
{ |
link_t *l; |
ipl_t ipl; |
ipl = interrupts_disable(); |