/SPARTAN/trunk/arch/amd64/Makefile.inc |
---|
60,13 → 60,13 |
# |
ifeq ($(CONFIG_SMP),y) |
DEFS += -DSMP |
DEFS += -DCONFIG_SMP |
endif |
ifeq ($(CONFIG_HT),y) |
DEFS += -DHT |
DEFS += -DCONFIG_HT |
endif |
ifeq ($(CONFIG_FPU_LAZY),y) |
DEFS += -DFPU_LAZY |
DEFS += -DCONFIG_FPU_LAZY |
endif |
ARCH_SOURCES = \ |
/SPARTAN/trunk/arch/amd64/src/amd64.c |
---|
75,10 → 75,10 |
trap_register(VECTOR_SYSCALL, syscall); |
#ifdef __SMP__ |
#ifdef CONFIG_SMP |
trap_register(VECTOR_TLB_SHOOTDOWN_IPI, tlb_shootdown_ipi); |
trap_register(VECTOR_WAKEUP_IPI, wakeup_ipi); |
#endif /* __SMP__ */ |
#endif /* CONFIG_SMP */ |
} |
} |
94,9 → 94,9 |
if (config.cpu_active == 1) { |
memory_print_map(); |
#ifdef __SMP__ |
#ifdef CONFIG_SMP |
acpi_init(); |
#endif /* __SMP__ */ |
#endif /* CONFIG_SMP */ |
} |
} |
/SPARTAN/trunk/arch/amd64/src/smp/ap.S |
---|
39,7 → 39,7 |
.section K_TEXT_START_2, "ax" |
#ifdef __SMP__ |
#ifdef CONFIG_SMP |
.global ap_boot |
100,4 → 100,4 |
call main_ap # never returns |
#endif /* __SMP__ */ |
#endif /* CONFIG_SMP */ |
/SPARTAN/trunk/arch/amd64/src/interrupt.c |
---|
139,7 → 139,7 |
void nm_fault(__u8 n, __native stack[]) |
{ |
#ifdef FPU_LAZY |
#ifdef CONFIG_FPU_LAZY |
scheduler_fpu_lazy_request(); |
#else |
panic("fpu fault"); |
/SPARTAN/trunk/arch/mips32/Makefile.inc |
---|
102,7 → 102,7 |
# |
ifeq ($(CONFIG_FPU_LAZY),y) |
DEFS += -DFPU_LAZY |
DEFS += -DCONFIG_FPU_LAZY |
endif |
ARCH_SOURCES = \ |
/SPARTAN/trunk/arch/mips32/src/exception.c |
---|
73,7 → 73,7 |
tlb_invalid(pstate); |
break; |
case EXC_CpU: |
#ifdef FPU_LAZY |
#ifdef CONFIG_FPU_LAZY |
if (cp0_cause_coperr(cause) == fpu_cop_id) |
scheduler_fpu_lazy_request(); |
else |
/SPARTAN/trunk/arch/ia32/include/atomic.h |
---|
32,19 → 32,19 |
#include <arch/types.h> |
static inline void atomic_inc(volatile int *val) { |
#ifdef __SMP__ |
#ifdef CONFIG_SMP |
__asm__ volatile ("lock incl %0\n" : "=m" (*val)); |
#else |
__asm__ volatile ("incl %0\n" : "=m" (*val)); |
#endif /* __SMP__ */ |
#endif /* CONFIG_SMP */ |
} |
static inline void atomic_dec(volatile int *val) { |
#ifdef __SMP__ |
#ifdef CONFIG_SMP |
__asm__ volatile ("lock decl %0\n" : "=m" (*val)); |
#else |
__asm__ volatile ("decl %0\n" : "=m" (*val)); |
#endif /* __SMP__ */ |
#endif /* CONFIG_SMP */ |
} |
static inline int test_and_set(volatile int *val) { |
/SPARTAN/trunk/arch/ia32/Makefile.inc |
---|
79,13 → 79,13 |
# |
ifeq ($(CONFIG_SMP),y) |
DEFS += -DSMP |
DEFS += -DCONFIG_SMP |
endif |
ifeq ($(CONFIG_HT),y) |
DEFS += -DHT |
DEFS += -DCONFIG_HT |
endif |
ifeq ($(CONFIG_FPU_LAZY),y) |
DEFS += -DFPU_LAZY |
DEFS += -DCONFIG_FPU_LAZY |
endif |
ARCH_SOURCES = \ |
/SPARTAN/trunk/arch/ia32/src/ia32.c |
---|
62,10 → 62,10 |
trap_register(VECTOR_SYSCALL, syscall); |
#ifdef __SMP__ |
#ifdef CONFIG_SMP |
trap_register(VECTOR_TLB_SHOOTDOWN_IPI, tlb_shootdown_ipi); |
trap_register(VECTOR_WAKEUP_IPI, wakeup_ipi); |
#endif /* __SMP__ */ |
#endif /* CONFIG_SMP */ |
} |
} |
81,9 → 81,9 |
if (config.cpu_active == 1) { |
memory_print_map(); |
#ifdef __SMP__ |
#ifdef CONFIG_SMP |
acpi_init(); |
#endif /* __SMP__ */ |
#endif /* CONFIG_SMP */ |
} |
} |
/SPARTAN/trunk/arch/ia32/src/smp/mps.c |
---|
26,7 → 26,7 |
* THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
*/ |
#ifdef __SMP__ |
#ifdef CONFIG_SMP |
#include <config.h> |
#include <print.h> |
422,4 → 422,4 |
return -1; |
} |
#endif /* __SMP__ */ |
#endif /* CONFIG_SMP */ |
/SPARTAN/trunk/arch/ia32/src/smp/smp.c |
---|
47,7 → 47,7 |
#include <memstr.h> |
#include <arch/i8259.h> |
#ifdef __SMP__ |
#ifdef CONFIG_SMP |
static struct smp_config_operations *ops = NULL; |
165,4 → 165,4 |
waitq_wakeup(&kmp_completion_wq, WAKEUP_FIRST); |
} |
#endif /* __SMP__ */ |
#endif /* CONFIG_SMP */ |
/SPARTAN/trunk/arch/ia32/src/smp/ap.S |
---|
32,7 → 32,7 |
.section K_TEXT_START_2, "ax" |
#ifdef __SMP__ |
#ifdef CONFIG_SMP |
.global ap_boot |
73,4 → 73,4 |
jmpl $KTEXT, $main_ap |
#endif /* __SMP__ */ |
#endif /* CONFIG_SMP */ |
/SPARTAN/trunk/arch/ia32/src/smp/ipi.c |
---|
26,7 → 26,7 |
* THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
*/ |
#ifdef __SMP__ |
#ifdef CONFIG_SMP |
#include <smp/ipi.h> |
#include <arch/smp/apic.h> |
36,4 → 36,4 |
(void) l_apic_broadcast_custom_ipi((__u8) ipi); |
} |
#endif /* __SMP__ */ |
#endif /* CONFIG_SMP */ |
/SPARTAN/trunk/arch/ia32/src/smp/apic.c |
---|
37,7 → 37,7 |
#include <arch/asm.h> |
#include <arch.h> |
#ifdef __SMP__ |
#ifdef CONFIG_SMP |
/* |
* This is functional, far-from-general-enough interface to the APIC. |
416,4 → 416,4 |
} |
#endif /* __SMP__ */ |
#endif /* CONFIG_SMP */ |
/SPARTAN/trunk/arch/ia32/src/interrupt.c |
---|
110,7 → 110,7 |
void nm_fault(__u8 n, __native stack[]) |
{ |
#ifdef FPU_LAZY |
#ifdef CONFIG_FPU_LAZY |
scheduler_fpu_lazy_request(); |
#else |
panic("fpu fault"); |
/SPARTAN/trunk/arch/ia32/src/atomic.S |
---|
28,7 → 28,7 |
.text |
#ifdef __SMP__ |
#ifdef CONFIG_SMP |
.global spinlock_arch |
42,7 → 42,7 |
movl 12(%esp),%ebx |
0: |
#ifdef __HT__ |
#ifdef CONFIG_HT |
pause # Pentium 4's with HT love this instruction |
#endif |
movl (%ebx),%eax |