769e0f974d
include: * Mutual exclusion is used instead of spl*(). See mutex(9). (Note: The alpha port is still in transition and currently uses both.) * Per-CPU idle processes. * Interrupts are run in their own separate kernel threads and can be preempted (i386 only). Partially contributed by: BSDi (BSD/OS) Submissions by (at least): cp, dfr, dillon, grog, jake, jhb, sheldonh
341 lines
8.6 KiB
C
341 lines
8.6 KiB
C
/*
|
|
* Copyright (c) 1996, by Steve Passe
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. The name of the developer may NOT be used to endorse or promote products
|
|
* derived from this software without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*
|
|
* $FreeBSD$
|
|
*/
|
|
|
|
#ifndef _MACHINE_SMPTESTS_H_
|
|
#define _MACHINE_SMPTESTS_H_
|
|
|
|
|
|
/*
|
|
* Various 'tests in progress' and configuration parameters.
|
|
*/
|
|
|
|
|
|
/*
|
|
* Tor's clock improvements.
|
|
*
|
|
* When the giant kernel lock disappears, a different strategy should
|
|
* probably be used, thus this patch can only be considered a temporary
|
|
* measure.
|
|
*
|
|
* This patch causes (NCPU-1)*(128+100) extra IPIs per second.
|
|
* During profiling, the number is (NCPU-1)*(1024+100) extra IPIs/s
|
|
* in addition to extra IPIs due to forwarding ASTs to other CPUs.
|
|
*
|
|
* Having a shared AST flag in an SMP configuration is wrong, and I've
|
|
* just kludged around it, based upon the kernel lock blocking other
|
|
* processors from entering the kernel while handling an AST for one
|
|
* processor. When the giant kernel lock disappers, this kludge breaks.
|
|
*
|
|
* -- Tor
|
|
*/
|
|
#define BETTER_CLOCK
|
|
|
|
|
|
/*
|
|
* Control the "giant lock" pushdown by logical steps.
|
|
*/
|
|
#define PUSHDOWN_LEVEL_1
|
|
#define PUSHDOWN_LEVEL_2
|
|
#define PUSHDOWN_LEVEL_3_NOT
|
|
#define PUSHDOWN_LEVEL_4_NOT
|
|
|
|
/*
|
|
* Debug version of simple_lock. This will store the CPU id of the
|
|
* holding CPU along with the lock. When a CPU fails to get the lock
|
|
* it compares its own id to the holder id. If they are the same it
|
|
* panic()s, as simple locks are binary, and this would cause a deadlock.
|
|
*
|
|
*/
|
|
#define SL_DEBUG
|
|
|
|
|
|
/*
|
|
* Put FAST_INTR() ISRs at an APIC priority above the regular INTs.
|
|
* Allow the mp_lock() routines to handle FAST interrupts while spinning.
|
|
*/
|
|
#ifdef PUSHDOWN_LEVEL_1
|
|
#define FAST_HI
|
|
#endif
|
|
|
|
|
|
/*
|
|
* These defines enable critical region locking of areas that were
|
|
* protected via cli/sti in the UP kernel.
|
|
*
|
|
* COMLOCK protects the sio/cy drivers.
|
|
* CLOCKLOCK protects clock hardware and data
|
|
* known to be incomplete:
|
|
* joystick lkm
|
|
* ?
|
|
*/
|
|
#ifdef PUSHDOWN_LEVEL_1
|
|
#define USE_COMLOCK
|
|
#define USE_CLOCKLOCK
|
|
#endif
|
|
|
|
|
|
/*
|
|
* INTR_SIMPLELOCK has been removed, as the interrupt mechanism will likely
|
|
* not use this sort of optimization if we move to interrupt threads.
|
|
*/
|
|
#ifdef PUSHDOWN_LEVEL_4
|
|
#endif
|
|
|
|
|
|
/*
|
|
* CPL_AND_CML has been removed. Interrupt threads will eventually not
|
|
* use either mechanism so there is no point trying to optimize it.
|
|
*/
|
|
#ifdef PUSHDOWN_LEVEL_3
|
|
#endif
|
|
|
|
|
|
/*
|
|
* SPL_DEBUG_POSTCODE/INTR_SPL/SPL_DEBUG - removed
|
|
*
|
|
* These functions were too expensive for the standard case but, more
|
|
* importantly, we should be able to come up with a much cleaner way
|
|
* to handle the cpl. Having to do any locking at all is a mistake
|
|
* for something that is modified as often as cpl is.
|
|
*/
|
|
|
|
/*
|
|
* FAST_WITHOUTCPL - now made the default (define removed). Text below
|
|
* contains the current discussion. I am confident we can find a solution
|
|
* that does not require us to process softints from a hard int, which can
|
|
* kill serial performance due to the lack of true hardware ipl's.
|
|
*
|
|
****
|
|
*
|
|
* Ignore the ipending bits when exiting FAST_INTR() routines.
|
|
*
|
|
* according to Bruce:
|
|
*
|
|
* setsoft*() may set ipending. setsofttty() is actually used in the
|
|
* FAST_INTR handler in some serial drivers. This is necessary to get
|
|
* output completions and other urgent events handled as soon as possible.
|
|
* The flag(s) could be set in a variable other than ipending, but they
|
|
* needs to be checked against cpl to decide whether the software interrupt
|
|
* handler can/should run.
|
|
*
|
|
* (FAST_INTR used to just return
|
|
* in all cases until rev.1.7 of vector.s. This worked OK provided there
|
|
* were no user-mode CPU hogs. CPU hogs caused an average latency of 1/2
|
|
* clock tick for output completions...)
|
|
***
|
|
*
|
|
* So I need to restore cpl handling someday, but AFTER
|
|
* I finish making spl/cpl MP-safe.
|
|
*/
|
|
#ifdef PUSHDOWN_LEVEL_1
|
|
#endif
|
|
|
|
|
|
/*
|
|
* FAST_SIMPLELOCK no longer exists, because it doesn't help us. The cpu
|
|
* is likely to already hold the MP lock and recursive MP locks are now
|
|
* very cheap, so we do not need this optimization. Eventually *ALL*
|
|
* interrupts will run in their own thread, so there is no sense complicating
|
|
* matters now.
|
|
*/
|
|
#ifdef PUSHDOWN_LEVEL_1
|
|
#endif
|
|
|
|
|
|
/*
|
|
* Portions of the old TEST_LOPRIO code, back from the grave!
|
|
*/
|
|
#define GRAB_LOPRIO
|
|
|
|
|
|
/*
|
|
* Send CPUSTOP IPI for stop/restart of other CPUs on DDB break.
|
|
*/
|
|
#define VERBOSE_CPUSTOP_ON_DDBBREAK
|
|
#define CPUSTOP_ON_DDBBREAK
|
|
|
|
|
|
/*
|
|
* Bracket code/comments relevant to the current 'giant lock' model.
|
|
* Everything is now the 'giant lock' model, but we will use this as
|
|
* we start to "push down" the lock.
|
|
*/
|
|
#define GIANT_LOCK
|
|
|
|
#ifdef APIC_IO
|
|
/*
|
|
* Enable extra counters for some selected locations in the interrupt handlers.
|
|
* Look in apic_vector.s, apic_ipl.s and ipl.s for APIC_ITRACE or
|
|
* APIC_INTR_DIAGNOSTIC.
|
|
*/
|
|
#undef APIC_INTR_DIAGNOSTIC
|
|
|
|
/*
|
|
* Add extra tracking of a specific interrupt. Look in apic_vector.s,
|
|
* apic_ipl.s and ipl.s for APIC_ITRACE and log_intr_event.
|
|
* APIC_INTR_DIAGNOSTIC must be defined for this to work.
|
|
*/
|
|
#ifdef APIC_INTR_DIAGNOSTIC
|
|
#define APIC_INTR_DIAGNOSTIC_IRQ 17
|
|
#endif
|
|
|
|
/*
|
|
* Don't assume that slow interrupt handler X is called from vector
|
|
* X + ICU_OFFSET.
|
|
*/
|
|
#define APIC_INTR_REORDER
|
|
|
|
/*
|
|
* Redirect clock interrupts to a higher priority (fast intr) vector,
|
|
* while still using the slow interrupt handler. Only effective when
|
|
* APIC_INTR_REORDER is defined.
|
|
*/
|
|
#define APIC_INTR_HIGHPRI_CLOCK
|
|
|
|
#endif /* APIC_IO */
|
|
|
|
/*
|
|
* Misc. counters.
|
|
*
|
|
#define COUNT_XINVLTLB_HITS
|
|
*/
|
|
|
|
|
|
/**
|
|
* Hack to "fake-out" kernel into thinking it is running on a 'default config'.
|
|
*
|
|
* value == default type
|
|
#define TEST_DEFAULT_CONFIG 6
|
|
*/
|
|
|
|
|
|
/*
|
|
* Simple test code for IPI interaction, save for future...
|
|
*
|
|
#define TEST_TEST1
|
|
#define IPI_TARGET_TEST1 1
|
|
*/
|
|
|
|
|
|
/*
|
|
* Address of POST hardware port.
|
|
* Defining this enables POSTCODE macros.
|
|
*
|
|
#define POST_ADDR 0x80
|
|
*/
|
|
|
|
|
|
/*
|
|
* POST hardware macros.
|
|
*/
|
|
#ifdef POST_ADDR
|
|
#define ASMPOSTCODE_INC \
|
|
pushl %eax ; \
|
|
movl _current_postcode, %eax ; \
|
|
incl %eax ; \
|
|
andl $0xff, %eax ; \
|
|
movl %eax, _current_postcode ; \
|
|
outb %al, $POST_ADDR ; \
|
|
popl %eax
|
|
|
|
/*
|
|
* Overwrite the current_postcode value.
|
|
*/
|
|
#define ASMPOSTCODE(X) \
|
|
pushl %eax ; \
|
|
movl $X, %eax ; \
|
|
movl %eax, _current_postcode ; \
|
|
outb %al, $POST_ADDR ; \
|
|
popl %eax
|
|
|
|
/*
|
|
* Overwrite the current_postcode low nibble.
|
|
*/
|
|
#define ASMPOSTCODE_LO(X) \
|
|
pushl %eax ; \
|
|
movl _current_postcode, %eax ; \
|
|
andl $0xf0, %eax ; \
|
|
orl $X, %eax ; \
|
|
movl %eax, _current_postcode ; \
|
|
outb %al, $POST_ADDR ; \
|
|
popl %eax
|
|
|
|
/*
|
|
* Overwrite the current_postcode high nibble.
|
|
*/
|
|
#define ASMPOSTCODE_HI(X) \
|
|
pushl %eax ; \
|
|
movl _current_postcode, %eax ; \
|
|
andl $0x0f, %eax ; \
|
|
orl $(X<<4), %eax ; \
|
|
movl %eax, _current_postcode ; \
|
|
outb %al, $POST_ADDR ; \
|
|
popl %eax
|
|
#else
|
|
#define ASMPOSTCODE_INC
|
|
#define ASMPOSTCODE(X)
|
|
#define ASMPOSTCODE_LO(X)
|
|
#define ASMPOSTCODE_HI(X)
|
|
#endif /* POST_ADDR */
|
|
|
|
|
|
/*
|
|
* These are all temps for debugging...
|
|
*
|
|
#define GUARD_INTS
|
|
*/
|
|
|
|
/*
|
|
* This macro traps unexpected INTs to a specific CPU, eg. GUARD_CPU.
|
|
*/
|
|
#ifdef GUARD_INTS
|
|
#define GUARD_CPU 1
|
|
#define MAYBE_PANIC(irq_num) \
|
|
cmpl $GUARD_CPU, _cpuid ; \
|
|
jne 9f ; \
|
|
cmpl $1, _ok_test1 ; \
|
|
jne 9f ; \
|
|
pushl lapic_isr3 ; \
|
|
pushl lapic_isr2 ; \
|
|
pushl lapic_isr1 ; \
|
|
pushl lapic_isr0 ; \
|
|
pushl lapic_irr3 ; \
|
|
pushl lapic_irr2 ; \
|
|
pushl lapic_irr1 ; \
|
|
pushl lapic_irr0 ; \
|
|
pushl $irq_num ; \
|
|
pushl _cpuid ; \
|
|
pushl $panic_msg ; \
|
|
call _printf ; \
|
|
addl $44, %esp ; \
|
|
9:
|
|
#else
|
|
#define MAYBE_PANIC(irq_num)
|
|
#endif /* GUARD_INTS */
|
|
|
|
#endif /* _MACHINE_SMPTESTS_H_ */
|