Make the spl oriented inline functions less likely to allow
potentially volatile memory to be kept in registers during the "call" (inline expansion.) Do the same for pmap_update.
This commit is contained in:
parent
3bef267d25
commit
8966b85c8f
Notes:
svn2git
2020-12-20 02:59:44 +00:00
svn path=/head/; revision=10004
@ -30,7 +30,7 @@
|
|||||||
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
||||||
* SUCH DAMAGE.
|
* SUCH DAMAGE.
|
||||||
*
|
*
|
||||||
* $Id: cpufunc.h,v 1.37 1995/05/30 08:00:30 rgrimes Exp $
|
* $Id: cpufunc.h,v 1.38 1995/07/25 21:28:47 bde Exp $
|
||||||
*/
|
*/
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@ -63,13 +63,13 @@ bdb(void)
|
|||||||
static __inline void
|
static __inline void
|
||||||
disable_intr(void)
|
disable_intr(void)
|
||||||
{
|
{
|
||||||
__asm __volatile("cli");
|
__asm __volatile("cli" : : : "memory");
|
||||||
}
|
}
|
||||||
|
|
||||||
static __inline void
|
static __inline void
|
||||||
enable_intr(void)
|
enable_intr(void)
|
||||||
{
|
{
|
||||||
__asm __volatile("sti");
|
__asm __volatile("sti" : : : "memory");
|
||||||
}
|
}
|
||||||
|
|
||||||
#define HAVE_INLINE_FFS
|
#define HAVE_INLINE_FFS
|
||||||
@ -268,7 +268,8 @@ pmap_update(void)
|
|||||||
* This should be implemented as load_cr3(rcr3()) when load_cr3()
|
* This should be implemented as load_cr3(rcr3()) when load_cr3()
|
||||||
* is inlined.
|
* is inlined.
|
||||||
*/
|
*/
|
||||||
__asm __volatile("movl %%cr3, %0; movl %0, %%cr3" : "=r" (temp));
|
__asm __volatile("movl %%cr3, %0; movl %0, %%cr3" : "=r" (temp) :
|
||||||
|
: "memory");
|
||||||
}
|
}
|
||||||
|
|
||||||
static __inline u_long
|
static __inline u_long
|
||||||
|
@ -30,7 +30,7 @@
|
|||||||
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
||||||
* SUCH DAMAGE.
|
* SUCH DAMAGE.
|
||||||
*
|
*
|
||||||
* $Id: cpufunc.h,v 1.37 1995/05/30 08:00:30 rgrimes Exp $
|
* $Id: cpufunc.h,v 1.38 1995/07/25 21:28:47 bde Exp $
|
||||||
*/
|
*/
|
||||||
|
|
||||||
/*
|
/*
|
||||||
@ -63,13 +63,13 @@ bdb(void)
|
|||||||
static __inline void
|
static __inline void
|
||||||
disable_intr(void)
|
disable_intr(void)
|
||||||
{
|
{
|
||||||
__asm __volatile("cli");
|
__asm __volatile("cli" : : : "memory");
|
||||||
}
|
}
|
||||||
|
|
||||||
static __inline void
|
static __inline void
|
||||||
enable_intr(void)
|
enable_intr(void)
|
||||||
{
|
{
|
||||||
__asm __volatile("sti");
|
__asm __volatile("sti" : : : "memory");
|
||||||
}
|
}
|
||||||
|
|
||||||
#define HAVE_INLINE_FFS
|
#define HAVE_INLINE_FFS
|
||||||
@ -268,7 +268,8 @@ pmap_update(void)
|
|||||||
* This should be implemented as load_cr3(rcr3()) when load_cr3()
|
* This should be implemented as load_cr3(rcr3()) when load_cr3()
|
||||||
* is inlined.
|
* is inlined.
|
||||||
*/
|
*/
|
||||||
__asm __volatile("movl %%cr3, %0; movl %0, %%cr3" : "=r" (temp));
|
__asm __volatile("movl %%cr3, %0; movl %0, %%cr3" : "=r" (temp) :
|
||||||
|
: "memory");
|
||||||
}
|
}
|
||||||
|
|
||||||
static __inline u_long
|
static __inline u_long
|
||||||
|
@ -30,7 +30,7 @@
|
|||||||
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
||||||
* SUCH DAMAGE.
|
* SUCH DAMAGE.
|
||||||
*
|
*
|
||||||
* $Id: spl.h,v 1.7 1995/05/11 00:13:01 wollman Exp $
|
* $Id: spl.h,v 1.8 1995/05/11 07:44:16 bde Exp $
|
||||||
*/
|
*/
|
||||||
|
|
||||||
#ifndef _MACHINE_IPL_H_
|
#ifndef _MACHINE_IPL_H_
|
||||||
@ -105,6 +105,7 @@ static __inline int name(void) \
|
|||||||
{ \
|
{ \
|
||||||
unsigned x; \
|
unsigned x; \
|
||||||
\
|
\
|
||||||
|
asm volatile("":::"memory"); \
|
||||||
x = cpl; \
|
x = cpl; \
|
||||||
set_cpl; \
|
set_cpl; \
|
||||||
return (x); \
|
return (x); \
|
||||||
|
Loading…
Reference in New Issue
Block a user