Kill some stale leftovers from the earlier attempts at SMP per-cpu pages

This commit is contained in:
peter 1997-06-22 15:47:16 +00:00
parent f4c30626de
commit e0245a10b2
5 changed files with 5 additions and 50 deletions

View File

@ -35,7 +35,7 @@
* SUCH DAMAGE.
*
* from: @(#)machdep.c 7.4 (Berkeley) 6/3/91
* $Id: machdep.c,v 1.248 1997/06/15 02:02:27 wollman Exp $
* $Id: machdep.c,v 1.249 1997/06/15 02:24:06 wollman Exp $
*/
#include "apm.h"
@ -359,23 +359,6 @@ again:
u_map = kmem_suballoc(kernel_map, &minaddr, &maxaddr,
(maxproc*UPAGES*PAGE_SIZE), FALSE);
#if defined(SMP) && defined(SMP_PRIVPAGES)
/* Per-cpu pages.. (the story so far is... subject to change)
* ========= For the per-cpu data page ========
* 1 private data page
* 1 PDE (per-cpu PTD entry page)
* 1 PT (per-cpu page table page)
* ============ For the idle loop =============
* 2 UPAGEs (per-cpu idle procs)
* 1 PTD (for per-cpu equiv of IdlePTD)
* ============================================
* = total of 6 pages per cpu. The BSP reuses the ones allocated
* by locore.s during boot to remove special cases at runtime.
*/
ppage_map = kmem_suballoc(kernel_map, &minaddr, &maxaddr,
(NCPU*6*PAGE_SIZE), FALSE);
#endif
/*
* Finally, allocate mbuf pool. Since mclrefcnt is an off-size
* we use the more space efficient malloc in place of kmem_alloc.

View File

@ -35,7 +35,7 @@
* SUCH DAMAGE.
*
* from: @(#)machdep.c 7.4 (Berkeley) 6/3/91
* $Id: machdep.c,v 1.248 1997/06/15 02:02:27 wollman Exp $
* $Id: machdep.c,v 1.249 1997/06/15 02:24:06 wollman Exp $
*/
#include "apm.h"
@ -359,23 +359,6 @@ again:
u_map = kmem_suballoc(kernel_map, &minaddr, &maxaddr,
(maxproc*UPAGES*PAGE_SIZE), FALSE);
#if defined(SMP) && defined(SMP_PRIVPAGES)
/* Per-cpu pages.. (the story so far is... subject to change)
* ========= For the per-cpu data page ========
* 1 private data page
* 1 PDE (per-cpu PTD entry page)
* 1 PT (per-cpu page table page)
* ============ For the idle loop =============
* 2 UPAGEs (per-cpu idle procs)
* 1 PTD (for per-cpu equiv of IdlePTD)
* ============================================
* = total of 6 pages per cpu. The BSP reuses the ones allocated
* by locore.s during boot to remove special cases at runtime.
*/
ppage_map = kmem_suballoc(kernel_map, &minaddr, &maxaddr,
(NCPU*6*PAGE_SIZE), FALSE);
#endif
/*
* Finally, allocate mbuf pool. Since mclrefcnt is an off-size
* we use the more space efficient malloc in place of kmem_alloc.

View File

@ -61,13 +61,12 @@
* any improvements or extensions that they make and grant Carnegie the
* rights to redistribute these changes.
*
* $Id: vm_kern.c,v 1.35 1997/04/26 11:46:23 peter Exp $
* $Id: vm_kern.c,v 1.36 1997/05/29 02:57:22 peter Exp $
*/
/*
* Kernel memory management.
*/
#include "opt_smp_privpages.h"
#include <sys/param.h>
#include <sys/systm.h>
@ -100,10 +99,6 @@ vm_map_t mb_map=0;
int mb_map_full=0;
vm_map_t io_map=0;
vm_map_t phys_map=0;
#if defined(SMP) && defined(SMP_PRIVPAGES)
vm_map_t ppage_map=0;
#endif
/*
* kmem_alloc_pageable:

View File

@ -61,7 +61,7 @@
* any improvements or extensions that they make and grant Carnegie the
* rights to redistribute these changes.
*
* $Id: vm_kern.h,v 1.13 1997/03/31 11:11:26 davidg Exp $
* $Id: vm_kern.h,v 1.14 1997/04/26 11:46:23 peter Exp $
*/
#ifndef _VM_VM_KERN_H_
@ -78,7 +78,6 @@ extern vm_map_t clean_map;
extern vm_map_t phys_map;
extern vm_map_t exec_map;
extern vm_map_t u_map;
extern vm_map_t ppage_map;
extern vm_offset_t kernel_vm_end;
/* XXX - elsewhere? */

View File

@ -61,13 +61,12 @@
* any improvements or extensions that they make and grant Carnegie the
* rights to redistribute these changes.
*
* $Id: vm_object.c,v 1.92 1997/05/29 02:57:22 peter Exp $
* $Id: vm_object.c,v 1.93 1997/06/22 03:00:24 dyson Exp $
*/
/*
* Virtual memory object module.
*/
#include "opt_smp_privpages.h"
#include <sys/param.h>
#include <sys/systm.h>
@ -1455,10 +1454,6 @@ vm_object_in_map( object)
return 1;
if( _vm_object_in_map( u_map, object, 0))
return 1;
#if defined(SMP) && defined(SMP_PRIVPAGES)
if( _vm_object_in_map( ppage_map, object, 0))
return 1;
#endif
return 0;
}