diff --git a/sys/vm/vm_glue.c b/sys/vm/vm_glue.c index 8471fd94e480..e82c4f5be413 100644 --- a/sys/vm/vm_glue.c +++ b/sys/vm/vm_glue.c @@ -59,7 +59,7 @@ * any improvements or extensions that they make and grant Carnegie the * rights to redistribute these changes. * - * $Id: vm_glue.c,v 1.27 1995/09/24 19:51:50 dyson Exp $ + * $Id: vm_glue.c,v 1.28 1995/10/16 05:45:49 dyson Exp $ */ #include @@ -213,7 +213,7 @@ vm_fork(p1, p2, isvfork) * non-inheritable */ (void) vm_map_inherit(&p1->p_vmspace->vm_map, - UPT_MIN_ADDRESS - UPAGES * NBPG, VM_MAX_ADDRESS, VM_INHERIT_NONE); + UPT_MIN_ADDRESS - UPAGES * PAGE_SIZE, VM_MAX_ADDRESS, VM_INHERIT_NONE); p2->p_vmspace = vmspace_fork(p1->p_vmspace); #ifdef SYSVSHM @@ -235,25 +235,25 @@ vm_fork(p1, p2, isvfork) /* force in the page table encompassing the UPAGES */ ptaddr = trunc_page((u_int) vtopte(addr)); - error = vm_map_pageable(vp, ptaddr, ptaddr + NBPG, FALSE); + error = vm_map_pageable(vp, ptaddr, ptaddr + PAGE_SIZE, FALSE); if (error) panic("vm_fork: wire of PT failed. error=%d", error); /* and force in (demand-zero) the UPAGES */ - error = vm_map_pageable(vp, addr, addr + UPAGES * NBPG, FALSE); + error = vm_map_pageable(vp, addr, addr + UPAGES * PAGE_SIZE, FALSE); if (error) panic("vm_fork: wire of UPAGES failed. error=%d", error); /* get a kernel virtual address for the UPAGES for this proc */ - up = (struct user *) kmem_alloc_pageable(u_map, UPAGES * NBPG); + up = (struct user *) kmem_alloc_pageable(u_map, UPAGES * PAGE_SIZE); if (up == NULL) panic("vm_fork: u_map allocation failed"); /* and force-map the upages into the kernel pmap */ for (i = 0; i < UPAGES; i++) pmap_enter(vm_map_pmap(u_map), - ((vm_offset_t) up) + NBPG * i, - pmap_extract(vp->pmap, addr + NBPG * i), + ((vm_offset_t) up) + PAGE_SIZE * i, + pmap_extract(vp->pmap, addr + PAGE_SIZE * i), VM_PROT_READ | VM_PROT_WRITE, 1); p2->p_addr = up; @@ -330,19 +330,19 @@ faultin(p) map = &p->p_vmspace->vm_map; /* force the page table encompassing the kernel stack (upages) */ ptaddr = trunc_page((u_int) vtopte(kstack)); - error = vm_map_pageable(map, ptaddr, ptaddr + NBPG, FALSE); + error = vm_map_pageable(map, ptaddr, ptaddr + PAGE_SIZE, FALSE); if (error) panic("faultin: wire of PT failed. error=%d", error); /* wire in the UPAGES */ error = vm_map_pageable(map, (vm_offset_t) kstack, - (vm_offset_t) kstack + UPAGES * NBPG, FALSE); + (vm_offset_t) kstack + UPAGES * PAGE_SIZE, FALSE); if (error) panic("faultin: wire of UPAGES failed. error=%d", error); /* and map them nicely into the kernel pmap */ for (i = 0; i < UPAGES; i++) { - vm_offset_t off = i * NBPG; + vm_offset_t off = i * PAGE_SIZE; vm_offset_t pa = (vm_offset_t) pmap_extract(&p->p_vmspace->vm_pmap, (vm_offset_t) kstack + off); @@ -525,13 +525,13 @@ swapout(p) * let the upages be paged */ pmap_remove(vm_map_pmap(u_map), - (vm_offset_t) p->p_addr, ((vm_offset_t) p->p_addr) + UPAGES * NBPG); + (vm_offset_t) p->p_addr, ((vm_offset_t) p->p_addr) + UPAGES * PAGE_SIZE); vm_map_pageable(map, (vm_offset_t) kstack, - (vm_offset_t) kstack + UPAGES * NBPG, TRUE); + (vm_offset_t) kstack + UPAGES * PAGE_SIZE, TRUE); ptaddr = trunc_page((u_int) vtopte(kstack)); - vm_map_pageable(map, ptaddr, ptaddr + NBPG, TRUE); + vm_map_pageable(map, ptaddr, ptaddr + PAGE_SIZE, TRUE); p->p_flag &= ~P_SWAPPING; p->p_swtime = 0; diff --git a/sys/vm/vm_map.c b/sys/vm/vm_map.c index 119e4ee6aea3..2c04dffdb91a 100644 --- a/sys/vm/vm_map.c +++ b/sys/vm/vm_map.c @@ -61,7 +61,7 @@ * any improvements or extensions that they make and grant Carnegie the * rights to redistribute these changes. * - * $Id: vm_map.c,v 1.23 1995/07/13 08:48:26 davidg Exp $ + * $Id: vm_map.c,v 1.24 1995/08/26 23:18:38 bde Exp $ */ /* @@ -325,7 +325,7 @@ vm_map_entry_create(map) if (m) { int newentries; - newentries = (NBPG / sizeof(struct vm_map_entry)); + newentries = (PAGE_SIZE / sizeof(struct vm_map_entry)); vm_page_wire(m); m->flags &= ~PG_BUSY; m->valid = VM_PAGE_BITS_ALL; @@ -334,7 +334,7 @@ vm_map_entry_create(map) m->flags |= PG_WRITEABLE|PG_MAPPED; entry = (vm_map_entry_t) mapvm; - mapvm += NBPG; + mapvm += PAGE_SIZE; --mapvmpgcnt; for (i = 0; i < newentries; i++) { diff --git a/sys/vm/vm_page.c b/sys/vm/vm_page.c index 4eaf273af4cc..6cfd5166b9bc 100644 --- a/sys/vm/vm_page.c +++ b/sys/vm/vm_page.c @@ -34,7 +34,7 @@ * SUCH DAMAGE. * * from: @(#)vm_page.c 7.4 (Berkeley) 5/7/91 - * $Id: vm_page.c,v 1.35 1995/09/03 19:57:25 dyson Exp $ + * $Id: vm_page.c,v 1.36 1995/09/03 20:40:43 dyson Exp $ */ /* @@ -344,7 +344,7 @@ vm_page_hash(object, offset) vm_object_t object; vm_offset_t offset; { - return ((unsigned) object + offset / NBPG) & vm_page_hash_mask; + return ((unsigned) object + (offset >> PAGE_SHIFT)) & vm_page_hash_mask; } /* diff --git a/sys/vm/vm_pageout.c b/sys/vm/vm_pageout.c index 1199c56f2aa3..38de74396871 100644 --- a/sys/vm/vm_pageout.c +++ b/sys/vm/vm_pageout.c @@ -65,7 +65,7 @@ * any improvements or extensions that they make and grant Carnegie the * rights to redistribute these changes. * - * $Id: vm_pageout.c,v 1.56 1995/10/06 09:42:11 phk Exp $ + * $Id: vm_pageout.c,v 1.57 1995/10/07 19:02:55 davidg Exp $ */ /* @@ -935,7 +935,7 @@ vm_daemon() if ((p->p_flag & P_INMEM) == 0) limit = 0; /* XXX */ - size = p->p_vmspace->vm_pmap.pm_stats.resident_count * NBPG; + size = p->p_vmspace->vm_pmap.pm_stats.resident_count * PAGE_SIZE; if (limit >= 0 && size >= limit) { overage = (size - limit) >> PAGE_SHIFT; vm_pageout_map_deactivate_pages(&p->p_vmspace->vm_map,