bcc23094ea
set improves performance and fixes the following problems (description from John Dyson): 1. Growing swap space problem in both static usage and in situations with lots of fork/execs in heavy paging situations. 2. Sparse swap space allocation (internal fragmentation.) 3. General swap_pager slowness. Additionally, the new swap_pager also provides hooks for multi-page I/O that is currently being developed (in early testing phases.) Problem #1 is a result of a problem where objects cannot be collapsed once a pager has been allocated for them. This problem has been solved by relaxing the restriction by allowing the pages contained in a shadow object's pager be copied to the parent object's pager. The copy is afforded by manipulating pointers to the disk blocks on the swap space. Since an improved swap_pager has already been developed with the data structures to support the copy operation, this new swap_pager has been introduced. Also, shadow object bypass in the collapse code has been enhanced to support checking for pages on disk. The vm_pageout daemon has also been modified to defer creation of an object's pager when the object's shadow is paging. This allows efficient immediate collapsing of a shadow into a parent object under many circumstances without the creation of an intermediate pager. Problem #2 is solved by the allocation data structures and algorithms in the new swap_pager. Additionally, a newer version of this new swap_pager is being tested that permits multiple page I/O and mitigation of the fragmentation problems associated with allocation of large contiguous blocks of swap space. Problem #3 is addressed by better algorithms and a fix of a couple of bugs in the swap_pager. Additionally, this new pager has a growth path allowing multi-page inputs from disk. Approximately 50% performance improvement can be expected under certain circumstances when using this pager in the standard single page mode. (Actually, I've seen more like twice the speed in my tests. -DLG)
291 lines
7.2 KiB
C
291 lines
7.2 KiB
C
/*
|
|
* Copyright (c) 1992 William F. Jolitz, TeleMuse
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
* 3. All advertising materials mentioning features or use of this software
|
|
* must display the following acknowledgement:
|
|
* This software is a component of "386BSD" developed by
|
|
William F. Jolitz, TeleMuse.
|
|
* 4. Neither the name of the developer nor the name "386BSD"
|
|
* may be used to endorse or promote products derived from this software
|
|
* without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS A COMPONENT OF 386BSD DEVELOPED BY WILLIAM F. JOLITZ
|
|
* AND IS INTENDED FOR RESEARCH AND EDUCATIONAL PURPOSES ONLY. THIS
|
|
* SOFTWARE SHOULD NOT BE CONSIDERED TO BE A COMMERCIAL PRODUCT.
|
|
* THE DEVELOPER URGES THAT USERS WHO REQUIRE A COMMERCIAL PRODUCT
|
|
* NOT MAKE USE THIS WORK.
|
|
*
|
|
* FOR USERS WHO WISH TO UNDERSTAND THE 386BSD SYSTEM DEVELOPED
|
|
* BY WILLIAM F. JOLITZ, WE RECOMMEND THE USER STUDY WRITTEN
|
|
* REFERENCES SUCH AS THE "PORTING UNIX TO THE 386" SERIES
|
|
* (BEGINNING JANUARY 1991 "DR. DOBBS JOURNAL", USA AND BEGINNING
|
|
* JUNE 1991 "UNIX MAGAZIN", GERMANY) BY WILLIAM F. JOLITZ AND
|
|
* LYNNE GREER JOLITZ, AS WELL AS OTHER BOOKS ON UNIX AND THE
|
|
* ON-LINE 386BSD USER MANUAL BEFORE USE. A BOOK DISCUSSING THE INTERNALS
|
|
* OF 386BSD ENTITLED "386BSD FROM THE INSIDE OUT" WILL BE AVAILABLE LATE 1992.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE DEVELOPER ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE DEVELOPER BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*
|
|
* $Id$
|
|
*/
|
|
|
|
#include "param.h"
|
|
#include "systm.h"
|
|
#include "cdefs.h"
|
|
#include "malloc.h"
|
|
#include "rlist.h"
|
|
#include "vm/vm.h"
|
|
#include "vm/vm_map.h"
|
|
|
|
extern vm_map_t kernel_map;
|
|
|
|
/*
|
|
* Resource lists.
|
|
*/
|
|
|
|
#define RLIST_MIN 128
|
|
static int rlist_count=0;
|
|
static struct rlist *rlfree;
|
|
int rlist_active;
|
|
|
|
static struct rlist *
|
|
rlist_malloc()
|
|
{
|
|
struct rlist *rl;
|
|
int i;
|
|
while( rlist_count < RLIST_MIN) {
|
|
extern vm_map_t kmem_map;
|
|
int s = splhigh();
|
|
rl = (struct rlist *)kmem_malloc(kmem_map, NBPG, 0);
|
|
splx(s);
|
|
if( !rl)
|
|
break;
|
|
|
|
for(i=0;i<(NBPG/(sizeof *rl));i++) {
|
|
rl->rl_next = rlfree;
|
|
rlfree = rl;
|
|
rlist_count++;
|
|
rl++;
|
|
}
|
|
}
|
|
|
|
if( (rl = rlfree) == 0 )
|
|
panic("Cannot get an rlist entry");
|
|
|
|
--rlist_count;
|
|
rlfree = rl->rl_next;
|
|
return rl;
|
|
}
|
|
|
|
inline static void
|
|
rlist_mfree( struct rlist *rl)
|
|
{
|
|
rl->rl_next = rlfree;
|
|
rlfree = rl;
|
|
++rlist_count;
|
|
}
|
|
|
|
|
|
/*
|
|
* Add space to a resource list. Used to either
|
|
* initialize a list or return free space to it.
|
|
*/
|
|
void
|
|
rlist_free (rlp, start, end)
|
|
register struct rlist **rlp;
|
|
unsigned start, end;
|
|
{
|
|
struct rlist *head;
|
|
register struct rlist *olp = 0;
|
|
int s;
|
|
|
|
s = splhigh();
|
|
while( rlist_active)
|
|
tsleep((caddr_t)&rlist_active, PSWP, "rlistf", 0);
|
|
rlist_active = 1;
|
|
splx(s);
|
|
|
|
head = *rlp;
|
|
|
|
loop:
|
|
/* if nothing here, insert (tail of list) */
|
|
if (*rlp == 0) {
|
|
*rlp = rlist_malloc();
|
|
(*rlp)->rl_start = start;
|
|
(*rlp)->rl_end = end;
|
|
(*rlp)->rl_next = 0;
|
|
rlist_active = 0;
|
|
wakeup((caddr_t)&rlist_active);
|
|
return;
|
|
}
|
|
|
|
/* if new region overlaps something currently present, panic */
|
|
if (start >= (*rlp)->rl_start && start <= (*rlp)->rl_end) {
|
|
printf("Frag %d:%d, ent %d:%d ", start, end,
|
|
(*rlp)->rl_start, (*rlp)->rl_end);
|
|
panic("overlapping front rlist_free: freed twice?");
|
|
}
|
|
if (end >= (*rlp)->rl_start && end <= (*rlp)->rl_end) {
|
|
printf("Frag %d:%d, ent %d:%d ", start, end,
|
|
(*rlp)->rl_start, (*rlp)->rl_end);
|
|
panic("overlapping tail rlist_free: freed twice?");
|
|
}
|
|
|
|
/* are we adjacent to this element? (in front) */
|
|
if (end+1 == (*rlp)->rl_start) {
|
|
/* coalesce */
|
|
(*rlp)->rl_start = start;
|
|
goto scan;
|
|
}
|
|
|
|
/* are we before this element? */
|
|
if (end < (*rlp)->rl_start) {
|
|
register struct rlist *nlp;
|
|
|
|
nlp = rlist_malloc();
|
|
nlp->rl_start = start;
|
|
nlp->rl_end = end;
|
|
nlp->rl_next = *rlp;
|
|
/*
|
|
* If the new element is in front of the list,
|
|
* adjust *rlp, else don't.
|
|
*/
|
|
if( olp) {
|
|
olp->rl_next = nlp;
|
|
} else {
|
|
*rlp = nlp;
|
|
}
|
|
rlist_active = 0;
|
|
wakeup((caddr_t)&rlist_active);
|
|
return;
|
|
}
|
|
|
|
/* are we adjacent to this element? (at tail) */
|
|
if ((*rlp)->rl_end + 1 == start) {
|
|
/* coalesce */
|
|
(*rlp)->rl_end = end;
|
|
goto scan;
|
|
}
|
|
|
|
/* are we after this element */
|
|
if (start > (*rlp)->rl_end) {
|
|
olp = *rlp;
|
|
rlp = &((*rlp)->rl_next);
|
|
goto loop;
|
|
} else
|
|
panic("rlist_free: can't happen");
|
|
|
|
scan:
|
|
/* can we coalesce list now that we've filled a void? */
|
|
{
|
|
register struct rlist *lp, *lpn;
|
|
|
|
for (lp = head; lp->rl_next ;) {
|
|
lpn = lp->rl_next;
|
|
|
|
/* coalesce ? */
|
|
if (lp->rl_end + 1 == lpn->rl_start) {
|
|
lp->rl_end = lpn->rl_end;
|
|
lp->rl_next = lpn->rl_next;
|
|
rlist_mfree(lpn);
|
|
} else
|
|
lp = lp->rl_next;
|
|
}
|
|
}
|
|
rlist_active = 0;
|
|
wakeup((caddr_t)&rlist_active);
|
|
}
|
|
|
|
/*
|
|
* Obtain a region of desired size from a resource list.
|
|
* If nothing available of that size, return 0. Otherwise,
|
|
* return a value of 1 and set resource start location with
|
|
* "*loc". (Note: loc can be zero if we don't wish the value)
|
|
*/
|
|
int rlist_alloc (rlp, size, loc)
|
|
struct rlist **rlp;
|
|
unsigned size, *loc;
|
|
{
|
|
register struct rlist *lp;
|
|
int s;
|
|
register struct rlist *olp = 0;
|
|
|
|
s = splhigh();
|
|
while( rlist_active)
|
|
tsleep((caddr_t)&rlist_active, PSWP, "rlista", 0);
|
|
rlist_active = 1;
|
|
splx(s);
|
|
|
|
/* walk list, allocating first thing that's big enough (first fit) */
|
|
for (; *rlp; rlp = &((*rlp)->rl_next))
|
|
if(size <= (*rlp)->rl_end - (*rlp)->rl_start + 1) {
|
|
|
|
/* hand it to the caller */
|
|
if (loc) *loc = (*rlp)->rl_start;
|
|
(*rlp)->rl_start += size;
|
|
|
|
/* did we eat this element entirely? */
|
|
if ((*rlp)->rl_start > (*rlp)->rl_end) {
|
|
lp = (*rlp)->rl_next;
|
|
rlist_mfree(*rlp);
|
|
/*
|
|
* if the deleted element was in fromt
|
|
* of the list, adjust *rlp, else don't.
|
|
*/
|
|
if (olp) {
|
|
olp->rl_next = lp;
|
|
} else {
|
|
*rlp = lp;
|
|
}
|
|
}
|
|
|
|
rlist_active = 0;
|
|
wakeup((caddr_t)&rlist_active);
|
|
return (1);
|
|
} else {
|
|
olp = *rlp;
|
|
}
|
|
|
|
rlist_active = 0;
|
|
wakeup((caddr_t)&rlist_active);
|
|
/* nothing in list that's big enough */
|
|
return (0);
|
|
}
|
|
|
|
/*
|
|
* Finished with this resource list, reclaim all space and
|
|
* mark it as being empty.
|
|
*/
|
|
void
|
|
rlist_destroy (rlp)
|
|
struct rlist **rlp;
|
|
{
|
|
struct rlist *lp, *nlp;
|
|
|
|
lp = *rlp;
|
|
*rlp = 0;
|
|
for (; lp; lp = nlp) {
|
|
nlp = lp->rl_next;
|
|
rlist_mfree(lp);
|
|
}
|
|
}
|