Remove L1 cache coloring optimization ( leave L2 cache coloring opt ).

Rewrite vm_page_list_find() and vm_page_select_free() - make inline out
    of nominal case.
This commit is contained in:
dillon 1999-02-07 20:45:15 +00:00
parent fad80fe485
commit b605e02c16
2 changed files with 92 additions and 197 deletions

View File

@ -34,7 +34,7 @@
* SUCH DAMAGE. * SUCH DAMAGE.
* *
* from: @(#)vm_page.c 7.4 (Berkeley) 5/7/91 * from: @(#)vm_page.c 7.4 (Berkeley) 5/7/91
* $Id: vm_page.c,v 1.122 1999/01/24 07:06:52 dillon Exp $ * $Id: vm_page.c,v 1.123 1999/01/28 00:57:57 dillon Exp $
*/ */
/* /*
@ -87,7 +87,7 @@
#include <vm/vm_extern.h> #include <vm/vm_extern.h>
static void vm_page_queue_init __P((void)); static void vm_page_queue_init __P((void));
static vm_page_t vm_page_select_free __P((vm_object_t object, static vm_page_t _vm_page_select_free __P((vm_object_t object,
vm_pindex_t pindex, int prefqueue)); vm_pindex_t pindex, int prefqueue));
static vm_page_t vm_page_select_cache __P((vm_object_t, vm_pindex_t)); static vm_page_t vm_page_select_cache __P((vm_object_t, vm_pindex_t));
@ -419,9 +419,6 @@ vm_page_insert(m, object, pindex)
*/ */
TAILQ_INSERT_TAIL(&object->memq, m, listq); TAILQ_INSERT_TAIL(&object->memq, m, listq);
#if 0
m->object->page_hint = m;
#endif
m->object->generation++; m->object->generation++;
if (m->wire_count) if (m->wire_count)
@ -547,12 +544,6 @@ vm_page_lookup(object, pindex)
* Search the hash table for this object/offset pair * Search the hash table for this object/offset pair
*/ */
#if 0
if (object->page_hint && (object->page_hint->pindex == pindex) &&
(object->page_hint->object == object))
return object->page_hint;
#endif
retry: retry:
generation = vm_page_bucket_generation; generation = vm_page_bucket_generation;
bucket = &vm_page_buckets[vm_page_hash(object, pindex)]; bucket = &vm_page_buckets[vm_page_hash(object, pindex)];
@ -560,9 +551,6 @@ vm_page_lookup(object, pindex)
if ((m->object == object) && (m->pindex == pindex)) { if ((m->object == object) && (m->pindex == pindex)) {
if (vm_page_bucket_generation != generation) if (vm_page_bucket_generation != generation)
goto retry; goto retry;
#if 0
m->object->page_hint = m;
#endif
return (m); return (m);
} }
} }
@ -671,6 +659,8 @@ vm_page_unqueue(m)
} }
} }
#if PQ_L2_SIZE > 1
/* /*
* vm_page_list_find: * vm_page_list_find:
* *
@ -684,84 +674,38 @@ vm_page_unqueue(m)
* *
* This routine must be called at splvm(). * This routine must be called at splvm().
* This routine may not block. * This routine may not block.
*
* This routine may only be called from the vm_page_list_find() macro
* in vm_page.h
*/ */
vm_page_t vm_page_t
vm_page_list_find(basequeue, index) _vm_page_list_find(basequeue, index)
int basequeue, index; int basequeue, index;
{ {
#if PQ_L2_SIZE > 1 int i;
vm_page_t m = NULL;
int i,j;
vm_page_t m;
int hindex;
struct vpgqueues *pq; struct vpgqueues *pq;
pq = &vm_page_queues[basequeue]; pq = &vm_page_queues[basequeue];
m = TAILQ_FIRST(pq[index].pl); /*
if (m) * Note that for the first loop, index+i and index-i wind up at the
return m; * same place. Even though this is not totally optimal, we've already
* blown it by missing the cache case so we do not care.
*/
for(j = 0; j < PQ_L1_SIZE; j++) { for(i = PQ_L2_SIZE / 2; i > 0; --i) {
int ij; if ((m = TAILQ_FIRST(pq[(index + i) & PQ_L2_MASK].pl)) != NULL)
for(i = (PQ_L2_SIZE / 2) - PQ_L1_SIZE; break;
(ij = i + j) > 0;
i -= PQ_L1_SIZE) {
hindex = index + ij; if ((m = TAILQ_FIRST(pq[(index - i) & PQ_L2_MASK].pl)) != NULL)
if (hindex >= PQ_L2_SIZE) break;
hindex -= PQ_L2_SIZE;
if ((m = TAILQ_FIRST(pq[hindex].pl)) != NULL)
return m;
hindex = index - ij;
if (hindex < 0)
hindex += PQ_L2_SIZE;
if ((m = TAILQ_FIRST(pq[hindex].pl)) != NULL)
return m;
}
} }
return(m);
hindex = index + PQ_L2_SIZE / 2;
if (hindex >= PQ_L2_SIZE)
hindex -= PQ_L2_SIZE;
m = TAILQ_FIRST(pq[hindex].pl);
if (m)
return m;
return NULL;
#else
return TAILQ_FIRST(vm_page_queues[basequeue].pl);
#endif
} }
/*
* vm_page_select:
*
* Find a page on the specified queue with color optimization.
*
* This routine must be called at splvm().
* This routine may not block.
*/
vm_page_t
vm_page_select(object, pindex, basequeue)
vm_object_t object;
vm_pindex_t pindex;
int basequeue;
{
#if PQ_L2_SIZE > 1
int index;
index = (pindex + object->pg_color) & PQ_L2_MASK;
return vm_page_list_find(basequeue, index);
#else
return TAILQ_FIRST(vm_page_queues[basequeue].pl);
#endif #endif
}
/* /*
* vm_page_select_cache: * vm_page_select_cache:
* *
@ -780,14 +724,10 @@ vm_page_select_cache(object, pindex)
vm_page_t m; vm_page_t m;
while (TRUE) { while (TRUE) {
#if PQ_L2_SIZE > 1 m = vm_page_list_find(
int index; PQ_CACHE,
index = (pindex + object->pg_color) & PQ_L2_MASK; (pindex + object->pg_color) & PQ_L2_MASK
m = vm_page_list_find(PQ_CACHE, index); );
#else
m = TAILQ_FIRST(vm_page_queues[PQ_CACHE].pl);
#endif
if (m && ((m->flags & PG_BUSY) || m->busy || if (m && ((m->flags & PG_BUSY) || m->busy ||
m->hold_count || m->wire_count)) { m->hold_count || m->wire_count)) {
vm_page_deactivate(m); vm_page_deactivate(m);
@ -800,107 +740,74 @@ vm_page_select_cache(object, pindex)
/* /*
* vm_page_select_free: * vm_page_select_free:
* *
* Find a free or zero page, with specified preference. * Find a free or zero page, with specified preference. We attempt to
* inline the nominal case and fall back to _vm_page_select_free()
* otherwise.
* *
* This routine must be called at splvm(). * This routine must be called at splvm().
* This routine may not block. * This routine may not block.
*/ */
static __inline vm_page_t
vm_page_select_free(vm_object_t object, vm_pindex_t pindex, int prefqueue)
{
vm_page_t m;
int otherq = (prefqueue == PQ_ZERO) ? PQ_FREE : PQ_ZERO;
#if PQ_L2_SIZE > 1
int i = (pindex + object->pg_color) & PQ_L2_MASK;
if ((m = TAILQ_FIRST(vm_page_queues[prefqueue+i].pl)) == NULL &&
(m = TAILQ_FIRST(vm_page_queues[otherq+i].pl)) == NULL
) {
m = _vm_page_select_free(object, pindex, prefqueue);
}
#else
if ((m = TAILQ_FIRST(vm_page_queues[prefqueue].pl)) == NULL)
m = TAILQ_FIRST(vm_page_queues[otherq].pl);
#endif
return(m);
}
#if PQ_L2_SIZE > 1
static vm_page_t static vm_page_t
vm_page_select_free(object, pindex, prefqueue) _vm_page_select_free(object, pindex, prefqueue)
vm_object_t object; vm_object_t object;
vm_pindex_t pindex; vm_pindex_t pindex;
int prefqueue; int prefqueue;
{ {
#if PQ_L2_SIZE > 1 int i;
int i,j; int index;
int index, hindex; vm_page_t m = NULL;
#endif
vm_page_t m;
#if 0
vm_page_t mh;
#endif
int oqueuediff;
struct vpgqueues *pq; struct vpgqueues *pq;
struct vpgqueues *po;
if (prefqueue == PQ_ZERO) if (prefqueue == PQ_ZERO) {
oqueuediff = PQ_FREE - PQ_ZERO; pq = &vm_page_queues[PQ_ZERO];
else po = &vm_page_queues[PQ_FREE];
oqueuediff = PQ_ZERO - PQ_FREE; } else {
pq = &vm_page_queues[PQ_FREE];
#if 0 po = &vm_page_queues[PQ_ZERO];
if (mh = object->page_hint) {
if (mh->pindex == (pindex - 1)) {
if ((mh->flags & PG_FICTITIOUS) == 0) {
if ((mh < &vm_page_array[cnt.v_page_count-1]) &&
(mh >= &vm_page_array[0])) {
int queue;
m = mh + 1;
if (VM_PAGE_TO_PHYS(m) == (VM_PAGE_TO_PHYS(mh) + PAGE_SIZE)) {
queue = m->queue - m->pc;
if (queue == PQ_FREE || queue == PQ_ZERO) {
return m;
}
}
}
}
}
}
#endif
pq = &vm_page_queues[prefqueue];
#if PQ_L2_SIZE > 1
index = (pindex + object->pg_color) & PQ_L2_MASK;
if ((m = TAILQ_FIRST(pq[index].pl)) != NULL)
return m;
if ((m = TAILQ_FIRST(pq[index + oqueuediff].pl)) != NULL)
return m;
for(j = 0; j < PQ_L1_SIZE; j++) {
int ij;
for(i = (PQ_L2_SIZE / 2) - PQ_L1_SIZE;
(ij = i + j) >= 0;
i -= PQ_L1_SIZE) {
hindex = index + ij;
if (hindex >= PQ_L2_SIZE)
hindex -= PQ_L2_SIZE;
if ((m = TAILQ_FIRST(pq[hindex].pl)) != NULL)
return m;
if ((m = TAILQ_FIRST(pq[hindex + oqueuediff].pl)) != NULL)
return m;
hindex = index - ij;
if (hindex < 0)
hindex += PQ_L2_SIZE;
if ((m = TAILQ_FIRST(pq[hindex].pl)) != NULL)
return m;
if ((m = TAILQ_FIRST(pq[hindex + oqueuediff].pl)) != NULL)
return m;
}
} }
hindex = index + PQ_L2_SIZE / 2; index = pindex + object->pg_color;
if (hindex >= PQ_L2_SIZE)
hindex -= PQ_L2_SIZE;
if ((m = TAILQ_FIRST(pq[hindex].pl)) != NULL)
return m;
if ((m = TAILQ_FIRST(pq[hindex+oqueuediff].pl)) != NULL)
return m;
#else for(i = PQ_L2_SIZE / 2; i > 0; --i) {
if ((m = TAILQ_FIRST(pq[0].pl)) != NULL) if ((m = TAILQ_FIRST(pq[(index+i) & PQ_L2_MASK].pl)) != NULL)
return m; break;
else if ((m = TAILQ_FIRST(po[(index+i) & PQ_L2_MASK].pl)) != NULL)
return TAILQ_FIRST(pq[oqueuediff].pl); break;
#endif if ((m = TAILQ_FIRST(pq[(index-i) & PQ_L2_MASK].pl)) != NULL)
break;
return NULL; if ((m = TAILQ_FIRST(po[(index-i) & PQ_L2_MASK].pl)) != NULL)
break;
}
return(m);
} }
#endif
/* /*
* vm_page_alloc: * vm_page_alloc:
* *
@ -1088,26 +995,6 @@ vm_page_alloc(object, pindex, page_req)
(cnt.v_free_count < cnt.v_pageout_free_min)) (cnt.v_free_count < cnt.v_pageout_free_min))
pagedaemon_wakeup(); pagedaemon_wakeup();
#if 0
/*
* (code removed - was previously a manual breakout of the act of
* freeing a page from cache. We now just call vm_page_free() on
* a cache page an loop so this code no longer needs to be here)
*/
if ((qtype == PQ_CACHE) &&
((page_req == VM_ALLOC_NORMAL) || (page_req == VM_ALLOC_ZERO)) &&
oldobject && (oldobject->type == OBJT_VNODE) &&
((oldobject->flags & OBJ_DEAD) == 0)) {
struct vnode *vp;
vp = (struct vnode *) oldobject->handle;
if (vp && VSHOULDFREE(vp)) {
if ((vp->v_flag & (VFREE|VTBFREE|VDOOMED)) == 0) {
TAILQ_INSERT_TAIL(&vnode_tobefree_list, vp, v_freelist);
vp->v_flag |= VTBFREE;
}
}
}
#endif
splx(s); splx(s);
return (m); return (m);

View File

@ -61,7 +61,7 @@
* any improvements or extensions that they make and grant Carnegie the * any improvements or extensions that they make and grant Carnegie the
* rights to redistribute these changes. * rights to redistribute these changes.
* *
* $Id: vm_page.h,v 1.52 1999/01/24 01:05:15 dillon Exp $ * $Id: vm_page.h,v 1.53 1999/01/24 05:57:50 dillon Exp $
*/ */
/* /*
@ -147,7 +147,6 @@ struct vm_page {
#define PQ_PRIME2 23 /* Prime number somewhat less than PQ_HASH_SIZE */ #define PQ_PRIME2 23 /* Prime number somewhat less than PQ_HASH_SIZE */
#define PQ_PRIME3 17 /* Prime number somewhat less than PQ_HASH_SIZE */ #define PQ_PRIME3 17 /* Prime number somewhat less than PQ_HASH_SIZE */
#define PQ_L2_SIZE 256 /* A number of colors opt for 1M cache */ #define PQ_L2_SIZE 256 /* A number of colors opt for 1M cache */
#define PQ_L1_SIZE 4 /* Four page L1 cache */
#endif #endif
/* Define one of the following */ /* Define one of the following */
@ -156,7 +155,6 @@ struct vm_page {
#define PQ_PRIME2 23 /* Prime number somewhat less than PQ_HASH_SIZE */ #define PQ_PRIME2 23 /* Prime number somewhat less than PQ_HASH_SIZE */
#define PQ_PRIME3 17 /* Prime number somewhat less than PQ_HASH_SIZE */ #define PQ_PRIME3 17 /* Prime number somewhat less than PQ_HASH_SIZE */
#define PQ_L2_SIZE 128 /* A number of colors opt for 512K cache */ #define PQ_L2_SIZE 128 /* A number of colors opt for 512K cache */
#define PQ_L1_SIZE 4 /* Four page L1 cache (for PII) */
#endif #endif
@ -168,7 +166,6 @@ struct vm_page {
#define PQ_PRIME2 1 #define PQ_PRIME2 1
#define PQ_PRIME3 1 #define PQ_PRIME3 1
#define PQ_L2_SIZE 1 #define PQ_L2_SIZE 1
#define PQ_L1_SIZE 1
#endif #endif
#if defined(PQ_NORMALCACHE) #if defined(PQ_NORMALCACHE)
@ -176,7 +173,6 @@ struct vm_page {
#define PQ_PRIME2 3 /* Prime number somewhat less than PQ_HASH_SIZE */ #define PQ_PRIME2 3 /* Prime number somewhat less than PQ_HASH_SIZE */
#define PQ_PRIME3 11 /* Prime number somewhat less than PQ_HASH_SIZE */ #define PQ_PRIME3 11 /* Prime number somewhat less than PQ_HASH_SIZE */
#define PQ_L2_SIZE 16 /* A reasonable number of colors (opt for 64K cache) */ #define PQ_L2_SIZE 16 /* A reasonable number of colors (opt for 64K cache) */
#define PQ_L1_SIZE 2 /* Two page L1 cache */
#endif #endif
#if defined(PQ_MEDIUMCACHE) || !defined(PQ_L2_SIZE) #if defined(PQ_MEDIUMCACHE) || !defined(PQ_L2_SIZE)
@ -184,7 +180,6 @@ struct vm_page {
#define PQ_PRIME2 7 /* Prime number somewhat less than PQ_HASH_SIZE */ #define PQ_PRIME2 7 /* Prime number somewhat less than PQ_HASH_SIZE */
#define PQ_PRIME3 5 /* Prime number somewhat less than PQ_HASH_SIZE */ #define PQ_PRIME3 5 /* Prime number somewhat less than PQ_HASH_SIZE */
#define PQ_L2_SIZE 64 /* A number of colors opt for 256K cache */ #define PQ_L2_SIZE 64 /* A number of colors opt for 256K cache */
#define PQ_L1_SIZE 2 /* Two page L1 cache */
#endif #endif
#define PQ_L2_MASK (PQ_L2_SIZE - 1) #define PQ_L2_MASK (PQ_L2_SIZE - 1)
@ -405,9 +400,8 @@ static __inline boolean_t vm_page_zero_fill __P((vm_page_t));
int vm_page_is_valid __P((vm_page_t, int, int)); int vm_page_is_valid __P((vm_page_t, int, int));
void vm_page_test_dirty __P((vm_page_t)); void vm_page_test_dirty __P((vm_page_t));
int vm_page_bits __P((int, int)); int vm_page_bits __P((int, int));
vm_page_t vm_page_list_find __P((int, int)); vm_page_t _vm_page_list_find __P((int, int));
int vm_page_queue_index __P((vm_offset_t, int)); int vm_page_queue_index __P((vm_offset_t, int));
vm_page_t vm_page_select __P((vm_object_t, vm_pindex_t, int));
#if 0 #if 0
int vm_page_sleep(vm_page_t m, char *msg, char *busy); int vm_page_sleep(vm_page_t m, char *msg, char *busy);
int vm_page_asleep(vm_page_t m, char *msg, char *busy); int vm_page_asleep(vm_page_t m, char *msg, char *busy);
@ -557,6 +551,20 @@ vm_page_dirty(vm_page_t m)
m->dirty = VM_PAGE_BITS_ALL; m->dirty = VM_PAGE_BITS_ALL;
} }
static __inline vm_page_t
vm_page_list_find(int basequeue, int index)
{
vm_page_t m;
#if PQ_L2_SIZE > 1
m = TAILQ_FIRST(vm_page_queues[basequeue+index].pl);
if (m == NULL)
m = _vm_page_list_find(basequeue, index);
#else
m = TAILQ_FIRST(vm_page_queues[basequeue].pl);
#endif
return(m);
}
#endif /* KERNEL */ #endif /* KERNEL */
#endif /* !_VM_PAGE_ */ #endif /* !_VM_PAGE_ */