Implement a bio-taskqueue to reduce number of context switches in

disk I/O processing.

        The intent is that the disk driver in its hardware interrupt
        routine will simply schedule the bio on the task queue with
        a routine to finish off whatever needs done.

        The g_up thread will then schedule this routine, the likely
        outcome of which is a biodone() which queues the bio on
        g_up's regular queue where it will be picked up and processed.

        Compared to the using the regular taskqueue, this saves one
        contextswitch.

Change our scheduling of the g_up and g_down queues to be water-tight,
at the cost of breaking the userland regression test-shims.

Input and ideas from:   scottl
This commit is contained in:
Poul-Henning Kamp 2003-02-11 22:30:26 +00:00
parent 55b84e8a54
commit f0e185d705
3 changed files with 65 additions and 21 deletions

View File

@ -59,6 +59,7 @@
static struct g_bioq g_bio_run_down;
static struct g_bioq g_bio_run_up;
static struct g_bioq g_bio_run_task;
static struct g_bioq g_bio_idle;
static u_int pace;
@ -101,13 +102,11 @@ g_bioq_first(struct g_bioq *bq)
{
struct bio *bp;
g_bioq_lock(bq);
bp = TAILQ_FIRST(&bq->bio_queue);
if (bp != NULL) {
TAILQ_REMOVE(&bq->bio_queue, bp, bio_queue);
bq->bio_queue_length--;
}
g_bioq_unlock(bq);
return (bp);
}
@ -126,7 +125,9 @@ g_new_bio(void)
{
struct bio *bp;
g_bioq_lock(&g_bio_idle);
bp = g_bioq_first(&g_bio_idle);
g_bioq_unlock(&g_bio_idle);
if (bp == NULL)
bp = g_malloc(sizeof *bp, M_NOWAIT | M_ZERO);
/* g_trace(G_T_BIO, "g_new_bio() = %p", bp); */
@ -167,6 +168,7 @@ g_io_init()
g_bioq_init(&g_bio_run_down);
g_bioq_init(&g_bio_run_up);
g_bioq_init(&g_bio_run_task);
g_bioq_init(&g_bio_idle);
}
@ -383,11 +385,20 @@ g_io_schedule_down(struct thread *tp __unused)
struct bio *bp;
off_t excess;
int error;
struct mtx mymutex;
bzero(&mymutex, sizeof mymutex);
mtx_init(&mymutex, "g_xdown", MTX_DEF, 0);
for(;;) {
g_bioq_lock(&g_bio_run_down);
bp = g_bioq_first(&g_bio_run_down);
if (bp == NULL)
break;
if (bp == NULL) {
msleep(&g_wait_down, &g_bio_run_down.bio_queue_lock,
PRIBIO | PDROP, "g_down", hz/10);
continue;
}
g_bioq_unlock(&g_bio_run_down);
error = g_io_check(bp);
if (error) {
g_io_deliver(bp, error);
@ -412,7 +423,9 @@ g_io_schedule_down(struct thread *tp __unused)
default:
break;
}
mtx_lock(&mymutex);
bp->bio_to->geom->start(bp);
mtx_unlock(&mymutex);
if (pace) {
pace--;
break;
@ -420,19 +433,51 @@ g_io_schedule_down(struct thread *tp __unused)
}
}
void
bio_taskqueue(struct bio *bp, bio_task_t *func, void *arg)
{
bp->bio_task = func;
bp->bio_task_arg = arg;
/*
* The taskqueue is actually just a second queue off the "up"
* queue, so we use the same lock.
*/
g_bioq_lock(&g_bio_run_up);
TAILQ_INSERT_TAIL(&g_bio_run_task.bio_queue, bp, bio_queue);
g_bio_run_task.bio_queue_length++;
wakeup(&g_wait_up);
g_bioq_unlock(&g_bio_run_up);
}
void
g_io_schedule_up(struct thread *tp __unused)
{
struct bio *bp;
struct g_consumer *cp;
struct mtx mymutex;
bzero(&mymutex, sizeof mymutex);
mtx_init(&mymutex, "g_xup", MTX_DEF, 0);
for(;;) {
g_bioq_lock(&g_bio_run_up);
bp = g_bioq_first(&g_bio_run_task);
if (bp != NULL) {
g_bioq_unlock(&g_bio_run_up);
mtx_lock(&mymutex);
bp->bio_task(bp, bp->bio_task_arg);
mtx_unlock(&mymutex);
continue;
}
bp = g_bioq_first(&g_bio_run_up);
if (bp == NULL)
break;
cp = bp->bio_from;
biodone(bp);
if (bp != NULL) {
g_bioq_unlock(&g_bio_run_up);
mtx_lock(&mymutex);
biodone(bp);
mtx_unlock(&mymutex);
continue;
}
msleep(&g_wait_up, &g_bio_run_up.bio_queue_lock,
PRIBIO | PDROP, "g_up", hz/10);
}
}

View File

@ -83,16 +83,11 @@ g_up_procbody(void)
{
struct proc *p = g_up_proc;
struct thread *tp = FIRST_THREAD_IN_PROC(p);
struct mtx mymutex;
mtx_assert(&Giant, MA_NOTOWNED);
bzero(&mymutex, sizeof mymutex);
mtx_init(&mymutex, "g_up", MTX_DEF, 0);
mtx_lock(&mymutex);
tp->td_base_pri = PRIBIO;
for(;;) {
g_io_schedule_up(tp);
msleep(&g_wait_up, &mymutex, PRIBIO, "g_up", hz/10);
}
}
@ -109,16 +104,11 @@ g_down_procbody(void)
{
struct proc *p = g_down_proc;
struct thread *tp = FIRST_THREAD_IN_PROC(p);
struct mtx mymutex;
mtx_assert(&Giant, MA_NOTOWNED);
bzero(&mymutex, sizeof mymutex);
mtx_init(&mymutex, "g_down", MTX_DEF, 0);
mtx_lock(&mymutex);
tp->td_base_pri = PRIBIO;
for(;;) {
g_io_schedule_down(tp);
msleep(&g_wait_down, &mymutex, PRIBIO, "g_down", hz/10);
}
}

View File

@ -45,6 +45,10 @@
#include <sys/queue.h>
struct disk;
struct bio;
typedef void bio_task_t(struct bio *, void *);
/*
* The bio structure describes an I/O operation in the kernel.
*/
@ -75,6 +79,9 @@ struct bio {
struct bio *bio_parent; /* Pointer to parent */
struct bintime bio_t0; /* Time request started */
bio_task_t *bio_task; /* Task_queue handler */
void *bio_task_arg; /* Argument to above */
/* XXX: these go away when bio chaining is introduced */
daddr_t bio_pblkno; /* physical block number */
};
@ -133,6 +140,8 @@ void bioq_disksort(struct bio_queue_head *ap, struct bio *bp);
void bioq_init(struct bio_queue_head *head);
void bioq_remove(struct bio_queue_head *head, struct bio *bp);
void bio_taskqueue(struct bio *bp, bio_task_t *fund, void *arg);
int physio(dev_t dev, struct uio *uio, int ioflag);
#define physread physio
#define physwrite physio