14e3eb7296
The work have been under testing and fixing since then, and it is mature enough to be put into HEAD for further testing. A lot have changed in this time, and here are the most important: - Gvinum now uses one single workerthread instead of one thread for each volume and each plex. The reason for this is that the previous scheme was very complex, and was the cause of many of the bugs discovered in gvinum. Instead, gvinum now uses one worker thread with an event queue, quite similar to what used in gmirror. - The rebuild/grow/initialize/parity check routines no longer runs in separate threads, but are run as regular I/O requests with special flags. This made it easier to support mounted growing and parity rebuild. - Support for growing striped and raid5-plexes, meaning that one can extend the volumes for these plex types in addition to the concat type. Also works while the volume is mounted. - Implementation of many of the missing commands from the old vinum: attach/detach, start (was partially implemented), stop (was partially implemented), concat, mirror, stripe, raid5 (shortcuts for creating volumes with one plex of these organizations). - The parity check and rebuild no longer goes between userland/kernel, meaning that the gvinum command will not stay and wait forever for the rebuild to finish. You can instead watch the status with the list command. - Many problems with gvinum have been reported since 5.x, and some has been hard to fix due to the complicated architecture. Hopefully, it should be more stable and better handle edge cases that previously made gvinum crash. - Failed drives no longer disappears entirely, but now leave behind a dummy drive that makes sure the original state is not forgotten in case the system is rebooted between drive failures/swaps. - Update manpage to reflect new commands and extend it with some examples. Sponsored by: Google Summer of Code 2007 Mentored by: le Tested by: Rick C. Petty <rick-freebsd2008 -at- kiwi-computer.com>
165 lines
4.2 KiB
C
165 lines
4.2 KiB
C
/*-
|
|
* Copyright (c) 2007 Lukas Ertl
|
|
* All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*/
|
|
|
|
#include <sys/cdefs.h>
|
|
__FBSDID("$FreeBSD$");
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/bio.h>
|
|
#include <sys/lock.h>
|
|
#include <sys/malloc.h>
|
|
#include <sys/systm.h>
|
|
|
|
#include <geom/geom.h>
|
|
#include <geom/vinum/geom_vinum_var.h>
|
|
#include <geom/vinum/geom_vinum.h>
|
|
|
|
void
|
|
gv_volume_flush(struct gv_volume *v)
|
|
{
|
|
struct gv_softc *sc;
|
|
struct bio *bp;
|
|
|
|
KASSERT(v != NULL, ("NULL v"));
|
|
sc = v->vinumconf;
|
|
KASSERT(sc != NULL, ("NULL sc"));
|
|
|
|
bp = bioq_takefirst(v->wqueue);
|
|
while (bp != NULL) {
|
|
gv_volume_start(sc, bp);
|
|
bp = bioq_takefirst(v->wqueue);
|
|
}
|
|
}
|
|
|
|
void
|
|
gv_volume_start(struct gv_softc *sc, struct bio *bp)
|
|
{
|
|
struct g_geom *gp;
|
|
struct gv_volume *v;
|
|
struct gv_plex *p, *lp;
|
|
int numwrites;
|
|
|
|
gp = sc->geom;
|
|
v = bp->bio_to->private;
|
|
if (v == NULL || v->state != GV_VOL_UP) {
|
|
g_io_deliver(bp, ENXIO);
|
|
return;
|
|
}
|
|
|
|
switch (bp->bio_cmd) {
|
|
case BIO_READ:
|
|
/*
|
|
* Try to find a good plex where we can send the request to,
|
|
* round-robin-style. The plex either has to be up, or it's a
|
|
* degraded RAID5 plex. Check if we have delayed requests. Put
|
|
* this request on the delayed queue if so. This makes sure that
|
|
* we don't read old values.
|
|
*/
|
|
if (bioq_first(v->wqueue) != NULL) {
|
|
bioq_insert_tail(v->wqueue, bp);
|
|
break;
|
|
}
|
|
lp = v->last_read_plex;
|
|
if (lp == NULL)
|
|
lp = LIST_FIRST(&v->plexes);
|
|
p = LIST_NEXT(lp, in_volume);
|
|
if (p == NULL)
|
|
p = LIST_FIRST(&v->plexes);
|
|
do {
|
|
if (p == NULL) {
|
|
p = lp;
|
|
break;
|
|
}
|
|
if ((p->state > GV_PLEX_DEGRADED) ||
|
|
(p->state >= GV_PLEX_DEGRADED &&
|
|
p->org == GV_PLEX_RAID5))
|
|
break;
|
|
p = LIST_NEXT(p, in_volume);
|
|
if (p == NULL)
|
|
p = LIST_FIRST(&v->plexes);
|
|
} while (p != lp);
|
|
|
|
if ((p == NULL) ||
|
|
(p->org == GV_PLEX_RAID5 && p->state < GV_PLEX_DEGRADED) ||
|
|
(p->org != GV_PLEX_RAID5 && p->state <= GV_PLEX_DEGRADED)) {
|
|
g_io_deliver(bp, ENXIO);
|
|
return;
|
|
}
|
|
v->last_read_plex = p;
|
|
|
|
/* Hand it down to the plex logic. */
|
|
gv_plex_start(p, bp);
|
|
break;
|
|
|
|
case BIO_WRITE:
|
|
case BIO_DELETE:
|
|
/* Delay write-requests if any plex is synchronizing. */
|
|
LIST_FOREACH(p, &v->plexes, in_volume) {
|
|
if (p->flags & GV_PLEX_SYNCING) {
|
|
bioq_insert_tail(v->wqueue, bp);
|
|
return;
|
|
}
|
|
}
|
|
|
|
numwrites = 0;
|
|
/* Give the BIO to each plex of this volume. */
|
|
LIST_FOREACH(p, &v->plexes, in_volume) {
|
|
if (p->state < GV_PLEX_DEGRADED)
|
|
continue;
|
|
gv_plex_start(p, bp);
|
|
numwrites++;
|
|
}
|
|
if (numwrites == 0)
|
|
g_io_deliver(bp, ENXIO);
|
|
break;
|
|
}
|
|
}
|
|
|
|
void
|
|
gv_bio_done(struct gv_softc *sc, struct bio *bp)
|
|
{
|
|
struct gv_volume *v;
|
|
struct gv_plex *p;
|
|
struct gv_sd *s;
|
|
|
|
s = bp->bio_caller1;
|
|
KASSERT(s != NULL, ("gv_bio_done: NULL s"));
|
|
p = s->plex_sc;
|
|
KASSERT(p != NULL, ("gv_bio_done: NULL p"));
|
|
v = p->vol_sc;
|
|
KASSERT(v != NULL, ("gv_bio_done: NULL v"));
|
|
|
|
switch (p->org) {
|
|
case GV_PLEX_CONCAT:
|
|
case GV_PLEX_STRIPED:
|
|
gv_plex_normal_done(p, bp);
|
|
break;
|
|
case GV_PLEX_RAID5:
|
|
gv_plex_raid5_done(p, bp);
|
|
break;
|
|
}
|
|
}
|