4064755812
The previous change involved calling check_cgmagic() twice in a row
for the same CG in order to differentiate when the CG was already ok vs.
when the CG was rebuilt, but that doesn't work because the second call
(which was supposed to rebuild the CG) returns 0 (indicating that
the CG was not rebuilt) due to the prevfailcg check causing an early
failure return. Fix this by moving the rebuild part of check_cgmagic()
out into a separate function which is called by pass1() when it wants to
rebuild a CG.
Fixes: da86e7a20d
Reported by: pho
Discussed with: mckusick
Sponsored by: Netflix
1475 lines
37 KiB
C
1475 lines
37 KiB
C
/*-
|
|
* SPDX-License-Identifier: BSD-3-Clause
|
|
*
|
|
* Copyright (c) 1980, 1986, 1993
|
|
* The Regents of the University of California. All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
* 3. Neither the name of the University nor the names of its contributors
|
|
* may be used to endorse or promote products derived from this software
|
|
* without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*/
|
|
|
|
#if 0
|
|
#ifndef lint
|
|
static const char sccsid[] = "@(#)utilities.c 8.6 (Berkeley) 5/19/95";
|
|
#endif /* not lint */
|
|
#endif
|
|
#include <sys/cdefs.h>
|
|
__FBSDID("$FreeBSD$");
|
|
|
|
#include <sys/param.h>
|
|
#include <sys/time.h>
|
|
#include <sys/types.h>
|
|
#include <sys/sysctl.h>
|
|
#include <sys/disk.h>
|
|
#include <sys/disklabel.h>
|
|
#include <sys/ioctl.h>
|
|
#include <sys/stat.h>
|
|
|
|
#include <ufs/ufs/dinode.h>
|
|
#include <ufs/ufs/dir.h>
|
|
#include <ufs/ffs/fs.h>
|
|
|
|
#include <err.h>
|
|
#include <errno.h>
|
|
#include <string.h>
|
|
#include <ctype.h>
|
|
#include <fstab.h>
|
|
#include <stdint.h>
|
|
#include <stdio.h>
|
|
#include <stdlib.h>
|
|
#include <time.h>
|
|
#include <unistd.h>
|
|
#include <libufs.h>
|
|
|
|
#include "fsck.h"
|
|
|
|
int sujrecovery = 0;
|
|
|
|
static struct bufarea *allocbuf(const char *);
|
|
static void cg_write(struct bufarea *);
|
|
static void slowio_start(void);
|
|
static void slowio_end(void);
|
|
static void printIOstats(void);
|
|
|
|
static long diskreads, totaldiskreads, totalreads; /* Disk cache statistics */
|
|
static struct timespec startpass, finishpass;
|
|
struct timeval slowio_starttime;
|
|
int slowio_delay_usec = 10000; /* Initial IO delay for background fsck */
|
|
int slowio_pollcnt;
|
|
static struct bufarea cgblk; /* backup buffer for cylinder group blocks */
|
|
static struct bufarea failedbuf; /* returned by failed getdatablk() */
|
|
static TAILQ_HEAD(bufqueue, bufarea) bufqueuehd; /* head of buffer cache LRU */
|
|
static LIST_HEAD(bufhash, bufarea) bufhashhd[HASHSIZE]; /* buffer hash list */
|
|
static struct bufhash freebufs; /* unused buffers */
|
|
static int numbufs; /* size of buffer cache */
|
|
static int cachelookups; /* number of cache lookups */
|
|
static int cachereads; /* number of cache reads */
|
|
static int flushtries; /* number of tries to reclaim memory */
|
|
|
|
char *buftype[BT_NUMBUFTYPES] = BT_NAMES;
|
|
|
|
void
|
|
fsutilinit(void)
|
|
{
|
|
diskreads = totaldiskreads = totalreads = 0;
|
|
bzero(&startpass, sizeof(struct timespec));
|
|
bzero(&finishpass, sizeof(struct timespec));
|
|
bzero(&slowio_starttime, sizeof(struct timeval));
|
|
slowio_delay_usec = 10000;
|
|
slowio_pollcnt = 0;
|
|
flushtries = 0;
|
|
}
|
|
|
|
int
|
|
ftypeok(union dinode *dp)
|
|
{
|
|
switch (DIP(dp, di_mode) & IFMT) {
|
|
|
|
case IFDIR:
|
|
case IFREG:
|
|
case IFBLK:
|
|
case IFCHR:
|
|
case IFLNK:
|
|
case IFSOCK:
|
|
case IFIFO:
|
|
return (1);
|
|
|
|
default:
|
|
if (debug)
|
|
printf("bad file type 0%o\n", DIP(dp, di_mode));
|
|
return (0);
|
|
}
|
|
}
|
|
|
|
int
|
|
reply(const char *question)
|
|
{
|
|
int persevere;
|
|
char c;
|
|
|
|
if (preen)
|
|
pfatal("INTERNAL ERROR: GOT TO reply()");
|
|
persevere = strcmp(question, "CONTINUE") == 0 ||
|
|
strcmp(question, "LOOK FOR ALTERNATE SUPERBLOCKS") == 0;
|
|
printf("\n");
|
|
if (!persevere && (nflag || (fswritefd < 0 && bkgrdflag == 0))) {
|
|
printf("%s? no\n\n", question);
|
|
resolved = 0;
|
|
return (0);
|
|
}
|
|
if (yflag || (persevere && nflag)) {
|
|
printf("%s? yes\n\n", question);
|
|
return (1);
|
|
}
|
|
do {
|
|
printf("%s? [yn] ", question);
|
|
(void) fflush(stdout);
|
|
c = getc(stdin);
|
|
while (c != '\n' && getc(stdin) != '\n') {
|
|
if (feof(stdin)) {
|
|
resolved = 0;
|
|
return (0);
|
|
}
|
|
}
|
|
} while (c != 'y' && c != 'Y' && c != 'n' && c != 'N');
|
|
printf("\n");
|
|
if (c == 'y' || c == 'Y')
|
|
return (1);
|
|
resolved = 0;
|
|
return (0);
|
|
}
|
|
|
|
/*
|
|
* Look up state information for an inode.
|
|
*/
|
|
struct inostat *
|
|
inoinfo(ino_t inum)
|
|
{
|
|
static struct inostat unallocated = { USTATE, 0, 0, 0 };
|
|
struct inostatlist *ilp;
|
|
int iloff;
|
|
|
|
if (inum > maxino)
|
|
errx(EEXIT, "inoinfo: inumber %ju out of range",
|
|
(uintmax_t)inum);
|
|
ilp = &inostathead[inum / sblock.fs_ipg];
|
|
iloff = inum % sblock.fs_ipg;
|
|
if (iloff >= ilp->il_numalloced)
|
|
return (&unallocated);
|
|
return (&ilp->il_stat[iloff]);
|
|
}
|
|
|
|
/*
|
|
* Malloc buffers and set up cache.
|
|
*/
|
|
void
|
|
bufinit(void)
|
|
{
|
|
int i;
|
|
|
|
initbarea(&failedbuf, BT_UNKNOWN);
|
|
failedbuf.b_errs = -1;
|
|
failedbuf.b_un.b_buf = NULL;
|
|
if ((cgblk.b_un.b_buf = Malloc((unsigned int)sblock.fs_bsize)) == NULL)
|
|
errx(EEXIT, "Initial malloc(%d) failed", sblock.fs_bsize);
|
|
initbarea(&cgblk, BT_CYLGRP);
|
|
numbufs = cachelookups = cachereads = 0;
|
|
TAILQ_INIT(&bufqueuehd);
|
|
LIST_INIT(&freebufs);
|
|
for (i = 0; i < HASHSIZE; i++)
|
|
LIST_INIT(&bufhashhd[i]);
|
|
for (i = 0; i < BT_NUMBUFTYPES; i++) {
|
|
readtime[i].tv_sec = totalreadtime[i].tv_sec = 0;
|
|
readtime[i].tv_nsec = totalreadtime[i].tv_nsec = 0;
|
|
readcnt[i] = totalreadcnt[i] = 0;
|
|
}
|
|
}
|
|
|
|
static struct bufarea *
|
|
allocbuf(const char *failreason)
|
|
{
|
|
struct bufarea *bp;
|
|
char *bufp;
|
|
|
|
bp = (struct bufarea *)Malloc(sizeof(struct bufarea));
|
|
bufp = Malloc((unsigned int)sblock.fs_bsize);
|
|
if (bp == NULL || bufp == NULL) {
|
|
errx(EEXIT, "%s", failreason);
|
|
/* NOTREACHED */
|
|
}
|
|
numbufs++;
|
|
bp->b_un.b_buf = bufp;
|
|
TAILQ_INSERT_HEAD(&bufqueuehd, bp, b_list);
|
|
initbarea(bp, BT_UNKNOWN);
|
|
return (bp);
|
|
}
|
|
|
|
/*
|
|
* Manage cylinder group buffers.
|
|
*
|
|
* Use getblk() here rather than cgget() because the cylinder group
|
|
* may be corrupted but we want it anyway so we can fix it.
|
|
*/
|
|
static struct bufarea *cgbufs; /* header for cylinder group cache */
|
|
static int flushtries; /* number of tries to reclaim memory */
|
|
|
|
struct bufarea *
|
|
cglookup(int cg)
|
|
{
|
|
struct bufarea *cgbp;
|
|
struct cg *cgp;
|
|
|
|
if ((unsigned) cg >= sblock.fs_ncg)
|
|
errx(EEXIT, "cglookup: out of range cylinder group %d", cg);
|
|
if (cgbufs == NULL) {
|
|
cgbufs = calloc(sblock.fs_ncg, sizeof(struct bufarea));
|
|
if (cgbufs == NULL)
|
|
errx(EEXIT, "Cannot allocate cylinder group buffers");
|
|
}
|
|
cgbp = &cgbufs[cg];
|
|
if (cgbp->b_un.b_cg != NULL)
|
|
return (cgbp);
|
|
cgp = NULL;
|
|
if (flushtries == 0)
|
|
cgp = Malloc((unsigned int)sblock.fs_cgsize);
|
|
if (cgp == NULL) {
|
|
if (sujrecovery)
|
|
errx(EEXIT,"Ran out of memory during journal recovery");
|
|
flush(fswritefd, &cgblk);
|
|
getblk(&cgblk, cgtod(&sblock, cg), sblock.fs_cgsize);
|
|
return (&cgblk);
|
|
}
|
|
cgbp->b_un.b_cg = cgp;
|
|
initbarea(cgbp, BT_CYLGRP);
|
|
getblk(cgbp, cgtod(&sblock, cg), sblock.fs_cgsize);
|
|
return (cgbp);
|
|
}
|
|
|
|
/*
|
|
* Mark a cylinder group buffer as dirty.
|
|
* Update its check-hash if they are enabled.
|
|
*/
|
|
void
|
|
cgdirty(struct bufarea *cgbp)
|
|
{
|
|
struct cg *cg;
|
|
|
|
cg = cgbp->b_un.b_cg;
|
|
if ((sblock.fs_metackhash & CK_CYLGRP) != 0) {
|
|
cg->cg_ckhash = 0;
|
|
cg->cg_ckhash =
|
|
calculate_crc32c(~0L, (void *)cg, sblock.fs_cgsize);
|
|
}
|
|
dirty(cgbp);
|
|
}
|
|
|
|
/*
|
|
* Attempt to flush a cylinder group cache entry.
|
|
* Return whether the flush was successful.
|
|
*/
|
|
int
|
|
flushentry(void)
|
|
{
|
|
struct bufarea *cgbp;
|
|
|
|
if (sujrecovery || flushtries == sblock.fs_ncg || cgbufs == NULL)
|
|
return (0);
|
|
cgbp = &cgbufs[flushtries++];
|
|
if (cgbp->b_un.b_cg == NULL)
|
|
return (0);
|
|
flush(fswritefd, cgbp);
|
|
free(cgbp->b_un.b_buf);
|
|
cgbp->b_un.b_buf = NULL;
|
|
return (1);
|
|
}
|
|
|
|
/*
|
|
* Manage a cache of filesystem disk blocks.
|
|
*/
|
|
struct bufarea *
|
|
getdatablk(ufs2_daddr_t blkno, long size, int type)
|
|
{
|
|
struct bufarea *bp;
|
|
struct bufhash *bhdp;
|
|
|
|
cachelookups++;
|
|
/*
|
|
* If out of range, return empty buffer with b_err == -1
|
|
*
|
|
* Skip check for inodes because chkrange() considers
|
|
* metadata areas invalid to write data.
|
|
*/
|
|
if (type != BT_INODES && chkrange(blkno, size / sblock.fs_fsize)) {
|
|
failedbuf.b_refcnt++;
|
|
return (&failedbuf);
|
|
}
|
|
bhdp = &bufhashhd[HASH(blkno)];
|
|
LIST_FOREACH(bp, bhdp, b_hash)
|
|
if (bp->b_bno == fsbtodb(&sblock, blkno)) {
|
|
if (debug && bp->b_size != size) {
|
|
prtbuf(bp, "getdatablk: size mismatch");
|
|
pfatal("getdatablk: b_size %d != size %ld\n",
|
|
bp->b_size, size);
|
|
}
|
|
TAILQ_REMOVE(&bufqueuehd, bp, b_list);
|
|
goto foundit;
|
|
}
|
|
/*
|
|
* Move long-term busy buffer back to the front of the LRU so we
|
|
* do not endless inspect them for recycling.
|
|
*/
|
|
bp = TAILQ_LAST(&bufqueuehd, bufqueue);
|
|
if (bp != NULL && bp->b_refcnt != 0) {
|
|
TAILQ_REMOVE(&bufqueuehd, bp, b_list);
|
|
TAILQ_INSERT_HEAD(&bufqueuehd, bp, b_list);
|
|
}
|
|
/*
|
|
* Allocate up to the minimum number of buffers before
|
|
* considering recycling any of them.
|
|
*/
|
|
if (size > sblock.fs_bsize)
|
|
errx(EEXIT, "Excessive buffer size %ld > %d\n", size,
|
|
sblock.fs_bsize);
|
|
if ((bp = LIST_FIRST(&freebufs)) != NULL) {
|
|
LIST_REMOVE(bp, b_hash);
|
|
} else if (numbufs < MINBUFS) {
|
|
bp = allocbuf("cannot create minimal buffer pool");
|
|
} else if (sujrecovery) {
|
|
/*
|
|
* SUJ recovery does not want anything written until it
|
|
* has successfully completed (so it can fail back to
|
|
* full fsck). Thus, we can only recycle clean buffers.
|
|
*/
|
|
TAILQ_FOREACH_REVERSE(bp, &bufqueuehd, bufqueue, b_list)
|
|
if ((bp->b_flags & B_DIRTY) == 0 && bp->b_refcnt == 0)
|
|
break;
|
|
if (bp == NULL)
|
|
bp = allocbuf("Ran out of memory during "
|
|
"journal recovery");
|
|
else
|
|
LIST_REMOVE(bp, b_hash);
|
|
} else {
|
|
/*
|
|
* Recycle oldest non-busy buffer.
|
|
*/
|
|
TAILQ_FOREACH_REVERSE(bp, &bufqueuehd, bufqueue, b_list)
|
|
if (bp->b_refcnt == 0)
|
|
break;
|
|
if (bp == NULL)
|
|
bp = allocbuf("Ran out of memory for buffers");
|
|
else
|
|
LIST_REMOVE(bp, b_hash);
|
|
}
|
|
TAILQ_REMOVE(&bufqueuehd, bp, b_list);
|
|
flush(fswritefd, bp);
|
|
bp->b_type = type;
|
|
LIST_INSERT_HEAD(bhdp, bp, b_hash);
|
|
getblk(bp, blkno, size);
|
|
cachereads++;
|
|
/* fall through */
|
|
foundit:
|
|
TAILQ_INSERT_HEAD(&bufqueuehd, bp, b_list);
|
|
if (debug && bp->b_type != type) {
|
|
printf("getdatablk: buffer type changed to %s",
|
|
BT_BUFTYPE(type));
|
|
prtbuf(bp, "");
|
|
}
|
|
if (bp->b_errs == 0)
|
|
bp->b_refcnt++;
|
|
return (bp);
|
|
}
|
|
|
|
void
|
|
getblk(struct bufarea *bp, ufs2_daddr_t blk, long size)
|
|
{
|
|
ufs2_daddr_t dblk;
|
|
struct timespec start, finish;
|
|
|
|
dblk = fsbtodb(&sblock, blk);
|
|
if (bp->b_bno == dblk) {
|
|
totalreads++;
|
|
} else {
|
|
if (debug) {
|
|
readcnt[bp->b_type]++;
|
|
clock_gettime(CLOCK_REALTIME_PRECISE, &start);
|
|
}
|
|
bp->b_errs = blread(fsreadfd, bp->b_un.b_buf, dblk, size);
|
|
if (debug) {
|
|
clock_gettime(CLOCK_REALTIME_PRECISE, &finish);
|
|
timespecsub(&finish, &start, &finish);
|
|
timespecadd(&readtime[bp->b_type], &finish,
|
|
&readtime[bp->b_type]);
|
|
}
|
|
bp->b_bno = dblk;
|
|
bp->b_size = size;
|
|
}
|
|
}
|
|
|
|
void
|
|
brelse(struct bufarea *bp)
|
|
{
|
|
|
|
if (bp->b_refcnt <= 0)
|
|
prtbuf(bp, "brelse: buffer with negative reference count");
|
|
bp->b_refcnt--;
|
|
}
|
|
|
|
void
|
|
binval(struct bufarea *bp)
|
|
{
|
|
|
|
bp->b_flags &= ~B_DIRTY;
|
|
LIST_REMOVE(bp, b_hash);
|
|
LIST_INSERT_HEAD(&freebufs, bp, b_hash);
|
|
}
|
|
|
|
void
|
|
flush(int fd, struct bufarea *bp)
|
|
{
|
|
struct inode ip;
|
|
|
|
if ((bp->b_flags & B_DIRTY) == 0)
|
|
return;
|
|
bp->b_flags &= ~B_DIRTY;
|
|
if (fswritefd < 0) {
|
|
pfatal("WRITING IN READ_ONLY MODE.\n");
|
|
return;
|
|
}
|
|
if (bp->b_errs != 0)
|
|
pfatal("WRITING %sZERO'ED BLOCK %lld TO DISK\n",
|
|
(bp->b_errs == bp->b_size / dev_bsize) ? "" : "PARTIALLY ",
|
|
(long long)bp->b_bno);
|
|
bp->b_errs = 0;
|
|
/*
|
|
* Write using the appropriate function.
|
|
*/
|
|
switch (bp->b_type) {
|
|
case BT_SUPERBLK:
|
|
if (bp != &sblk)
|
|
pfatal("BUFFER %p DOES NOT MATCH SBLK %p\n",
|
|
bp, &sblk);
|
|
/*
|
|
* Superblocks are always pre-copied so we do not need
|
|
* to check them for copy-on-write.
|
|
*/
|
|
if (sbput(fd, bp->b_un.b_fs, 0) == 0)
|
|
fsmodified = 1;
|
|
break;
|
|
case BT_CYLGRP:
|
|
/*
|
|
* Cylinder groups are always pre-copied so we do not
|
|
* need to check them for copy-on-write.
|
|
*/
|
|
if (sujrecovery)
|
|
cg_write(bp);
|
|
if (cgput(fswritefd, &sblock, bp->b_un.b_cg) == 0)
|
|
fsmodified = 1;
|
|
break;
|
|
case BT_INODES:
|
|
if (debug && sblock.fs_magic == FS_UFS2_MAGIC) {
|
|
struct ufs2_dinode *dp = bp->b_un.b_dinode2;
|
|
int i;
|
|
|
|
for (i = 0; i < bp->b_size; dp++, i += sizeof(*dp)) {
|
|
if (ffs_verify_dinode_ckhash(&sblock, dp) == 0)
|
|
continue;
|
|
pwarn("flush: INODE CHECK-HASH FAILED");
|
|
ip.i_bp = bp;
|
|
ip.i_dp = (union dinode *)dp;
|
|
ip.i_number = bp->b_index + (i / sizeof(*dp));
|
|
prtinode(&ip);
|
|
if (preen || reply("FIX") != 0) {
|
|
if (preen)
|
|
printf(" (FIXED)\n");
|
|
ffs_update_dinode_ckhash(&sblock, dp);
|
|
inodirty(&ip);
|
|
}
|
|
}
|
|
}
|
|
/* FALLTHROUGH */
|
|
default:
|
|
copyonwrite(&sblock, bp, std_checkblkavail);
|
|
blwrite(fd, bp->b_un.b_buf, bp->b_bno, bp->b_size);
|
|
break;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* If there are any snapshots, ensure that all the blocks that they
|
|
* care about have been copied, then release the snapshot inodes.
|
|
* These operations need to be done before we rebuild the cylinder
|
|
* groups so that any block allocations are properly recorded.
|
|
* Since all the cylinder group maps have already been copied in
|
|
* the snapshots, no further snapshot copies will need to be done.
|
|
*/
|
|
void
|
|
snapflush(ufs2_daddr_t (*checkblkavail)(ufs2_daddr_t, long))
|
|
{
|
|
struct bufarea *bp;
|
|
int cnt;
|
|
|
|
if (snapcnt > 0) {
|
|
if (debug)
|
|
printf("Check for snapshot copies\n");
|
|
TAILQ_FOREACH_REVERSE(bp, &bufqueuehd, bufqueue, b_list)
|
|
if ((bp->b_flags & B_DIRTY) != 0)
|
|
copyonwrite(&sblock, bp, checkblkavail);
|
|
for (cnt = 0; cnt < snapcnt; cnt++)
|
|
irelse(&snaplist[cnt]);
|
|
snapcnt = 0;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Journaled soft updates does not maintain cylinder group summary
|
|
* information during cleanup, so this routine recalculates the summary
|
|
* information and updates the superblock summary in preparation for
|
|
* writing out the cylinder group.
|
|
*/
|
|
static void
|
|
cg_write(struct bufarea *bp)
|
|
{
|
|
ufs1_daddr_t fragno, cgbno, maxbno;
|
|
u_int8_t *blksfree;
|
|
struct csum *csp;
|
|
struct cg *cgp;
|
|
int blk;
|
|
int i;
|
|
|
|
/*
|
|
* Fix the frag and cluster summary.
|
|
*/
|
|
cgp = bp->b_un.b_cg;
|
|
cgp->cg_cs.cs_nbfree = 0;
|
|
cgp->cg_cs.cs_nffree = 0;
|
|
bzero(&cgp->cg_frsum, sizeof(cgp->cg_frsum));
|
|
maxbno = fragstoblks(&sblock, sblock.fs_fpg);
|
|
if (sblock.fs_contigsumsize > 0) {
|
|
for (i = 1; i <= sblock.fs_contigsumsize; i++)
|
|
cg_clustersum(cgp)[i] = 0;
|
|
bzero(cg_clustersfree(cgp), howmany(maxbno, CHAR_BIT));
|
|
}
|
|
blksfree = cg_blksfree(cgp);
|
|
for (cgbno = 0; cgbno < maxbno; cgbno++) {
|
|
if (ffs_isfreeblock(&sblock, blksfree, cgbno))
|
|
continue;
|
|
if (ffs_isblock(&sblock, blksfree, cgbno)) {
|
|
ffs_clusteracct(&sblock, cgp, cgbno, 1);
|
|
cgp->cg_cs.cs_nbfree++;
|
|
continue;
|
|
}
|
|
fragno = blkstofrags(&sblock, cgbno);
|
|
blk = blkmap(&sblock, blksfree, fragno);
|
|
ffs_fragacct(&sblock, blk, cgp->cg_frsum, 1);
|
|
for (i = 0; i < sblock.fs_frag; i++)
|
|
if (isset(blksfree, fragno + i))
|
|
cgp->cg_cs.cs_nffree++;
|
|
}
|
|
/*
|
|
* Update the superblock cg summary from our now correct values
|
|
* before writing the block.
|
|
*/
|
|
csp = &sblock.fs_cs(&sblock, cgp->cg_cgx);
|
|
sblock.fs_cstotal.cs_ndir += cgp->cg_cs.cs_ndir - csp->cs_ndir;
|
|
sblock.fs_cstotal.cs_nbfree += cgp->cg_cs.cs_nbfree - csp->cs_nbfree;
|
|
sblock.fs_cstotal.cs_nifree += cgp->cg_cs.cs_nifree - csp->cs_nifree;
|
|
sblock.fs_cstotal.cs_nffree += cgp->cg_cs.cs_nffree - csp->cs_nffree;
|
|
sblock.fs_cs(&sblock, cgp->cg_cgx) = cgp->cg_cs;
|
|
}
|
|
|
|
void
|
|
rwerror(const char *mesg, ufs2_daddr_t blk)
|
|
{
|
|
|
|
if (bkgrdcheck)
|
|
exit(EEXIT);
|
|
if (preen == 0)
|
|
printf("\n");
|
|
pfatal("CANNOT %s: %ld", mesg, (long)blk);
|
|
if (reply("CONTINUE") == 0)
|
|
exit(EEXIT);
|
|
}
|
|
|
|
void
|
|
ckfini(int markclean)
|
|
{
|
|
struct bufarea *bp, *nbp;
|
|
int ofsmodified, cnt, cg;
|
|
|
|
if (bkgrdflag) {
|
|
unlink(snapname);
|
|
if ((!(sblock.fs_flags & FS_UNCLEAN)) != markclean) {
|
|
cmd.value = FS_UNCLEAN;
|
|
cmd.size = markclean ? -1 : 1;
|
|
if (sysctlbyname("vfs.ffs.setflags", 0, 0,
|
|
&cmd, sizeof cmd) == -1)
|
|
pwarn("CANNOT SET FILE SYSTEM DIRTY FLAG\n");
|
|
if (!preen) {
|
|
printf("\n***** FILE SYSTEM MARKED %s *****\n",
|
|
markclean ? "CLEAN" : "DIRTY");
|
|
if (!markclean)
|
|
rerun = 1;
|
|
}
|
|
} else if (!preen && !markclean) {
|
|
printf("\n***** FILE SYSTEM STILL DIRTY *****\n");
|
|
rerun = 1;
|
|
}
|
|
bkgrdflag = 0;
|
|
}
|
|
if (debug && cachelookups > 0)
|
|
printf("cache with %d buffers missed %d of %d (%d%%)\n",
|
|
numbufs, cachereads, cachelookups,
|
|
(int)(cachereads * 100 / cachelookups));
|
|
if (fswritefd < 0) {
|
|
(void)close(fsreadfd);
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* To remain idempotent with partial truncations the buffers
|
|
* must be flushed in this order:
|
|
* 1) cylinder groups (bitmaps)
|
|
* 2) indirect, directory, external attribute, and data blocks
|
|
* 3) inode blocks
|
|
* 4) superblock
|
|
* This ordering preserves access to the modified pointers
|
|
* until they are freed.
|
|
*/
|
|
/* Step 1: cylinder groups */
|
|
if (debug)
|
|
printf("Flush Cylinder groups\n");
|
|
if (cgbufs != NULL) {
|
|
for (cnt = 0; cnt < sblock.fs_ncg; cnt++) {
|
|
if (cgbufs[cnt].b_un.b_cg == NULL)
|
|
continue;
|
|
flush(fswritefd, &cgbufs[cnt]);
|
|
free(cgbufs[cnt].b_un.b_cg);
|
|
}
|
|
free(cgbufs);
|
|
cgbufs = NULL;
|
|
}
|
|
flush(fswritefd, &cgblk);
|
|
free(cgblk.b_un.b_buf);
|
|
cgblk.b_un.b_buf = NULL;
|
|
cnt = 0;
|
|
/* Step 2: indirect, directory, external attribute, and data blocks */
|
|
if (debug)
|
|
printf("Flush indirect, directory, external attribute, "
|
|
"and data blocks\n");
|
|
if (pdirbp != NULL) {
|
|
brelse(pdirbp);
|
|
pdirbp = NULL;
|
|
}
|
|
TAILQ_FOREACH_REVERSE_SAFE(bp, &bufqueuehd, bufqueue, b_list, nbp) {
|
|
switch (bp->b_type) {
|
|
/* These should not be in the buffer cache list */
|
|
case BT_UNKNOWN:
|
|
case BT_SUPERBLK:
|
|
case BT_CYLGRP:
|
|
default:
|
|
prtbuf(bp,"ckfini: improper buffer type on cache list");
|
|
continue;
|
|
/* These are the ones to flush in this step */
|
|
case BT_LEVEL1:
|
|
case BT_LEVEL2:
|
|
case BT_LEVEL3:
|
|
case BT_EXTATTR:
|
|
case BT_DIRDATA:
|
|
case BT_DATA:
|
|
break;
|
|
/* These are the ones to flush in the next step */
|
|
case BT_INODES:
|
|
continue;
|
|
}
|
|
if (debug && bp->b_refcnt != 0)
|
|
prtbuf(bp, "ckfini: clearing in-use buffer");
|
|
TAILQ_REMOVE(&bufqueuehd, bp, b_list);
|
|
LIST_REMOVE(bp, b_hash);
|
|
cnt++;
|
|
flush(fswritefd, bp);
|
|
free(bp->b_un.b_buf);
|
|
free((char *)bp);
|
|
}
|
|
/* Step 3: inode blocks */
|
|
if (debug)
|
|
printf("Flush inode blocks\n");
|
|
if (icachebp != NULL) {
|
|
brelse(icachebp);
|
|
icachebp = NULL;
|
|
}
|
|
TAILQ_FOREACH_REVERSE_SAFE(bp, &bufqueuehd, bufqueue, b_list, nbp) {
|
|
if (debug && bp->b_refcnt != 0)
|
|
prtbuf(bp, "ckfini: clearing in-use buffer");
|
|
TAILQ_REMOVE(&bufqueuehd, bp, b_list);
|
|
LIST_REMOVE(bp, b_hash);
|
|
cnt++;
|
|
flush(fswritefd, bp);
|
|
free(bp->b_un.b_buf);
|
|
free((char *)bp);
|
|
}
|
|
if (numbufs != cnt)
|
|
errx(EEXIT, "panic: lost %d buffers", numbufs - cnt);
|
|
/* Step 4: superblock */
|
|
if (debug)
|
|
printf("Flush the superblock\n");
|
|
flush(fswritefd, &sblk);
|
|
if (havesb && cursnapshot == 0 &&
|
|
sblk.b_bno != sblock.fs_sblockloc / dev_bsize) {
|
|
if (preen || reply("UPDATE STANDARD SUPERBLOCK")) {
|
|
/* Change write destination to standard superblock */
|
|
sblock.fs_sblockactualloc = sblock.fs_sblockloc;
|
|
sblk.b_bno = sblock.fs_sblockloc / dev_bsize;
|
|
sbdirty();
|
|
flush(fswritefd, &sblk);
|
|
} else {
|
|
markclean = 0;
|
|
}
|
|
}
|
|
if (cursnapshot == 0 && sblock.fs_clean != markclean) {
|
|
if ((sblock.fs_clean = markclean) != 0) {
|
|
sblock.fs_flags &= ~(FS_UNCLEAN | FS_NEEDSFSCK);
|
|
sblock.fs_pendingblocks = 0;
|
|
sblock.fs_pendinginodes = 0;
|
|
}
|
|
sbdirty();
|
|
ofsmodified = fsmodified;
|
|
flush(fswritefd, &sblk);
|
|
fsmodified = ofsmodified;
|
|
if (!preen) {
|
|
printf("\n***** FILE SYSTEM MARKED %s *****\n",
|
|
markclean ? "CLEAN" : "DIRTY");
|
|
if (!markclean)
|
|
rerun = 1;
|
|
}
|
|
} else if (!preen) {
|
|
if (markclean) {
|
|
printf("\n***** FILE SYSTEM IS CLEAN *****\n");
|
|
} else {
|
|
printf("\n***** FILE SYSTEM STILL DIRTY *****\n");
|
|
rerun = 1;
|
|
}
|
|
}
|
|
/*
|
|
* Free allocated tracking structures.
|
|
*/
|
|
if (blockmap != NULL)
|
|
free(blockmap);
|
|
blockmap = NULL;
|
|
if (inostathead != NULL) {
|
|
for (cg = 0; cg < sblock.fs_ncg; cg++)
|
|
if (inostathead[cg].il_stat != NULL)
|
|
free((char *)inostathead[cg].il_stat);
|
|
free(inostathead);
|
|
}
|
|
inostathead = NULL;
|
|
inocleanup();
|
|
finalIOstats();
|
|
(void)close(fsreadfd);
|
|
(void)close(fswritefd);
|
|
}
|
|
|
|
/*
|
|
* Print out I/O statistics.
|
|
*/
|
|
void
|
|
IOstats(char *what)
|
|
{
|
|
int i;
|
|
|
|
if (debug == 0)
|
|
return;
|
|
if (diskreads == 0) {
|
|
printf("%s: no I/O\n\n", what);
|
|
return;
|
|
}
|
|
if (startpass.tv_sec == 0)
|
|
startpass = startprog;
|
|
printf("%s: I/O statistics\n", what);
|
|
printIOstats();
|
|
totaldiskreads += diskreads;
|
|
diskreads = 0;
|
|
for (i = 0; i < BT_NUMBUFTYPES; i++) {
|
|
timespecadd(&totalreadtime[i], &readtime[i], &totalreadtime[i]);
|
|
totalreadcnt[i] += readcnt[i];
|
|
readtime[i].tv_sec = readtime[i].tv_nsec = 0;
|
|
readcnt[i] = 0;
|
|
}
|
|
clock_gettime(CLOCK_REALTIME_PRECISE, &startpass);
|
|
}
|
|
|
|
void
|
|
finalIOstats(void)
|
|
{
|
|
int i;
|
|
|
|
if (debug == 0)
|
|
return;
|
|
printf("Final I/O statistics\n");
|
|
totaldiskreads += diskreads;
|
|
diskreads = totaldiskreads;
|
|
startpass = startprog;
|
|
for (i = 0; i < BT_NUMBUFTYPES; i++) {
|
|
timespecadd(&totalreadtime[i], &readtime[i], &totalreadtime[i]);
|
|
totalreadcnt[i] += readcnt[i];
|
|
readtime[i] = totalreadtime[i];
|
|
readcnt[i] = totalreadcnt[i];
|
|
}
|
|
printIOstats();
|
|
}
|
|
|
|
static void printIOstats(void)
|
|
{
|
|
long long msec, totalmsec;
|
|
int i;
|
|
|
|
clock_gettime(CLOCK_REALTIME_PRECISE, &finishpass);
|
|
timespecsub(&finishpass, &startpass, &finishpass);
|
|
printf("Running time: %jd.%03ld sec\n",
|
|
(intmax_t)finishpass.tv_sec, finishpass.tv_nsec / 1000000);
|
|
printf("buffer reads by type:\n");
|
|
for (totalmsec = 0, i = 0; i < BT_NUMBUFTYPES; i++)
|
|
totalmsec += readtime[i].tv_sec * 1000 +
|
|
readtime[i].tv_nsec / 1000000;
|
|
if (totalmsec == 0)
|
|
totalmsec = 1;
|
|
for (i = 0; i < BT_NUMBUFTYPES; i++) {
|
|
if (readcnt[i] == 0)
|
|
continue;
|
|
msec =
|
|
readtime[i].tv_sec * 1000 + readtime[i].tv_nsec / 1000000;
|
|
printf("%21s:%8ld %2ld.%ld%% %4jd.%03ld sec %2lld.%lld%%\n",
|
|
buftype[i], readcnt[i], readcnt[i] * 100 / diskreads,
|
|
(readcnt[i] * 1000 / diskreads) % 10,
|
|
(intmax_t)readtime[i].tv_sec, readtime[i].tv_nsec / 1000000,
|
|
msec * 100 / totalmsec, (msec * 1000 / totalmsec) % 10);
|
|
}
|
|
printf("\n");
|
|
}
|
|
|
|
int
|
|
blread(int fd, char *buf, ufs2_daddr_t blk, long size)
|
|
{
|
|
char *cp;
|
|
int i, errs;
|
|
off_t offset;
|
|
|
|
offset = blk;
|
|
offset *= dev_bsize;
|
|
if (bkgrdflag)
|
|
slowio_start();
|
|
totalreads++;
|
|
diskreads++;
|
|
if (pread(fd, buf, (int)size, offset) == size) {
|
|
if (bkgrdflag)
|
|
slowio_end();
|
|
return (0);
|
|
}
|
|
|
|
/*
|
|
* This is handled specially here instead of in rwerror because
|
|
* rwerror is used for all sorts of errors, not just true read/write
|
|
* errors. It should be refactored and fixed.
|
|
*/
|
|
if (surrender) {
|
|
pfatal("CANNOT READ_BLK: %ld", (long)blk);
|
|
errx(EEXIT, "ABORTING DUE TO READ ERRORS");
|
|
} else
|
|
rwerror("READ BLK", blk);
|
|
|
|
errs = 0;
|
|
memset(buf, 0, (size_t)size);
|
|
printf("THE FOLLOWING DISK SECTORS COULD NOT BE READ:");
|
|
for (cp = buf, i = 0; i < size; i += secsize, cp += secsize) {
|
|
if (pread(fd, cp, (int)secsize, offset + i) != secsize) {
|
|
if (secsize != dev_bsize && dev_bsize != 1)
|
|
printf(" %jd (%jd),",
|
|
(intmax_t)(blk * dev_bsize + i) / secsize,
|
|
(intmax_t)blk + i / dev_bsize);
|
|
else
|
|
printf(" %jd,", (intmax_t)blk + i / dev_bsize);
|
|
errs++;
|
|
}
|
|
}
|
|
printf("\n");
|
|
if (errs)
|
|
resolved = 0;
|
|
return (errs);
|
|
}
|
|
|
|
void
|
|
blwrite(int fd, char *buf, ufs2_daddr_t blk, ssize_t size)
|
|
{
|
|
int i;
|
|
char *cp;
|
|
off_t offset;
|
|
|
|
if (fd < 0)
|
|
return;
|
|
offset = blk;
|
|
offset *= dev_bsize;
|
|
if (pwrite(fd, buf, size, offset) == size) {
|
|
fsmodified = 1;
|
|
return;
|
|
}
|
|
resolved = 0;
|
|
rwerror("WRITE BLK", blk);
|
|
printf("THE FOLLOWING SECTORS COULD NOT BE WRITTEN:");
|
|
for (cp = buf, i = 0; i < size; i += dev_bsize, cp += dev_bsize)
|
|
if (pwrite(fd, cp, dev_bsize, offset + i) != dev_bsize)
|
|
printf(" %jd,", (intmax_t)blk + i / dev_bsize);
|
|
printf("\n");
|
|
return;
|
|
}
|
|
|
|
void
|
|
blerase(int fd, ufs2_daddr_t blk, long size)
|
|
{
|
|
off_t ioarg[2];
|
|
|
|
if (fd < 0)
|
|
return;
|
|
ioarg[0] = blk * dev_bsize;
|
|
ioarg[1] = size;
|
|
ioctl(fd, DIOCGDELETE, ioarg);
|
|
/* we don't really care if we succeed or not */
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* Fill a contiguous region with all-zeroes. Note ZEROBUFSIZE is by
|
|
* definition a multiple of dev_bsize.
|
|
*/
|
|
void
|
|
blzero(int fd, ufs2_daddr_t blk, long size)
|
|
{
|
|
static char *zero;
|
|
off_t offset, len;
|
|
|
|
if (fd < 0)
|
|
return;
|
|
if (zero == NULL) {
|
|
zero = calloc(ZEROBUFSIZE, 1);
|
|
if (zero == NULL)
|
|
errx(EEXIT, "cannot allocate buffer pool");
|
|
}
|
|
offset = blk * dev_bsize;
|
|
if (lseek(fd, offset, 0) < 0)
|
|
rwerror("SEEK BLK", blk);
|
|
while (size > 0) {
|
|
len = MIN(ZEROBUFSIZE, size);
|
|
if (write(fd, zero, len) != len)
|
|
rwerror("WRITE BLK", blk);
|
|
blk += len / dev_bsize;
|
|
size -= len;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Verify cylinder group's magic number and other parameters. If the
|
|
* test fails, offer an option to rebuild the whole cylinder group.
|
|
*
|
|
* Return 1 if the cylinder group is good or return 0 if it is bad.
|
|
*/
|
|
#undef CHK
|
|
#define CHK(lhs, op, rhs, fmt) \
|
|
if (lhs op rhs) { \
|
|
pwarn("UFS%d cylinder group %d failed: " \
|
|
"%s (" #fmt ") %s %s (" #fmt ")\n", \
|
|
sblock.fs_magic == FS_UFS1_MAGIC ? 1 : 2, cg, \
|
|
#lhs, (intmax_t)lhs, #op, #rhs, (intmax_t)rhs); \
|
|
error = 1; \
|
|
}
|
|
int
|
|
check_cgmagic(int cg, struct bufarea *cgbp)
|
|
{
|
|
struct cg *cgp = cgbp->b_un.b_cg;
|
|
uint32_t cghash, calchash;
|
|
static int prevfailcg = -1;
|
|
long start;
|
|
int error;
|
|
|
|
/*
|
|
* Extended cylinder group checks.
|
|
*/
|
|
calchash = cgp->cg_ckhash;
|
|
if ((sblock.fs_metackhash & CK_CYLGRP) != 0 &&
|
|
(ckhashadd & CK_CYLGRP) == 0) {
|
|
cghash = cgp->cg_ckhash;
|
|
cgp->cg_ckhash = 0;
|
|
calchash = calculate_crc32c(~0L, (void *)cgp, sblock.fs_cgsize);
|
|
cgp->cg_ckhash = cghash;
|
|
}
|
|
error = 0;
|
|
CHK(cgp->cg_ckhash, !=, calchash, "%jd");
|
|
CHK(cg_chkmagic(cgp), ==, 0, "%jd");
|
|
CHK(cgp->cg_cgx, !=, cg, "%jd");
|
|
CHK(cgp->cg_ndblk, >, sblock.fs_fpg, "%jd");
|
|
if (sblock.fs_magic == FS_UFS1_MAGIC) {
|
|
CHK(cgp->cg_old_niblk, !=, sblock.fs_ipg, "%jd");
|
|
CHK(cgp->cg_old_ncyl, >, sblock.fs_old_cpg, "%jd");
|
|
} else if (sblock.fs_magic == FS_UFS2_MAGIC) {
|
|
CHK(cgp->cg_niblk, !=, sblock.fs_ipg, "%jd");
|
|
CHK(cgp->cg_initediblk, >, sblock.fs_ipg, "%jd");
|
|
}
|
|
if (cgbase(&sblock, cg) + sblock.fs_fpg < sblock.fs_size) {
|
|
CHK(cgp->cg_ndblk, !=, sblock.fs_fpg, "%jd");
|
|
} else {
|
|
CHK(cgp->cg_ndblk, !=, sblock.fs_size - cgbase(&sblock, cg),
|
|
"%jd");
|
|
}
|
|
start = &cgp->cg_space[0] - (u_char *)(&cgp->cg_firstfield);
|
|
if (sblock.fs_magic == FS_UFS2_MAGIC) {
|
|
CHK(cgp->cg_iusedoff, !=, start, "%jd");
|
|
} else if (sblock.fs_magic == FS_UFS1_MAGIC) {
|
|
CHK(cgp->cg_niblk, !=, 0, "%jd");
|
|
CHK(cgp->cg_initediblk, !=, 0, "%jd");
|
|
CHK(cgp->cg_old_ncyl, !=, sblock.fs_old_cpg, "%jd");
|
|
CHK(cgp->cg_old_niblk, !=, sblock.fs_ipg, "%jd");
|
|
CHK(cgp->cg_old_btotoff, !=, start, "%jd");
|
|
CHK(cgp->cg_old_boff, !=, cgp->cg_old_btotoff +
|
|
sblock.fs_old_cpg * sizeof(int32_t), "%jd");
|
|
CHK(cgp->cg_iusedoff, !=, cgp->cg_old_boff +
|
|
sblock.fs_old_cpg * sizeof(u_int16_t), "%jd");
|
|
}
|
|
CHK(cgp->cg_freeoff, !=,
|
|
cgp->cg_iusedoff + howmany(sblock.fs_ipg, CHAR_BIT), "%jd");
|
|
if (sblock.fs_contigsumsize == 0) {
|
|
CHK(cgp->cg_nextfreeoff, !=,
|
|
cgp->cg_freeoff + howmany(sblock.fs_fpg, CHAR_BIT), "%jd");
|
|
} else {
|
|
CHK(cgp->cg_nclusterblks, !=, cgp->cg_ndblk / sblock.fs_frag,
|
|
"%jd");
|
|
CHK(cgp->cg_clustersumoff, !=,
|
|
roundup(cgp->cg_freeoff + howmany(sblock.fs_fpg, CHAR_BIT),
|
|
sizeof(u_int32_t)) - sizeof(u_int32_t), "%jd");
|
|
CHK(cgp->cg_clusteroff, !=, cgp->cg_clustersumoff +
|
|
(sblock.fs_contigsumsize + 1) * sizeof(u_int32_t), "%jd");
|
|
CHK(cgp->cg_nextfreeoff, !=, cgp->cg_clusteroff +
|
|
howmany(fragstoblks(&sblock, sblock.fs_fpg), CHAR_BIT),
|
|
"%jd");
|
|
}
|
|
if (error == 0)
|
|
return (1);
|
|
if (prevfailcg == cg)
|
|
return (0);
|
|
prevfailcg = cg;
|
|
pfatal("CYLINDER GROUP %d: INTEGRITY CHECK FAILED", cg);
|
|
printf("\n");
|
|
return (0);
|
|
}
|
|
|
|
void
|
|
rebuild_cg(int cg, struct bufarea *cgbp)
|
|
{
|
|
struct cg *cgp = cgbp->b_un.b_cg;
|
|
long start;
|
|
|
|
/*
|
|
* Zero out the cylinder group and then initialize critical fields.
|
|
* Bit maps and summaries will be recalculated by later passes.
|
|
*/
|
|
memset(cgp, 0, (size_t)sblock.fs_cgsize);
|
|
cgp->cg_magic = CG_MAGIC;
|
|
cgp->cg_cgx = cg;
|
|
cgp->cg_niblk = sblock.fs_ipg;
|
|
cgp->cg_initediblk = MIN(sblock.fs_ipg, 2 * INOPB(&sblock));
|
|
if (cgbase(&sblock, cg) + sblock.fs_fpg < sblock.fs_size)
|
|
cgp->cg_ndblk = sblock.fs_fpg;
|
|
else
|
|
cgp->cg_ndblk = sblock.fs_size - cgbase(&sblock, cg);
|
|
start = &cgp->cg_space[0] - (u_char *)(&cgp->cg_firstfield);
|
|
if (sblock.fs_magic == FS_UFS2_MAGIC) {
|
|
cgp->cg_iusedoff = start;
|
|
} else if (sblock.fs_magic == FS_UFS1_MAGIC) {
|
|
cgp->cg_niblk = 0;
|
|
cgp->cg_initediblk = 0;
|
|
cgp->cg_old_ncyl = sblock.fs_old_cpg;
|
|
cgp->cg_old_niblk = sblock.fs_ipg;
|
|
cgp->cg_old_btotoff = start;
|
|
cgp->cg_old_boff = cgp->cg_old_btotoff +
|
|
sblock.fs_old_cpg * sizeof(int32_t);
|
|
cgp->cg_iusedoff = cgp->cg_old_boff +
|
|
sblock.fs_old_cpg * sizeof(u_int16_t);
|
|
}
|
|
cgp->cg_freeoff = cgp->cg_iusedoff + howmany(sblock.fs_ipg, CHAR_BIT);
|
|
cgp->cg_nextfreeoff = cgp->cg_freeoff + howmany(sblock.fs_fpg,CHAR_BIT);
|
|
if (sblock.fs_contigsumsize > 0) {
|
|
cgp->cg_nclusterblks = cgp->cg_ndblk / sblock.fs_frag;
|
|
cgp->cg_clustersumoff =
|
|
roundup(cgp->cg_nextfreeoff, sizeof(u_int32_t));
|
|
cgp->cg_clustersumoff -= sizeof(u_int32_t);
|
|
cgp->cg_clusteroff = cgp->cg_clustersumoff +
|
|
(sblock.fs_contigsumsize + 1) * sizeof(u_int32_t);
|
|
cgp->cg_nextfreeoff = cgp->cg_clusteroff +
|
|
howmany(fragstoblks(&sblock, sblock.fs_fpg), CHAR_BIT);
|
|
}
|
|
cgp->cg_ckhash = calculate_crc32c(~0L, (void *)cgp, sblock.fs_cgsize);
|
|
cgdirty(cgbp);
|
|
}
|
|
|
|
/*
|
|
* allocate a data block with the specified number of fragments
|
|
*/
|
|
ufs2_daddr_t
|
|
allocblk(long startcg, long frags,
|
|
ufs2_daddr_t (*checkblkavail)(ufs2_daddr_t blkno, long frags))
|
|
{
|
|
ufs2_daddr_t blkno, newblk;
|
|
|
|
if (sujrecovery && checkblkavail == std_checkblkavail) {
|
|
pfatal("allocblk: std_checkblkavail used for SUJ recovery\n");
|
|
return (0);
|
|
}
|
|
if (frags <= 0 || frags > sblock.fs_frag)
|
|
return (0);
|
|
for (blkno = MAX(cgdata(&sblock, startcg), 0);
|
|
blkno < maxfsblock - sblock.fs_frag;
|
|
blkno += sblock.fs_frag) {
|
|
if ((newblk = (*checkblkavail)(blkno, frags)) == 0)
|
|
continue;
|
|
if (newblk > 0)
|
|
return (newblk);
|
|
if (newblk < 0)
|
|
blkno = -newblk;
|
|
}
|
|
for (blkno = MAX(cgdata(&sblock, 0), 0);
|
|
blkno < cgbase(&sblock, startcg) - sblock.fs_frag;
|
|
blkno += sblock.fs_frag) {
|
|
if ((newblk = (*checkblkavail)(blkno, frags)) == 0)
|
|
continue;
|
|
if (newblk > 0)
|
|
return (newblk);
|
|
if (newblk < 0)
|
|
blkno = -newblk;
|
|
}
|
|
return (0);
|
|
}
|
|
|
|
ufs2_daddr_t
|
|
std_checkblkavail(ufs2_daddr_t blkno, long frags)
|
|
{
|
|
struct bufarea *cgbp;
|
|
struct cg *cgp;
|
|
ufs2_daddr_t j, k, baseblk;
|
|
long cg;
|
|
|
|
if ((u_int64_t)blkno > sblock.fs_size)
|
|
return (0);
|
|
for (j = 0; j <= sblock.fs_frag - frags; j++) {
|
|
if (testbmap(blkno + j))
|
|
continue;
|
|
for (k = 1; k < frags; k++)
|
|
if (testbmap(blkno + j + k))
|
|
break;
|
|
if (k < frags) {
|
|
j += k;
|
|
continue;
|
|
}
|
|
cg = dtog(&sblock, blkno + j);
|
|
cgbp = cglookup(cg);
|
|
cgp = cgbp->b_un.b_cg;
|
|
if (!check_cgmagic(cg, cgbp))
|
|
return (-((cg + 1) * sblock.fs_fpg - sblock.fs_frag));
|
|
baseblk = dtogd(&sblock, blkno + j);
|
|
for (k = 0; k < frags; k++) {
|
|
setbmap(blkno + j + k);
|
|
clrbit(cg_blksfree(cgp), baseblk + k);
|
|
}
|
|
n_blks += frags;
|
|
if (frags == sblock.fs_frag)
|
|
cgp->cg_cs.cs_nbfree--;
|
|
else
|
|
cgp->cg_cs.cs_nffree -= frags;
|
|
cgdirty(cgbp);
|
|
return (blkno + j);
|
|
}
|
|
return (0);
|
|
}
|
|
|
|
/*
|
|
* Slow down IO so as to leave some disk bandwidth for other processes
|
|
*/
|
|
void
|
|
slowio_start()
|
|
{
|
|
|
|
/* Delay one in every 8 operations */
|
|
slowio_pollcnt = (slowio_pollcnt + 1) & 7;
|
|
if (slowio_pollcnt == 0) {
|
|
gettimeofday(&slowio_starttime, NULL);
|
|
}
|
|
}
|
|
|
|
void
|
|
slowio_end()
|
|
{
|
|
struct timeval tv;
|
|
int delay_usec;
|
|
|
|
if (slowio_pollcnt != 0)
|
|
return;
|
|
|
|
/* Update the slowdown interval. */
|
|
gettimeofday(&tv, NULL);
|
|
delay_usec = (tv.tv_sec - slowio_starttime.tv_sec) * 1000000 +
|
|
(tv.tv_usec - slowio_starttime.tv_usec);
|
|
if (delay_usec < 64)
|
|
delay_usec = 64;
|
|
if (delay_usec > 2500000)
|
|
delay_usec = 2500000;
|
|
slowio_delay_usec = (slowio_delay_usec * 63 + delay_usec) >> 6;
|
|
/* delay by 8 times the average IO delay */
|
|
if (slowio_delay_usec > 64)
|
|
usleep(slowio_delay_usec * 8);
|
|
}
|
|
|
|
/*
|
|
* Find a pathname
|
|
*/
|
|
void
|
|
getpathname(char *namebuf, ino_t curdir, ino_t ino)
|
|
{
|
|
int len;
|
|
char *cp;
|
|
struct inode ip;
|
|
struct inodesc idesc;
|
|
static int busy = 0;
|
|
|
|
if (curdir == ino && ino == UFS_ROOTINO) {
|
|
(void)strcpy(namebuf, "/");
|
|
return;
|
|
}
|
|
if (busy || !INO_IS_DVALID(curdir)) {
|
|
(void)strcpy(namebuf, "?");
|
|
return;
|
|
}
|
|
busy = 1;
|
|
memset(&idesc, 0, sizeof(struct inodesc));
|
|
idesc.id_type = DATA;
|
|
idesc.id_fix = IGNORE;
|
|
cp = &namebuf[MAXPATHLEN - 1];
|
|
*cp = '\0';
|
|
if (curdir != ino) {
|
|
idesc.id_parent = curdir;
|
|
goto namelookup;
|
|
}
|
|
while (ino != UFS_ROOTINO) {
|
|
idesc.id_number = ino;
|
|
idesc.id_func = findino;
|
|
idesc.id_name = strdup("..");
|
|
ginode(ino, &ip);
|
|
if ((ckinode(ip.i_dp, &idesc) & FOUND) == 0) {
|
|
irelse(&ip);
|
|
free(idesc.id_name);
|
|
break;
|
|
}
|
|
irelse(&ip);
|
|
free(idesc.id_name);
|
|
namelookup:
|
|
idesc.id_number = idesc.id_parent;
|
|
idesc.id_parent = ino;
|
|
idesc.id_func = findname;
|
|
idesc.id_name = namebuf;
|
|
ginode(idesc.id_number, &ip);
|
|
if ((ckinode(ip.i_dp, &idesc) & FOUND) == 0) {
|
|
irelse(&ip);
|
|
break;
|
|
}
|
|
irelse(&ip);
|
|
len = strlen(namebuf);
|
|
cp -= len;
|
|
memmove(cp, namebuf, (size_t)len);
|
|
*--cp = '/';
|
|
if (cp < &namebuf[UFS_MAXNAMLEN])
|
|
break;
|
|
ino = idesc.id_number;
|
|
}
|
|
busy = 0;
|
|
if (ino != UFS_ROOTINO)
|
|
*--cp = '?';
|
|
memmove(namebuf, cp, (size_t)(&namebuf[MAXPATHLEN] - cp));
|
|
}
|
|
|
|
void
|
|
catch(int sig __unused)
|
|
{
|
|
|
|
ckfini(0);
|
|
exit(12);
|
|
}
|
|
|
|
/*
|
|
* When preening, allow a single quit to signal
|
|
* a special exit after file system checks complete
|
|
* so that reboot sequence may be interrupted.
|
|
*/
|
|
void
|
|
catchquit(int sig __unused)
|
|
{
|
|
printf("returning to single-user after file system check\n");
|
|
returntosingle = 1;
|
|
(void)signal(SIGQUIT, SIG_DFL);
|
|
}
|
|
|
|
/*
|
|
* determine whether an inode should be fixed.
|
|
*/
|
|
int
|
|
dofix(struct inodesc *idesc, const char *msg)
|
|
{
|
|
|
|
switch (idesc->id_fix) {
|
|
|
|
case DONTKNOW:
|
|
if (idesc->id_type == DATA)
|
|
direrror(idesc->id_number, msg);
|
|
else
|
|
pwarn("%s", msg);
|
|
if (preen) {
|
|
printf(" (SALVAGED)\n");
|
|
idesc->id_fix = FIX;
|
|
return (ALTERED);
|
|
}
|
|
if (reply("SALVAGE") == 0) {
|
|
idesc->id_fix = NOFIX;
|
|
return (0);
|
|
}
|
|
idesc->id_fix = FIX;
|
|
return (ALTERED);
|
|
|
|
case FIX:
|
|
return (ALTERED);
|
|
|
|
case NOFIX:
|
|
case IGNORE:
|
|
return (0);
|
|
|
|
default:
|
|
errx(EEXIT, "UNKNOWN INODESC FIX MODE %d", idesc->id_fix);
|
|
}
|
|
/* NOTREACHED */
|
|
return (0);
|
|
}
|
|
|
|
#include <stdarg.h>
|
|
|
|
/*
|
|
* Print details about a buffer.
|
|
*/
|
|
void
|
|
prtbuf(struct bufarea *bp, const char *fmt, ...)
|
|
{
|
|
va_list ap;
|
|
va_start(ap, fmt);
|
|
if (preen)
|
|
(void)fprintf(stdout, "%s: ", cdevname);
|
|
(void)vfprintf(stdout, fmt, ap);
|
|
va_end(ap);
|
|
printf(": bp %p, type %s, bno %jd, size %d, refcnt %d, flags %s, "
|
|
"index %jd\n", bp, BT_BUFTYPE(bp->b_type), (intmax_t) bp->b_bno,
|
|
bp->b_size, bp->b_refcnt, bp->b_flags & B_DIRTY ? "dirty" : "clean",
|
|
(intmax_t) bp->b_index);
|
|
}
|
|
|
|
/*
|
|
* An unexpected inconsistency occurred.
|
|
* Die if preening or file system is running with soft dependency protocol,
|
|
* otherwise just print message and continue.
|
|
*/
|
|
void
|
|
pfatal(const char *fmt, ...)
|
|
{
|
|
va_list ap;
|
|
va_start(ap, fmt);
|
|
if (!preen) {
|
|
(void)vfprintf(stdout, fmt, ap);
|
|
va_end(ap);
|
|
if (usedsoftdep)
|
|
(void)fprintf(stdout,
|
|
"\nUNEXPECTED SOFT UPDATE INCONSISTENCY\n");
|
|
/*
|
|
* Force foreground fsck to clean up inconsistency.
|
|
*/
|
|
if (bkgrdflag) {
|
|
cmd.value = FS_NEEDSFSCK;
|
|
cmd.size = 1;
|
|
if (sysctlbyname("vfs.ffs.setflags", 0, 0,
|
|
&cmd, sizeof cmd) == -1)
|
|
pwarn("CANNOT SET FS_NEEDSFSCK FLAG\n");
|
|
fprintf(stdout, "CANNOT RUN IN BACKGROUND\n");
|
|
ckfini(0);
|
|
exit(EEXIT);
|
|
}
|
|
return;
|
|
}
|
|
if (cdevname == NULL)
|
|
cdevname = strdup("fsck");
|
|
(void)fprintf(stdout, "%s: ", cdevname);
|
|
(void)vfprintf(stdout, fmt, ap);
|
|
(void)fprintf(stdout,
|
|
"\n%s: UNEXPECTED%sINCONSISTENCY; RUN fsck MANUALLY.\n",
|
|
cdevname, usedsoftdep ? " SOFT UPDATE " : " ");
|
|
/*
|
|
* Force foreground fsck to clean up inconsistency.
|
|
*/
|
|
if (bkgrdflag) {
|
|
cmd.value = FS_NEEDSFSCK;
|
|
cmd.size = 1;
|
|
if (sysctlbyname("vfs.ffs.setflags", 0, 0,
|
|
&cmd, sizeof cmd) == -1)
|
|
pwarn("CANNOT SET FS_NEEDSFSCK FLAG\n");
|
|
}
|
|
ckfini(0);
|
|
exit(EEXIT);
|
|
}
|
|
|
|
/*
|
|
* Pwarn just prints a message when not preening or running soft dependency
|
|
* protocol, or a warning (preceded by filename) when preening.
|
|
*/
|
|
void
|
|
pwarn(const char *fmt, ...)
|
|
{
|
|
va_list ap;
|
|
va_start(ap, fmt);
|
|
if (preen)
|
|
(void)fprintf(stdout, "%s: ", cdevname);
|
|
(void)vfprintf(stdout, fmt, ap);
|
|
va_end(ap);
|
|
}
|
|
|
|
/*
|
|
* Stub for routines from kernel.
|
|
*/
|
|
void
|
|
panic(const char *fmt, ...)
|
|
{
|
|
va_list ap;
|
|
va_start(ap, fmt);
|
|
pfatal("INTERNAL INCONSISTENCY:");
|
|
(void)vfprintf(stdout, fmt, ap);
|
|
va_end(ap);
|
|
exit(EEXIT);
|
|
}
|