/* * Copyright (c) 1994 John S. Dyson * All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice immediately at the beginning of the file, without modification, * this list of conditions, and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * 3. Absolutely no warranty of function or purpose is made by the author * John S. Dyson. * 4. Modifications may be freely made to this file if the above conditions * are met. * * $Id: vfs_bio.c,v 1.7 1994/08/07 14:53:20 davidg Exp $ */ #include #include #include #include #include #include #include #include #include #include #include #include struct buf *buf; /* buffer header pool */ int nbuf; /* number of buffer headers calculated elsewhere */ extern vm_map_t buffer_map, io_map; void vm_hold_free_pages(vm_offset_t from, vm_offset_t to); void vm_hold_load_pages(vm_offset_t from, vm_offset_t to); int needsbuffer; /* * Internal update daemon, process 3 * The variable vfs_update_wakeup allows for internal syncs. */ int vfs_update_wakeup; /* * Initialize buffer headers and related structures. */ void bufinit() { struct buf *bp; int i; TAILQ_INIT(&bswlist); LIST_INIT(&invalhash); /* first, make a null hash table */ for(i=0;ib_flags = B_INVAL; /* we're just an empty header */ bp->b_dev = NODEV; bp->b_vp = NULL; bp->b_rcred = NOCRED; bp->b_wcred = NOCRED; bp->b_qindex = QUEUE_EMPTY; bp->b_vnbufs.le_next = NOLIST; bp->b_data = (caddr_t)kmem_alloc_pageable(buffer_map, MAXBSIZE); TAILQ_INSERT_TAIL(&bufqueues[QUEUE_EMPTY], bp, b_freelist); LIST_INSERT_HEAD(&invalhash, bp, b_hash); } } /* * remove the buffer from the appropriate free list */ void bremfree(struct buf *bp) { int s = splbio(); if( bp->b_qindex != QUEUE_NONE) { TAILQ_REMOVE(&bufqueues[bp->b_qindex], bp, b_freelist); bp->b_qindex = QUEUE_NONE; } else { panic("bremfree: removing a buffer when not on a queue"); } splx(s); } /* * Get a buffer with the specified data. Look in the cache first. */ int bread(struct vnode *vp, daddr_t blkno, int size, struct ucred *cred, struct buf **bpp) { struct buf *bp; bp = getblk (vp, blkno, size, 0, 0); *bpp = bp; /* if not found in cache, do some I/O */ if ((bp->b_flags & B_CACHE) == 0) { if (curproc && curproc->p_stats) /* count block I/O */ curproc->p_stats->p_ru.ru_inblock++; bp->b_flags |= B_READ; bp->b_flags &= ~(B_DONE|B_ERROR|B_INVAL); if( bp->b_rcred == NOCRED) { if (cred != NOCRED) crhold(cred); bp->b_rcred = cred; } VOP_STRATEGY(bp); return( biowait (bp)); } return (0); } /* * Operates like bread, but also starts asynchronous I/O on * read-ahead blocks. */ int breadn(struct vnode *vp, daddr_t blkno, int size, daddr_t *rablkno, int *rabsize, int cnt, struct ucred *cred, struct buf **bpp) { struct buf *bp, *rabp; int i; int rv = 0, readwait = 0; *bpp = bp = getblk (vp, blkno, size, 0, 0); /* if not found in cache, do some I/O */ if ((bp->b_flags & B_CACHE) == 0) { if (curproc && curproc->p_stats) /* count block I/O */ curproc->p_stats->p_ru.ru_inblock++; bp->b_flags |= B_READ; bp->b_flags &= ~(B_DONE|B_ERROR|B_INVAL); if( bp->b_rcred == NOCRED) { if (cred != NOCRED) crhold(cred); bp->b_rcred = cred; } VOP_STRATEGY(bp); ++readwait; } for(i=0;ib_flags & B_CACHE) == 0) { if (curproc && curproc->p_stats) curproc->p_stats->p_ru.ru_inblock++; rabp->b_flags |= B_READ | B_ASYNC; rabp->b_flags &= ~(B_DONE|B_ERROR|B_INVAL); if( rabp->b_rcred == NOCRED) { if (cred != NOCRED) crhold(cred); rabp->b_rcred = cred; } VOP_STRATEGY(rabp); } else { brelse(rabp); } } if( readwait) { rv = biowait (bp); } return (rv); } /* * Write, release buffer on completion. (Done by iodone * if async.) */ int bwrite(struct buf *bp) { int oldflags = bp->b_flags; if(bp->b_flags & B_INVAL) { brelse(bp); return (0); } if(!(bp->b_flags & B_BUSY)) panic("bwrite: buffer is not busy???"); bp->b_flags &= ~(B_READ|B_DONE|B_ERROR|B_DELWRI); bp->b_flags |= B_WRITEINPROG; if (oldflags & B_ASYNC) { if (oldflags & B_DELWRI) { reassignbuf(bp, bp->b_vp); } else if( curproc) { ++curproc->p_stats->p_ru.ru_oublock; } } bp->b_vp->v_numoutput++; VOP_STRATEGY(bp); if( (oldflags & B_ASYNC) == 0) { int rtval = biowait(bp); if (oldflags & B_DELWRI) { reassignbuf(bp, bp->b_vp); } else if( curproc) { ++curproc->p_stats->p_ru.ru_oublock; } brelse(bp); return (rtval); } return(0); } int vn_bwrite(ap) struct vop_bwrite_args *ap; { return (bwrite(ap->a_bp)); } /* * Delayed write. (Buffer is marked dirty). */ void bdwrite(struct buf *bp) { if((bp->b_flags & B_BUSY) == 0) { panic("bdwrite: buffer is not busy"); } if(bp->b_flags & B_INVAL) { brelse(bp); return; } if(bp->b_flags & B_TAPE) { bawrite(bp); return; } bp->b_flags &= ~B_READ; if( (bp->b_flags & B_DELWRI) == 0) { if( curproc) ++curproc->p_stats->p_ru.ru_oublock; bp->b_flags |= B_DONE|B_DELWRI; reassignbuf(bp, bp->b_vp); } brelse(bp); return; } /* * Asynchronous write. * Start output on a buffer, but do not wait for it to complete. * The buffer is released when the output completes. */ void bawrite(struct buf *bp) { bp->b_flags |= B_ASYNC; (void) bwrite(bp); } /* * Release a buffer. */ void brelse(struct buf *bp) { int x; /* anyone need a "free" block? */ x=splbio(); if (needsbuffer) { needsbuffer = 0; wakeup((caddr_t)&needsbuffer); } /* anyone need this block? */ if (bp->b_flags & B_WANTED) { bp->b_flags &= ~(B_WANTED|B_AGE); wakeup((caddr_t)bp); } if (bp->b_flags & B_LOCKED) bp->b_flags &= ~B_ERROR; if ((bp->b_flags & (B_NOCACHE|B_INVAL|B_ERROR)) || (bp->b_bufsize <= 0)) { bp->b_flags |= B_INVAL; bp->b_flags &= ~(B_DELWRI|B_CACHE); if(bp->b_vp) brelvp(bp); } if( bp->b_qindex != QUEUE_NONE) panic("brelse: free buffer onto another queue???"); /* enqueue */ /* buffers with no memory */ if(bp->b_bufsize == 0) { bp->b_qindex = QUEUE_EMPTY; TAILQ_INSERT_HEAD(&bufqueues[QUEUE_EMPTY], bp, b_freelist); LIST_REMOVE(bp, b_hash); LIST_INSERT_HEAD(&invalhash, bp, b_hash); bp->b_dev = NODEV; /* buffers with junk contents */ } else if(bp->b_flags & (B_ERROR|B_INVAL|B_NOCACHE)) { bp->b_qindex = QUEUE_AGE; TAILQ_INSERT_HEAD(&bufqueues[QUEUE_AGE], bp, b_freelist); LIST_REMOVE(bp, b_hash); LIST_INSERT_HEAD(&invalhash, bp, b_hash); bp->b_dev = NODEV; /* buffers that are locked */ } else if(bp->b_flags & B_LOCKED) { bp->b_qindex = QUEUE_LOCKED; TAILQ_INSERT_TAIL(&bufqueues[QUEUE_LOCKED], bp, b_freelist); /* buffers with stale but valid contents */ } else if(bp->b_flags & B_AGE) { bp->b_qindex = QUEUE_AGE; TAILQ_INSERT_TAIL(&bufqueues[QUEUE_AGE], bp, b_freelist); /* buffers with valid and quite potentially reuseable contents */ } else { bp->b_qindex = QUEUE_LRU; TAILQ_INSERT_TAIL(&bufqueues[QUEUE_LRU], bp, b_freelist); } /* unlock */ bp->b_flags &= ~(B_WANTED|B_BUSY|B_ASYNC|B_NOCACHE|B_AGE); splx(x); } int freebufspace; int allocbufspace; /* * Find a buffer header which is available for use. */ struct buf * getnewbuf(int slpflag, int slptimeo) { struct buf *bp; int s; s = splbio(); start: /* can we constitute a new buffer? */ if (bp = bufqueues[QUEUE_EMPTY].tqh_first) { if( bp->b_qindex != QUEUE_EMPTY) panic("getnewbuf: inconsistent EMPTY queue"); bremfree(bp); goto fillbuf; } tryfree: if (bp = bufqueues[QUEUE_AGE].tqh_first) { if( bp->b_qindex != QUEUE_AGE) panic("getnewbuf: inconsistent AGE queue"); bremfree(bp); } else if (bp = bufqueues[QUEUE_LRU].tqh_first) { if( bp->b_qindex != QUEUE_LRU) panic("getnewbuf: inconsistent LRU queue"); bremfree(bp); } else { /* wait for a free buffer of any kind */ needsbuffer = 1; tsleep((caddr_t)&needsbuffer, PRIBIO, "newbuf", 0); splx(s); return (0); } /* if we are a delayed write, convert to an async write */ if (bp->b_flags & B_DELWRI) { bp->b_flags |= B_BUSY; bawrite (bp); goto start; } if(bp->b_vp) brelvp(bp); /* we are not free, nor do we contain interesting data */ if (bp->b_rcred != NOCRED) crfree(bp->b_rcred); if (bp->b_wcred != NOCRED) crfree(bp->b_wcred); fillbuf: bp->b_flags = B_BUSY; LIST_REMOVE(bp, b_hash); LIST_INSERT_HEAD(&invalhash, bp, b_hash); splx(s); bp->b_dev = NODEV; bp->b_vp = NULL; bp->b_blkno = bp->b_lblkno = 0; bp->b_iodone = 0; bp->b_error = 0; bp->b_resid = 0; bp->b_bcount = 0; bp->b_wcred = bp->b_rcred = NOCRED; bp->b_dirtyoff = bp->b_dirtyend = 0; bp->b_validoff = bp->b_validend = 0; return (bp); } /* * Check to see if a block is currently memory resident. */ struct buf * incore(struct vnode *vp, daddr_t blkno) { struct buf *bp; struct bufhashhdr *bh; int s = splbio(); bh = BUFHASH(vp, blkno); bp = bh->lh_first; /* Search hash chain */ while (bp) { if( (bp < buf) || (bp >= buf + nbuf)) { printf("incore: buf out of range: %lx, hash: %d\n", bp, bh - bufhashtbl); panic("incore: buf fault"); } /* hit */ if (bp->b_lblkno == blkno && bp->b_vp == vp && (bp->b_flags & B_INVAL) == 0) { splx(s); return (bp); } bp = bp->b_hash.le_next; } splx(s); return(0); } /* * Get a block given a specified block and offset into a file/device. */ struct buf * getblk(struct vnode *vp, daddr_t blkno, int size, int slpflag, int slptimeo) { struct buf *bp; int s; struct bufhashhdr *bh; s = splbio(); loop: if (bp = incore(vp, blkno)) { if (bp->b_flags & B_BUSY) { bp->b_flags |= B_WANTED; tsleep ((caddr_t)bp, PRIBIO, "getblk", 0); goto loop; } bp->b_flags |= B_BUSY | B_CACHE; bremfree(bp); /* * check for size inconsistancies */ if (bp->b_bcount != size) { printf("getblk: invalid buffer size: %d\n", bp->b_bcount); bp->b_flags |= B_INVAL; bwrite(bp); goto loop; } } else { if ((bp = getnewbuf(0, 0)) == 0) goto loop; allocbuf(bp, size); /* * have to check again, because of a possible * race condition. */ if (incore( vp, blkno)) { allocbuf(bp, 0); bp->b_flags |= B_INVAL; brelse(bp); goto loop; } bp->b_blkno = bp->b_lblkno = blkno; bgetvp(vp, bp); LIST_REMOVE(bp, b_hash); bh = BUFHASH(vp, blkno); LIST_INSERT_HEAD(bh, bp, b_hash); } splx(s); return (bp); } /* * Get an empty, disassociated buffer of given size. */ struct buf * geteblk(int size) { struct buf *bp; while ((bp = getnewbuf(0, 0)) == 0) ; allocbuf(bp, size); bp->b_flags |= B_INVAL; return (bp); } /* * Modify the length of a buffer's underlying buffer storage without * destroying information (unless, of course the buffer is shrinking). */ void allocbuf(struct buf *bp, int size) { int newbsize = round_page(size); if( newbsize == bp->b_bufsize) { bp->b_bcount = size; return; } else if( newbsize < bp->b_bufsize) { vm_hold_free_pages( (vm_offset_t) bp->b_data + newbsize, (vm_offset_t) bp->b_data + bp->b_bufsize); } else if( newbsize > bp->b_bufsize) { vm_hold_load_pages( (vm_offset_t) bp->b_data + bp->b_bufsize, (vm_offset_t) bp->b_data + newbsize); } /* adjust buffer cache's idea of memory allocated to buffer contents */ freebufspace -= newbsize - bp->b_bufsize; allocbufspace += newbsize - bp->b_bufsize; bp->b_bufsize = newbsize; bp->b_bcount = size; } /* * Wait for buffer I/O completion, returning error status. */ int biowait(register struct buf *bp) { int s; s = splbio(); while ((bp->b_flags & B_DONE) == 0) tsleep((caddr_t)bp, PRIBIO, "biowait", 0); if((bp->b_flags & B_ERROR) || bp->b_error) { if ((bp->b_flags & B_INVAL) == 0) { bp->b_flags |= B_INVAL; bp->b_dev = NODEV; LIST_REMOVE(bp, b_hash); LIST_INSERT_HEAD(&invalhash, bp, b_hash); } if (!bp->b_error) bp->b_error = EIO; else bp->b_flags |= B_ERROR; splx(s); return (bp->b_error); } else { splx(s); return (0); } } /* * Finish I/O on a buffer, calling an optional function. * This is usually called from interrupt level, so process blocking * is not *a good idea*. */ void biodone(register struct buf *bp) { int s; s = splbio(); bp->b_flags |= B_DONE; if ((bp->b_flags & B_READ) == 0) { vwakeup(bp); } if (bp->b_flags & B_BOUNCE) vm_bounce_free(bp); /* call optional completion function if requested */ if (bp->b_flags & B_CALL) { bp->b_flags &= ~B_CALL; (*bp->b_iodone)(bp); splx(s); return; } /* * For asynchronous completions, release the buffer now. The brelse * checks for B_WANTED and will do the wakeup there if necessary - * so no need to do a wakeup here in the async case. */ if (bp->b_flags & B_ASYNC) { brelse(bp); } else { bp->b_flags &= ~B_WANTED; wakeup((caddr_t) bp); } splx(s); } int count_lock_queue() { int count; struct buf *bp; count = 0; for(bp = bufqueues[QUEUE_LOCKED].tqh_first; bp != NULL; bp = bp->b_freelist.tqe_next) count++; return(count); } int vfs_update_interval = 30; void vfs_update() { (void) spl0(); while(1) { tsleep((caddr_t)&vfs_update_wakeup, PRIBIO, "update", hz * vfs_update_interval); vfs_update_wakeup = 0; sync(curproc, NULL, NULL); } } /* * these routines are not in the correct place (yet) * also they work *ONLY* for kernel_pmap!!! */ void vm_hold_load_pages(vm_offset_t froma, vm_offset_t toa) { vm_offset_t pg; vm_page_t p; vm_offset_t from = round_page(froma); vm_offset_t to = round_page(toa); for(pg = from ; pg < to ; pg += PAGE_SIZE) { vm_offset_t pa; tryagain: if (cnt.v_free_count <= cnt.v_free_reserved) { VM_WAIT; goto tryagain; } p = vm_page_alloc(kernel_object, pg - VM_MIN_KERNEL_ADDRESS); if( !p) { VM_WAIT; goto tryagain; } vm_page_wire(p); pmap_kenter( pg, VM_PAGE_TO_PHYS(p)); } } void vm_hold_free_pages(vm_offset_t froma, vm_offset_t toa) { vm_offset_t pg; vm_page_t p; vm_offset_t from = round_page(froma); vm_offset_t to = round_page(toa); for(pg = from ; pg < to ; pg += PAGE_SIZE) { p = PHYS_TO_VM_PAGE( pmap_kextract( pg)); pmap_kremove( pg); vm_page_free(p); } } void bufstats() { }