add USL's copyright notice
[unix-history] / usr / src / sys / kern / vfs_bio.c
index fac2dfc..43123da 100644 (file)
@@ -1,23 +1,29 @@
 /*-
 /*-
- * Copyright (c) 1982, 1986, 1989 The Regents of the University of California.
- * All rights reserved.
+ * Copyright (c) 1986, 1989, 1993
+ *     The Regents of the University of California.  All rights reserved.
+ * (c) UNIX System Laboratories, Inc.
+ * All or some portions of this file are derived from material licensed
+ * to the University of California by American Telephone and Telegraph
+ * Co. or Unix System Laboratories, Inc. and are reproduced herein with
+ * the permission of UNIX System Laboratories, Inc.
  *
  *
- * This module is believed to contain source code proprietary to AT&T.
- * Use and redistribution is subject to the Berkeley Software License
- * Agreement and your Software Agreement with AT&T (Western Electric).
+ * This code is derived from software contributed to Berkeley by
+ * Berkeley Software Design Inc.
  *
  *
- *     @(#)vfs_bio.c   7.59.1.1 (Berkeley) %G%
+ * %sccs.include.redist.c%
+ *
+ *     @(#)vfs_bio.c   8.7 (Berkeley) %G%
  */
 
 #include <sys/param.h>
  */
 
 #include <sys/param.h>
+#include <sys/systm.h>
 #include <sys/proc.h>
 #include <sys/buf.h>
 #include <sys/vnode.h>
 #include <sys/mount.h>
 #include <sys/trace.h>
 #include <sys/proc.h>
 #include <sys/buf.h>
 #include <sys/vnode.h>
 #include <sys/mount.h>
 #include <sys/trace.h>
-#include <sys/resourcevar.h>
 #include <sys/malloc.h>
 #include <sys/malloc.h>
-#include <libkern/libkern.h>
+#include <sys/resourcevar.h>
 #include <ufs/ufs/quota.h>
 #include <ufs/ufs/inode.h>
 
 #include <ufs/ufs/quota.h>
 #include <ufs/ufs/inode.h>
 
  */
 #define        BUFHASH(dvp, lbn)       \
        (&bufhashtbl[((int)(dvp) / sizeof(*(dvp)) + (int)(lbn)) & bufhash])
  */
 #define        BUFHASH(dvp, lbn)       \
        (&bufhashtbl[((int)(dvp) / sizeof(*(dvp)) + (int)(lbn)) & bufhash])
-struct list_entry *bufhashtbl, invalhash;
+LIST_HEAD(bufhashhdr, buf) *bufhashtbl, invalhash;
 u_long bufhash;
 
 /*
  * Insq/Remq for the buffer hash lists.
  */
 u_long bufhash;
 
 /*
  * Insq/Remq for the buffer hash lists.
  */
-#define        binshash(bp, dp)        list_enter_head(dp, bp, struct buf *, b_hash)
-#define        bremhash(bp)            list_remove(bp, struct buf *, b_hash)
+#define        binshash(bp, dp)        LIST_INSERT_HEAD(dp, bp, b_hash)
+#define        bremhash(bp)            LIST_REMOVE(bp, b_hash)
 
 /*
  * Definitions for the buffer free lists.
 
 /*
  * Definitions for the buffer free lists.
@@ -45,46 +51,36 @@ u_long      bufhash;
 #define        BQ_AGE          2               /* rubbish */
 #define        BQ_EMPTY        3               /* buffer headers with no memory */
 
 #define        BQ_AGE          2               /* rubbish */
 #define        BQ_EMPTY        3               /* buffer headers with no memory */
 
-struct queue_entry bufqueues[BQUEUES];
+TAILQ_HEAD(bqueues, buf) bufqueues[BQUEUES];
 int needbuffer;
 
 /*
  * Insq/Remq for the buffer free lists.
  */
 int needbuffer;
 
 /*
  * Insq/Remq for the buffer free lists.
  */
-#define        binsheadfree(bp, dp) \
-       queue_enter_head(dp, bp, struct buf *, b_freelist)
-#define        binstailfree(bp, dp) \
-       queue_enter_tail(dp, bp, struct buf *, b_freelist)
-
-/*
- * Local declarations
- */
-struct buf *cluster_newbuf __P((struct vnode *, struct buf *, long, daddr_t,
-           daddr_t, long, int));
-struct buf *cluster_rbuild __P((struct vnode *, u_quad_t, struct buf *,
-           daddr_t, daddr_t, long, int, long));
-void       cluster_wbuild __P((struct vnode *, struct buf *, long size,
-           daddr_t start_lbn, int len, daddr_t lbn));
+#define        binsheadfree(bp, dp)    TAILQ_INSERT_HEAD(dp, bp, b_freelist)
+#define        binstailfree(bp, dp)    TAILQ_INSERT_TAIL(dp, bp, b_freelist)
 
 void
 bremfree(bp)
        struct buf *bp;
 {
 
 void
 bremfree(bp)
        struct buf *bp;
 {
-       struct queue_entry *dp;
+       struct bqueues *dp = NULL;
 
        /*
         * We only calculate the head of the freelist when removing
         * the last element of the list as that is the only time that
         * it is needed (e.g. to reset the tail pointer).
 
        /*
         * We only calculate the head of the freelist when removing
         * the last element of the list as that is the only time that
         * it is needed (e.g. to reset the tail pointer).
+        *
+        * NB: This makes an assumption about how tailq's are implemented.
         */
         */
-       if (bp->b_freelist.qe_next == NULL) {
+       if (bp->b_freelist.tqe_next == NULL) {
                for (dp = bufqueues; dp < &bufqueues[BQUEUES]; dp++)
                for (dp = bufqueues; dp < &bufqueues[BQUEUES]; dp++)
-                       if (dp->qe_prev == &bp->b_freelist.qe_next)
+                       if (dp->tqh_last == &bp->b_freelist.tqe_next)
                                break;
                if (dp == &bufqueues[BQUEUES])
                        panic("bremfree: lost tail");
        }
                                break;
                if (dp == &bufqueues[BQUEUES])
                        panic("bremfree: lost tail");
        }
-       queue_remove(dp, bp, struct buf *, b_freelist);
+       TAILQ_REMOVE(dp, bp, b_freelist);
 }
 
 /*
 }
 
 /*
@@ -94,13 +90,13 @@ void
 bufinit()
 {
        register struct buf *bp;
 bufinit()
 {
        register struct buf *bp;
-       struct queue_entry *dp;
+       struct bqueues *dp;
        register int i;
        int base, residual;
 
        for (dp = bufqueues; dp < &bufqueues[BQUEUES]; dp++)
        register int i;
        int base, residual;
 
        for (dp = bufqueues; dp < &bufqueues[BQUEUES]; dp++)
-               queue_init(dp);
-       bufhashtbl = (struct list_entry *)hashinit(nbuf, M_CACHE, &bufhash);
+               TAILQ_INIT(dp);
+       bufhashtbl = hashinit(nbuf, M_CACHE, &bufhash);
        base = bufpages / nbuf;
        residual = bufpages % nbuf;
        for (i = 0; i < nbuf; i++) {
        base = bufpages / nbuf;
        residual = bufpages % nbuf;
        for (i = 0; i < nbuf; i++) {
@@ -109,7 +105,8 @@ bufinit()
                bp->b_dev = NODEV;
                bp->b_rcred = NOCRED;
                bp->b_wcred = NOCRED;
                bp->b_dev = NODEV;
                bp->b_rcred = NOCRED;
                bp->b_wcred = NOCRED;
-               bp->b_un.b_addr = buffers + i * MAXBSIZE;
+               bp->b_vnbufs.le_next = NOLIST;
+               bp->b_data = buffers + i * MAXBSIZE;
                if (i < residual)
                        bp->b_bufsize = (base + 1) * CLBYTES;
                else
                if (i < residual)
                        bp->b_bufsize = (base + 1) * CLBYTES;
                else
@@ -245,318 +242,6 @@ breadn(vp, blkno, size, rablkno, rabsize, num, cred, bpp)
        return (biowait(bp));
 }
 
        return (biowait(bp));
 }
 
-/*
- * We could optimize this by keeping track of where the last read-ahead
- * was, but it would involve adding fields to the vnode.  For now, let's
- * just get it working.
- *
- * This replaces bread.  If this is a bread at the beginning of a file and
- * lastr is 0, we assume this is the first read and we'll read up to two
- * blocks if they are sequential.  After that, we'll do regular read ahead
- * in clustered chunks.
- *
- * There are 4 or 5 cases depending on how you count:
- *     Desired block is in the cache:
- *         1 Not sequential access (0 I/Os).
- *         2 Access is sequential, do read-ahead (1 ASYNC).
- *     Desired block is not in cache:
- *         3 Not sequential access (1 SYNC).
- *         4 Sequential access, next block is contiguous (1 SYNC).
- *         5 Sequential access, next block is not contiguous (1 SYNC, 1 ASYNC)
- *
- * There are potentially two buffers that require I/O.
- *     bp is the block requested.
- *     rbp is the read-ahead block.
- *     If either is NULL, then you don't have to do the I/O.
- */
-cluster_read(vp, filesize, lblkno, size, cred, bpp)
-       struct vnode *vp;
-       u_quad_t filesize;
-       daddr_t lblkno;
-       long size;
-       struct ucred *cred;
-       struct buf **bpp;
-{
-       struct buf *bp, *rbp;
-       daddr_t blkno, ioblkno;
-       long flags;
-       int error, num_ra, alreadyincore;
-
-#ifdef DIAGNOSTIC
-       if (size == 0)
-               panic("cluster_read: size = 0");
-#endif
-
-       error = 0;
-       flags = B_READ;
-       *bpp = bp = getblk(vp, lblkno, size, 0, 0);
-       if (bp->b_flags & (B_CACHE | B_DONE | B_DELWRI)) {
-               /*
-                * Desired block is in cache; do any readahead ASYNC.
-                * Case 1, 2.
-                */
-               trace(TR_BREADHIT, pack(vp, size), lblkno);
-               flags |= B_ASYNC;
-               ioblkno = lblkno +
-                   (lblkno < vp->v_ralen ? vp->v_ralen >> 1 : vp->v_ralen);
-               alreadyincore = (int)incore(vp, ioblkno);
-               bp = NULL;
-       } else {
-               /* Block wasn't in cache, case 3, 4, 5. */
-               trace(TR_BREADMISS, pack(vp, size), lblkno);
-               ioblkno = lblkno;
-               bp->b_flags |= flags;
-               alreadyincore = 0;
-               curproc->p_stats->p_ru.ru_inblock++;            /* XXX */
-       }
-       /*
-        * XXX
-        * Replace 1 with a window size based on some permutation of
-        * maxcontig and rot_delay.  This will let you figure out how
-        * many blocks you should read-ahead (case 2, 4, 5).
-        *
-        * If the access isn't sequential, cut the window size in half.
-        */
-       rbp = NULL;
-       if (lblkno != vp->v_lastr + 1 && lblkno != 0)
-               vp->v_ralen = max(vp->v_ralen >> 1, 1);
-       else if ((ioblkno + 1) * size < filesize && !alreadyincore &&
-           !(error = VOP_BMAP(vp, ioblkno, NULL, &blkno, &num_ra))) {
-               /*
-                * Reading sequentially, and the next block is not in the
-                * cache.  We are going to try reading ahead. If this is
-                * the first read of a file, then limit read-ahead to a
-                * single block, else read as much as we're allowed.
-                */
-               if (num_ra > vp->v_ralen) {
-                       num_ra = vp->v_ralen;
-                       vp->v_ralen = min(MAXPHYS / size, vp->v_ralen << 1);
-               } else 
-                       vp->v_ralen = num_ra + 1;
-
-
-               if (num_ra)                             /* case 2, 4 */
-                       rbp = cluster_rbuild(vp, filesize,
-                           bp, ioblkno, blkno, size, num_ra, flags);
-               else if (lblkno != 0 && ioblkno == lblkno) {
-                       /* Case 5: check how many blocks to read ahead */
-                       ++ioblkno;
-                       if ((ioblkno + 1) * size > filesize ||
-                           (error = VOP_BMAP(vp,
-                           ioblkno, NULL, &blkno, &num_ra)))
-                               goto skip_readahead;
-                       flags |= B_ASYNC;
-                       if (num_ra)
-                               rbp = cluster_rbuild(vp, filesize,
-                                   NULL, ioblkno, blkno, size, num_ra, flags);
-                       else {
-                               rbp = getblk(vp, ioblkno, size, 0, 0);
-                               rbp->b_flags |= flags;
-                               rbp->b_blkno = blkno;
-                       }
-               } else if (lblkno != 0) {
-                       /* case 2; read ahead single block */
-                       rbp = getblk(vp, ioblkno, size, 0, 0);
-                       rbp->b_flags |= flags;
-                       rbp->b_blkno = blkno;
-               } else if (bp)                          /* case 1, 3, block 0 */
-                       bp->b_blkno = blkno;
-               /* Case 1 on block 0; not really doing sequential I/O */
-
-               if (rbp == bp)          /* case 4 */
-                       rbp = NULL;
-               else if (rbp) {                 /* case 2, 5 */
-                       trace(TR_BREADMISSRA,
-                           pack(vp, (num_ra + 1) * size), ioblkno);
-                       curproc->p_stats->p_ru.ru_inblock++;    /* XXX */
-               }
-       }
-
-       /* XXX Kirk, do we need to make sure the bp has creds? */
-skip_readahead:
-       if (bp)
-               if (bp->b_flags & (B_DONE | B_DELWRI))
-                       panic("cluster_read: DONE bp");
-               else 
-                       error = VOP_STRATEGY(bp);
-
-       if (rbp)
-               if (error || rbp->b_flags & (B_DONE | B_DELWRI)) {
-                       rbp->b_flags &= ~(B_ASYNC | B_READ);
-                       brelse(rbp);
-               } else
-                       (void) VOP_STRATEGY(rbp);
-
-       if (bp)
-               return(biowait(bp));
-       return(error);
-}
-
-/*
- * If blocks are contiguous on disk, use this to provide clustered
- * read ahead.  We will read as many blocks as possible sequentially
- * and then parcel them up into logical blocks in the buffer hash table.
- */
-struct buf *
-cluster_rbuild(vp, filesize, bp, lbn, blkno, size, run, flags)
-       struct vnode *vp;
-       u_quad_t filesize;
-       struct buf *bp;
-       daddr_t lbn;
-       daddr_t blkno;
-       long size;
-       int run;
-       long flags;
-{
-       struct cluster_save *b_save;
-       struct buf *tbp;
-       daddr_t bn;
-       int i, inc;
-
-#ifdef DIAGNOSTIC
-       if (size != vp->v_mount->mnt_stat.f_iosize)
-               panic("cluster_rbuild: size %d != filesize %d\n",
-                       size, vp->v_mount->mnt_stat.f_iosize);
-#endif
-       if (size * (lbn + run + 1) > filesize)
-               --run;
-       if (run == 0) {
-               if (!bp) {
-                       bp = getblk(vp, lbn, size, 0, 0);
-                       bp->b_blkno = blkno;
-                       bp->b_flags |= flags;
-               }
-               return(bp);
-       }
-
-       bp = cluster_newbuf(vp, bp, flags, blkno, lbn, size, run + 1);
-       if (bp->b_flags & (B_DONE | B_DELWRI))
-               return (bp);
-
-       b_save = malloc(sizeof(struct buf *) * run + sizeof(struct cluster_save),
-           M_SEGMENT, M_WAITOK);
-       b_save->bs_bufsize = b_save->bs_bcount = size;
-       b_save->bs_nchildren = 0;
-       b_save->bs_children = (struct buf **)(b_save + 1);
-       b_save->bs_saveaddr = bp->b_saveaddr;
-       bp->b_saveaddr = (caddr_t) b_save;
-
-       inc = size / DEV_BSIZE;
-       for (bn = blkno + inc, i = 1; i <= run; ++i, bn += inc) {
-               if (incore(vp, lbn + i)) {
-                       if (i == 1) {
-                               bp->b_saveaddr = b_save->bs_saveaddr;
-                               bp->b_flags &= ~B_CALL;
-                               bp->b_iodone = NULL;
-                               allocbuf(bp, size);
-                               free(b_save, M_SEGMENT);
-                       } else
-                               allocbuf(bp, size * i);
-                       break;
-               }
-               tbp = getblk(vp, lbn + i, 0, 0, 0);
-               tbp->b_bcount = tbp->b_bufsize = size;
-               tbp->b_blkno = bn;
-               {
-                       daddr_t temp;
-                       VOP_BMAP(tbp->b_vp, tbp->b_lblkno, NULL, &temp, NULL);
-                       if (temp != bn) {
-                               printf("Block: %d Assigned address: %x Bmap address: %x\n",
-                                           tbp->b_lblkno, tbp->b_blkno, temp);
-                               panic("cluster_rbuild: wrong disk address");
-                       }
-               }
-               tbp->b_flags |= flags | B_READ | B_ASYNC;
-               ++b_save->bs_nchildren;
-               b_save->bs_children[i - 1] = tbp;
-       }
-       if (!(bp->b_flags & B_ASYNC))
-               vp->v_ralen = max(vp->v_ralen - 1, 1);
-       return(bp);
-}
-
-/*
- * Either get a new buffer or grow the existing one.
- */
-struct buf *
-cluster_newbuf(vp, bp, flags, blkno, lblkno, size, run)
-       struct vnode *vp;
-       struct buf *bp;
-       long flags;
-       daddr_t blkno;
-       daddr_t lblkno;
-       long size;
-       int run;
-{
-       if (!bp) {
-               bp = getblk(vp, lblkno, size, 0, 0);
-               if (bp->b_flags & (B_DONE | B_DELWRI)) {
-                       bp->b_blkno = blkno;
-                       return(bp);
-               }
-       }
-       allocbuf(bp, run * size);
-       bp->b_blkno = blkno;
-       bp->b_iodone = cluster_callback;
-       bp->b_flags |= flags | B_CALL;
-       return(bp);
-}
-
-/*
- * Cleanup after a clustered read or write.
- */
-void
-cluster_callback(bp)
-       struct buf *bp;
-{
-       struct cluster_save *b_save;
-       struct buf **tbp;
-       long bsize;
-       caddr_t cp;
-       daddr_t daddr;
-       b_save = (struct cluster_save *)(bp->b_saveaddr);
-       bp->b_saveaddr = b_save->bs_saveaddr;
-
-       cp = bp->b_un.b_addr + b_save->bs_bufsize;
-       daddr = bp->b_blkno + b_save->bs_bufsize / DEV_BSIZE;
-       for (tbp = b_save->bs_children; b_save->bs_nchildren--; ++tbp) {
-               pagemove(cp, (*tbp)->b_un.b_addr, (*tbp)->b_bufsize);
-               cp += (*tbp)->b_bufsize;
-               bp->b_bufsize -= (*tbp)->b_bufsize;
-               if ((*tbp)->b_blkno != daddr) {
-                       struct inode *ip;
-                       printf("cluster_callback: bad disk address:\n");
-                       printf("Clustered Block: %d DiskAddr: %x bytes left: %d\n",
-                           bp->b_lblkno, bp->b_blkno, bp->b_bufsize);
-                       printf("\toriginal size: %d flags: %x\n", bp->b_bcount,
-                           bp->b_flags);
-                       printf("Child Block: %d DiskAddr: %x bytes: %d\n",
-                           (*tbp)->b_lblkno, (*tbp)->b_blkno,
-                           (*tbp)->b_bufsize);
-                       ip = VTOI((*tbp)->b_vp);
-                       printf("daddr: %x i_size %qd\n", daddr, ip->i_size);
-                       if ((*tbp)->b_lblkno < NDADDR)
-                               printf("Child block pointer from inode: %x\n",
-                                   ip->i_din.di_db[(*tbp)->b_lblkno]);
-                       spl0();
-                       panic ("cluster_callback: bad disk address");
-               }
-               daddr += (*tbp)->b_bufsize / DEV_BSIZE;
-               biodone(*tbp);
-       }
-#ifdef DIAGNOSTIC
-       if (bp->b_bufsize != b_save->bs_bufsize)
-               panic ("cluster_callback: more space to reclaim");
-#endif
-       bp->b_bcount = bp->b_bufsize;
-       bp->b_iodone = NULL;
-       free(b_save, M_SEGMENT);
-       if (bp->b_flags & B_ASYNC)
-               brelse(bp);
-       else
-               wakeup((caddr_t)bp);
-}
-
 /*
  * Synchronous write.
  * Release buffer on completion.
 /*
  * Synchronous write.
  * Release buffer on completion.
@@ -568,6 +253,8 @@ bwrite(bp)
        register int flag;
        int s, error = 0;
 
        register int flag;
        int s, error = 0;
 
+       if (bp->b_vp && (bp->b_vp->v_mount->mnt_flag & MNT_ASYNC))
+               bp->b_flags |= B_ASYNC;
        flag = bp->b_flags;
        bp->b_flags &= ~(B_READ | B_DONE | B_ERROR | B_DELWRI);
        if (flag & B_ASYNC) {
        flag = bp->b_flags;
        bp->b_flags &= ~(B_READ | B_DONE | B_ERROR | B_DELWRI);
        if (flag & B_ASYNC) {
@@ -613,6 +300,7 @@ int
 vn_bwrite(ap)
        struct vop_bwrite_args *ap;
 {
 vn_bwrite(ap)
        struct vop_bwrite_args *ap;
 {
+       
        return (bwrite(ap->a_bp));
 }
 
        return (bwrite(ap->a_bp));
 }
 
@@ -666,204 +354,6 @@ bawrite(bp)
        (void) VOP_BWRITE(bp);
 }
 
        (void) VOP_BWRITE(bp);
 }
 
-/*
- * Do clustered write for FFS.
- *
- * Three cases:
- *     1. Write is not sequential (write asynchronously)
- *     Write is sequential:
- *     2.      beginning of cluster - begin cluster
- *     3.      middle of a cluster - add to cluster
- *     4.      end of a cluster - asynchronously write cluster
- */
-void
-cluster_write(bp, filesize)
-        struct buf *bp;
-       u_quad_t filesize;
-{
-        struct vnode *vp;
-        daddr_t lbn;
-        int clen;
-
-        vp = bp->b_vp;
-        lbn = bp->b_lblkno;
-
-       /* Initialize vnode to beginning of file. */
-       if (lbn == 0)
-               vp->v_lasta = vp->v_clen = vp->v_cstart = vp->v_lastw = 0;
-
-        if (vp->v_clen == 0 || lbn != vp->v_lastw + 1 ||
-           (bp->b_blkno != vp->v_lasta + bp->b_bcount / DEV_BSIZE)) {
-               if (vp->v_clen != 0)
-                       /*
-                        * Write is not sequential.
-                        */
-                       cluster_wbuild(vp, NULL, bp->b_bcount, vp->v_cstart,
-                           vp->v_lastw - vp->v_cstart + 1, lbn);
-               /*
-                * Consider beginning a cluster.
-                */
-               if ((lbn + 1) * bp->b_bcount == filesize)
-                       /* End of file, make cluster as large as possible */
-                       clen = MAXBSIZE / vp->v_mount->mnt_stat.f_iosize - 1;
-               else if (VOP_BMAP(vp, lbn, NULL, &bp->b_blkno, &clen)) {
-                       bawrite(bp);
-                       vp->v_clen = 0;
-                       vp->v_lasta = bp->b_blkno;
-                       vp->v_cstart = lbn + 1;
-                       vp->v_lastw = lbn;
-                       return;
-               } else
-                       clen = 0;
-                vp->v_clen = clen;
-                if (clen == 0) {               /* I/O not contiguous */
-                       vp->v_cstart = lbn + 1;
-                        bawrite(bp);
-                } else {                       /* Wait for rest of cluster */
-                       vp->v_cstart = lbn;
-                        bdwrite(bp);
-               }
-        } else if (lbn == vp->v_cstart + vp->v_clen) {
-               /*
-                * At end of cluster, write it out.
-                */
-               cluster_wbuild(vp, bp, bp->b_bcount, vp->v_cstart,
-                   vp->v_clen + 1, lbn);
-               vp->v_clen = 0;
-               vp->v_cstart = lbn + 1;
-        } else
-               /*
-                * In the middle of a cluster, so just delay the
-                * I/O for now.
-                */
-                bdwrite(bp);
-        vp->v_lastw = lbn;
-       vp->v_lasta = bp->b_blkno;
-}
-
-
-/*
- * This is an awful lot like cluster_rbuild...wish they could be combined.
- * The last lbn argument is the current block on which I/O is being
- * performed.  Check to see that it doesn't fall in the middle of
- * the current block.
- */
-void
-cluster_wbuild(vp, last_bp, size, start_lbn, len, lbn)
-       struct vnode *vp;
-       struct buf *last_bp;
-       long size;
-       daddr_t start_lbn;
-       int len;
-       daddr_t lbn;
-{
-       struct cluster_save *b_save;
-       struct buf *bp, *tbp;
-       caddr_t cp;
-       int i, s;
-
-#ifdef DIAGNOSTIC
-       if (size != vp->v_mount->mnt_stat.f_iosize)
-               panic("cluster_wbuild: size %d != filesize %d\n",
-                       size, vp->v_mount->mnt_stat.f_iosize);
-#endif
-redo:
-       while ((!incore(vp, start_lbn) || start_lbn == lbn) && len) {
-               ++start_lbn;
-               --len;
-       }
-
-       /* Get more memory for current buffer */
-       if (len <= 1) {
-               if (last_bp) {
-                       bawrite(last_bp);
-               } else if (len) {
-                       bp = getblk(vp, start_lbn, size, 0, 0);
-                       bawrite(bp);
-               }
-               return;
-       }
-
-       bp = getblk(vp, start_lbn, size, 0, 0);
-       if (!(bp->b_flags & B_DELWRI)) {
-               ++start_lbn;
-               --len;
-               brelse(bp);
-               goto redo;
-       }
-
-       --len;
-       b_save = malloc(sizeof(struct buf *) * len + sizeof(struct cluster_save),
-           M_SEGMENT, M_WAITOK);
-       b_save->bs_bcount = bp->b_bcount;
-       b_save->bs_bufsize = bp->b_bufsize;
-       b_save->bs_nchildren = 0;
-       b_save->bs_children = (struct buf **)(b_save + 1);
-       b_save->bs_saveaddr = bp->b_saveaddr;
-       bp->b_saveaddr = (caddr_t) b_save;
-
-
-       bp->b_flags |= B_CALL;
-       bp->b_iodone = cluster_callback;
-       cp = bp->b_un.b_addr + bp->b_bufsize;
-       for (++start_lbn, i = 0; i < len; ++i, ++start_lbn) {
-               if (!incore(vp, start_lbn) || start_lbn == lbn)
-                       break;
-
-               if (last_bp == NULL || start_lbn != last_bp->b_lblkno) {
-                       tbp = getblk(vp, start_lbn, size, 0, 0);
-#ifdef DIAGNOSTIC
-                       if (tbp->b_bcount != tbp->b_bufsize)
-                               panic("cluster_wbuild: Buffer too big");
-#endif
-                       if (!(tbp->b_flags & B_DELWRI)) {
-                               brelse(tbp);
-                               break;
-                       }
-               } else
-                       tbp = last_bp;
-
-               ++b_save->bs_nchildren;
-
-               /* Move memory from children to parent */
-               if (tbp->b_blkno != (bp->b_blkno + bp->b_bufsize / DEV_BSIZE)) {
-                       printf("Clustered Block: %d addr %x bufsize: %d\n",
-                           bp->b_lblkno, bp->b_blkno, bp->b_bufsize);
-                       printf("Child Block: %d addr: %x\n", tbp->b_lblkno,
-                           tbp->b_blkno);
-                       panic("Clustered write to wrong blocks");
-               }
-
-               pagemove(tbp->b_un.b_daddr, cp, size);
-               bp->b_bcount += size;
-               bp->b_bufsize += size;
-
-               tbp->b_flags &= ~(B_READ | B_DONE | B_ERROR | B_DELWRI);
-               tbp->b_flags |= B_ASYNC;
-               s = splbio();
-               reassignbuf(tbp, tbp->b_vp);            /* put on clean list */
-               ++tbp->b_vp->v_numoutput;
-               splx(s);
-               b_save->bs_children[i] = tbp;
-
-               cp += tbp->b_bufsize;
-       }
-
-       if (i == 0) {
-               /* None to cluster */
-               bp->b_saveaddr = b_save->bs_saveaddr;
-               bp->b_flags &= ~B_CALL;
-               bp->b_iodone = NULL;
-               free(b_save, M_SEGMENT);
-       }
-       bawrite(bp);
-       if (i < len) {
-               len -= i + 1;
-               start_lbn += 1;
-               goto redo;
-       }
-}
-
 /*
  * Release a buffer.
  * Even if the buffer is dirty, no I/O is started.
 /*
  * Release a buffer.
  * Even if the buffer is dirty, no I/O is started.
@@ -871,7 +361,7 @@ redo:
 brelse(bp)
        register struct buf *bp;
 {
 brelse(bp)
        register struct buf *bp;
 {
-       register struct queue_entry *flist;
+       register struct bqueues *flist;
        int s;
 
        trace(TR_BRELSE, pack(bp->b_vp, bp->b_bufsize), bp->b_lblkno);
        int s;
 
        trace(TR_BRELSE, pack(bp->b_vp, bp->b_bufsize), bp->b_lblkno);
@@ -935,7 +425,7 @@ incore(vp, blkno)
 {
        register struct buf *bp;
 
 {
        register struct buf *bp;
 
-       for (bp = BUFHASH(vp, blkno)->le_next; bp; bp = bp->b_hash.qe_next)
+       for (bp = BUFHASH(vp, blkno)->lh_first; bp; bp = bp->b_hash.le_next)
                if (bp->b_lblkno == blkno && bp->b_vp == vp &&
                    (bp->b_flags & B_INVAL) == 0)
                        return (bp);
                if (bp->b_lblkno == blkno && bp->b_vp == vp &&
                    (bp->b_flags & B_INVAL) == 0)
                        return (bp);
@@ -960,7 +450,7 @@ getblk(vp, blkno, size, slpflag, slptimeo)
 #endif SECSIZE
 {
        register struct buf *bp;
 #endif SECSIZE
 {
        register struct buf *bp;
-       struct list_entry *dp;
+       struct bufhashhdr *dp;
        int s, error;
 
        if (size > MAXBSIZE)
        int s, error;
 
        if (size > MAXBSIZE)
@@ -972,7 +462,7 @@ getblk(vp, blkno, size, slpflag, slptimeo)
         */
        dp = BUFHASH(vp, blkno);
 loop:
         */
        dp = BUFHASH(vp, blkno);
 loop:
-       for (bp = dp->le_next; bp; bp = bp->b_hash.qe_next) {
+       for (bp = dp->lh_first; bp; bp = bp->b_hash.le_next) {
                if (bp->b_lblkno != blkno || bp->b_vp != vp)
                        continue;
                s = splbio();
                if (bp->b_lblkno != blkno || bp->b_vp != vp)
                        continue;
                s = splbio();
@@ -1085,13 +575,13 @@ allocbuf(tp, size)
         * extra space in the present buffer.
         */
        if (sizealloc < tp->b_bufsize) {
         * extra space in the present buffer.
         */
        if (sizealloc < tp->b_bufsize) {
-               if ((ep = bufqueues[BQ_EMPTY].qe_next) == NULL)
+               if ((ep = bufqueues[BQ_EMPTY].tqh_first) == NULL)
                        goto out;
                s = splbio();
                bremfree(ep);
                ep->b_flags |= B_BUSY;
                splx(s);
                        goto out;
                s = splbio();
                bremfree(ep);
                ep->b_flags |= B_BUSY;
                splx(s);
-               pagemove(tp->b_un.b_addr + sizealloc, ep->b_un.b_addr,
+               pagemove((char *)tp->b_data + sizealloc, ep->b_data,
                    (int)tp->b_bufsize - sizealloc);
                ep->b_bufsize = tp->b_bufsize - sizealloc;
                tp->b_bufsize = sizealloc;
                    (int)tp->b_bufsize - sizealloc);
                ep->b_bufsize = tp->b_bufsize - sizealloc;
                tp->b_bufsize = sizealloc;
@@ -1111,8 +601,8 @@ allocbuf(tp, size)
                        /* void */;
                if (take >= bp->b_bufsize)
                        take = bp->b_bufsize;
                        /* void */;
                if (take >= bp->b_bufsize)
                        take = bp->b_bufsize;
-               pagemove(&bp->b_un.b_addr[bp->b_bufsize - take],
-                   &tp->b_un.b_addr[tp->b_bufsize], take);
+               pagemove(&((char *)bp->b_data)[bp->b_bufsize - take],
+                   &((char *)tp->b_data)[tp->b_bufsize], take);
                tp->b_bufsize += take;
                bp->b_bufsize = bp->b_bufsize - take;
                if (bp->b_bcount > bp->b_bufsize)
                tp->b_bufsize += take;
                bp->b_bufsize = bp->b_bufsize - take;
                if (bp->b_bcount > bp->b_bufsize)
@@ -1141,7 +631,7 @@ getnewbuf(slpflag, slptimeo)
        int slpflag, slptimeo;
 {
        register struct buf *bp;
        int slpflag, slptimeo;
 {
        register struct buf *bp;
-       register struct queue_entry *dp;
+       register struct bqueues *dp;
        register struct ucred *cred;
        int s;
        struct buf *abp;
        register struct ucred *cred;
        int s;
        struct buf *abp;
@@ -1262,8 +752,8 @@ count_lock_queue()
        register struct buf *bp;
        register int ret;
 
        register struct buf *bp;
        register int ret;
 
-       for (ret = 0, bp = (struct buf *)bufqueues[BQ_LOCKED].qe_next;
-           bp; bp = (struct buf *)bp->b_freelist.qe_next)
+       for (ret = 0, bp = (struct buf *)bufqueues[BQ_LOCKED].tqh_first;
+           bp; bp = (struct buf *)bp->b_freelist.tqe_next)
                ++ret;
        return(ret);
 }
                ++ret;
        return(ret);
 }
@@ -1272,14 +762,14 @@ count_lock_queue()
 /*
  * Print out statistics on the current allocation of the buffer pool.
  * Can be enabled to print out on every ``sync'' by setting "syncprt"
 /*
  * Print out statistics on the current allocation of the buffer pool.
  * Can be enabled to print out on every ``sync'' by setting "syncprt"
- * above.
+ * in vfs_syscalls.c using sysctl.
  */
 void
 vfs_bufstats()
 {
        int s, i, j, count;
        register struct buf *bp;
  */
 void
 vfs_bufstats()
 {
        int s, i, j, count;
        register struct buf *bp;
-       register struct queue_entry *dp;
+       register struct bqueues *dp;
        int counts[MAXBSIZE/CLBYTES+1];
        static char *bname[BQUEUES] = { "LOCKED", "LRU", "AGE", "EMPTY" };
 
        int counts[MAXBSIZE/CLBYTES+1];
        static char *bname[BQUEUES] = { "LOCKED", "LRU", "AGE", "EMPTY" };
 
@@ -1288,7 +778,7 @@ vfs_bufstats()
                for (j = 0; j <= MAXBSIZE/CLBYTES; j++)
                        counts[j] = 0;
                s = splbio();
                for (j = 0; j <= MAXBSIZE/CLBYTES; j++)
                        counts[j] = 0;
                s = splbio();
-               for (bp = dp->qe_next; bp; bp = bp->b_freelist.qe_next) {
+               for (bp = dp->tqh_first; bp; bp = bp->b_freelist.tqe_next) {
                        counts[bp->b_bufsize/CLBYTES]++;
                        count++;
                }
                        counts[bp->b_bufsize/CLBYTES]++;
                        count++;
                }