summaryrefslogtreecommitdiffstats
path: root/sys
diff options
context:
space:
mode:
authorphk <phk@FreeBSD.org>2005-02-08 20:29:10 +0000
committerphk <phk@FreeBSD.org>2005-02-08 20:29:10 +0000
commitaf5ef3f262889299ffe39d2c97f6ef14b8a66b70 (patch)
tree9fcf46127f5e7e982c2869763b1e03ed0cc2c784 /sys
parentad1ee10f6d12f8bbdd826c263511342a4a45857a (diff)
downloadFreeBSD-src-af5ef3f262889299ffe39d2c97f6ef14b8a66b70.zip
FreeBSD-src-af5ef3f262889299ffe39d2c97f6ef14b8a66b70.tar.gz
Background writes are entirely an FFS/Softupdates thing.
Give FFS vnodes a specific bufwrite method which contains all the background write stuff and then calls into the default bufwrite() for the rest of the job. Remove all the background write related stuff from the normal bufwrite. This drags the softdep_move_dependencies() back into FFS. Long term, it is worth looking at simply copying the data into allocated memory and issuing the bio directly and not create the "shadow buf" in the first place (just like copy-on-write is done in snapshots for instance). I don't think we really gain anything but complexity from doing this with a buf.
Diffstat (limited to 'sys')
-rw-r--r--sys/kern/vfs_bio.c131
-rw-r--r--sys/sys/buf.h8
-rw-r--r--sys/ufs/ffs/ffs_extern.h1
-rw-r--r--sys/ufs/ffs/ffs_softdep.c4
-rw-r--r--sys/ufs/ffs/ffs_vfsops.c173
5 files changed, 179 insertions, 138 deletions
diff --git a/sys/kern/vfs_bio.c b/sys/kern/vfs_bio.c
index d36ceae..e53d10d 100644
--- a/sys/kern/vfs_bio.c
+++ b/sys/kern/vfs_bio.c
@@ -100,7 +100,6 @@ static void vfs_page_set_valid(struct buf *bp, vm_ooffset_t off,
static void vfs_clean_pages(struct buf *bp);
static void vfs_setdirty(struct buf *bp);
static void vfs_vmio_release(struct buf *bp);
-static void vfs_backgroundwritedone(struct buf *bp);
static int vfs_bio_clcheck(struct vnode *vp, int size,
daddr_t lblkno, daddr_t blkno);
static int flushbufqueues(int flushdeps);
@@ -182,9 +181,6 @@ SYSCTL_INT(_vfs, OID_AUTO, getnewbufcalls, CTLFLAG_RW, &getnewbufcalls, 0,
static int getnewbufrestarts;
SYSCTL_INT(_vfs, OID_AUTO, getnewbufrestarts, CTLFLAG_RW, &getnewbufrestarts, 0,
"Number of times getnewbuf has had to restart a buffer aquisition");
-static int dobkgrdwrite = 1;
-SYSCTL_INT(_debug, OID_AUTO, dobkgrdwrite, CTLFLAG_RW, &dobkgrdwrite, 0,
- "Do background writes (honoring the BV_BKGRDWRITE flag)?");
/*
* Wakeup point for bufdaemon, as well as indicator of whether it is already
@@ -803,7 +799,6 @@ int
bufwrite(struct buf *bp)
{
int oldflags, s;
- struct buf *newbp;
CTR3(KTR_BUF, "bufwrite(%p) vp %p flags %X", bp, bp->b_vp, bp->b_flags);
if (bp->b_flags & B_INVAL) {
@@ -816,81 +811,15 @@ bufwrite(struct buf *bp)
if (BUF_REFCNT(bp) == 0)
panic("bufwrite: buffer is not busy???");
s = splbio();
- /*
- * If a background write is already in progress, delay
- * writing this block if it is asynchronous. Otherwise
- * wait for the background write to complete.
- */
- BO_LOCK(bp->b_bufobj);
- if (bp->b_vflags & BV_BKGRDINPROG) {
- if (bp->b_flags & B_ASYNC) {
- BO_UNLOCK(bp->b_bufobj);
- splx(s);
- bdwrite(bp);
- return (0);
- }
- bp->b_vflags |= BV_BKGRDWAIT;
- msleep(&bp->b_xflags, BO_MTX(bp->b_bufobj), PRIBIO, "bwrbg", 0);
- if (bp->b_vflags & BV_BKGRDINPROG)
- panic("bufwrite: still writing");
- }
- BO_UNLOCK(bp->b_bufobj);
+
+ KASSERT(!(bp->b_vflags & BV_BKGRDINPROG),
+ ("FFS background buffer should not get here %p", bp));
/* Mark the buffer clean */
bundirty(bp);
- /*
- * If this buffer is marked for background writing and we
- * do not have to wait for it, make a copy and write the
- * copy so as to leave this buffer ready for further use.
- *
- * This optimization eats a lot of memory. If we have a page
- * or buffer shortfall we can't do it.
- */
- if (dobkgrdwrite && (bp->b_xflags & BX_BKGRDWRITE) &&
- (bp->b_flags & B_ASYNC) &&
- !vm_page_count_severe() &&
- !buf_dirty_count_severe()) {
- KASSERT(bp->b_iodone == NULL,
- ("bufwrite: needs chained iodone (%p)", bp->b_iodone));
-
- /* get a new block */
- newbp = geteblk(bp->b_bufsize);
-
- /*
- * set it to be identical to the old block. We have to
- * set b_lblkno and BKGRDMARKER before calling bgetvp()
- * to avoid confusing the splay tree and gbincore().
- */
- memcpy(newbp->b_data, bp->b_data, bp->b_bufsize);
- newbp->b_lblkno = bp->b_lblkno;
- newbp->b_xflags |= BX_BKGRDMARKER;
- BO_LOCK(bp->b_bufobj);
- bp->b_vflags |= BV_BKGRDINPROG;
- bgetvp(bp->b_vp, newbp);
- BO_UNLOCK(bp->b_bufobj);
- newbp->b_bufobj = &bp->b_vp->v_bufobj;
- newbp->b_blkno = bp->b_blkno;
- newbp->b_offset = bp->b_offset;
- newbp->b_iodone = vfs_backgroundwritedone;
- newbp->b_flags |= B_ASYNC;
- newbp->b_flags &= ~B_INVAL;
-
- /* move over the dependencies */
- if (LIST_FIRST(&bp->b_dep) != NULL)
- buf_movedeps(bp, newbp);
-
- /*
- * Initiate write on the copy, release the original to
- * the B_LOCKED queue so that it cannot go away until
- * the background write completes. If not locked it could go
- * away and then be reconstituted while it was being written.
- * If the reconstituted buffer were written, we could end up
- * with two background copies being written at the same time.
- */
- bqrelse(bp);
- bp = newbp;
- }
+ KASSERT(!(bp->b_xflags & BX_BKGRDWRITE),
+ ("FFS background buffer should not get here %p", bp));
bp->b_flags &= ~B_DONE;
bp->b_ioflags &= ~BIO_ERROR;
@@ -936,56 +865,6 @@ bufwrite(struct buf *bp)
}
/*
- * Complete a background write started from bwrite.
- */
-static void
-vfs_backgroundwritedone(struct buf *bp)
-{
- struct buf *origbp;
-
- /*
- * Find the original buffer that we are writing.
- */
- BO_LOCK(bp->b_bufobj);
- if ((origbp = gbincore(bp->b_bufobj, bp->b_lblkno)) == NULL)
- panic("backgroundwritedone: lost buffer");
- BO_UNLOCK(bp->b_bufobj);
- /*
- * Process dependencies then return any unfinished ones.
- */
- if (LIST_FIRST(&bp->b_dep) != NULL)
- buf_complete(bp);
- if (LIST_FIRST(&bp->b_dep) != NULL)
- buf_movedeps(bp, origbp);
-
- /*
- * This buffer is marked B_NOCACHE, so when it is released
- * by biodone, it will be tossed. We mark it with BIO_READ
- * to avoid biodone doing a second bufobj_wdrop.
- */
- bp->b_flags |= B_NOCACHE;
- bp->b_iocmd = BIO_READ;
- bp->b_flags &= ~(B_CACHE | B_DONE);
- bp->b_iodone = 0;
- bufdone(bp);
- BO_LOCK(origbp->b_bufobj);
- /*
- * Clear the BV_BKGRDINPROG flag in the original buffer
- * and awaken it if it is waiting for the write to complete.
- * If BV_BKGRDINPROG is not set in the original buffer it must
- * have been released and re-instantiated - which is not legal.
- */
- KASSERT((origbp->b_vflags & BV_BKGRDINPROG),
- ("backgroundwritedone: lost buffer2"));
- origbp->b_vflags &= ~BV_BKGRDINPROG;
- if (origbp->b_vflags & BV_BKGRDWAIT) {
- origbp->b_vflags &= ~BV_BKGRDWAIT;
- wakeup(&origbp->b_xflags);
- }
- BO_UNLOCK(origbp->b_bufobj);
-}
-
-/*
* Delayed write. (Buffer is marked dirty). Do not bother writing
* anything if the buffer is marked invalid.
*
diff --git a/sys/sys/buf.h b/sys/sys/buf.h
index 25a9133..f80316c 100644
--- a/sys/sys/buf.h
+++ b/sys/sys/buf.h
@@ -64,7 +64,6 @@ extern struct bio_ops {
void (*io_start)(struct buf *);
void (*io_complete)(struct buf *);
void (*io_deallocate)(struct buf *);
- void (*io_movedeps)(struct buf *, struct buf *);
int (*io_countdeps)(struct buf *, int);
} bioops;
@@ -436,13 +435,6 @@ buf_deallocate(struct buf *bp)
BUF_LOCKFREE(bp);
}
-static __inline void
-buf_movedeps(struct buf *bp, struct buf *bp2)
-{
- if (bioops.io_movedeps)
- (*bioops.io_movedeps)(bp, bp2);
-}
-
static __inline int
buf_countdeps(struct buf *bp, int i)
{
diff --git a/sys/ufs/ffs/ffs_extern.h b/sys/ufs/ffs/ffs_extern.h
index b25f10e..63fc3dd 100644
--- a/sys/ufs/ffs/ffs_extern.h
+++ b/sys/ufs/ffs/ffs_extern.h
@@ -101,6 +101,7 @@ void softdep_initialize(void);
void softdep_uninitialize(void);
int softdep_mount(struct vnode *, struct mount *, struct fs *,
struct ucred *);
+void softdep_move_dependencies(struct buf *, struct buf *);
int softdep_flushworklist(struct mount *, int *, struct thread *);
int softdep_flushfiles(struct mount *, int, struct thread *);
void softdep_update_inodeblock(struct inode *, struct buf *, int);
diff --git a/sys/ufs/ffs/ffs_softdep.c b/sys/ufs/ffs/ffs_softdep.c
index 55c3035..f90a497 100644
--- a/sys/ufs/ffs/ffs_softdep.c
+++ b/sys/ufs/ffs/ffs_softdep.c
@@ -222,7 +222,6 @@ static void add_to_worklist(struct worklist *);
static void softdep_disk_io_initiation(struct buf *);
static void softdep_disk_write_complete(struct buf *);
static void softdep_deallocate_dependencies(struct buf *);
-static void softdep_move_dependencies(struct buf *, struct buf *);
static int softdep_count_dependencies(struct buf *bp, int);
static struct mtx lk;
@@ -597,7 +596,7 @@ process_worklist_item(matchmnt, flags)
/*
* Move dependencies from one buffer to another.
*/
-static void
+void
softdep_move_dependencies(oldbp, newbp)
struct buf *oldbp;
struct buf *newbp;
@@ -992,7 +991,6 @@ softdep_initialize()
bioops.io_start = softdep_disk_io_initiation;
bioops.io_complete = softdep_disk_write_complete;
bioops.io_deallocate = softdep_deallocate_dependencies;
- bioops.io_movedeps = softdep_move_dependencies;
bioops.io_countdeps = softdep_count_dependencies;
}
diff --git a/sys/ufs/ffs/ffs_vfsops.c b/sys/ufs/ffs/ffs_vfsops.c
index aae9207..6434f62 100644
--- a/sys/ufs/ffs/ffs_vfsops.c
+++ b/sys/ufs/ffs/ffs_vfsops.c
@@ -102,10 +102,11 @@ static struct vfsops ufs_vfsops = {
VFS_SET(ufs_vfsops, ufs, 0);
static b_strategy_t ffs_geom_strategy;
+static b_write_t ffs_bufwrite;
static struct buf_ops ffs_ops = {
.bop_name = "FFS",
- .bop_write = bufwrite,
+ .bop_write = ffs_bufwrite,
.bop_strategy = ffs_geom_strategy,
.bop_sync = bufsync,
};
@@ -1505,6 +1506,176 @@ ffs_ifree(struct ufsmount *ump, struct inode *ip)
uma_zfree(uma_inode, ip);
}
+static int dobkgrdwrite = 1;
+SYSCTL_INT(_debug, OID_AUTO, dobkgrdwrite, CTLFLAG_RW, &dobkgrdwrite, 0,
+ "Do background writes (honoring the BV_BKGRDWRITE flag)?");
+
+/*
+ * Complete a background write started from bwrite.
+ */
+static void
+ffs_backgroundwritedone(struct buf *bp)
+{
+ struct buf *origbp;
+
+ /*
+ * Find the original buffer that we are writing.
+ */
+ BO_LOCK(bp->b_bufobj);
+ if ((origbp = gbincore(bp->b_bufobj, bp->b_lblkno)) == NULL)
+ panic("backgroundwritedone: lost buffer");
+ BO_UNLOCK(bp->b_bufobj);
+ /*
+ * Process dependencies then return any unfinished ones.
+ */
+ if (LIST_FIRST(&bp->b_dep) != NULL)
+ buf_complete(bp);
+#ifdef SOFTUPDATES
+ if (LIST_FIRST(&bp->b_dep) != NULL)
+ softdep_move_dependencies(bp, origbp);
+#endif
+
+ /*
+ * This buffer is marked B_NOCACHE, so when it is released
+ * by biodone, it will be tossed. We mark it with BIO_READ
+ * to avoid biodone doing a second bufobj_wdrop.
+ */
+ bp->b_flags |= B_NOCACHE;
+ bp->b_iocmd = BIO_READ;
+ bp->b_flags &= ~(B_CACHE | B_DONE);
+ bp->b_iodone = 0;
+ bufdone(bp);
+ BO_LOCK(origbp->b_bufobj);
+ /*
+ * Clear the BV_BKGRDINPROG flag in the original buffer
+ * and awaken it if it is waiting for the write to complete.
+ * If BV_BKGRDINPROG is not set in the original buffer it must
+ * have been released and re-instantiated - which is not legal.
+ */
+ KASSERT((origbp->b_vflags & BV_BKGRDINPROG),
+ ("backgroundwritedone: lost buffer2"));
+ origbp->b_vflags &= ~BV_BKGRDINPROG;
+ if (origbp->b_vflags & BV_BKGRDWAIT) {
+ origbp->b_vflags &= ~BV_BKGRDWAIT;
+ wakeup(&origbp->b_xflags);
+ }
+ BO_UNLOCK(origbp->b_bufobj);
+}
+
+
+/*
+ * Write, release buffer on completion. (Done by iodone
+ * if async). Do not bother writing anything if the buffer
+ * is invalid.
+ *
+ * Note that we set B_CACHE here, indicating that buffer is
+ * fully valid and thus cacheable. This is true even of NFS
+ * now so we set it generally. This could be set either here
+ * or in biodone() since the I/O is synchronous. We put it
+ * here.
+ */
+static int
+ffs_bufwrite(struct buf *bp)
+{
+ int oldflags, s;
+ struct buf *newbp;
+
+ CTR3(KTR_BUF, "bufwrite(%p) vp %p flags %X", bp, bp->b_vp, bp->b_flags);
+ if (bp->b_flags & B_INVAL) {
+ brelse(bp);
+ return (0);
+ }
+
+ oldflags = bp->b_flags;
+
+ if (BUF_REFCNT(bp) == 0)
+ panic("bufwrite: buffer is not busy???");
+ s = splbio();
+ /*
+ * If a background write is already in progress, delay
+ * writing this block if it is asynchronous. Otherwise
+ * wait for the background write to complete.
+ */
+ BO_LOCK(bp->b_bufobj);
+ if (bp->b_vflags & BV_BKGRDINPROG) {
+ if (bp->b_flags & B_ASYNC) {
+ BO_UNLOCK(bp->b_bufobj);
+ splx(s);
+ bdwrite(bp);
+ return (0);
+ }
+ bp->b_vflags |= BV_BKGRDWAIT;
+ msleep(&bp->b_xflags, BO_MTX(bp->b_bufobj), PRIBIO, "bwrbg", 0);
+ if (bp->b_vflags & BV_BKGRDINPROG)
+ panic("bufwrite: still writing");
+ }
+ BO_UNLOCK(bp->b_bufobj);
+
+ /* Mark the buffer clean */
+ bundirty(bp);
+
+ /*
+ * If this buffer is marked for background writing and we
+ * do not have to wait for it, make a copy and write the
+ * copy so as to leave this buffer ready for further use.
+ *
+ * This optimization eats a lot of memory. If we have a page
+ * or buffer shortfall we can't do it.
+ */
+ if (dobkgrdwrite && (bp->b_xflags & BX_BKGRDWRITE) &&
+ (bp->b_flags & B_ASYNC) &&
+ !vm_page_count_severe() &&
+ !buf_dirty_count_severe()) {
+ KASSERT(bp->b_iodone == NULL,
+ ("bufwrite: needs chained iodone (%p)", bp->b_iodone));
+
+ /* get a new block */
+ newbp = geteblk(bp->b_bufsize);
+
+ /*
+ * set it to be identical to the old block. We have to
+ * set b_lblkno and BKGRDMARKER before calling bgetvp()
+ * to avoid confusing the splay tree and gbincore().
+ */
+ memcpy(newbp->b_data, bp->b_data, bp->b_bufsize);
+ newbp->b_lblkno = bp->b_lblkno;
+ newbp->b_xflags |= BX_BKGRDMARKER;
+ BO_LOCK(bp->b_bufobj);
+ bp->b_vflags |= BV_BKGRDINPROG;
+ bgetvp(bp->b_vp, newbp);
+ BO_UNLOCK(bp->b_bufobj);
+ newbp->b_bufobj = &bp->b_vp->v_bufobj;
+ newbp->b_blkno = bp->b_blkno;
+ newbp->b_offset = bp->b_offset;
+ newbp->b_iodone = ffs_backgroundwritedone;
+ newbp->b_flags |= B_ASYNC;
+ newbp->b_flags &= ~B_INVAL;
+
+#ifdef SOFTUPDATES
+ /* move over the dependencies */
+ if (LIST_FIRST(&bp->b_dep) != NULL)
+ softdep_move_dependencies(bp, newbp);
+#endif
+
+ /*
+ * Initiate write on the copy, release the original to
+ * the B_LOCKED queue so that it cannot go away until
+ * the background write completes. If not locked it could go
+ * away and then be reconstituted while it was being written.
+ * If the reconstituted buffer were written, we could end up
+ * with two background copies being written at the same time.
+ */
+ bqrelse(bp);
+ bp = newbp;
+ }
+
+ /* Let the normal bufwrite do the rest for us */
+ bufwrite(bp);
+
+ return (0);
+}
+
+
static void
ffs_geom_strategy(struct bufobj *bo, struct buf *bp)
{
OpenPOWER on IntegriCloud