svn commit: r254668 - in head/sys: kern sys
Konstantin Belousov
kib at FreeBSD.org
Thu Aug 22 18:26:46 UTC 2013
Author: kib
Date: Thu Aug 22 18:26:45 2013
New Revision: 254668
URL: http://svnweb.freebsd.org/changeset/base/254668
Log:
Both cluster_rbuild() and cluster_wbuild() sometimes set the pages
shared busy without first draining the hard busy state. Previously it
went unnoticed since VPO_BUSY and m->busy fields were distinct, and
vm_page_io_start() did not verified that the passed page has VPO_BUSY
flag cleared, but such page state is wrong. New implementation is
more strict and catched this case.
Drain the busy state as needed, before calling vm_page_sbusy().
Tested by: pho, jkim
Sponsored by: The FreeBSD Foundation
Modified:
head/sys/kern/vfs_bio.c
head/sys/kern/vfs_cluster.c
head/sys/sys/buf.h
Modified: head/sys/kern/vfs_bio.c
==============================================================================
--- head/sys/kern/vfs_bio.c Thu Aug 22 18:12:24 2013 (r254667)
+++ head/sys/kern/vfs_bio.c Thu Aug 22 18:26:45 2013 (r254668)
@@ -108,7 +108,6 @@ static void vm_hold_load_pages(struct bu
static void vfs_page_set_valid(struct buf *bp, vm_ooffset_t off, vm_page_t m);
static void vfs_page_set_validclean(struct buf *bp, vm_ooffset_t off,
vm_page_t m);
-static void vfs_drain_busy_pages(struct buf *bp);
static void vfs_clean_pages_dirty_buf(struct buf *bp);
static void vfs_setdirty_locked_object(struct buf *bp);
static void vfs_vmio_release(struct buf *bp);
@@ -3983,7 +3982,7 @@ vfs_page_set_validclean(struct buf *bp,
* Ensure that all buffer pages are not exclusive busied. If any page is
* exclusive busy, drain it.
*/
-static void
+void
vfs_drain_busy_pages(struct buf *bp)
{
vm_page_t m;
Modified: head/sys/kern/vfs_cluster.c
==============================================================================
--- head/sys/kern/vfs_cluster.c Thu Aug 22 18:12:24 2013 (r254667)
+++ head/sys/kern/vfs_cluster.c Thu Aug 22 18:26:45 2013 (r254668)
@@ -315,7 +315,7 @@ cluster_rbuild(struct vnode *vp, u_quad_
daddr_t bn;
off_t off;
long tinc, tsize;
- int i, inc, j, toff;
+ int i, inc, j, k, toff;
KASSERT(size == vp->v_mount->mnt_stat.f_iosize,
("cluster_rbuild: size %ld != f_iosize %jd\n",
@@ -378,7 +378,15 @@ cluster_rbuild(struct vnode *vp, u_quad_
inc = btodb(size);
bo = &vp->v_bufobj;
for (bn = blkno, i = 0; i < run; ++i, bn += inc) {
- if (i != 0) {
+ if (i == 0) {
+ VM_OBJECT_WLOCK(tbp->b_bufobj->bo_object);
+ vfs_drain_busy_pages(tbp);
+ vm_object_pip_add(tbp->b_bufobj->bo_object,
+ tbp->b_npages);
+ for (k = 0; k < tbp->b_npages; k++)
+ vm_page_sbusy(tbp->b_pages[k]);
+ VM_OBJECT_WUNLOCK(tbp->b_bufobj->bo_object);
+ } else {
if ((bp->b_npages * PAGE_SIZE) +
round_page(size) > vp->v_mount->mnt_iosize_max) {
break;
@@ -424,14 +432,23 @@ cluster_rbuild(struct vnode *vp, u_quad_
if ((tbp->b_pages[j]->valid &
vm_page_bits(toff, tinc)) != 0)
break;
+ if (vm_page_xbusied(tbp->b_pages[j]))
+ break;
+ vm_object_pip_add(tbp->b_bufobj->bo_object, 1);
+ vm_page_sbusy(tbp->b_pages[j]);
off += tinc;
tsize -= tinc;
}
- VM_OBJECT_WUNLOCK(tbp->b_bufobj->bo_object);
if (tsize > 0) {
+clean_sbusy:
+ vm_object_pip_add(tbp->b_bufobj->bo_object, -j);
+ for (k = 0; k < j; k++)
+ vm_page_sunbusy(tbp->b_pages[k]);
+ VM_OBJECT_WUNLOCK(tbp->b_bufobj->bo_object);
bqrelse(tbp);
break;
}
+ VM_OBJECT_WUNLOCK(tbp->b_bufobj->bo_object);
/*
* Set a read-ahead mark as appropriate
@@ -451,8 +468,8 @@ cluster_rbuild(struct vnode *vp, u_quad_
if (tbp->b_blkno == tbp->b_lblkno) {
tbp->b_blkno = bn;
} else if (tbp->b_blkno != bn) {
- brelse(tbp);
- break;
+ VM_OBJECT_WLOCK(tbp->b_bufobj->bo_object);
+ goto clean_sbusy;
}
}
/*
@@ -466,10 +483,8 @@ cluster_rbuild(struct vnode *vp, u_quad_
for (j = 0; j < tbp->b_npages; j += 1) {
vm_page_t m;
m = tbp->b_pages[j];
- vm_page_sbusy(m);
- vm_object_pip_add(m->object, 1);
if ((bp->b_npages == 0) ||
- (bp->b_pages[bp->b_npages-1] != m)) {
+ (bp->b_pages[bp->b_npages-1] != m)) {
bp->b_pages[bp->b_npages] = m;
bp->b_npages++;
}
@@ -944,7 +959,9 @@ cluster_wbuild(struct vnode *vp, long si
vm_page_t m;
VM_OBJECT_WLOCK(tbp->b_bufobj->bo_object);
- if (i != 0) { /* if not first buffer */
+ if (i == 0) {
+ vfs_drain_busy_pages(tbp);
+ } else { /* if not first buffer */
for (j = 0; j < tbp->b_npages; j += 1) {
m = tbp->b_pages[j];
if (vm_page_xbusied(m)) {
Modified: head/sys/sys/buf.h
==============================================================================
--- head/sys/sys/buf.h Thu Aug 22 18:12:24 2013 (r254667)
+++ head/sys/sys/buf.h Thu Aug 22 18:26:45 2013 (r254668)
@@ -501,6 +501,7 @@ void bufstrategy(struct bufobj *, struct
void brelse(struct buf *);
void bqrelse(struct buf *);
int vfs_bio_awrite(struct buf *);
+void vfs_drain_busy_pages(struct buf *bp);
struct buf * getpbuf(int *);
struct buf *incore(struct bufobj *, daddr_t);
struct buf *gbincore(struct bufobj *, daddr_t);
More information about the svn-src-head
mailing list