summaryrefslogtreecommitdiffstats
diff options
context:
space:
mode:
-rw-r--r--fs/xfs/linux-2.6/xfs_sync.c75
1 files changed, 18 insertions, 57 deletions
diff --git a/fs/xfs/linux-2.6/xfs_sync.c b/fs/xfs/linux-2.6/xfs_sync.c
index ee1648b179f7..fb5cca3df840 100644
--- a/fs/xfs/linux-2.6/xfs_sync.c
+++ b/fs/xfs/linux-2.6/xfs_sync.c
@@ -63,25 +63,16 @@ xfs_sync_inodes_ag(
int error = 0;
int last_error = 0;
int fflag = XFS_B_ASYNC;
- int lock_flags = XFS_ILOCK_SHARED;
if (flags & SYNC_DELWRI)
fflag = XFS_B_DELWRI;
if (flags & SYNC_WAIT)
fflag = 0; /* synchronous overrides all */
- if (flags & SYNC_DELWRI) {
- /*
- * We need the I/O lock if we're going to call any of
- * the flush/inval routines.
- */
- lock_flags |= XFS_IOLOCK_SHARED;
- }
-
do {
struct inode *inode;
- boolean_t inode_refed;
xfs_inode_t *ip = NULL;
+ int lock_flags = XFS_ILOCK_SHARED;
/*
* use a gang lookup to find the next inode in the tree
@@ -109,22 +100,6 @@ xfs_sync_inodes_ag(
break;
}
- /*
- * skip inodes in reclaim. Let xfs_syncsub do that for
- * us so we don't need to worry.
- */
- if (xfs_iflags_test(ip, (XFS_IRECLAIM|XFS_IRECLAIMABLE))) {
- read_unlock(&pag->pag_ici_lock);
- continue;
- }
-
- /* bad inodes are dealt with elsewhere */
- inode = VFS_I(ip);
- if (is_bad_inode(inode)) {
- read_unlock(&pag->pag_ici_lock);
- continue;
- }
-
/* nothing to sync during shutdown */
if (XFS_FORCED_SHUTDOWN(mp)) {
read_unlock(&pag->pag_ici_lock);
@@ -132,42 +107,34 @@ xfs_sync_inodes_ag(
}
/*
- * If we can't get a reference on the VFS_I, the inode must be
- * in reclaim. If we can get the inode lock without blocking,
- * it is safe to flush the inode because we hold the tree lock
- * and xfs_iextract will block right now. Hence if we lock the
- * inode while holding the tree lock, xfs_ireclaim() is
- * guaranteed to block on the inode lock we now hold and hence
- * it is safe to reference the inode until we drop the inode
- * locks completely.
+ * If we can't get a reference on the inode, it must be
+ * in reclaim. Leave it for the reclaim code to flush.
*/
- inode_refed = B_FALSE;
- if (igrab(inode)) {
- read_unlock(&pag->pag_ici_lock);
- xfs_ilock(ip, lock_flags);
- inode_refed = B_TRUE;
- } else {
- if (!xfs_ilock_nowait(ip, lock_flags)) {
- /* leave it to reclaim */
- read_unlock(&pag->pag_ici_lock);
- continue;
- }
+ inode = VFS_I(ip);
+ if (!igrab(inode)) {
read_unlock(&pag->pag_ici_lock);
+ continue;
+ }
+ read_unlock(&pag->pag_ici_lock);
+
+ /* bad inodes are dealt with elsewhere */
+ if (is_bad_inode(inode)) {
+ IRELE(ip);
+ continue;
}
/*
* If we have to flush data or wait for I/O completion
- * we need to drop the ilock that we currently hold.
- * If we need to drop the lock, insert a marker if we
- * have not already done so.
+ * we need to hold the iolock.
*/
if ((flags & SYNC_DELWRI) && VN_DIRTY(inode)) {
- xfs_iunlock(ip, XFS_ILOCK_SHARED);
+ xfs_ilock(ip, XFS_IOLOCK_SHARED);
+ lock_flags |= XFS_IOLOCK_SHARED;
error = xfs_flush_pages(ip, 0, -1, fflag, FI_NONE);
if (flags & SYNC_IOWAIT)
vn_iowait(ip);
- xfs_ilock(ip, XFS_ILOCK_SHARED);
}
+ xfs_ilock(ip, XFS_ILOCK_SHARED);
if ((flags & SYNC_ATTR) && !xfs_inode_clean(ip)) {
if (flags & SYNC_WAIT) {
@@ -183,13 +150,7 @@ xfs_sync_inodes_ag(
xfs_ifunlock(ip);
}
}
-
- if (lock_flags)
- xfs_iunlock(ip, lock_flags);
-
- if (inode_refed) {
- IRELE(ip);
- }
+ xfs_iput(ip, lock_flags);
if (error)
last_error = error;