Message ID | 20210726060716.3295008-4-david@fromorbit.com (mailing list archive) |
---|---|
State | Superseded |
Headers | show |
Series | xfs: fix log cache flush regressions and bugs | expand |
On Mon, Jul 26, 2021 at 04:07:09PM +1000, Dave Chinner wrote: > From: Dave Chinner <dchinner@redhat.com> > > Fold __xlog_state_release_iclog intos it's only caller to prepare s/it's/its/ > make an upcomding fix easier. s/upcomding/upcoming/ With those fixed, Reviewed-by: Darrick J. Wong <djwong@kernel.org> --D > > Signed-off-by: Dave Chinner <dchinner@redhat.com> > [hch: split from a larger patch] > Signed-off-by: Christoph Hellwig <hch@lst.de> > --- > fs/xfs/xfs_log.c | 45 +++++++++++++++++---------------------------- > 1 file changed, 17 insertions(+), 28 deletions(-) > > diff --git a/fs/xfs/xfs_log.c b/fs/xfs/xfs_log.c > index a3c4d48195d9..82f5996d3889 100644 > --- a/fs/xfs/xfs_log.c > +++ b/fs/xfs/xfs_log.c > @@ -487,29 +487,6 @@ xfs_log_reserve( > return error; > } > > -static bool > -__xlog_state_release_iclog( > - struct xlog *log, > - struct xlog_in_core *iclog) > -{ > - lockdep_assert_held(&log->l_icloglock); > - > - if (iclog->ic_state == XLOG_STATE_WANT_SYNC) { > - /* update tail before writing to iclog */ > - xfs_lsn_t tail_lsn = xlog_assign_tail_lsn(log->l_mp); > - > - iclog->ic_state = XLOG_STATE_SYNCING; > - iclog->ic_header.h_tail_lsn = cpu_to_be64(tail_lsn); > - xlog_verify_tail_lsn(log, iclog, tail_lsn); > - /* cycle incremented when incrementing curr_block */ > - trace_xlog_iclog_syncing(iclog, _RET_IP_); > - return true; > - } > - > - ASSERT(iclog->ic_state == XLOG_STATE_ACTIVE); > - return false; > -} > - > /* > * Flush iclog to disk if this is the last reference to the given iclog and the > * it is in the WANT_SYNC state. > @@ -519,19 +496,31 @@ xlog_state_release_iclog( > struct xlog *log, > struct xlog_in_core *iclog) > { > + xfs_lsn_t tail_lsn; > lockdep_assert_held(&log->l_icloglock); > > trace_xlog_iclog_release(iclog, _RET_IP_); > if (iclog->ic_state == XLOG_STATE_IOERROR) > return -EIO; > > - if (atomic_dec_and_test(&iclog->ic_refcnt) && > - __xlog_state_release_iclog(log, iclog)) { > - spin_unlock(&log->l_icloglock); > - xlog_sync(log, iclog); > - spin_lock(&log->l_icloglock); > + if (!atomic_dec_and_test(&iclog->ic_refcnt)) > + return 0; > + > + if (iclog->ic_state != XLOG_STATE_WANT_SYNC) { > + ASSERT(iclog->ic_state == XLOG_STATE_ACTIVE); > + return 0; > } > > + /* update tail before writing to iclog */ > + tail_lsn = xlog_assign_tail_lsn(log->l_mp); > + iclog->ic_state = XLOG_STATE_SYNCING; > + iclog->ic_header.h_tail_lsn = cpu_to_be64(tail_lsn); > + xlog_verify_tail_lsn(log, iclog, tail_lsn); > + trace_xlog_iclog_syncing(iclog, _RET_IP_); > + > + spin_unlock(&log->l_icloglock); > + xlog_sync(log, iclog); > + spin_lock(&log->l_icloglock); > return 0; > } > > -- > 2.31.1 >
diff --git a/fs/xfs/xfs_log.c b/fs/xfs/xfs_log.c index a3c4d48195d9..82f5996d3889 100644 --- a/fs/xfs/xfs_log.c +++ b/fs/xfs/xfs_log.c @@ -487,29 +487,6 @@ xfs_log_reserve( return error; } -static bool -__xlog_state_release_iclog( - struct xlog *log, - struct xlog_in_core *iclog) -{ - lockdep_assert_held(&log->l_icloglock); - - if (iclog->ic_state == XLOG_STATE_WANT_SYNC) { - /* update tail before writing to iclog */ - xfs_lsn_t tail_lsn = xlog_assign_tail_lsn(log->l_mp); - - iclog->ic_state = XLOG_STATE_SYNCING; - iclog->ic_header.h_tail_lsn = cpu_to_be64(tail_lsn); - xlog_verify_tail_lsn(log, iclog, tail_lsn); - /* cycle incremented when incrementing curr_block */ - trace_xlog_iclog_syncing(iclog, _RET_IP_); - return true; - } - - ASSERT(iclog->ic_state == XLOG_STATE_ACTIVE); - return false; -} - /* * Flush iclog to disk if this is the last reference to the given iclog and the * it is in the WANT_SYNC state. @@ -519,19 +496,31 @@ xlog_state_release_iclog( struct xlog *log, struct xlog_in_core *iclog) { + xfs_lsn_t tail_lsn; lockdep_assert_held(&log->l_icloglock); trace_xlog_iclog_release(iclog, _RET_IP_); if (iclog->ic_state == XLOG_STATE_IOERROR) return -EIO; - if (atomic_dec_and_test(&iclog->ic_refcnt) && - __xlog_state_release_iclog(log, iclog)) { - spin_unlock(&log->l_icloglock); - xlog_sync(log, iclog); - spin_lock(&log->l_icloglock); + if (!atomic_dec_and_test(&iclog->ic_refcnt)) + return 0; + + if (iclog->ic_state != XLOG_STATE_WANT_SYNC) { + ASSERT(iclog->ic_state == XLOG_STATE_ACTIVE); + return 0; } + /* update tail before writing to iclog */ + tail_lsn = xlog_assign_tail_lsn(log->l_mp); + iclog->ic_state = XLOG_STATE_SYNCING; + iclog->ic_header.h_tail_lsn = cpu_to_be64(tail_lsn); + xlog_verify_tail_lsn(log, iclog, tail_lsn); + trace_xlog_iclog_syncing(iclog, _RET_IP_); + + spin_unlock(&log->l_icloglock); + xlog_sync(log, iclog); + spin_lock(&log->l_icloglock); return 0; }