Message ID | 20171213232805.GI6896@magnolia (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
On Wed, Dec 13, 2017 at 03:28:05PM -0800, Darrick J. Wong wrote: > From: Darrick J. Wong <darrick.wong@oracle.com> > > In this test we use a fixed sequence of operations in fsstress to create > some number of files and dirs and then exercise xfsdump/xfsrestore on > them. Since clonerange/deduperange are not supported on all xfs > configurations, detect if they're in fsstress and disable them so that > we always execute exactly the same sequence of operations no matter how > the filesystem is configured. > > Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com> > --- > tests/xfs/068 | 8 ++++++++ > 1 file changed, 8 insertions(+) > > diff --git a/tests/xfs/068 b/tests/xfs/068 > index 7151e28..f95a539 100755 > --- a/tests/xfs/068 > +++ b/tests/xfs/068 > @@ -43,6 +43,14 @@ trap "rm -rf $tmp.*; exit \$status" 0 1 2 3 15 > _supported_fs xfs > _supported_os Linux > > +# Remove fsstress commands that aren't supported on all xfs configs > +if $FSSTRESS_PROG | grep -q clonerange; then > + FSSTRESS_AVOID="-f clonerange=0 $FSSTRESS_AVOID" > +fi > +if $FSSTRESS_PROG | grep -q deduperange; then > + FSSTRESS_AVOID="-f deduperange=0 $FSSTRESS_AVOID" > +fi > + I'd put this inside _create_dumpdir_stress_num as it's supposed to DTRT for the dump/restore that follows. Otherwise looks fine. Cheers, Dave.
On Thu, Dec 14, 2017 at 1:44 AM, Dave Chinner <david@fromorbit.com> wrote: > On Wed, Dec 13, 2017 at 03:28:05PM -0800, Darrick J. Wong wrote: >> From: Darrick J. Wong <darrick.wong@oracle.com> >> >> In this test we use a fixed sequence of operations in fsstress to create >> some number of files and dirs and then exercise xfsdump/xfsrestore on >> them. Since clonerange/deduperange are not supported on all xfs >> configurations, detect if they're in fsstress and disable them so that >> we always execute exactly the same sequence of operations no matter how >> the filesystem is configured. >> >> Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com> >> --- >> tests/xfs/068 | 8 ++++++++ >> 1 file changed, 8 insertions(+) >> >> diff --git a/tests/xfs/068 b/tests/xfs/068 >> index 7151e28..f95a539 100755 >> --- a/tests/xfs/068 >> +++ b/tests/xfs/068 >> @@ -43,6 +43,14 @@ trap "rm -rf $tmp.*; exit \$status" 0 1 2 3 15 >> _supported_fs xfs >> _supported_os Linux >> >> +# Remove fsstress commands that aren't supported on all xfs configs >> +if $FSSTRESS_PROG | grep -q clonerange; then >> + FSSTRESS_AVOID="-f clonerange=0 $FSSTRESS_AVOID" >> +fi >> +if $FSSTRESS_PROG | grep -q deduperange; then >> + FSSTRESS_AVOID="-f deduperange=0 $FSSTRESS_AVOID" >> +fi >> + > > I'd put this inside _create_dumpdir_stress_num as it's supposed to > DTRT for the dump/restore that follows. Otherwise looks fine. > Guys, Please take a look at the only 2 changes in the history of this test. I would like to make sure we are not in a loop: 5d36d85 xfs/068: update golden output due to new operations in fsstress 6e5194d fsstress: Add fallocate insert range operation The first change excludes the new insert op (by dchinner on commit) The second change re-includes insert op, does not exclude new mread/mwrite ops and updates golden output, following this discussion: https://marc.info/?l=fstests&m=149014697111838&w=2 (the referenced thread ends with a ? to Dave, but was followed by v6..v8 that were "silently acked" by Dave). I personally argued that the blacklist approach to xfs/068 is fragile and indeed this is the third time the test breaks in the history I know of, because of added fsstress ops. Fine. As long as we at least stay consistent with a decision about update golden output vs. exclude ops and document the decision in a comment with the reasoning, so we won't have to repeat this discussion next time. Darrick, IMO, we should follow the path of updating golden output and instead of dropping clone/dedupe from ops table in runtime, you should make them a noop or ignore the error, keeping the random sequence unchanged. This is more or less what happens with insert/collapse (error is ignored) already, so it would be weird to make exceptions. For reference, fsx does disable insert/collapse/zero/punch at runtime and that does change the random sequence of fsx. Cheers, Amir. -- To unsubscribe from this list: send the line "unsubscribe fstests" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
On Thu, Dec 14, 2017 at 8:52 AM, Amir Goldstein <amir73il@gmail.com> wrote: > On Thu, Dec 14, 2017 at 1:44 AM, Dave Chinner <david@fromorbit.com> wrote: >> On Wed, Dec 13, 2017 at 03:28:05PM -0800, Darrick J. Wong wrote: >>> From: Darrick J. Wong <darrick.wong@oracle.com> >>> >>> In this test we use a fixed sequence of operations in fsstress to create >>> some number of files and dirs and then exercise xfsdump/xfsrestore on >>> them. Since clonerange/deduperange are not supported on all xfs >>> configurations, detect if they're in fsstress and disable them so that >>> we always execute exactly the same sequence of operations no matter how >>> the filesystem is configured. >>> >>> Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com> >>> --- >>> tests/xfs/068 | 8 ++++++++ >>> 1 file changed, 8 insertions(+) >>> >>> diff --git a/tests/xfs/068 b/tests/xfs/068 >>> index 7151e28..f95a539 100755 >>> --- a/tests/xfs/068 >>> +++ b/tests/xfs/068 >>> @@ -43,6 +43,14 @@ trap "rm -rf $tmp.*; exit \$status" 0 1 2 3 15 >>> _supported_fs xfs >>> _supported_os Linux >>> >>> +# Remove fsstress commands that aren't supported on all xfs configs >>> +if $FSSTRESS_PROG | grep -q clonerange; then >>> + FSSTRESS_AVOID="-f clonerange=0 $FSSTRESS_AVOID" >>> +fi >>> +if $FSSTRESS_PROG | grep -q deduperange; then >>> + FSSTRESS_AVOID="-f deduperange=0 $FSSTRESS_AVOID" >>> +fi >>> + >> >> I'd put this inside _create_dumpdir_stress_num as it's supposed to >> DTRT for the dump/restore that follows. Otherwise looks fine. >> > > Guys, > > Please take a look at the only 2 changes in the history of this test. > I would like to make sure we are not in a loop: > > 5d36d85 xfs/068: update golden output due to new operations in fsstress > 6e5194d fsstress: Add fallocate insert range operation > > The first change excludes the new insert op (by dchinner on commit) > The second change re-includes insert op, does not exclude new > mread/mwrite ops and updates golden output, following this discussion: > https://marc.info/?l=fstests&m=149014697111838&w=2 > (the referenced thread ends with a ? to Dave, but was followed by v6..v8 > that were "silently acked" by Dave). > > I personally argued that the blacklist approach to xfs/068 is fragile and indeed > this is the third time the test breaks in the history I know of, > because of added > fsstress ops. Fine. As long as we at least stay consistent with a decision about > update golden output vs. exclude ops and document the decision in a comment > with the reasoning, so we won't have to repeat this discussion next time. > > Darrick, > > IMO, we should follow the path of updating golden output and instead of > dropping clone/dedupe from ops table in runtime, you should make them > a noop or ignore the error, keeping the random sequence unchanged. > This is more or less what happens with insert/collapse (error is ignored) > already, so it would be weird to make exceptions. > > For reference, fsx does disable insert/collapse/zero/punch at runtime > and that does change the random sequence of fsx. > Looking again, I think this test is currently broken if Fallocate is disabled at build time and will also be Broken if we fix golden output for clonerange and it Will be disabled in build time, because rand() on the Operation params won't happen. Amir. -- To unsubscribe from this list: send the line "unsubscribe fstests" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
On Thu, Dec 14, 2017 at 08:52:32AM +0200, Amir Goldstein wrote: > On Thu, Dec 14, 2017 at 1:44 AM, Dave Chinner <david@fromorbit.com> wrote: > > On Wed, Dec 13, 2017 at 03:28:05PM -0800, Darrick J. Wong wrote: > >> From: Darrick J. Wong <darrick.wong@oracle.com> > >> > >> In this test we use a fixed sequence of operations in fsstress to create > >> some number of files and dirs and then exercise xfsdump/xfsrestore on > >> them. Since clonerange/deduperange are not supported on all xfs > >> configurations, detect if they're in fsstress and disable them so that > >> we always execute exactly the same sequence of operations no matter how > >> the filesystem is configured. > >> > >> Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com> > >> --- > >> tests/xfs/068 | 8 ++++++++ > >> 1 file changed, 8 insertions(+) > >> > >> diff --git a/tests/xfs/068 b/tests/xfs/068 > >> index 7151e28..f95a539 100755 > >> --- a/tests/xfs/068 > >> +++ b/tests/xfs/068 > >> @@ -43,6 +43,14 @@ trap "rm -rf $tmp.*; exit \$status" 0 1 2 3 15 > >> _supported_fs xfs > >> _supported_os Linux > >> > >> +# Remove fsstress commands that aren't supported on all xfs configs > >> +if $FSSTRESS_PROG | grep -q clonerange; then > >> + FSSTRESS_AVOID="-f clonerange=0 $FSSTRESS_AVOID" > >> +fi > >> +if $FSSTRESS_PROG | grep -q deduperange; then > >> + FSSTRESS_AVOID="-f deduperange=0 $FSSTRESS_AVOID" > >> +fi > >> + > > > > I'd put this inside _create_dumpdir_stress_num as it's supposed to > > DTRT for the dump/restore that follows. Otherwise looks fine. > > > > Guys, > > Please take a look at the only 2 changes in the history of this test. > I would like to make sure we are not in a loop: > > 5d36d85 xfs/068: update golden output due to new operations in fsstress > 6e5194d fsstress: Add fallocate insert range operation > > The first change excludes the new insert op (by dchinner on commit) > The second change re-includes insert op, does not exclude new > mread/mwrite ops and updates golden output, following this discussion: > https://marc.info/?l=fstests&m=149014697111838&w=2 > (the referenced thread ends with a ? to Dave, but was followed by v6..v8 > that were "silently acked" by Dave). > > I personally argued that the blacklist approach to xfs/068 is fragile and indeed > this is the third time the test breaks in the history I know of, > because of added > fsstress ops. Fine. As long as we at least stay consistent with a decision about > update golden output vs. exclude ops and document the decision in a comment > with the reasoning, so we won't have to repeat this discussion next time. I think the fundamental problem of xfs/068 is the hardcoded file numbers in .out file, perhaps we should calculate the expected number of files/dirs to be dumped/restored before the dump test and extract the actual restored number of files/dirs from xfsrestore output and do a comparison. (or save the whole tree structure for comparison? I haven't done any test yet, just some random thoughts for now.) Currently, xfs/068 will easily break if there's user-defined FSSTRESS_AVOID, e.g. FSSTRESS_AVOID="-ffallocate=0", and that's totally legal test configuration. IHMO we really should fix xfs/068 first to avoid hitting the same problem again and again. Thanks, Eryu > > Darrick, > > IMO, we should follow the path of updating golden output and instead of > dropping clone/dedupe from ops table in runtime, you should make them > a noop or ignore the error, keeping the random sequence unchanged. > This is more or less what happens with insert/collapse (error is ignored) > already, so it would be weird to make exceptions. > > For reference, fsx does disable insert/collapse/zero/punch at runtime > and that does change the random sequence of fsx. > > Cheers, > Amir. -- To unsubscribe from this list: send the line "unsubscribe fstests" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
On Thu, Dec 14, 2017 at 9:49 AM, Eryu Guan <eguan@redhat.com> wrote: > On Thu, Dec 14, 2017 at 08:52:32AM +0200, Amir Goldstein wrote: >> On Thu, Dec 14, 2017 at 1:44 AM, Dave Chinner <david@fromorbit.com> wrote: >> > On Wed, Dec 13, 2017 at 03:28:05PM -0800, Darrick J. Wong wrote: >> >> From: Darrick J. Wong <darrick.wong@oracle.com> >> >> >> >> In this test we use a fixed sequence of operations in fsstress to create >> >> some number of files and dirs and then exercise xfsdump/xfsrestore on >> >> them. Since clonerange/deduperange are not supported on all xfs >> >> configurations, detect if they're in fsstress and disable them so that >> >> we always execute exactly the same sequence of operations no matter how >> >> the filesystem is configured. >> >> >> >> Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com> >> >> --- >> >> tests/xfs/068 | 8 ++++++++ >> >> 1 file changed, 8 insertions(+) >> >> >> >> diff --git a/tests/xfs/068 b/tests/xfs/068 >> >> index 7151e28..f95a539 100755 >> >> --- a/tests/xfs/068 >> >> +++ b/tests/xfs/068 >> >> @@ -43,6 +43,14 @@ trap "rm -rf $tmp.*; exit \$status" 0 1 2 3 15 >> >> _supported_fs xfs >> >> _supported_os Linux >> >> >> >> +# Remove fsstress commands that aren't supported on all xfs configs >> >> +if $FSSTRESS_PROG | grep -q clonerange; then >> >> + FSSTRESS_AVOID="-f clonerange=0 $FSSTRESS_AVOID" >> >> +fi >> >> +if $FSSTRESS_PROG | grep -q deduperange; then >> >> + FSSTRESS_AVOID="-f deduperange=0 $FSSTRESS_AVOID" >> >> +fi >> >> + >> > >> > I'd put this inside _create_dumpdir_stress_num as it's supposed to >> > DTRT for the dump/restore that follows. Otherwise looks fine. >> > >> >> Guys, >> >> Please take a look at the only 2 changes in the history of this test. >> I would like to make sure we are not in a loop: >> >> 5d36d85 xfs/068: update golden output due to new operations in fsstress >> 6e5194d fsstress: Add fallocate insert range operation >> >> The first change excludes the new insert op (by dchinner on commit) >> The second change re-includes insert op, does not exclude new >> mread/mwrite ops and updates golden output, following this discussion: >> https://marc.info/?l=fstests&m=149014697111838&w=2 >> (the referenced thread ends with a ? to Dave, but was followed by v6..v8 >> that were "silently acked" by Dave). >> >> I personally argued that the blacklist approach to xfs/068 is fragile and indeed >> this is the third time the test breaks in the history I know of, >> because of added >> fsstress ops. Fine. As long as we at least stay consistent with a decision about >> update golden output vs. exclude ops and document the decision in a comment >> with the reasoning, so we won't have to repeat this discussion next time. > > I think the fundamental problem of xfs/068 is the hardcoded file numbers > in .out file, perhaps we should calculate the expected number of > files/dirs to be dumped/restored before the dump test and extract the > actual restored number of files/dirs from xfsrestore output and do a > comparison. (or save the whole tree structure for comparison? I haven't > done any test yet, just some random thoughts for now.) > > Currently, xfs/068 will easily break if there's user-defined > FSSTRESS_AVOID, e.g. FSSTRESS_AVOID="-ffallocate=0", and that's totally > legal test configuration. > > IHMO we really should fix xfs/068 first to avoid hitting the same > problem again and again. > Agreed. And while at it, need to redirect: echo "fsstress : $_param" to $seq.full and remove it from golden output and print the actual parameters in the log including $FSSTRESS_AVOID, not the make believe params. Amir. -- To unsubscribe from this list: send the line "unsubscribe fstests" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
On Thu, Dec 14, 2017 at 03:49:47PM +0800, Eryu Guan wrote: > On Thu, Dec 14, 2017 at 08:52:32AM +0200, Amir Goldstein wrote: > > On Thu, Dec 14, 2017 at 1:44 AM, Dave Chinner <david@fromorbit.com> wrote: > > > On Wed, Dec 13, 2017 at 03:28:05PM -0800, Darrick J. Wong wrote: > > >> From: Darrick J. Wong <darrick.wong@oracle.com> > > >> > > >> In this test we use a fixed sequence of operations in fsstress to create > > >> some number of files and dirs and then exercise xfsdump/xfsrestore on > > >> them. Since clonerange/deduperange are not supported on all xfs > > >> configurations, detect if they're in fsstress and disable them so that > > >> we always execute exactly the same sequence of operations no matter how > > >> the filesystem is configured. > > >> > > >> Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com> > > >> --- > > >> tests/xfs/068 | 8 ++++++++ > > >> 1 file changed, 8 insertions(+) > > >> > > >> diff --git a/tests/xfs/068 b/tests/xfs/068 > > >> index 7151e28..f95a539 100755 > > >> --- a/tests/xfs/068 > > >> +++ b/tests/xfs/068 > > >> @@ -43,6 +43,14 @@ trap "rm -rf $tmp.*; exit \$status" 0 1 2 3 15 > > >> _supported_fs xfs > > >> _supported_os Linux > > >> > > >> +# Remove fsstress commands that aren't supported on all xfs configs > > >> +if $FSSTRESS_PROG | grep -q clonerange; then > > >> + FSSTRESS_AVOID="-f clonerange=0 $FSSTRESS_AVOID" > > >> +fi > > >> +if $FSSTRESS_PROG | grep -q deduperange; then > > >> + FSSTRESS_AVOID="-f deduperange=0 $FSSTRESS_AVOID" > > >> +fi > > >> + > > > > > > I'd put this inside _create_dumpdir_stress_num as it's supposed to > > > DTRT for the dump/restore that follows. Otherwise looks fine. > > > > > > > Guys, > > > > Please take a look at the only 2 changes in the history of this test. > > I would like to make sure we are not in a loop: > > > > 5d36d85 xfs/068: update golden output due to new operations in fsstress > > 6e5194d fsstress: Add fallocate insert range operation > > > > The first change excludes the new insert op (by dchinner on commit) > > The second change re-includes insert op, does not exclude new > > mread/mwrite ops and updates golden output, following this discussion: > > https://marc.info/?l=fstests&m=149014697111838&w=2 > > (the referenced thread ends with a ? to Dave, but was followed by v6..v8 > > that were "silently acked" by Dave). > > > > I personally argued that the blacklist approach to xfs/068 is fragile and indeed > > this is the third time the test breaks in the history I know of, > > because of added > > fsstress ops. Fine. As long as we at least stay consistent with a decision about > > update golden output vs. exclude ops and document the decision in a comment > > with the reasoning, so we won't have to repeat this discussion next time. > > I think the fundamental problem of xfs/068 is the hardcoded file numbers > in .out file, perhaps we should calculate the expected number of > files/dirs to be dumped/restored before the dump test and extract the > actual restored number of files/dirs from xfsrestore output and do a > comparison. (or save the whole tree structure for comparison? I haven't > done any test yet, just some random thoughts for now.) Or we don't waste any more time on trying to make a reliable, stable regression test that has a history of detecting bulkstat regressions work differently? Indeed, the problem here is our "turn on new functionality in fsstress as it is added" process will always break older tests that require fixed functionality to test. Having tests fail when we do this is perfectly reasonable - it means we have to consider whether that new fsstress operation is valid for the test being run. Making tests silently accept new operations that may not be valid for the thing being tested doesn't improve our test coverage. What it does is take away a warning canary that tells us we may have broken something we didn't intend to break. e.g. maybe this test is telling us reflink breaks xfsdump or xfsrestore? That's the point of having hard coded numbers in the golden output - any change whether intended, expected or otherwise requires us to go look at whether that new functionality has broken xfsdump/restore. That's what regression tests are for, and taking that away from the test under the guise of "easier test maintenance" is misguided. Regression tests require validation and checking when new functionality is added to the tools they use. Having old tests fail when new features are added is exactly what we want the regression tests to do, otherwise we'll miss regressions that the current code actually catches. > Currently, xfs/068 will easily break if there's user-defined > FSSTRESS_AVOID, e.g. FSSTRESS_AVOID="-ffallocate=0", and that's totally > legal test configuration. I think that's a strawman argument - who tests XFS without fallocate enabled these days? Indeed, fsstress will still be doing preallocation via the old ALLOCSP and RESVSP ioctls that predate fallocate.... > IHMO we really should fix xfs/068 first to avoid hitting the same > problem again and again. IMO, we should not be changing the way old tests work, especially those that have, in the past, been very good at exposing bugs in kernel interfaces. Cheers, Dave.
On Fri, Dec 15, 2017 at 08:35:41AM +1100, Dave Chinner wrote: > On Thu, Dec 14, 2017 at 03:49:47PM +0800, Eryu Guan wrote: > > On Thu, Dec 14, 2017 at 08:52:32AM +0200, Amir Goldstein wrote: > > > On Thu, Dec 14, 2017 at 1:44 AM, Dave Chinner <david@fromorbit.com> wrote: > > > > On Wed, Dec 13, 2017 at 03:28:05PM -0800, Darrick J. Wong wrote: > > > >> From: Darrick J. Wong <darrick.wong@oracle.com> > > > >> > > > >> In this test we use a fixed sequence of operations in fsstress to create > > > >> some number of files and dirs and then exercise xfsdump/xfsrestore on > > > >> them. Since clonerange/deduperange are not supported on all xfs > > > >> configurations, detect if they're in fsstress and disable them so that > > > >> we always execute exactly the same sequence of operations no matter how > > > >> the filesystem is configured. > > > >> > > > >> Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com> > > > >> --- > > > >> tests/xfs/068 | 8 ++++++++ > > > >> 1 file changed, 8 insertions(+) > > > >> > > > >> diff --git a/tests/xfs/068 b/tests/xfs/068 > > > >> index 7151e28..f95a539 100755 > > > >> --- a/tests/xfs/068 > > > >> +++ b/tests/xfs/068 > > > >> @@ -43,6 +43,14 @@ trap "rm -rf $tmp.*; exit \$status" 0 1 2 3 15 > > > >> _supported_fs xfs > > > >> _supported_os Linux > > > >> > > > >> +# Remove fsstress commands that aren't supported on all xfs configs > > > >> +if $FSSTRESS_PROG | grep -q clonerange; then > > > >> + FSSTRESS_AVOID="-f clonerange=0 $FSSTRESS_AVOID" > > > >> +fi > > > >> +if $FSSTRESS_PROG | grep -q deduperange; then > > > >> + FSSTRESS_AVOID="-f deduperange=0 $FSSTRESS_AVOID" > > > >> +fi > > > >> + > > > > > > > > I'd put this inside _create_dumpdir_stress_num as it's supposed to > > > > DTRT for the dump/restore that follows. Otherwise looks fine. > > > > > > > > > > Guys, > > > > > > Please take a look at the only 2 changes in the history of this test. > > > I would like to make sure we are not in a loop: > > > > > > 5d36d85 xfs/068: update golden output due to new operations in fsstress > > > 6e5194d fsstress: Add fallocate insert range operation > > > > > > The first change excludes the new insert op (by dchinner on commit) > > > The second change re-includes insert op, does not exclude new > > > mread/mwrite ops and updates golden output, following this discussion: > > > https://marc.info/?l=fstests&m=149014697111838&w=2 > > > (the referenced thread ends with a ? to Dave, but was followed by v6..v8 > > > that were "silently acked" by Dave). > > > > > > I personally argued that the blacklist approach to xfs/068 is fragile and indeed > > > this is the third time the test breaks in the history I know of, > > > because of added > > > fsstress ops. Fine. As long as we at least stay consistent with a decision about > > > update golden output vs. exclude ops and document the decision in a comment > > > with the reasoning, so we won't have to repeat this discussion next time. > > > > I think the fundamental problem of xfs/068 is the hardcoded file numbers > > in .out file, perhaps we should calculate the expected number of > > files/dirs to be dumped/restored before the dump test and extract the > > actual restored number of files/dirs from xfsrestore output and do a > > comparison. (or save the whole tree structure for comparison? I haven't > > done any test yet, just some random thoughts for now.) > > Or we don't waste any more time on trying to make a reliable, stable > regression test that has a history of detecting bulkstat regressions > work differently? <shrug> See now, the frustrating part about fixing this testcase is that I still don't feel like I have a good grasp on what this thing is trying to test -- apparently we're checking for bulkstat regressions, dump problems, and restore problems? Are we also looking for problems that might crop up with the newer APIs, whatever those might be? Currently I have a reworked version of this patch that runs fsstress, measures the number of directories and inodes in $dump_dir, then programmatically compares that to whatever xfsrestore tells us it restored. This ought to be enough that we can create a sufficiently messy filesystem with whatever sequence of syscalls we want, and make sure that dump/restore actually work on them. First we run fsstress, then we count the number of dirs, the number of fs objects, take a snapshot of the 'find .' output, and md5sum every file in the dump directory. If fsstress creates fewer than 100 dirs or 600 inodes, we fail the test because that wasn't enough. If bulkstat fails to iterate all the inodes, restore's output will reflect fewer files than was expected. If dump fails to generate a full dump, restore's output will reflect fewer files than was expected. If restore fails to restore the full dump, restore's output will reflect fewer files than was expected. If the restore output doesn't reflect the number of dirs/inodes we counted at the beginning, we fail the test. If the 'find .' output of the restored dir doesn't match the original, we fail the test. If the md5sum -c output shows corrupt files, we fail the test. So now I really have no idea -- is that enough to check that everything works? I felt like it does, but given all the back and forth now I'm wondering if even this is enough. (Yeah, I'm frustrated because the fsstress additions have been very helpful at flushing out more reflink bugs and I feel like I'm making very little progress on this xfs/068 thing. Sorry.) --D > Indeed, the problem here is our "turn on new functionality in > fsstress as it is added" process will always break older tests that > require fixed functionality to test. Having tests fail when we do > this is perfectly reasonable - it means we have to consider whether > that new fsstress operation is valid for the test being run. > > Making tests silently accept new operations that may not be valid > for the thing being tested doesn't improve our test coverage. What > it does is take away a warning canary that tells us we may have > broken something we didn't intend to break. e.g. maybe this test is > telling us reflink breaks xfsdump or xfsrestore? That's the point of > having hard coded numbers in the golden output - any change whether > intended, expected or otherwise requires us to go look at whether > that new functionality has broken xfsdump/restore. > > That's what regression tests are for, and taking that away from the > test under the guise of "easier test maintenance" is misguided. > Regression tests require validation and checking when new > functionality is added to the tools they use. Having old tests fail > when new features are added is exactly what we want the regression > tests to do, otherwise we'll miss regressions that the current code > actually catches. > > > Currently, xfs/068 will easily break if there's user-defined > > FSSTRESS_AVOID, e.g. FSSTRESS_AVOID="-ffallocate=0", and that's totally > > legal test configuration. > > I think that's a strawman argument - who tests XFS without fallocate > enabled these days? Indeed, fsstress will still be doing > preallocation via the old ALLOCSP and RESVSP ioctls that predate > fallocate.... > > > IHMO we really should fix xfs/068 first to avoid hitting the same > > problem again and again. > > IMO, we should not be changing the way old tests work, especially > those that have, in the past, been very good at exposing bugs in > kernel interfaces. > > Cheers, > > Dave. > -- > Dave Chinner > david@fromorbit.com > -- > To unsubscribe from this list: send the line "unsubscribe fstests" in > the body of a message to majordomo@vger.kernel.org > More majordomo info at http://vger.kernel.org/majordomo-info.html -- To unsubscribe from this list: send the line "unsubscribe fstests" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
On Thu, Dec 14, 2017 at 06:04:19PM -0800, Darrick J. Wong wrote: > On Fri, Dec 15, 2017 at 08:35:41AM +1100, Dave Chinner wrote: > > On Thu, Dec 14, 2017 at 03:49:47PM +0800, Eryu Guan wrote: > > > On Thu, Dec 14, 2017 at 08:52:32AM +0200, Amir Goldstein wrote: > > > > On Thu, Dec 14, 2017 at 1:44 AM, Dave Chinner <david@fromorbit.com> wrote: > > > > > On Wed, Dec 13, 2017 at 03:28:05PM -0800, Darrick J. Wong wrote: > > > > >> From: Darrick J. Wong <darrick.wong@oracle.com> > > > > >> > > > > >> In this test we use a fixed sequence of operations in fsstress to create > > > > >> some number of files and dirs and then exercise xfsdump/xfsrestore on > > > > >> them. Since clonerange/deduperange are not supported on all xfs > > > > >> configurations, detect if they're in fsstress and disable them so that > > > > >> we always execute exactly the same sequence of operations no matter how > > > > >> the filesystem is configured. > > > > >> > > > > >> Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com> > > > > >> --- > > > > >> tests/xfs/068 | 8 ++++++++ > > > > >> 1 file changed, 8 insertions(+) > > > > >> > > > > >> diff --git a/tests/xfs/068 b/tests/xfs/068 > > > > >> index 7151e28..f95a539 100755 > > > > >> --- a/tests/xfs/068 > > > > >> +++ b/tests/xfs/068 > > > > >> @@ -43,6 +43,14 @@ trap "rm -rf $tmp.*; exit \$status" 0 1 2 3 15 > > > > >> _supported_fs xfs > > > > >> _supported_os Linux > > > > >> > > > > >> +# Remove fsstress commands that aren't supported on all xfs configs > > > > >> +if $FSSTRESS_PROG | grep -q clonerange; then > > > > >> + FSSTRESS_AVOID="-f clonerange=0 $FSSTRESS_AVOID" > > > > >> +fi > > > > >> +if $FSSTRESS_PROG | grep -q deduperange; then > > > > >> + FSSTRESS_AVOID="-f deduperange=0 $FSSTRESS_AVOID" > > > > >> +fi > > > > >> + > > > > > > > > > > I'd put this inside _create_dumpdir_stress_num as it's supposed to > > > > > DTRT for the dump/restore that follows. Otherwise looks fine. > > > > > > > > > > > > > Guys, > > > > > > > > Please take a look at the only 2 changes in the history of this test. > > > > I would like to make sure we are not in a loop: > > > > > > > > 5d36d85 xfs/068: update golden output due to new operations in fsstress > > > > 6e5194d fsstress: Add fallocate insert range operation > > > > > > > > The first change excludes the new insert op (by dchinner on commit) > > > > The second change re-includes insert op, does not exclude new > > > > mread/mwrite ops and updates golden output, following this discussion: > > > > https://marc.info/?l=fstests&m=149014697111838&w=2 > > > > (the referenced thread ends with a ? to Dave, but was followed by v6..v8 > > > > that were "silently acked" by Dave). > > > > > > > > I personally argued that the blacklist approach to xfs/068 is fragile and indeed > > > > this is the third time the test breaks in the history I know of, > > > > because of added > > > > fsstress ops. Fine. As long as we at least stay consistent with a decision about > > > > update golden output vs. exclude ops and document the decision in a comment > > > > with the reasoning, so we won't have to repeat this discussion next time. > > > > > > I think the fundamental problem of xfs/068 is the hardcoded file numbers > > > in .out file, perhaps we should calculate the expected number of > > > files/dirs to be dumped/restored before the dump test and extract the > > > actual restored number of files/dirs from xfsrestore output and do a > > > comparison. (or save the whole tree structure for comparison? I haven't > > > done any test yet, just some random thoughts for now.) > > > > Or we don't waste any more time on trying to make a reliable, stable > > regression test that has a history of detecting bulkstat regressions > > work differently? > > <shrug> See now, the frustrating part about fixing this testcase is that > I still don't feel like I have a good grasp on what this thing is trying > to test -- apparently we're checking for bulkstat regressions, dump > problems, and restore problems? commit 481c28f52fd4ed3976f2733a1c65f92760138258 Author: Eric Sandeen <sandeen@redhat.com> Date: Tue Oct 14 22:59:39 2014 +1100 xfs: test larger dump/restore to/from file This test creates a large-ish directory structure using fsstress, and does a dump/restore to make sure we dump all the files. Without the fix for the regression caused by: c7cb51d xfs: fix error handling at xfs_inumbers we will see failures like: -xfsrestore: 486 directories and 1590 entries processed +xfsrestore: 30 directories and 227 entries processed as it fails to process all inodes. I think that existing tests have a much smaller set of files, and so don't trip the bug. I don't do a file-by-file comparison here, because for some reason the diff output gets garbled; this test only checks that we've dumped & restored the correct number of files. Signed-off-by: Eric Sandeen <sandeen@redhat.com> Reviewed-by: Dave Chinner <dchinner@redhat.com> Signed-off-by: Dave Chinner <david@fromorbit.com> FWIW, I'm pretty sure the diff problems were related to binary file contents, so it was dropped as it wasn't critical to validting that bulkstat and inode number iteration worked correctly. > Are we also looking for problems that > might crop up with the newer APIs, whatever those might be? No, we're explicitly using fsstress to generate a dataset large enough to cause iteration over various API xfsdump relies on to work correctly. i.e. the features fsstress have are irrelevant to the functioning of this test - we want it to generate a specific, consistent, deterministic data set and that's it. Really, all I care about is that we don't overcomplicate the problem and the solution. Just adding commands to the avoid list for fsstress is a perfectly acceptible, simple solutioni - we've done it twice in 3 years for this test, and we've done it for other tests, too. It's hardly a crippling maintenance burden. And, FWIW, we check the file count from xfsrestore in teh golden output of pretty much every xfsdump/restore test: $ git grep "entries processed" tests/xfs tests/xfs/022:_do_restore | sed -e "/entries processed$/s/[0-9][0-9]*/NUM/g" tests/xfs/022.out:xfsrestore: NUM directories and NUM entries processed tests/xfs/023.out:xfsrestore: 3 directories and 38 entries processed tests/xfs/024.out:xfsrestore: 3 directories and 38 entries processed tests/xfs/025.out:xfsrestore: 3 directories and 38 entries processed tests/xfs/026.out:xfsrestore: 3 directories and 38 entries processed tests/xfs/027.out:xfsrestore: 3 directories and 39 entries processed tests/xfs/035.out:xfsrestore: 3 directories and 6 entries processed tests/xfs/036.out:xfsrestore: 3 directories and 38 entries processed tests/xfs/037.out:xfsrestore: 3 directories and 38 entries processed tests/xfs/038.out:xfsrestore: 3 directories and 38 entries processed tests/xfs/039.out:xfsrestore: 3 directories and 38 entries processed tests/xfs/043.out:xfsrestore: 3 directories and 38 entries processed tests/xfs/046.out:xfsrestore: 3 directories and 10 entries processed tests/xfs/055.out:xfsrestore: 3 directories and 38 entries processed tests/xfs/056.out:xfsrestore: 7 directories and 11 entries processed tests/xfs/060.out:xfsrestore: 3 directories and 41 entries processed tests/xfs/061.out:xfsrestore: 7 directories and 11 entries processed tests/xfs/063.out:xfsrestore: 4 directories and 21 entries processed ..... Really, I don't see a need to do anything else than avoid the fsstress ops that caused the change of behaviour. All the other xfsdump/restore tests do file and directory tree validations, so they are going to catch any regression on that side of things. This test just exercises iteration of various APIs that we've broken in the past... > Currently I have a reworked version of this patch that runs > fsstress, measures the number of directories and inodes in > $dump_dir, then programmatically compares that to whatever > xfsrestore tells us it restored. This ought to be enough that we > can create a sufficiently messy filesystem with whatever sequence > of syscalls we want, and make sure that dump/restore actually work > on them. > > First we run fsstress, then we count the number of dirs, the > number of fs objects, take a snapshot of the 'find .' output, and > md5sum every file in the dump directory. > > If fsstress creates fewer than 100 dirs or 600 inodes, we fail the > test because that wasn't enough. > > If bulkstat fails to iterate all the inodes, restore's output will > reflect fewer files than was expected. > > If dump fails to generate a full dump, restore's output will > reflect fewer files than was expected. > > If restore fails to restore the full dump, restore's output will > reflect fewer files than was expected. > > If the restore output doesn't reflect the number of dirs/inodes we > counted at the beginning, we fail the test. > > If the 'find .' output of the restored dir doesn't match the > original, we fail the test. > > If the md5sum -c output shows corrupt files, we fail the test. > > So now I really have no idea -- is that enough to check that > everything works? I felt like it does, but given all the back and > forth now I'm wondering if even this is enough. What did I say about not wanting to overcomplicate the problem and the solution? :/ Folks, I don't say "leave it alone, it's fine" without a good reason. If you've never tried to debug xfsdump or xfsrestore, and you aren't familiar with the ancient xfsdump and restore unit tests that were written before anyone here was working on linux, then don't suggest we rewrite them to make them nicer. Their value is in the fact they've been around almost entirely unchanged for 15 years and they still catch bugs.... Make whatever changes are necessary to keep them running exactly as they are and don't change them unless xfsdump/restore testing requires them to be changed. > (Yeah, I'm frustrated because the fsstress additions have been > very helpful at flushing out more reflink bugs and I feel like I'm > making very little progress on this xfs/068 thing. Sorry.) Well, I thought it was all sorted until people started suggesting we do crazy things like you've now gone and done. Cheers, Dave.
On Fri, Dec 15, 2017 at 6:37 AM, Dave Chinner <david@fromorbit.com> wrote: > On Thu, Dec 14, 2017 at 06:04:19PM -0800, Darrick J. Wong wrote: >> On Fri, Dec 15, 2017 at 08:35:41AM +1100, Dave Chinner wrote: >> > On Thu, Dec 14, 2017 at 03:49:47PM +0800, Eryu Guan wrote: >> > > On Thu, Dec 14, 2017 at 08:52:32AM +0200, Amir Goldstein wrote: >> > > > On Thu, Dec 14, 2017 at 1:44 AM, Dave Chinner <david@fromorbit.com> wrote: >> > > > > On Wed, Dec 13, 2017 at 03:28:05PM -0800, Darrick J. Wong wrote: >> > > > >> From: Darrick J. Wong <darrick.wong@oracle.com> >> > > > >> >> > > > >> In this test we use a fixed sequence of operations in fsstress to create >> > > > >> some number of files and dirs and then exercise xfsdump/xfsrestore on >> > > > >> them. Since clonerange/deduperange are not supported on all xfs >> > > > >> configurations, detect if they're in fsstress and disable them so that >> > > > >> we always execute exactly the same sequence of operations no matter how >> > > > >> the filesystem is configured. >> > > > >> >> > > > >> Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com> >> > > > >> --- >> > > > >> tests/xfs/068 | 8 ++++++++ >> > > > >> 1 file changed, 8 insertions(+) >> > > > >> >> > > > >> diff --git a/tests/xfs/068 b/tests/xfs/068 >> > > > >> index 7151e28..f95a539 100755 >> > > > >> --- a/tests/xfs/068 >> > > > >> +++ b/tests/xfs/068 >> > > > >> @@ -43,6 +43,14 @@ trap "rm -rf $tmp.*; exit \$status" 0 1 2 3 15 >> > > > >> _supported_fs xfs >> > > > >> _supported_os Linux >> > > > >> >> > > > >> +# Remove fsstress commands that aren't supported on all xfs configs >> > > > >> +if $FSSTRESS_PROG | grep -q clonerange; then >> > > > >> + FSSTRESS_AVOID="-f clonerange=0 $FSSTRESS_AVOID" >> > > > >> +fi >> > > > >> +if $FSSTRESS_PROG | grep -q deduperange; then >> > > > >> + FSSTRESS_AVOID="-f deduperange=0 $FSSTRESS_AVOID" >> > > > >> +fi >> > > > >> + >> > > > > >> > > > > I'd put this inside _create_dumpdir_stress_num as it's supposed to >> > > > > DTRT for the dump/restore that follows. Otherwise looks fine. >> > > > > >> > > > >> > > > Guys, >> > > > >> > > > Please take a look at the only 2 changes in the history of this test. >> > > > I would like to make sure we are not in a loop: >> > > > >> > > > 5d36d85 xfs/068: update golden output due to new operations in fsstress >> > > > 6e5194d fsstress: Add fallocate insert range operation >> > > > >> > > > The first change excludes the new insert op (by dchinner on commit) >> > > > The second change re-includes insert op, does not exclude new >> > > > mread/mwrite ops and updates golden output, following this discussion: >> > > > https://marc.info/?l=fstests&m=149014697111838&w=2 >> > > > (the referenced thread ends with a ? to Dave, but was followed by v6..v8 >> > > > that were "silently acked" by Dave). >> > > > >> > > > I personally argued that the blacklist approach to xfs/068 is fragile and indeed >> > > > this is the third time the test breaks in the history I know of, >> > > > because of added >> > > > fsstress ops. Fine. As long as we at least stay consistent with a decision about >> > > > update golden output vs. exclude ops and document the decision in a comment >> > > > with the reasoning, so we won't have to repeat this discussion next time. >> > > >> > > I think the fundamental problem of xfs/068 is the hardcoded file numbers >> > > in .out file, perhaps we should calculate the expected number of >> > > files/dirs to be dumped/restored before the dump test and extract the >> > > actual restored number of files/dirs from xfsrestore output and do a >> > > comparison. (or save the whole tree structure for comparison? I haven't >> > > done any test yet, just some random thoughts for now.) >> > >> > Or we don't waste any more time on trying to make a reliable, stable >> > regression test that has a history of detecting bulkstat regressions >> > work differently? >> >> <shrug> See now, the frustrating part about fixing this testcase is that >> I still don't feel like I have a good grasp on what this thing is trying >> to test -- apparently we're checking for bulkstat regressions, dump >> problems, and restore problems? > > commit 481c28f52fd4ed3976f2733a1c65f92760138258 > Author: Eric Sandeen <sandeen@redhat.com> > Date: Tue Oct 14 22:59:39 2014 +1100 > > xfs: test larger dump/restore to/from file > > This test creates a large-ish directory structure using > fsstress, and does a dump/restore to make sure we dump > all the files. > > Without the fix for the regression caused by: > c7cb51d xfs: fix error handling at xfs_inumbers > > we will see failures like: > > -xfsrestore: 486 directories and 1590 entries processed > +xfsrestore: 30 directories and 227 entries processed > > as it fails to process all inodes. > > I think that existing tests have a much smaller set of files, > and so don't trip the bug. > > I don't do a file-by-file comparison here, because for some > reason the diff output gets garbled; this test only checks > that we've dumped & restored the correct number of files. > > Signed-off-by: Eric Sandeen <sandeen@redhat.com> > Reviewed-by: Dave Chinner <dchinner@redhat.com> > Signed-off-by: Dave Chinner <david@fromorbit.com> > > FWIW, I'm pretty sure the diff problems were related to binary file > contents, so it was dropped as it wasn't critical to validting > that bulkstat and inode number iteration worked correctly. > >> Are we also looking for problems that >> might crop up with the newer APIs, whatever those might be? > > No, we're explicitly using fsstress to generate a dataset large > enough to cause iteration over various API xfsdump relies on to > work correctly. i.e. the features fsstress have are irrelevant to > the functioning of this test - we want it to generate a specific, > consistent, deterministic data set and that's it. > > Really, all I care about is that we don't overcomplicate the > problem and the solution. Just adding commands to the avoid list > for fsstress is a perfectly acceptible, simple solutioni - we've > done it twice in 3 years for this test, and we've done it for other > tests, too. It's hardly a crippling maintenance burden. > > And, FWIW, we check the file count from xfsrestore in teh golden > output of pretty much every xfsdump/restore test: > > $ git grep "entries processed" tests/xfs > tests/xfs/022:_do_restore | sed -e "/entries processed$/s/[0-9][0-9]*/NUM/g" > tests/xfs/022.out:xfsrestore: NUM directories and NUM entries processed > tests/xfs/023.out:xfsrestore: 3 directories and 38 entries processed > tests/xfs/024.out:xfsrestore: 3 directories and 38 entries processed > tests/xfs/025.out:xfsrestore: 3 directories and 38 entries processed > tests/xfs/026.out:xfsrestore: 3 directories and 38 entries processed > tests/xfs/027.out:xfsrestore: 3 directories and 39 entries processed > tests/xfs/035.out:xfsrestore: 3 directories and 6 entries processed > tests/xfs/036.out:xfsrestore: 3 directories and 38 entries processed > tests/xfs/037.out:xfsrestore: 3 directories and 38 entries processed > tests/xfs/038.out:xfsrestore: 3 directories and 38 entries processed > tests/xfs/039.out:xfsrestore: 3 directories and 38 entries processed > tests/xfs/043.out:xfsrestore: 3 directories and 38 entries processed > tests/xfs/046.out:xfsrestore: 3 directories and 10 entries processed > tests/xfs/055.out:xfsrestore: 3 directories and 38 entries processed > tests/xfs/056.out:xfsrestore: 7 directories and 11 entries processed > tests/xfs/060.out:xfsrestore: 3 directories and 41 entries processed > tests/xfs/061.out:xfsrestore: 7 directories and 11 entries processed > tests/xfs/063.out:xfsrestore: 4 directories and 21 entries processed > ..... > > Really, I don't see a need to do anything else than avoid the > fsstress ops that caused the change of behaviour. All the other > xfsdump/restore tests do file and directory tree validations, so > they are going to catch any regression on that side of things. This > test just exercises iteration of various APIs that we've broken in > the past... > >> Currently I have a reworked version of this patch that runs >> fsstress, measures the number of directories and inodes in >> $dump_dir, then programmatically compares that to whatever >> xfsrestore tells us it restored. This ought to be enough that we >> can create a sufficiently messy filesystem with whatever sequence >> of syscalls we want, and make sure that dump/restore actually work >> on them. >> >> First we run fsstress, then we count the number of dirs, the >> number of fs objects, take a snapshot of the 'find .' output, and >> md5sum every file in the dump directory. >> >> If fsstress creates fewer than 100 dirs or 600 inodes, we fail the >> test because that wasn't enough. >> >> If bulkstat fails to iterate all the inodes, restore's output will >> reflect fewer files than was expected. >> >> If dump fails to generate a full dump, restore's output will >> reflect fewer files than was expected. >> >> If restore fails to restore the full dump, restore's output will >> reflect fewer files than was expected. >> >> If the restore output doesn't reflect the number of dirs/inodes we >> counted at the beginning, we fail the test. >> >> If the 'find .' output of the restored dir doesn't match the >> original, we fail the test. >> >> If the md5sum -c output shows corrupt files, we fail the test. >> >> So now I really have no idea -- is that enough to check that >> everything works? I felt like it does, but given all the back and >> forth now I'm wondering if even this is enough. > > What did I say about not wanting to overcomplicate the problem and > the solution? :/ > > Folks, I don't say "leave it alone, it's fine" without a good > reason. If you've never tried to debug xfsdump or xfsrestore, and > you aren't familiar with the ancient xfsdump and restore unit tests > that were written before anyone here was working on linux, then > don't suggest we rewrite them to make them nicer. Their value is in > the fact they've been around almost entirely unchanged for 15 years > and they still catch bugs.... > > Make whatever changes are necessary to keep them running > exactly as they are and don't change them unless xfsdump/restore > testing requires them to be changed. > >> (Yeah, I'm frustrated because the fsstress additions have been >> very helpful at flushing out more reflink bugs and I feel like I'm >> making very little progress on this xfs/068 thing. Sorry.) > > Well, I thought it was all sorted until people started suggesting we > do crazy things like you've now gone and done. > Please read this test description out loud and tell yourself this is a good sanity test: 1. perform a $series of fs operations 2. backup & restore 3. count the number of file&dir 4. makes sure the $count didn't change across $kernel & user $tools releases Now developer changes the test variable $series, so the result $count obviously changes, so developer changes the expected count WITHOUT verifying that the expected count is actually correct. There is NOTHING in the current test (before Darrick's "crazy" changes) that promise we are not actually validating that bulkstat *is* broken. Amir. -- To unsubscribe from this list: send the line "unsubscribe fstests" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
diff --git a/tests/xfs/068 b/tests/xfs/068 index 7151e28..f95a539 100755 --- a/tests/xfs/068 +++ b/tests/xfs/068 @@ -43,6 +43,14 @@ trap "rm -rf $tmp.*; exit \$status" 0 1 2 3 15 _supported_fs xfs _supported_os Linux +# Remove fsstress commands that aren't supported on all xfs configs +if $FSSTRESS_PROG | grep -q clonerange; then + FSSTRESS_AVOID="-f clonerange=0 $FSSTRESS_AVOID" +fi +if $FSSTRESS_PROG | grep -q deduperange; then + FSSTRESS_AVOID="-f deduperange=0 $FSSTRESS_AVOID" +fi + _create_dumpdir_stress_num 4096 _do_dump_restore