@@ -58,31 +58,37 @@ test_clone_and_read_compressed_extent()
_scratch_mkfs >>$seqres.full 2>&1
_scratch_mount $mount_opts
- # Create our test file with a single extent of 64Kb that is going to be
- # compressed no matter which compression algorithm is used (zlib/lzo).
- $XFS_IO_PROG -f -c "pwrite -S 0xaa 0K 64K" \
- $SCRATCH_MNT/foo | _filter_xfs_io
-
+ PAGE_SIZE=$(get_page_size)
+
+ # Create our test file with 16 pages worth of data in a single extent
+ # that is going to be compressed no matter which compression algorithm
+ # is used (zlib/lzo).
+ $XFS_IO_PROG -f -c "pwrite -S 0xaa 0K $((16 * $PAGE_SIZE))" \
+ $SCRATCH_MNT/foo | _filter_xfs_io_pages_modified
+
# Now clone the compressed extent into an adjacent file offset.
- $CLONER_PROG -s 0 -d $((64 * 1024)) -l $((64 * 1024)) \
+ $CLONER_PROG -s 0 -d $((16 * $PAGE_SIZE)) -l $((16 * $PAGE_SIZE)) \
$SCRATCH_MNT/foo $SCRATCH_MNT/foo
- echo "File digest before unmount:"
- md5sum $SCRATCH_MNT/foo | _filter_scratch
+ orig_hash=$(md5sum $SCRATCH_MNT/foo | cut -f 1 -d ' ')
# Remount the fs or clear the page cache to trigger the bug in btrfs.
- # Because the extent has an uncompressed length that is a multiple of
- # 16 pages, all the pages belonging to the second range of the file
- # (64K to 128K), which points to the same extent as the first range
- # (0K to 64K), had their contents full of zeroes instead of the byte
- # 0xaa. This was a bug exclusively in the read path of compressed
- # extents, the correct data was stored on disk, btrfs just failed to
- # fill in the pages correctly.
+ # Because the extent has an uncompressed length that is a multiple of 16
+ # pages, all the pages belonging to the second range of the file that is
+ # mapped by the page index range [16, 31], which points to the same
+ # extent as the first file range mapped by the page index range [0, 15],
+ # had their contents full of zeroes instead of the byte 0xaa. This was a
+ # bug exclusively in the read path of compressed extents, the correct
+ # data was stored on disk, btrfs just failed to fill in the pages
+ # correctly.
_scratch_remount
- echo "File digest after remount:"
- # Must match the digest we got before.
- md5sum $SCRATCH_MNT/foo | _filter_scratch
+ hash=$(md5sum $SCRATCH_MNT/foo | cut -f 1 -d ' ')
+
+ if [ $orig_hash != $hash ]; then
+ echo "Read operation failed on $SCRATCH_MNT/foo: "\
+ "Mimatching hash values detected."
+ fi
}
echo -e "\nTesting with zlib compression..."
@@ -1,17 +1,7 @@
QA output created by 106
Testing with zlib compression...
-wrote 65536/65536 bytes at offset 0
-XXX Bytes, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec)
-File digest before unmount:
-be68df46e3cf60b559376a35f9fbb05d SCRATCH_MNT/foo
-File digest after remount:
-be68df46e3cf60b559376a35f9fbb05d SCRATCH_MNT/foo
+Pages modified: [0 - 15]
Testing with lzo compression...
-wrote 65536/65536 bytes at offset 0
-XXX Bytes, X ops; XX:XX:XX.X (XXX YYY/sec and XXX ops/sec)
-File digest before unmount:
-be68df46e3cf60b559376a35f9fbb05d SCRATCH_MNT/foo
-File digest after remount:
-be68df46e3cf60b559376a35f9fbb05d SCRATCH_MNT/foo
+Pages modified: [0 - 15]
This commit makes use of the new _filter_xfs_io_pages_modified filtering function to print information in terms of file blocks rather than file offset. Signed-off-by: Chandan Rajendra <chandan@linux.vnet.ibm.com> --- tests/btrfs/106 | 42 ++++++++++++++++++++++++------------------ tests/btrfs/106.out | 14 ++------------ 2 files changed, 26 insertions(+), 30 deletions(-)