Message ID | 20131003031738.GA27418@kmo-pixel (mailing list archive) |
---|---|
State | Awaiting Upstream, archived |
Delegated to: | Mike Snitzer |
Headers | show |
On Wed, Oct 02 2013 at 11:17pm -0400, Kent Overstreet <kmo@daterainc.com> wrote: > On Sat, Sep 28, 2013 at 12:59:09AM -0400, Mike Snitzer wrote: > > Hey Kent, > > > > I haven't been able to pinpoint the issue yet, but using your for-jens > > branch, if I create a dm-thin volume with this lvm command: > > lvcreate -L20G -V20G -T vg/pool --name thinlv > > > > and try to format /dev/vg/thinlv with XFS the kernel warns and then > > hangs with the following: > > Thanks, for whatever reason ext4 wasn't tickling that codepath but xfs > does. I folded the fix into my for-jens branch, here's what changed: > > diff --git a/drivers/md/dm.c b/drivers/md/dm.c > index b60b350..79cee1a 100644 > --- a/drivers/md/dm.c > +++ b/drivers/md/dm.c > @@ -1098,7 +1098,7 @@ static void clone_bio(struct dm_target_io *tio, struct bio *bio, > bio_integrity_clone(clone, bio, GFP_NOIO); > > bio_advance(clone, to_bytes(sector - clone->bi_iter.bi_sector)); > - bio->bi_iter.bi_size = to_bytes(len); > + clone->bi_iter.bi_size = to_bytes(len); > > if (bio_integrity(bio)) > bio_integrity_trim(clone, 0, len); The above change matches my findings during review (I clearly should've shared it sooner) > @@ -1267,7 +1267,7 @@ static int __split_and_process_non_flush(struct clone_info *ci) > if (!dm_target_is_valid(ti)) > return -EIO; > > - len = min_t(unsigned, max_io_len(ci->sector, ti), bio_sectors(bio)); > + len = min_t(sector_t, max_io_len(ci->sector, ti), ci->sector_count); > > __clone_and_map_data_bio(ci, ti, ci->sector, len); > Great, thanks for finding this, I'll test and review the code further. -- dm-devel mailing list dm-devel@redhat.com https://www.redhat.com/mailman/listinfo/dm-devel
On Wed, Oct 02, 2013 at 11:23:21PM -0400, Mike Snitzer wrote:
> Great, thanks for finding this, I'll test and review the code further.
Cool - let me know if you find anything else (how thoroughly do you
think you've tested it so far?)
--
dm-devel mailing list
dm-devel@redhat.com
https://www.redhat.com/mailman/listinfo/dm-devel
On Thu, Oct 03 2013 at 5:45pm -0400, Kent Overstreet <kmo@daterainc.com> wrote: > On Wed, Oct 02, 2013 at 11:23:21PM -0400, Mike Snitzer wrote: > > Great, thanks for finding this, I'll test and review the code further. > > Cool - let me know if you find anything else (how thoroughly do you > think you've tested it so far?) Not very, the issue you resolved was a blocker for DM. But I haven't even verified your fix yet though... will do tomorrow with more extensive testing using the dm-thinp and dm-cache test-suite. -- dm-devel mailing list dm-devel@redhat.com https://www.redhat.com/mailman/listinfo/dm-devel
On Thu, Oct 03 2013 at 6:50pm -0400, Mike Snitzer <snitzer@redhat.com> wrote: > On Thu, Oct 03 2013 at 5:45pm -0400, > Kent Overstreet <kmo@daterainc.com> wrote: > > > On Wed, Oct 02, 2013 at 11:23:21PM -0400, Mike Snitzer wrote: > > > Great, thanks for finding this, I'll test and review the code further. > > > > Cool - let me know if you find anything else (how thoroughly do you > > think you've tested it so far?) > > Not very, the issue you resolved was a blocker for DM. But I haven't > even verified your fix yet though... will do tomorrow with more > extensive testing using the dm-thinp and dm-cache test-suite. With your latest fix I was able to create a thin device and format with XFS. Unfortunately, when I tried to run the thinp-test-suite the very first BasicTests test (test_dd_benchmark) fails -- need to look closer but it would seem to me the thinp saved bio_endio path isn't happy. We likely need an appropriately placed atomic_inc(&bio->bi_remaining); like you did in dm-cache-target.c ------------[ cut here ]------------ kernel BUG at fs/bio.c:1722! invalid opcode: 0000 [#1] SMP Modules linked in: dm_cache_cleaner dm_cache_mq dm_cache dm_thin_pool dm_bio_prison dm_persistent_data dm_bufio dm_mod xfs exportfs libcrc32c ebtable_nat ebtables xt_C HECKSUM iptable_mangle bridge autofs4 target_core_iblock target_core_file target_core_pscsi target_core_mod configfs bnx2fc fcoe 8021q libfcoe libfc garp stp llc scsi_ transport_fc scsi_tgt sunrpc cpufreq_ondemand ipt_REJECT nf_conntrack_ipv4 nf_defrag_ipv4 iptable_filter ip_tables ip6t_REJECT nf_conntrack_ipv6 nf_defrag_ipv6 xt_stat e nf_conntrack ip6table_filter ip6_tables bnx2i cnic uio ipv6 cxgb4i cxgb4 cxgb3i libcxgbi cxgb3 iscsi_tcp libiscsi_tcp libiscsi scsi_transport_iscsi vhost_net macvtap macvlan vhost tun kvm_intel kvm iTCO_wdt iTCO_vendor_support microcode i2c_i801 lpc_ich mfd_core igb i2c_algo_bit i2c_core i7core_edac edac_core ixgbe dca ptp pps_cor e mdio sg ses enclosure acpi_cpufreq freq_table ext4 jbd2 mbcache sr_mod cdrom pata_acpi ata_generic ata_piix sd_mod crc_t10dif crct10dif_common megaraid_sas [last unl oaded: dm_cache_mq] CPU: 0 PID: 342 Comm: kworker/u24:7 Tainted: G W 3.12.0-rc2.snitm+ #76 Hardware name: FUJITSU PRIMERGY RX300 S6 /D2619, BIOS 6.00 Rev. 1.10.2619.N1 05/24/2011 Workqueue: dm-thin do_worker [dm_thin_pool] task: ffff88032cd88ab0 ti: ffff88032c5e6000 task.ti: ffff88032c5e6000 RIP: 0010:[<ffffffff811b1464>] [<ffffffff811b1464>] bio_endio+0x74/0x80 RSP: 0018:ffff88032c5e7d48 EFLAGS: 00010246 RAX: ffff88032f10a784 RBX: ffff88032d097450 RCX: 00000000ffffffff RDX: 0000000000000000 RSI: 0000000000000000 RDI: ffff88032f10a740 RBP: ffff88032c5e7d48 R08: ffff88033fc12d00 R09: 0000000000000000 R10: 0000000000000000 R11: 0000000000000002 R12: ffff88032f10a740 R13: ffff88032def5880 R14: ffff88032d0c7c00 R15: ffff88032d457a05 FS: 0000000000000000(0000) GS:ffff88033fc00000(0000) knlGS:0000000000000000 CS: 0010 DS: 0000 ES: 0000 CR0: 000000008005003b CR2: 0000003bfea13000 CR3: 0000000001a0b000 CR4: 00000000000007f0 Stack: ffff88032c5e7d98 ffffffffa05f2ef9 ffff88032c5e7da8 0000000000000000 ffff88032d0c7c00 ffff88032c5e7da8 ffff88032d0c7d98 ffff88032d0974a8 ffff88032d0c7d14 ffff88032d457a05 ffff88032c5e7dd8 ffffffffa05f2ba7 Call Trace: [<ffffffffa05f2ef9>] process_prepared_mapping+0x79/0x150 [dm_thin_pool] [<ffffffffa05f2ba7>] process_prepared+0x87/0xa0 [dm_thin_pool] [<ffffffffa05f58f3>] do_worker+0x33/0x60 [dm_thin_pool] [<ffffffff81067862>] process_one_work+0x182/0x3b0 [<ffffffff81068c80>] worker_thread+0x120/0x3a0 [<ffffffff81068b60>] ? manage_workers+0x160/0x160 [<ffffffff8106eace>] kthread+0xce/0xe0 [<ffffffff8106ea00>] ? kthread_freezable_should_stop+0x70/0x70 [<ffffffff8152b1ac>] ret_from_fork+0x7c/0xb0 [<ffffffff8106ea00>] ? kthread_freezable_should_stop+0x70/0x70 Code: 1f 84 00 00 00 00 00 48 8b 57 10 83 e2 01 0f 44 f1 eb cd 0f 1f 40 00 48 8b 7f 50 48 85 ff 74 dd 8b 57 44 48 8d 47 44 85 d2 7f ac <0f> 0b eb fe 0f 1f 84 00 00 00 00 00 55 48 89 e5 66 66 66 66 90 RIP [<ffffffff811b1464>] bio_endio+0x74/0x80 RSP <ffff88032c5e7d48> ---[ end trace acb5a7d638591b7b ]--- -- dm-devel mailing list dm-devel@redhat.com https://www.redhat.com/mailman/listinfo/dm-devel
diff --git a/drivers/md/dm.c b/drivers/md/dm.c index b60b350..79cee1a 100644 --- a/drivers/md/dm.c +++ b/drivers/md/dm.c @@ -1098,7 +1098,7 @@ static void clone_bio(struct dm_target_io *tio, struct bio *bio, bio_integrity_clone(clone, bio, GFP_NOIO); bio_advance(clone, to_bytes(sector - clone->bi_iter.bi_sector)); - bio->bi_iter.bi_size = to_bytes(len); + clone->bi_iter.bi_size = to_bytes(len); if (bio_integrity(bio)) bio_integrity_trim(clone, 0, len); @@ -1267,7 +1267,7 @@ static int __split_and_process_non_flush(struct clone_info *ci) if (!dm_target_is_valid(ti)) return -EIO; - len = min_t(unsigned, max_io_len(ci->sector, ti), bio_sectors(bio)); + len = min_t(sector_t, max_io_len(ci->sector, ti), ci->sector_count); __clone_and_map_data_bio(ci, ti, ci->sector, len);