Message ID | 20230327084103.21601-7-anuj20.g@samsung.com (mailing list archive) |
---|---|
State | New, archived |
Headers | show |
Series | [v8,1/9] block: Introduce queue limits for copy-offload support | expand |
Hi Anuj, Thank you for the patch! Perhaps something to improve: [auto build test WARNING on axboe-block/for-next] [also build test WARNING on linus/master v6.3-rc4 next-20230329] [cannot apply to device-mapper-dm/for-next] [If your patch is applied to the wrong git tree, kindly drop us a note. And when submitting patch, we suggest to use '--base' as documented in https://git-scm.com/docs/git-format-patch#_base_tree_information] url: https://github.com/intel-lab-lkp/linux/commits/Anuj-Gupta/block-Add-copy-offload-support-infrastructure/20230329-162018 base: https://git.kernel.org/pub/scm/linux/kernel/git/axboe/linux-block.git for-next patch link: https://lore.kernel.org/r/20230327084103.21601-7-anuj20.g%40samsung.com patch subject: [PATCH v8 6/9] nvmet: add copy command support for bdev and file ns config: arm64-randconfig-s041-20230329 (https://download.01.org/0day-ci/archive/20230329/202303292148.Pbx4mDpS-lkp@intel.com/config) compiler: aarch64-linux-gcc (GCC) 12.1.0 reproduce: wget https://raw.githubusercontent.com/intel/lkp-tests/master/sbin/make.cross -O ~/bin/make.cross chmod +x ~/bin/make.cross # apt-get install sparse # sparse version: v0.6.4-39-gce1a6720-dirty # https://github.com/intel-lab-lkp/linux/commit/f846a8ac40882d9d42532e9e2b43560650ef8510 git remote add linux-review https://github.com/intel-lab-lkp/linux git fetch --no-tags linux-review Anuj-Gupta/block-Add-copy-offload-support-infrastructure/20230329-162018 git checkout f846a8ac40882d9d42532e9e2b43560650ef8510 # save the config file mkdir build_dir && cp config build_dir/.config COMPILER_INSTALL_PATH=$HOME/0day COMPILER=gcc-12.1.0 make.cross C=1 CF='-fdiagnostic-prefix -D__CHECK_ENDIAN__' O=build_dir ARCH=arm64 olddefconfig COMPILER_INSTALL_PATH=$HOME/0day COMPILER=gcc-12.1.0 make.cross C=1 CF='-fdiagnostic-prefix -D__CHECK_ENDIAN__' O=build_dir ARCH=arm64 SHELL=/bin/bash drivers/nvme/target/ If you fix the issue, kindly add following tag where applicable | Reported-by: kernel test robot <lkp@intel.com> | Link: https://lore.kernel.org/oe-kbuild-all/202303292148.Pbx4mDpS-lkp@intel.com/ sparse warnings: (new ones prefixed by >>) >> drivers/nvme/target/io-cmd-bdev.c:55:27: sparse: sparse: incorrect type in argument 1 (different base types) @@ expected unsigned int [usertype] val @@ got restricted __le16 [usertype] mssrl @@ drivers/nvme/target/io-cmd-bdev.c:55:27: sparse: expected unsigned int [usertype] val drivers/nvme/target/io-cmd-bdev.c:55:27: sparse: got restricted __le16 [usertype] mssrl >> drivers/nvme/target/io-cmd-bdev.c:55:27: sparse: sparse: cast from restricted __le16 >> drivers/nvme/target/io-cmd-bdev.c:55:27: sparse: sparse: cast from restricted __le16 >> drivers/nvme/target/io-cmd-bdev.c:55:27: sparse: sparse: cast from restricted __le16 >> drivers/nvme/target/io-cmd-bdev.c:55:27: sparse: sparse: cast from restricted __le16 drivers/nvme/target/io-cmd-bdev.c:57:29: sparse: sparse: cast from restricted __le16 drivers/nvme/target/io-cmd-bdev.c:60:27: sparse: sparse: incorrect type in argument 1 (different base types) @@ expected unsigned int [usertype] val @@ got restricted __le16 [usertype] mssrl @@ drivers/nvme/target/io-cmd-bdev.c:60:27: sparse: expected unsigned int [usertype] val drivers/nvme/target/io-cmd-bdev.c:60:27: sparse: got restricted __le16 [usertype] mssrl drivers/nvme/target/io-cmd-bdev.c:60:27: sparse: sparse: cast from restricted __le16 drivers/nvme/target/io-cmd-bdev.c:60:27: sparse: sparse: cast from restricted __le16 drivers/nvme/target/io-cmd-bdev.c:60:27: sparse: sparse: cast from restricted __le16 drivers/nvme/target/io-cmd-bdev.c:60:27: sparse: sparse: cast from restricted __le16 vim +55 drivers/nvme/target/io-cmd-bdev.c 12 13 void nvmet_bdev_set_limits(struct block_device *bdev, struct nvme_id_ns *id) 14 { 15 /* Logical blocks per physical block, 0's based. */ 16 const __le16 lpp0b = to0based(bdev_physical_block_size(bdev) / 17 bdev_logical_block_size(bdev)); 18 19 /* 20 * For NVMe 1.2 and later, bit 1 indicates that the fields NAWUN, 21 * NAWUPF, and NACWU are defined for this namespace and should be 22 * used by the host for this namespace instead of the AWUN, AWUPF, 23 * and ACWU fields in the Identify Controller data structure. If 24 * any of these fields are zero that means that the corresponding 25 * field from the identify controller data structure should be used. 26 */ 27 id->nsfeat |= 1 << 1; 28 id->nawun = lpp0b; 29 id->nawupf = lpp0b; 30 id->nacwu = lpp0b; 31 32 /* 33 * Bit 4 indicates that the fields NPWG, NPWA, NPDG, NPDA, and 34 * NOWS are defined for this namespace and should be used by 35 * the host for I/O optimization. 36 */ 37 id->nsfeat |= 1 << 4; 38 /* NPWG = Namespace Preferred Write Granularity. 0's based */ 39 id->npwg = lpp0b; 40 /* NPWA = Namespace Preferred Write Alignment. 0's based */ 41 id->npwa = id->npwg; 42 /* NPDG = Namespace Preferred Deallocate Granularity. 0's based */ 43 id->npdg = to0based(bdev_discard_granularity(bdev) / 44 bdev_logical_block_size(bdev)); 45 /* NPDG = Namespace Preferred Deallocate Alignment */ 46 id->npda = id->npdg; 47 /* NOWS = Namespace Optimal Write Size */ 48 id->nows = to0based(bdev_io_opt(bdev) / bdev_logical_block_size(bdev)); 49 50 /*Copy limits*/ 51 if (bdev_max_copy_sectors(bdev)) { 52 id->msrc = id->msrc; 53 id->mssrl = cpu_to_le16((bdev_max_copy_sectors(bdev) << 54 SECTOR_SHIFT) / bdev_logical_block_size(bdev)); > 55 id->mcl = cpu_to_le32(id->mssrl); 56 } else { 57 id->msrc = (u8)to0based(BIO_MAX_VECS - 1); 58 id->mssrl = cpu_to_le16((BIO_MAX_VECS << PAGE_SHIFT) / 59 bdev_logical_block_size(bdev)); 60 id->mcl = cpu_to_le32(id->mssrl); 61 } 62 } 63
Hi Anuj, Thank you for the patch! Perhaps something to improve: [auto build test WARNING on axboe-block/for-next] [also build test WARNING on linus/master v6.3-rc4 next-20230329] [cannot apply to device-mapper-dm/for-next] [If your patch is applied to the wrong git tree, kindly drop us a note. And when submitting patch, we suggest to use '--base' as documented in https://git-scm.com/docs/git-format-patch#_base_tree_information] url: https://github.com/intel-lab-lkp/linux/commits/Anuj-Gupta/block-Add-copy-offload-support-infrastructure/20230329-162018 base: https://git.kernel.org/pub/scm/linux/kernel/git/axboe/linux-block.git for-next patch link: https://lore.kernel.org/r/20230327084103.21601-7-anuj20.g%40samsung.com patch subject: [PATCH v8 6/9] nvmet: add copy command support for bdev and file ns config: arm64-randconfig-s041-20230329 (https://download.01.org/0day-ci/archive/20230330/202303300238.vmt9ne37-lkp@intel.com/config) compiler: aarch64-linux-gcc (GCC) 12.1.0 reproduce: wget https://raw.githubusercontent.com/intel/lkp-tests/master/sbin/make.cross -O ~/bin/make.cross chmod +x ~/bin/make.cross # apt-get install sparse # sparse version: v0.6.4-39-gce1a6720-dirty # https://github.com/intel-lab-lkp/linux/commit/f846a8ac40882d9d42532e9e2b43560650ef8510 git remote add linux-review https://github.com/intel-lab-lkp/linux git fetch --no-tags linux-review Anuj-Gupta/block-Add-copy-offload-support-infrastructure/20230329-162018 git checkout f846a8ac40882d9d42532e9e2b43560650ef8510 # save the config file mkdir build_dir && cp config build_dir/.config COMPILER_INSTALL_PATH=$HOME/0day COMPILER=gcc-12.1.0 make.cross C=1 CF='-fdiagnostic-prefix -D__CHECK_ENDIAN__' O=build_dir ARCH=arm64 olddefconfig COMPILER_INSTALL_PATH=$HOME/0day COMPILER=gcc-12.1.0 make.cross C=1 CF='-fdiagnostic-prefix -D__CHECK_ENDIAN__' O=build_dir ARCH=arm64 SHELL=/bin/bash drivers/nvme/target/ If you fix the issue, kindly add following tag where applicable | Reported-by: kernel test robot <lkp@intel.com> | Link: https://lore.kernel.org/oe-kbuild-all/202303300238.vmt9ne37-lkp@intel.com/ sparse warnings: (new ones prefixed by >>) >> drivers/nvme/target/admin-cmd.c:539:29: sparse: sparse: cast from restricted __le16 vim +539 drivers/nvme/target/admin-cmd.c 490 491 static void nvmet_execute_identify_ns(struct nvmet_req *req) 492 { 493 struct nvme_id_ns *id; 494 u16 status; 495 496 if (le32_to_cpu(req->cmd->identify.nsid) == NVME_NSID_ALL) { 497 req->error_loc = offsetof(struct nvme_identify, nsid); 498 status = NVME_SC_INVALID_NS | NVME_SC_DNR; 499 goto out; 500 } 501 502 id = kzalloc(sizeof(*id), GFP_KERNEL); 503 if (!id) { 504 status = NVME_SC_INTERNAL; 505 goto out; 506 } 507 508 /* return an all zeroed buffer if we can't find an active namespace */ 509 status = nvmet_req_find_ns(req); 510 if (status) { 511 status = 0; 512 goto done; 513 } 514 515 if (nvmet_ns_revalidate(req->ns)) { 516 mutex_lock(&req->ns->subsys->lock); 517 nvmet_ns_changed(req->ns->subsys, req->ns->nsid); 518 mutex_unlock(&req->ns->subsys->lock); 519 } 520 521 /* 522 * nuse = ncap = nsze isn't always true, but we have no way to find 523 * that out from the underlying device. 524 */ 525 id->ncap = id->nsze = 526 cpu_to_le64(req->ns->size >> req->ns->blksize_shift); 527 switch (req->port->ana_state[req->ns->anagrpid]) { 528 case NVME_ANA_INACCESSIBLE: 529 case NVME_ANA_PERSISTENT_LOSS: 530 break; 531 default: 532 id->nuse = id->nsze; 533 break; 534 } 535 536 if (req->ns->bdev) 537 nvmet_bdev_set_limits(req->ns->bdev, id); 538 else { > 539 id->msrc = (u8)to0based(BIO_MAX_VECS - 1); 540 id->mssrl = cpu_to_le16(BIO_MAX_VECS << 541 (PAGE_SHIFT - SECTOR_SHIFT)); 542 id->mcl = cpu_to_le32(le16_to_cpu(id->mssrl)); 543 } 544 545 /* 546 * We just provide a single LBA format that matches what the 547 * underlying device reports. 548 */ 549 id->nlbaf = 0; 550 id->flbas = 0; 551 552 /* 553 * Our namespace might always be shared. Not just with other 554 * controllers, but also with any other user of the block device. 555 */ 556 id->nmic = NVME_NS_NMIC_SHARED; 557 id->anagrpid = cpu_to_le32(req->ns->anagrpid); 558 559 memcpy(&id->nguid, &req->ns->nguid, sizeof(id->nguid)); 560 561 id->lbaf[0].ds = req->ns->blksize_shift; 562 563 if (req->sq->ctrl->pi_support && nvmet_ns_has_pi(req->ns)) { 564 id->dpc = NVME_NS_DPC_PI_FIRST | NVME_NS_DPC_PI_LAST | 565 NVME_NS_DPC_PI_TYPE1 | NVME_NS_DPC_PI_TYPE2 | 566 NVME_NS_DPC_PI_TYPE3; 567 id->mc = NVME_MC_EXTENDED_LBA; 568 id->dps = req->ns->pi_type; 569 id->flbas = NVME_NS_FLBAS_META_EXT; 570 id->lbaf[0].ms = cpu_to_le16(req->ns->metadata_size); 571 } 572 573 if (req->ns->readonly) 574 id->nsattr |= NVME_NS_ATTR_RO; 575 done: 576 if (!status) 577 status = nvmet_copy_to_sgl(req, 0, id, sizeof(*id)); 578 579 kfree(id); 580 out: 581 nvmet_req_complete(req, status); 582 } 583
diff --git a/drivers/nvme/target/admin-cmd.c b/drivers/nvme/target/admin-cmd.c index 80099df37314..978786ec6a9e 100644 --- a/drivers/nvme/target/admin-cmd.c +++ b/drivers/nvme/target/admin-cmd.c @@ -433,8 +433,7 @@ static void nvmet_execute_identify_ctrl(struct nvmet_req *req) id->nn = cpu_to_le32(NVMET_MAX_NAMESPACES); id->mnan = cpu_to_le32(NVMET_MAX_NAMESPACES); id->oncs = cpu_to_le16(NVME_CTRL_ONCS_DSM | - NVME_CTRL_ONCS_WRITE_ZEROES); - + NVME_CTRL_ONCS_WRITE_ZEROES | NVME_CTRL_ONCS_COPY); /* XXX: don't report vwc if the underlying device is write through */ id->vwc = NVME_CTRL_VWC_PRESENT; @@ -536,6 +535,12 @@ static void nvmet_execute_identify_ns(struct nvmet_req *req) if (req->ns->bdev) nvmet_bdev_set_limits(req->ns->bdev, id); + else { + id->msrc = (u8)to0based(BIO_MAX_VECS - 1); + id->mssrl = cpu_to_le16(BIO_MAX_VECS << + (PAGE_SHIFT - SECTOR_SHIFT)); + id->mcl = cpu_to_le32(le16_to_cpu(id->mssrl)); + } /* * We just provide a single LBA format that matches what the diff --git a/drivers/nvme/target/io-cmd-bdev.c b/drivers/nvme/target/io-cmd-bdev.c index c2d6cea0236b..0af273097aa4 100644 --- a/drivers/nvme/target/io-cmd-bdev.c +++ b/drivers/nvme/target/io-cmd-bdev.c @@ -46,6 +46,19 @@ void nvmet_bdev_set_limits(struct block_device *bdev, struct nvme_id_ns *id) id->npda = id->npdg; /* NOWS = Namespace Optimal Write Size */ id->nows = to0based(bdev_io_opt(bdev) / bdev_logical_block_size(bdev)); + + /*Copy limits*/ + if (bdev_max_copy_sectors(bdev)) { + id->msrc = id->msrc; + id->mssrl = cpu_to_le16((bdev_max_copy_sectors(bdev) << + SECTOR_SHIFT) / bdev_logical_block_size(bdev)); + id->mcl = cpu_to_le32(id->mssrl); + } else { + id->msrc = (u8)to0based(BIO_MAX_VECS - 1); + id->mssrl = cpu_to_le16((BIO_MAX_VECS << PAGE_SHIFT) / + bdev_logical_block_size(bdev)); + id->mcl = cpu_to_le32(id->mssrl); + } } void nvmet_bdev_ns_disable(struct nvmet_ns *ns) @@ -184,6 +197,19 @@ static void nvmet_bio_done(struct bio *bio) nvmet_req_bio_put(req, bio); } +static void nvmet_bdev_copy_end_io(void *private, int comp_len) +{ + struct nvmet_req *req = (struct nvmet_req *)private; + + if (comp_len == req->copy_len) { + req->cqe->result.u32 = cpu_to_le32(1); + nvmet_req_complete(req, errno_to_nvme_status(req, 0)); + } else { + req->cqe->result.u32 = cpu_to_le32(0); + nvmet_req_complete(req, blk_to_nvme_status(req, BLK_STS_IOERR)); + } +} + #ifdef CONFIG_BLK_DEV_INTEGRITY static int nvmet_bdev_alloc_bip(struct nvmet_req *req, struct bio *bio, struct sg_mapping_iter *miter) @@ -450,6 +476,34 @@ static void nvmet_bdev_execute_write_zeroes(struct nvmet_req *req) } } +/* At present we handle only one range entry */ +static void nvmet_bdev_execute_copy(struct nvmet_req *req) +{ + struct nvme_copy_range range; + struct nvme_command *cmnd = req->cmd; + int ret; + + + ret = nvmet_copy_from_sgl(req, 0, &range, sizeof(range)); + if (ret) + goto out; + + ret = blkdev_issue_copy(req->ns->bdev, + le64_to_cpu(cmnd->copy.sdlba) << req->ns->blksize_shift, + req->ns->bdev, + le64_to_cpu(range.slba) << req->ns->blksize_shift, + (le16_to_cpu(range.nlb) + 1) << req->ns->blksize_shift, + nvmet_bdev_copy_end_io, (void *)req, GFP_KERNEL); + if (ret) { + req->cqe->result.u32 = cpu_to_le32(0); + nvmet_req_complete(req, blk_to_nvme_status(req, BLK_STS_IOERR)); + } + + return; +out: + nvmet_req_complete(req, errno_to_nvme_status(req, ret)); +} + u16 nvmet_bdev_parse_io_cmd(struct nvmet_req *req) { switch (req->cmd->common.opcode) { @@ -468,6 +522,10 @@ u16 nvmet_bdev_parse_io_cmd(struct nvmet_req *req) case nvme_cmd_write_zeroes: req->execute = nvmet_bdev_execute_write_zeroes; return 0; + case nvme_cmd_copy: + req->execute = nvmet_bdev_execute_copy; + return 0; + default: return nvmet_report_invalid_opcode(req); } diff --git a/drivers/nvme/target/io-cmd-file.c b/drivers/nvme/target/io-cmd-file.c index 2d068439b129..69f198ecec77 100644 --- a/drivers/nvme/target/io-cmd-file.c +++ b/drivers/nvme/target/io-cmd-file.c @@ -322,6 +322,49 @@ static void nvmet_file_dsm_work(struct work_struct *w) } } +static void nvmet_file_copy_work(struct work_struct *w) +{ + struct nvmet_req *req = container_of(w, struct nvmet_req, f.work); + int nr_range; + loff_t pos; + struct nvme_command *cmnd = req->cmd; + int ret = 0, len = 0, src, id; + + nr_range = cmnd->copy.nr_range + 1; + pos = le64_to_cpu(req->cmd->copy.sdlba) << req->ns->blksize_shift; + if (unlikely(pos + req->transfer_len > req->ns->size)) { + nvmet_req_complete(req, errno_to_nvme_status(req, -ENOSPC)); + return; + } + + for (id = 0 ; id < nr_range; id++) { + struct nvme_copy_range range; + + ret = nvmet_copy_from_sgl(req, id * sizeof(range), &range, + sizeof(range)); + if (ret) + goto out; + + len = (le16_to_cpu(range.nlb) + 1) << (req->ns->blksize_shift); + src = (le64_to_cpu(range.slba) << (req->ns->blksize_shift)); + ret = vfs_copy_file_range(req->ns->file, src, req->ns->file, + pos, len, 0); +out: + if (ret != len) { + pos += ret; + req->cqe->result.u32 = cpu_to_le32(id); + nvmet_req_complete(req, ret < 0 ? + errno_to_nvme_status(req, ret) : + errno_to_nvme_status(req, -EIO)); + return; + + } else + pos += len; + } + + nvmet_req_complete(req, 0); + +} static void nvmet_file_execute_dsm(struct nvmet_req *req) { if (!nvmet_check_data_len_lte(req, nvmet_dsm_len(req))) @@ -330,6 +373,12 @@ static void nvmet_file_execute_dsm(struct nvmet_req *req) queue_work(nvmet_wq, &req->f.work); } +static void nvmet_file_execute_copy(struct nvmet_req *req) +{ + INIT_WORK(&req->f.work, nvmet_file_copy_work); + queue_work(nvmet_wq, &req->f.work); +} + static void nvmet_file_write_zeroes_work(struct work_struct *w) { struct nvmet_req *req = container_of(w, struct nvmet_req, f.work); @@ -376,6 +425,9 @@ u16 nvmet_file_parse_io_cmd(struct nvmet_req *req) case nvme_cmd_write_zeroes: req->execute = nvmet_file_execute_write_zeroes; return 0; + case nvme_cmd_copy: + req->execute = nvmet_file_execute_copy; + return 0; default: return nvmet_report_invalid_opcode(req); } diff --git a/drivers/nvme/target/loop.c b/drivers/nvme/target/loop.c index f2d24b2d992f..d18ed8067a15 100644 --- a/drivers/nvme/target/loop.c +++ b/drivers/nvme/target/loop.c @@ -146,6 +146,12 @@ static blk_status_t nvme_loop_queue_rq(struct blk_mq_hw_ctx *hctx, return ret; nvme_start_request(req); + if (unlikely((req->cmd_flags & REQ_COPY) && + (req_op(req) == REQ_OP_READ))) { + blk_mq_set_request_complete(req); + blk_mq_end_request(req, BLK_STS_OK); + return BLK_STS_OK; + } iod->cmd.common.flags |= NVME_CMD_SGL_METABUF; iod->req.port = queue->ctrl->port; if (!nvmet_req_init(&iod->req, &queue->nvme_cq, diff --git a/drivers/nvme/target/nvmet.h b/drivers/nvme/target/nvmet.h index 89bedfcd974c..69ed4c8469e5 100644 --- a/drivers/nvme/target/nvmet.h +++ b/drivers/nvme/target/nvmet.h @@ -393,6 +393,7 @@ struct nvmet_req { struct device *p2p_client; u16 error_loc; u64 error_slba; + size_t copy_len; }; #define NVMET_MAX_MPOOL_BVEC 16