From patchwork Sun Jan 20 13:05:56 2019 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Leon Romanovsky X-Patchwork-Id: 10772389 Return-Path: Received: from mail.wl.linuxfoundation.org (pdx-wl-mail.web.codeaurora.org [172.30.200.125]) by pdx-korg-patchwork-2.web.codeaurora.org (Postfix) with ESMTP id 4130B6C2 for ; Sun, 20 Jan 2019 13:06:14 +0000 (UTC) Received: from mail.wl.linuxfoundation.org (localhost [127.0.0.1]) by mail.wl.linuxfoundation.org (Postfix) with ESMTP id 2FE592A6F1 for ; Sun, 20 Jan 2019 13:06:14 +0000 (UTC) Received: by mail.wl.linuxfoundation.org (Postfix, from userid 486) id 242512AA7A; Sun, 20 Jan 2019 13:06:14 +0000 (UTC) X-Spam-Checker-Version: SpamAssassin 3.3.1 (2010-03-16) on pdx-wl-mail.web.codeaurora.org X-Spam-Level: X-Spam-Status: No, score=-8.0 required=2.0 tests=BAYES_00,DKIM_SIGNED, DKIM_VALID,DKIM_VALID_AU,MAILING_LIST_MULTI,RCVD_IN_DNSWL_HI autolearn=ham version=3.3.1 Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by mail.wl.linuxfoundation.org (Postfix) with ESMTP id 54BBC2AA81 for ; Sun, 20 Jan 2019 13:06:13 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1730585AbfATNGN (ORCPT ); Sun, 20 Jan 2019 08:06:13 -0500 Received: from mail.kernel.org ([198.145.29.99]:36348 "EHLO mail.kernel.org" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1730542AbfATNGM (ORCPT ); Sun, 20 Jan 2019 08:06:12 -0500 Received: from localhost (unknown [193.47.165.251]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-GCM-SHA384 (256/256 bits)) (No client certificate requested) by mail.kernel.org (Postfix) with ESMTPSA id 3A4AD20880; Sun, 20 Jan 2019 13:06:11 +0000 (UTC) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=kernel.org; s=default; t=1547989571; bh=MsXgQRrd0sRdQ2T5fx1NLniGPwd+Yiz0G8beRgCQodc=; h=From:To:Cc:Subject:Date:In-Reply-To:References:From; b=RFc1REX7FBJMeCuZWbPqXyglC/7fEuQWvKk5NKdEEG14M41ff5zgarWakgObSYtih TF7jt3GiOAo0D3V63r70b/6CoL3fmufvj6wbkrC+dvtbipl0yQ6D+ajZW++WCNBi4L onhRddrMCv2E2dGheRxpreW1QNnvkkj2+zZVtpGs= From: Leon Romanovsky To: Doug Ledford , Jason Gunthorpe Cc: Leon Romanovsky , RDMA mailing list Subject: [PATCH rdma-next v1 2/6] RDMA/mlx5: Use restrack allocation PD scheme Date: Sun, 20 Jan 2019 15:05:56 +0200 Message-Id: <20190120130600.8788-3-leon@kernel.org> X-Mailer: git-send-email 2.20.1 In-Reply-To: <20190120130600.8788-1-leon@kernel.org> References: <20190120130600.8788-1-leon@kernel.org> MIME-Version: 1.0 Sender: linux-rdma-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-rdma@vger.kernel.org X-Virus-Scanned: ClamAV using ClamSMTP From: Leon Romanovsky Adapt mlx5 to share restrack ID instead of local variable. Signed-off-by: Leon Romanovsky --- drivers/infiniband/hw/mlx5/main.c | 39 +++++++++++++++++++--------- drivers/infiniband/hw/mlx5/mlx5_ib.h | 1 - drivers/infiniband/hw/mlx5/mr.c | 16 ++++++------ drivers/infiniband/hw/mlx5/qp.c | 22 +++++++++------- drivers/infiniband/hw/mlx5/srq.c | 2 +- 5 files changed, 49 insertions(+), 31 deletions(-) diff --git a/drivers/infiniband/hw/mlx5/main.c b/drivers/infiniband/hw/mlx5/main.c index b32b5c7f1cfd..1348e4e9e119 100644 --- a/drivers/infiniband/hw/mlx5/main.c +++ b/drivers/infiniband/hw/mlx5/main.c @@ -2262,33 +2262,46 @@ static struct ib_pd *mlx5_ib_alloc_pd(struct ib_device *ibdev, u32 out[MLX5_ST_SZ_DW(alloc_pd_out)] = {}; u32 in[MLX5_ST_SZ_DW(alloc_pd_in)] = {}; u16 uid = 0; + u32 pdn; pd = kzalloc(sizeof(*pd), GFP_KERNEL); if (!pd) return ERR_PTR(-ENOMEM); + pd->ibpd.device = ibdev; + uid = context ? to_mucontext(context)->devx_uid : 0; MLX5_SET(alloc_pd_in, in, opcode, MLX5_CMD_OP_ALLOC_PD); MLX5_SET(alloc_pd_in, in, uid, uid); err = mlx5_cmd_exec(to_mdev(ibdev)->mdev, in, sizeof(in), out, sizeof(out)); - if (err) { - kfree(pd); - return ERR_PTR(err); - } + if (err) + goto out; - pd->pdn = MLX5_GET(alloc_pd_out, out, pd); + pdn = MLX5_GET(alloc_pd_out, out, pd); pd->uid = uid; + rdma_rt_set_type(&pd->ibpd.res, RDMA_RESTRACK_PD); + rdma_rt_set_id(&pd->ibpd.res, pdn); + err = rdma_restrack_add(&pd->ibpd.res); + if (err) + goto out; + if (context) { - resp.pdn = pd->pdn; + resp.pdn = pdn; if (ib_copy_to_udata(udata, &resp, sizeof(resp))) { - mlx5_cmd_dealloc_pd(to_mdev(ibdev)->mdev, pd->pdn, uid); - kfree(pd); - return ERR_PTR(-EFAULT); + err = -EFAULT; + goto err_copy; } } return &pd->ibpd; + +err_copy: + mlx5_cmd_dealloc_pd(to_mdev(ibdev)->mdev, pdn, uid); + rdma_restrack_del(&pd->ibpd.res); + +out: kfree(pd); + return ERR_PTR(err); } static int mlx5_ib_dealloc_pd(struct ib_pd *pd) @@ -2296,7 +2309,8 @@ static int mlx5_ib_dealloc_pd(struct ib_pd *pd) struct mlx5_ib_dev *mdev = to_mdev(pd->device); struct mlx5_ib_pd *mpd = to_mpd(pd); - mlx5_cmd_dealloc_pd(mdev->mdev, mpd->pdn, mpd->uid); + rdma_restrack_del(&pd->res); + mlx5_cmd_dealloc_pd(mdev->mdev, rdma_res_to_id(&pd->res), mpd->uid); kfree(mpd); return 0; @@ -4665,8 +4679,6 @@ static int create_dev_resources(struct mlx5_ib_resources *devr) ret = PTR_ERR(devr->p0); goto error0; } - devr->p0->device = &dev->ib_dev; - devr->p0->uobject = NULL; atomic_set(&devr->p0->usecnt, 0); devr->c0 = mlx5_ib_create_cq(&dev->ib_dev, &cq_attr, NULL, NULL); @@ -6513,6 +6525,9 @@ static void *mlx5_ib_add(struct mlx5_core_dev *mdev) if (!dev) return NULL; + /* PD ID managed by FW */ + rdma_rt_set_id_range(&dev->ib_dev, RDMA_RESTRACK_PD, 0, 0); + dev->mdev = mdev; dev->num_ports = max(MLX5_CAP_GEN(mdev, num_ports), MLX5_CAP_GEN(mdev, num_vhca_ports)); diff --git a/drivers/infiniband/hw/mlx5/mlx5_ib.h b/drivers/infiniband/hw/mlx5/mlx5_ib.h index cead0f97a7b1..8b235ceb9574 100644 --- a/drivers/infiniband/hw/mlx5/mlx5_ib.h +++ b/drivers/infiniband/hw/mlx5/mlx5_ib.h @@ -143,7 +143,6 @@ static inline struct mlx5_ib_ucontext *to_mucontext(struct ib_ucontext *ibuconte struct mlx5_ib_pd { struct ib_pd ibpd; - u32 pdn; u16 uid; }; diff --git a/drivers/infiniband/hw/mlx5/mr.c b/drivers/infiniband/hw/mlx5/mr.c index 6d763d6a189f..8f557f90ad7f 100644 --- a/drivers/infiniband/hw/mlx5/mr.c +++ b/drivers/infiniband/hw/mlx5/mr.c @@ -800,7 +800,7 @@ struct ib_mr *mlx5_ib_get_dma_mr(struct ib_pd *pd, int acc) MLX5_SET(mkc, mkc, lr, 1); MLX5_SET(mkc, mkc, length64, 1); - MLX5_SET(mkc, mkc, pd, to_mpd(pd)->pdn); + MLX5_SET(mkc, mkc, pd, rdma_res_to_id(&pd->res)); MLX5_SET(mkc, mkc, qpn, 0xffffff); MLX5_SET64(mkc, mkc, start_addr, 0); @@ -950,7 +950,7 @@ static struct mlx5_ib_mr *alloc_mr_from_cache( mr->desc_size = sizeof(struct mlx5_mtt); mr->mmkey.iova = virt_addr; mr->mmkey.size = len; - mr->mmkey.pd = to_mpd(pd)->pdn; + mr->mmkey.pd = rdma_res_to_id(&pd->res); return mr; } @@ -1171,7 +1171,7 @@ static struct mlx5_ib_mr *reg_create(struct ib_mr *ibmr, struct ib_pd *pd, MLX5_SET64(mkc, mkc, start_addr, virt_addr); MLX5_SET64(mkc, mkc, len, length); - MLX5_SET(mkc, mkc, pd, to_mpd(pd)->pdn); + MLX5_SET(mkc, mkc, pd, rdma_res_to_id(&pd->res)); MLX5_SET(mkc, mkc, bsf_octword_size, 0); MLX5_SET(mkc, mkc, translations_octword_size, get_octo_len(virt_addr, length, page_shift)); @@ -1250,7 +1250,7 @@ static struct ib_mr *mlx5_ib_get_memic_mr(struct ib_pd *pd, u64 memic_addr, MLX5_SET(mkc, mkc, lr, 1); MLX5_SET64(mkc, mkc, len, length); - MLX5_SET(mkc, mkc, pd, to_mpd(pd)->pdn); + MLX5_SET(mkc, mkc, pd, rdma_res_to_id(&pd->res)); MLX5_SET(mkc, mkc, qpn, 0xffffff); MLX5_SET64(mkc, mkc, start_addr, memic_addr - pci_resource_start(dev->mdev->pdev, 0)); @@ -1525,7 +1525,7 @@ int mlx5_ib_rereg_user_mr(struct ib_mr *ib_mr, int flags, u64 start, mr->access_flags = access_flags; mr->mmkey.iova = addr; mr->mmkey.size = len; - mr->mmkey.pd = to_mpd(pd)->pdn; + mr->mmkey.pd = rdma_res_to_id(&pd->res); if (flags & IB_MR_REREG_TRANS) { upd_flags = MLX5_IB_UPD_XLT_ADDR; @@ -1703,7 +1703,7 @@ struct ib_mr *mlx5_ib_alloc_mr(struct ib_pd *pd, MLX5_SET(mkc, mkc, free, 1); MLX5_SET(mkc, mkc, translations_octword_size, ndescs); MLX5_SET(mkc, mkc, qpn, 0xffffff); - MLX5_SET(mkc, mkc, pd, to_mpd(pd)->pdn); + MLX5_SET(mkc, mkc, pd, rdma_res_to_id(&pd->res)); if (mr_type == IB_MR_TYPE_MEM_REG) { mr->access_mode = MLX5_MKC_ACCESS_MODE_MTT; @@ -1736,7 +1736,7 @@ struct ib_mr *mlx5_ib_alloc_mr(struct ib_pd *pd, } /* create mem & wire PSVs */ - err = mlx5_core_create_psv(dev->mdev, to_mpd(pd)->pdn, + err = mlx5_core_create_psv(dev->mdev, rdma_res_to_id(&pd->res), 2, psv_index); if (err) goto err_free_sig; @@ -1834,7 +1834,7 @@ struct ib_mw *mlx5_ib_alloc_mw(struct ib_pd *pd, enum ib_mw_type type, MLX5_SET(mkc, mkc, free, 1); MLX5_SET(mkc, mkc, translations_octword_size, ndescs); - MLX5_SET(mkc, mkc, pd, to_mpd(pd)->pdn); + MLX5_SET(mkc, mkc, pd, rdma_res_to_id(&pd->res)); MLX5_SET(mkc, mkc, umr_en, 1); MLX5_SET(mkc, mkc, lr, 1); MLX5_SET(mkc, mkc, access_mode_1_0, MLX5_MKC_ACCESS_MODE_KLMS); diff --git a/drivers/infiniband/hw/mlx5/qp.c b/drivers/infiniband/hw/mlx5/qp.c index 529e76f67cb6..44bde8dfebf6 100644 --- a/drivers/infiniband/hw/mlx5/qp.c +++ b/drivers/infiniband/hw/mlx5/qp.c @@ -2038,11 +2038,13 @@ static int create_qp_common(struct mlx5_ib_dev *dev, struct ib_pd *pd, MLX5_SET(qpc, qpc, st, mlx5_st); MLX5_SET(qpc, qpc, pm_state, MLX5_QP_PM_MIGRATED); - if (init_attr->qp_type != MLX5_IB_QPT_REG_UMR) - MLX5_SET(qpc, qpc, pd, to_mpd(pd ? pd : devr->p0)->pdn); - else - MLX5_SET(qpc, qpc, latency_sensitive, 1); + if (init_attr->qp_type != MLX5_IB_QPT_REG_UMR) { + struct ib_pd *t = pd ? pd : devr->p0; + MLX5_SET(qpc, qpc, pd, rdma_res_to_id(&t->res)); + } else { + MLX5_SET(qpc, qpc, latency_sensitive, 1); + } if (qp->wq_sig) MLX5_SET(qpc, qpc, wq_signature, 1); @@ -2431,7 +2433,7 @@ static struct ib_qp *mlx5_ib_create_dct(struct ib_pd *pd, MLX5_SET(create_dct_in, qp->dct.in, uid, to_mpd(pd)->uid); dctc = MLX5_ADDR_OF(create_dct_in, qp->dct.in, dct_context_entry); qp->qp_sub_type = MLX5_IB_QPT_DCT; - MLX5_SET(dctc, dctc, pd, to_mpd(pd)->pdn); + MLX5_SET(dctc, dctc, pd, rdma_res_to_id(&pd->res)); MLX5_SET(dctc, dctc, srqn_xrqn, to_msrq(attr->srq)->msrq.srqn); MLX5_SET(dctc, dctc, cqn, to_mcq(attr->recv_cq)->mcq.cqn); MLX5_SET64(dctc, dctc, dc_access_key, ucmd->access_key); @@ -3357,7 +3359,9 @@ static int __mlx5_ib_modify_qp(struct ib_qp *ibqp, get_cqs(qp->ibqp.qp_type, qp->ibqp.send_cq, qp->ibqp.recv_cq, &send_cq, &recv_cq); - context->flags_pd = cpu_to_be32(pd ? pd->pdn : to_mpd(dev->devr.p0)->pdn); + context->flags_pd = + cpu_to_be32(pd ? rdma_res_to_id(&pd->ibpd.res) : + rdma_res_to_id(&dev->devr.p0->res)); context->cqn_send = send_cq ? cpu_to_be32(send_cq->mcq.cqn) : 0; context->cqn_recv = recv_cq ? cpu_to_be32(recv_cq->mcq.cqn) : 0; context->params1 = cpu_to_be32(MLX5_IB_ACK_REQ_FREQ << 28); @@ -4148,7 +4152,7 @@ static void set_reg_mkey_segment(struct mlx5_mkey_seg *seg, seg->flags = convert_access(umrwr->access_flags); if (umrwr->pd) - seg->flags_pd = cpu_to_be32(to_mpd(umrwr->pd)->pdn); + seg->flags_pd = cpu_to_be32(rdma_res_to_id(&umrwr->pd->res)); if (wr->send_flags & MLX5_IB_SEND_UMR_UPDATE_TRANSLATION && !umrwr->length) seg->flags_pd |= cpu_to_be32(MLX5_MKEY_LEN64); @@ -4495,7 +4499,7 @@ static int set_sig_umr_wr(const struct ib_send_wr *send_wr, { const struct ib_sig_handover_wr *wr = sig_handover_wr(send_wr); struct mlx5_ib_mr *sig_mr = to_mmr(wr->sig_mr); - u32 pdn = get_pd(qp)->pdn; + u32 pdn = rdma_res_to_id(&get_pd(qp)->ibpd.res); u32 xlt_size; int region_len, ret; @@ -5647,7 +5651,7 @@ static int create_rq(struct mlx5_ib_rwq *rwq, struct ib_pd *pd, MLX5_MIN_SINGLE_WQE_LOG_NUM_STRIDES); } MLX5_SET(wq, wq, log_wq_sz, rwq->log_rq_size); - MLX5_SET(wq, wq, pd, to_mpd(pd)->pdn); + MLX5_SET(wq, wq, pd, rdma_res_to_id(&pd->res)); MLX5_SET(wq, wq, page_offset, rwq->rq_page_offset); MLX5_SET(wq, wq, log_wq_pg_sz, rwq->log_page_size); MLX5_SET(wq, wq, wq_signature, rwq->wq_sig); diff --git a/drivers/infiniband/hw/mlx5/srq.c b/drivers/infiniband/hw/mlx5/srq.c index 22bd774e0b4e..57c7df0dabdf 100644 --- a/drivers/infiniband/hw/mlx5/srq.c +++ b/drivers/infiniband/hw/mlx5/srq.c @@ -297,7 +297,7 @@ struct ib_srq *mlx5_ib_create_srq(struct ib_pd *pd, else in.cqn = to_mcq(dev->devr.c0)->mcq.cqn; - in.pd = to_mpd(pd)->pdn; + in.pd = rdma_res_to_id(&pd->res); in.db_record = srq->db.dma; err = mlx5_cmd_create_srq(dev, &srq->msrq, &in); kvfree(in.pas);