From patchwork Sat Oct 4 11:45:47 2014 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Yuval Shaia X-Patchwork-Id: 5032891 Return-Path: X-Original-To: patchwork-linux-rdma@patchwork.kernel.org Delivered-To: patchwork-parsemail@patchwork1.web.kernel.org Received: from mail.kernel.org (mail.kernel.org [198.145.19.201]) by patchwork1.web.kernel.org (Postfix) with ESMTP id 3F3C69F327 for ; Sat, 4 Oct 2014 18:51:27 +0000 (UTC) Received: from mail.kernel.org (localhost [127.0.0.1]) by mail.kernel.org (Postfix) with ESMTP id BE1E420122 for ; Sat, 4 Oct 2014 18:51:25 +0000 (UTC) Received: from vger.kernel.org (vger.kernel.org [209.132.180.67]) by mail.kernel.org (Postfix) with ESMTP id CA2D22009C for ; Sat, 4 Oct 2014 18:51:23 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S1751036AbaJDSvT (ORCPT ); Sat, 4 Oct 2014 14:51:19 -0400 Received: from userp1040.oracle.com ([156.151.31.81]:41187 "EHLO userp1040.oracle.com" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S1750928AbaJDSvS (ORCPT ); Sat, 4 Oct 2014 14:51:18 -0400 Received: from acsinet21.oracle.com (acsinet21.oracle.com [141.146.126.237]) by userp1040.oracle.com (Sentrion-MTA-4.3.2/Sentrion-MTA-4.3.2) with ESMTP id s94IpHC3025269 (version=TLSv1/SSLv3 cipher=DHE-RSA-AES256-SHA bits=256 verify=OK) for ; Sat, 4 Oct 2014 18:51:18 GMT Received: from userz7021.oracle.com (userz7021.oracle.com [156.151.31.85]) by acsinet21.oracle.com (8.14.4+Sun/8.14.4) with ESMTP id s94IpFeJ025729 (version=TLSv1/SSLv3 cipher=DHE-RSA-AES256-SHA bits=256 verify=FAIL) for ; Sat, 4 Oct 2014 18:51:16 GMT Received: from abhmp0016.oracle.com (abhmp0016.oracle.com [141.146.116.22]) by userz7021.oracle.com (8.14.4+Sun/8.14.4) with ESMTP id s94IpFIS015783 for ; Sat, 4 Oct 2014 18:51:15 GMT Received: from yuval-net-srv-ca.us.oracle.com (/10.211.3.111) by default (Oracle Beehive Gateway v4.0) with ESMTP ; Sat, 04 Oct 2014 11:51:15 -0700 From: Yuval Shaia To: yuval.shaia@oracle.com, linux-rdma@vger.kernel.org Subject: [PATCH] ib_ipoib: CSUM support in connected mode Date: Sat, 4 Oct 2014 04:45:47 -0700 Message-Id: <1412423147-8402-1-git-send-email-yuval.shaia@oracle.com> X-Mailer: git-send-email 1.7.1 X-Source-IP: acsinet21.oracle.com [141.146.126.237] Sender: linux-rdma-owner@vger.kernel.org Precedence: bulk List-ID: X-Mailing-List: linux-rdma@vger.kernel.org X-Spam-Status: No, score=-5.4 required=5.0 tests=BAYES_00, DATE_IN_PAST_06_12, RCVD_IN_DNSWL_HI, T_RP_MATCHES_RCVD, UNPARSEABLE_RELAY autolearn=unavailable version=3.3.1 X-Spam-Checker-Version: SpamAssassin 3.3.1 (2010-03-16) on mail.kernel.org X-Virus-Scanned: ClamAV using ClamSMTP This enhancement suggest the usage of IB CRC instead of CSUM in IPoIB CM. IPoIB Connected Mode driver uses RC (Reliable Connection) which guarantees the corruption free delivery of the packet. InfiniBand uses 32b CRC which provides stronger data integrity protection compare to 16b IP Checksum. So, there is no added value that IP Checksum provides in the IB world. The proposal is to tell that network stack that IPoIB-CM supports IP Checksum offload. This enables Linux to save the time of checksum calculation of IPoIB CM packets. Network sends the IP packet without adding the IP Checksum to the header. On the receive side, IPoIB driver again tells the network stack that IP Checksum is good for the incoming packets and network stack avoids the IP Checksum calculations. During connection establishment the driver determine if the other end supports IB CRC as checksum. This is done so driver will be able to calculate checksum before transmiting the packet in case the other end does not support this feature. A support for fragmented skb is added to transmit path. Signed-off-by: Yuval Shaia --- drivers/infiniband/ulp/ipoib/ipoib.h | 25 ++++++ drivers/infiniband/ulp/ipoib/ipoib_cm.c | 117 ++++++++++++++++++++++++---- drivers/infiniband/ulp/ipoib/ipoib_ib.c | 3 +- drivers/infiniband/ulp/ipoib/ipoib_main.c | 19 ++++- 4 files changed, 143 insertions(+), 21 deletions(-) diff --git a/drivers/infiniband/ulp/ipoib/ipoib.h b/drivers/infiniband/ulp/ipoib/ipoib.h index 0ed9aed..6fedf83 100644 --- a/drivers/infiniband/ulp/ipoib/ipoib.h +++ b/drivers/infiniband/ulp/ipoib/ipoib.h @@ -100,6 +100,7 @@ enum { IPOIB_FLAG_AUTO_MODER = 13, /*indicates moderation is running*/ /*indicates if event handler was registered*/ IPOIB_FLAG_EVENTS_REGISTERED = 14, + IPOIB_FLAG_CSUM = 15, IPOIB_MAX_BACKOFF_SECONDS = 16, @@ -192,9 +193,28 @@ struct ipoib_pmtu_update { struct ib_cm_id; +#define IPOIB_CM_PROTO_SIG 0x2211 +#define IPOIB_CM_PROTO_VER (1UL << 12) + +static inline int ipoib_cm_check_proto_sig(u16 proto_sig) +{ + return proto_sig & IPOIB_CM_PROTO_SIG; +}; + +static inline int ipoib_cm_check_proto_ver(u16 caps) +{ + return caps & IPOIB_CM_PROTO_VER; +}; + +enum ipoib_cm_data_caps { + IPOIB_CM_CAPS_IBCRC_AS_CSUM = 1UL << 0, +}; + struct ipoib_cm_data { __be32 qpn; /* High byte MUST be ignored on receive */ __be32 mtu; + __be16 sig; /* must be IPOIB_CM_PROTO_SIG */ + __be16 caps; /* 4 bits proto ver and 12 bits capabilities */ }; /* @@ -241,6 +261,7 @@ struct ipoib_cm_rx { int recv_count; u32 qpn; int index; /* For ring counters */ + u16 caps; }; struct ipoib_cm_tx { @@ -255,6 +276,7 @@ struct ipoib_cm_tx { unsigned tx_tail; unsigned long flags; u32 mtu; + u16 caps; }; struct ipoib_cm_rx_buf { @@ -558,6 +580,8 @@ void ipoib_del_neighs_by_gid(struct net_device *dev, u8 *gid); extern struct workqueue_struct *ipoib_workqueue; extern struct workqueue_struct *ipoib_auto_moder_workqueue; +extern int cm_ibcrc_as_csum; + /* functions */ int ipoib_poll(struct napi_struct *napi, int budget); @@ -613,6 +637,7 @@ int ipoib_mcast_stop_thread(struct net_device *dev, int flush); void ipoib_mcast_dev_down(struct net_device *dev); void ipoib_mcast_dev_flush(struct net_device *dev); +int ipoib_dma_map_tx(struct ib_device *ca, struct ipoib_tx_buf *tx_req); #ifdef CONFIG_INFINIBAND_IPOIB_DEBUG struct ipoib_mcast_iter *ipoib_mcast_iter_init(struct net_device *dev); diff --git a/drivers/infiniband/ulp/ipoib/ipoib_cm.c b/drivers/infiniband/ulp/ipoib/ipoib_cm.c index db0303e..553569a 100644 --- a/drivers/infiniband/ulp/ipoib/ipoib_cm.c +++ b/drivers/infiniband/ulp/ipoib/ipoib_cm.c @@ -436,9 +436,16 @@ static int ipoib_cm_send_rep(struct net_device *dev, struct ib_cm_id *cm_id, struct ipoib_dev_priv *priv = netdev_priv(dev); struct ipoib_cm_data data = {}; struct ib_cm_rep_param rep = {}; + u16 caps = 0; + + caps |= IPOIB_CM_PROTO_VER; + if (cm_ibcrc_as_csum) + caps |= IPOIB_CM_CAPS_IBCRC_AS_CSUM; data.qpn = cpu_to_be32(priv->qp->qp_num); data.mtu = cpu_to_be32(IPOIB_CM_BUF_SIZE); + data.sig = cpu_to_be16(IPOIB_CM_PROTO_SIG); + data.caps = cpu_to_be16(caps); rep.private_data = &data; rep.private_data_len = sizeof data; @@ -458,6 +465,7 @@ static int ipoib_cm_req_handler(struct ib_cm_id *cm_id, struct ib_cm_event *even struct ipoib_cm_data *data = event->private_data; unsigned psn; int ret; + struct ipoib_cm_data *cm_data; ipoib_dbg(priv, "REQ arrived\n"); p = kzalloc(sizeof *p, GFP_KERNEL); @@ -480,6 +488,13 @@ static int ipoib_cm_req_handler(struct ib_cm_id *cm_id, struct ib_cm_event *even goto err_qp; } + cm_data = (struct ipoib_cm_data *)event->private_data; + ipoib_dbg(priv, "Otherend sig=0x%x\n", be16_to_cpu(cm_data->sig)); + if (ipoib_cm_check_proto_sig(be16_to_cpu(cm_data->sig)) && + ipoib_cm_check_proto_ver(be16_to_cpu(cm_data->caps))) + p->caps = be16_to_cpu(cm_data->caps); + ipoib_dbg(priv, "Otherend caps=0x%x\n", p->caps); + psn = random32() & 0xffffff; ret = ipoib_cm_modify_rx_qp(dev, cm_id, p->qp, psn); if (ret) @@ -696,6 +711,9 @@ copied: if (skb->dev->priv_flags & IFF_EIPOIB_VIF) set_skb_oob_cb_data(skb, wc, NULL); + if (cm_ibcrc_as_csum) + skb->ip_summed = CHECKSUM_UNNECESSARY; + netif_receive_skb(skb); repost: @@ -732,14 +750,43 @@ static inline int post_send(struct ipoib_dev_priv *priv, return ib_post_send(tx->qp, &send_ring->tx_wr, &bad_wr); } +static inline int post_send_sg(struct ipoib_dev_priv *priv, + struct ipoib_cm_tx *tx, + unsigned int wr_id, + struct sk_buff *skb, + u64 mapping[MAX_SKB_FRAGS + 1], + struct ipoib_send_ring *send_ring) +{ + struct ib_send_wr *bad_wr; + int i, off; + skb_frag_t *frags = skb_shinfo(skb)->frags; + int nr_frags = skb_shinfo(skb)->nr_frags; + if (skb_headlen(skb)) { + send_ring->tx_sge[0].addr = mapping[0]; + send_ring->tx_sge[0].length = skb_headlen(skb); + off = 1; + } else + off = 0; + + for (i = 0; i < nr_frags; ++i) { + send_ring->tx_sge[i + off].addr = mapping[i + off]; + send_ring->tx_sge[i + off].length = frags[i].size; + } + send_ring->tx_wr.num_sge = nr_frags + off; + send_ring->tx_wr.wr_id = wr_id | IPOIB_OP_CM; + + return ib_post_send(tx->qp, &send_ring->tx_wr, &bad_wr); +} + void ipoib_cm_send(struct net_device *dev, struct sk_buff *skb, struct ipoib_cm_tx *tx) { struct ipoib_dev_priv *priv = netdev_priv(dev); struct ipoib_cm_tx_buf *tx_req; - u64 addr; + u64 addr = 0; int rc; struct ipoib_send_ring *send_ring; u16 queue_index; + struct ipoib_tx_buf sg_tx_req; queue_index = skb_get_queue_mapping(skb); send_ring = priv->send_ring + queue_index; @@ -768,26 +815,45 @@ void ipoib_cm_send(struct net_device *dev, struct sk_buff *skb, struct ipoib_cm_ tx_req = &tx->tx_ring[tx->tx_head & (ipoib_sendq_size - 1)]; tx_req->skb = skb; - if (skb->len < ipoib_inline_thold && !skb_shinfo(skb)->nr_frags) { - addr = (u64) skb->data; - send_ring->tx_wr.send_flags |= IB_SEND_INLINE; - tx_req->is_inline = 1; - } else { - addr = ib_dma_map_single(priv->ca, skb->data, - skb->len, DMA_TO_DEVICE); - if (unlikely(ib_dma_mapping_error(priv->ca, addr))) { - ++send_ring->stats.tx_errors; + /* Calculate checksum if we support ibcrc_as_csum but peer does not */ + if ((skb->ip_summed != CHECKSUM_NONE) && cm_ibcrc_as_csum && + !(tx->caps & IPOIB_CM_CAPS_IBCRC_AS_CSUM)) + skb_checksum_help(skb); + + if (skb_shinfo(skb)->nr_frags) { + sg_tx_req.skb = skb; + if (unlikely(ipoib_dma_map_tx(priv->ca, &sg_tx_req))) { + ++dev->stats.tx_errors; dev_kfree_skb_any(skb); return; } - - tx_req->mapping = addr; tx_req->is_inline = 0; - send_ring->tx_wr.send_flags &= ~IB_SEND_INLINE; - } + rc = post_send_sg(priv, tx, tx->tx_head & + (ipoib_sendq_size - 1), + skb, sg_tx_req.mapping, send_ring); + } else { + if (skb->len < ipoib_inline_thold && + !skb_shinfo(skb)->nr_frags) { + addr = (u64) skb->data; + send_ring->tx_wr.send_flags |= IB_SEND_INLINE; + tx_req->is_inline = 1; + } else { + addr = ib_dma_map_single(priv->ca, skb->data, + skb->len, DMA_TO_DEVICE); + if (unlikely(ib_dma_mapping_error(priv->ca, addr))) { + ++send_ring->stats.tx_errors; + dev_kfree_skb_any(skb); + return; + } - rc = post_send(priv, tx, tx->tx_head & (ipoib_sendq_size - 1), - addr, skb->len, send_ring); + tx_req->mapping = addr; + tx_req->is_inline = 0; + send_ring->tx_wr.send_flags &= ~IB_SEND_INLINE; + } + + rc = post_send(priv, tx, tx->tx_head & (ipoib_sendq_size - 1), + addr, skb->len, send_ring); + } if (unlikely(rc)) { ipoib_warn(priv, "%s: post_send failed, error %d queue_index: %d skb->len: %d\n", __func__, rc, queue_index, skb->len); @@ -1012,6 +1078,7 @@ static int ipoib_cm_rep_handler(struct ib_cm_id *cm_id, struct ib_cm_event *even struct ib_qp_attr qp_attr; int qp_attr_mask, ret; struct sk_buff *skb; + struct ipoib_cm_data *cm_data; p->mtu = be32_to_cpu(data->mtu); @@ -1021,6 +1088,13 @@ static int ipoib_cm_rep_handler(struct ib_cm_id *cm_id, struct ib_cm_event *even return -EINVAL; } + cm_data = (struct ipoib_cm_data *)event->private_data; + ipoib_dbg(priv, "Otherend sig=0x%x\n", be16_to_cpu(cm_data->sig)); + if (ipoib_cm_check_proto_sig(be16_to_cpu(cm_data->sig)) && + ipoib_cm_check_proto_ver(be16_to_cpu(cm_data->caps))) + p->caps = be16_to_cpu(cm_data->caps); + ipoib_dbg(priv, "Otherend caps=0x%x\n", p->caps); + qp_attr.qp_state = IB_QPS_RTR; ret = ib_cm_init_qp_attr(cm_id, &qp_attr, &qp_attr_mask); if (ret) { @@ -1105,6 +1179,9 @@ static struct ib_qp *ipoib_cm_create_tx_qp(struct net_device *dev, struct ipoib_ spin_unlock_irqrestore(&priv->lock, flags); + if (dev->features & NETIF_F_SG) + attr.cap.max_send_sge = MAX_SKB_FRAGS + 1; + return ib_create_qp(priv->pd, &attr); } @@ -1116,9 +1193,16 @@ static int ipoib_cm_send_req(struct net_device *dev, struct ipoib_dev_priv *priv = netdev_priv(dev); struct ipoib_cm_data data = {}; struct ib_cm_req_param req = {}; + u16 caps = 0; + + caps |= IPOIB_CM_PROTO_VER; + if (cm_ibcrc_as_csum) + caps |= IPOIB_CM_CAPS_IBCRC_AS_CSUM; data.qpn = cpu_to_be32(priv->qp->qp_num); data.mtu = cpu_to_be32(IPOIB_CM_BUF_SIZE); + data.sig = cpu_to_be16(IPOIB_CM_PROTO_SIG); + data.caps = cpu_to_be16(caps); req.primary_path = pathrec; req.alternate_path = NULL; @@ -1594,7 +1678,6 @@ static void ipoib_cm_stale_task(struct work_struct *work) spin_unlock_irq(&priv->lock); } - static ssize_t show_mode(struct device *d, struct device_attribute *attr, char *buf) { diff --git a/drivers/infiniband/ulp/ipoib/ipoib_ib.c b/drivers/infiniband/ulp/ipoib/ipoib_ib.c index 9e7d824..28f36dd 100644 --- a/drivers/infiniband/ulp/ipoib/ipoib_ib.c +++ b/drivers/infiniband/ulp/ipoib/ipoib_ib.c @@ -349,8 +349,7 @@ repost: "for buf %d\n", wr_id); } -static int ipoib_dma_map_tx(struct ib_device *ca, - struct ipoib_tx_buf *tx_req) +int ipoib_dma_map_tx(struct ib_device *ca, struct ipoib_tx_buf *tx_req) { struct sk_buff *skb = tx_req->skb; u64 *mapping = tx_req->mapping; diff --git a/drivers/infiniband/ulp/ipoib/ipoib_main.c b/drivers/infiniband/ulp/ipoib/ipoib_main.c index dcf18df..b17c7ab 100644 --- a/drivers/infiniband/ulp/ipoib/ipoib_main.c +++ b/drivers/infiniband/ulp/ipoib/ipoib_main.c @@ -69,6 +69,12 @@ module_param_named(debug_level, ipoib_debug_level, int, 0644); MODULE_PARM_DESC(debug_level, "Enable debug tracing if > 0 (default: 0) (0-1)"); #endif +int cm_ibcrc_as_csum; + +module_param_named(cm_ibcrc_as_csum, cm_ibcrc_as_csum, int, 0444); +MODULE_PARM_DESC(cm_ibcrc_as_csum, "Indicates whether to utilize IB-CRC as " + "CSUM in connected mode,(default: 0)"); + struct ipoib_path_iter { struct net_device *dev; struct ipoib_path path; @@ -195,7 +201,12 @@ static netdev_features_t ipoib_fix_features(struct net_device *dev, netdev_featu struct ipoib_dev_priv *priv = netdev_priv(dev); if (test_bit(IPOIB_FLAG_ADMIN_CM, &priv->flags)) - features &= ~(NETIF_F_SG | NETIF_F_IP_CSUM | NETIF_F_TSO); + if (cm_ibcrc_as_csum && (test_bit(IPOIB_FLAG_CSUM, + &priv->flags))) + dev->features |= NETIF_F_IP_CSUM | NETIF_F_SG; + else + features &= ~(NETIF_F_SG | NETIF_F_IP_CSUM | + NETIF_F_TSO); return features; } @@ -249,7 +260,11 @@ int ipoib_set_mode(struct net_device *dev, const char *buf) send_ring = priv->send_ring; for (i = 0; i < priv->num_tx_queues; i++) { - send_ring->tx_wr.send_flags &= ~IB_SEND_IP_CSUM; + if (cm_ibcrc_as_csum && (test_bit(IPOIB_FLAG_CSUM, + &priv->flags))) + send_ring->tx_wr.send_flags |= IB_SEND_IP_CSUM; + else + send_ring->tx_wr.send_flags &= ~IB_SEND_IP_CSUM; send_ring++; }