From patchwork Fri Feb 17 19:15:07 2023 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Tirthendu Sarkar X-Patchwork-Id: 13145185 Return-Path: X-Spam-Checker-Version: SpamAssassin 3.4.0 (2014-02-07) on aws-us-west-2-korg-lkml-1.web.codeaurora.org Received: from vger.kernel.org (vger.kernel.org [23.128.96.18]) by smtp.lore.kernel.org (Postfix) with ESMTP id C9202C05027 for ; Fri, 17 Feb 2023 19:29:49 +0000 (UTC) Received: (majordomo@vger.kernel.org) by vger.kernel.org via listexpand id S229672AbjBQT3s (ORCPT ); Fri, 17 Feb 2023 14:29:48 -0500 Received: from lindbergh.monkeyblade.net ([23.128.96.19]:48508 "EHLO lindbergh.monkeyblade.net" rhost-flags-OK-OK-OK-OK) by vger.kernel.org with ESMTP id S229463AbjBQT3r (ORCPT ); Fri, 17 Feb 2023 14:29:47 -0500 Received: from mga06.intel.com (mga06b.intel.com [134.134.136.31]) by lindbergh.monkeyblade.net (Postfix) with ESMTPS id 7C0FC2FCFE; Fri, 17 Feb 2023 11:29:46 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/simple; d=intel.com; i=@intel.com; q=dns/txt; s=Intel; t=1676662186; x=1708198186; h=from:to:cc:subject:date:message-id:mime-version: content-transfer-encoding; bh=CEnd+qF+2FYATu7ckoPg+BqcjVuV/8rQiz89geC2aUg=; b=FjvpsHW+ff01Di+U9BLIEGEeVIYF0nKvn7715uIthVpiUE0OMy6ZTg+a hzkgciYkcviyYZZbogcZeow7qTRpXpv60NnDB+Mt09sZKDvRAKskOBzRx RgfIBVRsqbz+EMCOwGGNW6sEwTsTm35AGefUvIzf8gWZCqvhPzwjnUoFF tLBFMbskMRn0+K57RmqfFZA6B0PPsmoVBUSEQubrqqdZJRTHFD2kZwkiT NxwaKFJfMbX4fjlvf4QvUuBjVlSUecGromgiLzc+HVb04XcZZ+fX7BCkS K2hQ/Fwseqld79PhxPOwYsAIpMchjBtY4zbXMClEHgTUOdZ+kZvzFKKN4 A==; X-IronPort-AV: E=McAfee;i="6500,9779,10624"; a="394550033" X-IronPort-AV: E=Sophos;i="5.97,306,1669104000"; d="scan'208";a="394550033" Received: from orsmga008.jf.intel.com ([10.7.209.65]) by orsmga104.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Feb 2023 11:29:46 -0800 X-IronPort-AV: E=McAfee;i="6500,9779,10624"; a="701013375" X-IronPort-AV: E=Sophos;i="5.97,306,1669104000"; d="scan'208";a="701013375" Received: from unknown (HELO paamrpdk12-S2600BPB.aw.intel.com) ([10.228.151.145]) by orsmga008-auth.jf.intel.com with ESMTP/TLS/ECDHE-RSA-AES256-GCM-SHA384; 17 Feb 2023 11:29:45 -0800 From: Tirthendu Sarkar To: intel-wired-lan@lists.osuosl.org Cc: jesse.brandeburg@intel.com, anthony.l.nguyen@intel.com, netdev@vger.kernel.org, bpf@vger.kernel.org, magnus.karlsson@intel.com, maciej.fijalkowski@intel.com, tirthendu.sarkar@intel.com Subject: [PATCH intel-next v6 0/8] i40e: support XDP multi-buffer Date: Sat, 18 Feb 2023 00:45:07 +0530 Message-Id: <20230217191515.166819-1-tirthendu.sarkar@intel.com> X-Mailer: git-send-email 2.34.1 MIME-Version: 1.0 Precedence: bulk List-ID: X-Mailing-List: bpf@vger.kernel.org This patchset adds multi-buffer support for XDP. Tx side already has support for multi-buffer. This patchset focuses on Rx side. The last patch contains actual multi-buffer changes while the previous ones are preparatory patches. On receiving the first buffer of a packet, xdp_buff is built and its subsequent buffers are added to it as frags. While 'next_to_clean' keeps pointing to the first descriptor, the newly introduced 'next_to_process' keeps track of every descriptor for the packet. On receiving EOP buffer the XDP program is called and appropriate action is taken (building skb for XDP_PASS, reusing page for XDP_DROP, adjusting page offsets for XDP_{REDIRECT,TX}). The patchset also streamlines page offset adjustments for buffer reuse to make it easier to post process the rx_buffers after running XDP prog. With this patchset there does not seem to be any performance degradation for XDP_PASS and some improvement (~1% for XDP_TX, ~5% for XDP_DROP) when measured using xdp_rxq_info program from samples/bpf/ for 64B packets. Changelog: v5 -> v6: - Rebased on top of next-queue commit ce45ffb815e8 ("i40e: add double of VLAN header when computing the max MTU") v4 -> v5: - Change s/size/truesize [Tony] - Rebased on top of commit 9dd6e53ef63d ("i40e: check vsi type before setting xdp_features flag") [Lorenzo] - Changed size of on stack variable to u32 from u16. v3 -> v4: - Added non-linear XDP buffer support to xdp_features. [Maciej] - Removed double space. [Maciej] v2 -> v3: - Fixed buffer cleanup for single buffer packets on skb alloc failure. - Better naming of cleanup function. - Stop incrementing nr_frags for overflowing packets. v1 -> v2: - Instead of building xdp_buff on eop now it is built incrementally. - xdp_buff is now added to i40e_ring struct for preserving across napi calls. [Alexander Duyck] - Post XDP program rx_buffer processing has been simplified. - Rx buffer allocation pull out is reverted to avoid performance issues for smaller ring sizes and now done when at least half of the ring has been cleaned. With v1 there was ~75% drop for XDP_PASS with the smallest ring size of 64 which is mitigated by v2 [Alexander Duyck] - Instead of retrying skb allocation on previous failure now the packet is dropped. [Maciej] - Simplified page offset adjustments by using xdp->frame_sz instead of recalculating truesize. [Maciej] - Change i40e_trace() to use xdp instead of skb [Maciej] - Reserve tailroom for legacy-rx [Maciej] - Centralize max frame size calculation Tirthendu Sarkar (8): i40e: consolidate maximum frame size calculation for vsi i40e: change Rx buffer size for legacy-rx to support XDP multi-buffer i40e: add pre-xdp page_count in rx_buffer i40e: Change size to truesize when using i40e_rx_buffer_flip() i40e: use frame_sz instead of recalculating truesize for building skb i40e: introduce next_to_process to i40e_ring i40e: add xdp_buff to i40e_ring struct i40e: add support for XDP multi-buffer Rx drivers/net/ethernet/intel/i40e/i40e_main.c | 78 ++-- drivers/net/ethernet/intel/i40e/i40e_trace.h | 20 +- drivers/net/ethernet/intel/i40e/i40e_txrx.c | 420 +++++++++++-------- drivers/net/ethernet/intel/i40e/i40e_txrx.h | 21 +- 4 files changed, 307 insertions(+), 232 deletions(-)