diff mbox series

[RFC,bpf-next,27/52] net, xdp: add &sk_buff <-> &xdp_meta_generic converters

Message ID 20220628194812.1453059-28-alexandr.lobakin@intel.com (mailing list archive)
State RFC
Delegated to: BPF
Headers show
Series bpf, xdp: introduce and use Generic Hints/metadata | expand

Checks

Context Check Description
bpf/vmtest-bpf-next-PR fail PR summary
bpf/vmtest-bpf-next-VM_Test-1 fail Logs for Kernel LATEST on ubuntu-latest with gcc
bpf/vmtest-bpf-next-VM_Test-2 fail Logs for Kernel LATEST on ubuntu-latest with llvm-15
bpf/vmtest-bpf-next-VM_Test-3 fail Logs for Kernel LATEST on z15 with gcc
netdev/tree_selection success Clearly marked for bpf-next, async
netdev/apply fail Patch does not apply to bpf-next

Commit Message

Alexander Lobakin June 28, 2022, 7:47 p.m. UTC
Add two functions (with their underscored versions) to pass
HW-origined info (checksums, hashes, Rx queue ID etc.) from an skb
to an XDP generic metadata and vice versa. They can be used to carry
that info between hardware, xdp_buff/xdp_frame and sk_buff.
The &sk_buff -> &xdp_meta_generic converter uses a static,
init-time filled &xdp_meta_tail to not query BTF info on hotpath.
For the fields which values are being assigned directly, make sure
they match with the help of static asserts.
Also add a wrapper bpf_match_type_btf_id() designed especially for
drivers and taking care of corner-cases.

Signed-off-by: Alexander Lobakin <alexandr.lobakin@intel.com>
---
 include/net/xdp_meta.h | 112 +++++++++++++++++++++++++++++++
 net/bpf/core.c         | 148 ++++++++++++++++++++++++++++++++++++++++-
 2 files changed, 259 insertions(+), 1 deletion(-)
diff mbox series

Patch

diff --git a/include/net/xdp_meta.h b/include/net/xdp_meta.h
index f61831e39eb0..d37ea873a6a8 100644
--- a/include/net/xdp_meta.h
+++ b/include/net/xdp_meta.h
@@ -46,6 +46,17 @@  static inline bool xdp_metalen_invalid(unsigned long metalen)
 	return (metalen & (sizeof(u32) - 1)) || metalen > max;
 }
 
+/* We use direct assignments from &xdp_meta_generic to &sk_buff fields,
+ * thus they must match.
+ */
+static_assert((u32)XDP_META_RX_CSUM_NONE == (u32)CHECKSUM_NONE);
+static_assert((u32)XDP_META_RX_CSUM_OK == (u32)CHECKSUM_UNNECESSARY);
+static_assert((u32)XDP_META_RX_CSUM_COMP == (u32)CHECKSUM_COMPLETE);
+static_assert((u32)XDP_META_RX_HASH_NONE == (u32)PKT_HASH_TYPE_NONE);
+static_assert((u32)XDP_META_RX_HASH_L2 == (u32)PKT_HASH_TYPE_L2);
+static_assert((u32)XDP_META_RX_HASH_L3 == (u32)PKT_HASH_TYPE_L3);
+static_assert((u32)XDP_META_RX_HASH_L4 == (u32)PKT_HASH_TYPE_L4);
+
 /* This builds _get(), _set() and _rep() for each bitfield.
  * If you know for sure the field is empty (e.g. you zeroed the struct
  * previously), use faster _set() op to save several cycles, otherwise
@@ -283,4 +294,105 @@  static inline bool xdp_meta_skb_has_generic(const struct sk_buff *skb)
 	return xdp_meta_has_generic(skb_metadata_end(skb));
 }
 
+/**
+ * xdp_meta_init - initialize a metadata structure
+ * @md: pointer to xdp_meta_generic or its ::rx_full or its ::id member
+ * @id: full BTF + type ID for the metadata type (can be u* or __le64)
+ *
+ * Zeroes the passed metadata struct (or part) and initializes its tail, so
+ * it becomes ready for further processing. If a driver is responsible for
+ * composing metadata, it is important to zero the space it occupies in each
+ * Rx buffer as `xdp->data - xdp->data_hard_start` doesn't get initialized
+ * by default.
+ */
+#define _xdp_meta_init(md, id, locmd, locid) ({				      \
+	typeof(md) locmd = (md);					      \
+	typeof(id) locid = (id);					      \
+									      \
+	if (offsetof(typeof(*locmd), full_id))				      \
+		memset(locmd, 0, offsetof(typeof(*locmd), full_id));	      \
+									      \
+	locmd->full_id = __same_type(locid, __le64) ? (__force __le64)locid : \
+			 cpu_to_le64((__force u64)locid);		      \
+	locmd->magic_id = cpu_to_le16(XDP_META_GENERIC_MAGIC);		      \
+})
+#define xdp_meta_init(md, id)						      \
+	_xdp_meta_init((md), (id), __UNIQUE_ID(md_), __UNIQUE_ID(id_))
+
+void ___xdp_build_meta_generic_from_skb(struct xdp_meta_generic_rx *rx_md,
+					const struct sk_buff *skb);
+void ___xdp_populate_skb_meta_generic(struct sk_buff *skb,
+				      const struct xdp_meta_generic_rx *rx_md);
+
+#define _xdp_build_meta_generic_from_skb(md, skb, locmd) ({		      \
+	typeof(md) locmd = (md);					      \
+									      \
+	if (offsetof(typeof(*locmd), rx))				      \
+		memset(locmd, 0, offsetof(typeof(*locmd), rx));		      \
+									      \
+	___xdp_build_meta_generic_from_skb(to_rx_md(locmd), skb);	      \
+})
+#define __xdp_build_meta_generic_from_skb(md, skb)			      \
+	_xdp_build_meta_generic_from_skb((md), (skb), __UNIQUE_ID(md_))
+
+#define __xdp_populate_skb_meta_generic(skb, md)			      \
+	___xdp_populate_skb_meta_generic((skb), to_rx_md(md))
+
+/**
+ * xdp_build_meta_generic_from_skb - build the generic meta before the skb data
+ * @skb: a pointer to the &sk_buff
+ *
+ * Builds an XDP generic metadata in front of the skb data from its fields.
+ * Note: skb->mac_header must be set and valid.
+ */
+static inline void xdp_build_meta_generic_from_skb(struct sk_buff *skb)
+{
+	struct xdp_meta_generic *md;
+	u32 needed;
+
+	/* skb_headroom() is `skb->data - skb->head`, i.e. it doesn't account
+	 * for the pulled headers, e.g. MAC header. Metadata resides in front
+	 * of the MAC header, so counting starts from there, not the current
+	 * data pointer position.
+	 * CoW won't happen in here when coming from Generic XDP path as it
+	 * ensures that an skb has at least %XDP_PACKET_HEADROOM beforehand.
+	 * It won't be happening also as long as `sizeof(*md) <= NET_SKB_PAD`.
+	 */
+	needed = (void *)skb->data - skb_metadata_end(skb) + sizeof(*md);
+	if (unlikely(skb_cow_head(skb, needed)))
+		return;
+
+	md = xdp_meta_generic_ptr(skb_metadata_end(skb));
+	__xdp_build_meta_generic_from_skb(md, skb);
+
+	skb_metadata_set(skb, sizeof(*md));
+	skb_metadata_nocomp_set(skb);
+}
+
+/**
+ * xdp_populate_skb_meta_generic - fill an skb from the metadata in front of it
+ * @skb: a pointer to the &sk_buff
+ *
+ * Fills the skb fields from the metadata in front of its MAC header and marks
+ * its metadata as "non-comparable".
+ * Note: skb->mac_header must be set and valid.
+ */
+static inline void xdp_populate_skb_meta_generic(struct sk_buff *skb)
+{
+	const struct xdp_meta_generic *md;
+
+	if (skb_metadata_len(skb) < sizeof(*md))
+		return;
+
+	md = xdp_meta_generic_ptr(skb_metadata_end(skb));
+	__xdp_populate_skb_meta_generic(skb, md);
+
+	/* We know at this point that skb metadata may contain
+	 * unique values, mark it as nocomp to not confuse GRO.
+	 */
+	skb_metadata_nocomp_set(skb);
+}
+
+int xdp_meta_match_id(const char * const *list, u64 id);
+
 #endif /* __LINUX_NET_XDP_META_H__ */
diff --git a/net/bpf/core.c b/net/bpf/core.c
index 18174d6d8687..a8685bcc6e00 100644
--- a/net/bpf/core.c
+++ b/net/bpf/core.c
@@ -3,7 +3,7 @@ 
  *
  * Copyright (c) 2017 Jesper Dangaard Brouer, Red Hat Inc.
  */
-#include <linux/bpf.h>
+#include <linux/btf.h>
 #include <linux/filter.h>
 #include <linux/types.h>
 #include <linux/mm.h>
@@ -713,3 +713,149 @@  struct xdp_frame *xdpf_clone(struct xdp_frame *xdpf)
 
 	return nxdpf;
 }
+
+/**
+ * xdp_meta_match_id - find a type name corresponding to a given full ID
+ * @list: pointer to the %NULL-terminated list of type names
+ * @id: full ID (BTF ID + type ID) of the type to look
+ *
+ * Convenience wrapper over bpf_match_type_btf_id() for usage in drivers which
+ * takes care of zeroed ID and BPF syscall being not compiled in (to not break
+ * code flow and return "no meta").
+ *
+ * Returns a string array element index on success, an error code otherwise.
+ */
+int xdp_meta_match_id(const char * const *list, u64 id)
+{
+	int ret;
+
+	if (unlikely(!list || !*list))
+		return id ? -EINVAL : 0;
+
+	ret = bpf_match_type_btf_id(list, id);
+	if (ret == -ENOSYS || !id) {
+		for (ret = 0; list[ret]; ret++)
+			;
+	}
+
+	return ret;
+}
+EXPORT_SYMBOL_GPL(xdp_meta_match_id);
+
+/* Used in __xdp_build_meta_generic_from_skb() to quickly get the ID
+ * on hotpath.
+ */
+static __le64 xdp_meta_generic_id __ro_after_init;
+
+static int __init xdp_meta_generic_id_init(void)
+{
+	int ret;
+	u64 id;
+
+	ret = bpf_get_type_btf_id("struct xdp_meta_generic", &id);
+	xdp_meta_generic_id = cpu_to_le64(id);
+
+	return ret;
+}
+late_initcall(xdp_meta_generic_id_init);
+
+#define _xdp_meta_rx_hash_type_from_skb(skb, locskb) ({		\
+	typeof(skb) locskb = (skb);				\
+								\
+	likely((locskb)->l4_hash) ? XDP_META_RX_HASH_L4 :	\
+	skb_get_hash_raw(locskb) ? XDP_META_RX_HASH_L3 :	\
+	XDP_META_RX_HASH_NONE;					\
+})
+#define xdp_meta_rx_hash_type_from_skb(skb)			\
+	_xdp_meta_rx_hash_type_from_skb((skb), __UNIQUE_ID(skb_))
+
+#define xdp_meta_rx_vlan_from_prot(skb) ({			\
+	(skb)->vlan_proto == htons(ETH_P_8021Q) ?		\
+	XDP_META_RX_CVID : XDP_META_RX_SVID;			\
+})
+
+#define xdp_meta_rx_vlan_to_prot(md) ({				\
+	xdp_meta_rx_vlan_type_get(md) == XDP_META_RX_CVID ?	\
+	htons(ETH_P_8021Q) : htons(ETH_P_8021AD);		\
+})
+
+/**
+ * ___xdp_build_meta_generic_from_skb - fill a generic metadata from an skb
+ * @rx_md: a pointer to the XDP generic metadata to be filled
+ * @skb: a pointer to the skb to take the info from
+ *
+ * Fills a given generic metadata struct with the info set previously in
+ * an skb. @md can point to anywhere and the function doesn't use the
+ * skb_metadata_{end,len}().
+ */
+void ___xdp_build_meta_generic_from_skb(struct xdp_meta_generic_rx *rx_md,
+					const struct sk_buff *skb)
+{
+	struct xdp_meta_generic *md = to_gen_md(rx_md);
+	ktime_t ts;
+
+	xdp_meta_init(rx_md, xdp_meta_generic_id);
+
+	xdp_meta_rx_csum_level_set(md, skb->csum_level);
+	xdp_meta_rx_csum_status_set(md, skb->ip_summed);
+	xdp_meta_rx_csum_set(md, skb->csum);
+
+	xdp_meta_rx_hash_set(md, skb_get_hash_raw(skb));
+	xdp_meta_rx_hash_type_set(md, xdp_meta_rx_hash_type_from_skb(skb));
+
+	if (likely(skb_rx_queue_recorded(skb))) {
+		xdp_meta_rx_qid_present_set(md, 1);
+		xdp_meta_rx_qid_set(md, skb_get_rx_queue(skb));
+	}
+
+	if (skb_vlan_tag_present(skb)) {
+		xdp_meta_rx_vlan_type_set(md, xdp_meta_rx_vlan_from_prot(skb));
+		xdp_meta_rx_vid_set(md, skb_vlan_tag_get(skb));
+	}
+
+	ts = skb_hwtstamps(skb)->hwtstamp;
+	if (ts) {
+		xdp_meta_rx_tstamp_present_set(md, 1);
+		xdp_meta_rx_tstamp_set(md, ktime_to_ns(ts));
+	}
+}
+EXPORT_SYMBOL_GPL(___xdp_build_meta_generic_from_skb);
+
+/**
+ * ___xdp_populate_skb_meta_generic - fill the skb fields from a generic meta
+ * @skb: a pointer to the skb to be filled
+ * @rx_md: a pointer to the generic metadata to take the values from
+ *
+ * Populates the &sk_buff fields from a given XDP generic metadata. A meta
+ * can be from anywhere, the function doesn't use skb_metadata_{end,len}().
+ * Checks whether the metadata is generic-compatible before accessing other
+ * fields.
+ */
+void ___xdp_populate_skb_meta_generic(struct sk_buff *skb,
+				      const struct xdp_meta_generic_rx *rx_md)
+{
+	const struct xdp_meta_generic *md = to_gen_md(rx_md);
+
+	if (unlikely(!xdp_meta_has_generic(md + 1)))
+		return;
+
+	skb->csum_level = xdp_meta_rx_csum_level_get(md);
+	skb->ip_summed = xdp_meta_rx_csum_status_get(md);
+	skb->csum = xdp_meta_rx_csum_get(md);
+
+	skb_set_hash(skb, xdp_meta_rx_hash_get(md),
+		     xdp_meta_rx_hash_type_get(md));
+
+	if (likely(xdp_meta_rx_qid_present_get(md)))
+		skb_record_rx_queue(skb, xdp_meta_rx_qid_get(md));
+
+	if (xdp_meta_rx_vlan_type_get(md))
+		__vlan_hwaccel_put_tag(skb, xdp_meta_rx_vlan_to_prot(md),
+				       xdp_meta_rx_vid_get(md));
+
+	if (xdp_meta_rx_tstamp_present_get(md))
+		*skb_hwtstamps(skb) = (struct skb_shared_hwtstamps){
+			.hwtstamp = ns_to_ktime(xdp_meta_rx_tstamp_get(md)),
+		};
+}
+EXPORT_SYMBOL_GPL(___xdp_populate_skb_meta_generic);