Re: [PATCH net-next 3/4] mvpp2: add basic XDP support
On Thu, Jul 2, 2020 at 11:14 AM Maciej Fijalkowski wrote: > > > + if (port->dev->mtu > ETH_DATA_LEN) { > > > + netdev_err(port->dev, "Jumbo frames are not supported by XDP, > > > current MTU %d.\n", > > > + port->dev->mtu); > > > > ditto > > Here I agree and for every other netdev_err within mvpp2_xdp_setup(). > Nice idea, I'll add extack error reporting where possible. -- per aspera ad upstream
Re: [PATCH net-next 3/4] mvpp2: add basic XDP support
On Thu, Jul 02, 2020 at 11:08:19AM +0300, ilias.apalodi...@linaro.org wrote: > On Tue, Jun 30, 2020 at 08:09:29PM +0200, Matteo Croce wrote: > > From: Matteo Croce > > > > Add XDP native support. > > By now only XDP_DROP, XDP_PASS and XDP_REDIRECT > > verdicts are supported. > > > > Co-developed-by: Sven Auhagen > > Signed-off-by: Sven Auhagen > > Signed-off-by: Matteo Croce > > --- > > [...] > > > } > > > > +static int > > +mvpp2_run_xdp(struct mvpp2_port *port, struct mvpp2_rx_queue *rxq, > > + struct bpf_prog *prog, struct xdp_buff *xdp, > > + struct page_pool *pp) > > +{ > > + unsigned int len, sync, err; > > + struct page *page; > > + u32 ret, act; > > + > > + len = xdp->data_end - xdp->data_hard_start - MVPP2_SKB_HEADROOM; > > + act = bpf_prog_run_xdp(prog, xdp); > > + > > + /* Due xdp_adjust_tail: DMA sync for_device cover max len CPU touch */ > > + sync = xdp->data_end - xdp->data_hard_start - MVPP2_SKB_HEADROOM; > > + sync = max(sync, len); > > + > > + switch (act) { > > + case XDP_PASS: > > + ret = MVPP2_XDP_PASS; > > + break; > > + case XDP_REDIRECT: > > + err = xdp_do_redirect(port->dev, xdp, prog); > > + if (unlikely(err)) { > > + ret = MVPP2_XDP_DROPPED; > > + page = virt_to_head_page(xdp->data); > > + page_pool_put_page(pp, page, sync, true); > > + } else { > > + ret = MVPP2_XDP_REDIR; > > + } > > + break; > > + default: > > + bpf_warn_invalid_xdp_action(act); > > + fallthrough; > > + case XDP_ABORTED: > > + trace_xdp_exception(port->dev, prog, act); > > + fallthrough; > > + case XDP_DROP: > > + page = virt_to_head_page(xdp->data); > > + page_pool_put_page(pp, page, sync, true); > > + ret = MVPP2_XDP_DROPPED; > > + break; > > + } > > + > > + return ret; > > +} > > + > > /* Main rx processing */ > > static int mvpp2_rx(struct mvpp2_port *port, struct napi_struct *napi, > > int rx_todo, struct mvpp2_rx_queue *rxq) > > { > > struct net_device *dev = port->dev; > > + struct bpf_prog *xdp_prog; > > + struct xdp_buff xdp; > > int rx_received; > > int rx_done = 0; > > + u32 xdp_ret = 0; > > u32 rcvd_pkts = 0; > > u32 rcvd_bytes = 0; > > > > + rcu_read_lock(); > > + > > + xdp_prog = READ_ONCE(port->xdp_prog); > > + > > /* Get number of received packets and clamp the to-do */ > > rx_received = mvpp2_rxq_received(port, rxq->id); > > if (rx_todo > rx_received) > > @@ -3060,7 +3115,7 @@ static int mvpp2_rx(struct mvpp2_port *port, struct > > napi_struct *napi, > > dma_addr_t dma_addr; > > phys_addr_t phys_addr; > > u32 rx_status; > > - int pool, rx_bytes, err; > > + int pool, rx_bytes, err, ret; > > void *data; > > > > rx_done++; > > @@ -3096,6 +3151,33 @@ static int mvpp2_rx(struct mvpp2_port *port, struct > > napi_struct *napi, > > else > > frag_size = bm_pool->frag_size; > > > > + if (xdp_prog) { > > + xdp.data_hard_start = data; > > + xdp.data = data + MVPP2_MH_SIZE + MVPP2_SKB_HEADROOM; > > + xdp.data_end = xdp.data + rx_bytes; > > + xdp.frame_sz = PAGE_SIZE; > > + > > + if (bm_pool->pkt_size == MVPP2_BM_SHORT_PKT_SIZE) > > + xdp.rxq = >xdp_rxq_short; > > + else > > + xdp.rxq = >xdp_rxq_long; > > + > > + xdp_set_data_meta_invalid(); > > + > > + ret = mvpp2_run_xdp(port, rxq, xdp_prog, , pp); > > + > > + if (ret) { > > + xdp_ret |= ret; > > + err = mvpp2_rx_refill(port, bm_pool, pp, pool); > > + if (err) { > > + netdev_err(port->dev, "failed to refill > > BM pools\n"); > > + goto err_drop_frame; > > + } > > + > > + continue; > > + } > > + } > > + > > skb = build_skb(data, frag_size); > > if (!skb) { > > netdev_warn(port->dev, "skb build failed\n"); > > @@ -3118,7 +3200,7 @@ static int mvpp2_rx(struct mvpp2_port *port, struct > > napi_struct *napi, > > rcvd_pkts++; > > rcvd_bytes += rx_bytes; > > > > - skb_reserve(skb, MVPP2_MH_SIZE + NET_SKB_PAD); > > + skb_reserve(skb, MVPP2_MH_SIZE + MVPP2_SKB_HEADROOM); > > skb_put(skb, rx_bytes); > > skb->protocol = eth_type_trans(skb, dev); > > mvpp2_rx_csum(port, rx_status, skb); > > @@ -3133,6 +3215,8 @@ static int mvpp2_rx(struct
Re: [PATCH net-next 3/4] mvpp2: add basic XDP support
On Tue, Jun 30, 2020 at 08:09:29PM +0200, Matteo Croce wrote: > From: Matteo Croce > > Add XDP native support. > By now only XDP_DROP, XDP_PASS and XDP_REDIRECT > verdicts are supported. > > Co-developed-by: Sven Auhagen > Signed-off-by: Sven Auhagen > Signed-off-by: Matteo Croce > --- [...] > } > > +static int > +mvpp2_run_xdp(struct mvpp2_port *port, struct mvpp2_rx_queue *rxq, > + struct bpf_prog *prog, struct xdp_buff *xdp, > + struct page_pool *pp) > +{ > + unsigned int len, sync, err; > + struct page *page; > + u32 ret, act; > + > + len = xdp->data_end - xdp->data_hard_start - MVPP2_SKB_HEADROOM; > + act = bpf_prog_run_xdp(prog, xdp); > + > + /* Due xdp_adjust_tail: DMA sync for_device cover max len CPU touch */ > + sync = xdp->data_end - xdp->data_hard_start - MVPP2_SKB_HEADROOM; > + sync = max(sync, len); > + > + switch (act) { > + case XDP_PASS: > + ret = MVPP2_XDP_PASS; > + break; > + case XDP_REDIRECT: > + err = xdp_do_redirect(port->dev, xdp, prog); > + if (unlikely(err)) { > + ret = MVPP2_XDP_DROPPED; > + page = virt_to_head_page(xdp->data); > + page_pool_put_page(pp, page, sync, true); > + } else { > + ret = MVPP2_XDP_REDIR; > + } > + break; > + default: > + bpf_warn_invalid_xdp_action(act); > + fallthrough; > + case XDP_ABORTED: > + trace_xdp_exception(port->dev, prog, act); > + fallthrough; > + case XDP_DROP: > + page = virt_to_head_page(xdp->data); > + page_pool_put_page(pp, page, sync, true); > + ret = MVPP2_XDP_DROPPED; > + break; > + } > + > + return ret; > +} > + > /* Main rx processing */ > static int mvpp2_rx(struct mvpp2_port *port, struct napi_struct *napi, > int rx_todo, struct mvpp2_rx_queue *rxq) > { > struct net_device *dev = port->dev; > + struct bpf_prog *xdp_prog; > + struct xdp_buff xdp; > int rx_received; > int rx_done = 0; > + u32 xdp_ret = 0; > u32 rcvd_pkts = 0; > u32 rcvd_bytes = 0; > > + rcu_read_lock(); > + > + xdp_prog = READ_ONCE(port->xdp_prog); > + > /* Get number of received packets and clamp the to-do */ > rx_received = mvpp2_rxq_received(port, rxq->id); > if (rx_todo > rx_received) > @@ -3060,7 +3115,7 @@ static int mvpp2_rx(struct mvpp2_port *port, struct > napi_struct *napi, > dma_addr_t dma_addr; > phys_addr_t phys_addr; > u32 rx_status; > - int pool, rx_bytes, err; > + int pool, rx_bytes, err, ret; > void *data; > > rx_done++; > @@ -3096,6 +3151,33 @@ static int mvpp2_rx(struct mvpp2_port *port, struct > napi_struct *napi, > else > frag_size = bm_pool->frag_size; > > + if (xdp_prog) { > + xdp.data_hard_start = data; > + xdp.data = data + MVPP2_MH_SIZE + MVPP2_SKB_HEADROOM; > + xdp.data_end = xdp.data + rx_bytes; > + xdp.frame_sz = PAGE_SIZE; > + > + if (bm_pool->pkt_size == MVPP2_BM_SHORT_PKT_SIZE) > + xdp.rxq = >xdp_rxq_short; > + else > + xdp.rxq = >xdp_rxq_long; > + > + xdp_set_data_meta_invalid(); > + > + ret = mvpp2_run_xdp(port, rxq, xdp_prog, , pp); > + > + if (ret) { > + xdp_ret |= ret; > + err = mvpp2_rx_refill(port, bm_pool, pp, pool); > + if (err) { > + netdev_err(port->dev, "failed to refill > BM pools\n"); > + goto err_drop_frame; > + } > + > + continue; > + } > + } > + > skb = build_skb(data, frag_size); > if (!skb) { > netdev_warn(port->dev, "skb build failed\n"); > @@ -3118,7 +3200,7 @@ static int mvpp2_rx(struct mvpp2_port *port, struct > napi_struct *napi, > rcvd_pkts++; > rcvd_bytes += rx_bytes; > > - skb_reserve(skb, MVPP2_MH_SIZE + NET_SKB_PAD); > + skb_reserve(skb, MVPP2_MH_SIZE + MVPP2_SKB_HEADROOM); > skb_put(skb, rx_bytes); > skb->protocol = eth_type_trans(skb, dev); > mvpp2_rx_csum(port, rx_status, skb); > @@ -3133,6 +3215,8 @@ static int mvpp2_rx(struct mvpp2_port *port, struct > napi_struct *napi, > mvpp2_bm_pool_put(port, pool, dma_addr, phys_addr); > } > > + rcu_read_unlock(); > + > if
[PATCH net-next 3/4] mvpp2: add basic XDP support
From: Matteo Croce Add XDP native support. By now only XDP_DROP, XDP_PASS and XDP_REDIRECT verdicts are supported. Co-developed-by: Sven Auhagen Signed-off-by: Sven Auhagen Signed-off-by: Matteo Croce --- drivers/net/ethernet/marvell/mvpp2/mvpp2.h| 28 ++- .../net/ethernet/marvell/mvpp2/mvpp2_main.c | 166 +- 2 files changed, 186 insertions(+), 8 deletions(-) diff --git a/drivers/net/ethernet/marvell/mvpp2/mvpp2.h b/drivers/net/ethernet/marvell/mvpp2/mvpp2.h index 4c16c9e9c1e5..f351e41c9da6 100644 --- a/drivers/net/ethernet/marvell/mvpp2/mvpp2.h +++ b/drivers/net/ethernet/marvell/mvpp2/mvpp2.h @@ -16,6 +16,18 @@ #include #include #include +#include +#include + +/* The PacketOffset field is measured in units of 32 bytes and is 3 bits wide, + * so the maximum offset is 7 * 32 = 224 + */ +#define MVPP2_SKB_HEADROOM min(max(XDP_PACKET_HEADROOM, NET_SKB_PAD), 224) + +#define MVPP2_XDP_PASS 0 +#define MVPP2_XDP_DROPPED BIT(0) +#define MVPP2_XDP_TX BIT(1) +#define MVPP2_XDP_REDIRBIT(2) /* Fifo Registers */ #define MVPP2_RX_DATA_FIFO_SIZE_REG(port) (0x00 + 4 * (port)) @@ -629,10 +641,12 @@ ALIGN((mtu) + MVPP2_MH_SIZE + MVPP2_VLAN_TAG_LEN + \ ETH_HLEN + ETH_FCS_LEN, cache_line_size()) -#define MVPP2_RX_BUF_SIZE(pkt_size)((pkt_size) + NET_SKB_PAD) +#define MVPP2_RX_BUF_SIZE(pkt_size)((pkt_size) + MVPP2_SKB_HEADROOM) #define MVPP2_RX_TOTAL_SIZE(buf_size) ((buf_size) + MVPP2_SKB_SHINFO_SIZE) #define MVPP2_RX_MAX_PKT_SIZE(total_size) \ - ((total_size) - NET_SKB_PAD - MVPP2_SKB_SHINFO_SIZE) + ((total_size) - MVPP2_SKB_HEADROOM - MVPP2_SKB_SHINFO_SIZE) + +#define MVPP2_MAX_RX_BUF_SIZE (PAGE_SIZE - MVPP2_SKB_SHINFO_SIZE - MVPP2_SKB_HEADROOM) #define MVPP2_BIT_TO_BYTE(bit) ((bit) / 8) #define MVPP2_BIT_TO_WORD(bit) ((bit) / 32) @@ -690,9 +704,9 @@ enum mvpp2_prs_l3_cast { #define MVPP2_BM_COOKIE_POOL_OFFS 8 #define MVPP2_BM_COOKIE_CPU_OFFS 24 -#define MVPP2_BM_SHORT_FRAME_SIZE 512 -#define MVPP2_BM_LONG_FRAME_SIZE 2048 -#define MVPP2_BM_JUMBO_FRAME_SIZE 10240 +#define MVPP2_BM_SHORT_FRAME_SIZE 704 /* frame size 128 */ +#define MVPP2_BM_LONG_FRAME_SIZE 2240/* frame size 1664 */ +#define MVPP2_BM_JUMBO_FRAME_SIZE 10432 /* frame size 9856 */ /* BM short pool packet size * These value assure that for SWF the total number * of bytes allocated for each buffer will be 512 @@ -913,6 +927,8 @@ struct mvpp2_port { unsigned int ntxqs; struct net_device *dev; + struct bpf_prog *xdp_prog; + int pkt_size; /* Per-CPU port control */ @@ -932,6 +948,8 @@ struct mvpp2_port { struct mvpp2_pcpu_stats __percpu *stats; u64 *ethtool_stats; + unsigned long state; + /* Per-port work and its lock to gather hardware statistics */ struct mutex gather_stats_lock; struct delayed_work stats_work; diff --git a/drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c b/drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c index 9e2e8fb0a0b8..864d4789a0b3 100644 --- a/drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c +++ b/drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c @@ -36,6 +36,7 @@ #include #include #include +#include #include "mvpp2.h" #include "mvpp2_prs.h" @@ -105,6 +106,7 @@ mvpp2_create_page_pool(struct device *dev, int num, int len) .nid = NUMA_NO_NODE, .dev = dev, .dma_dir = DMA_FROM_DEVICE, + .offset = MVPP2_SKB_HEADROOM, .max_len = len, }; @@ -2463,7 +2465,7 @@ static int mvpp2_rxq_init(struct mvpp2_port *port, put_cpu(); /* Set Offset */ - mvpp2_rxq_offset_set(port, rxq->id, NET_SKB_PAD); + mvpp2_rxq_offset_set(port, rxq->id, MVPP2_SKB_HEADROOM); /* Set coalescing pkts and time */ mvpp2_rx_pkts_coal_set(port, rxq); @@ -3036,16 +3038,69 @@ static u32 mvpp2_skb_tx_csum(struct mvpp2_port *port, struct sk_buff *skb) return MVPP2_TXD_L4_CSUM_NOT | MVPP2_TXD_IP_CSUM_DISABLE; } +static int +mvpp2_run_xdp(struct mvpp2_port *port, struct mvpp2_rx_queue *rxq, + struct bpf_prog *prog, struct xdp_buff *xdp, + struct page_pool *pp) +{ + unsigned int len, sync, err; + struct page *page; + u32 ret, act; + + len = xdp->data_end - xdp->data_hard_start - MVPP2_SKB_HEADROOM; + act = bpf_prog_run_xdp(prog, xdp); + + /* Due xdp_adjust_tail: DMA sync for_device cover max len CPU touch */ + sync = xdp->data_end - xdp->data_hard_start - MVPP2_SKB_HEADROOM; + sync = max(sync, len); + + switch (act) { + case XDP_PASS: + ret = MVPP2_XDP_PASS; + break; + case XDP_REDIRECT: + err = xdp_do_redirect(port->dev, xdp, prog); + if (unlikely(err)) { +