Re: [PATCH net-next 3/4] mvpp2: add basic XDP support

2020-07-02 Thread Matteo Croce
On Thu, Jul 2, 2020 at 11:14 AM Maciej Fijalkowski
 wrote:
> > > +   if (port->dev->mtu > ETH_DATA_LEN) {
> > > +   netdev_err(port->dev, "Jumbo frames are not supported by XDP, 
> > > current MTU %d.\n",
> > > +  port->dev->mtu);
> >
> > ditto
>
> Here I agree and for every other netdev_err within mvpp2_xdp_setup().
>

Nice idea, I'll add extack error reporting where possible.

-- 
per aspera ad upstream


Re: [PATCH net-next 3/4] mvpp2: add basic XDP support

2020-07-02 Thread Maciej Fijalkowski
On Thu, Jul 02, 2020 at 11:08:19AM +0300, ilias.apalodi...@linaro.org wrote:
> On Tue, Jun 30, 2020 at 08:09:29PM +0200, Matteo Croce wrote:
> > From: Matteo Croce 
> > 
> > Add XDP native support.
> > By now only XDP_DROP, XDP_PASS and XDP_REDIRECT
> > verdicts are supported.
> > 
> > Co-developed-by: Sven Auhagen 
> > Signed-off-by: Sven Auhagen 
> > Signed-off-by: Matteo Croce 
> > ---
> 
> [...]
> 
> >  }
> >  
> > +static int
> > +mvpp2_run_xdp(struct mvpp2_port *port, struct mvpp2_rx_queue *rxq,
> > + struct bpf_prog *prog, struct xdp_buff *xdp,
> > + struct page_pool *pp)
> > +{
> > +   unsigned int len, sync, err;
> > +   struct page *page;
> > +   u32 ret, act;
> > +
> > +   len = xdp->data_end - xdp->data_hard_start - MVPP2_SKB_HEADROOM;
> > +   act = bpf_prog_run_xdp(prog, xdp);
> > +
> > +   /* Due xdp_adjust_tail: DMA sync for_device cover max len CPU touch */
> > +   sync = xdp->data_end - xdp->data_hard_start - MVPP2_SKB_HEADROOM;
> > +   sync = max(sync, len);
> > +
> > +   switch (act) {
> > +   case XDP_PASS:
> > +   ret = MVPP2_XDP_PASS;
> > +   break;
> > +   case XDP_REDIRECT:
> > +   err = xdp_do_redirect(port->dev, xdp, prog);
> > +   if (unlikely(err)) {
> > +   ret = MVPP2_XDP_DROPPED;
> > +   page = virt_to_head_page(xdp->data);
> > +   page_pool_put_page(pp, page, sync, true);
> > +   } else {
> > +   ret = MVPP2_XDP_REDIR;
> > +   }
> > +   break;
> > +   default:
> > +   bpf_warn_invalid_xdp_action(act);
> > +   fallthrough;
> > +   case XDP_ABORTED:
> > +   trace_xdp_exception(port->dev, prog, act);
> > +   fallthrough;
> > +   case XDP_DROP:
> > +   page = virt_to_head_page(xdp->data);
> > +   page_pool_put_page(pp, page, sync, true);
> > +   ret = MVPP2_XDP_DROPPED;
> > +   break;
> > +   }
> > +
> > +   return ret;
> > +}
> > +
> >  /* Main rx processing */
> >  static int mvpp2_rx(struct mvpp2_port *port, struct napi_struct *napi,
> > int rx_todo, struct mvpp2_rx_queue *rxq)
> >  {
> > struct net_device *dev = port->dev;
> > +   struct bpf_prog *xdp_prog;
> > +   struct xdp_buff xdp;
> > int rx_received;
> > int rx_done = 0;
> > +   u32 xdp_ret = 0;
> > u32 rcvd_pkts = 0;
> > u32 rcvd_bytes = 0;
> >  
> > +   rcu_read_lock();
> > +
> > +   xdp_prog = READ_ONCE(port->xdp_prog);
> > +
> > /* Get number of received packets and clamp the to-do */
> > rx_received = mvpp2_rxq_received(port, rxq->id);
> > if (rx_todo > rx_received)
> > @@ -3060,7 +3115,7 @@ static int mvpp2_rx(struct mvpp2_port *port, struct 
> > napi_struct *napi,
> > dma_addr_t dma_addr;
> > phys_addr_t phys_addr;
> > u32 rx_status;
> > -   int pool, rx_bytes, err;
> > +   int pool, rx_bytes, err, ret;
> > void *data;
> >  
> > rx_done++;
> > @@ -3096,6 +3151,33 @@ static int mvpp2_rx(struct mvpp2_port *port, struct 
> > napi_struct *napi,
> > else
> > frag_size = bm_pool->frag_size;
> >  
> > +   if (xdp_prog) {
> > +   xdp.data_hard_start = data;
> > +   xdp.data = data + MVPP2_MH_SIZE + MVPP2_SKB_HEADROOM;
> > +   xdp.data_end = xdp.data + rx_bytes;
> > +   xdp.frame_sz = PAGE_SIZE;
> > +
> > +   if (bm_pool->pkt_size == MVPP2_BM_SHORT_PKT_SIZE)
> > +   xdp.rxq = >xdp_rxq_short;
> > +   else
> > +   xdp.rxq = >xdp_rxq_long;
> > +
> > +   xdp_set_data_meta_invalid();
> > +
> > +   ret = mvpp2_run_xdp(port, rxq, xdp_prog, , pp);
> > +
> > +   if (ret) {
> > +   xdp_ret |= ret;
> > +   err = mvpp2_rx_refill(port, bm_pool, pp, pool);
> > +   if (err) {
> > +   netdev_err(port->dev, "failed to refill 
> > BM pools\n");
> > +   goto err_drop_frame;
> > +   }
> > +
> > +   continue;
> > +   }
> > +   }
> > +
> > skb = build_skb(data, frag_size);
> > if (!skb) {
> > netdev_warn(port->dev, "skb build failed\n");
> > @@ -3118,7 +3200,7 @@ static int mvpp2_rx(struct mvpp2_port *port, struct 
> > napi_struct *napi,
> > rcvd_pkts++;
> > rcvd_bytes += rx_bytes;
> >  
> > -   skb_reserve(skb, MVPP2_MH_SIZE + NET_SKB_PAD);
> > +   skb_reserve(skb, MVPP2_MH_SIZE + MVPP2_SKB_HEADROOM);
> > skb_put(skb, rx_bytes);
> > skb->protocol = eth_type_trans(skb, dev);
> > mvpp2_rx_csum(port, rx_status, skb);
> > @@ -3133,6 +3215,8 @@ static int mvpp2_rx(struct 

Re: [PATCH net-next 3/4] mvpp2: add basic XDP support

2020-07-02 Thread ilias . apalodimas
On Tue, Jun 30, 2020 at 08:09:29PM +0200, Matteo Croce wrote:
> From: Matteo Croce 
> 
> Add XDP native support.
> By now only XDP_DROP, XDP_PASS and XDP_REDIRECT
> verdicts are supported.
> 
> Co-developed-by: Sven Auhagen 
> Signed-off-by: Sven Auhagen 
> Signed-off-by: Matteo Croce 
> ---

[...]

>  }
>  
> +static int
> +mvpp2_run_xdp(struct mvpp2_port *port, struct mvpp2_rx_queue *rxq,
> +   struct bpf_prog *prog, struct xdp_buff *xdp,
> +   struct page_pool *pp)
> +{
> + unsigned int len, sync, err;
> + struct page *page;
> + u32 ret, act;
> +
> + len = xdp->data_end - xdp->data_hard_start - MVPP2_SKB_HEADROOM;
> + act = bpf_prog_run_xdp(prog, xdp);
> +
> + /* Due xdp_adjust_tail: DMA sync for_device cover max len CPU touch */
> + sync = xdp->data_end - xdp->data_hard_start - MVPP2_SKB_HEADROOM;
> + sync = max(sync, len);
> +
> + switch (act) {
> + case XDP_PASS:
> + ret = MVPP2_XDP_PASS;
> + break;
> + case XDP_REDIRECT:
> + err = xdp_do_redirect(port->dev, xdp, prog);
> + if (unlikely(err)) {
> + ret = MVPP2_XDP_DROPPED;
> + page = virt_to_head_page(xdp->data);
> + page_pool_put_page(pp, page, sync, true);
> + } else {
> + ret = MVPP2_XDP_REDIR;
> + }
> + break;
> + default:
> + bpf_warn_invalid_xdp_action(act);
> + fallthrough;
> + case XDP_ABORTED:
> + trace_xdp_exception(port->dev, prog, act);
> + fallthrough;
> + case XDP_DROP:
> + page = virt_to_head_page(xdp->data);
> + page_pool_put_page(pp, page, sync, true);
> + ret = MVPP2_XDP_DROPPED;
> + break;
> + }
> +
> + return ret;
> +}
> +
>  /* Main rx processing */
>  static int mvpp2_rx(struct mvpp2_port *port, struct napi_struct *napi,
>   int rx_todo, struct mvpp2_rx_queue *rxq)
>  {
>   struct net_device *dev = port->dev;
> + struct bpf_prog *xdp_prog;
> + struct xdp_buff xdp;
>   int rx_received;
>   int rx_done = 0;
> + u32 xdp_ret = 0;
>   u32 rcvd_pkts = 0;
>   u32 rcvd_bytes = 0;
>  
> + rcu_read_lock();
> +
> + xdp_prog = READ_ONCE(port->xdp_prog);
> +
>   /* Get number of received packets and clamp the to-do */
>   rx_received = mvpp2_rxq_received(port, rxq->id);
>   if (rx_todo > rx_received)
> @@ -3060,7 +3115,7 @@ static int mvpp2_rx(struct mvpp2_port *port, struct 
> napi_struct *napi,
>   dma_addr_t dma_addr;
>   phys_addr_t phys_addr;
>   u32 rx_status;
> - int pool, rx_bytes, err;
> + int pool, rx_bytes, err, ret;
>   void *data;
>  
>   rx_done++;
> @@ -3096,6 +3151,33 @@ static int mvpp2_rx(struct mvpp2_port *port, struct 
> napi_struct *napi,
>   else
>   frag_size = bm_pool->frag_size;
>  
> + if (xdp_prog) {
> + xdp.data_hard_start = data;
> + xdp.data = data + MVPP2_MH_SIZE + MVPP2_SKB_HEADROOM;
> + xdp.data_end = xdp.data + rx_bytes;
> + xdp.frame_sz = PAGE_SIZE;
> +
> + if (bm_pool->pkt_size == MVPP2_BM_SHORT_PKT_SIZE)
> + xdp.rxq = >xdp_rxq_short;
> + else
> + xdp.rxq = >xdp_rxq_long;
> +
> + xdp_set_data_meta_invalid();
> +
> + ret = mvpp2_run_xdp(port, rxq, xdp_prog, , pp);
> +
> + if (ret) {
> + xdp_ret |= ret;
> + err = mvpp2_rx_refill(port, bm_pool, pp, pool);
> + if (err) {
> + netdev_err(port->dev, "failed to refill 
> BM pools\n");
> + goto err_drop_frame;
> + }
> +
> + continue;
> + }
> + }
> +
>   skb = build_skb(data, frag_size);
>   if (!skb) {
>   netdev_warn(port->dev, "skb build failed\n");
> @@ -3118,7 +3200,7 @@ static int mvpp2_rx(struct mvpp2_port *port, struct 
> napi_struct *napi,
>   rcvd_pkts++;
>   rcvd_bytes += rx_bytes;
>  
> - skb_reserve(skb, MVPP2_MH_SIZE + NET_SKB_PAD);
> + skb_reserve(skb, MVPP2_MH_SIZE + MVPP2_SKB_HEADROOM);
>   skb_put(skb, rx_bytes);
>   skb->protocol = eth_type_trans(skb, dev);
>   mvpp2_rx_csum(port, rx_status, skb);
> @@ -3133,6 +3215,8 @@ static int mvpp2_rx(struct mvpp2_port *port, struct 
> napi_struct *napi,
>   mvpp2_bm_pool_put(port, pool, dma_addr, phys_addr);
>   }
>  
> + rcu_read_unlock();
> +
>   if 

[PATCH net-next 3/4] mvpp2: add basic XDP support

2020-06-30 Thread Matteo Croce
From: Matteo Croce 

Add XDP native support.
By now only XDP_DROP, XDP_PASS and XDP_REDIRECT
verdicts are supported.

Co-developed-by: Sven Auhagen 
Signed-off-by: Sven Auhagen 
Signed-off-by: Matteo Croce 
---
 drivers/net/ethernet/marvell/mvpp2/mvpp2.h|  28 ++-
 .../net/ethernet/marvell/mvpp2/mvpp2_main.c   | 166 +-
 2 files changed, 186 insertions(+), 8 deletions(-)

diff --git a/drivers/net/ethernet/marvell/mvpp2/mvpp2.h 
b/drivers/net/ethernet/marvell/mvpp2/mvpp2.h
index 4c16c9e9c1e5..f351e41c9da6 100644
--- a/drivers/net/ethernet/marvell/mvpp2/mvpp2.h
+++ b/drivers/net/ethernet/marvell/mvpp2/mvpp2.h
@@ -16,6 +16,18 @@
 #include 
 #include 
 #include 
+#include 
+#include 
+
+/* The PacketOffset field is measured in units of 32 bytes and is 3 bits wide,
+ * so the maximum offset is 7 * 32 = 224
+ */
+#define MVPP2_SKB_HEADROOM min(max(XDP_PACKET_HEADROOM, NET_SKB_PAD), 224)
+
+#define MVPP2_XDP_PASS 0
+#define MVPP2_XDP_DROPPED  BIT(0)
+#define MVPP2_XDP_TX   BIT(1)
+#define MVPP2_XDP_REDIRBIT(2)
 
 /* Fifo Registers */
 #define MVPP2_RX_DATA_FIFO_SIZE_REG(port)  (0x00 + 4 * (port))
@@ -629,10 +641,12 @@
ALIGN((mtu) + MVPP2_MH_SIZE + MVPP2_VLAN_TAG_LEN + \
  ETH_HLEN + ETH_FCS_LEN, cache_line_size())
 
-#define MVPP2_RX_BUF_SIZE(pkt_size)((pkt_size) + NET_SKB_PAD)
+#define MVPP2_RX_BUF_SIZE(pkt_size)((pkt_size) + MVPP2_SKB_HEADROOM)
 #define MVPP2_RX_TOTAL_SIZE(buf_size)  ((buf_size) + MVPP2_SKB_SHINFO_SIZE)
 #define MVPP2_RX_MAX_PKT_SIZE(total_size) \
-   ((total_size) - NET_SKB_PAD - MVPP2_SKB_SHINFO_SIZE)
+   ((total_size) - MVPP2_SKB_HEADROOM - MVPP2_SKB_SHINFO_SIZE)
+
+#define MVPP2_MAX_RX_BUF_SIZE  (PAGE_SIZE - MVPP2_SKB_SHINFO_SIZE - 
MVPP2_SKB_HEADROOM)
 
 #define MVPP2_BIT_TO_BYTE(bit) ((bit) / 8)
 #define MVPP2_BIT_TO_WORD(bit) ((bit) / 32)
@@ -690,9 +704,9 @@ enum mvpp2_prs_l3_cast {
 #define MVPP2_BM_COOKIE_POOL_OFFS  8
 #define MVPP2_BM_COOKIE_CPU_OFFS   24
 
-#define MVPP2_BM_SHORT_FRAME_SIZE  512
-#define MVPP2_BM_LONG_FRAME_SIZE   2048
-#define MVPP2_BM_JUMBO_FRAME_SIZE  10240
+#define MVPP2_BM_SHORT_FRAME_SIZE  704 /* frame size 128 */
+#define MVPP2_BM_LONG_FRAME_SIZE   2240/* frame size 1664 */
+#define MVPP2_BM_JUMBO_FRAME_SIZE  10432   /* frame size 9856 */
 /* BM short pool packet size
  * These value assure that for SWF the total number
  * of bytes allocated for each buffer will be 512
@@ -913,6 +927,8 @@ struct mvpp2_port {
unsigned int ntxqs;
struct net_device *dev;
 
+   struct bpf_prog *xdp_prog;
+
int pkt_size;
 
/* Per-CPU port control */
@@ -932,6 +948,8 @@ struct mvpp2_port {
struct mvpp2_pcpu_stats __percpu *stats;
u64 *ethtool_stats;
 
+   unsigned long state;
+
/* Per-port work and its lock to gather hardware statistics */
struct mutex gather_stats_lock;
struct delayed_work stats_work;
diff --git a/drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c 
b/drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
index 9e2e8fb0a0b8..864d4789a0b3 100644
--- a/drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
+++ b/drivers/net/ethernet/marvell/mvpp2/mvpp2_main.c
@@ -36,6 +36,7 @@
 #include 
 #include 
 #include 
+#include 
 
 #include "mvpp2.h"
 #include "mvpp2_prs.h"
@@ -105,6 +106,7 @@ mvpp2_create_page_pool(struct device *dev, int num, int len)
.nid = NUMA_NO_NODE,
.dev = dev,
.dma_dir = DMA_FROM_DEVICE,
+   .offset = MVPP2_SKB_HEADROOM,
.max_len = len,
};
 
@@ -2463,7 +2465,7 @@ static int mvpp2_rxq_init(struct mvpp2_port *port,
put_cpu();
 
/* Set Offset */
-   mvpp2_rxq_offset_set(port, rxq->id, NET_SKB_PAD);
+   mvpp2_rxq_offset_set(port, rxq->id, MVPP2_SKB_HEADROOM);
 
/* Set coalescing pkts and time */
mvpp2_rx_pkts_coal_set(port, rxq);
@@ -3036,16 +3038,69 @@ static u32 mvpp2_skb_tx_csum(struct mvpp2_port *port, 
struct sk_buff *skb)
return MVPP2_TXD_L4_CSUM_NOT | MVPP2_TXD_IP_CSUM_DISABLE;
 }
 
+static int
+mvpp2_run_xdp(struct mvpp2_port *port, struct mvpp2_rx_queue *rxq,
+ struct bpf_prog *prog, struct xdp_buff *xdp,
+ struct page_pool *pp)
+{
+   unsigned int len, sync, err;
+   struct page *page;
+   u32 ret, act;
+
+   len = xdp->data_end - xdp->data_hard_start - MVPP2_SKB_HEADROOM;
+   act = bpf_prog_run_xdp(prog, xdp);
+
+   /* Due xdp_adjust_tail: DMA sync for_device cover max len CPU touch */
+   sync = xdp->data_end - xdp->data_hard_start - MVPP2_SKB_HEADROOM;
+   sync = max(sync, len);
+
+   switch (act) {
+   case XDP_PASS:
+   ret = MVPP2_XDP_PASS;
+   break;
+   case XDP_REDIRECT:
+   err = xdp_do_redirect(port->dev, xdp, prog);
+   if (unlikely(err)) {
+