Hi, > -----Original Message----- > From: dev <dev-boun...@dpdk.org> On Behalf Of Dekel Peled > Sent: Monday, July 22, 2019 6:37 PM > To: Yongseok Koh <ys...@mellanox.com>; Slava Ovsiienko > <viachesl...@mellanox.com>; Shahaf Shuler <shah...@mellanox.com> > Cc: Jack Min <jack...@mellanox.com>; Ori Kam <or...@mellanox.com>; > dev@dpdk.org; sta...@dpdk.org > Subject: [dpdk-dev] [PATCH v2] net/mlx5: fix NVGRE matching > > NVGRE has a GRE header with c_rsvd0_ver value 0x2000 and protocol value > 0x6558. > These should be matched when item_nvgre is provided. > > This patch adds validation function of NVGRE item. > It also updates the translate function of NVGRE item, to add the required > values, if they were not specified. > > Original work by Xiaoyu Min <jack...@mellanox.com> > > Fixes: fc2c498ccb94 ("net/mlx5: add Direct Verbs translate items") > Cc: sta...@dpdk.org > > Signed-off-by: Dekel Peled <dek...@mellanox.com> > --- > drivers/net/mlx5/mlx5_flow.c | 63 > +++++++++++++++++++++++++++++++++++++++++ > drivers/net/mlx5/mlx5_flow.h | 10 +++++-- > drivers/net/mlx5/mlx5_flow_dv.c | 25 ++++++++++++++-- > drivers/net/mlx5/mlx5_rxtx.h | 2 +- > 4 files changed, 95 insertions(+), 5 deletions(-) > > diff --git a/drivers/net/mlx5/mlx5_flow.c b/drivers/net/mlx5/mlx5_flow.c > index e082cbb..3d2d5fc 100644 > --- a/drivers/net/mlx5/mlx5_flow.c > +++ b/drivers/net/mlx5/mlx5_flow.c > @@ -298,6 +298,10 @@ struct mlx5_flow_tunnel_info { > .tunnel = MLX5_FLOW_LAYER_MPLS, > .ptype = RTE_PTYPE_TUNNEL_MPLS_IN_GRE, > }, > + { > + .tunnel = MLX5_FLOW_LAYER_NVGRE, > + .ptype = RTE_PTYPE_TUNNEL_NVGRE, > + }, > }; > > /** > @@ -1323,6 +1327,11 @@ uint32_t mlx5_flow_adjust_priority(struct > rte_eth_dev *dev, int32_t priority, > return rte_flow_error_set(error, EINVAL, > RTE_FLOW_ERROR_TYPE_ITEM, > item, > "L3 cannot follow an L4 layer."); > + else if ((item_flags & MLX5_FLOW_LAYER_NVGRE) && > + !(item_flags & MLX5_FLOW_LAYER_INNER_L2)) > + return rte_flow_error_set(error, EINVAL, > + RTE_FLOW_ERROR_TYPE_ITEM, > item, > + "L3 cannot follow an NVGRE layer."); > if (!mask) > mask = &rte_flow_item_ipv4_mask; > else if (mask->hdr.next_proto_id != 0 && @@ -1409,6 +1418,11 @@ > uint32_t mlx5_flow_adjust_priority(struct rte_eth_dev *dev, int32_t > priority, > return rte_flow_error_set(error, EINVAL, > RTE_FLOW_ERROR_TYPE_ITEM, > item, > "L3 cannot follow an L4 layer."); > + else if ((item_flags & MLX5_FLOW_LAYER_NVGRE) && > + !(item_flags & MLX5_FLOW_LAYER_INNER_L2)) > + return rte_flow_error_set(error, EINVAL, > + RTE_FLOW_ERROR_TYPE_ITEM, > item, > + "L3 cannot follow an NVGRE layer."); > if (!mask) > mask = &rte_flow_item_ipv6_mask; > ret = mlx5_flow_item_acceptable(item, (const uint8_t *)mask, @@ - > 1887,6 +1901,55 @@ uint32_t mlx5_flow_adjust_priority(struct rte_eth_dev > *dev, int32_t priority, > " update."); > } > > +/** > + * Validate NVGRE item. > + * > + * @param[in] item > + * Item specification. > + * @param[in] item_flags > + * Bit flags to mark detected items. > + * @param[in] target_protocol > + * The next protocol in the previous item. > + * @param[out] error > + * Pointer to error structure. > + * > + * @return > + * 0 on success, a negative errno value otherwise and rte_errno is set. > + */ > +int > +mlx5_flow_validate_item_nvgre(const struct rte_flow_item *item, > + uint64_t item_flags, > + uint8_t target_protocol, > + struct rte_flow_error *error) > +{ > + const struct rte_flow_item_nvgre *mask = item->mask; > + int ret; > + > + if (target_protocol != 0xff && target_protocol != IPPROTO_GRE) > + return rte_flow_error_set(error, EINVAL, > + RTE_FLOW_ERROR_TYPE_ITEM, > item, > + "protocol filtering not compatible" > + " with this GRE layer"); > + if (item_flags & MLX5_FLOW_LAYER_TUNNEL) > + return rte_flow_error_set(error, ENOTSUP, > + RTE_FLOW_ERROR_TYPE_ITEM, > item, > + "multiple tunnel layers not" > + " supported"); > + if (!(item_flags & MLX5_FLOW_LAYER_OUTER_L3)) > + return rte_flow_error_set(error, ENOTSUP, > + RTE_FLOW_ERROR_TYPE_ITEM, > item, > + "L3 Layer is missing"); > + if (!mask) > + mask = &rte_flow_item_nvgre_mask; > + ret = mlx5_flow_item_acceptable > + (item, (const uint8_t *)mask, > + (const uint8_t *)&rte_flow_item_nvgre_mask, > + sizeof(struct rte_flow_item_nvgre), error); > + if (ret < 0) > + return ret; > + return 0; > +} > + > static int > flow_null_validate(struct rte_eth_dev *dev __rte_unused, > const struct rte_flow_attr *attr __rte_unused, diff --git > a/drivers/net/mlx5/mlx5_flow.h b/drivers/net/mlx5/mlx5_flow.h index > 3f96bec..24da74b 100644 > --- a/drivers/net/mlx5/mlx5_flow.h > +++ b/drivers/net/mlx5/mlx5_flow.h > @@ -48,6 +48,7 @@ > #define MLX5_FLOW_LAYER_VXLAN_GPE (1u << 13) #define > MLX5_FLOW_LAYER_GRE (1u << 14) #define MLX5_FLOW_LAYER_MPLS (1u > << 15) > +/* List of tunnel Layer bits continued below. */ > > /* General pattern items bits. */ > #define MLX5_FLOW_ITEM_METADATA (1u << 16) @@ -58,8 +59,10 @@ > #define MLX5_FLOW_LAYER_ICMP6 (1u << 19) #define > MLX5_FLOW_LAYER_GRE_KEY (1u << 20) > > +/* Pattern tunnel Layer bits (continued). */ > #define MLX5_FLOW_LAYER_IPIP (1u << 21) #define > MLX5_FLOW_LAYER_IPV6_ENCAP (1u << 22) > +#define MLX5_FLOW_LAYER_NVGRE (1u << 23) > > /* Outer Masks. */ > #define MLX5_FLOW_LAYER_OUTER_L3 \ > @@ -79,7 +82,7 @@ > /* Tunnel Masks. */ > #define MLX5_FLOW_LAYER_TUNNEL \ > (MLX5_FLOW_LAYER_VXLAN | MLX5_FLOW_LAYER_VXLAN_GPE | \ > - MLX5_FLOW_LAYER_GRE | MLX5_FLOW_LAYER_MPLS | \ > + MLX5_FLOW_LAYER_GRE | MLX5_FLOW_LAYER_NVGRE | > MLX5_FLOW_LAYER_MPLS | > +\ > MLX5_FLOW_LAYER_IPIP | MLX5_FLOW_LAYER_IPV6_ENCAP) > > /* Inner Masks. */ > @@ -518,5 +521,8 @@ int mlx5_flow_validate_item_icmp6(const struct > rte_flow_item *item, > uint64_t item_flags, > uint8_t target_protocol, > struct rte_flow_error *error); > - > +int mlx5_flow_validate_item_nvgre(const struct rte_flow_item *item, > + uint64_t item_flags, > + uint8_t target_protocol, > + struct rte_flow_error *error); > #endif /* RTE_PMD_MLX5_FLOW_H_ */ > diff --git a/drivers/net/mlx5/mlx5_flow_dv.c > b/drivers/net/mlx5/mlx5_flow_dv.c index 7240d3b..f1d32bd 100644 > --- a/drivers/net/mlx5/mlx5_flow_dv.c > +++ b/drivers/net/mlx5/mlx5_flow_dv.c > @@ -2966,7 +2966,6 @@ struct field_modify_info modify_tcp[] = { > > MLX5_FLOW_LAYER_OUTER_L4_UDP; > break; > case RTE_FLOW_ITEM_TYPE_GRE: > - case RTE_FLOW_ITEM_TYPE_NVGRE: > ret = mlx5_flow_validate_item_gre(items, > item_flags, > next_protocol, > error); > if (ret < 0) > @@ -2974,6 +2973,14 @@ struct field_modify_info modify_tcp[] = { > gre_item = items; > last_item = MLX5_FLOW_LAYER_GRE; > break; > + case RTE_FLOW_ITEM_TYPE_NVGRE: > + ret = mlx5_flow_validate_item_nvgre(items, > item_flags, > + next_protocol, > + error); > + if (ret < 0) > + return ret; > + last_item = MLX5_FLOW_LAYER_NVGRE; > + break; > case RTE_FLOW_ITEM_TYPE_GRE_KEY: > ret = mlx5_flow_validate_item_gre_key > (items, item_flags, gre_item, error); @@ - > 3919,7 +3926,21 @@ struct field_modify_info modify_tcp[] = { > int size; > int i; > > - flow_dv_translate_item_gre(matcher, key, item, inner); > + /* For NVGRE, GRE header fields must be set with defined values. */ > + const struct rte_flow_item_gre gre_spec = { > + .c_rsvd0_ver = RTE_BE16(0x2000), > + .protocol = RTE_BE16(RTE_ETHER_TYPE_TEB) > + }; > + const struct rte_flow_item_gre gre_mask = { > + .c_rsvd0_ver = RTE_BE16(0xB000), > + .protocol = RTE_BE16(UINT16_MAX), > + }; > + const struct rte_flow_item gre_item = { > + .spec = &gre_spec, > + .mask = &gre_mask, > + .last = NULL, > + }; > + flow_dv_translate_item_gre(matcher, key, &gre_item, inner); > if (!nvgre_v) > return; > if (!nvgre_m) > diff --git a/drivers/net/mlx5/mlx5_rxtx.h b/drivers/net/mlx5/mlx5_rxtx.h > index 4252832..928d6c3 100644 > --- a/drivers/net/mlx5/mlx5_rxtx.h > +++ b/drivers/net/mlx5/mlx5_rxtx.h > @@ -40,7 +40,7 @@ > #include "mlx5_glue.h" > > /* Support tunnel matching. */ > -#define MLX5_FLOW_TUNNEL 5 > +#define MLX5_FLOW_TUNNEL 6 > > struct mlx5_rxq_stats { > #ifdef MLX5_PMD_SOFT_COUNTERS > -- > 1.8.3.1
Patch applied to next-net-mlx, Kindest regards, Raslan Darawsheh