net/mlx5e: Support RSS for IP-in-IP and IPv6 tunneled packets
authorAya Levin <ayal@mellanox.com>
Mon, 29 Apr 2019 14:45:52 +0000 (17:45 +0300)
committerSaeed Mahameed <saeedm@mellanox.com>
Wed, 28 Aug 2019 18:49:04 +0000 (11:49 -0700)
Add support for inner header RSS on IP-in-IP and IPv6 tunneled packets.

Add rules to the steering table regarding outer IP header, with
IPv4/6->IP-in-IP. Tunneled packets with protocol numbers: 0x4 (IP-in-IP)
and 0x29 (IPv6) are RSS-ed on the inner IP header.
Separate FW dependencies between flow table inner IP capabilities and
GRE offload support. Allowing this feature even if GRE offload is not
supported.  Tested with multi stream TCP traffic tunneled with IPnIP.
Verified that:
Without this patch, only a single RX ring was processing the traffic.
With this patch, multiple RX rings were processing the traffic.
Verified with and without GRE offload support.

Signed-off-by: Aya Levin <ayal@mellanox.com>
Reviewed-by: Tariq Toukan <tariqt@mellanox.com>
Signed-off-by: Saeed Mahameed <saeedm@mellanox.com>
drivers/net/ethernet/mellanox/mlx5/core/en/fs.h
drivers/net/ethernet/mellanox/mlx5/core/en_fs.c

index 5acd982ff228614ca9c7762ca18db144db23bec8..5aae3a7a5497330c359b0cc7831aef4e7b1bc998 100644 (file)
@@ -95,6 +95,10 @@ struct mlx5e_tirc_config {
 enum mlx5e_tunnel_types {
        MLX5E_TT_IPV4_GRE,
        MLX5E_TT_IPV6_GRE,
+       MLX5E_TT_IPV4_IPIP,
+       MLX5E_TT_IPV6_IPIP,
+       MLX5E_TT_IPV4_IPV6,
+       MLX5E_TT_IPV6_IPV6,
        MLX5E_NUM_TUNNEL_TT,
 };
 
index a8340e4fb0b92d3a89b56e4f78b5e4c5dbcc3969..b99b17957543fd827ec82bc657ade5954d31abee 100644 (file)
@@ -747,11 +747,52 @@ static struct mlx5e_etype_proto ttc_tunnel_rules[] = {
                .etype = ETH_P_IPV6,
                .proto = IPPROTO_GRE,
        },
+       [MLX5E_TT_IPV4_IPIP] = {
+               .etype = ETH_P_IP,
+               .proto = IPPROTO_IPIP,
+       },
+       [MLX5E_TT_IPV6_IPIP] = {
+               .etype = ETH_P_IPV6,
+               .proto = IPPROTO_IPIP,
+       },
+       [MLX5E_TT_IPV4_IPV6] = {
+               .etype = ETH_P_IP,
+               .proto = IPPROTO_IPV6,
+       },
+       [MLX5E_TT_IPV6_IPV6] = {
+               .etype = ETH_P_IPV6,
+               .proto = IPPROTO_IPV6,
+       },
+
 };
 
+static bool mlx5e_tunnel_proto_supported(struct mlx5_core_dev *mdev, u8 proto_type)
+{
+       switch (proto_type) {
+       case IPPROTO_GRE:
+               return MLX5_CAP_ETH(mdev, tunnel_stateless_gre);
+       case IPPROTO_IPIP:
+       case IPPROTO_IPV6:
+               return MLX5_CAP_ETH(mdev, tunnel_stateless_ip_over_ip);
+       default:
+               return false;
+       }
+}
+
+static bool mlx5e_any_tunnel_proto_supported(struct mlx5_core_dev *mdev)
+{
+       int tt;
+
+       for (tt = 0; tt < MLX5E_NUM_TUNNEL_TT; tt++) {
+               if (mlx5e_tunnel_proto_supported(mdev, ttc_tunnel_rules[tt].proto))
+                       return true;
+       }
+       return false;
+}
+
 bool mlx5e_tunnel_inner_ft_supported(struct mlx5_core_dev *mdev)
 {
-       return (MLX5_CAP_ETH(mdev, tunnel_stateless_gre) &&
+       return (mlx5e_any_tunnel_proto_supported(mdev) &&
                MLX5_CAP_FLOWTABLE_NIC_RX(mdev, ft_field_support.inner_ip_version));
 }
 
@@ -844,6 +885,9 @@ static int mlx5e_generate_ttc_table_rules(struct mlx5e_priv *priv,
        dest.type = MLX5_FLOW_DESTINATION_TYPE_FLOW_TABLE;
        dest.ft   = params->inner_ttc->ft.t;
        for (tt = 0; tt < MLX5E_NUM_TUNNEL_TT; tt++) {
+               if (!mlx5e_tunnel_proto_supported(priv->mdev,
+                                                 ttc_tunnel_rules[tt].proto))
+                       continue;
                rules[tt] = mlx5e_generate_ttc_rule(priv, ft, &dest,
                                                    ttc_tunnel_rules[tt].etype,
                                                    ttc_tunnel_rules[tt].proto);