1/*
2 * Copyright (c) 2016, Mellanox Technologies. All rights reserved.
3 *
4 * This software is available to you under a choice of one of two
5 * licenses. You may choose to be licensed under the terms of the GNU
6 * General Public License (GPL) Version 2, available from the file
7 * COPYING in the main directory of this source tree, or the
8 * OpenIB.org BSD license below:
9 *
10 * Redistribution and use in source and binary forms, with or
11 * without modification, are permitted provided that the following
12 * conditions are met:
13 *
14 * - Redistributions of source code must retain the above
15 * copyright notice, this list of conditions and the following
16 * disclaimer.
17 *
18 * - Redistributions in binary form must reproduce the above
19 * copyright notice, this list of conditions and the following
20 * disclaimer in the documentation and/or other materials
21 * provided with the distribution.
22 *
23 * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
24 * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
25 * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
26 * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
27 * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
28 * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
29 * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
30 * SOFTWARE.
31 */
32
33#include <linux/mlx5/fs.h>
34#include "en.h"
35#include "en/params.h"
36#include "en/xsk/pool.h"
37#include "en/fs_ethtool.h"
38
39struct mlx5e_ethtool_table {
40 struct mlx5_flow_table *ft;
41 int num_rules;
42};
43
44#define ETHTOOL_NUM_L3_L4_FTS 7
45#define ETHTOOL_NUM_L2_FTS 4
46
47struct mlx5e_ethtool_steering {
48 struct mlx5e_ethtool_table l3_l4_ft[ETHTOOL_NUM_L3_L4_FTS];
49 struct mlx5e_ethtool_table l2_ft[ETHTOOL_NUM_L2_FTS];
50 struct list_head rules;
51 int tot_num_rules;
52};
53
54static int flow_type_to_traffic_type(u32 flow_type);
55
56static u32 flow_type_mask(u32 flow_type)
57{
58 return flow_type & ~(FLOW_EXT | FLOW_MAC_EXT | FLOW_RSS);
59}
60
61struct mlx5e_ethtool_rule {
62 struct list_head list;
63 struct ethtool_rx_flow_spec flow_spec;
64 struct mlx5_flow_handle *rule;
65 struct mlx5e_ethtool_table *eth_ft;
66 struct mlx5e_rss *rss;
67};
68
69static void put_flow_table(struct mlx5e_ethtool_table *eth_ft)
70{
71 if (!--eth_ft->num_rules) {
72 mlx5_destroy_flow_table(ft: eth_ft->ft);
73 eth_ft->ft = NULL;
74 }
75}
76
77#define MLX5E_ETHTOOL_L3_L4_PRIO 0
78#define MLX5E_ETHTOOL_L2_PRIO (MLX5E_ETHTOOL_L3_L4_PRIO + ETHTOOL_NUM_L3_L4_FTS)
79#define MLX5E_ETHTOOL_NUM_ENTRIES 64000
80#define MLX5E_ETHTOOL_NUM_GROUPS 10
81static struct mlx5e_ethtool_table *get_flow_table(struct mlx5e_priv *priv,
82 struct ethtool_rx_flow_spec *fs,
83 int num_tuples)
84{
85 struct mlx5e_ethtool_steering *ethtool = mlx5e_fs_get_ethtool(fs: priv->fs);
86 struct mlx5_flow_table_attr ft_attr = {};
87 struct mlx5e_ethtool_table *eth_ft;
88 struct mlx5_flow_namespace *ns;
89 struct mlx5_flow_table *ft;
90 int max_tuples;
91 int table_size;
92 int prio;
93
94 switch (flow_type_mask(flow_type: fs->flow_type)) {
95 case TCP_V4_FLOW:
96 case UDP_V4_FLOW:
97 case TCP_V6_FLOW:
98 case UDP_V6_FLOW:
99 case IP_USER_FLOW:
100 case IPV6_USER_FLOW:
101 max_tuples = ETHTOOL_NUM_L3_L4_FTS;
102 prio = MLX5E_ETHTOOL_L3_L4_PRIO + (max_tuples - num_tuples);
103 eth_ft = &ethtool->l3_l4_ft[prio];
104 break;
105 case ETHER_FLOW:
106 max_tuples = ETHTOOL_NUM_L2_FTS;
107 prio = max_tuples - num_tuples;
108 eth_ft = &ethtool->l2_ft[prio];
109 prio += MLX5E_ETHTOOL_L2_PRIO;
110 break;
111 default:
112 return ERR_PTR(error: -EINVAL);
113 }
114
115 eth_ft->num_rules++;
116 if (eth_ft->ft)
117 return eth_ft;
118
119 ns = mlx5_get_flow_namespace(dev: priv->mdev,
120 type: MLX5_FLOW_NAMESPACE_ETHTOOL);
121 if (!ns)
122 return ERR_PTR(error: -EOPNOTSUPP);
123
124 table_size = min_t(u32, BIT(MLX5_CAP_FLOWTABLE(priv->mdev,
125 flow_table_properties_nic_receive.log_max_ft_size)),
126 MLX5E_ETHTOOL_NUM_ENTRIES);
127
128 ft_attr.prio = prio;
129 ft_attr.max_fte = table_size;
130 ft_attr.autogroup.max_num_groups = MLX5E_ETHTOOL_NUM_GROUPS;
131 ft = mlx5_create_auto_grouped_flow_table(ns, ft_attr: &ft_attr);
132 if (IS_ERR(ptr: ft))
133 return (void *)ft;
134
135 eth_ft->ft = ft;
136 return eth_ft;
137}
138
139static void mask_spec(u8 *mask, u8 *val, size_t size)
140{
141 unsigned int i;
142
143 for (i = 0; i < size; i++, mask++, val++)
144 *((u8 *)val) = *((u8 *)mask) & *((u8 *)val);
145}
146
147#define MLX5E_FTE_SET(header_p, fld, v) \
148 MLX5_SET(fte_match_set_lyr_2_4, header_p, fld, v)
149
150#define MLX5E_FTE_ADDR_OF(header_p, fld) \
151 MLX5_ADDR_OF(fte_match_set_lyr_2_4, header_p, fld)
152
153static void
154set_ip4(void *headers_c, void *headers_v, __be32 ip4src_m,
155 __be32 ip4src_v, __be32 ip4dst_m, __be32 ip4dst_v)
156{
157 if (ip4src_m) {
158 memcpy(MLX5E_FTE_ADDR_OF(headers_v, src_ipv4_src_ipv6.ipv4_layout.ipv4),
159 &ip4src_v, sizeof(ip4src_v));
160 memcpy(MLX5E_FTE_ADDR_OF(headers_c, src_ipv4_src_ipv6.ipv4_layout.ipv4),
161 &ip4src_m, sizeof(ip4src_m));
162 }
163 if (ip4dst_m) {
164 memcpy(MLX5E_FTE_ADDR_OF(headers_v, dst_ipv4_dst_ipv6.ipv4_layout.ipv4),
165 &ip4dst_v, sizeof(ip4dst_v));
166 memcpy(MLX5E_FTE_ADDR_OF(headers_c, dst_ipv4_dst_ipv6.ipv4_layout.ipv4),
167 &ip4dst_m, sizeof(ip4dst_m));
168 }
169
170 MLX5E_FTE_SET(headers_c, ethertype, 0xffff);
171 MLX5E_FTE_SET(headers_v, ethertype, ETH_P_IP);
172}
173
174static void
175set_ip6(void *headers_c, void *headers_v, __be32 ip6src_m[4],
176 __be32 ip6src_v[4], __be32 ip6dst_m[4], __be32 ip6dst_v[4])
177{
178 u8 ip6_sz = MLX5_FLD_SZ_BYTES(ipv6_layout, ipv6);
179
180 if (!ipv6_addr_any(a: (struct in6_addr *)ip6src_m)) {
181 memcpy(MLX5E_FTE_ADDR_OF(headers_v, src_ipv4_src_ipv6.ipv6_layout.ipv6),
182 ip6src_v, ip6_sz);
183 memcpy(MLX5E_FTE_ADDR_OF(headers_c, src_ipv4_src_ipv6.ipv6_layout.ipv6),
184 ip6src_m, ip6_sz);
185 }
186 if (!ipv6_addr_any(a: (struct in6_addr *)ip6dst_m)) {
187 memcpy(MLX5E_FTE_ADDR_OF(headers_v, dst_ipv4_dst_ipv6.ipv6_layout.ipv6),
188 ip6dst_v, ip6_sz);
189 memcpy(MLX5E_FTE_ADDR_OF(headers_c, dst_ipv4_dst_ipv6.ipv6_layout.ipv6),
190 ip6dst_m, ip6_sz);
191 }
192
193 MLX5E_FTE_SET(headers_c, ethertype, 0xffff);
194 MLX5E_FTE_SET(headers_v, ethertype, ETH_P_IPV6);
195}
196
197static void
198set_tcp(void *headers_c, void *headers_v, __be16 psrc_m, __be16 psrc_v,
199 __be16 pdst_m, __be16 pdst_v)
200{
201 if (psrc_m) {
202 MLX5E_FTE_SET(headers_c, tcp_sport, ntohs(psrc_m));
203 MLX5E_FTE_SET(headers_v, tcp_sport, ntohs(psrc_v));
204 }
205 if (pdst_m) {
206 MLX5E_FTE_SET(headers_c, tcp_dport, ntohs(pdst_m));
207 MLX5E_FTE_SET(headers_v, tcp_dport, ntohs(pdst_v));
208 }
209
210 MLX5E_FTE_SET(headers_c, ip_protocol, 0xffff);
211 MLX5E_FTE_SET(headers_v, ip_protocol, IPPROTO_TCP);
212}
213
214static void
215set_udp(void *headers_c, void *headers_v, __be16 psrc_m, __be16 psrc_v,
216 __be16 pdst_m, __be16 pdst_v)
217{
218 if (psrc_m) {
219 MLX5E_FTE_SET(headers_c, udp_sport, ntohs(psrc_m));
220 MLX5E_FTE_SET(headers_v, udp_sport, ntohs(psrc_v));
221 }
222
223 if (pdst_m) {
224 MLX5E_FTE_SET(headers_c, udp_dport, ntohs(pdst_m));
225 MLX5E_FTE_SET(headers_v, udp_dport, ntohs(pdst_v));
226 }
227
228 MLX5E_FTE_SET(headers_c, ip_protocol, 0xffff);
229 MLX5E_FTE_SET(headers_v, ip_protocol, IPPROTO_UDP);
230}
231
232static void
233parse_tcp4(void *headers_c, void *headers_v, struct ethtool_rx_flow_spec *fs)
234{
235 struct ethtool_tcpip4_spec *l4_mask = &fs->m_u.tcp_ip4_spec;
236 struct ethtool_tcpip4_spec *l4_val = &fs->h_u.tcp_ip4_spec;
237
238 set_ip4(headers_c, headers_v, ip4src_m: l4_mask->ip4src, ip4src_v: l4_val->ip4src,
239 ip4dst_m: l4_mask->ip4dst, ip4dst_v: l4_val->ip4dst);
240
241 set_tcp(headers_c, headers_v, psrc_m: l4_mask->psrc, psrc_v: l4_val->psrc,
242 pdst_m: l4_mask->pdst, pdst_v: l4_val->pdst);
243}
244
245static void
246parse_udp4(void *headers_c, void *headers_v, struct ethtool_rx_flow_spec *fs)
247{
248 struct ethtool_tcpip4_spec *l4_mask = &fs->m_u.udp_ip4_spec;
249 struct ethtool_tcpip4_spec *l4_val = &fs->h_u.udp_ip4_spec;
250
251 set_ip4(headers_c, headers_v, ip4src_m: l4_mask->ip4src, ip4src_v: l4_val->ip4src,
252 ip4dst_m: l4_mask->ip4dst, ip4dst_v: l4_val->ip4dst);
253
254 set_udp(headers_c, headers_v, psrc_m: l4_mask->psrc, psrc_v: l4_val->psrc,
255 pdst_m: l4_mask->pdst, pdst_v: l4_val->pdst);
256}
257
258static void
259parse_ip4(void *headers_c, void *headers_v, struct ethtool_rx_flow_spec *fs)
260{
261 struct ethtool_usrip4_spec *l3_mask = &fs->m_u.usr_ip4_spec;
262 struct ethtool_usrip4_spec *l3_val = &fs->h_u.usr_ip4_spec;
263
264 set_ip4(headers_c, headers_v, ip4src_m: l3_mask->ip4src, ip4src_v: l3_val->ip4src,
265 ip4dst_m: l3_mask->ip4dst, ip4dst_v: l3_val->ip4dst);
266
267 if (l3_mask->proto) {
268 MLX5E_FTE_SET(headers_c, ip_protocol, l3_mask->proto);
269 MLX5E_FTE_SET(headers_v, ip_protocol, l3_val->proto);
270 }
271}
272
273static void
274parse_ip6(void *headers_c, void *headers_v, struct ethtool_rx_flow_spec *fs)
275{
276 struct ethtool_usrip6_spec *l3_mask = &fs->m_u.usr_ip6_spec;
277 struct ethtool_usrip6_spec *l3_val = &fs->h_u.usr_ip6_spec;
278
279 set_ip6(headers_c, headers_v, ip6src_m: l3_mask->ip6src,
280 ip6src_v: l3_val->ip6src, ip6dst_m: l3_mask->ip6dst, ip6dst_v: l3_val->ip6dst);
281
282 if (l3_mask->l4_proto) {
283 MLX5E_FTE_SET(headers_c, ip_protocol, l3_mask->l4_proto);
284 MLX5E_FTE_SET(headers_v, ip_protocol, l3_val->l4_proto);
285 }
286}
287
288static void
289parse_tcp6(void *headers_c, void *headers_v, struct ethtool_rx_flow_spec *fs)
290{
291 struct ethtool_tcpip6_spec *l4_mask = &fs->m_u.tcp_ip6_spec;
292 struct ethtool_tcpip6_spec *l4_val = &fs->h_u.tcp_ip6_spec;
293
294 set_ip6(headers_c, headers_v, ip6src_m: l4_mask->ip6src,
295 ip6src_v: l4_val->ip6src, ip6dst_m: l4_mask->ip6dst, ip6dst_v: l4_val->ip6dst);
296
297 set_tcp(headers_c, headers_v, psrc_m: l4_mask->psrc, psrc_v: l4_val->psrc,
298 pdst_m: l4_mask->pdst, pdst_v: l4_val->pdst);
299}
300
301static void
302parse_udp6(void *headers_c, void *headers_v, struct ethtool_rx_flow_spec *fs)
303{
304 struct ethtool_tcpip6_spec *l4_mask = &fs->m_u.udp_ip6_spec;
305 struct ethtool_tcpip6_spec *l4_val = &fs->h_u.udp_ip6_spec;
306
307 set_ip6(headers_c, headers_v, ip6src_m: l4_mask->ip6src,
308 ip6src_v: l4_val->ip6src, ip6dst_m: l4_mask->ip6dst, ip6dst_v: l4_val->ip6dst);
309
310 set_udp(headers_c, headers_v, psrc_m: l4_mask->psrc, psrc_v: l4_val->psrc,
311 pdst_m: l4_mask->pdst, pdst_v: l4_val->pdst);
312}
313
314static void
315parse_ether(void *headers_c, void *headers_v, struct ethtool_rx_flow_spec *fs)
316{
317 struct ethhdr *eth_mask = &fs->m_u.ether_spec;
318 struct ethhdr *eth_val = &fs->h_u.ether_spec;
319
320 mask_spec(mask: (u8 *)eth_mask, val: (u8 *)eth_val, size: sizeof(*eth_mask));
321 ether_addr_copy(MLX5E_FTE_ADDR_OF(headers_c, smac_47_16), src: eth_mask->h_source);
322 ether_addr_copy(MLX5E_FTE_ADDR_OF(headers_v, smac_47_16), src: eth_val->h_source);
323 ether_addr_copy(MLX5E_FTE_ADDR_OF(headers_c, dmac_47_16), src: eth_mask->h_dest);
324 ether_addr_copy(MLX5E_FTE_ADDR_OF(headers_v, dmac_47_16), src: eth_val->h_dest);
325 MLX5E_FTE_SET(headers_c, ethertype, ntohs(eth_mask->h_proto));
326 MLX5E_FTE_SET(headers_v, ethertype, ntohs(eth_val->h_proto));
327}
328
329static void
330set_cvlan(void *headers_c, void *headers_v, __be16 vlan_tci)
331{
332 MLX5E_FTE_SET(headers_c, cvlan_tag, 1);
333 MLX5E_FTE_SET(headers_v, cvlan_tag, 1);
334 MLX5E_FTE_SET(headers_c, first_vid, 0xfff);
335 MLX5E_FTE_SET(headers_v, first_vid, ntohs(vlan_tci));
336}
337
338static void
339set_dmac(void *headers_c, void *headers_v,
340 unsigned char m_dest[ETH_ALEN], unsigned char v_dest[ETH_ALEN])
341{
342 ether_addr_copy(MLX5E_FTE_ADDR_OF(headers_c, dmac_47_16), src: m_dest);
343 ether_addr_copy(MLX5E_FTE_ADDR_OF(headers_v, dmac_47_16), src: v_dest);
344}
345
346static int set_flow_attrs(u32 *match_c, u32 *match_v,
347 struct ethtool_rx_flow_spec *fs)
348{
349 void *outer_headers_c = MLX5_ADDR_OF(fte_match_param, match_c,
350 outer_headers);
351 void *outer_headers_v = MLX5_ADDR_OF(fte_match_param, match_v,
352 outer_headers);
353 u32 flow_type = flow_type_mask(flow_type: fs->flow_type);
354
355 switch (flow_type) {
356 case TCP_V4_FLOW:
357 parse_tcp4(headers_c: outer_headers_c, headers_v: outer_headers_v, fs);
358 break;
359 case UDP_V4_FLOW:
360 parse_udp4(headers_c: outer_headers_c, headers_v: outer_headers_v, fs);
361 break;
362 case IP_USER_FLOW:
363 parse_ip4(headers_c: outer_headers_c, headers_v: outer_headers_v, fs);
364 break;
365 case TCP_V6_FLOW:
366 parse_tcp6(headers_c: outer_headers_c, headers_v: outer_headers_v, fs);
367 break;
368 case UDP_V6_FLOW:
369 parse_udp6(headers_c: outer_headers_c, headers_v: outer_headers_v, fs);
370 break;
371 case IPV6_USER_FLOW:
372 parse_ip6(headers_c: outer_headers_c, headers_v: outer_headers_v, fs);
373 break;
374 case ETHER_FLOW:
375 parse_ether(headers_c: outer_headers_c, headers_v: outer_headers_v, fs);
376 break;
377 default:
378 return -EINVAL;
379 }
380
381 if ((fs->flow_type & FLOW_EXT) &&
382 (fs->m_ext.vlan_tci & cpu_to_be16(VLAN_VID_MASK)))
383 set_cvlan(headers_c: outer_headers_c, headers_v: outer_headers_v, vlan_tci: fs->h_ext.vlan_tci);
384
385 if (fs->flow_type & FLOW_MAC_EXT &&
386 !is_zero_ether_addr(addr: fs->m_ext.h_dest)) {
387 mask_spec(mask: fs->m_ext.h_dest, val: fs->h_ext.h_dest, ETH_ALEN);
388 set_dmac(headers_c: outer_headers_c, headers_v: outer_headers_v, m_dest: fs->m_ext.h_dest,
389 v_dest: fs->h_ext.h_dest);
390 }
391
392 return 0;
393}
394
395static void add_rule_to_list(struct mlx5e_priv *priv,
396 struct mlx5e_ethtool_rule *rule)
397{
398 struct mlx5e_ethtool_steering *ethtool = mlx5e_fs_get_ethtool(fs: priv->fs);
399 struct list_head *head = &ethtool->rules;
400 struct mlx5e_ethtool_rule *iter;
401
402 list_for_each_entry(iter, &ethtool->rules, list) {
403 if (iter->flow_spec.location > rule->flow_spec.location)
404 break;
405 head = &iter->list;
406 }
407 ethtool->tot_num_rules++;
408 list_add(new: &rule->list, head);
409}
410
411static bool outer_header_zero(u32 *match_criteria)
412{
413 int size = MLX5_FLD_SZ_BYTES(fte_match_param, outer_headers);
414 char *outer_headers_c = MLX5_ADDR_OF(fte_match_param, match_criteria,
415 outer_headers);
416
417 return outer_headers_c[0] == 0 && !memcmp(p: outer_headers_c,
418 q: outer_headers_c + 1,
419 size: size - 1);
420}
421
422static int flow_get_tirn(struct mlx5e_priv *priv,
423 struct mlx5e_ethtool_rule *eth_rule,
424 struct ethtool_rx_flow_spec *fs,
425 u32 rss_context, u32 *tirn)
426{
427 if (fs->flow_type & FLOW_RSS) {
428 struct mlx5e_packet_merge_param pkt_merge_param;
429 struct mlx5e_rss *rss;
430 u32 flow_type;
431 int err;
432 int tt;
433
434 rss = mlx5e_rx_res_rss_get(res: priv->rx_res, rss_idx: rss_context);
435 if (!rss)
436 return -ENOENT;
437
438 flow_type = flow_type_mask(flow_type: fs->flow_type);
439 tt = flow_type_to_traffic_type(flow_type);
440 if (tt < 0)
441 return -EINVAL;
442
443 pkt_merge_param = priv->channels.params.packet_merge;
444 err = mlx5e_rss_obtain_tirn(rss, tt, init_pkt_merge_param: &pkt_merge_param, inner: false, tirn);
445 if (err)
446 return err;
447 eth_rule->rss = rss;
448 mlx5e_rss_refcnt_inc(rss: eth_rule->rss);
449 } else {
450 *tirn = mlx5e_rx_res_get_tirn_direct(res: priv->rx_res, ix: fs->ring_cookie);
451 }
452
453 return 0;
454}
455
456static struct mlx5_flow_handle *
457add_ethtool_flow_rule(struct mlx5e_priv *priv,
458 struct mlx5e_ethtool_rule *eth_rule,
459 struct mlx5_flow_table *ft,
460 struct ethtool_rx_flow_spec *fs, u32 rss_context)
461{
462 struct mlx5_flow_act flow_act = { .flags = FLOW_ACT_NO_APPEND };
463 struct mlx5_flow_destination *dst = NULL;
464 struct mlx5_flow_handle *rule;
465 struct mlx5_flow_spec *spec;
466 int err = 0;
467
468 spec = kvzalloc(size: sizeof(*spec), GFP_KERNEL);
469 if (!spec)
470 return ERR_PTR(error: -ENOMEM);
471 err = set_flow_attrs(match_c: spec->match_criteria, match_v: spec->match_value,
472 fs);
473 if (err)
474 goto free;
475
476 if (fs->ring_cookie == RX_CLS_FLOW_DISC) {
477 flow_act.action = MLX5_FLOW_CONTEXT_ACTION_DROP;
478 } else {
479 dst = kzalloc(size: sizeof(*dst), GFP_KERNEL);
480 if (!dst) {
481 err = -ENOMEM;
482 goto free;
483 }
484
485 err = flow_get_tirn(priv, eth_rule, fs, rss_context, tirn: &dst->tir_num);
486 if (err)
487 goto free;
488
489 dst->type = MLX5_FLOW_DESTINATION_TYPE_TIR;
490 flow_act.action = MLX5_FLOW_CONTEXT_ACTION_FWD_DEST;
491 }
492
493 spec->match_criteria_enable = (!outer_header_zero(match_criteria: spec->match_criteria));
494 spec->flow_context.flow_tag = MLX5_FS_DEFAULT_FLOW_TAG;
495 rule = mlx5_add_flow_rules(ft, spec, flow_act: &flow_act, dest: dst, num_dest: dst ? 1 : 0);
496 if (IS_ERR(ptr: rule)) {
497 err = PTR_ERR(ptr: rule);
498 netdev_err(dev: priv->netdev, format: "%s: failed to add ethtool steering rule: %d\n",
499 __func__, err);
500 goto free;
501 }
502free:
503 kvfree(addr: spec);
504 kfree(objp: dst);
505 return err ? ERR_PTR(error: err) : rule;
506}
507
508static void del_ethtool_rule(struct mlx5e_flow_steering *fs,
509 struct mlx5e_ethtool_rule *eth_rule)
510{
511 struct mlx5e_ethtool_steering *ethtool = mlx5e_fs_get_ethtool(fs);
512 if (eth_rule->rule)
513 mlx5_del_flow_rules(fr: eth_rule->rule);
514 if (eth_rule->rss)
515 mlx5e_rss_refcnt_dec(rss: eth_rule->rss);
516 list_del(entry: &eth_rule->list);
517 ethtool->tot_num_rules--;
518 put_flow_table(eth_ft: eth_rule->eth_ft);
519 kfree(objp: eth_rule);
520}
521
522static struct mlx5e_ethtool_rule *find_ethtool_rule(struct mlx5e_priv *priv,
523 int location)
524{
525 struct mlx5e_ethtool_steering *ethtool = mlx5e_fs_get_ethtool(fs: priv->fs);
526 struct mlx5e_ethtool_rule *iter;
527
528 list_for_each_entry(iter, &ethtool->rules, list) {
529 if (iter->flow_spec.location == location)
530 return iter;
531 }
532 return NULL;
533}
534
535static struct mlx5e_ethtool_rule *get_ethtool_rule(struct mlx5e_priv *priv,
536 int location)
537{
538 struct mlx5e_ethtool_rule *eth_rule;
539
540 eth_rule = find_ethtool_rule(priv, location);
541 if (eth_rule)
542 del_ethtool_rule(fs: priv->fs, eth_rule);
543
544 eth_rule = kzalloc(size: sizeof(*eth_rule), GFP_KERNEL);
545 if (!eth_rule)
546 return ERR_PTR(error: -ENOMEM);
547
548 add_rule_to_list(priv, rule: eth_rule);
549 return eth_rule;
550}
551
552#define MAX_NUM_OF_ETHTOOL_RULES BIT(10)
553
554#define all_ones(field) (field == (__force typeof(field))-1)
555#define all_zeros_or_all_ones(field) \
556 ((field) == 0 || (field) == (__force typeof(field))-1)
557
558static int validate_ethter(struct ethtool_rx_flow_spec *fs)
559{
560 struct ethhdr *eth_mask = &fs->m_u.ether_spec;
561 int ntuples = 0;
562
563 if (!is_zero_ether_addr(addr: eth_mask->h_dest))
564 ntuples++;
565 if (!is_zero_ether_addr(addr: eth_mask->h_source))
566 ntuples++;
567 if (eth_mask->h_proto)
568 ntuples++;
569 return ntuples;
570}
571
572static int validate_tcpudp4(struct ethtool_rx_flow_spec *fs)
573{
574 struct ethtool_tcpip4_spec *l4_mask = &fs->m_u.tcp_ip4_spec;
575 int ntuples = 0;
576
577 if (l4_mask->tos)
578 return -EINVAL;
579
580 if (l4_mask->ip4src)
581 ntuples++;
582 if (l4_mask->ip4dst)
583 ntuples++;
584 if (l4_mask->psrc)
585 ntuples++;
586 if (l4_mask->pdst)
587 ntuples++;
588 /* Flow is TCP/UDP */
589 return ++ntuples;
590}
591
592static int validate_ip4(struct ethtool_rx_flow_spec *fs)
593{
594 struct ethtool_usrip4_spec *l3_mask = &fs->m_u.usr_ip4_spec;
595 int ntuples = 0;
596
597 if (l3_mask->l4_4_bytes || l3_mask->tos ||
598 fs->h_u.usr_ip4_spec.ip_ver != ETH_RX_NFC_IP4)
599 return -EINVAL;
600 if (l3_mask->ip4src)
601 ntuples++;
602 if (l3_mask->ip4dst)
603 ntuples++;
604 if (l3_mask->proto)
605 ntuples++;
606 /* Flow is IPv4 */
607 return ++ntuples;
608}
609
610static int validate_ip6(struct ethtool_rx_flow_spec *fs)
611{
612 struct ethtool_usrip6_spec *l3_mask = &fs->m_u.usr_ip6_spec;
613 int ntuples = 0;
614
615 if (l3_mask->l4_4_bytes || l3_mask->tclass)
616 return -EINVAL;
617 if (!ipv6_addr_any(a: (struct in6_addr *)l3_mask->ip6src))
618 ntuples++;
619
620 if (!ipv6_addr_any(a: (struct in6_addr *)l3_mask->ip6dst))
621 ntuples++;
622 if (l3_mask->l4_proto)
623 ntuples++;
624 /* Flow is IPv6 */
625 return ++ntuples;
626}
627
628static int validate_tcpudp6(struct ethtool_rx_flow_spec *fs)
629{
630 struct ethtool_tcpip6_spec *l4_mask = &fs->m_u.tcp_ip6_spec;
631 int ntuples = 0;
632
633 if (l4_mask->tclass)
634 return -EINVAL;
635
636 if (!ipv6_addr_any(a: (struct in6_addr *)l4_mask->ip6src))
637 ntuples++;
638
639 if (!ipv6_addr_any(a: (struct in6_addr *)l4_mask->ip6dst))
640 ntuples++;
641
642 if (l4_mask->psrc)
643 ntuples++;
644 if (l4_mask->pdst)
645 ntuples++;
646 /* Flow is TCP/UDP */
647 return ++ntuples;
648}
649
650static int validate_vlan(struct ethtool_rx_flow_spec *fs)
651{
652 if (fs->m_ext.vlan_etype ||
653 fs->m_ext.vlan_tci != cpu_to_be16(VLAN_VID_MASK))
654 return -EINVAL;
655
656 if (fs->m_ext.vlan_tci &&
657 (be16_to_cpu(fs->h_ext.vlan_tci) >= VLAN_N_VID))
658 return -EINVAL;
659
660 return 1;
661}
662
663static int validate_flow(struct mlx5e_priv *priv,
664 struct ethtool_rx_flow_spec *fs)
665{
666 int num_tuples = 0;
667 int ret = 0;
668
669 if (fs->location >= MAX_NUM_OF_ETHTOOL_RULES)
670 return -ENOSPC;
671
672 if (fs->ring_cookie != RX_CLS_FLOW_DISC)
673 if (fs->ring_cookie >= priv->channels.params.num_channels)
674 return -EINVAL;
675
676 switch (flow_type_mask(flow_type: fs->flow_type)) {
677 case ETHER_FLOW:
678 num_tuples += validate_ethter(fs);
679 break;
680 case TCP_V4_FLOW:
681 case UDP_V4_FLOW:
682 ret = validate_tcpudp4(fs);
683 if (ret < 0)
684 return ret;
685 num_tuples += ret;
686 break;
687 case IP_USER_FLOW:
688 ret = validate_ip4(fs);
689 if (ret < 0)
690 return ret;
691 num_tuples += ret;
692 break;
693 case TCP_V6_FLOW:
694 case UDP_V6_FLOW:
695 ret = validate_tcpudp6(fs);
696 if (ret < 0)
697 return ret;
698 num_tuples += ret;
699 break;
700 case IPV6_USER_FLOW:
701 ret = validate_ip6(fs);
702 if (ret < 0)
703 return ret;
704 num_tuples += ret;
705 break;
706 default:
707 return -ENOTSUPP;
708 }
709 if ((fs->flow_type & FLOW_EXT)) {
710 ret = validate_vlan(fs);
711 if (ret < 0)
712 return ret;
713 num_tuples += ret;
714 }
715
716 if (fs->flow_type & FLOW_MAC_EXT &&
717 !is_zero_ether_addr(addr: fs->m_ext.h_dest))
718 num_tuples++;
719
720 return num_tuples;
721}
722
723static int
724mlx5e_ethtool_flow_replace(struct mlx5e_priv *priv,
725 struct ethtool_rx_flow_spec *fs, u32 rss_context)
726{
727 struct mlx5e_ethtool_table *eth_ft;
728 struct mlx5e_ethtool_rule *eth_rule;
729 struct mlx5_flow_handle *rule;
730 int num_tuples;
731 int err;
732
733 num_tuples = validate_flow(priv, fs);
734 if (num_tuples <= 0) {
735 netdev_warn(dev: priv->netdev, format: "%s: flow is not valid %d\n",
736 __func__, num_tuples);
737 return num_tuples;
738 }
739
740 eth_ft = get_flow_table(priv, fs, num_tuples);
741 if (IS_ERR(ptr: eth_ft))
742 return PTR_ERR(ptr: eth_ft);
743
744 eth_rule = get_ethtool_rule(priv, location: fs->location);
745 if (IS_ERR(ptr: eth_rule)) {
746 put_flow_table(eth_ft);
747 return PTR_ERR(ptr: eth_rule);
748 }
749
750 eth_rule->flow_spec = *fs;
751 eth_rule->eth_ft = eth_ft;
752
753 rule = add_ethtool_flow_rule(priv, eth_rule, ft: eth_ft->ft, fs, rss_context);
754 if (IS_ERR(ptr: rule)) {
755 err = PTR_ERR(ptr: rule);
756 goto del_ethtool_rule;
757 }
758
759 eth_rule->rule = rule;
760
761 return 0;
762
763del_ethtool_rule:
764 del_ethtool_rule(fs: priv->fs, eth_rule);
765
766 return err;
767}
768
769static int
770mlx5e_ethtool_flow_remove(struct mlx5e_priv *priv, int location)
771{
772 struct mlx5e_ethtool_rule *eth_rule;
773 int err = 0;
774
775 if (location >= MAX_NUM_OF_ETHTOOL_RULES)
776 return -ENOSPC;
777
778 eth_rule = find_ethtool_rule(priv, location);
779 if (!eth_rule) {
780 err = -ENOENT;
781 goto out;
782 }
783
784 del_ethtool_rule(fs: priv->fs, eth_rule);
785out:
786 return err;
787}
788
789static int
790mlx5e_ethtool_get_flow(struct mlx5e_priv *priv,
791 struct ethtool_rxnfc *info, int location)
792{
793 struct mlx5e_ethtool_steering *ethtool = mlx5e_fs_get_ethtool(fs: priv->fs);
794 struct mlx5e_ethtool_rule *eth_rule;
795
796 if (location < 0 || location >= MAX_NUM_OF_ETHTOOL_RULES)
797 return -EINVAL;
798
799 list_for_each_entry(eth_rule, &ethtool->rules, list) {
800 int index;
801
802 if (eth_rule->flow_spec.location != location)
803 continue;
804 if (!info)
805 return 0;
806 info->fs = eth_rule->flow_spec;
807 if (!eth_rule->rss)
808 return 0;
809 index = mlx5e_rx_res_rss_index(res: priv->rx_res, rss: eth_rule->rss);
810 if (index < 0)
811 return index;
812 info->rss_context = index;
813 return 0;
814 }
815
816 return -ENOENT;
817}
818
819static int
820mlx5e_ethtool_get_all_flows(struct mlx5e_priv *priv,
821 struct ethtool_rxnfc *info, u32 *rule_locs)
822{
823 int location = 0;
824 int idx = 0;
825 int err = 0;
826
827 info->data = MAX_NUM_OF_ETHTOOL_RULES;
828 while ((!err || err == -ENOENT) && idx < info->rule_cnt) {
829 err = mlx5e_ethtool_get_flow(priv, NULL, location);
830 if (!err)
831 rule_locs[idx++] = location;
832 location++;
833 }
834 return err;
835}
836
837int mlx5e_ethtool_alloc(struct mlx5e_ethtool_steering **ethtool)
838{
839 *ethtool = kvzalloc(size: sizeof(**ethtool), GFP_KERNEL);
840 if (!*ethtool)
841 return -ENOMEM;
842 return 0;
843}
844
845void mlx5e_ethtool_free(struct mlx5e_ethtool_steering *ethtool)
846{
847 kvfree(addr: ethtool);
848}
849
850void mlx5e_ethtool_cleanup_steering(struct mlx5e_flow_steering *fs)
851{
852 struct mlx5e_ethtool_steering *ethtool = mlx5e_fs_get_ethtool(fs);
853 struct mlx5e_ethtool_rule *iter;
854 struct mlx5e_ethtool_rule *temp;
855
856 list_for_each_entry_safe(iter, temp, &ethtool->rules, list)
857 del_ethtool_rule(fs, eth_rule: iter);
858}
859
860void mlx5e_ethtool_init_steering(struct mlx5e_flow_steering *fs)
861{
862 struct mlx5e_ethtool_steering *ethtool = mlx5e_fs_get_ethtool(fs);
863
864 INIT_LIST_HEAD(list: &ethtool->rules);
865}
866
867static int flow_type_to_traffic_type(u32 flow_type)
868{
869 switch (flow_type) {
870 case TCP_V4_FLOW:
871 return MLX5_TT_IPV4_TCP;
872 case TCP_V6_FLOW:
873 return MLX5_TT_IPV6_TCP;
874 case UDP_V4_FLOW:
875 return MLX5_TT_IPV4_UDP;
876 case UDP_V6_FLOW:
877 return MLX5_TT_IPV6_UDP;
878 case AH_V4_FLOW:
879 return MLX5_TT_IPV4_IPSEC_AH;
880 case AH_V6_FLOW:
881 return MLX5_TT_IPV6_IPSEC_AH;
882 case ESP_V4_FLOW:
883 return MLX5_TT_IPV4_IPSEC_ESP;
884 case ESP_V6_FLOW:
885 return MLX5_TT_IPV6_IPSEC_ESP;
886 case IPV4_FLOW:
887 return MLX5_TT_IPV4;
888 case IPV6_FLOW:
889 return MLX5_TT_IPV6;
890 default:
891 return -EINVAL;
892 }
893}
894
895static int mlx5e_set_rss_hash_opt(struct mlx5e_priv *priv,
896 struct ethtool_rxnfc *nfc)
897{
898 u8 rx_hash_field = 0;
899 u32 flow_type = 0;
900 u32 rss_idx = 0;
901 int err;
902 int tt;
903
904 if (nfc->flow_type & FLOW_RSS)
905 rss_idx = nfc->rss_context;
906
907 flow_type = flow_type_mask(flow_type: nfc->flow_type);
908 tt = flow_type_to_traffic_type(flow_type);
909 if (tt < 0)
910 return tt;
911
912 /* RSS does not support anything other than hashing to queues
913 * on src IP, dest IP, TCP/UDP src port and TCP/UDP dest
914 * port.
915 */
916 if (flow_type != TCP_V4_FLOW &&
917 flow_type != TCP_V6_FLOW &&
918 flow_type != UDP_V4_FLOW &&
919 flow_type != UDP_V6_FLOW)
920 return -EOPNOTSUPP;
921
922 if (nfc->data & ~(RXH_IP_SRC | RXH_IP_DST |
923 RXH_L4_B_0_1 | RXH_L4_B_2_3))
924 return -EOPNOTSUPP;
925
926 if (nfc->data & RXH_IP_SRC)
927 rx_hash_field |= MLX5_HASH_FIELD_SEL_SRC_IP;
928 if (nfc->data & RXH_IP_DST)
929 rx_hash_field |= MLX5_HASH_FIELD_SEL_DST_IP;
930 if (nfc->data & RXH_L4_B_0_1)
931 rx_hash_field |= MLX5_HASH_FIELD_SEL_L4_SPORT;
932 if (nfc->data & RXH_L4_B_2_3)
933 rx_hash_field |= MLX5_HASH_FIELD_SEL_L4_DPORT;
934
935 mutex_lock(&priv->state_lock);
936 err = mlx5e_rx_res_rss_set_hash_fields(res: priv->rx_res, rss_idx, tt, rx_hash_fields: rx_hash_field);
937 mutex_unlock(lock: &priv->state_lock);
938
939 return err;
940}
941
942static int mlx5e_get_rss_hash_opt(struct mlx5e_priv *priv,
943 struct ethtool_rxnfc *nfc)
944{
945 int hash_field = 0;
946 u32 flow_type = 0;
947 u32 rss_idx = 0;
948 int tt;
949
950 if (nfc->flow_type & FLOW_RSS)
951 rss_idx = nfc->rss_context;
952
953 flow_type = flow_type_mask(flow_type: nfc->flow_type);
954 tt = flow_type_to_traffic_type(flow_type);
955 if (tt < 0)
956 return tt;
957
958 hash_field = mlx5e_rx_res_rss_get_hash_fields(res: priv->rx_res, rss_idx, tt);
959 if (hash_field < 0)
960 return hash_field;
961
962 nfc->data = 0;
963
964 if (hash_field & MLX5_HASH_FIELD_SEL_SRC_IP)
965 nfc->data |= RXH_IP_SRC;
966 if (hash_field & MLX5_HASH_FIELD_SEL_DST_IP)
967 nfc->data |= RXH_IP_DST;
968 if (hash_field & MLX5_HASH_FIELD_SEL_L4_SPORT)
969 nfc->data |= RXH_L4_B_0_1;
970 if (hash_field & MLX5_HASH_FIELD_SEL_L4_DPORT)
971 nfc->data |= RXH_L4_B_2_3;
972
973 return 0;
974}
975
976int mlx5e_ethtool_set_rxnfc(struct mlx5e_priv *priv, struct ethtool_rxnfc *cmd)
977{
978 int err = 0;
979
980 switch (cmd->cmd) {
981 case ETHTOOL_SRXCLSRLINS:
982 err = mlx5e_ethtool_flow_replace(priv, fs: &cmd->fs, rss_context: cmd->rss_context);
983 break;
984 case ETHTOOL_SRXCLSRLDEL:
985 err = mlx5e_ethtool_flow_remove(priv, location: cmd->fs.location);
986 break;
987 case ETHTOOL_SRXFH:
988 err = mlx5e_set_rss_hash_opt(priv, nfc: cmd);
989 break;
990 default:
991 err = -EOPNOTSUPP;
992 break;
993 }
994
995 return err;
996}
997
998int mlx5e_ethtool_get_rxnfc(struct mlx5e_priv *priv,
999 struct ethtool_rxnfc *info, u32 *rule_locs)
1000{
1001 struct mlx5e_ethtool_steering *ethtool = mlx5e_fs_get_ethtool(fs: priv->fs);
1002 int err = 0;
1003
1004 switch (info->cmd) {
1005 case ETHTOOL_GRXCLSRLCNT:
1006 info->rule_cnt = ethtool->tot_num_rules;
1007 break;
1008 case ETHTOOL_GRXCLSRULE:
1009 err = mlx5e_ethtool_get_flow(priv, info, location: info->fs.location);
1010 break;
1011 case ETHTOOL_GRXCLSRLALL:
1012 err = mlx5e_ethtool_get_all_flows(priv, info, rule_locs);
1013 break;
1014 case ETHTOOL_GRXFH:
1015 err = mlx5e_get_rss_hash_opt(priv, nfc: info);
1016 break;
1017 default:
1018 err = -EOPNOTSUPP;
1019 break;
1020 }
1021
1022 return err;
1023}
1024
1025

source code of linux/drivers/net/ethernet/mellanox/mlx5/core/en_fs_ethtool.c