1
2
3
4#include <linux/netdevice.h>
5#include "en_accel/fs_tcp.h"
6#include "fs_core.h"
7
8enum accel_fs_tcp_type {
9 ACCEL_FS_IPV4_TCP,
10 ACCEL_FS_IPV6_TCP,
11 ACCEL_FS_TCP_NUM_TYPES,
12};
13
14struct mlx5e_accel_fs_tcp {
15 struct mlx5e_flow_table tables[ACCEL_FS_TCP_NUM_TYPES];
16 struct mlx5_flow_handle *default_rules[ACCEL_FS_TCP_NUM_TYPES];
17};
18
19static enum mlx5_traffic_types fs_accel2tt(enum accel_fs_tcp_type i)
20{
21 switch (i) {
22 case ACCEL_FS_IPV4_TCP:
23 return MLX5_TT_IPV4_TCP;
24 default:
25 return MLX5_TT_IPV6_TCP;
26 }
27}
28
29static void accel_fs_tcp_set_ipv4_flow(struct mlx5_flow_spec *spec, struct sock *sk)
30{
31 MLX5_SET_TO_ONES(fte_match_param, spec->match_criteria, outer_headers.ip_protocol);
32 MLX5_SET(fte_match_param, spec->match_value, outer_headers.ip_protocol, IPPROTO_TCP);
33 MLX5_SET_TO_ONES(fte_match_param, spec->match_criteria, outer_headers.ip_version);
34 MLX5_SET(fte_match_param, spec->match_value, outer_headers.ip_version, 4);
35 memcpy(MLX5_ADDR_OF(fte_match_param, spec->match_value,
36 outer_headers.src_ipv4_src_ipv6.ipv4_layout.ipv4),
37 &inet_sk(sk)->inet_daddr, 4);
38 memcpy(MLX5_ADDR_OF(fte_match_param, spec->match_value,
39 outer_headers.dst_ipv4_dst_ipv6.ipv4_layout.ipv4),
40 &inet_sk(sk)->inet_rcv_saddr, 4);
41 MLX5_SET_TO_ONES(fte_match_param, spec->match_criteria,
42 outer_headers.src_ipv4_src_ipv6.ipv4_layout.ipv4);
43 MLX5_SET_TO_ONES(fte_match_param, spec->match_criteria,
44 outer_headers.dst_ipv4_dst_ipv6.ipv4_layout.ipv4);
45}
46
47#if IS_ENABLED(CONFIG_IPV6)
48static void accel_fs_tcp_set_ipv6_flow(struct mlx5_flow_spec *spec, struct sock *sk)
49{
50 MLX5_SET_TO_ONES(fte_match_param, spec->match_criteria, outer_headers.ip_protocol);
51 MLX5_SET(fte_match_param, spec->match_value, outer_headers.ip_protocol, IPPROTO_TCP);
52 MLX5_SET_TO_ONES(fte_match_param, spec->match_criteria, outer_headers.ip_version);
53 MLX5_SET(fte_match_param, spec->match_value, outer_headers.ip_version, 6);
54 memcpy(MLX5_ADDR_OF(fte_match_param, spec->match_value,
55 outer_headers.src_ipv4_src_ipv6.ipv6_layout.ipv6),
56 &sk->sk_v6_daddr, 16);
57 memcpy(MLX5_ADDR_OF(fte_match_param, spec->match_value,
58 outer_headers.dst_ipv4_dst_ipv6.ipv6_layout.ipv6),
59 &inet6_sk(sk)->saddr, 16);
60 memset(MLX5_ADDR_OF(fte_match_param, spec->match_criteria,
61 outer_headers.src_ipv4_src_ipv6.ipv6_layout.ipv6),
62 0xff, 16);
63 memset(MLX5_ADDR_OF(fte_match_param, spec->match_criteria,
64 outer_headers.dst_ipv4_dst_ipv6.ipv6_layout.ipv6),
65 0xff, 16);
66}
67#endif
68
69void mlx5e_accel_fs_del_sk(struct mlx5_flow_handle *rule)
70{
71 mlx5_del_flow_rules(rule);
72}
73
74struct mlx5_flow_handle *mlx5e_accel_fs_add_sk(struct mlx5e_priv *priv,
75 struct sock *sk, u32 tirn,
76 uint32_t flow_tag)
77{
78 struct mlx5_flow_destination dest = {};
79 struct mlx5e_flow_table *ft = NULL;
80 struct mlx5e_accel_fs_tcp *fs_tcp;
81 MLX5_DECLARE_FLOW_ACT(flow_act);
82 struct mlx5_flow_handle *flow;
83 struct mlx5_flow_spec *spec;
84
85 spec = kvzalloc(sizeof(*spec), GFP_KERNEL);
86 if (!spec)
87 return ERR_PTR(-ENOMEM);
88
89 fs_tcp = priv->fs.accel_tcp;
90
91 spec->match_criteria_enable = MLX5_MATCH_OUTER_HEADERS;
92
93 switch (sk->sk_family) {
94 case AF_INET:
95 accel_fs_tcp_set_ipv4_flow(spec, sk);
96 ft = &fs_tcp->tables[ACCEL_FS_IPV4_TCP];
97 mlx5e_dbg(HW, priv, "%s flow is %pI4:%d -> %pI4:%d\n", __func__,
98 &inet_sk(sk)->inet_rcv_saddr,
99 inet_sk(sk)->inet_sport,
100 &inet_sk(sk)->inet_daddr,
101 inet_sk(sk)->inet_dport);
102 break;
103#if IS_ENABLED(CONFIG_IPV6)
104 case AF_INET6:
105 if (!sk->sk_ipv6only &&
106 ipv6_addr_type(&sk->sk_v6_daddr) == IPV6_ADDR_MAPPED) {
107 accel_fs_tcp_set_ipv4_flow(spec, sk);
108 ft = &fs_tcp->tables[ACCEL_FS_IPV4_TCP];
109 } else {
110 accel_fs_tcp_set_ipv6_flow(spec, sk);
111 ft = &fs_tcp->tables[ACCEL_FS_IPV6_TCP];
112 }
113 break;
114#endif
115 default:
116 break;
117 }
118
119 if (!ft) {
120 flow = ERR_PTR(-EINVAL);
121 goto out;
122 }
123
124 MLX5_SET_TO_ONES(fte_match_param, spec->match_criteria,
125 outer_headers.tcp_dport);
126 MLX5_SET_TO_ONES(fte_match_param, spec->match_criteria,
127 outer_headers.tcp_sport);
128 MLX5_SET(fte_match_param, spec->match_value, outer_headers.tcp_dport,
129 ntohs(inet_sk(sk)->inet_sport));
130 MLX5_SET(fte_match_param, spec->match_value, outer_headers.tcp_sport,
131 ntohs(inet_sk(sk)->inet_dport));
132
133 dest.type = MLX5_FLOW_DESTINATION_TYPE_TIR;
134 dest.tir_num = tirn;
135 if (flow_tag != MLX5_FS_DEFAULT_FLOW_TAG) {
136 spec->flow_context.flow_tag = flow_tag;
137 spec->flow_context.flags = FLOW_CONTEXT_HAS_TAG;
138 }
139
140 flow = mlx5_add_flow_rules(ft->t, spec, &flow_act, &dest, 1);
141
142 if (IS_ERR(flow))
143 netdev_err(priv->netdev, "mlx5_add_flow_rules() failed, flow is %ld\n",
144 PTR_ERR(flow));
145
146out:
147 kvfree(spec);
148 return flow;
149}
150
151static int accel_fs_tcp_add_default_rule(struct mlx5e_priv *priv,
152 enum accel_fs_tcp_type type)
153{
154 struct mlx5e_flow_table *accel_fs_t;
155 struct mlx5_flow_destination dest;
156 struct mlx5e_accel_fs_tcp *fs_tcp;
157 MLX5_DECLARE_FLOW_ACT(flow_act);
158 struct mlx5_flow_handle *rule;
159 int err = 0;
160
161 fs_tcp = priv->fs.accel_tcp;
162 accel_fs_t = &fs_tcp->tables[type];
163
164 dest = mlx5_ttc_get_default_dest(priv->fs.ttc, fs_accel2tt(type));
165 rule = mlx5_add_flow_rules(accel_fs_t->t, NULL, &flow_act, &dest, 1);
166 if (IS_ERR(rule)) {
167 err = PTR_ERR(rule);
168 netdev_err(priv->netdev,
169 "%s: add default rule failed, accel_fs type=%d, err %d\n",
170 __func__, type, err);
171 return err;
172 }
173
174 fs_tcp->default_rules[type] = rule;
175 return 0;
176}
177
178#define MLX5E_ACCEL_FS_TCP_NUM_GROUPS (2)
179#define MLX5E_ACCEL_FS_TCP_GROUP1_SIZE (BIT(16) - 1)
180#define MLX5E_ACCEL_FS_TCP_GROUP2_SIZE (BIT(0))
181#define MLX5E_ACCEL_FS_TCP_TABLE_SIZE (MLX5E_ACCEL_FS_TCP_GROUP1_SIZE +\
182 MLX5E_ACCEL_FS_TCP_GROUP2_SIZE)
183static int accel_fs_tcp_create_groups(struct mlx5e_flow_table *ft,
184 enum accel_fs_tcp_type type)
185{
186 int inlen = MLX5_ST_SZ_BYTES(create_flow_group_in);
187 void *outer_headers_c;
188 int ix = 0;
189 u32 *in;
190 int err;
191 u8 *mc;
192
193 ft->g = kcalloc(MLX5E_ACCEL_FS_TCP_NUM_GROUPS, sizeof(*ft->g), GFP_KERNEL);
194 in = kvzalloc(inlen, GFP_KERNEL);
195 if (!in || !ft->g) {
196 kfree(ft->g);
197 kvfree(in);
198 return -ENOMEM;
199 }
200
201 mc = MLX5_ADDR_OF(create_flow_group_in, in, match_criteria);
202 outer_headers_c = MLX5_ADDR_OF(fte_match_param, mc, outer_headers);
203 MLX5_SET_TO_ONES(fte_match_set_lyr_2_4, outer_headers_c, ip_protocol);
204 MLX5_SET_TO_ONES(fte_match_set_lyr_2_4, outer_headers_c, ip_version);
205
206 switch (type) {
207 case ACCEL_FS_IPV4_TCP:
208 case ACCEL_FS_IPV6_TCP:
209 MLX5_SET_TO_ONES(fte_match_set_lyr_2_4, outer_headers_c, tcp_dport);
210 MLX5_SET_TO_ONES(fte_match_set_lyr_2_4, outer_headers_c, tcp_sport);
211 break;
212 default:
213 err = -EINVAL;
214 goto out;
215 }
216
217 switch (type) {
218 case ACCEL_FS_IPV4_TCP:
219 MLX5_SET_TO_ONES(fte_match_set_lyr_2_4, outer_headers_c,
220 src_ipv4_src_ipv6.ipv4_layout.ipv4);
221 MLX5_SET_TO_ONES(fte_match_set_lyr_2_4, outer_headers_c,
222 dst_ipv4_dst_ipv6.ipv4_layout.ipv4);
223 break;
224 case ACCEL_FS_IPV6_TCP:
225 memset(MLX5_ADDR_OF(fte_match_set_lyr_2_4, outer_headers_c,
226 src_ipv4_src_ipv6.ipv6_layout.ipv6),
227 0xff, 16);
228 memset(MLX5_ADDR_OF(fte_match_set_lyr_2_4, outer_headers_c,
229 dst_ipv4_dst_ipv6.ipv6_layout.ipv6),
230 0xff, 16);
231 break;
232 default:
233 err = -EINVAL;
234 goto out;
235 }
236
237 MLX5_SET_CFG(in, match_criteria_enable, MLX5_MATCH_OUTER_HEADERS);
238 MLX5_SET_CFG(in, start_flow_index, ix);
239 ix += MLX5E_ACCEL_FS_TCP_GROUP1_SIZE;
240 MLX5_SET_CFG(in, end_flow_index, ix - 1);
241 ft->g[ft->num_groups] = mlx5_create_flow_group(ft->t, in);
242 if (IS_ERR(ft->g[ft->num_groups]))
243 goto err;
244 ft->num_groups++;
245
246
247 memset(in, 0, inlen);
248 MLX5_SET_CFG(in, start_flow_index, ix);
249 ix += MLX5E_ACCEL_FS_TCP_GROUP2_SIZE;
250 MLX5_SET_CFG(in, end_flow_index, ix - 1);
251 ft->g[ft->num_groups] = mlx5_create_flow_group(ft->t, in);
252 if (IS_ERR(ft->g[ft->num_groups]))
253 goto err;
254 ft->num_groups++;
255
256 kvfree(in);
257 return 0;
258
259err:
260 err = PTR_ERR(ft->g[ft->num_groups]);
261 ft->g[ft->num_groups] = NULL;
262out:
263 kvfree(in);
264
265 return err;
266}
267
268static int accel_fs_tcp_create_table(struct mlx5e_priv *priv, enum accel_fs_tcp_type type)
269{
270 struct mlx5e_flow_table *ft = &priv->fs.accel_tcp->tables[type];
271 struct mlx5_flow_table_attr ft_attr = {};
272 int err;
273
274 ft->num_groups = 0;
275
276 ft_attr.max_fte = MLX5E_ACCEL_FS_TCP_TABLE_SIZE;
277 ft_attr.level = MLX5E_ACCEL_FS_TCP_FT_LEVEL;
278 ft_attr.prio = MLX5E_NIC_PRIO;
279
280 ft->t = mlx5_create_flow_table(priv->fs.ns, &ft_attr);
281 if (IS_ERR(ft->t)) {
282 err = PTR_ERR(ft->t);
283 ft->t = NULL;
284 return err;
285 }
286
287 netdev_dbg(priv->netdev, "Created fs accel table id %u level %u\n",
288 ft->t->id, ft->t->level);
289
290 err = accel_fs_tcp_create_groups(ft, type);
291 if (err)
292 goto err;
293
294 err = accel_fs_tcp_add_default_rule(priv, type);
295 if (err)
296 goto err;
297
298 return 0;
299err:
300 mlx5e_destroy_flow_table(ft);
301 return err;
302}
303
304static int accel_fs_tcp_disable(struct mlx5e_priv *priv)
305{
306 int err, i;
307
308 for (i = 0; i < ACCEL_FS_TCP_NUM_TYPES; i++) {
309
310 err = mlx5_ttc_fwd_default_dest(priv->fs.ttc, fs_accel2tt(i));
311 if (err) {
312 netdev_err(priv->netdev,
313 "%s: modify ttc[%d] default destination failed, err(%d)\n",
314 __func__, fs_accel2tt(i), err);
315 return err;
316 }
317 }
318
319 return 0;
320}
321
322static int accel_fs_tcp_enable(struct mlx5e_priv *priv)
323{
324 struct mlx5_flow_destination dest = {};
325 int err, i;
326
327 dest.type = MLX5_FLOW_DESTINATION_TYPE_FLOW_TABLE;
328 for (i = 0; i < ACCEL_FS_TCP_NUM_TYPES; i++) {
329 dest.ft = priv->fs.accel_tcp->tables[i].t;
330
331
332 err = mlx5_ttc_fwd_dest(priv->fs.ttc, fs_accel2tt(i), &dest);
333 if (err) {
334 netdev_err(priv->netdev,
335 "%s: modify ttc[%d] destination to accel failed, err(%d)\n",
336 __func__, fs_accel2tt(i), err);
337 return err;
338 }
339 }
340 return 0;
341}
342
343static void accel_fs_tcp_destroy_table(struct mlx5e_priv *priv, int i)
344{
345 struct mlx5e_accel_fs_tcp *fs_tcp;
346
347 fs_tcp = priv->fs.accel_tcp;
348 if (IS_ERR_OR_NULL(fs_tcp->tables[i].t))
349 return;
350
351 mlx5_del_flow_rules(fs_tcp->default_rules[i]);
352 mlx5e_destroy_flow_table(&fs_tcp->tables[i]);
353 fs_tcp->tables[i].t = NULL;
354}
355
356void mlx5e_accel_fs_tcp_destroy(struct mlx5e_priv *priv)
357{
358 int i;
359
360 if (!priv->fs.accel_tcp)
361 return;
362
363 accel_fs_tcp_disable(priv);
364
365 for (i = 0; i < ACCEL_FS_TCP_NUM_TYPES; i++)
366 accel_fs_tcp_destroy_table(priv, i);
367
368 kfree(priv->fs.accel_tcp);
369 priv->fs.accel_tcp = NULL;
370}
371
372int mlx5e_accel_fs_tcp_create(struct mlx5e_priv *priv)
373{
374 int i, err;
375
376 if (!MLX5_CAP_FLOWTABLE_NIC_RX(priv->mdev, ft_field_support.outer_ip_version))
377 return -EOPNOTSUPP;
378
379 priv->fs.accel_tcp = kzalloc(sizeof(*priv->fs.accel_tcp), GFP_KERNEL);
380 if (!priv->fs.accel_tcp)
381 return -ENOMEM;
382
383 for (i = 0; i < ACCEL_FS_TCP_NUM_TYPES; i++) {
384 err = accel_fs_tcp_create_table(priv, i);
385 if (err)
386 goto err_destroy_tables;
387 }
388
389 err = accel_fs_tcp_enable(priv);
390 if (err)
391 goto err_destroy_tables;
392
393 return 0;
394
395err_destroy_tables:
396 while (--i >= 0)
397 accel_fs_tcp_destroy_table(priv, i);
398
399 kfree(priv->fs.accel_tcp);
400 priv->fs.accel_tcp = NULL;
401 return err;
402}
403