1// SPDX-License-Identifier: (GPL-2.0-only OR BSD-2-Clause)
2/* Copyright (C) 2017-2018 Netronome Systems, Inc. */
3
4#include <linux/bpf.h>
5#include <linux/bitops.h>
6#include <linux/bug.h>
7#include <linux/jiffies.h>
8#include <linux/skbuff.h>
9#include <linux/timekeeping.h>
10
11#include "../ccm.h"
12#include "../nfp_app.h"
13#include "../nfp_net.h"
14#include "fw.h"
15#include "main.h"
16
17static struct sk_buff *
18nfp_bpf_cmsg_alloc(struct nfp_app_bpf *bpf, unsigned int size)
19{
20 struct sk_buff *skb;
21
22 skb = nfp_app_ctrl_msg_alloc(app: bpf->app, size, GFP_KERNEL);
23 skb_put(skb, len: size);
24
25 return skb;
26}
27
28static unsigned int
29nfp_bpf_cmsg_map_req_size(struct nfp_app_bpf *bpf, unsigned int n)
30{
31 unsigned int size;
32
33 size = sizeof(struct cmsg_req_map_op);
34 size += (bpf->cmsg_key_sz + bpf->cmsg_val_sz) * n;
35
36 return size;
37}
38
39static struct sk_buff *
40nfp_bpf_cmsg_map_req_alloc(struct nfp_app_bpf *bpf, unsigned int n)
41{
42 return nfp_bpf_cmsg_alloc(bpf, size: nfp_bpf_cmsg_map_req_size(bpf, n));
43}
44
45static unsigned int
46nfp_bpf_cmsg_map_reply_size(struct nfp_app_bpf *bpf, unsigned int n)
47{
48 unsigned int size;
49
50 size = sizeof(struct cmsg_reply_map_op);
51 size += (bpf->cmsg_key_sz + bpf->cmsg_val_sz) * n;
52
53 return size;
54}
55
56static int
57nfp_bpf_ctrl_rc_to_errno(struct nfp_app_bpf *bpf,
58 struct cmsg_reply_map_simple *reply)
59{
60 static const int res_table[] = {
61 [CMSG_RC_SUCCESS] = 0,
62 [CMSG_RC_ERR_MAP_FD] = -EBADFD,
63 [CMSG_RC_ERR_MAP_NOENT] = -ENOENT,
64 [CMSG_RC_ERR_MAP_ERR] = -EINVAL,
65 [CMSG_RC_ERR_MAP_PARSE] = -EIO,
66 [CMSG_RC_ERR_MAP_EXIST] = -EEXIST,
67 [CMSG_RC_ERR_MAP_NOMEM] = -ENOMEM,
68 [CMSG_RC_ERR_MAP_E2BIG] = -E2BIG,
69 };
70 u32 rc;
71
72 rc = be32_to_cpu(reply->rc);
73 if (rc >= ARRAY_SIZE(res_table)) {
74 cmsg_warn(bpf, "FW responded with invalid status: %u\n", rc);
75 return -EIO;
76 }
77
78 return res_table[rc];
79}
80
81long long int
82nfp_bpf_ctrl_alloc_map(struct nfp_app_bpf *bpf, struct bpf_map *map)
83{
84 struct cmsg_reply_map_alloc_tbl *reply;
85 struct cmsg_req_map_alloc_tbl *req;
86 struct sk_buff *skb;
87 u32 tid;
88 int err;
89
90 skb = nfp_bpf_cmsg_alloc(bpf, size: sizeof(*req));
91 if (!skb)
92 return -ENOMEM;
93
94 req = (void *)skb->data;
95 req->key_size = cpu_to_be32(map->key_size);
96 req->value_size = cpu_to_be32(map->value_size);
97 req->max_entries = cpu_to_be32(map->max_entries);
98 req->map_type = cpu_to_be32(map->map_type);
99 req->map_flags = 0;
100
101 skb = nfp_ccm_communicate(ccm: &bpf->ccm, skb, type: NFP_CCM_TYPE_BPF_MAP_ALLOC,
102 reply_size: sizeof(*reply));
103 if (IS_ERR(ptr: skb))
104 return PTR_ERR(ptr: skb);
105
106 reply = (void *)skb->data;
107 err = nfp_bpf_ctrl_rc_to_errno(bpf, reply: &reply->reply_hdr);
108 if (err)
109 goto err_free;
110
111 tid = be32_to_cpu(reply->tid);
112 dev_consume_skb_any(skb);
113
114 return tid;
115err_free:
116 dev_kfree_skb_any(skb);
117 return err;
118}
119
120void nfp_bpf_ctrl_free_map(struct nfp_app_bpf *bpf, struct nfp_bpf_map *nfp_map)
121{
122 struct cmsg_reply_map_free_tbl *reply;
123 struct cmsg_req_map_free_tbl *req;
124 struct sk_buff *skb;
125 int err;
126
127 skb = nfp_bpf_cmsg_alloc(bpf, size: sizeof(*req));
128 if (!skb) {
129 cmsg_warn(bpf, "leaking map - failed to allocate msg\n");
130 return;
131 }
132
133 req = (void *)skb->data;
134 req->tid = cpu_to_be32(nfp_map->tid);
135
136 skb = nfp_ccm_communicate(ccm: &bpf->ccm, skb, type: NFP_CCM_TYPE_BPF_MAP_FREE,
137 reply_size: sizeof(*reply));
138 if (IS_ERR(ptr: skb)) {
139 cmsg_warn(bpf, "leaking map - I/O error\n");
140 return;
141 }
142
143 reply = (void *)skb->data;
144 err = nfp_bpf_ctrl_rc_to_errno(bpf, reply: &reply->reply_hdr);
145 if (err)
146 cmsg_warn(bpf, "leaking map - FW responded with: %d\n", err);
147
148 dev_consume_skb_any(skb);
149}
150
151static void *
152nfp_bpf_ctrl_req_key(struct nfp_app_bpf *bpf, struct cmsg_req_map_op *req,
153 unsigned int n)
154{
155 return &req->data[bpf->cmsg_key_sz * n + bpf->cmsg_val_sz * n];
156}
157
158static void *
159nfp_bpf_ctrl_req_val(struct nfp_app_bpf *bpf, struct cmsg_req_map_op *req,
160 unsigned int n)
161{
162 return &req->data[bpf->cmsg_key_sz * (n + 1) + bpf->cmsg_val_sz * n];
163}
164
165static void *
166nfp_bpf_ctrl_reply_key(struct nfp_app_bpf *bpf, struct cmsg_reply_map_op *reply,
167 unsigned int n)
168{
169 return &reply->data[bpf->cmsg_key_sz * n + bpf->cmsg_val_sz * n];
170}
171
172static void *
173nfp_bpf_ctrl_reply_val(struct nfp_app_bpf *bpf, struct cmsg_reply_map_op *reply,
174 unsigned int n)
175{
176 return &reply->data[bpf->cmsg_key_sz * (n + 1) + bpf->cmsg_val_sz * n];
177}
178
179static bool nfp_bpf_ctrl_op_cache_invalidate(enum nfp_ccm_type op)
180{
181 return op == NFP_CCM_TYPE_BPF_MAP_UPDATE ||
182 op == NFP_CCM_TYPE_BPF_MAP_DELETE;
183}
184
185static bool nfp_bpf_ctrl_op_cache_capable(enum nfp_ccm_type op)
186{
187 return op == NFP_CCM_TYPE_BPF_MAP_LOOKUP ||
188 op == NFP_CCM_TYPE_BPF_MAP_GETNEXT;
189}
190
191static bool nfp_bpf_ctrl_op_cache_fill(enum nfp_ccm_type op)
192{
193 return op == NFP_CCM_TYPE_BPF_MAP_GETFIRST ||
194 op == NFP_CCM_TYPE_BPF_MAP_GETNEXT;
195}
196
197static unsigned int
198nfp_bpf_ctrl_op_cache_get(struct nfp_bpf_map *nfp_map, enum nfp_ccm_type op,
199 const u8 *key, u8 *out_key, u8 *out_value,
200 u32 *cache_gen)
201{
202 struct bpf_map *map = &nfp_map->offmap->map;
203 struct nfp_app_bpf *bpf = nfp_map->bpf;
204 unsigned int i, count, n_entries;
205 struct cmsg_reply_map_op *reply;
206
207 n_entries = nfp_bpf_ctrl_op_cache_fill(op) ? bpf->cmsg_cache_cnt : 1;
208
209 spin_lock(lock: &nfp_map->cache_lock);
210 *cache_gen = nfp_map->cache_gen;
211 if (nfp_map->cache_blockers)
212 n_entries = 1;
213
214 if (nfp_bpf_ctrl_op_cache_invalidate(op))
215 goto exit_block;
216 if (!nfp_bpf_ctrl_op_cache_capable(op))
217 goto exit_unlock;
218
219 if (!nfp_map->cache)
220 goto exit_unlock;
221 if (nfp_map->cache_to < ktime_get_ns())
222 goto exit_invalidate;
223
224 reply = (void *)nfp_map->cache->data;
225 count = be32_to_cpu(reply->count);
226
227 for (i = 0; i < count; i++) {
228 void *cached_key;
229
230 cached_key = nfp_bpf_ctrl_reply_key(bpf, reply, n: i);
231 if (memcmp(p: cached_key, q: key, size: map->key_size))
232 continue;
233
234 if (op == NFP_CCM_TYPE_BPF_MAP_LOOKUP)
235 memcpy(out_value, nfp_bpf_ctrl_reply_val(bpf, reply, i),
236 map->value_size);
237 if (op == NFP_CCM_TYPE_BPF_MAP_GETNEXT) {
238 if (i + 1 == count)
239 break;
240
241 memcpy(out_key,
242 nfp_bpf_ctrl_reply_key(bpf, reply, i + 1),
243 map->key_size);
244 }
245
246 n_entries = 0;
247 goto exit_unlock;
248 }
249 goto exit_unlock;
250
251exit_block:
252 nfp_map->cache_blockers++;
253exit_invalidate:
254 dev_consume_skb_any(skb: nfp_map->cache);
255 nfp_map->cache = NULL;
256exit_unlock:
257 spin_unlock(lock: &nfp_map->cache_lock);
258 return n_entries;
259}
260
261static void
262nfp_bpf_ctrl_op_cache_put(struct nfp_bpf_map *nfp_map, enum nfp_ccm_type op,
263 struct sk_buff *skb, u32 cache_gen)
264{
265 bool blocker, filler;
266
267 blocker = nfp_bpf_ctrl_op_cache_invalidate(op);
268 filler = nfp_bpf_ctrl_op_cache_fill(op);
269 if (blocker || filler) {
270 u64 to = 0;
271
272 if (filler)
273 to = ktime_get_ns() + NFP_BPF_MAP_CACHE_TIME_NS;
274
275 spin_lock(lock: &nfp_map->cache_lock);
276 if (blocker) {
277 nfp_map->cache_blockers--;
278 nfp_map->cache_gen++;
279 }
280 if (filler && !nfp_map->cache_blockers &&
281 nfp_map->cache_gen == cache_gen) {
282 nfp_map->cache_to = to;
283 swap(nfp_map->cache, skb);
284 }
285 spin_unlock(lock: &nfp_map->cache_lock);
286 }
287
288 dev_consume_skb_any(skb);
289}
290
291static int
292nfp_bpf_ctrl_entry_op(struct bpf_offloaded_map *offmap, enum nfp_ccm_type op,
293 u8 *key, u8 *value, u64 flags, u8 *out_key, u8 *out_value)
294{
295 struct nfp_bpf_map *nfp_map = offmap->dev_priv;
296 unsigned int n_entries, reply_entries, count;
297 struct nfp_app_bpf *bpf = nfp_map->bpf;
298 struct bpf_map *map = &offmap->map;
299 struct cmsg_reply_map_op *reply;
300 struct cmsg_req_map_op *req;
301 struct sk_buff *skb;
302 u32 cache_gen;
303 int err;
304
305 /* FW messages have no space for more than 32 bits of flags */
306 if (flags >> 32)
307 return -EOPNOTSUPP;
308
309 /* Handle op cache */
310 n_entries = nfp_bpf_ctrl_op_cache_get(nfp_map, op, key, out_key,
311 out_value, cache_gen: &cache_gen);
312 if (!n_entries)
313 return 0;
314
315 skb = nfp_bpf_cmsg_map_req_alloc(bpf, n: 1);
316 if (!skb) {
317 err = -ENOMEM;
318 goto err_cache_put;
319 }
320
321 req = (void *)skb->data;
322 req->tid = cpu_to_be32(nfp_map->tid);
323 req->count = cpu_to_be32(n_entries);
324 req->flags = cpu_to_be32(flags);
325
326 /* Copy inputs */
327 if (key)
328 memcpy(nfp_bpf_ctrl_req_key(bpf, req, 0), key, map->key_size);
329 if (value)
330 memcpy(nfp_bpf_ctrl_req_val(bpf, req, 0), value,
331 map->value_size);
332
333 skb = nfp_ccm_communicate(ccm: &bpf->ccm, skb, type: op, reply_size: 0);
334 if (IS_ERR(ptr: skb)) {
335 err = PTR_ERR(ptr: skb);
336 goto err_cache_put;
337 }
338
339 if (skb->len < sizeof(*reply)) {
340 cmsg_warn(bpf, "cmsg drop - type 0x%02x too short %d!\n",
341 op, skb->len);
342 err = -EIO;
343 goto err_free;
344 }
345
346 reply = (void *)skb->data;
347 count = be32_to_cpu(reply->count);
348 err = nfp_bpf_ctrl_rc_to_errno(bpf, reply: &reply->reply_hdr);
349 /* FW responds with message sized to hold the good entries,
350 * plus one extra entry if there was an error.
351 */
352 reply_entries = count + !!err;
353 if (n_entries > 1 && count)
354 err = 0;
355 if (err)
356 goto err_free;
357
358 if (skb->len != nfp_bpf_cmsg_map_reply_size(bpf, n: reply_entries)) {
359 cmsg_warn(bpf, "cmsg drop - type 0x%02x too short %d for %d entries!\n",
360 op, skb->len, reply_entries);
361 err = -EIO;
362 goto err_free;
363 }
364
365 /* Copy outputs */
366 if (out_key)
367 memcpy(out_key, nfp_bpf_ctrl_reply_key(bpf, reply, 0),
368 map->key_size);
369 if (out_value)
370 memcpy(out_value, nfp_bpf_ctrl_reply_val(bpf, reply, 0),
371 map->value_size);
372
373 nfp_bpf_ctrl_op_cache_put(nfp_map, op, skb, cache_gen);
374
375 return 0;
376err_free:
377 dev_kfree_skb_any(skb);
378err_cache_put:
379 nfp_bpf_ctrl_op_cache_put(nfp_map, op, NULL, cache_gen);
380 return err;
381}
382
383int nfp_bpf_ctrl_update_entry(struct bpf_offloaded_map *offmap,
384 void *key, void *value, u64 flags)
385{
386 return nfp_bpf_ctrl_entry_op(offmap, op: NFP_CCM_TYPE_BPF_MAP_UPDATE,
387 key, value, flags, NULL, NULL);
388}
389
390int nfp_bpf_ctrl_del_entry(struct bpf_offloaded_map *offmap, void *key)
391{
392 return nfp_bpf_ctrl_entry_op(offmap, op: NFP_CCM_TYPE_BPF_MAP_DELETE,
393 key, NULL, flags: 0, NULL, NULL);
394}
395
396int nfp_bpf_ctrl_lookup_entry(struct bpf_offloaded_map *offmap,
397 void *key, void *value)
398{
399 return nfp_bpf_ctrl_entry_op(offmap, op: NFP_CCM_TYPE_BPF_MAP_LOOKUP,
400 key, NULL, flags: 0, NULL, out_value: value);
401}
402
403int nfp_bpf_ctrl_getfirst_entry(struct bpf_offloaded_map *offmap,
404 void *next_key)
405{
406 return nfp_bpf_ctrl_entry_op(offmap, op: NFP_CCM_TYPE_BPF_MAP_GETFIRST,
407 NULL, NULL, flags: 0, out_key: next_key, NULL);
408}
409
410int nfp_bpf_ctrl_getnext_entry(struct bpf_offloaded_map *offmap,
411 void *key, void *next_key)
412{
413 return nfp_bpf_ctrl_entry_op(offmap, op: NFP_CCM_TYPE_BPF_MAP_GETNEXT,
414 key, NULL, flags: 0, out_key: next_key, NULL);
415}
416
417unsigned int nfp_bpf_ctrl_cmsg_min_mtu(struct nfp_app_bpf *bpf)
418{
419 return max(nfp_bpf_cmsg_map_req_size(bpf, 1),
420 nfp_bpf_cmsg_map_reply_size(bpf, 1));
421}
422
423unsigned int nfp_bpf_ctrl_cmsg_mtu(struct nfp_app_bpf *bpf)
424{
425 return max3(NFP_NET_DEFAULT_MTU,
426 nfp_bpf_cmsg_map_req_size(bpf, NFP_BPF_MAP_CACHE_CNT),
427 nfp_bpf_cmsg_map_reply_size(bpf, NFP_BPF_MAP_CACHE_CNT));
428}
429
430unsigned int nfp_bpf_ctrl_cmsg_cache_cnt(struct nfp_app_bpf *bpf)
431{
432 unsigned int mtu, req_max, reply_max, entry_sz;
433
434 mtu = bpf->app->ctrl->dp.mtu;
435 entry_sz = bpf->cmsg_key_sz + bpf->cmsg_val_sz;
436 req_max = (mtu - sizeof(struct cmsg_req_map_op)) / entry_sz;
437 reply_max = (mtu - sizeof(struct cmsg_reply_map_op)) / entry_sz;
438
439 return min3(req_max, reply_max, NFP_BPF_MAP_CACHE_CNT);
440}
441
442void nfp_bpf_ctrl_msg_rx(struct nfp_app *app, struct sk_buff *skb)
443{
444 struct nfp_app_bpf *bpf = app->priv;
445
446 if (unlikely(skb->len < sizeof(struct cmsg_reply_map_simple))) {
447 cmsg_warn(bpf, "cmsg drop - too short %d!\n", skb->len);
448 dev_kfree_skb_any(skb);
449 return;
450 }
451
452 if (nfp_ccm_get_type(skb) == NFP_CCM_TYPE_BPF_BPF_EVENT) {
453 if (!nfp_bpf_event_output(bpf, data: skb->data, len: skb->len))
454 dev_consume_skb_any(skb);
455 else
456 dev_kfree_skb_any(skb);
457 return;
458 }
459
460 nfp_ccm_rx(ccm: &bpf->ccm, skb);
461}
462
463void
464nfp_bpf_ctrl_msg_rx_raw(struct nfp_app *app, const void *data, unsigned int len)
465{
466 const struct nfp_ccm_hdr *hdr = data;
467 struct nfp_app_bpf *bpf = app->priv;
468
469 if (unlikely(len < sizeof(struct cmsg_reply_map_simple))) {
470 cmsg_warn(bpf, "cmsg drop - too short %d!\n", len);
471 return;
472 }
473
474 if (hdr->type == NFP_CCM_TYPE_BPF_BPF_EVENT)
475 nfp_bpf_event_output(bpf, data, len);
476 else
477 cmsg_warn(bpf, "cmsg drop - msg type %d with raw buffer!\n",
478 hdr->type);
479}
480

source code of linux/drivers/net/ethernet/netronome/nfp/bpf/cmsg.c