1 | /* |
2 | * Copyright (c) 2016 Citrix Systems Inc. |
3 | * |
4 | * This program is free software; you can redistribute it and/or |
5 | * modify it under the terms of the GNU General Public License version 2 |
6 | * as published by the Free Softare Foundation; or, when distributed |
7 | * separately from the Linux kernel or incorporated into other |
8 | * software packages, subject to the following license: |
9 | * |
10 | * Permission is hereby granted, free of charge, to any person obtaining a copy |
11 | * of this source file (the "Software"), to deal in the Software without |
12 | * restriction, including without limitation the rights to use, copy, modify, |
13 | * merge, publish, distribute, sublicense, and/or sell copies of the Software, |
14 | * and to permit persons to whom the Software is furnished to do so, subject to |
15 | * the following conditions: |
16 | * |
17 | * The above copyright notice and this permission notice shall be included in |
18 | * all copies or substantial portions of the Software. |
19 | * |
20 | * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR |
21 | * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, |
22 | * FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE |
23 | * AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER |
24 | * LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING |
25 | * FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS |
26 | * IN THE SOFTWARE. |
27 | */ |
28 | |
29 | #define XEN_NETIF_DEFINE_TOEPLITZ |
30 | |
31 | #include "common.h" |
32 | #include <linux/vmalloc.h> |
33 | #include <linux/rculist.h> |
34 | |
35 | static void xenvif_add_hash(struct xenvif *vif, const u8 *tag, |
36 | unsigned int len, u32 val) |
37 | { |
38 | struct xenvif_hash_cache_entry *new, *entry, *oldest; |
39 | unsigned long flags; |
40 | bool found; |
41 | |
42 | new = kmalloc(size: sizeof(*entry), GFP_ATOMIC); |
43 | if (!new) |
44 | return; |
45 | |
46 | memcpy(new->tag, tag, len); |
47 | new->len = len; |
48 | new->val = val; |
49 | |
50 | spin_lock_irqsave(&vif->hash.cache.lock, flags); |
51 | |
52 | found = false; |
53 | oldest = NULL; |
54 | list_for_each_entry_rcu(entry, &vif->hash.cache.list, link, |
55 | lockdep_is_held(&vif->hash.cache.lock)) { |
56 | /* Make sure we don't add duplicate entries */ |
57 | if (entry->len == len && |
58 | memcmp(p: entry->tag, q: tag, size: len) == 0) |
59 | found = true; |
60 | if (!oldest || entry->seq < oldest->seq) |
61 | oldest = entry; |
62 | } |
63 | |
64 | if (!found) { |
65 | new->seq = atomic_inc_return(v: &vif->hash.cache.seq); |
66 | list_add_rcu(new: &new->link, head: &vif->hash.cache.list); |
67 | |
68 | if (++vif->hash.cache.count > xenvif_hash_cache_size) { |
69 | list_del_rcu(entry: &oldest->link); |
70 | vif->hash.cache.count--; |
71 | kfree_rcu(oldest, rcu); |
72 | } |
73 | } |
74 | |
75 | spin_unlock_irqrestore(lock: &vif->hash.cache.lock, flags); |
76 | |
77 | if (found) |
78 | kfree(objp: new); |
79 | } |
80 | |
81 | static u32 xenvif_new_hash(struct xenvif *vif, const u8 *data, |
82 | unsigned int len) |
83 | { |
84 | u32 val; |
85 | |
86 | val = xen_netif_toeplitz_hash(key: vif->hash.key, |
87 | keylen: sizeof(vif->hash.key), |
88 | buf: data, buflen: len); |
89 | |
90 | if (xenvif_hash_cache_size != 0) |
91 | xenvif_add_hash(vif, tag: data, len, val); |
92 | |
93 | return val; |
94 | } |
95 | |
96 | static void xenvif_flush_hash(struct xenvif *vif) |
97 | { |
98 | struct xenvif_hash_cache_entry *entry; |
99 | unsigned long flags; |
100 | |
101 | if (xenvif_hash_cache_size == 0) |
102 | return; |
103 | |
104 | spin_lock_irqsave(&vif->hash.cache.lock, flags); |
105 | |
106 | list_for_each_entry_rcu(entry, &vif->hash.cache.list, link, |
107 | lockdep_is_held(&vif->hash.cache.lock)) { |
108 | list_del_rcu(entry: &entry->link); |
109 | vif->hash.cache.count--; |
110 | kfree_rcu(entry, rcu); |
111 | } |
112 | |
113 | spin_unlock_irqrestore(lock: &vif->hash.cache.lock, flags); |
114 | } |
115 | |
116 | static u32 xenvif_find_hash(struct xenvif *vif, const u8 *data, |
117 | unsigned int len) |
118 | { |
119 | struct xenvif_hash_cache_entry *entry; |
120 | u32 val; |
121 | bool found; |
122 | |
123 | if (len >= XEN_NETBK_HASH_TAG_SIZE) |
124 | return 0; |
125 | |
126 | if (xenvif_hash_cache_size == 0) |
127 | return xenvif_new_hash(vif, data, len); |
128 | |
129 | rcu_read_lock(); |
130 | |
131 | found = false; |
132 | |
133 | list_for_each_entry_rcu(entry, &vif->hash.cache.list, link) { |
134 | if (entry->len == len && |
135 | memcmp(p: entry->tag, q: data, size: len) == 0) { |
136 | val = entry->val; |
137 | entry->seq = atomic_inc_return(v: &vif->hash.cache.seq); |
138 | found = true; |
139 | break; |
140 | } |
141 | } |
142 | |
143 | rcu_read_unlock(); |
144 | |
145 | if (!found) |
146 | val = xenvif_new_hash(vif, data, len); |
147 | |
148 | return val; |
149 | } |
150 | |
151 | void xenvif_set_skb_hash(struct xenvif *vif, struct sk_buff *skb) |
152 | { |
153 | struct flow_keys flow; |
154 | u32 hash = 0; |
155 | enum pkt_hash_types type = PKT_HASH_TYPE_NONE; |
156 | u32 flags = vif->hash.flags; |
157 | bool has_tcp_hdr; |
158 | |
159 | /* Quick rejection test: If the network protocol doesn't |
160 | * correspond to any enabled hash type then there's no point |
161 | * in parsing the packet header. |
162 | */ |
163 | switch (skb->protocol) { |
164 | case htons(ETH_P_IP): |
165 | if (flags & (XEN_NETIF_CTRL_HASH_TYPE_IPV4_TCP | |
166 | XEN_NETIF_CTRL_HASH_TYPE_IPV4)) |
167 | break; |
168 | |
169 | goto done; |
170 | |
171 | case htons(ETH_P_IPV6): |
172 | if (flags & (XEN_NETIF_CTRL_HASH_TYPE_IPV6_TCP | |
173 | XEN_NETIF_CTRL_HASH_TYPE_IPV6)) |
174 | break; |
175 | |
176 | goto done; |
177 | |
178 | default: |
179 | goto done; |
180 | } |
181 | |
182 | memset(&flow, 0, sizeof(flow)); |
183 | if (!skb_flow_dissect_flow_keys(skb, flow: &flow, flags: 0)) |
184 | goto done; |
185 | |
186 | has_tcp_hdr = (flow.basic.ip_proto == IPPROTO_TCP) && |
187 | !(flow.control.flags & FLOW_DIS_IS_FRAGMENT); |
188 | |
189 | switch (skb->protocol) { |
190 | case htons(ETH_P_IP): |
191 | if (has_tcp_hdr && |
192 | (flags & XEN_NETIF_CTRL_HASH_TYPE_IPV4_TCP)) { |
193 | u8 data[12]; |
194 | |
195 | memcpy(&data[0], &flow.addrs.v4addrs.src, 4); |
196 | memcpy(&data[4], &flow.addrs.v4addrs.dst, 4); |
197 | memcpy(&data[8], &flow.ports.src, 2); |
198 | memcpy(&data[10], &flow.ports.dst, 2); |
199 | |
200 | hash = xenvif_find_hash(vif, data, len: sizeof(data)); |
201 | type = PKT_HASH_TYPE_L4; |
202 | } else if (flags & XEN_NETIF_CTRL_HASH_TYPE_IPV4) { |
203 | u8 data[8]; |
204 | |
205 | memcpy(&data[0], &flow.addrs.v4addrs.src, 4); |
206 | memcpy(&data[4], &flow.addrs.v4addrs.dst, 4); |
207 | |
208 | hash = xenvif_find_hash(vif, data, len: sizeof(data)); |
209 | type = PKT_HASH_TYPE_L3; |
210 | } |
211 | |
212 | break; |
213 | |
214 | case htons(ETH_P_IPV6): |
215 | if (has_tcp_hdr && |
216 | (flags & XEN_NETIF_CTRL_HASH_TYPE_IPV6_TCP)) { |
217 | u8 data[36]; |
218 | |
219 | memcpy(&data[0], &flow.addrs.v6addrs.src, 16); |
220 | memcpy(&data[16], &flow.addrs.v6addrs.dst, 16); |
221 | memcpy(&data[32], &flow.ports.src, 2); |
222 | memcpy(&data[34], &flow.ports.dst, 2); |
223 | |
224 | hash = xenvif_find_hash(vif, data, len: sizeof(data)); |
225 | type = PKT_HASH_TYPE_L4; |
226 | } else if (flags & XEN_NETIF_CTRL_HASH_TYPE_IPV6) { |
227 | u8 data[32]; |
228 | |
229 | memcpy(&data[0], &flow.addrs.v6addrs.src, 16); |
230 | memcpy(&data[16], &flow.addrs.v6addrs.dst, 16); |
231 | |
232 | hash = xenvif_find_hash(vif, data, len: sizeof(data)); |
233 | type = PKT_HASH_TYPE_L3; |
234 | } |
235 | |
236 | break; |
237 | } |
238 | |
239 | done: |
240 | if (type == PKT_HASH_TYPE_NONE) |
241 | skb_clear_hash(skb); |
242 | else |
243 | __skb_set_sw_hash(skb, hash, is_l4: type == PKT_HASH_TYPE_L4); |
244 | } |
245 | |
246 | u32 xenvif_set_hash_alg(struct xenvif *vif, u32 alg) |
247 | { |
248 | switch (alg) { |
249 | case XEN_NETIF_CTRL_HASH_ALGORITHM_NONE: |
250 | case XEN_NETIF_CTRL_HASH_ALGORITHM_TOEPLITZ: |
251 | break; |
252 | |
253 | default: |
254 | return XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER; |
255 | } |
256 | |
257 | vif->hash.alg = alg; |
258 | |
259 | return XEN_NETIF_CTRL_STATUS_SUCCESS; |
260 | } |
261 | |
262 | u32 xenvif_get_hash_flags(struct xenvif *vif, u32 *flags) |
263 | { |
264 | if (vif->hash.alg == XEN_NETIF_CTRL_HASH_ALGORITHM_NONE) |
265 | return XEN_NETIF_CTRL_STATUS_NOT_SUPPORTED; |
266 | |
267 | *flags = XEN_NETIF_CTRL_HASH_TYPE_IPV4 | |
268 | XEN_NETIF_CTRL_HASH_TYPE_IPV4_TCP | |
269 | XEN_NETIF_CTRL_HASH_TYPE_IPV6 | |
270 | XEN_NETIF_CTRL_HASH_TYPE_IPV6_TCP; |
271 | |
272 | return XEN_NETIF_CTRL_STATUS_SUCCESS; |
273 | } |
274 | |
275 | u32 xenvif_set_hash_flags(struct xenvif *vif, u32 flags) |
276 | { |
277 | if (flags & ~(XEN_NETIF_CTRL_HASH_TYPE_IPV4 | |
278 | XEN_NETIF_CTRL_HASH_TYPE_IPV4_TCP | |
279 | XEN_NETIF_CTRL_HASH_TYPE_IPV6 | |
280 | XEN_NETIF_CTRL_HASH_TYPE_IPV6_TCP)) |
281 | return XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER; |
282 | |
283 | if (vif->hash.alg == XEN_NETIF_CTRL_HASH_ALGORITHM_NONE) |
284 | return XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER; |
285 | |
286 | vif->hash.flags = flags; |
287 | |
288 | return XEN_NETIF_CTRL_STATUS_SUCCESS; |
289 | } |
290 | |
291 | u32 xenvif_set_hash_key(struct xenvif *vif, u32 gref, u32 len) |
292 | { |
293 | u8 *key = vif->hash.key; |
294 | struct gnttab_copy copy_op = { |
295 | .source.u.ref = gref, |
296 | .source.domid = vif->domid, |
297 | .dest.u.gmfn = virt_to_gfn(key), |
298 | .dest.domid = DOMID_SELF, |
299 | .dest.offset = xen_offset_in_page(key), |
300 | .len = len, |
301 | .flags = GNTCOPY_source_gref |
302 | }; |
303 | |
304 | if (len > XEN_NETBK_MAX_HASH_KEY_SIZE) |
305 | return XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER; |
306 | |
307 | if (copy_op.len != 0) { |
308 | gnttab_batch_copy(batch: ©_op, count: 1); |
309 | |
310 | if (copy_op.status != GNTST_okay) |
311 | return XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER; |
312 | } |
313 | |
314 | /* Clear any remaining key octets */ |
315 | if (len < XEN_NETBK_MAX_HASH_KEY_SIZE) |
316 | memset(key + len, 0, XEN_NETBK_MAX_HASH_KEY_SIZE - len); |
317 | |
318 | xenvif_flush_hash(vif); |
319 | |
320 | return XEN_NETIF_CTRL_STATUS_SUCCESS; |
321 | } |
322 | |
323 | u32 xenvif_set_hash_mapping_size(struct xenvif *vif, u32 size) |
324 | { |
325 | if (size > XEN_NETBK_MAX_HASH_MAPPING_SIZE) |
326 | return XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER; |
327 | |
328 | vif->hash.size = size; |
329 | memset(vif->hash.mapping[vif->hash.mapping_sel], 0, |
330 | sizeof(u32) * size); |
331 | |
332 | return XEN_NETIF_CTRL_STATUS_SUCCESS; |
333 | } |
334 | |
335 | u32 xenvif_set_hash_mapping(struct xenvif *vif, u32 gref, u32 len, |
336 | u32 off) |
337 | { |
338 | u32 *mapping = vif->hash.mapping[!vif->hash.mapping_sel]; |
339 | unsigned int nr = 1; |
340 | struct gnttab_copy copy_op[2] = {{ |
341 | .source.u.ref = gref, |
342 | .source.domid = vif->domid, |
343 | .dest.domid = DOMID_SELF, |
344 | .len = len * sizeof(*mapping), |
345 | .flags = GNTCOPY_source_gref |
346 | }}; |
347 | |
348 | if ((off + len < off) || (off + len > vif->hash.size) || |
349 | len > XEN_PAGE_SIZE / sizeof(*mapping)) |
350 | return XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER; |
351 | |
352 | copy_op[0].dest.u.gmfn = virt_to_gfn(mapping + off); |
353 | copy_op[0].dest.offset = xen_offset_in_page(mapping + off); |
354 | if (copy_op[0].dest.offset + copy_op[0].len > XEN_PAGE_SIZE) { |
355 | copy_op[1] = copy_op[0]; |
356 | copy_op[1].source.offset = XEN_PAGE_SIZE - copy_op[0].dest.offset; |
357 | copy_op[1].dest.u.gmfn = virt_to_gfn(mapping + off + len); |
358 | copy_op[1].dest.offset = 0; |
359 | copy_op[1].len = copy_op[0].len - copy_op[1].source.offset; |
360 | copy_op[0].len = copy_op[1].source.offset; |
361 | nr = 2; |
362 | } |
363 | |
364 | memcpy(mapping, vif->hash.mapping[vif->hash.mapping_sel], |
365 | vif->hash.size * sizeof(*mapping)); |
366 | |
367 | if (copy_op[0].len != 0) { |
368 | gnttab_batch_copy(batch: copy_op, count: nr); |
369 | |
370 | if (copy_op[0].status != GNTST_okay || |
371 | copy_op[nr - 1].status != GNTST_okay) |
372 | return XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER; |
373 | } |
374 | |
375 | while (len-- != 0) |
376 | if (mapping[off++] >= vif->num_queues) |
377 | return XEN_NETIF_CTRL_STATUS_INVALID_PARAMETER; |
378 | |
379 | vif->hash.mapping_sel = !vif->hash.mapping_sel; |
380 | |
381 | return XEN_NETIF_CTRL_STATUS_SUCCESS; |
382 | } |
383 | |
384 | #ifdef CONFIG_DEBUG_FS |
385 | void xenvif_dump_hash_info(struct xenvif *vif, struct seq_file *m) |
386 | { |
387 | unsigned int i; |
388 | |
389 | switch (vif->hash.alg) { |
390 | case XEN_NETIF_CTRL_HASH_ALGORITHM_TOEPLITZ: |
391 | seq_puts(m, s: "Hash Algorithm: TOEPLITZ\n" ); |
392 | break; |
393 | |
394 | case XEN_NETIF_CTRL_HASH_ALGORITHM_NONE: |
395 | seq_puts(m, s: "Hash Algorithm: NONE\n" ); |
396 | fallthrough; |
397 | default: |
398 | return; |
399 | } |
400 | |
401 | if (vif->hash.flags) { |
402 | seq_puts(m, s: "\nHash Flags:\n" ); |
403 | |
404 | if (vif->hash.flags & XEN_NETIF_CTRL_HASH_TYPE_IPV4) |
405 | seq_puts(m, s: "- IPv4\n" ); |
406 | if (vif->hash.flags & XEN_NETIF_CTRL_HASH_TYPE_IPV4_TCP) |
407 | seq_puts(m, s: "- IPv4 + TCP\n" ); |
408 | if (vif->hash.flags & XEN_NETIF_CTRL_HASH_TYPE_IPV6) |
409 | seq_puts(m, s: "- IPv6\n" ); |
410 | if (vif->hash.flags & XEN_NETIF_CTRL_HASH_TYPE_IPV6_TCP) |
411 | seq_puts(m, s: "- IPv6 + TCP\n" ); |
412 | } |
413 | |
414 | seq_puts(m, s: "\nHash Key:\n" ); |
415 | |
416 | for (i = 0; i < XEN_NETBK_MAX_HASH_KEY_SIZE; ) { |
417 | unsigned int j, n; |
418 | |
419 | n = 8; |
420 | if (i + n >= XEN_NETBK_MAX_HASH_KEY_SIZE) |
421 | n = XEN_NETBK_MAX_HASH_KEY_SIZE - i; |
422 | |
423 | seq_printf(m, fmt: "[%2u - %2u]: " , i, i + n - 1); |
424 | |
425 | for (j = 0; j < n; j++, i++) |
426 | seq_printf(m, fmt: "%02x " , vif->hash.key[i]); |
427 | |
428 | seq_puts(m, s: "\n" ); |
429 | } |
430 | |
431 | if (vif->hash.size != 0) { |
432 | const u32 *mapping = vif->hash.mapping[vif->hash.mapping_sel]; |
433 | |
434 | seq_puts(m, s: "\nHash Mapping:\n" ); |
435 | |
436 | for (i = 0; i < vif->hash.size; ) { |
437 | unsigned int j, n; |
438 | |
439 | n = 8; |
440 | if (i + n >= vif->hash.size) |
441 | n = vif->hash.size - i; |
442 | |
443 | seq_printf(m, fmt: "[%4u - %4u]: " , i, i + n - 1); |
444 | |
445 | for (j = 0; j < n; j++, i++) |
446 | seq_printf(m, fmt: "%4u " , mapping[i]); |
447 | |
448 | seq_puts(m, s: "\n" ); |
449 | } |
450 | } |
451 | } |
452 | #endif /* CONFIG_DEBUG_FS */ |
453 | |
454 | void xenvif_init_hash(struct xenvif *vif) |
455 | { |
456 | if (xenvif_hash_cache_size == 0) |
457 | return; |
458 | |
459 | BUG_ON(vif->hash.cache.count); |
460 | |
461 | spin_lock_init(&vif->hash.cache.lock); |
462 | INIT_LIST_HEAD(list: &vif->hash.cache.list); |
463 | } |
464 | |
465 | void xenvif_deinit_hash(struct xenvif *vif) |
466 | { |
467 | xenvif_flush_hash(vif); |
468 | } |
469 | |