1 | /* SPDX-License-Identifier: GPL-2.0-or-later */ |
2 | /* AF_RXRPC internal definitions |
3 | * |
4 | * Copyright (C) 2007 Red Hat, Inc. All Rights Reserved. |
5 | * Written by David Howells (dhowells@redhat.com) |
6 | */ |
7 | |
8 | #include <linux/atomic.h> |
9 | #include <linux/seqlock.h> |
10 | #include <linux/win_minmax.h> |
11 | #include <net/net_namespace.h> |
12 | #include <net/netns/generic.h> |
13 | #include <net/sock.h> |
14 | #include <net/af_rxrpc.h> |
15 | #include <keys/rxrpc-type.h> |
16 | #include "protocol.h" |
17 | |
18 | #define FCRYPT_BSIZE 8 |
19 | struct rxrpc_crypt { |
20 | union { |
21 | u8 x[FCRYPT_BSIZE]; |
22 | __be32 n[2]; |
23 | }; |
24 | } __attribute__((aligned(8))); |
25 | |
26 | #define rxrpc_queue_work(WS) queue_work(rxrpc_workqueue, (WS)) |
27 | #define rxrpc_queue_delayed_work(WS,D) \ |
28 | queue_delayed_work(rxrpc_workqueue, (WS), (D)) |
29 | |
30 | struct key_preparsed_payload; |
31 | struct rxrpc_connection; |
32 | struct rxrpc_txbuf; |
33 | |
34 | /* |
35 | * Mark applied to socket buffers in skb->mark. skb->priority is used |
36 | * to pass supplementary information. |
37 | */ |
38 | enum rxrpc_skb_mark { |
39 | RXRPC_SKB_MARK_PACKET, /* Received packet */ |
40 | RXRPC_SKB_MARK_ERROR, /* Error notification */ |
41 | RXRPC_SKB_MARK_SERVICE_CONN_SECURED, /* Service connection response has been verified */ |
42 | RXRPC_SKB_MARK_REJECT_BUSY, /* Reject with BUSY */ |
43 | RXRPC_SKB_MARK_REJECT_ABORT, /* Reject with ABORT (code in skb->priority) */ |
44 | }; |
45 | |
46 | /* |
47 | * sk_state for RxRPC sockets |
48 | */ |
49 | enum { |
50 | RXRPC_UNBOUND = 0, |
51 | RXRPC_CLIENT_UNBOUND, /* Unbound socket used as client */ |
52 | RXRPC_CLIENT_BOUND, /* client local address bound */ |
53 | RXRPC_SERVER_BOUND, /* server local address bound */ |
54 | RXRPC_SERVER_BOUND2, /* second server local address bound */ |
55 | RXRPC_SERVER_LISTENING, /* server listening for connections */ |
56 | RXRPC_SERVER_LISTEN_DISABLED, /* server listening disabled */ |
57 | RXRPC_CLOSE, /* socket is being closed */ |
58 | }; |
59 | |
60 | /* |
61 | * Per-network namespace data. |
62 | */ |
63 | struct rxrpc_net { |
64 | struct proc_dir_entry *proc_net; /* Subdir in /proc/net */ |
65 | u32 epoch; /* Local epoch for detecting local-end reset */ |
66 | struct list_head calls; /* List of calls active in this namespace */ |
67 | spinlock_t call_lock; /* Lock for ->calls */ |
68 | atomic_t nr_calls; /* Count of allocated calls */ |
69 | |
70 | atomic_t nr_conns; |
71 | struct list_head conn_proc_list; /* List of conns in this namespace for proc */ |
72 | struct list_head service_conns; /* Service conns in this namespace */ |
73 | rwlock_t conn_lock; /* Lock for ->conn_proc_list, ->service_conns */ |
74 | struct work_struct service_conn_reaper; |
75 | struct timer_list service_conn_reap_timer; |
76 | |
77 | bool live; |
78 | |
79 | atomic_t nr_client_conns; |
80 | |
81 | struct hlist_head local_endpoints; |
82 | struct mutex local_mutex; /* Lock for ->local_endpoints */ |
83 | |
84 | DECLARE_HASHTABLE (peer_hash, 10); |
85 | spinlock_t peer_hash_lock; /* Lock for ->peer_hash */ |
86 | |
87 | #define RXRPC_KEEPALIVE_TIME 20 /* NAT keepalive time in seconds */ |
88 | u8 peer_keepalive_cursor; |
89 | time64_t peer_keepalive_base; |
90 | struct list_head peer_keepalive[32]; |
91 | struct list_head peer_keepalive_new; |
92 | struct timer_list peer_keepalive_timer; |
93 | struct work_struct peer_keepalive_work; |
94 | |
95 | atomic_t stat_tx_data; |
96 | atomic_t stat_tx_data_retrans; |
97 | atomic_t stat_tx_data_send; |
98 | atomic_t stat_tx_data_send_frag; |
99 | atomic_t stat_tx_data_send_fail; |
100 | atomic_t stat_tx_data_underflow; |
101 | atomic_t stat_tx_data_cwnd_reset; |
102 | atomic_t stat_rx_data; |
103 | atomic_t stat_rx_data_reqack; |
104 | atomic_t stat_rx_data_jumbo; |
105 | |
106 | atomic_t stat_tx_ack_fill; |
107 | atomic_t stat_tx_ack_send; |
108 | atomic_t stat_tx_ack_skip; |
109 | atomic_t stat_tx_acks[256]; |
110 | atomic_t stat_rx_acks[256]; |
111 | |
112 | atomic_t stat_why_req_ack[8]; |
113 | |
114 | atomic_t stat_io_loop; |
115 | }; |
116 | |
117 | /* |
118 | * Service backlog preallocation. |
119 | * |
120 | * This contains circular buffers of preallocated peers, connections and calls |
121 | * for incoming service calls and their head and tail pointers. This allows |
122 | * calls to be set up in the data_ready handler, thereby avoiding the need to |
123 | * shuffle packets around so much. |
124 | */ |
125 | struct rxrpc_backlog { |
126 | unsigned short peer_backlog_head; |
127 | unsigned short peer_backlog_tail; |
128 | unsigned short conn_backlog_head; |
129 | unsigned short conn_backlog_tail; |
130 | unsigned short call_backlog_head; |
131 | unsigned short call_backlog_tail; |
132 | #define RXRPC_BACKLOG_MAX 32 |
133 | struct rxrpc_peer *peer_backlog[RXRPC_BACKLOG_MAX]; |
134 | struct rxrpc_connection *conn_backlog[RXRPC_BACKLOG_MAX]; |
135 | struct rxrpc_call *call_backlog[RXRPC_BACKLOG_MAX]; |
136 | }; |
137 | |
138 | /* |
139 | * RxRPC socket definition |
140 | */ |
141 | struct rxrpc_sock { |
142 | /* WARNING: sk has to be the first member */ |
143 | struct sock sk; |
144 | rxrpc_notify_new_call_t notify_new_call; /* Func to notify of new call */ |
145 | rxrpc_discard_new_call_t discard_new_call; /* Func to discard a new call */ |
146 | struct rxrpc_local *local; /* local endpoint */ |
147 | struct rxrpc_backlog *backlog; /* Preallocation for services */ |
148 | spinlock_t incoming_lock; /* Incoming call vs service shutdown lock */ |
149 | struct list_head sock_calls; /* List of calls owned by this socket */ |
150 | struct list_head to_be_accepted; /* calls awaiting acceptance */ |
151 | struct list_head recvmsg_q; /* Calls awaiting recvmsg's attention */ |
152 | spinlock_t recvmsg_lock; /* Lock for recvmsg_q */ |
153 | struct key *key; /* security for this socket */ |
154 | struct key *securities; /* list of server security descriptors */ |
155 | struct rb_root calls; /* User ID -> call mapping */ |
156 | unsigned long flags; |
157 | #define RXRPC_SOCK_CONNECTED 0 /* connect_srx is set */ |
158 | rwlock_t call_lock; /* lock for calls */ |
159 | u32 min_sec_level; /* minimum security level */ |
160 | #define RXRPC_SECURITY_MAX RXRPC_SECURITY_ENCRYPT |
161 | bool exclusive; /* Exclusive connection for a client socket */ |
162 | u16 second_service; /* Additional service bound to the endpoint */ |
163 | struct { |
164 | /* Service upgrade information */ |
165 | u16 from; /* Service ID to upgrade (if not 0) */ |
166 | u16 to; /* service ID to upgrade to */ |
167 | } service_upgrade; |
168 | sa_family_t family; /* Protocol family created with */ |
169 | struct sockaddr_rxrpc srx; /* Primary Service/local addresses */ |
170 | struct sockaddr_rxrpc connect_srx; /* Default client address from connect() */ |
171 | }; |
172 | |
173 | #define rxrpc_sk(__sk) container_of((__sk), struct rxrpc_sock, sk) |
174 | |
175 | /* |
176 | * CPU-byteorder normalised Rx packet header. |
177 | */ |
178 | struct { |
179 | u32 ; /* client boot timestamp */ |
180 | u32 ; /* connection and channel ID */ |
181 | u32 ; /* call ID (0 for connection-level packets) */ |
182 | u32 ; /* sequence number of pkt in call stream */ |
183 | u32 ; /* serial number of pkt sent to network */ |
184 | u8 ; /* packet type */ |
185 | u8 ; /* packet flags */ |
186 | u8 ; /* app-layer defined status */ |
187 | u8 ; /* security protocol ID */ |
188 | union { |
189 | u16 ; /* reserved */ |
190 | u16 ; /* kerberos security checksum */ |
191 | }; |
192 | u16 ; /* service ID */ |
193 | } __packed; |
194 | |
195 | /* |
196 | * RxRPC socket buffer private variables |
197 | * - max 48 bytes (struct sk_buff::cb) |
198 | */ |
199 | struct rxrpc_skb_priv { |
200 | struct rxrpc_connection *conn; /* Connection referred to (poke packet) */ |
201 | u16 offset; /* Offset of data */ |
202 | u16 len; /* Length of data */ |
203 | u8 flags; |
204 | #define RXRPC_RX_VERIFIED 0x01 |
205 | |
206 | struct rxrpc_host_header hdr; /* RxRPC packet header from this packet */ |
207 | }; |
208 | |
209 | #define rxrpc_skb(__skb) ((struct rxrpc_skb_priv *) &(__skb)->cb) |
210 | |
211 | /* |
212 | * RxRPC security module interface |
213 | */ |
214 | struct rxrpc_security { |
215 | const char *name; /* name of this service */ |
216 | u8 security_index; /* security type provided */ |
217 | u32 no_key_abort; /* Abort code indicating no key */ |
218 | |
219 | /* Initialise a security service */ |
220 | int (*init)(void); |
221 | |
222 | /* Clean up a security service */ |
223 | void (*exit)(void); |
224 | |
225 | /* Parse the information from a server key */ |
226 | int (*preparse_server_key)(struct key_preparsed_payload *); |
227 | |
228 | /* Clean up the preparse buffer after parsing a server key */ |
229 | void (*free_preparse_server_key)(struct key_preparsed_payload *); |
230 | |
231 | /* Destroy the payload of a server key */ |
232 | void (*destroy_server_key)(struct key *); |
233 | |
234 | /* Describe a server key */ |
235 | void (*describe_server_key)(const struct key *, struct seq_file *); |
236 | |
237 | /* initialise a connection's security */ |
238 | int (*init_connection_security)(struct rxrpc_connection *, |
239 | struct rxrpc_key_token *); |
240 | |
241 | /* Work out how much data we can store in a packet, given an estimate |
242 | * of the amount of data remaining. |
243 | */ |
244 | int (*how_much_data)(struct rxrpc_call *, size_t, |
245 | size_t *, size_t *, size_t *); |
246 | |
247 | /* impose security on a packet */ |
248 | int (*secure_packet)(struct rxrpc_call *, struct rxrpc_txbuf *); |
249 | |
250 | /* verify the security on a received packet */ |
251 | int (*verify_packet)(struct rxrpc_call *, struct sk_buff *); |
252 | |
253 | /* Free crypto request on a call */ |
254 | void (*free_call_crypto)(struct rxrpc_call *); |
255 | |
256 | /* issue a challenge */ |
257 | int (*issue_challenge)(struct rxrpc_connection *); |
258 | |
259 | /* respond to a challenge */ |
260 | int (*respond_to_challenge)(struct rxrpc_connection *, |
261 | struct sk_buff *); |
262 | |
263 | /* verify a response */ |
264 | int (*verify_response)(struct rxrpc_connection *, |
265 | struct sk_buff *); |
266 | |
267 | /* clear connection security */ |
268 | void (*clear)(struct rxrpc_connection *); |
269 | }; |
270 | |
271 | /* |
272 | * RxRPC local transport endpoint description |
273 | * - owned by a single AF_RXRPC socket |
274 | * - pointed to by transport socket struct sk_user_data |
275 | */ |
276 | struct rxrpc_local { |
277 | struct rcu_head rcu; |
278 | atomic_t active_users; /* Number of users of the local endpoint */ |
279 | refcount_t ref; /* Number of references to the structure */ |
280 | struct net *net; /* The network namespace */ |
281 | struct rxrpc_net *rxnet; /* Our bits in the network namespace */ |
282 | struct hlist_node link; |
283 | struct socket *socket; /* my UDP socket */ |
284 | struct task_struct *io_thread; |
285 | struct completion io_thread_ready; /* Indication that the I/O thread started */ |
286 | struct rxrpc_sock *service; /* Service(s) listening on this endpoint */ |
287 | #ifdef CONFIG_AF_RXRPC_INJECT_RX_DELAY |
288 | struct sk_buff_head rx_delay_queue; /* Delay injection queue */ |
289 | #endif |
290 | struct sk_buff_head rx_queue; /* Received packets */ |
291 | struct list_head conn_attend_q; /* Conns requiring immediate attention */ |
292 | struct list_head call_attend_q; /* Calls requiring immediate attention */ |
293 | |
294 | struct rb_root client_bundles; /* Client connection bundles by socket params */ |
295 | spinlock_t client_bundles_lock; /* Lock for client_bundles */ |
296 | bool kill_all_client_conns; |
297 | struct list_head idle_client_conns; |
298 | struct timer_list client_conn_reap_timer; |
299 | unsigned long client_conn_flags; |
300 | #define RXRPC_CLIENT_CONN_REAP_TIMER 0 /* The client conn reap timer expired */ |
301 | |
302 | spinlock_t lock; /* access lock */ |
303 | rwlock_t services_lock; /* lock for services list */ |
304 | int debug_id; /* debug ID for printks */ |
305 | bool dead; |
306 | bool service_closed; /* Service socket closed */ |
307 | struct idr conn_ids; /* List of connection IDs */ |
308 | struct list_head new_client_calls; /* Newly created client calls need connection */ |
309 | spinlock_t client_call_lock; /* Lock for ->new_client_calls */ |
310 | struct sockaddr_rxrpc srx; /* local address */ |
311 | }; |
312 | |
313 | /* |
314 | * RxRPC remote transport endpoint definition |
315 | * - matched by local endpoint, remote port, address and protocol type |
316 | */ |
317 | struct rxrpc_peer { |
318 | struct rcu_head rcu; /* This must be first */ |
319 | refcount_t ref; |
320 | unsigned long hash_key; |
321 | struct hlist_node hash_link; |
322 | struct rxrpc_local *local; |
323 | struct hlist_head error_targets; /* targets for net error distribution */ |
324 | struct rb_root service_conns; /* Service connections */ |
325 | struct list_head keepalive_link; /* Link in net->peer_keepalive[] */ |
326 | time64_t last_tx_at; /* Last time packet sent here */ |
327 | seqlock_t service_conn_lock; |
328 | spinlock_t lock; /* access lock */ |
329 | unsigned int if_mtu; /* interface MTU for this peer */ |
330 | unsigned int mtu; /* network MTU for this peer */ |
331 | unsigned int maxdata; /* data size (MTU - hdrsize) */ |
332 | unsigned short hdrsize; /* header size (IP + UDP + RxRPC) */ |
333 | int debug_id; /* debug ID for printks */ |
334 | struct sockaddr_rxrpc srx; /* remote address */ |
335 | |
336 | /* calculated RTT cache */ |
337 | #define RXRPC_RTT_CACHE_SIZE 32 |
338 | spinlock_t rtt_input_lock; /* RTT lock for input routine */ |
339 | ktime_t rtt_last_req; /* Time of last RTT request */ |
340 | unsigned int rtt_count; /* Number of samples we've got */ |
341 | |
342 | u32 srtt_us; /* smoothed round trip time << 3 in usecs */ |
343 | u32 mdev_us; /* medium deviation */ |
344 | u32 mdev_max_us; /* maximal mdev for the last rtt period */ |
345 | u32 rttvar_us; /* smoothed mdev_max */ |
346 | u32 rto_j; /* Retransmission timeout in jiffies */ |
347 | u8 backoff; /* Backoff timeout */ |
348 | |
349 | u8 cong_ssthresh; /* Congestion slow-start threshold */ |
350 | }; |
351 | |
352 | /* |
353 | * Keys for matching a connection. |
354 | */ |
355 | struct rxrpc_conn_proto { |
356 | union { |
357 | struct { |
358 | u32 epoch; /* epoch of this connection */ |
359 | u32 cid; /* connection ID */ |
360 | }; |
361 | u64 index_key; |
362 | }; |
363 | }; |
364 | |
365 | struct rxrpc_conn_parameters { |
366 | struct rxrpc_local *local; /* Representation of local endpoint */ |
367 | struct key *key; /* Security details */ |
368 | bool exclusive; /* T if conn is exclusive */ |
369 | bool upgrade; /* T if service ID can be upgraded */ |
370 | u16 service_id; /* Service ID for this connection */ |
371 | u32 security_level; /* Security level selected */ |
372 | }; |
373 | |
374 | /* |
375 | * Call completion condition (state == RXRPC_CALL_COMPLETE). |
376 | */ |
377 | enum rxrpc_call_completion { |
378 | RXRPC_CALL_SUCCEEDED, /* - Normal termination */ |
379 | RXRPC_CALL_REMOTELY_ABORTED, /* - call aborted by peer */ |
380 | RXRPC_CALL_LOCALLY_ABORTED, /* - call aborted locally on error or close */ |
381 | RXRPC_CALL_LOCAL_ERROR, /* - call failed due to local error */ |
382 | RXRPC_CALL_NETWORK_ERROR, /* - call terminated by network error */ |
383 | NR__RXRPC_CALL_COMPLETIONS |
384 | }; |
385 | |
386 | /* |
387 | * Bits in the connection flags. |
388 | */ |
389 | enum rxrpc_conn_flag { |
390 | RXRPC_CONN_IN_SERVICE_CONNS, /* Conn is in peer->service_conns */ |
391 | RXRPC_CONN_DONT_REUSE, /* Don't reuse this connection */ |
392 | RXRPC_CONN_PROBING_FOR_UPGRADE, /* Probing for service upgrade */ |
393 | RXRPC_CONN_FINAL_ACK_0, /* Need final ACK for channel 0 */ |
394 | RXRPC_CONN_FINAL_ACK_1, /* Need final ACK for channel 1 */ |
395 | RXRPC_CONN_FINAL_ACK_2, /* Need final ACK for channel 2 */ |
396 | RXRPC_CONN_FINAL_ACK_3, /* Need final ACK for channel 3 */ |
397 | }; |
398 | |
399 | #define RXRPC_CONN_FINAL_ACK_MASK ((1UL << RXRPC_CONN_FINAL_ACK_0) | \ |
400 | (1UL << RXRPC_CONN_FINAL_ACK_1) | \ |
401 | (1UL << RXRPC_CONN_FINAL_ACK_2) | \ |
402 | (1UL << RXRPC_CONN_FINAL_ACK_3)) |
403 | |
404 | /* |
405 | * Events that can be raised upon a connection. |
406 | */ |
407 | enum rxrpc_conn_event { |
408 | RXRPC_CONN_EV_CHALLENGE, /* Send challenge packet */ |
409 | RXRPC_CONN_EV_ABORT_CALLS, /* Abort attached calls */ |
410 | }; |
411 | |
412 | /* |
413 | * The connection protocol state. |
414 | */ |
415 | enum rxrpc_conn_proto_state { |
416 | RXRPC_CONN_UNUSED, /* Connection not yet attempted */ |
417 | RXRPC_CONN_CLIENT_UNSECURED, /* Client connection needs security init */ |
418 | RXRPC_CONN_CLIENT, /* Client connection */ |
419 | RXRPC_CONN_SERVICE_PREALLOC, /* Service connection preallocation */ |
420 | RXRPC_CONN_SERVICE_UNSECURED, /* Service unsecured connection */ |
421 | RXRPC_CONN_SERVICE_CHALLENGING, /* Service challenging for security */ |
422 | RXRPC_CONN_SERVICE, /* Service secured connection */ |
423 | RXRPC_CONN_ABORTED, /* Conn aborted */ |
424 | RXRPC_CONN__NR_STATES |
425 | }; |
426 | |
427 | /* |
428 | * RxRPC client connection bundle. |
429 | */ |
430 | struct rxrpc_bundle { |
431 | struct rxrpc_local *local; /* Representation of local endpoint */ |
432 | struct rxrpc_peer *peer; /* Remote endpoint */ |
433 | struct key *key; /* Security details */ |
434 | const struct rxrpc_security *security; /* applied security module */ |
435 | refcount_t ref; |
436 | atomic_t active; /* Number of active users */ |
437 | unsigned int debug_id; |
438 | u32 security_level; /* Security level selected */ |
439 | u16 service_id; /* Service ID for this connection */ |
440 | bool try_upgrade; /* True if the bundle is attempting upgrade */ |
441 | bool exclusive; /* T if conn is exclusive */ |
442 | bool upgrade; /* T if service ID can be upgraded */ |
443 | unsigned short alloc_error; /* Error from last conn allocation */ |
444 | struct rb_node local_node; /* Node in local->client_conns */ |
445 | struct list_head waiting_calls; /* Calls waiting for channels */ |
446 | unsigned long avail_chans; /* Mask of available channels */ |
447 | struct rxrpc_connection *conns[4]; /* The connections in the bundle (max 4) */ |
448 | }; |
449 | |
450 | /* |
451 | * RxRPC connection definition |
452 | * - matched by { local, peer, epoch, conn_id, direction } |
453 | * - each connection can only handle four simultaneous calls |
454 | */ |
455 | struct rxrpc_connection { |
456 | struct rxrpc_conn_proto proto; |
457 | struct rxrpc_local *local; /* Representation of local endpoint */ |
458 | struct rxrpc_peer *peer; /* Remote endpoint */ |
459 | struct rxrpc_net *rxnet; /* Network namespace to which call belongs */ |
460 | struct key *key; /* Security details */ |
461 | struct list_head attend_link; /* Link in local->conn_attend_q */ |
462 | |
463 | refcount_t ref; |
464 | atomic_t active; /* Active count for service conns */ |
465 | struct rcu_head rcu; |
466 | struct list_head cache_link; |
467 | |
468 | unsigned char act_chans; /* Mask of active channels */ |
469 | struct rxrpc_channel { |
470 | unsigned long final_ack_at; /* Time at which to issue final ACK */ |
471 | struct rxrpc_call *call; /* Active call */ |
472 | unsigned int call_debug_id; /* call->debug_id */ |
473 | u32 call_id; /* ID of current call */ |
474 | u32 call_counter; /* Call ID counter */ |
475 | u32 last_call; /* ID of last call */ |
476 | u8 last_type; /* Type of last packet */ |
477 | union { |
478 | u32 last_seq; |
479 | u32 last_abort; |
480 | }; |
481 | } channels[RXRPC_MAXCALLS]; |
482 | |
483 | struct timer_list timer; /* Conn event timer */ |
484 | struct work_struct processor; /* connection event processor */ |
485 | struct work_struct destructor; /* In-process-context destroyer */ |
486 | struct rxrpc_bundle *bundle; /* Client connection bundle */ |
487 | struct rb_node service_node; /* Node in peer->service_conns */ |
488 | struct list_head proc_link; /* link in procfs list */ |
489 | struct list_head link; /* link in master connection list */ |
490 | struct sk_buff_head rx_queue; /* received conn-level packets */ |
491 | |
492 | struct mutex security_lock; /* Lock for security management */ |
493 | const struct rxrpc_security *security; /* applied security module */ |
494 | union { |
495 | struct { |
496 | struct crypto_sync_skcipher *cipher; /* encryption handle */ |
497 | struct rxrpc_crypt csum_iv; /* packet checksum base */ |
498 | u32 nonce; /* response re-use preventer */ |
499 | } rxkad; |
500 | }; |
501 | unsigned long flags; |
502 | unsigned long events; |
503 | unsigned long idle_timestamp; /* Time at which last became idle */ |
504 | spinlock_t state_lock; /* state-change lock */ |
505 | enum rxrpc_conn_proto_state state; /* current state of connection */ |
506 | enum rxrpc_call_completion completion; /* Completion condition */ |
507 | s32 abort_code; /* Abort code of connection abort */ |
508 | int debug_id; /* debug ID for printks */ |
509 | atomic_t serial; /* packet serial number counter */ |
510 | unsigned int hi_serial; /* highest serial number received */ |
511 | u32 service_id; /* Service ID, possibly upgraded */ |
512 | u32 security_level; /* Security level selected */ |
513 | u8 security_ix; /* security type */ |
514 | u8 out_clientflag; /* RXRPC_CLIENT_INITIATED if we are client */ |
515 | u8 bundle_shift; /* Index into bundle->avail_chans */ |
516 | bool exclusive; /* T if conn is exclusive */ |
517 | bool upgrade; /* T if service ID can be upgraded */ |
518 | u16 orig_service_id; /* Originally requested service ID */ |
519 | short error; /* Local error code */ |
520 | }; |
521 | |
522 | static inline bool rxrpc_to_server(const struct rxrpc_skb_priv *sp) |
523 | { |
524 | return sp->hdr.flags & RXRPC_CLIENT_INITIATED; |
525 | } |
526 | |
527 | static inline bool rxrpc_to_client(const struct rxrpc_skb_priv *sp) |
528 | { |
529 | return !rxrpc_to_server(sp); |
530 | } |
531 | |
532 | /* |
533 | * Flags in call->flags. |
534 | */ |
535 | enum rxrpc_call_flag { |
536 | RXRPC_CALL_RELEASED, /* call has been released - no more message to userspace */ |
537 | RXRPC_CALL_HAS_USERID, /* has a user ID attached */ |
538 | RXRPC_CALL_IS_SERVICE, /* Call is service call */ |
539 | RXRPC_CALL_EXPOSED, /* The call was exposed to the world */ |
540 | RXRPC_CALL_RX_LAST, /* Received the last packet (at rxtx_top) */ |
541 | RXRPC_CALL_TX_LAST, /* Last packet in Tx buffer (at rxtx_top) */ |
542 | RXRPC_CALL_TX_ALL_ACKED, /* Last packet has been hard-acked */ |
543 | RXRPC_CALL_SEND_PING, /* A ping will need to be sent */ |
544 | RXRPC_CALL_RETRANS_TIMEOUT, /* Retransmission due to timeout occurred */ |
545 | RXRPC_CALL_BEGAN_RX_TIMER, /* We began the expect_rx_by timer */ |
546 | RXRPC_CALL_RX_HEARD, /* The peer responded at least once to this call */ |
547 | RXRPC_CALL_DISCONNECTED, /* The call has been disconnected */ |
548 | RXRPC_CALL_KERNEL, /* The call was made by the kernel */ |
549 | RXRPC_CALL_UPGRADE, /* Service upgrade was requested for the call */ |
550 | RXRPC_CALL_EXCLUSIVE, /* The call uses a once-only connection */ |
551 | RXRPC_CALL_RX_IS_IDLE, /* recvmsg() is idle - send an ACK */ |
552 | RXRPC_CALL_RECVMSG_READ_ALL, /* recvmsg() read all of the received data */ |
553 | }; |
554 | |
555 | /* |
556 | * Events that can be raised on a call. |
557 | */ |
558 | enum rxrpc_call_event { |
559 | RXRPC_CALL_EV_ACK_LOST, /* ACK may be lost, send ping */ |
560 | RXRPC_CALL_EV_INITIAL_PING, /* Send initial ping for a new service call */ |
561 | }; |
562 | |
563 | /* |
564 | * The states that a call can be in. |
565 | */ |
566 | enum rxrpc_call_state { |
567 | RXRPC_CALL_UNINITIALISED, |
568 | RXRPC_CALL_CLIENT_AWAIT_CONN, /* - client waiting for connection to become available */ |
569 | RXRPC_CALL_CLIENT_SEND_REQUEST, /* - client sending request phase */ |
570 | RXRPC_CALL_CLIENT_AWAIT_REPLY, /* - client awaiting reply */ |
571 | RXRPC_CALL_CLIENT_RECV_REPLY, /* - client receiving reply phase */ |
572 | RXRPC_CALL_SERVER_PREALLOC, /* - service preallocation */ |
573 | RXRPC_CALL_SERVER_SECURING, /* - server securing request connection */ |
574 | RXRPC_CALL_SERVER_RECV_REQUEST, /* - server receiving request */ |
575 | RXRPC_CALL_SERVER_ACK_REQUEST, /* - server pending ACK of request */ |
576 | RXRPC_CALL_SERVER_SEND_REPLY, /* - server sending reply */ |
577 | RXRPC_CALL_SERVER_AWAIT_ACK, /* - server awaiting final ACK */ |
578 | RXRPC_CALL_COMPLETE, /* - call complete */ |
579 | NR__RXRPC_CALL_STATES |
580 | }; |
581 | |
582 | /* |
583 | * Call Tx congestion management modes. |
584 | */ |
585 | enum rxrpc_congest_mode { |
586 | RXRPC_CALL_SLOW_START, |
587 | RXRPC_CALL_CONGEST_AVOIDANCE, |
588 | RXRPC_CALL_PACKET_LOSS, |
589 | RXRPC_CALL_FAST_RETRANSMIT, |
590 | NR__RXRPC_CONGEST_MODES |
591 | }; |
592 | |
593 | /* |
594 | * RxRPC call definition |
595 | * - matched by { connection, call_id } |
596 | */ |
597 | struct rxrpc_call { |
598 | struct rcu_head rcu; |
599 | struct rxrpc_connection *conn; /* connection carrying call */ |
600 | struct rxrpc_bundle *bundle; /* Connection bundle to use */ |
601 | struct rxrpc_peer *peer; /* Peer record for remote address */ |
602 | struct rxrpc_local *local; /* Representation of local endpoint */ |
603 | struct rxrpc_sock __rcu *socket; /* socket responsible */ |
604 | struct rxrpc_net *rxnet; /* Network namespace to which call belongs */ |
605 | struct key *key; /* Security details */ |
606 | const struct rxrpc_security *security; /* applied security module */ |
607 | struct mutex user_mutex; /* User access mutex */ |
608 | struct sockaddr_rxrpc dest_srx; /* Destination address */ |
609 | unsigned long delay_ack_at; /* When DELAY ACK needs to happen */ |
610 | unsigned long ack_lost_at; /* When ACK is figured as lost */ |
611 | unsigned long resend_at; /* When next resend needs to happen */ |
612 | unsigned long ping_at; /* When next to send a ping */ |
613 | unsigned long keepalive_at; /* When next to send a keepalive ping */ |
614 | unsigned long expect_rx_by; /* When we expect to get a packet by */ |
615 | unsigned long expect_req_by; /* When we expect to get a request DATA packet by */ |
616 | unsigned long expect_term_by; /* When we expect call termination by */ |
617 | u32 next_rx_timo; /* Timeout for next Rx packet (jif) */ |
618 | u32 next_req_timo; /* Timeout for next Rx request packet (jif) */ |
619 | u32 hard_timo; /* Maximum lifetime or 0 (jif) */ |
620 | struct timer_list timer; /* Combined event timer */ |
621 | struct work_struct destroyer; /* In-process-context destroyer */ |
622 | rxrpc_notify_rx_t notify_rx; /* kernel service Rx notification function */ |
623 | struct list_head link; /* link in master call list */ |
624 | struct list_head wait_link; /* Link in local->new_client_calls */ |
625 | struct hlist_node error_link; /* link in error distribution list */ |
626 | struct list_head accept_link; /* Link in rx->acceptq */ |
627 | struct list_head recvmsg_link; /* Link in rx->recvmsg_q */ |
628 | struct list_head sock_link; /* Link in rx->sock_calls */ |
629 | struct rb_node sock_node; /* Node in rx->calls */ |
630 | struct list_head attend_link; /* Link in local->call_attend_q */ |
631 | struct rxrpc_txbuf *tx_pending; /* Tx buffer being filled */ |
632 | wait_queue_head_t waitq; /* Wait queue for channel or Tx */ |
633 | s64 tx_total_len; /* Total length left to be transmitted (or -1) */ |
634 | unsigned long user_call_ID; /* user-defined call ID */ |
635 | unsigned long flags; |
636 | unsigned long events; |
637 | spinlock_t notify_lock; /* Kernel notification lock */ |
638 | unsigned int send_abort_why; /* Why the abort [enum rxrpc_abort_reason] */ |
639 | s32 send_abort; /* Abort code to be sent */ |
640 | short send_abort_err; /* Error to be associated with the abort */ |
641 | rxrpc_seq_t send_abort_seq; /* DATA packet that incurred the abort (or 0) */ |
642 | s32 abort_code; /* Local/remote abort code */ |
643 | int error; /* Local error incurred */ |
644 | enum rxrpc_call_state _state; /* Current state of call (needs barrier) */ |
645 | enum rxrpc_call_completion completion; /* Call completion condition */ |
646 | refcount_t ref; |
647 | u8 security_ix; /* Security type */ |
648 | enum rxrpc_interruptibility interruptibility; /* At what point call may be interrupted */ |
649 | u32 call_id; /* call ID on connection */ |
650 | u32 cid; /* connection ID plus channel index */ |
651 | u32 security_level; /* Security level selected */ |
652 | int debug_id; /* debug ID for printks */ |
653 | unsigned short rx_pkt_offset; /* Current recvmsg packet offset */ |
654 | unsigned short rx_pkt_len; /* Current recvmsg packet len */ |
655 | |
656 | /* Transmitted data tracking. */ |
657 | spinlock_t tx_lock; /* Transmit queue lock */ |
658 | struct list_head tx_sendmsg; /* Sendmsg prepared packets */ |
659 | struct list_head tx_buffer; /* Buffer of transmissible packets */ |
660 | rxrpc_seq_t tx_bottom; /* First packet in buffer */ |
661 | rxrpc_seq_t tx_transmitted; /* Highest packet transmitted */ |
662 | rxrpc_seq_t tx_prepared; /* Highest Tx slot prepared. */ |
663 | rxrpc_seq_t tx_top; /* Highest Tx slot allocated. */ |
664 | u16 tx_backoff; /* Delay to insert due to Tx failure */ |
665 | u8 tx_winsize; /* Maximum size of Tx window */ |
666 | #define RXRPC_TX_MAX_WINDOW 128 |
667 | ktime_t tx_last_sent; /* Last time a transmission occurred */ |
668 | |
669 | /* Received data tracking */ |
670 | struct sk_buff_head recvmsg_queue; /* Queue of packets ready for recvmsg() */ |
671 | struct sk_buff_head rx_oos_queue; /* Queue of out of sequence packets */ |
672 | |
673 | rxrpc_seq_t rx_highest_seq; /* Higest sequence number received */ |
674 | rxrpc_seq_t rx_consumed; /* Highest packet consumed */ |
675 | rxrpc_serial_t rx_serial; /* Highest serial received for this call */ |
676 | u8 rx_winsize; /* Size of Rx window */ |
677 | |
678 | /* TCP-style slow-start congestion control [RFC5681]. Since the SMSS |
679 | * is fixed, we keep these numbers in terms of segments (ie. DATA |
680 | * packets) rather than bytes. |
681 | */ |
682 | #define RXRPC_TX_SMSS RXRPC_JUMBO_DATALEN |
683 | #define RXRPC_MIN_CWND (RXRPC_TX_SMSS > 2190 ? 2 : RXRPC_TX_SMSS > 1095 ? 3 : 4) |
684 | u8 cong_cwnd; /* Congestion window size */ |
685 | u8 ; /* Extra to send for congestion management */ |
686 | u8 cong_ssthresh; /* Slow-start threshold */ |
687 | enum rxrpc_congest_mode cong_mode:8; /* Congestion management mode */ |
688 | u8 cong_dup_acks; /* Count of ACKs showing missing packets */ |
689 | u8 cong_cumul_acks; /* Cumulative ACK count */ |
690 | ktime_t cong_tstamp; /* Last time cwnd was changed */ |
691 | |
692 | /* Receive-phase ACK management (ACKs we send). */ |
693 | u8 ackr_reason; /* reason to ACK */ |
694 | u16 ackr_sack_base; /* Starting slot in SACK table ring */ |
695 | rxrpc_serial_t ackr_serial; /* serial of packet being ACK'd */ |
696 | rxrpc_seq_t ackr_window; /* Base of SACK window */ |
697 | rxrpc_seq_t ackr_wtop; /* Base of SACK window */ |
698 | unsigned int ackr_nr_unacked; /* Number of unacked packets */ |
699 | atomic_t ackr_nr_consumed; /* Number of packets needing hard ACK */ |
700 | struct { |
701 | #define RXRPC_SACK_SIZE 256 |
702 | /* SACK table for soft-acked packets */ |
703 | u8 ackr_sack_table[RXRPC_SACK_SIZE]; |
704 | } __aligned(8); |
705 | |
706 | /* RTT management */ |
707 | rxrpc_serial_t rtt_serial[4]; /* Serial number of DATA or PING sent */ |
708 | ktime_t rtt_sent_at[4]; /* Time packet sent */ |
709 | unsigned long rtt_avail; /* Mask of available slots in bits 0-3, |
710 | * Mask of pending samples in 8-11 */ |
711 | #define RXRPC_CALL_RTT_AVAIL_MASK 0xf |
712 | #define RXRPC_CALL_RTT_PEND_SHIFT 8 |
713 | |
714 | /* Transmission-phase ACK management (ACKs we've received). */ |
715 | ktime_t acks_latest_ts; /* Timestamp of latest ACK received */ |
716 | rxrpc_seq_t acks_first_seq; /* first sequence number received */ |
717 | rxrpc_seq_t acks_prev_seq; /* Highest previousPacket received */ |
718 | rxrpc_seq_t acks_hard_ack; /* Latest hard-ack point */ |
719 | rxrpc_seq_t acks_lowest_nak; /* Lowest NACK in the buffer (or ==tx_hard_ack) */ |
720 | rxrpc_serial_t acks_highest_serial; /* Highest serial number ACK'd */ |
721 | }; |
722 | |
723 | /* |
724 | * Summary of a new ACK and the changes it made to the Tx buffer packet states. |
725 | */ |
726 | struct rxrpc_ack_summary { |
727 | u16 nr_acks; /* Number of ACKs in packet */ |
728 | u16 nr_new_acks; /* Number of new ACKs in packet */ |
729 | u16 nr_rot_new_acks; /* Number of rotated new ACKs */ |
730 | u8 ack_reason; |
731 | bool saw_nacks; /* Saw NACKs in packet */ |
732 | bool new_low_nack; /* T if new low NACK found */ |
733 | bool retrans_timeo; /* T if reTx due to timeout happened */ |
734 | u8 flight_size; /* Number of unreceived transmissions */ |
735 | /* Place to stash values for tracing */ |
736 | enum rxrpc_congest_mode mode:8; |
737 | u8 cwnd; |
738 | u8 ssthresh; |
739 | u8 dup_acks; |
740 | u8 cumulative_acks; |
741 | }; |
742 | |
743 | /* |
744 | * sendmsg() cmsg-specified parameters. |
745 | */ |
746 | enum rxrpc_command { |
747 | RXRPC_CMD_SEND_DATA, /* send data message */ |
748 | RXRPC_CMD_SEND_ABORT, /* request abort generation */ |
749 | RXRPC_CMD_REJECT_BUSY, /* [server] reject a call as busy */ |
750 | RXRPC_CMD_CHARGE_ACCEPT, /* [server] charge accept preallocation */ |
751 | }; |
752 | |
753 | struct rxrpc_call_params { |
754 | s64 tx_total_len; /* Total Tx data length (if send data) */ |
755 | unsigned long user_call_ID; /* User's call ID */ |
756 | struct { |
757 | u32 hard; /* Maximum lifetime (sec) */ |
758 | u32 idle; /* Max time since last data packet (msec) */ |
759 | u32 normal; /* Max time since last call packet (msec) */ |
760 | } timeouts; |
761 | u8 nr_timeouts; /* Number of timeouts specified */ |
762 | bool kernel; /* T if kernel is making the call */ |
763 | enum rxrpc_interruptibility interruptibility; /* How is interruptible is the call? */ |
764 | }; |
765 | |
766 | struct rxrpc_send_params { |
767 | struct rxrpc_call_params call; |
768 | u32 abort_code; /* Abort code to Tx (if abort) */ |
769 | enum rxrpc_command command : 8; /* The command to implement */ |
770 | bool exclusive; /* Shared or exclusive call */ |
771 | bool upgrade; /* If the connection is upgradeable */ |
772 | }; |
773 | |
774 | /* |
775 | * Buffer of data to be output as a packet. |
776 | */ |
777 | struct rxrpc_txbuf { |
778 | struct rcu_head rcu; |
779 | struct list_head call_link; /* Link in call->tx_sendmsg/tx_buffer */ |
780 | struct list_head tx_link; /* Link in live Enc queue or Tx queue */ |
781 | ktime_t last_sent; /* Time at which last transmitted */ |
782 | refcount_t ref; |
783 | rxrpc_seq_t seq; /* Sequence number of this packet */ |
784 | unsigned int call_debug_id; |
785 | unsigned int debug_id; |
786 | unsigned int len; /* Amount of data in buffer */ |
787 | unsigned int space; /* Remaining data space */ |
788 | unsigned int offset; /* Offset of fill point */ |
789 | unsigned long flags; |
790 | #define RXRPC_TXBUF_LAST 0 /* Set if last packet in Tx phase */ |
791 | #define RXRPC_TXBUF_RESENT 1 /* Set if has been resent */ |
792 | u8 /*enum rxrpc_propose_ack_trace*/ ack_why; /* If ack, why */ |
793 | struct { |
794 | /* The packet for encrypting and DMA'ing. We align it such |
795 | * that data[] aligns correctly for any crypto blocksize. |
796 | */ |
797 | u8 pad[64 - sizeof(struct rxrpc_wire_header)]; |
798 | struct rxrpc_wire_header wire; /* Network-ready header */ |
799 | union { |
800 | u8 data[RXRPC_JUMBO_DATALEN]; /* Data packet */ |
801 | struct { |
802 | struct rxrpc_ackpacket ack; |
803 | DECLARE_FLEX_ARRAY(u8, acks); |
804 | }; |
805 | }; |
806 | } __aligned(64); |
807 | }; |
808 | |
809 | static inline bool rxrpc_sending_to_server(const struct rxrpc_txbuf *txb) |
810 | { |
811 | return txb->wire.flags & RXRPC_CLIENT_INITIATED; |
812 | } |
813 | |
814 | static inline bool rxrpc_sending_to_client(const struct rxrpc_txbuf *txb) |
815 | { |
816 | return !rxrpc_sending_to_server(txb); |
817 | } |
818 | |
819 | #include <trace/events/rxrpc.h> |
820 | |
821 | /* |
822 | * af_rxrpc.c |
823 | */ |
824 | extern atomic_t rxrpc_n_rx_skbs; |
825 | extern struct workqueue_struct *rxrpc_workqueue; |
826 | |
827 | /* |
828 | * call_accept.c |
829 | */ |
830 | int rxrpc_service_prealloc(struct rxrpc_sock *, gfp_t); |
831 | void rxrpc_discard_prealloc(struct rxrpc_sock *); |
832 | bool rxrpc_new_incoming_call(struct rxrpc_local *local, |
833 | struct rxrpc_peer *peer, |
834 | struct rxrpc_connection *conn, |
835 | struct sockaddr_rxrpc *peer_srx, |
836 | struct sk_buff *skb); |
837 | void rxrpc_accept_incoming_calls(struct rxrpc_local *); |
838 | int rxrpc_user_charge_accept(struct rxrpc_sock *, unsigned long); |
839 | |
840 | /* |
841 | * call_event.c |
842 | */ |
843 | void rxrpc_propose_ping(struct rxrpc_call *call, u32 serial, |
844 | enum rxrpc_propose_ack_trace why); |
845 | void rxrpc_send_ACK(struct rxrpc_call *, u8, rxrpc_serial_t, enum rxrpc_propose_ack_trace); |
846 | void rxrpc_propose_delay_ACK(struct rxrpc_call *, rxrpc_serial_t, |
847 | enum rxrpc_propose_ack_trace); |
848 | void rxrpc_shrink_call_tx_buffer(struct rxrpc_call *); |
849 | void rxrpc_resend(struct rxrpc_call *call, struct sk_buff *ack_skb); |
850 | |
851 | void rxrpc_reduce_call_timer(struct rxrpc_call *call, |
852 | unsigned long expire_at, |
853 | unsigned long now, |
854 | enum rxrpc_timer_trace why); |
855 | |
856 | bool rxrpc_input_call_event(struct rxrpc_call *call, struct sk_buff *skb); |
857 | |
858 | /* |
859 | * call_object.c |
860 | */ |
861 | extern const char *const rxrpc_call_states[]; |
862 | extern const char *const rxrpc_call_completions[]; |
863 | extern struct kmem_cache *rxrpc_call_jar; |
864 | |
865 | void rxrpc_poke_call(struct rxrpc_call *call, enum rxrpc_call_poke_trace what); |
866 | struct rxrpc_call *rxrpc_find_call_by_user_ID(struct rxrpc_sock *, unsigned long); |
867 | struct rxrpc_call *rxrpc_alloc_call(struct rxrpc_sock *, gfp_t, unsigned int); |
868 | struct rxrpc_call *rxrpc_new_client_call(struct rxrpc_sock *, |
869 | struct rxrpc_conn_parameters *, |
870 | struct sockaddr_rxrpc *, |
871 | struct rxrpc_call_params *, gfp_t, |
872 | unsigned int); |
873 | void rxrpc_start_call_timer(struct rxrpc_call *call); |
874 | void rxrpc_incoming_call(struct rxrpc_sock *, struct rxrpc_call *, |
875 | struct sk_buff *); |
876 | void rxrpc_release_call(struct rxrpc_sock *, struct rxrpc_call *); |
877 | void rxrpc_release_calls_on_socket(struct rxrpc_sock *); |
878 | void rxrpc_see_call(struct rxrpc_call *, enum rxrpc_call_trace); |
879 | struct rxrpc_call *rxrpc_try_get_call(struct rxrpc_call *, enum rxrpc_call_trace); |
880 | void rxrpc_get_call(struct rxrpc_call *, enum rxrpc_call_trace); |
881 | void rxrpc_put_call(struct rxrpc_call *, enum rxrpc_call_trace); |
882 | void rxrpc_cleanup_call(struct rxrpc_call *); |
883 | void rxrpc_destroy_all_calls(struct rxrpc_net *); |
884 | |
885 | static inline bool rxrpc_is_service_call(const struct rxrpc_call *call) |
886 | { |
887 | return test_bit(RXRPC_CALL_IS_SERVICE, &call->flags); |
888 | } |
889 | |
890 | static inline bool rxrpc_is_client_call(const struct rxrpc_call *call) |
891 | { |
892 | return !rxrpc_is_service_call(call); |
893 | } |
894 | |
895 | /* |
896 | * call_state.c |
897 | */ |
898 | bool rxrpc_set_call_completion(struct rxrpc_call *call, |
899 | enum rxrpc_call_completion compl, |
900 | u32 abort_code, |
901 | int error); |
902 | bool rxrpc_call_completed(struct rxrpc_call *call); |
903 | bool rxrpc_abort_call(struct rxrpc_call *call, rxrpc_seq_t seq, |
904 | u32 abort_code, int error, enum rxrpc_abort_reason why); |
905 | void rxrpc_prefail_call(struct rxrpc_call *call, enum rxrpc_call_completion compl, |
906 | int error); |
907 | |
908 | static inline void rxrpc_set_call_state(struct rxrpc_call *call, |
909 | enum rxrpc_call_state state) |
910 | { |
911 | /* Order write of completion info before write of ->state. */ |
912 | smp_store_release(&call->_state, state); |
913 | wake_up(&call->waitq); |
914 | } |
915 | |
916 | static inline enum rxrpc_call_state __rxrpc_call_state(const struct rxrpc_call *call) |
917 | { |
918 | return call->_state; /* Only inside I/O thread */ |
919 | } |
920 | |
921 | static inline bool __rxrpc_call_is_complete(const struct rxrpc_call *call) |
922 | { |
923 | return __rxrpc_call_state(call) == RXRPC_CALL_COMPLETE; |
924 | } |
925 | |
926 | static inline enum rxrpc_call_state rxrpc_call_state(const struct rxrpc_call *call) |
927 | { |
928 | /* Order read ->state before read of completion info. */ |
929 | return smp_load_acquire(&call->_state); |
930 | } |
931 | |
932 | static inline bool rxrpc_call_is_complete(const struct rxrpc_call *call) |
933 | { |
934 | return rxrpc_call_state(call) == RXRPC_CALL_COMPLETE; |
935 | } |
936 | |
937 | static inline bool rxrpc_call_has_failed(const struct rxrpc_call *call) |
938 | { |
939 | return rxrpc_call_is_complete(call) && call->completion != RXRPC_CALL_SUCCEEDED; |
940 | } |
941 | |
942 | /* |
943 | * conn_client.c |
944 | */ |
945 | extern unsigned int rxrpc_reap_client_connections; |
946 | extern unsigned long rxrpc_conn_idle_client_expiry; |
947 | extern unsigned long rxrpc_conn_idle_client_fast_expiry; |
948 | |
949 | void rxrpc_purge_client_connections(struct rxrpc_local *local); |
950 | struct rxrpc_bundle *rxrpc_get_bundle(struct rxrpc_bundle *, enum rxrpc_bundle_trace); |
951 | void rxrpc_put_bundle(struct rxrpc_bundle *, enum rxrpc_bundle_trace); |
952 | int rxrpc_look_up_bundle(struct rxrpc_call *call, gfp_t gfp); |
953 | void rxrpc_connect_client_calls(struct rxrpc_local *local); |
954 | void rxrpc_expose_client_call(struct rxrpc_call *); |
955 | void rxrpc_disconnect_client_call(struct rxrpc_bundle *, struct rxrpc_call *); |
956 | void rxrpc_deactivate_bundle(struct rxrpc_bundle *bundle); |
957 | void rxrpc_put_client_conn(struct rxrpc_connection *, enum rxrpc_conn_trace); |
958 | void rxrpc_discard_expired_client_conns(struct rxrpc_local *local); |
959 | void rxrpc_clean_up_local_conns(struct rxrpc_local *); |
960 | |
961 | /* |
962 | * conn_event.c |
963 | */ |
964 | void rxrpc_conn_retransmit_call(struct rxrpc_connection *conn, struct sk_buff *skb, |
965 | unsigned int channel); |
966 | int rxrpc_abort_conn(struct rxrpc_connection *conn, struct sk_buff *skb, |
967 | s32 abort_code, int err, enum rxrpc_abort_reason why); |
968 | void rxrpc_process_connection(struct work_struct *); |
969 | void rxrpc_process_delayed_final_acks(struct rxrpc_connection *, bool); |
970 | bool rxrpc_input_conn_packet(struct rxrpc_connection *conn, struct sk_buff *skb); |
971 | void rxrpc_input_conn_event(struct rxrpc_connection *conn, struct sk_buff *skb); |
972 | |
973 | static inline bool rxrpc_is_conn_aborted(const struct rxrpc_connection *conn) |
974 | { |
975 | /* Order reading the abort info after the state check. */ |
976 | return smp_load_acquire(&conn->state) == RXRPC_CONN_ABORTED; |
977 | } |
978 | |
979 | /* |
980 | * conn_object.c |
981 | */ |
982 | extern unsigned int rxrpc_connection_expiry; |
983 | extern unsigned int rxrpc_closed_conn_expiry; |
984 | |
985 | void rxrpc_poke_conn(struct rxrpc_connection *conn, enum rxrpc_conn_trace why); |
986 | struct rxrpc_connection *rxrpc_alloc_connection(struct rxrpc_net *, gfp_t); |
987 | struct rxrpc_connection *rxrpc_find_client_connection_rcu(struct rxrpc_local *, |
988 | struct sockaddr_rxrpc *, |
989 | struct sk_buff *); |
990 | void __rxrpc_disconnect_call(struct rxrpc_connection *, struct rxrpc_call *); |
991 | void rxrpc_disconnect_call(struct rxrpc_call *); |
992 | void rxrpc_kill_client_conn(struct rxrpc_connection *); |
993 | void rxrpc_queue_conn(struct rxrpc_connection *, enum rxrpc_conn_trace); |
994 | void rxrpc_see_connection(struct rxrpc_connection *, enum rxrpc_conn_trace); |
995 | struct rxrpc_connection *rxrpc_get_connection(struct rxrpc_connection *, |
996 | enum rxrpc_conn_trace); |
997 | struct rxrpc_connection *rxrpc_get_connection_maybe(struct rxrpc_connection *, |
998 | enum rxrpc_conn_trace); |
999 | void rxrpc_put_connection(struct rxrpc_connection *, enum rxrpc_conn_trace); |
1000 | void rxrpc_service_connection_reaper(struct work_struct *); |
1001 | void rxrpc_destroy_all_connections(struct rxrpc_net *); |
1002 | |
1003 | static inline bool rxrpc_conn_is_client(const struct rxrpc_connection *conn) |
1004 | { |
1005 | return conn->out_clientflag; |
1006 | } |
1007 | |
1008 | static inline bool rxrpc_conn_is_service(const struct rxrpc_connection *conn) |
1009 | { |
1010 | return !rxrpc_conn_is_client(conn); |
1011 | } |
1012 | |
1013 | static inline void rxrpc_reduce_conn_timer(struct rxrpc_connection *conn, |
1014 | unsigned long expire_at) |
1015 | { |
1016 | timer_reduce(timer: &conn->timer, expires: expire_at); |
1017 | } |
1018 | |
1019 | /* |
1020 | * conn_service.c |
1021 | */ |
1022 | struct rxrpc_connection *rxrpc_find_service_conn_rcu(struct rxrpc_peer *, |
1023 | struct sk_buff *); |
1024 | struct rxrpc_connection *rxrpc_prealloc_service_connection(struct rxrpc_net *, gfp_t); |
1025 | void rxrpc_new_incoming_connection(struct rxrpc_sock *, struct rxrpc_connection *, |
1026 | const struct rxrpc_security *, struct sk_buff *); |
1027 | void rxrpc_unpublish_service_conn(struct rxrpc_connection *); |
1028 | |
1029 | /* |
1030 | * input.c |
1031 | */ |
1032 | void rxrpc_congestion_degrade(struct rxrpc_call *); |
1033 | void rxrpc_input_call_packet(struct rxrpc_call *, struct sk_buff *); |
1034 | void rxrpc_implicit_end_call(struct rxrpc_call *, struct sk_buff *); |
1035 | |
1036 | /* |
1037 | * io_thread.c |
1038 | */ |
1039 | int rxrpc_encap_rcv(struct sock *, struct sk_buff *); |
1040 | void rxrpc_error_report(struct sock *); |
1041 | bool rxrpc_direct_abort(struct sk_buff *skb, enum rxrpc_abort_reason why, |
1042 | s32 abort_code, int err); |
1043 | int rxrpc_io_thread(void *data); |
1044 | static inline void rxrpc_wake_up_io_thread(struct rxrpc_local *local) |
1045 | { |
1046 | wake_up_process(tsk: local->io_thread); |
1047 | } |
1048 | |
1049 | static inline bool rxrpc_protocol_error(struct sk_buff *skb, enum rxrpc_abort_reason why) |
1050 | { |
1051 | return rxrpc_direct_abort(skb, why, RX_PROTOCOL_ERROR, err: -EPROTO); |
1052 | } |
1053 | |
1054 | /* |
1055 | * insecure.c |
1056 | */ |
1057 | extern const struct rxrpc_security rxrpc_no_security; |
1058 | |
1059 | /* |
1060 | * key.c |
1061 | */ |
1062 | extern struct key_type key_type_rxrpc; |
1063 | |
1064 | int rxrpc_request_key(struct rxrpc_sock *, sockptr_t , int); |
1065 | int rxrpc_get_server_data_key(struct rxrpc_connection *, const void *, time64_t, |
1066 | u32); |
1067 | |
1068 | /* |
1069 | * local_event.c |
1070 | */ |
1071 | void rxrpc_gen_version_string(void); |
1072 | void rxrpc_send_version_request(struct rxrpc_local *local, |
1073 | struct rxrpc_host_header *hdr, |
1074 | struct sk_buff *skb); |
1075 | |
1076 | /* |
1077 | * local_object.c |
1078 | */ |
1079 | struct rxrpc_local *rxrpc_lookup_local(struct net *, const struct sockaddr_rxrpc *); |
1080 | struct rxrpc_local *rxrpc_get_local(struct rxrpc_local *, enum rxrpc_local_trace); |
1081 | struct rxrpc_local *rxrpc_get_local_maybe(struct rxrpc_local *, enum rxrpc_local_trace); |
1082 | void rxrpc_put_local(struct rxrpc_local *, enum rxrpc_local_trace); |
1083 | struct rxrpc_local *rxrpc_use_local(struct rxrpc_local *, enum rxrpc_local_trace); |
1084 | void rxrpc_unuse_local(struct rxrpc_local *, enum rxrpc_local_trace); |
1085 | void rxrpc_destroy_local(struct rxrpc_local *local); |
1086 | void rxrpc_destroy_all_locals(struct rxrpc_net *); |
1087 | |
1088 | static inline bool __rxrpc_use_local(struct rxrpc_local *local, |
1089 | enum rxrpc_local_trace why) |
1090 | { |
1091 | int r, u; |
1092 | |
1093 | r = refcount_read(r: &local->ref); |
1094 | u = atomic_fetch_add_unless(v: &local->active_users, a: 1, u: 0); |
1095 | trace_rxrpc_local(local_debug_id: local->debug_id, op: why, ref: r, usage: u); |
1096 | return u != 0; |
1097 | } |
1098 | |
1099 | static inline void rxrpc_see_local(struct rxrpc_local *local, |
1100 | enum rxrpc_local_trace why) |
1101 | { |
1102 | int r, u; |
1103 | |
1104 | r = refcount_read(r: &local->ref); |
1105 | u = atomic_read(v: &local->active_users); |
1106 | trace_rxrpc_local(local_debug_id: local->debug_id, op: why, ref: r, usage: u); |
1107 | } |
1108 | |
1109 | /* |
1110 | * misc.c |
1111 | */ |
1112 | extern unsigned int rxrpc_max_backlog __read_mostly; |
1113 | extern unsigned long rxrpc_soft_ack_delay; |
1114 | extern unsigned long rxrpc_idle_ack_delay; |
1115 | extern unsigned int rxrpc_rx_window_size; |
1116 | extern unsigned int rxrpc_rx_mtu; |
1117 | extern unsigned int rxrpc_rx_jumbo_max; |
1118 | #ifdef CONFIG_AF_RXRPC_INJECT_RX_DELAY |
1119 | extern unsigned long rxrpc_inject_rx_delay; |
1120 | #endif |
1121 | |
1122 | /* |
1123 | * net_ns.c |
1124 | */ |
1125 | extern unsigned int rxrpc_net_id; |
1126 | extern struct pernet_operations rxrpc_net_ops; |
1127 | |
1128 | static inline struct rxrpc_net *rxrpc_net(struct net *net) |
1129 | { |
1130 | return net_generic(net, id: rxrpc_net_id); |
1131 | } |
1132 | |
1133 | /* |
1134 | * output.c |
1135 | */ |
1136 | int rxrpc_send_ack_packet(struct rxrpc_call *call, struct rxrpc_txbuf *txb); |
1137 | int rxrpc_send_abort_packet(struct rxrpc_call *); |
1138 | int rxrpc_send_data_packet(struct rxrpc_call *, struct rxrpc_txbuf *); |
1139 | void rxrpc_send_conn_abort(struct rxrpc_connection *conn); |
1140 | void rxrpc_reject_packet(struct rxrpc_local *local, struct sk_buff *skb); |
1141 | void rxrpc_send_keepalive(struct rxrpc_peer *); |
1142 | void rxrpc_transmit_one(struct rxrpc_call *call, struct rxrpc_txbuf *txb); |
1143 | |
1144 | /* |
1145 | * peer_event.c |
1146 | */ |
1147 | void rxrpc_input_error(struct rxrpc_local *, struct sk_buff *); |
1148 | void rxrpc_peer_keepalive_worker(struct work_struct *); |
1149 | |
1150 | /* |
1151 | * peer_object.c |
1152 | */ |
1153 | struct rxrpc_peer *rxrpc_lookup_peer_rcu(struct rxrpc_local *, |
1154 | const struct sockaddr_rxrpc *); |
1155 | struct rxrpc_peer *rxrpc_lookup_peer(struct rxrpc_local *local, |
1156 | struct sockaddr_rxrpc *srx, gfp_t gfp); |
1157 | struct rxrpc_peer *rxrpc_alloc_peer(struct rxrpc_local *, gfp_t, |
1158 | enum rxrpc_peer_trace); |
1159 | void rxrpc_new_incoming_peer(struct rxrpc_local *local, struct rxrpc_peer *peer); |
1160 | void rxrpc_destroy_all_peers(struct rxrpc_net *); |
1161 | struct rxrpc_peer *rxrpc_get_peer(struct rxrpc_peer *, enum rxrpc_peer_trace); |
1162 | struct rxrpc_peer *rxrpc_get_peer_maybe(struct rxrpc_peer *, enum rxrpc_peer_trace); |
1163 | void rxrpc_put_peer(struct rxrpc_peer *, enum rxrpc_peer_trace); |
1164 | |
1165 | /* |
1166 | * proc.c |
1167 | */ |
1168 | extern const struct seq_operations rxrpc_call_seq_ops; |
1169 | extern const struct seq_operations rxrpc_connection_seq_ops; |
1170 | extern const struct seq_operations rxrpc_peer_seq_ops; |
1171 | extern const struct seq_operations rxrpc_local_seq_ops; |
1172 | |
1173 | /* |
1174 | * recvmsg.c |
1175 | */ |
1176 | void rxrpc_notify_socket(struct rxrpc_call *); |
1177 | int rxrpc_recvmsg(struct socket *, struct msghdr *, size_t, int); |
1178 | |
1179 | /* |
1180 | * Abort a call due to a protocol error. |
1181 | */ |
1182 | static inline int rxrpc_abort_eproto(struct rxrpc_call *call, |
1183 | struct sk_buff *skb, |
1184 | s32 abort_code, |
1185 | enum rxrpc_abort_reason why) |
1186 | { |
1187 | struct rxrpc_skb_priv *sp = rxrpc_skb(skb); |
1188 | |
1189 | rxrpc_abort_call(call, seq: sp->hdr.seq, abort_code, error: -EPROTO, why); |
1190 | return -EPROTO; |
1191 | } |
1192 | |
1193 | /* |
1194 | * rtt.c |
1195 | */ |
1196 | void rxrpc_peer_add_rtt(struct rxrpc_call *, enum rxrpc_rtt_rx_trace, int, |
1197 | rxrpc_serial_t, rxrpc_serial_t, ktime_t, ktime_t); |
1198 | unsigned long rxrpc_get_rto_backoff(struct rxrpc_peer *, bool); |
1199 | void rxrpc_peer_init_rtt(struct rxrpc_peer *); |
1200 | |
1201 | /* |
1202 | * rxkad.c |
1203 | */ |
1204 | #ifdef CONFIG_RXKAD |
1205 | extern const struct rxrpc_security rxkad; |
1206 | #endif |
1207 | |
1208 | /* |
1209 | * security.c |
1210 | */ |
1211 | int __init rxrpc_init_security(void); |
1212 | const struct rxrpc_security *rxrpc_security_lookup(u8); |
1213 | void rxrpc_exit_security(void); |
1214 | int rxrpc_init_client_call_security(struct rxrpc_call *); |
1215 | int rxrpc_init_client_conn_security(struct rxrpc_connection *); |
1216 | const struct rxrpc_security *rxrpc_get_incoming_security(struct rxrpc_sock *, |
1217 | struct sk_buff *); |
1218 | struct key *rxrpc_look_up_server_security(struct rxrpc_connection *, |
1219 | struct sk_buff *, u32, u32); |
1220 | |
1221 | /* |
1222 | * sendmsg.c |
1223 | */ |
1224 | bool rxrpc_propose_abort(struct rxrpc_call *call, s32 abort_code, int error, |
1225 | enum rxrpc_abort_reason why); |
1226 | int rxrpc_do_sendmsg(struct rxrpc_sock *, struct msghdr *, size_t); |
1227 | |
1228 | /* |
1229 | * server_key.c |
1230 | */ |
1231 | extern struct key_type key_type_rxrpc_s; |
1232 | |
1233 | int rxrpc_server_keyring(struct rxrpc_sock *, sockptr_t, int); |
1234 | |
1235 | /* |
1236 | * skbuff.c |
1237 | */ |
1238 | void rxrpc_kernel_data_consumed(struct rxrpc_call *, struct sk_buff *); |
1239 | void rxrpc_new_skb(struct sk_buff *, enum rxrpc_skb_trace); |
1240 | void rxrpc_see_skb(struct sk_buff *, enum rxrpc_skb_trace); |
1241 | void rxrpc_eaten_skb(struct sk_buff *, enum rxrpc_skb_trace); |
1242 | void rxrpc_get_skb(struct sk_buff *, enum rxrpc_skb_trace); |
1243 | void rxrpc_free_skb(struct sk_buff *, enum rxrpc_skb_trace); |
1244 | void rxrpc_purge_queue(struct sk_buff_head *); |
1245 | |
1246 | /* |
1247 | * stats.c |
1248 | */ |
1249 | int rxrpc_stats_show(struct seq_file *seq, void *v); |
1250 | int rxrpc_stats_clear(struct file *file, char *buf, size_t size); |
1251 | |
1252 | #define rxrpc_inc_stat(rxnet, s) atomic_inc(&(rxnet)->s) |
1253 | #define rxrpc_dec_stat(rxnet, s) atomic_dec(&(rxnet)->s) |
1254 | |
1255 | /* |
1256 | * sysctl.c |
1257 | */ |
1258 | #ifdef CONFIG_SYSCTL |
1259 | extern int __init rxrpc_sysctl_init(void); |
1260 | extern void rxrpc_sysctl_exit(void); |
1261 | #else |
1262 | static inline int __init rxrpc_sysctl_init(void) { return 0; } |
1263 | static inline void rxrpc_sysctl_exit(void) {} |
1264 | #endif |
1265 | |
1266 | /* |
1267 | * txbuf.c |
1268 | */ |
1269 | extern atomic_t rxrpc_nr_txbuf; |
1270 | struct rxrpc_txbuf *rxrpc_alloc_txbuf(struct rxrpc_call *call, u8 packet_type, |
1271 | gfp_t gfp); |
1272 | void rxrpc_get_txbuf(struct rxrpc_txbuf *txb, enum rxrpc_txbuf_trace what); |
1273 | void rxrpc_see_txbuf(struct rxrpc_txbuf *txb, enum rxrpc_txbuf_trace what); |
1274 | void rxrpc_put_txbuf(struct rxrpc_txbuf *txb, enum rxrpc_txbuf_trace what); |
1275 | |
1276 | /* |
1277 | * utils.c |
1278 | */ |
1279 | int (struct sockaddr_rxrpc *, struct sk_buff *); |
1280 | |
1281 | static inline bool before(u32 seq1, u32 seq2) |
1282 | { |
1283 | return (s32)(seq1 - seq2) < 0; |
1284 | } |
1285 | static inline bool before_eq(u32 seq1, u32 seq2) |
1286 | { |
1287 | return (s32)(seq1 - seq2) <= 0; |
1288 | } |
1289 | static inline bool after(u32 seq1, u32 seq2) |
1290 | { |
1291 | return (s32)(seq1 - seq2) > 0; |
1292 | } |
1293 | static inline bool after_eq(u32 seq1, u32 seq2) |
1294 | { |
1295 | return (s32)(seq1 - seq2) >= 0; |
1296 | } |
1297 | |
1298 | /* |
1299 | * debug tracing |
1300 | */ |
1301 | extern unsigned int rxrpc_debug; |
1302 | |
1303 | #define dbgprintk(FMT,...) \ |
1304 | printk("[%-6.6s] "FMT"\n", current->comm ,##__VA_ARGS__) |
1305 | |
1306 | #define kenter(FMT,...) dbgprintk("==> %s("FMT")",__func__ ,##__VA_ARGS__) |
1307 | #define kleave(FMT,...) dbgprintk("<== %s()"FMT"",__func__ ,##__VA_ARGS__) |
1308 | #define kdebug(FMT,...) dbgprintk(" "FMT ,##__VA_ARGS__) |
1309 | |
1310 | |
1311 | #if defined(__KDEBUG) |
1312 | #define _enter(FMT,...) kenter(FMT,##__VA_ARGS__) |
1313 | #define _leave(FMT,...) kleave(FMT,##__VA_ARGS__) |
1314 | #define _debug(FMT,...) kdebug(FMT,##__VA_ARGS__) |
1315 | |
1316 | #elif defined(CONFIG_AF_RXRPC_DEBUG) |
1317 | #define RXRPC_DEBUG_KENTER 0x01 |
1318 | #define RXRPC_DEBUG_KLEAVE 0x02 |
1319 | #define RXRPC_DEBUG_KDEBUG 0x04 |
1320 | |
1321 | #define _enter(FMT,...) \ |
1322 | do { \ |
1323 | if (unlikely(rxrpc_debug & RXRPC_DEBUG_KENTER)) \ |
1324 | kenter(FMT,##__VA_ARGS__); \ |
1325 | } while (0) |
1326 | |
1327 | #define _leave(FMT,...) \ |
1328 | do { \ |
1329 | if (unlikely(rxrpc_debug & RXRPC_DEBUG_KLEAVE)) \ |
1330 | kleave(FMT,##__VA_ARGS__); \ |
1331 | } while (0) |
1332 | |
1333 | #define _debug(FMT,...) \ |
1334 | do { \ |
1335 | if (unlikely(rxrpc_debug & RXRPC_DEBUG_KDEBUG)) \ |
1336 | kdebug(FMT,##__VA_ARGS__); \ |
1337 | } while (0) |
1338 | |
1339 | #else |
1340 | #define _enter(FMT,...) no_printk("==> %s("FMT")",__func__ ,##__VA_ARGS__) |
1341 | #define _leave(FMT,...) no_printk("<== %s()"FMT"",__func__ ,##__VA_ARGS__) |
1342 | #define _debug(FMT,...) no_printk(" "FMT ,##__VA_ARGS__) |
1343 | #endif |
1344 | |
1345 | /* |
1346 | * debug assertion checking |
1347 | */ |
1348 | #if 1 // defined(__KDEBUGALL) |
1349 | |
1350 | #define ASSERT(X) \ |
1351 | do { \ |
1352 | if (unlikely(!(X))) { \ |
1353 | pr_err("Assertion failed\n"); \ |
1354 | BUG(); \ |
1355 | } \ |
1356 | } while (0) |
1357 | |
1358 | #define ASSERTCMP(X, OP, Y) \ |
1359 | do { \ |
1360 | __typeof__(X) _x = (X); \ |
1361 | __typeof__(Y) _y = (__typeof__(X))(Y); \ |
1362 | if (unlikely(!(_x OP _y))) { \ |
1363 | pr_err("Assertion failed - %lu(0x%lx) %s %lu(0x%lx) is false\n", \ |
1364 | (unsigned long)_x, (unsigned long)_x, #OP, \ |
1365 | (unsigned long)_y, (unsigned long)_y); \ |
1366 | BUG(); \ |
1367 | } \ |
1368 | } while (0) |
1369 | |
1370 | #define ASSERTIF(C, X) \ |
1371 | do { \ |
1372 | if (unlikely((C) && !(X))) { \ |
1373 | pr_err("Assertion failed\n"); \ |
1374 | BUG(); \ |
1375 | } \ |
1376 | } while (0) |
1377 | |
1378 | #define ASSERTIFCMP(C, X, OP, Y) \ |
1379 | do { \ |
1380 | __typeof__(X) _x = (X); \ |
1381 | __typeof__(Y) _y = (__typeof__(X))(Y); \ |
1382 | if (unlikely((C) && !(_x OP _y))) { \ |
1383 | pr_err("Assertion failed - %lu(0x%lx) %s %lu(0x%lx) is false\n", \ |
1384 | (unsigned long)_x, (unsigned long)_x, #OP, \ |
1385 | (unsigned long)_y, (unsigned long)_y); \ |
1386 | BUG(); \ |
1387 | } \ |
1388 | } while (0) |
1389 | |
1390 | #else |
1391 | |
1392 | #define ASSERT(X) \ |
1393 | do { \ |
1394 | } while (0) |
1395 | |
1396 | #define ASSERTCMP(X, OP, Y) \ |
1397 | do { \ |
1398 | } while (0) |
1399 | |
1400 | #define ASSERTIF(C, X) \ |
1401 | do { \ |
1402 | } while (0) |
1403 | |
1404 | #define ASSERTIFCMP(C, X, OP, Y) \ |
1405 | do { \ |
1406 | } while (0) |
1407 | |
1408 | #endif /* __KDEBUGALL */ |
1409 | |