1 | /* SPDX-License-Identifier: GPL-2.0-or-later */ |
2 | /* AF_RXRPC internal definitions |
3 | * |
4 | * Copyright (C) 2007 Red Hat, Inc. All Rights Reserved. |
5 | * Written by David Howells (dhowells@redhat.com) |
6 | */ |
7 | |
8 | #include <linux/atomic.h> |
9 | #include <linux/seqlock.h> |
10 | #include <linux/win_minmax.h> |
11 | #include <net/net_namespace.h> |
12 | #include <net/netns/generic.h> |
13 | #include <net/sock.h> |
14 | #include <net/af_rxrpc.h> |
15 | #include <keys/rxrpc-type.h> |
16 | #include "protocol.h" |
17 | |
18 | #define FCRYPT_BSIZE 8 |
19 | struct rxrpc_crypt { |
20 | union { |
21 | u8 x[FCRYPT_BSIZE]; |
22 | __be32 n[2]; |
23 | }; |
24 | } __attribute__((aligned(8))); |
25 | |
26 | #define rxrpc_queue_work(WS) queue_work(rxrpc_workqueue, (WS)) |
27 | #define rxrpc_queue_delayed_work(WS,D) \ |
28 | queue_delayed_work(rxrpc_workqueue, (WS), (D)) |
29 | |
30 | struct key_preparsed_payload; |
31 | struct rxrpc_connection; |
32 | struct rxrpc_txbuf; |
33 | |
34 | /* |
35 | * Mark applied to socket buffers in skb->mark. skb->priority is used |
36 | * to pass supplementary information. |
37 | */ |
38 | enum rxrpc_skb_mark { |
39 | RXRPC_SKB_MARK_PACKET, /* Received packet */ |
40 | RXRPC_SKB_MARK_ERROR, /* Error notification */ |
41 | RXRPC_SKB_MARK_SERVICE_CONN_SECURED, /* Service connection response has been verified */ |
42 | RXRPC_SKB_MARK_REJECT_BUSY, /* Reject with BUSY */ |
43 | RXRPC_SKB_MARK_REJECT_ABORT, /* Reject with ABORT (code in skb->priority) */ |
44 | }; |
45 | |
46 | /* |
47 | * sk_state for RxRPC sockets |
48 | */ |
49 | enum { |
50 | RXRPC_UNBOUND = 0, |
51 | RXRPC_CLIENT_UNBOUND, /* Unbound socket used as client */ |
52 | RXRPC_CLIENT_BOUND, /* client local address bound */ |
53 | RXRPC_SERVER_BOUND, /* server local address bound */ |
54 | RXRPC_SERVER_BOUND2, /* second server local address bound */ |
55 | RXRPC_SERVER_LISTENING, /* server listening for connections */ |
56 | RXRPC_SERVER_LISTEN_DISABLED, /* server listening disabled */ |
57 | RXRPC_CLOSE, /* socket is being closed */ |
58 | }; |
59 | |
60 | /* |
61 | * Per-network namespace data. |
62 | */ |
63 | struct rxrpc_net { |
64 | struct proc_dir_entry *proc_net; /* Subdir in /proc/net */ |
65 | u32 epoch; /* Local epoch for detecting local-end reset */ |
66 | struct list_head calls; /* List of calls active in this namespace */ |
67 | spinlock_t call_lock; /* Lock for ->calls */ |
68 | atomic_t nr_calls; /* Count of allocated calls */ |
69 | |
70 | atomic_t nr_conns; |
71 | struct list_head bundle_proc_list; /* List of bundles for proc */ |
72 | struct list_head conn_proc_list; /* List of conns in this namespace for proc */ |
73 | struct list_head service_conns; /* Service conns in this namespace */ |
74 | rwlock_t conn_lock; /* Lock for ->conn_proc_list, ->service_conns */ |
75 | struct work_struct service_conn_reaper; |
76 | struct timer_list service_conn_reap_timer; |
77 | |
78 | bool live; |
79 | |
80 | atomic_t nr_client_conns; |
81 | |
82 | struct hlist_head local_endpoints; |
83 | struct mutex local_mutex; /* Lock for ->local_endpoints */ |
84 | |
85 | DECLARE_HASHTABLE (peer_hash, 10); |
86 | spinlock_t peer_hash_lock; /* Lock for ->peer_hash */ |
87 | |
88 | #define RXRPC_KEEPALIVE_TIME 20 /* NAT keepalive time in seconds */ |
89 | u8 peer_keepalive_cursor; |
90 | time64_t peer_keepalive_base; |
91 | struct list_head peer_keepalive[32]; |
92 | struct list_head peer_keepalive_new; |
93 | struct timer_list peer_keepalive_timer; |
94 | struct work_struct peer_keepalive_work; |
95 | |
96 | atomic_t stat_tx_data; |
97 | atomic_t stat_tx_data_retrans; |
98 | atomic_t stat_tx_data_send; |
99 | atomic_t stat_tx_data_send_frag; |
100 | atomic_t stat_tx_data_send_fail; |
101 | atomic_t stat_tx_data_underflow; |
102 | atomic_t stat_tx_data_cwnd_reset; |
103 | atomic_t stat_rx_data; |
104 | atomic_t stat_rx_data_reqack; |
105 | atomic_t stat_rx_data_jumbo; |
106 | |
107 | atomic_t stat_tx_ack_fill; |
108 | atomic_t stat_tx_ack_send; |
109 | atomic_t stat_tx_ack_skip; |
110 | atomic_t stat_tx_acks[256]; |
111 | atomic_t stat_rx_acks[256]; |
112 | |
113 | atomic_t stat_why_req_ack[8]; |
114 | |
115 | atomic_t stat_io_loop; |
116 | }; |
117 | |
118 | /* |
119 | * Service backlog preallocation. |
120 | * |
121 | * This contains circular buffers of preallocated peers, connections and calls |
122 | * for incoming service calls and their head and tail pointers. This allows |
123 | * calls to be set up in the data_ready handler, thereby avoiding the need to |
124 | * shuffle packets around so much. |
125 | */ |
126 | struct rxrpc_backlog { |
127 | unsigned short peer_backlog_head; |
128 | unsigned short peer_backlog_tail; |
129 | unsigned short conn_backlog_head; |
130 | unsigned short conn_backlog_tail; |
131 | unsigned short call_backlog_head; |
132 | unsigned short call_backlog_tail; |
133 | #define RXRPC_BACKLOG_MAX 32 |
134 | struct rxrpc_peer *peer_backlog[RXRPC_BACKLOG_MAX]; |
135 | struct rxrpc_connection *conn_backlog[RXRPC_BACKLOG_MAX]; |
136 | struct rxrpc_call *call_backlog[RXRPC_BACKLOG_MAX]; |
137 | }; |
138 | |
139 | /* |
140 | * RxRPC socket definition |
141 | */ |
142 | struct rxrpc_sock { |
143 | /* WARNING: sk has to be the first member */ |
144 | struct sock sk; |
145 | rxrpc_notify_new_call_t notify_new_call; /* Func to notify of new call */ |
146 | rxrpc_discard_new_call_t discard_new_call; /* Func to discard a new call */ |
147 | struct rxrpc_local *local; /* local endpoint */ |
148 | struct rxrpc_backlog *backlog; /* Preallocation for services */ |
149 | spinlock_t incoming_lock; /* Incoming call vs service shutdown lock */ |
150 | struct list_head sock_calls; /* List of calls owned by this socket */ |
151 | struct list_head to_be_accepted; /* calls awaiting acceptance */ |
152 | struct list_head recvmsg_q; /* Calls awaiting recvmsg's attention */ |
153 | spinlock_t recvmsg_lock; /* Lock for recvmsg_q */ |
154 | struct key *key; /* security for this socket */ |
155 | struct key *securities; /* list of server security descriptors */ |
156 | struct rb_root calls; /* User ID -> call mapping */ |
157 | unsigned long flags; |
158 | #define RXRPC_SOCK_CONNECTED 0 /* connect_srx is set */ |
159 | rwlock_t call_lock; /* lock for calls */ |
160 | u32 min_sec_level; /* minimum security level */ |
161 | #define RXRPC_SECURITY_MAX RXRPC_SECURITY_ENCRYPT |
162 | bool exclusive; /* Exclusive connection for a client socket */ |
163 | u16 second_service; /* Additional service bound to the endpoint */ |
164 | struct { |
165 | /* Service upgrade information */ |
166 | u16 from; /* Service ID to upgrade (if not 0) */ |
167 | u16 to; /* service ID to upgrade to */ |
168 | } service_upgrade; |
169 | sa_family_t family; /* Protocol family created with */ |
170 | struct sockaddr_rxrpc srx; /* Primary Service/local addresses */ |
171 | struct sockaddr_rxrpc connect_srx; /* Default client address from connect() */ |
172 | }; |
173 | |
174 | #define rxrpc_sk(__sk) container_of((__sk), struct rxrpc_sock, sk) |
175 | |
176 | /* |
177 | * CPU-byteorder normalised Rx packet header. |
178 | */ |
179 | struct { |
180 | u32 ; /* client boot timestamp */ |
181 | u32 ; /* connection and channel ID */ |
182 | u32 ; /* call ID (0 for connection-level packets) */ |
183 | u32 ; /* sequence number of pkt in call stream */ |
184 | u32 ; /* serial number of pkt sent to network */ |
185 | u8 ; /* packet type */ |
186 | u8 ; /* packet flags */ |
187 | u8 ; /* app-layer defined status */ |
188 | u8 ; /* security protocol ID */ |
189 | union { |
190 | u16 ; /* reserved */ |
191 | u16 ; /* kerberos security checksum */ |
192 | }; |
193 | u16 ; /* service ID */ |
194 | } __packed; |
195 | |
196 | /* |
197 | * RxRPC socket buffer private variables |
198 | * - max 48 bytes (struct sk_buff::cb) |
199 | */ |
200 | struct rxrpc_skb_priv { |
201 | union { |
202 | struct rxrpc_connection *conn; /* Connection referred to (poke packet) */ |
203 | struct { |
204 | u16 offset; /* Offset of data */ |
205 | u16 len; /* Length of data */ |
206 | u8 flags; |
207 | #define RXRPC_RX_VERIFIED 0x01 |
208 | }; |
209 | struct { |
210 | rxrpc_seq_t first_ack; /* First packet in acks table */ |
211 | rxrpc_seq_t prev_ack; /* Highest seq seen */ |
212 | rxrpc_serial_t acked_serial; /* Packet in response to (or 0) */ |
213 | u8 reason; /* Reason for ack */ |
214 | u8 nr_acks; /* Number of acks+nacks */ |
215 | u8 nr_nacks; /* Number of nacks */ |
216 | } ack; |
217 | }; |
218 | struct rxrpc_host_header hdr; /* RxRPC packet header from this packet */ |
219 | }; |
220 | |
221 | #define rxrpc_skb(__skb) ((struct rxrpc_skb_priv *) &(__skb)->cb) |
222 | |
223 | /* |
224 | * RxRPC security module interface |
225 | */ |
226 | struct rxrpc_security { |
227 | const char *name; /* name of this service */ |
228 | u8 security_index; /* security type provided */ |
229 | u32 no_key_abort; /* Abort code indicating no key */ |
230 | |
231 | /* Initialise a security service */ |
232 | int (*init)(void); |
233 | |
234 | /* Clean up a security service */ |
235 | void (*exit)(void); |
236 | |
237 | /* Parse the information from a server key */ |
238 | int (*preparse_server_key)(struct key_preparsed_payload *); |
239 | |
240 | /* Clean up the preparse buffer after parsing a server key */ |
241 | void (*free_preparse_server_key)(struct key_preparsed_payload *); |
242 | |
243 | /* Destroy the payload of a server key */ |
244 | void (*destroy_server_key)(struct key *); |
245 | |
246 | /* Describe a server key */ |
247 | void (*describe_server_key)(const struct key *, struct seq_file *); |
248 | |
249 | /* initialise a connection's security */ |
250 | int (*init_connection_security)(struct rxrpc_connection *, |
251 | struct rxrpc_key_token *); |
252 | |
253 | /* Work out how much data we can store in a packet, given an estimate |
254 | * of the amount of data remaining and allocate a data buffer. |
255 | */ |
256 | struct rxrpc_txbuf *(*alloc_txbuf)(struct rxrpc_call *call, size_t remaining, gfp_t gfp); |
257 | |
258 | /* impose security on a packet */ |
259 | int (*secure_packet)(struct rxrpc_call *, struct rxrpc_txbuf *); |
260 | |
261 | /* verify the security on a received packet */ |
262 | int (*verify_packet)(struct rxrpc_call *, struct sk_buff *); |
263 | |
264 | /* Free crypto request on a call */ |
265 | void (*free_call_crypto)(struct rxrpc_call *); |
266 | |
267 | /* issue a challenge */ |
268 | int (*issue_challenge)(struct rxrpc_connection *); |
269 | |
270 | /* respond to a challenge */ |
271 | int (*respond_to_challenge)(struct rxrpc_connection *, |
272 | struct sk_buff *); |
273 | |
274 | /* verify a response */ |
275 | int (*verify_response)(struct rxrpc_connection *, |
276 | struct sk_buff *); |
277 | |
278 | /* clear connection security */ |
279 | void (*clear)(struct rxrpc_connection *); |
280 | }; |
281 | |
282 | /* |
283 | * RxRPC local transport endpoint description |
284 | * - owned by a single AF_RXRPC socket |
285 | * - pointed to by transport socket struct sk_user_data |
286 | */ |
287 | struct rxrpc_local { |
288 | struct rcu_head rcu; |
289 | atomic_t active_users; /* Number of users of the local endpoint */ |
290 | refcount_t ref; /* Number of references to the structure */ |
291 | struct net *net; /* The network namespace */ |
292 | struct rxrpc_net *rxnet; /* Our bits in the network namespace */ |
293 | struct hlist_node link; |
294 | struct socket *socket; /* my UDP socket */ |
295 | struct task_struct *io_thread; |
296 | struct completion io_thread_ready; /* Indication that the I/O thread started */ |
297 | struct page_frag_cache tx_alloc; /* Tx control packet allocation (I/O thread only) */ |
298 | struct rxrpc_sock *service; /* Service(s) listening on this endpoint */ |
299 | #ifdef CONFIG_AF_RXRPC_INJECT_RX_DELAY |
300 | struct sk_buff_head rx_delay_queue; /* Delay injection queue */ |
301 | #endif |
302 | struct sk_buff_head rx_queue; /* Received packets */ |
303 | struct list_head conn_attend_q; /* Conns requiring immediate attention */ |
304 | struct list_head call_attend_q; /* Calls requiring immediate attention */ |
305 | |
306 | struct rb_root client_bundles; /* Client connection bundles by socket params */ |
307 | spinlock_t client_bundles_lock; /* Lock for client_bundles */ |
308 | bool kill_all_client_conns; |
309 | struct list_head idle_client_conns; |
310 | struct timer_list client_conn_reap_timer; |
311 | unsigned long client_conn_flags; |
312 | #define RXRPC_CLIENT_CONN_REAP_TIMER 0 /* The client conn reap timer expired */ |
313 | |
314 | spinlock_t lock; /* access lock */ |
315 | rwlock_t services_lock; /* lock for services list */ |
316 | int debug_id; /* debug ID for printks */ |
317 | bool dead; |
318 | bool service_closed; /* Service socket closed */ |
319 | struct idr conn_ids; /* List of connection IDs */ |
320 | struct list_head new_client_calls; /* Newly created client calls need connection */ |
321 | spinlock_t client_call_lock; /* Lock for ->new_client_calls */ |
322 | struct sockaddr_rxrpc srx; /* local address */ |
323 | }; |
324 | |
325 | /* |
326 | * RxRPC remote transport endpoint definition |
327 | * - matched by local endpoint, remote port, address and protocol type |
328 | */ |
329 | struct rxrpc_peer { |
330 | struct rcu_head rcu; /* This must be first */ |
331 | refcount_t ref; |
332 | unsigned long hash_key; |
333 | struct hlist_node hash_link; |
334 | struct rxrpc_local *local; |
335 | struct hlist_head error_targets; /* targets for net error distribution */ |
336 | struct rb_root service_conns; /* Service connections */ |
337 | struct list_head keepalive_link; /* Link in net->peer_keepalive[] */ |
338 | time64_t last_tx_at; /* Last time packet sent here */ |
339 | seqlock_t service_conn_lock; |
340 | spinlock_t lock; /* access lock */ |
341 | unsigned int if_mtu; /* interface MTU for this peer */ |
342 | unsigned int mtu; /* network MTU for this peer */ |
343 | unsigned int maxdata; /* data size (MTU - hdrsize) */ |
344 | unsigned short hdrsize; /* header size (IP + UDP + RxRPC) */ |
345 | int debug_id; /* debug ID for printks */ |
346 | struct sockaddr_rxrpc srx; /* remote address */ |
347 | |
348 | /* calculated RTT cache */ |
349 | #define RXRPC_RTT_CACHE_SIZE 32 |
350 | spinlock_t rtt_input_lock; /* RTT lock for input routine */ |
351 | ktime_t rtt_last_req; /* Time of last RTT request */ |
352 | unsigned int rtt_count; /* Number of samples we've got */ |
353 | |
354 | u32 srtt_us; /* smoothed round trip time << 3 in usecs */ |
355 | u32 mdev_us; /* medium deviation */ |
356 | u32 mdev_max_us; /* maximal mdev for the last rtt period */ |
357 | u32 rttvar_us; /* smoothed mdev_max */ |
358 | u32 rto_us; /* Retransmission timeout in usec */ |
359 | u8 backoff; /* Backoff timeout (as shift) */ |
360 | |
361 | u8 cong_ssthresh; /* Congestion slow-start threshold */ |
362 | }; |
363 | |
364 | /* |
365 | * Keys for matching a connection. |
366 | */ |
367 | struct rxrpc_conn_proto { |
368 | union { |
369 | struct { |
370 | u32 epoch; /* epoch of this connection */ |
371 | u32 cid; /* connection ID */ |
372 | }; |
373 | u64 index_key; |
374 | }; |
375 | }; |
376 | |
377 | struct rxrpc_conn_parameters { |
378 | struct rxrpc_local *local; /* Representation of local endpoint */ |
379 | struct rxrpc_peer *peer; /* Representation of remote endpoint */ |
380 | struct key *key; /* Security details */ |
381 | bool exclusive; /* T if conn is exclusive */ |
382 | bool upgrade; /* T if service ID can be upgraded */ |
383 | u16 service_id; /* Service ID for this connection */ |
384 | u32 security_level; /* Security level selected */ |
385 | }; |
386 | |
387 | /* |
388 | * Call completion condition (state == RXRPC_CALL_COMPLETE). |
389 | */ |
390 | enum rxrpc_call_completion { |
391 | RXRPC_CALL_SUCCEEDED, /* - Normal termination */ |
392 | RXRPC_CALL_REMOTELY_ABORTED, /* - call aborted by peer */ |
393 | RXRPC_CALL_LOCALLY_ABORTED, /* - call aborted locally on error or close */ |
394 | RXRPC_CALL_LOCAL_ERROR, /* - call failed due to local error */ |
395 | RXRPC_CALL_NETWORK_ERROR, /* - call terminated by network error */ |
396 | NR__RXRPC_CALL_COMPLETIONS |
397 | }; |
398 | |
399 | /* |
400 | * Bits in the connection flags. |
401 | */ |
402 | enum rxrpc_conn_flag { |
403 | RXRPC_CONN_IN_SERVICE_CONNS, /* Conn is in peer->service_conns */ |
404 | RXRPC_CONN_DONT_REUSE, /* Don't reuse this connection */ |
405 | RXRPC_CONN_PROBING_FOR_UPGRADE, /* Probing for service upgrade */ |
406 | RXRPC_CONN_FINAL_ACK_0, /* Need final ACK for channel 0 */ |
407 | RXRPC_CONN_FINAL_ACK_1, /* Need final ACK for channel 1 */ |
408 | RXRPC_CONN_FINAL_ACK_2, /* Need final ACK for channel 2 */ |
409 | RXRPC_CONN_FINAL_ACK_3, /* Need final ACK for channel 3 */ |
410 | }; |
411 | |
412 | #define RXRPC_CONN_FINAL_ACK_MASK ((1UL << RXRPC_CONN_FINAL_ACK_0) | \ |
413 | (1UL << RXRPC_CONN_FINAL_ACK_1) | \ |
414 | (1UL << RXRPC_CONN_FINAL_ACK_2) | \ |
415 | (1UL << RXRPC_CONN_FINAL_ACK_3)) |
416 | |
417 | /* |
418 | * Events that can be raised upon a connection. |
419 | */ |
420 | enum rxrpc_conn_event { |
421 | RXRPC_CONN_EV_CHALLENGE, /* Send challenge packet */ |
422 | RXRPC_CONN_EV_ABORT_CALLS, /* Abort attached calls */ |
423 | }; |
424 | |
425 | /* |
426 | * The connection protocol state. |
427 | */ |
428 | enum rxrpc_conn_proto_state { |
429 | RXRPC_CONN_UNUSED, /* Connection not yet attempted */ |
430 | RXRPC_CONN_CLIENT_UNSECURED, /* Client connection needs security init */ |
431 | RXRPC_CONN_CLIENT, /* Client connection */ |
432 | RXRPC_CONN_SERVICE_PREALLOC, /* Service connection preallocation */ |
433 | RXRPC_CONN_SERVICE_UNSECURED, /* Service unsecured connection */ |
434 | RXRPC_CONN_SERVICE_CHALLENGING, /* Service challenging for security */ |
435 | RXRPC_CONN_SERVICE, /* Service secured connection */ |
436 | RXRPC_CONN_ABORTED, /* Conn aborted */ |
437 | RXRPC_CONN__NR_STATES |
438 | }; |
439 | |
440 | /* |
441 | * RxRPC client connection bundle. |
442 | */ |
443 | struct rxrpc_bundle { |
444 | struct rxrpc_local *local; /* Representation of local endpoint */ |
445 | struct rxrpc_peer *peer; /* Remote endpoint */ |
446 | struct key *key; /* Security details */ |
447 | struct list_head proc_link; /* Link in net->bundle_proc_list */ |
448 | const struct rxrpc_security *security; /* applied security module */ |
449 | refcount_t ref; |
450 | atomic_t active; /* Number of active users */ |
451 | unsigned int debug_id; |
452 | u32 security_level; /* Security level selected */ |
453 | u16 service_id; /* Service ID for this connection */ |
454 | bool try_upgrade; /* True if the bundle is attempting upgrade */ |
455 | bool exclusive; /* T if conn is exclusive */ |
456 | bool upgrade; /* T if service ID can be upgraded */ |
457 | unsigned short alloc_error; /* Error from last conn allocation */ |
458 | struct rb_node local_node; /* Node in local->client_conns */ |
459 | struct list_head waiting_calls; /* Calls waiting for channels */ |
460 | unsigned long avail_chans; /* Mask of available channels */ |
461 | unsigned int conn_ids[4]; /* Connection IDs. */ |
462 | struct rxrpc_connection *conns[4]; /* The connections in the bundle (max 4) */ |
463 | }; |
464 | |
465 | /* |
466 | * RxRPC connection definition |
467 | * - matched by { local, peer, epoch, conn_id, direction } |
468 | * - each connection can only handle four simultaneous calls |
469 | */ |
470 | struct rxrpc_connection { |
471 | struct rxrpc_conn_proto proto; |
472 | struct rxrpc_local *local; /* Representation of local endpoint */ |
473 | struct rxrpc_peer *peer; /* Remote endpoint */ |
474 | struct rxrpc_net *rxnet; /* Network namespace to which call belongs */ |
475 | struct key *key; /* Security details */ |
476 | struct list_head attend_link; /* Link in local->conn_attend_q */ |
477 | |
478 | refcount_t ref; |
479 | atomic_t active; /* Active count for service conns */ |
480 | struct rcu_head rcu; |
481 | struct list_head cache_link; |
482 | |
483 | unsigned char act_chans; /* Mask of active channels */ |
484 | struct rxrpc_channel { |
485 | unsigned long final_ack_at; /* Time at which to issue final ACK */ |
486 | struct rxrpc_call *call; /* Active call */ |
487 | unsigned int call_debug_id; /* call->debug_id */ |
488 | u32 call_id; /* ID of current call */ |
489 | u32 call_counter; /* Call ID counter */ |
490 | u32 last_call; /* ID of last call */ |
491 | u8 last_type; /* Type of last packet */ |
492 | union { |
493 | u32 last_seq; |
494 | u32 last_abort; |
495 | }; |
496 | } channels[RXRPC_MAXCALLS]; |
497 | |
498 | struct timer_list timer; /* Conn event timer */ |
499 | struct work_struct processor; /* connection event processor */ |
500 | struct work_struct destructor; /* In-process-context destroyer */ |
501 | struct rxrpc_bundle *bundle; /* Client connection bundle */ |
502 | struct rb_node service_node; /* Node in peer->service_conns */ |
503 | struct list_head proc_link; /* link in procfs list */ |
504 | struct list_head link; /* link in master connection list */ |
505 | struct sk_buff_head rx_queue; /* received conn-level packets */ |
506 | struct page_frag_cache tx_data_alloc; /* Tx DATA packet allocation */ |
507 | struct mutex tx_data_alloc_lock; |
508 | |
509 | struct mutex security_lock; /* Lock for security management */ |
510 | const struct rxrpc_security *security; /* applied security module */ |
511 | union { |
512 | struct { |
513 | struct crypto_sync_skcipher *cipher; /* encryption handle */ |
514 | struct rxrpc_crypt csum_iv; /* packet checksum base */ |
515 | u32 nonce; /* response re-use preventer */ |
516 | } rxkad; |
517 | }; |
518 | unsigned long flags; |
519 | unsigned long events; |
520 | unsigned long idle_timestamp; /* Time at which last became idle */ |
521 | spinlock_t state_lock; /* state-change lock */ |
522 | enum rxrpc_conn_proto_state state; /* current state of connection */ |
523 | enum rxrpc_call_completion completion; /* Completion condition */ |
524 | s32 abort_code; /* Abort code of connection abort */ |
525 | int debug_id; /* debug ID for printks */ |
526 | rxrpc_serial_t tx_serial; /* Outgoing packet serial number counter */ |
527 | unsigned int hi_serial; /* highest serial number received */ |
528 | u32 service_id; /* Service ID, possibly upgraded */ |
529 | u32 security_level; /* Security level selected */ |
530 | u8 security_ix; /* security type */ |
531 | u8 out_clientflag; /* RXRPC_CLIENT_INITIATED if we are client */ |
532 | u8 bundle_shift; /* Index into bundle->avail_chans */ |
533 | bool exclusive; /* T if conn is exclusive */ |
534 | bool upgrade; /* T if service ID can be upgraded */ |
535 | u16 orig_service_id; /* Originally requested service ID */ |
536 | short error; /* Local error code */ |
537 | }; |
538 | |
539 | static inline bool rxrpc_to_server(const struct rxrpc_skb_priv *sp) |
540 | { |
541 | return sp->hdr.flags & RXRPC_CLIENT_INITIATED; |
542 | } |
543 | |
544 | static inline bool rxrpc_to_client(const struct rxrpc_skb_priv *sp) |
545 | { |
546 | return !rxrpc_to_server(sp); |
547 | } |
548 | |
549 | /* |
550 | * Flags in call->flags. |
551 | */ |
552 | enum rxrpc_call_flag { |
553 | RXRPC_CALL_RELEASED, /* call has been released - no more message to userspace */ |
554 | RXRPC_CALL_HAS_USERID, /* has a user ID attached */ |
555 | RXRPC_CALL_IS_SERVICE, /* Call is service call */ |
556 | RXRPC_CALL_EXPOSED, /* The call was exposed to the world */ |
557 | RXRPC_CALL_RX_LAST, /* Received the last packet (at rxtx_top) */ |
558 | RXRPC_CALL_TX_LAST, /* Last packet in Tx buffer (at rxtx_top) */ |
559 | RXRPC_CALL_TX_ALL_ACKED, /* Last packet has been hard-acked */ |
560 | RXRPC_CALL_SEND_PING, /* A ping will need to be sent */ |
561 | RXRPC_CALL_RETRANS_TIMEOUT, /* Retransmission due to timeout occurred */ |
562 | RXRPC_CALL_BEGAN_RX_TIMER, /* We began the expect_rx_by timer */ |
563 | RXRPC_CALL_RX_HEARD, /* The peer responded at least once to this call */ |
564 | RXRPC_CALL_DISCONNECTED, /* The call has been disconnected */ |
565 | RXRPC_CALL_KERNEL, /* The call was made by the kernel */ |
566 | RXRPC_CALL_UPGRADE, /* Service upgrade was requested for the call */ |
567 | RXRPC_CALL_EXCLUSIVE, /* The call uses a once-only connection */ |
568 | RXRPC_CALL_RX_IS_IDLE, /* recvmsg() is idle - send an ACK */ |
569 | RXRPC_CALL_RECVMSG_READ_ALL, /* recvmsg() read all of the received data */ |
570 | }; |
571 | |
572 | /* |
573 | * Events that can be raised on a call. |
574 | */ |
575 | enum rxrpc_call_event { |
576 | RXRPC_CALL_EV_ACK_LOST, /* ACK may be lost, send ping */ |
577 | RXRPC_CALL_EV_INITIAL_PING, /* Send initial ping for a new service call */ |
578 | }; |
579 | |
580 | /* |
581 | * The states that a call can be in. |
582 | */ |
583 | enum rxrpc_call_state { |
584 | RXRPC_CALL_UNINITIALISED, |
585 | RXRPC_CALL_CLIENT_AWAIT_CONN, /* - client waiting for connection to become available */ |
586 | RXRPC_CALL_CLIENT_SEND_REQUEST, /* - client sending request phase */ |
587 | RXRPC_CALL_CLIENT_AWAIT_REPLY, /* - client awaiting reply */ |
588 | RXRPC_CALL_CLIENT_RECV_REPLY, /* - client receiving reply phase */ |
589 | RXRPC_CALL_SERVER_PREALLOC, /* - service preallocation */ |
590 | RXRPC_CALL_SERVER_SECURING, /* - server securing request connection */ |
591 | RXRPC_CALL_SERVER_RECV_REQUEST, /* - server receiving request */ |
592 | RXRPC_CALL_SERVER_ACK_REQUEST, /* - server pending ACK of request */ |
593 | RXRPC_CALL_SERVER_SEND_REPLY, /* - server sending reply */ |
594 | RXRPC_CALL_SERVER_AWAIT_ACK, /* - server awaiting final ACK */ |
595 | RXRPC_CALL_COMPLETE, /* - call complete */ |
596 | NR__RXRPC_CALL_STATES |
597 | }; |
598 | |
599 | /* |
600 | * Call Tx congestion management modes. |
601 | */ |
602 | enum rxrpc_congest_mode { |
603 | RXRPC_CALL_SLOW_START, |
604 | RXRPC_CALL_CONGEST_AVOIDANCE, |
605 | RXRPC_CALL_PACKET_LOSS, |
606 | RXRPC_CALL_FAST_RETRANSMIT, |
607 | NR__RXRPC_CONGEST_MODES |
608 | }; |
609 | |
610 | /* |
611 | * RxRPC call definition |
612 | * - matched by { connection, call_id } |
613 | */ |
614 | struct rxrpc_call { |
615 | struct rcu_head rcu; |
616 | struct rxrpc_connection *conn; /* connection carrying call */ |
617 | struct rxrpc_bundle *bundle; /* Connection bundle to use */ |
618 | struct rxrpc_peer *peer; /* Peer record for remote address */ |
619 | struct rxrpc_local *local; /* Representation of local endpoint */ |
620 | struct rxrpc_sock __rcu *socket; /* socket responsible */ |
621 | struct rxrpc_net *rxnet; /* Network namespace to which call belongs */ |
622 | struct key *key; /* Security details */ |
623 | const struct rxrpc_security *security; /* applied security module */ |
624 | struct mutex user_mutex; /* User access mutex */ |
625 | struct sockaddr_rxrpc dest_srx; /* Destination address */ |
626 | ktime_t delay_ack_at; /* When DELAY ACK needs to happen */ |
627 | ktime_t ack_lost_at; /* When ACK is figured as lost */ |
628 | ktime_t resend_at; /* When next resend needs to happen */ |
629 | ktime_t ping_at; /* When next to send a ping */ |
630 | ktime_t keepalive_at; /* When next to send a keepalive ping */ |
631 | ktime_t expect_rx_by; /* When we expect to get a packet by */ |
632 | ktime_t expect_req_by; /* When we expect to get a request DATA packet by */ |
633 | ktime_t expect_term_by; /* When we expect call termination by */ |
634 | u32 next_rx_timo; /* Timeout for next Rx packet (ms) */ |
635 | u32 next_req_timo; /* Timeout for next Rx request packet (ms) */ |
636 | u32 hard_timo; /* Maximum lifetime or 0 (s) */ |
637 | struct timer_list timer; /* Combined event timer */ |
638 | struct work_struct destroyer; /* In-process-context destroyer */ |
639 | rxrpc_notify_rx_t notify_rx; /* kernel service Rx notification function */ |
640 | struct list_head link; /* link in master call list */ |
641 | struct list_head wait_link; /* Link in local->new_client_calls */ |
642 | struct hlist_node error_link; /* link in error distribution list */ |
643 | struct list_head accept_link; /* Link in rx->acceptq */ |
644 | struct list_head recvmsg_link; /* Link in rx->recvmsg_q */ |
645 | struct list_head sock_link; /* Link in rx->sock_calls */ |
646 | struct rb_node sock_node; /* Node in rx->calls */ |
647 | struct list_head attend_link; /* Link in local->call_attend_q */ |
648 | struct rxrpc_txbuf *tx_pending; /* Tx buffer being filled */ |
649 | wait_queue_head_t waitq; /* Wait queue for channel or Tx */ |
650 | s64 tx_total_len; /* Total length left to be transmitted (or -1) */ |
651 | unsigned long user_call_ID; /* user-defined call ID */ |
652 | unsigned long flags; |
653 | unsigned long events; |
654 | spinlock_t notify_lock; /* Kernel notification lock */ |
655 | unsigned int send_abort_why; /* Why the abort [enum rxrpc_abort_reason] */ |
656 | s32 send_abort; /* Abort code to be sent */ |
657 | short send_abort_err; /* Error to be associated with the abort */ |
658 | rxrpc_seq_t send_abort_seq; /* DATA packet that incurred the abort (or 0) */ |
659 | s32 abort_code; /* Local/remote abort code */ |
660 | int error; /* Local error incurred */ |
661 | enum rxrpc_call_state _state; /* Current state of call (needs barrier) */ |
662 | enum rxrpc_call_completion completion; /* Call completion condition */ |
663 | refcount_t ref; |
664 | u8 security_ix; /* Security type */ |
665 | enum rxrpc_interruptibility interruptibility; /* At what point call may be interrupted */ |
666 | u32 call_id; /* call ID on connection */ |
667 | u32 cid; /* connection ID plus channel index */ |
668 | u32 security_level; /* Security level selected */ |
669 | int debug_id; /* debug ID for printks */ |
670 | unsigned short rx_pkt_offset; /* Current recvmsg packet offset */ |
671 | unsigned short rx_pkt_len; /* Current recvmsg packet len */ |
672 | |
673 | /* Transmitted data tracking. */ |
674 | spinlock_t tx_lock; /* Transmit queue lock */ |
675 | struct list_head tx_sendmsg; /* Sendmsg prepared packets */ |
676 | struct list_head tx_buffer; /* Buffer of transmissible packets */ |
677 | rxrpc_seq_t tx_bottom; /* First packet in buffer */ |
678 | rxrpc_seq_t tx_transmitted; /* Highest packet transmitted */ |
679 | rxrpc_seq_t tx_prepared; /* Highest Tx slot prepared. */ |
680 | rxrpc_seq_t tx_top; /* Highest Tx slot allocated. */ |
681 | u16 tx_backoff; /* Delay to insert due to Tx failure (ms) */ |
682 | u8 tx_winsize; /* Maximum size of Tx window */ |
683 | #define RXRPC_TX_MAX_WINDOW 128 |
684 | ktime_t tx_last_sent; /* Last time a transmission occurred */ |
685 | |
686 | /* Received data tracking */ |
687 | struct sk_buff_head recvmsg_queue; /* Queue of packets ready for recvmsg() */ |
688 | struct sk_buff_head rx_oos_queue; /* Queue of out of sequence packets */ |
689 | |
690 | rxrpc_seq_t rx_highest_seq; /* Higest sequence number received */ |
691 | rxrpc_seq_t rx_consumed; /* Highest packet consumed */ |
692 | rxrpc_serial_t rx_serial; /* Highest serial received for this call */ |
693 | u8 rx_winsize; /* Size of Rx window */ |
694 | |
695 | /* TCP-style slow-start congestion control [RFC5681]. Since the SMSS |
696 | * is fixed, we keep these numbers in terms of segments (ie. DATA |
697 | * packets) rather than bytes. |
698 | */ |
699 | #define RXRPC_TX_SMSS RXRPC_JUMBO_DATALEN |
700 | #define RXRPC_MIN_CWND (RXRPC_TX_SMSS > 2190 ? 2 : RXRPC_TX_SMSS > 1095 ? 3 : 4) |
701 | u8 cong_cwnd; /* Congestion window size */ |
702 | u8 ; /* Extra to send for congestion management */ |
703 | u8 cong_ssthresh; /* Slow-start threshold */ |
704 | enum rxrpc_congest_mode cong_mode:8; /* Congestion management mode */ |
705 | u8 cong_dup_acks; /* Count of ACKs showing missing packets */ |
706 | u8 cong_cumul_acks; /* Cumulative ACK count */ |
707 | ktime_t cong_tstamp; /* Last time cwnd was changed */ |
708 | struct sk_buff *cong_last_nack; /* Last ACK with nacks received */ |
709 | |
710 | /* Receive-phase ACK management (ACKs we send). */ |
711 | u8 ackr_reason; /* reason to ACK */ |
712 | u16 ackr_sack_base; /* Starting slot in SACK table ring */ |
713 | rxrpc_seq_t ackr_window; /* Base of SACK window */ |
714 | rxrpc_seq_t ackr_wtop; /* Base of SACK window */ |
715 | unsigned int ackr_nr_unacked; /* Number of unacked packets */ |
716 | atomic_t ackr_nr_consumed; /* Number of packets needing hard ACK */ |
717 | struct { |
718 | #define RXRPC_SACK_SIZE 256 |
719 | /* SACK table for soft-acked packets */ |
720 | u8 ackr_sack_table[RXRPC_SACK_SIZE]; |
721 | } __aligned(8); |
722 | |
723 | /* RTT management */ |
724 | rxrpc_serial_t rtt_serial[4]; /* Serial number of DATA or PING sent */ |
725 | ktime_t rtt_sent_at[4]; /* Time packet sent */ |
726 | unsigned long rtt_avail; /* Mask of available slots in bits 0-3, |
727 | * Mask of pending samples in 8-11 */ |
728 | #define RXRPC_CALL_RTT_AVAIL_MASK 0xf |
729 | #define RXRPC_CALL_RTT_PEND_SHIFT 8 |
730 | |
731 | /* Transmission-phase ACK management (ACKs we've received). */ |
732 | ktime_t acks_latest_ts; /* Timestamp of latest ACK received */ |
733 | rxrpc_seq_t acks_first_seq; /* first sequence number received */ |
734 | rxrpc_seq_t acks_prev_seq; /* Highest previousPacket received */ |
735 | rxrpc_seq_t acks_hard_ack; /* Latest hard-ack point */ |
736 | rxrpc_seq_t acks_lowest_nak; /* Lowest NACK in the buffer (or ==tx_hard_ack) */ |
737 | rxrpc_serial_t acks_highest_serial; /* Highest serial number ACK'd */ |
738 | }; |
739 | |
740 | /* |
741 | * Summary of a new ACK and the changes it made to the Tx buffer packet states. |
742 | */ |
743 | struct rxrpc_ack_summary { |
744 | u16 nr_acks; /* Number of ACKs in packet */ |
745 | u16 nr_new_acks; /* Number of new ACKs in packet */ |
746 | u16 nr_new_nacks; /* Number of new nacks in packet */ |
747 | u16 nr_retained_nacks; /* Number of nacks retained between ACKs */ |
748 | u8 ack_reason; |
749 | bool saw_nacks; /* Saw NACKs in packet */ |
750 | bool new_low_nack; /* T if new low NACK found */ |
751 | bool retrans_timeo; /* T if reTx due to timeout happened */ |
752 | u8 flight_size; /* Number of unreceived transmissions */ |
753 | /* Place to stash values for tracing */ |
754 | enum rxrpc_congest_mode mode:8; |
755 | u8 cwnd; |
756 | u8 ssthresh; |
757 | u8 dup_acks; |
758 | u8 cumulative_acks; |
759 | }; |
760 | |
761 | /* |
762 | * sendmsg() cmsg-specified parameters. |
763 | */ |
764 | enum rxrpc_command { |
765 | RXRPC_CMD_SEND_DATA, /* send data message */ |
766 | RXRPC_CMD_SEND_ABORT, /* request abort generation */ |
767 | RXRPC_CMD_REJECT_BUSY, /* [server] reject a call as busy */ |
768 | RXRPC_CMD_CHARGE_ACCEPT, /* [server] charge accept preallocation */ |
769 | }; |
770 | |
771 | struct rxrpc_call_params { |
772 | s64 tx_total_len; /* Total Tx data length (if send data) */ |
773 | unsigned long user_call_ID; /* User's call ID */ |
774 | struct { |
775 | u32 hard; /* Maximum lifetime (sec) */ |
776 | u32 idle; /* Max time since last data packet (msec) */ |
777 | u32 normal; /* Max time since last call packet (msec) */ |
778 | } timeouts; |
779 | u8 nr_timeouts; /* Number of timeouts specified */ |
780 | bool kernel; /* T if kernel is making the call */ |
781 | enum rxrpc_interruptibility interruptibility; /* How is interruptible is the call? */ |
782 | }; |
783 | |
784 | struct rxrpc_send_params { |
785 | struct rxrpc_call_params call; |
786 | u32 abort_code; /* Abort code to Tx (if abort) */ |
787 | enum rxrpc_command command : 8; /* The command to implement */ |
788 | bool exclusive; /* Shared or exclusive call */ |
789 | bool upgrade; /* If the connection is upgradeable */ |
790 | }; |
791 | |
792 | /* |
793 | * Buffer of data to be output as a packet. |
794 | */ |
795 | struct rxrpc_txbuf { |
796 | struct list_head call_link; /* Link in call->tx_sendmsg/tx_buffer */ |
797 | struct list_head tx_link; /* Link in live Enc queue or Tx queue */ |
798 | ktime_t last_sent; /* Time at which last transmitted */ |
799 | refcount_t ref; |
800 | rxrpc_seq_t seq; /* Sequence number of this packet */ |
801 | rxrpc_serial_t serial; /* Last serial number transmitted with */ |
802 | unsigned int call_debug_id; |
803 | unsigned int debug_id; |
804 | unsigned int len; /* Amount of data in buffer */ |
805 | unsigned int space; /* Remaining data space */ |
806 | unsigned int offset; /* Offset of fill point */ |
807 | unsigned int flags; |
808 | #define RXRPC_TXBUF_WIRE_FLAGS 0xff /* The wire protocol flags */ |
809 | #define RXRPC_TXBUF_RESENT 0x100 /* Set if has been resent */ |
810 | __be16 cksum; /* Checksum to go in header */ |
811 | unsigned short ack_rwind; /* ACK receive window */ |
812 | u8 /*enum rxrpc_propose_ack_trace*/ ack_why; /* If ack, why */ |
813 | u8 nr_kvec; /* Amount of kvec[] used */ |
814 | struct kvec kvec[3]; |
815 | }; |
816 | |
817 | static inline bool rxrpc_sending_to_server(const struct rxrpc_txbuf *txb) |
818 | { |
819 | return txb->flags & RXRPC_CLIENT_INITIATED; |
820 | } |
821 | |
822 | static inline bool rxrpc_sending_to_client(const struct rxrpc_txbuf *txb) |
823 | { |
824 | return !rxrpc_sending_to_server(txb); |
825 | } |
826 | |
827 | #include <trace/events/rxrpc.h> |
828 | |
829 | /* |
830 | * Allocate the next serial number on a connection. 0 must be skipped. |
831 | */ |
832 | static inline rxrpc_serial_t rxrpc_get_next_serial(struct rxrpc_connection *conn) |
833 | { |
834 | rxrpc_serial_t serial; |
835 | |
836 | serial = conn->tx_serial; |
837 | if (serial == 0) |
838 | serial = 1; |
839 | conn->tx_serial = serial + 1; |
840 | return serial; |
841 | } |
842 | |
843 | /* |
844 | * af_rxrpc.c |
845 | */ |
846 | extern atomic_t rxrpc_n_rx_skbs; |
847 | extern struct workqueue_struct *rxrpc_workqueue; |
848 | |
849 | /* |
850 | * call_accept.c |
851 | */ |
852 | int rxrpc_service_prealloc(struct rxrpc_sock *, gfp_t); |
853 | void rxrpc_discard_prealloc(struct rxrpc_sock *); |
854 | bool rxrpc_new_incoming_call(struct rxrpc_local *local, |
855 | struct rxrpc_peer *peer, |
856 | struct rxrpc_connection *conn, |
857 | struct sockaddr_rxrpc *peer_srx, |
858 | struct sk_buff *skb); |
859 | void rxrpc_accept_incoming_calls(struct rxrpc_local *); |
860 | int rxrpc_user_charge_accept(struct rxrpc_sock *, unsigned long); |
861 | |
862 | /* |
863 | * call_event.c |
864 | */ |
865 | void rxrpc_propose_ping(struct rxrpc_call *call, u32 serial, |
866 | enum rxrpc_propose_ack_trace why); |
867 | void rxrpc_propose_delay_ACK(struct rxrpc_call *, rxrpc_serial_t, |
868 | enum rxrpc_propose_ack_trace); |
869 | void rxrpc_shrink_call_tx_buffer(struct rxrpc_call *); |
870 | void rxrpc_resend(struct rxrpc_call *call, struct sk_buff *ack_skb); |
871 | |
872 | bool rxrpc_input_call_event(struct rxrpc_call *call, struct sk_buff *skb); |
873 | |
874 | /* |
875 | * call_object.c |
876 | */ |
877 | extern const char *const rxrpc_call_states[]; |
878 | extern const char *const rxrpc_call_completions[]; |
879 | extern struct kmem_cache *rxrpc_call_jar; |
880 | |
881 | void rxrpc_poke_call(struct rxrpc_call *call, enum rxrpc_call_poke_trace what); |
882 | struct rxrpc_call *rxrpc_find_call_by_user_ID(struct rxrpc_sock *, unsigned long); |
883 | struct rxrpc_call *rxrpc_alloc_call(struct rxrpc_sock *, gfp_t, unsigned int); |
884 | struct rxrpc_call *rxrpc_new_client_call(struct rxrpc_sock *, |
885 | struct rxrpc_conn_parameters *, |
886 | struct rxrpc_call_params *, gfp_t, |
887 | unsigned int); |
888 | void rxrpc_start_call_timer(struct rxrpc_call *call); |
889 | void rxrpc_incoming_call(struct rxrpc_sock *, struct rxrpc_call *, |
890 | struct sk_buff *); |
891 | void rxrpc_release_call(struct rxrpc_sock *, struct rxrpc_call *); |
892 | void rxrpc_release_calls_on_socket(struct rxrpc_sock *); |
893 | void rxrpc_see_call(struct rxrpc_call *, enum rxrpc_call_trace); |
894 | struct rxrpc_call *rxrpc_try_get_call(struct rxrpc_call *, enum rxrpc_call_trace); |
895 | void rxrpc_get_call(struct rxrpc_call *, enum rxrpc_call_trace); |
896 | void rxrpc_put_call(struct rxrpc_call *, enum rxrpc_call_trace); |
897 | void rxrpc_cleanup_call(struct rxrpc_call *); |
898 | void rxrpc_destroy_all_calls(struct rxrpc_net *); |
899 | |
900 | static inline bool rxrpc_is_service_call(const struct rxrpc_call *call) |
901 | { |
902 | return test_bit(RXRPC_CALL_IS_SERVICE, &call->flags); |
903 | } |
904 | |
905 | static inline bool rxrpc_is_client_call(const struct rxrpc_call *call) |
906 | { |
907 | return !rxrpc_is_service_call(call); |
908 | } |
909 | |
910 | /* |
911 | * call_state.c |
912 | */ |
913 | bool rxrpc_set_call_completion(struct rxrpc_call *call, |
914 | enum rxrpc_call_completion compl, |
915 | u32 abort_code, |
916 | int error); |
917 | bool rxrpc_call_completed(struct rxrpc_call *call); |
918 | bool rxrpc_abort_call(struct rxrpc_call *call, rxrpc_seq_t seq, |
919 | u32 abort_code, int error, enum rxrpc_abort_reason why); |
920 | void rxrpc_prefail_call(struct rxrpc_call *call, enum rxrpc_call_completion compl, |
921 | int error); |
922 | |
923 | static inline void rxrpc_set_call_state(struct rxrpc_call *call, |
924 | enum rxrpc_call_state state) |
925 | { |
926 | /* Order write of completion info before write of ->state. */ |
927 | smp_store_release(&call->_state, state); |
928 | wake_up(&call->waitq); |
929 | } |
930 | |
931 | static inline enum rxrpc_call_state __rxrpc_call_state(const struct rxrpc_call *call) |
932 | { |
933 | return call->_state; /* Only inside I/O thread */ |
934 | } |
935 | |
936 | static inline bool __rxrpc_call_is_complete(const struct rxrpc_call *call) |
937 | { |
938 | return __rxrpc_call_state(call) == RXRPC_CALL_COMPLETE; |
939 | } |
940 | |
941 | static inline enum rxrpc_call_state rxrpc_call_state(const struct rxrpc_call *call) |
942 | { |
943 | /* Order read ->state before read of completion info. */ |
944 | return smp_load_acquire(&call->_state); |
945 | } |
946 | |
947 | static inline bool rxrpc_call_is_complete(const struct rxrpc_call *call) |
948 | { |
949 | return rxrpc_call_state(call) == RXRPC_CALL_COMPLETE; |
950 | } |
951 | |
952 | static inline bool rxrpc_call_has_failed(const struct rxrpc_call *call) |
953 | { |
954 | return rxrpc_call_is_complete(call) && call->completion != RXRPC_CALL_SUCCEEDED; |
955 | } |
956 | |
957 | /* |
958 | * conn_client.c |
959 | */ |
960 | extern unsigned int rxrpc_reap_client_connections; |
961 | extern unsigned long rxrpc_conn_idle_client_expiry; |
962 | extern unsigned long rxrpc_conn_idle_client_fast_expiry; |
963 | |
964 | void rxrpc_purge_client_connections(struct rxrpc_local *local); |
965 | struct rxrpc_bundle *rxrpc_get_bundle(struct rxrpc_bundle *, enum rxrpc_bundle_trace); |
966 | void rxrpc_put_bundle(struct rxrpc_bundle *, enum rxrpc_bundle_trace); |
967 | int rxrpc_look_up_bundle(struct rxrpc_call *call, gfp_t gfp); |
968 | void rxrpc_connect_client_calls(struct rxrpc_local *local); |
969 | void rxrpc_expose_client_call(struct rxrpc_call *); |
970 | void rxrpc_disconnect_client_call(struct rxrpc_bundle *, struct rxrpc_call *); |
971 | void rxrpc_deactivate_bundle(struct rxrpc_bundle *bundle); |
972 | void rxrpc_put_client_conn(struct rxrpc_connection *, enum rxrpc_conn_trace); |
973 | void rxrpc_discard_expired_client_conns(struct rxrpc_local *local); |
974 | void rxrpc_clean_up_local_conns(struct rxrpc_local *); |
975 | |
976 | /* |
977 | * conn_event.c |
978 | */ |
979 | void rxrpc_conn_retransmit_call(struct rxrpc_connection *conn, struct sk_buff *skb, |
980 | unsigned int channel); |
981 | int rxrpc_abort_conn(struct rxrpc_connection *conn, struct sk_buff *skb, |
982 | s32 abort_code, int err, enum rxrpc_abort_reason why); |
983 | void rxrpc_process_connection(struct work_struct *); |
984 | void rxrpc_process_delayed_final_acks(struct rxrpc_connection *, bool); |
985 | bool rxrpc_input_conn_packet(struct rxrpc_connection *conn, struct sk_buff *skb); |
986 | void rxrpc_input_conn_event(struct rxrpc_connection *conn, struct sk_buff *skb); |
987 | |
988 | static inline bool rxrpc_is_conn_aborted(const struct rxrpc_connection *conn) |
989 | { |
990 | /* Order reading the abort info after the state check. */ |
991 | return smp_load_acquire(&conn->state) == RXRPC_CONN_ABORTED; |
992 | } |
993 | |
994 | /* |
995 | * conn_object.c |
996 | */ |
997 | extern unsigned int rxrpc_connection_expiry; |
998 | extern unsigned int rxrpc_closed_conn_expiry; |
999 | |
1000 | void rxrpc_poke_conn(struct rxrpc_connection *conn, enum rxrpc_conn_trace why); |
1001 | struct rxrpc_connection *rxrpc_alloc_connection(struct rxrpc_net *, gfp_t); |
1002 | struct rxrpc_connection *rxrpc_find_client_connection_rcu(struct rxrpc_local *, |
1003 | struct sockaddr_rxrpc *, |
1004 | struct sk_buff *); |
1005 | void __rxrpc_disconnect_call(struct rxrpc_connection *, struct rxrpc_call *); |
1006 | void rxrpc_disconnect_call(struct rxrpc_call *); |
1007 | void rxrpc_kill_client_conn(struct rxrpc_connection *); |
1008 | void rxrpc_queue_conn(struct rxrpc_connection *, enum rxrpc_conn_trace); |
1009 | void rxrpc_see_connection(struct rxrpc_connection *, enum rxrpc_conn_trace); |
1010 | struct rxrpc_connection *rxrpc_get_connection(struct rxrpc_connection *, |
1011 | enum rxrpc_conn_trace); |
1012 | struct rxrpc_connection *rxrpc_get_connection_maybe(struct rxrpc_connection *, |
1013 | enum rxrpc_conn_trace); |
1014 | void rxrpc_put_connection(struct rxrpc_connection *, enum rxrpc_conn_trace); |
1015 | void rxrpc_service_connection_reaper(struct work_struct *); |
1016 | void rxrpc_destroy_all_connections(struct rxrpc_net *); |
1017 | |
1018 | static inline bool rxrpc_conn_is_client(const struct rxrpc_connection *conn) |
1019 | { |
1020 | return conn->out_clientflag; |
1021 | } |
1022 | |
1023 | static inline bool rxrpc_conn_is_service(const struct rxrpc_connection *conn) |
1024 | { |
1025 | return !rxrpc_conn_is_client(conn); |
1026 | } |
1027 | |
1028 | static inline void rxrpc_reduce_conn_timer(struct rxrpc_connection *conn, |
1029 | unsigned long expire_at) |
1030 | { |
1031 | timer_reduce(timer: &conn->timer, expires: expire_at); |
1032 | } |
1033 | |
1034 | /* |
1035 | * conn_service.c |
1036 | */ |
1037 | struct rxrpc_connection *rxrpc_find_service_conn_rcu(struct rxrpc_peer *, |
1038 | struct sk_buff *); |
1039 | struct rxrpc_connection *rxrpc_prealloc_service_connection(struct rxrpc_net *, gfp_t); |
1040 | void rxrpc_new_incoming_connection(struct rxrpc_sock *, struct rxrpc_connection *, |
1041 | const struct rxrpc_security *, struct sk_buff *); |
1042 | void rxrpc_unpublish_service_conn(struct rxrpc_connection *); |
1043 | |
1044 | /* |
1045 | * input.c |
1046 | */ |
1047 | void rxrpc_congestion_degrade(struct rxrpc_call *); |
1048 | void rxrpc_input_call_packet(struct rxrpc_call *, struct sk_buff *); |
1049 | void rxrpc_implicit_end_call(struct rxrpc_call *, struct sk_buff *); |
1050 | |
1051 | /* |
1052 | * io_thread.c |
1053 | */ |
1054 | int rxrpc_encap_rcv(struct sock *, struct sk_buff *); |
1055 | void rxrpc_error_report(struct sock *); |
1056 | bool rxrpc_direct_abort(struct sk_buff *skb, enum rxrpc_abort_reason why, |
1057 | s32 abort_code, int err); |
1058 | int rxrpc_io_thread(void *data); |
1059 | static inline void rxrpc_wake_up_io_thread(struct rxrpc_local *local) |
1060 | { |
1061 | wake_up_process(tsk: local->io_thread); |
1062 | } |
1063 | |
1064 | static inline bool rxrpc_protocol_error(struct sk_buff *skb, enum rxrpc_abort_reason why) |
1065 | { |
1066 | return rxrpc_direct_abort(skb, why, RX_PROTOCOL_ERROR, err: -EPROTO); |
1067 | } |
1068 | |
1069 | /* |
1070 | * insecure.c |
1071 | */ |
1072 | extern const struct rxrpc_security rxrpc_no_security; |
1073 | |
1074 | /* |
1075 | * key.c |
1076 | */ |
1077 | extern struct key_type key_type_rxrpc; |
1078 | |
1079 | int rxrpc_request_key(struct rxrpc_sock *, sockptr_t , int); |
1080 | int rxrpc_get_server_data_key(struct rxrpc_connection *, const void *, time64_t, |
1081 | u32); |
1082 | |
1083 | /* |
1084 | * local_event.c |
1085 | */ |
1086 | void rxrpc_gen_version_string(void); |
1087 | void rxrpc_send_version_request(struct rxrpc_local *local, |
1088 | struct rxrpc_host_header *hdr, |
1089 | struct sk_buff *skb); |
1090 | |
1091 | /* |
1092 | * local_object.c |
1093 | */ |
1094 | void rxrpc_local_dont_fragment(const struct rxrpc_local *local, bool set); |
1095 | struct rxrpc_local *rxrpc_lookup_local(struct net *, const struct sockaddr_rxrpc *); |
1096 | struct rxrpc_local *rxrpc_get_local(struct rxrpc_local *, enum rxrpc_local_trace); |
1097 | struct rxrpc_local *rxrpc_get_local_maybe(struct rxrpc_local *, enum rxrpc_local_trace); |
1098 | void rxrpc_put_local(struct rxrpc_local *, enum rxrpc_local_trace); |
1099 | struct rxrpc_local *rxrpc_use_local(struct rxrpc_local *, enum rxrpc_local_trace); |
1100 | void rxrpc_unuse_local(struct rxrpc_local *, enum rxrpc_local_trace); |
1101 | void rxrpc_destroy_local(struct rxrpc_local *local); |
1102 | void rxrpc_destroy_all_locals(struct rxrpc_net *); |
1103 | |
1104 | static inline bool __rxrpc_use_local(struct rxrpc_local *local, |
1105 | enum rxrpc_local_trace why) |
1106 | { |
1107 | int r, u; |
1108 | |
1109 | r = refcount_read(r: &local->ref); |
1110 | u = atomic_fetch_add_unless(v: &local->active_users, a: 1, u: 0); |
1111 | trace_rxrpc_local(local_debug_id: local->debug_id, op: why, ref: r, usage: u); |
1112 | return u != 0; |
1113 | } |
1114 | |
1115 | static inline void rxrpc_see_local(struct rxrpc_local *local, |
1116 | enum rxrpc_local_trace why) |
1117 | { |
1118 | int r, u; |
1119 | |
1120 | r = refcount_read(r: &local->ref); |
1121 | u = atomic_read(v: &local->active_users); |
1122 | trace_rxrpc_local(local_debug_id: local->debug_id, op: why, ref: r, usage: u); |
1123 | } |
1124 | |
1125 | /* |
1126 | * misc.c |
1127 | */ |
1128 | extern unsigned int rxrpc_max_backlog __read_mostly; |
1129 | extern unsigned long rxrpc_soft_ack_delay; |
1130 | extern unsigned long rxrpc_idle_ack_delay; |
1131 | extern unsigned int rxrpc_rx_window_size; |
1132 | extern unsigned int rxrpc_rx_mtu; |
1133 | extern unsigned int rxrpc_rx_jumbo_max; |
1134 | #ifdef CONFIG_AF_RXRPC_INJECT_RX_DELAY |
1135 | extern unsigned long rxrpc_inject_rx_delay; |
1136 | #endif |
1137 | |
1138 | /* |
1139 | * net_ns.c |
1140 | */ |
1141 | extern unsigned int rxrpc_net_id; |
1142 | extern struct pernet_operations rxrpc_net_ops; |
1143 | |
1144 | static inline struct rxrpc_net *rxrpc_net(struct net *net) |
1145 | { |
1146 | return net_generic(net, id: rxrpc_net_id); |
1147 | } |
1148 | |
1149 | /* |
1150 | * output.c |
1151 | */ |
1152 | void rxrpc_send_ACK(struct rxrpc_call *call, u8 ack_reason, |
1153 | rxrpc_serial_t serial, enum rxrpc_propose_ack_trace why); |
1154 | int rxrpc_send_abort_packet(struct rxrpc_call *); |
1155 | void rxrpc_send_conn_abort(struct rxrpc_connection *conn); |
1156 | void rxrpc_reject_packet(struct rxrpc_local *local, struct sk_buff *skb); |
1157 | void rxrpc_send_keepalive(struct rxrpc_peer *); |
1158 | void rxrpc_transmit_one(struct rxrpc_call *call, struct rxrpc_txbuf *txb); |
1159 | |
1160 | /* |
1161 | * peer_event.c |
1162 | */ |
1163 | void rxrpc_input_error(struct rxrpc_local *, struct sk_buff *); |
1164 | void rxrpc_peer_keepalive_worker(struct work_struct *); |
1165 | |
1166 | /* |
1167 | * peer_object.c |
1168 | */ |
1169 | struct rxrpc_peer *rxrpc_lookup_peer_rcu(struct rxrpc_local *, |
1170 | const struct sockaddr_rxrpc *); |
1171 | struct rxrpc_peer *rxrpc_lookup_peer(struct rxrpc_local *local, |
1172 | struct sockaddr_rxrpc *srx, gfp_t gfp); |
1173 | struct rxrpc_peer *rxrpc_alloc_peer(struct rxrpc_local *, gfp_t, |
1174 | enum rxrpc_peer_trace); |
1175 | void rxrpc_new_incoming_peer(struct rxrpc_local *local, struct rxrpc_peer *peer); |
1176 | void rxrpc_destroy_all_peers(struct rxrpc_net *); |
1177 | struct rxrpc_peer *rxrpc_get_peer(struct rxrpc_peer *, enum rxrpc_peer_trace); |
1178 | struct rxrpc_peer *rxrpc_get_peer_maybe(struct rxrpc_peer *, enum rxrpc_peer_trace); |
1179 | void rxrpc_put_peer(struct rxrpc_peer *, enum rxrpc_peer_trace); |
1180 | |
1181 | /* |
1182 | * proc.c |
1183 | */ |
1184 | extern const struct seq_operations rxrpc_call_seq_ops; |
1185 | extern const struct seq_operations rxrpc_connection_seq_ops; |
1186 | extern const struct seq_operations rxrpc_bundle_seq_ops; |
1187 | extern const struct seq_operations rxrpc_peer_seq_ops; |
1188 | extern const struct seq_operations rxrpc_local_seq_ops; |
1189 | |
1190 | /* |
1191 | * recvmsg.c |
1192 | */ |
1193 | void rxrpc_notify_socket(struct rxrpc_call *); |
1194 | int rxrpc_recvmsg(struct socket *, struct msghdr *, size_t, int); |
1195 | |
1196 | /* |
1197 | * Abort a call due to a protocol error. |
1198 | */ |
1199 | static inline int rxrpc_abort_eproto(struct rxrpc_call *call, |
1200 | struct sk_buff *skb, |
1201 | s32 abort_code, |
1202 | enum rxrpc_abort_reason why) |
1203 | { |
1204 | struct rxrpc_skb_priv *sp = rxrpc_skb(skb); |
1205 | |
1206 | rxrpc_abort_call(call, seq: sp->hdr.seq, abort_code, error: -EPROTO, why); |
1207 | return -EPROTO; |
1208 | } |
1209 | |
1210 | /* |
1211 | * rtt.c |
1212 | */ |
1213 | void rxrpc_peer_add_rtt(struct rxrpc_call *, enum rxrpc_rtt_rx_trace, int, |
1214 | rxrpc_serial_t, rxrpc_serial_t, ktime_t, ktime_t); |
1215 | ktime_t rxrpc_get_rto_backoff(struct rxrpc_peer *peer, bool retrans); |
1216 | void rxrpc_peer_init_rtt(struct rxrpc_peer *); |
1217 | |
1218 | /* |
1219 | * rxkad.c |
1220 | */ |
1221 | #ifdef CONFIG_RXKAD |
1222 | extern const struct rxrpc_security rxkad; |
1223 | #endif |
1224 | |
1225 | /* |
1226 | * security.c |
1227 | */ |
1228 | int __init rxrpc_init_security(void); |
1229 | const struct rxrpc_security *rxrpc_security_lookup(u8); |
1230 | void rxrpc_exit_security(void); |
1231 | int rxrpc_init_client_call_security(struct rxrpc_call *); |
1232 | int rxrpc_init_client_conn_security(struct rxrpc_connection *); |
1233 | const struct rxrpc_security *rxrpc_get_incoming_security(struct rxrpc_sock *, |
1234 | struct sk_buff *); |
1235 | struct key *rxrpc_look_up_server_security(struct rxrpc_connection *, |
1236 | struct sk_buff *, u32, u32); |
1237 | |
1238 | /* |
1239 | * sendmsg.c |
1240 | */ |
1241 | bool rxrpc_propose_abort(struct rxrpc_call *call, s32 abort_code, int error, |
1242 | enum rxrpc_abort_reason why); |
1243 | int rxrpc_do_sendmsg(struct rxrpc_sock *, struct msghdr *, size_t); |
1244 | |
1245 | /* |
1246 | * server_key.c |
1247 | */ |
1248 | extern struct key_type key_type_rxrpc_s; |
1249 | |
1250 | int rxrpc_server_keyring(struct rxrpc_sock *, sockptr_t, int); |
1251 | |
1252 | /* |
1253 | * skbuff.c |
1254 | */ |
1255 | void rxrpc_kernel_data_consumed(struct rxrpc_call *, struct sk_buff *); |
1256 | void rxrpc_new_skb(struct sk_buff *, enum rxrpc_skb_trace); |
1257 | void rxrpc_see_skb(struct sk_buff *, enum rxrpc_skb_trace); |
1258 | void rxrpc_eaten_skb(struct sk_buff *, enum rxrpc_skb_trace); |
1259 | void rxrpc_get_skb(struct sk_buff *, enum rxrpc_skb_trace); |
1260 | void rxrpc_free_skb(struct sk_buff *, enum rxrpc_skb_trace); |
1261 | void rxrpc_purge_queue(struct sk_buff_head *); |
1262 | |
1263 | /* |
1264 | * stats.c |
1265 | */ |
1266 | int rxrpc_stats_show(struct seq_file *seq, void *v); |
1267 | int rxrpc_stats_clear(struct file *file, char *buf, size_t size); |
1268 | |
1269 | #define rxrpc_inc_stat(rxnet, s) atomic_inc(&(rxnet)->s) |
1270 | #define rxrpc_dec_stat(rxnet, s) atomic_dec(&(rxnet)->s) |
1271 | |
1272 | /* |
1273 | * sysctl.c |
1274 | */ |
1275 | #ifdef CONFIG_SYSCTL |
1276 | extern int __init rxrpc_sysctl_init(void); |
1277 | extern void rxrpc_sysctl_exit(void); |
1278 | #else |
1279 | static inline int __init rxrpc_sysctl_init(void) { return 0; } |
1280 | static inline void rxrpc_sysctl_exit(void) {} |
1281 | #endif |
1282 | |
1283 | /* |
1284 | * txbuf.c |
1285 | */ |
1286 | extern atomic_t rxrpc_nr_txbuf; |
1287 | struct rxrpc_txbuf *rxrpc_alloc_data_txbuf(struct rxrpc_call *call, size_t data_size, |
1288 | size_t data_align, gfp_t gfp); |
1289 | struct rxrpc_txbuf *rxrpc_alloc_ack_txbuf(struct rxrpc_call *call, size_t sack_size); |
1290 | void rxrpc_get_txbuf(struct rxrpc_txbuf *txb, enum rxrpc_txbuf_trace what); |
1291 | void rxrpc_see_txbuf(struct rxrpc_txbuf *txb, enum rxrpc_txbuf_trace what); |
1292 | void rxrpc_put_txbuf(struct rxrpc_txbuf *txb, enum rxrpc_txbuf_trace what); |
1293 | |
1294 | /* |
1295 | * utils.c |
1296 | */ |
1297 | int (struct sockaddr_rxrpc *, struct sk_buff *); |
1298 | |
1299 | static inline bool before(u32 seq1, u32 seq2) |
1300 | { |
1301 | return (s32)(seq1 - seq2) < 0; |
1302 | } |
1303 | static inline bool before_eq(u32 seq1, u32 seq2) |
1304 | { |
1305 | return (s32)(seq1 - seq2) <= 0; |
1306 | } |
1307 | static inline bool after(u32 seq1, u32 seq2) |
1308 | { |
1309 | return (s32)(seq1 - seq2) > 0; |
1310 | } |
1311 | static inline bool after_eq(u32 seq1, u32 seq2) |
1312 | { |
1313 | return (s32)(seq1 - seq2) >= 0; |
1314 | } |
1315 | |
1316 | /* |
1317 | * debug tracing |
1318 | */ |
1319 | extern unsigned int rxrpc_debug; |
1320 | |
1321 | #define dbgprintk(FMT,...) \ |
1322 | printk("[%-6.6s] "FMT"\n", current->comm ,##__VA_ARGS__) |
1323 | |
1324 | #define kenter(FMT,...) dbgprintk("==> %s("FMT")",__func__ ,##__VA_ARGS__) |
1325 | #define kleave(FMT,...) dbgprintk("<== %s()"FMT"",__func__ ,##__VA_ARGS__) |
1326 | #define kdebug(FMT,...) dbgprintk(" "FMT ,##__VA_ARGS__) |
1327 | |
1328 | |
1329 | #if defined(__KDEBUG) |
1330 | #define _enter(FMT,...) kenter(FMT,##__VA_ARGS__) |
1331 | #define _leave(FMT,...) kleave(FMT,##__VA_ARGS__) |
1332 | #define _debug(FMT,...) kdebug(FMT,##__VA_ARGS__) |
1333 | |
1334 | #elif defined(CONFIG_AF_RXRPC_DEBUG) |
1335 | #define RXRPC_DEBUG_KENTER 0x01 |
1336 | #define RXRPC_DEBUG_KLEAVE 0x02 |
1337 | #define RXRPC_DEBUG_KDEBUG 0x04 |
1338 | |
1339 | #define _enter(FMT,...) \ |
1340 | do { \ |
1341 | if (unlikely(rxrpc_debug & RXRPC_DEBUG_KENTER)) \ |
1342 | kenter(FMT,##__VA_ARGS__); \ |
1343 | } while (0) |
1344 | |
1345 | #define _leave(FMT,...) \ |
1346 | do { \ |
1347 | if (unlikely(rxrpc_debug & RXRPC_DEBUG_KLEAVE)) \ |
1348 | kleave(FMT,##__VA_ARGS__); \ |
1349 | } while (0) |
1350 | |
1351 | #define _debug(FMT,...) \ |
1352 | do { \ |
1353 | if (unlikely(rxrpc_debug & RXRPC_DEBUG_KDEBUG)) \ |
1354 | kdebug(FMT,##__VA_ARGS__); \ |
1355 | } while (0) |
1356 | |
1357 | #else |
1358 | #define _enter(FMT,...) no_printk("==> %s("FMT")",__func__ ,##__VA_ARGS__) |
1359 | #define _leave(FMT,...) no_printk("<== %s()"FMT"",__func__ ,##__VA_ARGS__) |
1360 | #define _debug(FMT,...) no_printk(" "FMT ,##__VA_ARGS__) |
1361 | #endif |
1362 | |
1363 | /* |
1364 | * debug assertion checking |
1365 | */ |
1366 | #if 1 // defined(__KDEBUGALL) |
1367 | |
1368 | #define ASSERT(X) \ |
1369 | do { \ |
1370 | if (unlikely(!(X))) { \ |
1371 | pr_err("Assertion failed\n"); \ |
1372 | BUG(); \ |
1373 | } \ |
1374 | } while (0) |
1375 | |
1376 | #define ASSERTCMP(X, OP, Y) \ |
1377 | do { \ |
1378 | __typeof__(X) _x = (X); \ |
1379 | __typeof__(Y) _y = (__typeof__(X))(Y); \ |
1380 | if (unlikely(!(_x OP _y))) { \ |
1381 | pr_err("Assertion failed - %lu(0x%lx) %s %lu(0x%lx) is false\n", \ |
1382 | (unsigned long)_x, (unsigned long)_x, #OP, \ |
1383 | (unsigned long)_y, (unsigned long)_y); \ |
1384 | BUG(); \ |
1385 | } \ |
1386 | } while (0) |
1387 | |
1388 | #define ASSERTIF(C, X) \ |
1389 | do { \ |
1390 | if (unlikely((C) && !(X))) { \ |
1391 | pr_err("Assertion failed\n"); \ |
1392 | BUG(); \ |
1393 | } \ |
1394 | } while (0) |
1395 | |
1396 | #define ASSERTIFCMP(C, X, OP, Y) \ |
1397 | do { \ |
1398 | __typeof__(X) _x = (X); \ |
1399 | __typeof__(Y) _y = (__typeof__(X))(Y); \ |
1400 | if (unlikely((C) && !(_x OP _y))) { \ |
1401 | pr_err("Assertion failed - %lu(0x%lx) %s %lu(0x%lx) is false\n", \ |
1402 | (unsigned long)_x, (unsigned long)_x, #OP, \ |
1403 | (unsigned long)_y, (unsigned long)_y); \ |
1404 | BUG(); \ |
1405 | } \ |
1406 | } while (0) |
1407 | |
1408 | #else |
1409 | |
1410 | #define ASSERT(X) \ |
1411 | do { \ |
1412 | } while (0) |
1413 | |
1414 | #define ASSERTCMP(X, OP, Y) \ |
1415 | do { \ |
1416 | } while (0) |
1417 | |
1418 | #define ASSERTIF(C, X) \ |
1419 | do { \ |
1420 | } while (0) |
1421 | |
1422 | #define ASSERTIFCMP(C, X, OP, Y) \ |
1423 | do { \ |
1424 | } while (0) |
1425 | |
1426 | #endif /* __KDEBUGALL */ |
1427 | |