1 | // SPDX-License-Identifier: GPL-2.0-only |
2 | /* xfrm_user.c: User interface to configure xfrm engine. |
3 | * |
4 | * Copyright (C) 2002 David S. Miller (davem@redhat.com) |
5 | * |
6 | * Changes: |
7 | * Mitsuru KANDA @USAGI |
8 | * Kazunori MIYAZAWA @USAGI |
9 | * Kunihiro Ishiguro <kunihiro@ipinfusion.com> |
10 | * IPv6 support |
11 | * |
12 | */ |
13 | |
14 | #include <linux/compat.h> |
15 | #include <linux/crypto.h> |
16 | #include <linux/module.h> |
17 | #include <linux/kernel.h> |
18 | #include <linux/types.h> |
19 | #include <linux/slab.h> |
20 | #include <linux/socket.h> |
21 | #include <linux/string.h> |
22 | #include <linux/net.h> |
23 | #include <linux/skbuff.h> |
24 | #include <linux/pfkeyv2.h> |
25 | #include <linux/ipsec.h> |
26 | #include <linux/init.h> |
27 | #include <linux/security.h> |
28 | #include <net/sock.h> |
29 | #include <net/xfrm.h> |
30 | #include <net/netlink.h> |
31 | #include <net/ah.h> |
32 | #include <linux/uaccess.h> |
33 | #if IS_ENABLED(CONFIG_IPV6) |
34 | #include <linux/in6.h> |
35 | #endif |
36 | #include <asm/unaligned.h> |
37 | |
38 | static int verify_one_alg(struct nlattr **attrs, enum xfrm_attr_type_t type, |
39 | struct netlink_ext_ack *extack) |
40 | { |
41 | struct nlattr *rt = attrs[type]; |
42 | struct xfrm_algo *algp; |
43 | |
44 | if (!rt) |
45 | return 0; |
46 | |
47 | algp = nla_data(nla: rt); |
48 | if (nla_len(nla: rt) < (int)xfrm_alg_len(alg: algp)) { |
49 | NL_SET_ERR_MSG(extack, "Invalid AUTH/CRYPT/COMP attribute length" ); |
50 | return -EINVAL; |
51 | } |
52 | |
53 | switch (type) { |
54 | case XFRMA_ALG_AUTH: |
55 | case XFRMA_ALG_CRYPT: |
56 | case XFRMA_ALG_COMP: |
57 | break; |
58 | |
59 | default: |
60 | NL_SET_ERR_MSG(extack, "Invalid algorithm attribute type" ); |
61 | return -EINVAL; |
62 | } |
63 | |
64 | algp->alg_name[sizeof(algp->alg_name) - 1] = '\0'; |
65 | return 0; |
66 | } |
67 | |
68 | static int verify_auth_trunc(struct nlattr **attrs, |
69 | struct netlink_ext_ack *extack) |
70 | { |
71 | struct nlattr *rt = attrs[XFRMA_ALG_AUTH_TRUNC]; |
72 | struct xfrm_algo_auth *algp; |
73 | |
74 | if (!rt) |
75 | return 0; |
76 | |
77 | algp = nla_data(nla: rt); |
78 | if (nla_len(nla: rt) < (int)xfrm_alg_auth_len(alg: algp)) { |
79 | NL_SET_ERR_MSG(extack, "Invalid AUTH_TRUNC attribute length" ); |
80 | return -EINVAL; |
81 | } |
82 | |
83 | algp->alg_name[sizeof(algp->alg_name) - 1] = '\0'; |
84 | return 0; |
85 | } |
86 | |
87 | static int verify_aead(struct nlattr **attrs, struct netlink_ext_ack *extack) |
88 | { |
89 | struct nlattr *rt = attrs[XFRMA_ALG_AEAD]; |
90 | struct xfrm_algo_aead *algp; |
91 | |
92 | if (!rt) |
93 | return 0; |
94 | |
95 | algp = nla_data(nla: rt); |
96 | if (nla_len(nla: rt) < (int)aead_len(alg: algp)) { |
97 | NL_SET_ERR_MSG(extack, "Invalid AEAD attribute length" ); |
98 | return -EINVAL; |
99 | } |
100 | |
101 | algp->alg_name[sizeof(algp->alg_name) - 1] = '\0'; |
102 | return 0; |
103 | } |
104 | |
105 | static void verify_one_addr(struct nlattr **attrs, enum xfrm_attr_type_t type, |
106 | xfrm_address_t **addrp) |
107 | { |
108 | struct nlattr *rt = attrs[type]; |
109 | |
110 | if (rt && addrp) |
111 | *addrp = nla_data(nla: rt); |
112 | } |
113 | |
114 | static inline int verify_sec_ctx_len(struct nlattr **attrs, struct netlink_ext_ack *extack) |
115 | { |
116 | struct nlattr *rt = attrs[XFRMA_SEC_CTX]; |
117 | struct xfrm_user_sec_ctx *uctx; |
118 | |
119 | if (!rt) |
120 | return 0; |
121 | |
122 | uctx = nla_data(nla: rt); |
123 | if (uctx->len > nla_len(nla: rt) || |
124 | uctx->len != (sizeof(struct xfrm_user_sec_ctx) + uctx->ctx_len)) { |
125 | NL_SET_ERR_MSG(extack, "Invalid security context length" ); |
126 | return -EINVAL; |
127 | } |
128 | |
129 | return 0; |
130 | } |
131 | |
132 | static inline int verify_replay(struct xfrm_usersa_info *p, |
133 | struct nlattr **attrs, |
134 | struct netlink_ext_ack *extack) |
135 | { |
136 | struct nlattr *rt = attrs[XFRMA_REPLAY_ESN_VAL]; |
137 | struct xfrm_replay_state_esn *rs; |
138 | |
139 | if (!rt) { |
140 | if (p->flags & XFRM_STATE_ESN) { |
141 | NL_SET_ERR_MSG(extack, "Missing required attribute for ESN" ); |
142 | return -EINVAL; |
143 | } |
144 | return 0; |
145 | } |
146 | |
147 | rs = nla_data(nla: rt); |
148 | |
149 | if (rs->bmp_len > XFRMA_REPLAY_ESN_MAX / sizeof(rs->bmp[0]) / 8) { |
150 | NL_SET_ERR_MSG(extack, "ESN bitmap length must be <= 128" ); |
151 | return -EINVAL; |
152 | } |
153 | |
154 | if (nla_len(nla: rt) < (int)xfrm_replay_state_esn_len(replay_esn: rs) && |
155 | nla_len(nla: rt) != sizeof(*rs)) { |
156 | NL_SET_ERR_MSG(extack, "ESN attribute is too short to fit the full bitmap length" ); |
157 | return -EINVAL; |
158 | } |
159 | |
160 | /* As only ESP and AH support ESN feature. */ |
161 | if ((p->id.proto != IPPROTO_ESP) && (p->id.proto != IPPROTO_AH)) { |
162 | NL_SET_ERR_MSG(extack, "ESN only supported for ESP and AH" ); |
163 | return -EINVAL; |
164 | } |
165 | |
166 | if (p->replay_window != 0) { |
167 | NL_SET_ERR_MSG(extack, "ESN not compatible with legacy replay_window" ); |
168 | return -EINVAL; |
169 | } |
170 | |
171 | return 0; |
172 | } |
173 | |
174 | static int verify_newsa_info(struct xfrm_usersa_info *p, |
175 | struct nlattr **attrs, |
176 | struct netlink_ext_ack *extack) |
177 | { |
178 | int err; |
179 | |
180 | err = -EINVAL; |
181 | switch (p->family) { |
182 | case AF_INET: |
183 | break; |
184 | |
185 | case AF_INET6: |
186 | #if IS_ENABLED(CONFIG_IPV6) |
187 | break; |
188 | #else |
189 | err = -EAFNOSUPPORT; |
190 | NL_SET_ERR_MSG(extack, "IPv6 support disabled" ); |
191 | goto out; |
192 | #endif |
193 | |
194 | default: |
195 | NL_SET_ERR_MSG(extack, "Invalid address family" ); |
196 | goto out; |
197 | } |
198 | |
199 | switch (p->sel.family) { |
200 | case AF_UNSPEC: |
201 | break; |
202 | |
203 | case AF_INET: |
204 | if (p->sel.prefixlen_d > 32 || p->sel.prefixlen_s > 32) { |
205 | NL_SET_ERR_MSG(extack, "Invalid prefix length in selector (must be <= 32 for IPv4)" ); |
206 | goto out; |
207 | } |
208 | |
209 | break; |
210 | |
211 | case AF_INET6: |
212 | #if IS_ENABLED(CONFIG_IPV6) |
213 | if (p->sel.prefixlen_d > 128 || p->sel.prefixlen_s > 128) { |
214 | NL_SET_ERR_MSG(extack, "Invalid prefix length in selector (must be <= 128 for IPv6)" ); |
215 | goto out; |
216 | } |
217 | |
218 | break; |
219 | #else |
220 | NL_SET_ERR_MSG(extack, "IPv6 support disabled" ); |
221 | err = -EAFNOSUPPORT; |
222 | goto out; |
223 | #endif |
224 | |
225 | default: |
226 | NL_SET_ERR_MSG(extack, "Invalid address family in selector" ); |
227 | goto out; |
228 | } |
229 | |
230 | err = -EINVAL; |
231 | switch (p->id.proto) { |
232 | case IPPROTO_AH: |
233 | if (!attrs[XFRMA_ALG_AUTH] && |
234 | !attrs[XFRMA_ALG_AUTH_TRUNC]) { |
235 | NL_SET_ERR_MSG(extack, "Missing required attribute for AH: AUTH_TRUNC or AUTH" ); |
236 | goto out; |
237 | } |
238 | |
239 | if (attrs[XFRMA_ALG_AEAD] || |
240 | attrs[XFRMA_ALG_CRYPT] || |
241 | attrs[XFRMA_ALG_COMP] || |
242 | attrs[XFRMA_TFCPAD]) { |
243 | NL_SET_ERR_MSG(extack, "Invalid attributes for AH: AEAD, CRYPT, COMP, TFCPAD" ); |
244 | goto out; |
245 | } |
246 | break; |
247 | |
248 | case IPPROTO_ESP: |
249 | if (attrs[XFRMA_ALG_COMP]) { |
250 | NL_SET_ERR_MSG(extack, "Invalid attribute for ESP: COMP" ); |
251 | goto out; |
252 | } |
253 | |
254 | if (!attrs[XFRMA_ALG_AUTH] && |
255 | !attrs[XFRMA_ALG_AUTH_TRUNC] && |
256 | !attrs[XFRMA_ALG_CRYPT] && |
257 | !attrs[XFRMA_ALG_AEAD]) { |
258 | NL_SET_ERR_MSG(extack, "Missing required attribute for ESP: at least one of AUTH, AUTH_TRUNC, CRYPT, AEAD" ); |
259 | goto out; |
260 | } |
261 | |
262 | if ((attrs[XFRMA_ALG_AUTH] || |
263 | attrs[XFRMA_ALG_AUTH_TRUNC] || |
264 | attrs[XFRMA_ALG_CRYPT]) && |
265 | attrs[XFRMA_ALG_AEAD]) { |
266 | NL_SET_ERR_MSG(extack, "Invalid attribute combination for ESP: AEAD can't be used with AUTH, AUTH_TRUNC, CRYPT" ); |
267 | goto out; |
268 | } |
269 | |
270 | if (attrs[XFRMA_TFCPAD] && |
271 | p->mode != XFRM_MODE_TUNNEL) { |
272 | NL_SET_ERR_MSG(extack, "TFC padding can only be used in tunnel mode" ); |
273 | goto out; |
274 | } |
275 | break; |
276 | |
277 | case IPPROTO_COMP: |
278 | if (!attrs[XFRMA_ALG_COMP]) { |
279 | NL_SET_ERR_MSG(extack, "Missing required attribute for COMP: COMP" ); |
280 | goto out; |
281 | } |
282 | |
283 | if (attrs[XFRMA_ALG_AEAD] || |
284 | attrs[XFRMA_ALG_AUTH] || |
285 | attrs[XFRMA_ALG_AUTH_TRUNC] || |
286 | attrs[XFRMA_ALG_CRYPT] || |
287 | attrs[XFRMA_TFCPAD]) { |
288 | NL_SET_ERR_MSG(extack, "Invalid attributes for COMP: AEAD, AUTH, AUTH_TRUNC, CRYPT, TFCPAD" ); |
289 | goto out; |
290 | } |
291 | |
292 | if (ntohl(p->id.spi) >= 0x10000) { |
293 | NL_SET_ERR_MSG(extack, "SPI is too large for COMP (must be < 0x10000)" ); |
294 | goto out; |
295 | } |
296 | break; |
297 | |
298 | #if IS_ENABLED(CONFIG_IPV6) |
299 | case IPPROTO_DSTOPTS: |
300 | case IPPROTO_ROUTING: |
301 | if (attrs[XFRMA_ALG_COMP] || |
302 | attrs[XFRMA_ALG_AUTH] || |
303 | attrs[XFRMA_ALG_AUTH_TRUNC] || |
304 | attrs[XFRMA_ALG_AEAD] || |
305 | attrs[XFRMA_ALG_CRYPT] || |
306 | attrs[XFRMA_ENCAP] || |
307 | attrs[XFRMA_SEC_CTX] || |
308 | attrs[XFRMA_TFCPAD]) { |
309 | NL_SET_ERR_MSG(extack, "Invalid attributes for DSTOPTS/ROUTING" ); |
310 | goto out; |
311 | } |
312 | |
313 | if (!attrs[XFRMA_COADDR]) { |
314 | NL_SET_ERR_MSG(extack, "Missing required COADDR attribute for DSTOPTS/ROUTING" ); |
315 | goto out; |
316 | } |
317 | break; |
318 | #endif |
319 | |
320 | default: |
321 | NL_SET_ERR_MSG(extack, "Unsupported protocol" ); |
322 | goto out; |
323 | } |
324 | |
325 | if ((err = verify_aead(attrs, extack))) |
326 | goto out; |
327 | if ((err = verify_auth_trunc(attrs, extack))) |
328 | goto out; |
329 | if ((err = verify_one_alg(attrs, type: XFRMA_ALG_AUTH, extack))) |
330 | goto out; |
331 | if ((err = verify_one_alg(attrs, type: XFRMA_ALG_CRYPT, extack))) |
332 | goto out; |
333 | if ((err = verify_one_alg(attrs, type: XFRMA_ALG_COMP, extack))) |
334 | goto out; |
335 | if ((err = verify_sec_ctx_len(attrs, extack))) |
336 | goto out; |
337 | if ((err = verify_replay(p, attrs, extack))) |
338 | goto out; |
339 | |
340 | err = -EINVAL; |
341 | switch (p->mode) { |
342 | case XFRM_MODE_TRANSPORT: |
343 | case XFRM_MODE_TUNNEL: |
344 | case XFRM_MODE_ROUTEOPTIMIZATION: |
345 | case XFRM_MODE_BEET: |
346 | break; |
347 | |
348 | default: |
349 | NL_SET_ERR_MSG(extack, "Unsupported mode" ); |
350 | goto out; |
351 | } |
352 | |
353 | err = 0; |
354 | |
355 | if (attrs[XFRMA_MTIMER_THRESH]) { |
356 | if (!attrs[XFRMA_ENCAP]) { |
357 | NL_SET_ERR_MSG(extack, "MTIMER_THRESH attribute can only be set on ENCAP states" ); |
358 | err = -EINVAL; |
359 | goto out; |
360 | } |
361 | } |
362 | |
363 | out: |
364 | return err; |
365 | } |
366 | |
367 | static int attach_one_algo(struct xfrm_algo **algpp, u8 *props, |
368 | struct xfrm_algo_desc *(*get_byname)(const char *, int), |
369 | struct nlattr *rta, struct netlink_ext_ack *extack) |
370 | { |
371 | struct xfrm_algo *p, *ualg; |
372 | struct xfrm_algo_desc *algo; |
373 | |
374 | if (!rta) |
375 | return 0; |
376 | |
377 | ualg = nla_data(nla: rta); |
378 | |
379 | algo = get_byname(ualg->alg_name, 1); |
380 | if (!algo) { |
381 | NL_SET_ERR_MSG(extack, "Requested COMP algorithm not found" ); |
382 | return -ENOSYS; |
383 | } |
384 | *props = algo->desc.sadb_alg_id; |
385 | |
386 | p = kmemdup(p: ualg, size: xfrm_alg_len(alg: ualg), GFP_KERNEL); |
387 | if (!p) |
388 | return -ENOMEM; |
389 | |
390 | strcpy(p: p->alg_name, q: algo->name); |
391 | *algpp = p; |
392 | return 0; |
393 | } |
394 | |
395 | static int attach_crypt(struct xfrm_state *x, struct nlattr *rta, |
396 | struct netlink_ext_ack *extack) |
397 | { |
398 | struct xfrm_algo *p, *ualg; |
399 | struct xfrm_algo_desc *algo; |
400 | |
401 | if (!rta) |
402 | return 0; |
403 | |
404 | ualg = nla_data(nla: rta); |
405 | |
406 | algo = xfrm_ealg_get_byname(name: ualg->alg_name, probe: 1); |
407 | if (!algo) { |
408 | NL_SET_ERR_MSG(extack, "Requested CRYPT algorithm not found" ); |
409 | return -ENOSYS; |
410 | } |
411 | x->props.ealgo = algo->desc.sadb_alg_id; |
412 | |
413 | p = kmemdup(p: ualg, size: xfrm_alg_len(alg: ualg), GFP_KERNEL); |
414 | if (!p) |
415 | return -ENOMEM; |
416 | |
417 | strcpy(p: p->alg_name, q: algo->name); |
418 | x->ealg = p; |
419 | x->geniv = algo->uinfo.encr.geniv; |
420 | return 0; |
421 | } |
422 | |
423 | static int attach_auth(struct xfrm_algo_auth **algpp, u8 *props, |
424 | struct nlattr *rta, struct netlink_ext_ack *extack) |
425 | { |
426 | struct xfrm_algo *ualg; |
427 | struct xfrm_algo_auth *p; |
428 | struct xfrm_algo_desc *algo; |
429 | |
430 | if (!rta) |
431 | return 0; |
432 | |
433 | ualg = nla_data(nla: rta); |
434 | |
435 | algo = xfrm_aalg_get_byname(name: ualg->alg_name, probe: 1); |
436 | if (!algo) { |
437 | NL_SET_ERR_MSG(extack, "Requested AUTH algorithm not found" ); |
438 | return -ENOSYS; |
439 | } |
440 | *props = algo->desc.sadb_alg_id; |
441 | |
442 | p = kmalloc(size: sizeof(*p) + (ualg->alg_key_len + 7) / 8, GFP_KERNEL); |
443 | if (!p) |
444 | return -ENOMEM; |
445 | |
446 | strcpy(p: p->alg_name, q: algo->name); |
447 | p->alg_key_len = ualg->alg_key_len; |
448 | p->alg_trunc_len = algo->uinfo.auth.icv_truncbits; |
449 | memcpy(p->alg_key, ualg->alg_key, (ualg->alg_key_len + 7) / 8); |
450 | |
451 | *algpp = p; |
452 | return 0; |
453 | } |
454 | |
455 | static int attach_auth_trunc(struct xfrm_algo_auth **algpp, u8 *props, |
456 | struct nlattr *rta, struct netlink_ext_ack *extack) |
457 | { |
458 | struct xfrm_algo_auth *p, *ualg; |
459 | struct xfrm_algo_desc *algo; |
460 | |
461 | if (!rta) |
462 | return 0; |
463 | |
464 | ualg = nla_data(nla: rta); |
465 | |
466 | algo = xfrm_aalg_get_byname(name: ualg->alg_name, probe: 1); |
467 | if (!algo) { |
468 | NL_SET_ERR_MSG(extack, "Requested AUTH_TRUNC algorithm not found" ); |
469 | return -ENOSYS; |
470 | } |
471 | if (ualg->alg_trunc_len > algo->uinfo.auth.icv_fullbits) { |
472 | NL_SET_ERR_MSG(extack, "Invalid length requested for truncated ICV" ); |
473 | return -EINVAL; |
474 | } |
475 | *props = algo->desc.sadb_alg_id; |
476 | |
477 | p = kmemdup(p: ualg, size: xfrm_alg_auth_len(alg: ualg), GFP_KERNEL); |
478 | if (!p) |
479 | return -ENOMEM; |
480 | |
481 | strcpy(p: p->alg_name, q: algo->name); |
482 | if (!p->alg_trunc_len) |
483 | p->alg_trunc_len = algo->uinfo.auth.icv_truncbits; |
484 | |
485 | *algpp = p; |
486 | return 0; |
487 | } |
488 | |
489 | static int attach_aead(struct xfrm_state *x, struct nlattr *rta, |
490 | struct netlink_ext_ack *extack) |
491 | { |
492 | struct xfrm_algo_aead *p, *ualg; |
493 | struct xfrm_algo_desc *algo; |
494 | |
495 | if (!rta) |
496 | return 0; |
497 | |
498 | ualg = nla_data(nla: rta); |
499 | |
500 | algo = xfrm_aead_get_byname(name: ualg->alg_name, icv_len: ualg->alg_icv_len, probe: 1); |
501 | if (!algo) { |
502 | NL_SET_ERR_MSG(extack, "Requested AEAD algorithm not found" ); |
503 | return -ENOSYS; |
504 | } |
505 | x->props.ealgo = algo->desc.sadb_alg_id; |
506 | |
507 | p = kmemdup(p: ualg, size: aead_len(alg: ualg), GFP_KERNEL); |
508 | if (!p) |
509 | return -ENOMEM; |
510 | |
511 | strcpy(p: p->alg_name, q: algo->name); |
512 | x->aead = p; |
513 | x->geniv = algo->uinfo.aead.geniv; |
514 | return 0; |
515 | } |
516 | |
517 | static inline int xfrm_replay_verify_len(struct xfrm_replay_state_esn *replay_esn, |
518 | struct nlattr *rp, |
519 | struct netlink_ext_ack *extack) |
520 | { |
521 | struct xfrm_replay_state_esn *up; |
522 | unsigned int ulen; |
523 | |
524 | if (!replay_esn || !rp) |
525 | return 0; |
526 | |
527 | up = nla_data(nla: rp); |
528 | ulen = xfrm_replay_state_esn_len(replay_esn: up); |
529 | |
530 | /* Check the overall length and the internal bitmap length to avoid |
531 | * potential overflow. */ |
532 | if (nla_len(nla: rp) < (int)ulen) { |
533 | NL_SET_ERR_MSG(extack, "ESN attribute is too short" ); |
534 | return -EINVAL; |
535 | } |
536 | |
537 | if (xfrm_replay_state_esn_len(replay_esn) != ulen) { |
538 | NL_SET_ERR_MSG(extack, "New ESN size doesn't match the existing SA's ESN size" ); |
539 | return -EINVAL; |
540 | } |
541 | |
542 | if (replay_esn->bmp_len != up->bmp_len) { |
543 | NL_SET_ERR_MSG(extack, "New ESN bitmap size doesn't match the existing SA's ESN bitmap" ); |
544 | return -EINVAL; |
545 | } |
546 | |
547 | if (up->replay_window > up->bmp_len * sizeof(__u32) * 8) { |
548 | NL_SET_ERR_MSG(extack, "ESN replay window is longer than the bitmap" ); |
549 | return -EINVAL; |
550 | } |
551 | |
552 | return 0; |
553 | } |
554 | |
555 | static int xfrm_alloc_replay_state_esn(struct xfrm_replay_state_esn **replay_esn, |
556 | struct xfrm_replay_state_esn **preplay_esn, |
557 | struct nlattr *rta) |
558 | { |
559 | struct xfrm_replay_state_esn *p, *pp, *up; |
560 | unsigned int klen, ulen; |
561 | |
562 | if (!rta) |
563 | return 0; |
564 | |
565 | up = nla_data(nla: rta); |
566 | klen = xfrm_replay_state_esn_len(replay_esn: up); |
567 | ulen = nla_len(nla: rta) >= (int)klen ? klen : sizeof(*up); |
568 | |
569 | p = kzalloc(size: klen, GFP_KERNEL); |
570 | if (!p) |
571 | return -ENOMEM; |
572 | |
573 | pp = kzalloc(size: klen, GFP_KERNEL); |
574 | if (!pp) { |
575 | kfree(objp: p); |
576 | return -ENOMEM; |
577 | } |
578 | |
579 | memcpy(p, up, ulen); |
580 | memcpy(pp, up, ulen); |
581 | |
582 | *replay_esn = p; |
583 | *preplay_esn = pp; |
584 | |
585 | return 0; |
586 | } |
587 | |
588 | static inline unsigned int xfrm_user_sec_ctx_size(struct xfrm_sec_ctx *xfrm_ctx) |
589 | { |
590 | unsigned int len = 0; |
591 | |
592 | if (xfrm_ctx) { |
593 | len += sizeof(struct xfrm_user_sec_ctx); |
594 | len += xfrm_ctx->ctx_len; |
595 | } |
596 | return len; |
597 | } |
598 | |
599 | static void copy_from_user_state(struct xfrm_state *x, struct xfrm_usersa_info *p) |
600 | { |
601 | memcpy(&x->id, &p->id, sizeof(x->id)); |
602 | memcpy(&x->sel, &p->sel, sizeof(x->sel)); |
603 | memcpy(&x->lft, &p->lft, sizeof(x->lft)); |
604 | x->props.mode = p->mode; |
605 | x->props.replay_window = min_t(unsigned int, p->replay_window, |
606 | sizeof(x->replay.bitmap) * 8); |
607 | x->props.reqid = p->reqid; |
608 | x->props.family = p->family; |
609 | memcpy(&x->props.saddr, &p->saddr, sizeof(x->props.saddr)); |
610 | x->props.flags = p->flags; |
611 | |
612 | if (!x->sel.family && !(p->flags & XFRM_STATE_AF_UNSPEC)) |
613 | x->sel.family = p->family; |
614 | } |
615 | |
616 | /* |
617 | * someday when pfkey also has support, we could have the code |
618 | * somehow made shareable and move it to xfrm_state.c - JHS |
619 | * |
620 | */ |
621 | static void xfrm_update_ae_params(struct xfrm_state *x, struct nlattr **attrs, |
622 | int update_esn) |
623 | { |
624 | struct nlattr *rp = attrs[XFRMA_REPLAY_VAL]; |
625 | struct nlattr *re = update_esn ? attrs[XFRMA_REPLAY_ESN_VAL] : NULL; |
626 | struct nlattr *lt = attrs[XFRMA_LTIME_VAL]; |
627 | struct nlattr *et = attrs[XFRMA_ETIMER_THRESH]; |
628 | struct nlattr *rt = attrs[XFRMA_REPLAY_THRESH]; |
629 | struct nlattr *mt = attrs[XFRMA_MTIMER_THRESH]; |
630 | |
631 | if (re && x->replay_esn && x->preplay_esn) { |
632 | struct xfrm_replay_state_esn *replay_esn; |
633 | replay_esn = nla_data(nla: re); |
634 | memcpy(x->replay_esn, replay_esn, |
635 | xfrm_replay_state_esn_len(replay_esn)); |
636 | memcpy(x->preplay_esn, replay_esn, |
637 | xfrm_replay_state_esn_len(replay_esn)); |
638 | } |
639 | |
640 | if (rp) { |
641 | struct xfrm_replay_state *replay; |
642 | replay = nla_data(nla: rp); |
643 | memcpy(&x->replay, replay, sizeof(*replay)); |
644 | memcpy(&x->preplay, replay, sizeof(*replay)); |
645 | } |
646 | |
647 | if (lt) { |
648 | struct xfrm_lifetime_cur *ltime; |
649 | ltime = nla_data(nla: lt); |
650 | x->curlft.bytes = ltime->bytes; |
651 | x->curlft.packets = ltime->packets; |
652 | x->curlft.add_time = ltime->add_time; |
653 | x->curlft.use_time = ltime->use_time; |
654 | } |
655 | |
656 | if (et) |
657 | x->replay_maxage = nla_get_u32(nla: et); |
658 | |
659 | if (rt) |
660 | x->replay_maxdiff = nla_get_u32(nla: rt); |
661 | |
662 | if (mt) |
663 | x->mapping_maxage = nla_get_u32(nla: mt); |
664 | } |
665 | |
666 | static void xfrm_smark_init(struct nlattr **attrs, struct xfrm_mark *m) |
667 | { |
668 | if (attrs[XFRMA_SET_MARK]) { |
669 | m->v = nla_get_u32(nla: attrs[XFRMA_SET_MARK]); |
670 | if (attrs[XFRMA_SET_MARK_MASK]) |
671 | m->m = nla_get_u32(nla: attrs[XFRMA_SET_MARK_MASK]); |
672 | else |
673 | m->m = 0xffffffff; |
674 | } else { |
675 | m->v = m->m = 0; |
676 | } |
677 | } |
678 | |
679 | static struct xfrm_state *xfrm_state_construct(struct net *net, |
680 | struct xfrm_usersa_info *p, |
681 | struct nlattr **attrs, |
682 | int *errp, |
683 | struct netlink_ext_ack *extack) |
684 | { |
685 | struct xfrm_state *x = xfrm_state_alloc(net); |
686 | int err = -ENOMEM; |
687 | |
688 | if (!x) |
689 | goto error_no_put; |
690 | |
691 | copy_from_user_state(x, p); |
692 | |
693 | if (attrs[XFRMA_ENCAP]) { |
694 | x->encap = kmemdup(p: nla_data(nla: attrs[XFRMA_ENCAP]), |
695 | size: sizeof(*x->encap), GFP_KERNEL); |
696 | if (x->encap == NULL) |
697 | goto error; |
698 | } |
699 | |
700 | if (attrs[XFRMA_COADDR]) { |
701 | x->coaddr = kmemdup(p: nla_data(nla: attrs[XFRMA_COADDR]), |
702 | size: sizeof(*x->coaddr), GFP_KERNEL); |
703 | if (x->coaddr == NULL) |
704 | goto error; |
705 | } |
706 | |
707 | if (attrs[XFRMA_SA_EXTRA_FLAGS]) |
708 | x->props.extra_flags = nla_get_u32(nla: attrs[XFRMA_SA_EXTRA_FLAGS]); |
709 | |
710 | if ((err = attach_aead(x, rta: attrs[XFRMA_ALG_AEAD], extack))) |
711 | goto error; |
712 | if ((err = attach_auth_trunc(algpp: &x->aalg, props: &x->props.aalgo, |
713 | rta: attrs[XFRMA_ALG_AUTH_TRUNC], extack))) |
714 | goto error; |
715 | if (!x->props.aalgo) { |
716 | if ((err = attach_auth(algpp: &x->aalg, props: &x->props.aalgo, |
717 | rta: attrs[XFRMA_ALG_AUTH], extack))) |
718 | goto error; |
719 | } |
720 | if ((err = attach_crypt(x, rta: attrs[XFRMA_ALG_CRYPT], extack))) |
721 | goto error; |
722 | if ((err = attach_one_algo(algpp: &x->calg, props: &x->props.calgo, |
723 | get_byname: xfrm_calg_get_byname, |
724 | rta: attrs[XFRMA_ALG_COMP], extack))) |
725 | goto error; |
726 | |
727 | if (attrs[XFRMA_TFCPAD]) |
728 | x->tfcpad = nla_get_u32(nla: attrs[XFRMA_TFCPAD]); |
729 | |
730 | xfrm_mark_get(attrs, m: &x->mark); |
731 | |
732 | xfrm_smark_init(attrs, m: &x->props.smark); |
733 | |
734 | if (attrs[XFRMA_IF_ID]) |
735 | x->if_id = nla_get_u32(nla: attrs[XFRMA_IF_ID]); |
736 | |
737 | err = __xfrm_init_state(x, init_replay: false, offload: attrs[XFRMA_OFFLOAD_DEV], extack); |
738 | if (err) |
739 | goto error; |
740 | |
741 | if (attrs[XFRMA_SEC_CTX]) { |
742 | err = security_xfrm_state_alloc(x, |
743 | sec_ctx: nla_data(nla: attrs[XFRMA_SEC_CTX])); |
744 | if (err) |
745 | goto error; |
746 | } |
747 | |
748 | if ((err = xfrm_alloc_replay_state_esn(replay_esn: &x->replay_esn, preplay_esn: &x->preplay_esn, |
749 | rta: attrs[XFRMA_REPLAY_ESN_VAL]))) |
750 | goto error; |
751 | |
752 | x->km.seq = p->seq; |
753 | x->replay_maxdiff = net->xfrm.sysctl_aevent_rseqth; |
754 | /* sysctl_xfrm_aevent_etime is in 100ms units */ |
755 | x->replay_maxage = (net->xfrm.sysctl_aevent_etime*HZ)/XFRM_AE_ETH_M; |
756 | |
757 | if ((err = xfrm_init_replay(x, extack))) |
758 | goto error; |
759 | |
760 | /* override default values from above */ |
761 | xfrm_update_ae_params(x, attrs, update_esn: 0); |
762 | |
763 | /* configure the hardware if offload is requested */ |
764 | if (attrs[XFRMA_OFFLOAD_DEV]) { |
765 | err = xfrm_dev_state_add(net, x, |
766 | xuo: nla_data(nla: attrs[XFRMA_OFFLOAD_DEV]), |
767 | extack); |
768 | if (err) |
769 | goto error; |
770 | } |
771 | |
772 | return x; |
773 | |
774 | error: |
775 | x->km.state = XFRM_STATE_DEAD; |
776 | xfrm_state_put(x); |
777 | error_no_put: |
778 | *errp = err; |
779 | return NULL; |
780 | } |
781 | |
782 | static int xfrm_add_sa(struct sk_buff *skb, struct nlmsghdr *nlh, |
783 | struct nlattr **attrs, struct netlink_ext_ack *extack) |
784 | { |
785 | struct net *net = sock_net(sk: skb->sk); |
786 | struct xfrm_usersa_info *p = nlmsg_data(nlh); |
787 | struct xfrm_state *x; |
788 | int err; |
789 | struct km_event c; |
790 | |
791 | err = verify_newsa_info(p, attrs, extack); |
792 | if (err) |
793 | return err; |
794 | |
795 | x = xfrm_state_construct(net, p, attrs, errp: &err, extack); |
796 | if (!x) |
797 | return err; |
798 | |
799 | xfrm_state_hold(x); |
800 | if (nlh->nlmsg_type == XFRM_MSG_NEWSA) |
801 | err = xfrm_state_add(x); |
802 | else |
803 | err = xfrm_state_update(x); |
804 | |
805 | xfrm_audit_state_add(x, result: err ? 0 : 1, task_valid: true); |
806 | |
807 | if (err < 0) { |
808 | x->km.state = XFRM_STATE_DEAD; |
809 | xfrm_dev_state_delete(x); |
810 | __xfrm_state_put(x); |
811 | goto out; |
812 | } |
813 | |
814 | if (x->km.state == XFRM_STATE_VOID) |
815 | x->km.state = XFRM_STATE_VALID; |
816 | |
817 | c.seq = nlh->nlmsg_seq; |
818 | c.portid = nlh->nlmsg_pid; |
819 | c.event = nlh->nlmsg_type; |
820 | |
821 | km_state_notify(x, c: &c); |
822 | out: |
823 | xfrm_state_put(x); |
824 | return err; |
825 | } |
826 | |
827 | static struct xfrm_state *xfrm_user_state_lookup(struct net *net, |
828 | struct xfrm_usersa_id *p, |
829 | struct nlattr **attrs, |
830 | int *errp) |
831 | { |
832 | struct xfrm_state *x = NULL; |
833 | struct xfrm_mark m; |
834 | int err; |
835 | u32 mark = xfrm_mark_get(attrs, m: &m); |
836 | |
837 | if (xfrm_id_proto_match(proto: p->proto, IPSEC_PROTO_ANY)) { |
838 | err = -ESRCH; |
839 | x = xfrm_state_lookup(net, mark, daddr: &p->daddr, spi: p->spi, proto: p->proto, family: p->family); |
840 | } else { |
841 | xfrm_address_t *saddr = NULL; |
842 | |
843 | verify_one_addr(attrs, type: XFRMA_SRCADDR, addrp: &saddr); |
844 | if (!saddr) { |
845 | err = -EINVAL; |
846 | goto out; |
847 | } |
848 | |
849 | err = -ESRCH; |
850 | x = xfrm_state_lookup_byaddr(net, mark, |
851 | daddr: &p->daddr, saddr, |
852 | proto: p->proto, family: p->family); |
853 | } |
854 | |
855 | out: |
856 | if (!x && errp) |
857 | *errp = err; |
858 | return x; |
859 | } |
860 | |
861 | static int xfrm_del_sa(struct sk_buff *skb, struct nlmsghdr *nlh, |
862 | struct nlattr **attrs, struct netlink_ext_ack *extack) |
863 | { |
864 | struct net *net = sock_net(sk: skb->sk); |
865 | struct xfrm_state *x; |
866 | int err = -ESRCH; |
867 | struct km_event c; |
868 | struct xfrm_usersa_id *p = nlmsg_data(nlh); |
869 | |
870 | x = xfrm_user_state_lookup(net, p, attrs, errp: &err); |
871 | if (x == NULL) |
872 | return err; |
873 | |
874 | if ((err = security_xfrm_state_delete(x)) != 0) |
875 | goto out; |
876 | |
877 | if (xfrm_state_kern(x)) { |
878 | NL_SET_ERR_MSG(extack, "SA is in use by tunnels" ); |
879 | err = -EPERM; |
880 | goto out; |
881 | } |
882 | |
883 | err = xfrm_state_delete(x); |
884 | if (err < 0) |
885 | goto out; |
886 | |
887 | c.seq = nlh->nlmsg_seq; |
888 | c.portid = nlh->nlmsg_pid; |
889 | c.event = nlh->nlmsg_type; |
890 | km_state_notify(x, c: &c); |
891 | |
892 | out: |
893 | xfrm_audit_state_delete(x, result: err ? 0 : 1, task_valid: true); |
894 | xfrm_state_put(x); |
895 | return err; |
896 | } |
897 | |
898 | static void copy_to_user_state(struct xfrm_state *x, struct xfrm_usersa_info *p) |
899 | { |
900 | memset(p, 0, sizeof(*p)); |
901 | memcpy(&p->id, &x->id, sizeof(p->id)); |
902 | memcpy(&p->sel, &x->sel, sizeof(p->sel)); |
903 | memcpy(&p->lft, &x->lft, sizeof(p->lft)); |
904 | if (x->xso.dev) |
905 | xfrm_dev_state_update_stats(x); |
906 | memcpy(&p->curlft, &x->curlft, sizeof(p->curlft)); |
907 | put_unaligned(x->stats.replay_window, &p->stats.replay_window); |
908 | put_unaligned(x->stats.replay, &p->stats.replay); |
909 | put_unaligned(x->stats.integrity_failed, &p->stats.integrity_failed); |
910 | memcpy(&p->saddr, &x->props.saddr, sizeof(p->saddr)); |
911 | p->mode = x->props.mode; |
912 | p->replay_window = x->props.replay_window; |
913 | p->reqid = x->props.reqid; |
914 | p->family = x->props.family; |
915 | p->flags = x->props.flags; |
916 | p->seq = x->km.seq; |
917 | } |
918 | |
919 | struct xfrm_dump_info { |
920 | struct sk_buff *in_skb; |
921 | struct sk_buff *out_skb; |
922 | u32 nlmsg_seq; |
923 | u16 nlmsg_flags; |
924 | }; |
925 | |
926 | static int copy_sec_ctx(struct xfrm_sec_ctx *s, struct sk_buff *skb) |
927 | { |
928 | struct xfrm_user_sec_ctx *uctx; |
929 | struct nlattr *attr; |
930 | int ctx_size = sizeof(*uctx) + s->ctx_len; |
931 | |
932 | attr = nla_reserve(skb, attrtype: XFRMA_SEC_CTX, attrlen: ctx_size); |
933 | if (attr == NULL) |
934 | return -EMSGSIZE; |
935 | |
936 | uctx = nla_data(nla: attr); |
937 | uctx->exttype = XFRMA_SEC_CTX; |
938 | uctx->len = ctx_size; |
939 | uctx->ctx_doi = s->ctx_doi; |
940 | uctx->ctx_alg = s->ctx_alg; |
941 | uctx->ctx_len = s->ctx_len; |
942 | memcpy(uctx + 1, s->ctx_str, s->ctx_len); |
943 | |
944 | return 0; |
945 | } |
946 | |
947 | static int copy_user_offload(struct xfrm_dev_offload *xso, struct sk_buff *skb) |
948 | { |
949 | struct xfrm_user_offload *xuo; |
950 | struct nlattr *attr; |
951 | |
952 | attr = nla_reserve(skb, attrtype: XFRMA_OFFLOAD_DEV, attrlen: sizeof(*xuo)); |
953 | if (attr == NULL) |
954 | return -EMSGSIZE; |
955 | |
956 | xuo = nla_data(nla: attr); |
957 | memset(xuo, 0, sizeof(*xuo)); |
958 | xuo->ifindex = xso->dev->ifindex; |
959 | if (xso->dir == XFRM_DEV_OFFLOAD_IN) |
960 | xuo->flags = XFRM_OFFLOAD_INBOUND; |
961 | if (xso->type == XFRM_DEV_OFFLOAD_PACKET) |
962 | xuo->flags |= XFRM_OFFLOAD_PACKET; |
963 | |
964 | return 0; |
965 | } |
966 | |
967 | static bool xfrm_redact(void) |
968 | { |
969 | return IS_ENABLED(CONFIG_SECURITY) && |
970 | security_locked_down(what: LOCKDOWN_XFRM_SECRET); |
971 | } |
972 | |
973 | static int copy_to_user_auth(struct xfrm_algo_auth *auth, struct sk_buff *skb) |
974 | { |
975 | struct xfrm_algo *algo; |
976 | struct xfrm_algo_auth *ap; |
977 | struct nlattr *nla; |
978 | bool redact_secret = xfrm_redact(); |
979 | |
980 | nla = nla_reserve(skb, attrtype: XFRMA_ALG_AUTH, |
981 | attrlen: sizeof(*algo) + (auth->alg_key_len + 7) / 8); |
982 | if (!nla) |
983 | return -EMSGSIZE; |
984 | algo = nla_data(nla); |
985 | strncpy(p: algo->alg_name, q: auth->alg_name, size: sizeof(algo->alg_name)); |
986 | |
987 | if (redact_secret && auth->alg_key_len) |
988 | memset(algo->alg_key, 0, (auth->alg_key_len + 7) / 8); |
989 | else |
990 | memcpy(algo->alg_key, auth->alg_key, |
991 | (auth->alg_key_len + 7) / 8); |
992 | algo->alg_key_len = auth->alg_key_len; |
993 | |
994 | nla = nla_reserve(skb, attrtype: XFRMA_ALG_AUTH_TRUNC, attrlen: xfrm_alg_auth_len(alg: auth)); |
995 | if (!nla) |
996 | return -EMSGSIZE; |
997 | ap = nla_data(nla); |
998 | memcpy(ap, auth, sizeof(struct xfrm_algo_auth)); |
999 | if (redact_secret && auth->alg_key_len) |
1000 | memset(ap->alg_key, 0, (auth->alg_key_len + 7) / 8); |
1001 | else |
1002 | memcpy(ap->alg_key, auth->alg_key, |
1003 | (auth->alg_key_len + 7) / 8); |
1004 | return 0; |
1005 | } |
1006 | |
1007 | static int copy_to_user_aead(struct xfrm_algo_aead *aead, struct sk_buff *skb) |
1008 | { |
1009 | struct nlattr *nla = nla_reserve(skb, attrtype: XFRMA_ALG_AEAD, attrlen: aead_len(alg: aead)); |
1010 | struct xfrm_algo_aead *ap; |
1011 | bool redact_secret = xfrm_redact(); |
1012 | |
1013 | if (!nla) |
1014 | return -EMSGSIZE; |
1015 | |
1016 | ap = nla_data(nla); |
1017 | strscpy_pad(ap->alg_name, aead->alg_name, sizeof(ap->alg_name)); |
1018 | ap->alg_key_len = aead->alg_key_len; |
1019 | ap->alg_icv_len = aead->alg_icv_len; |
1020 | |
1021 | if (redact_secret && aead->alg_key_len) |
1022 | memset(ap->alg_key, 0, (aead->alg_key_len + 7) / 8); |
1023 | else |
1024 | memcpy(ap->alg_key, aead->alg_key, |
1025 | (aead->alg_key_len + 7) / 8); |
1026 | return 0; |
1027 | } |
1028 | |
1029 | static int copy_to_user_ealg(struct xfrm_algo *ealg, struct sk_buff *skb) |
1030 | { |
1031 | struct xfrm_algo *ap; |
1032 | bool redact_secret = xfrm_redact(); |
1033 | struct nlattr *nla = nla_reserve(skb, attrtype: XFRMA_ALG_CRYPT, |
1034 | attrlen: xfrm_alg_len(alg: ealg)); |
1035 | if (!nla) |
1036 | return -EMSGSIZE; |
1037 | |
1038 | ap = nla_data(nla); |
1039 | strscpy_pad(ap->alg_name, ealg->alg_name, sizeof(ap->alg_name)); |
1040 | ap->alg_key_len = ealg->alg_key_len; |
1041 | |
1042 | if (redact_secret && ealg->alg_key_len) |
1043 | memset(ap->alg_key, 0, (ealg->alg_key_len + 7) / 8); |
1044 | else |
1045 | memcpy(ap->alg_key, ealg->alg_key, |
1046 | (ealg->alg_key_len + 7) / 8); |
1047 | |
1048 | return 0; |
1049 | } |
1050 | |
1051 | static int copy_to_user_calg(struct xfrm_algo *calg, struct sk_buff *skb) |
1052 | { |
1053 | struct nlattr *nla = nla_reserve(skb, attrtype: XFRMA_ALG_COMP, attrlen: sizeof(*calg)); |
1054 | struct xfrm_algo *ap; |
1055 | |
1056 | if (!nla) |
1057 | return -EMSGSIZE; |
1058 | |
1059 | ap = nla_data(nla); |
1060 | strscpy_pad(ap->alg_name, calg->alg_name, sizeof(ap->alg_name)); |
1061 | ap->alg_key_len = 0; |
1062 | |
1063 | return 0; |
1064 | } |
1065 | |
1066 | static int copy_to_user_encap(struct xfrm_encap_tmpl *ep, struct sk_buff *skb) |
1067 | { |
1068 | struct nlattr *nla = nla_reserve(skb, attrtype: XFRMA_ENCAP, attrlen: sizeof(*ep)); |
1069 | struct xfrm_encap_tmpl *uep; |
1070 | |
1071 | if (!nla) |
1072 | return -EMSGSIZE; |
1073 | |
1074 | uep = nla_data(nla); |
1075 | memset(uep, 0, sizeof(*uep)); |
1076 | |
1077 | uep->encap_type = ep->encap_type; |
1078 | uep->encap_sport = ep->encap_sport; |
1079 | uep->encap_dport = ep->encap_dport; |
1080 | uep->encap_oa = ep->encap_oa; |
1081 | |
1082 | return 0; |
1083 | } |
1084 | |
1085 | static int xfrm_smark_put(struct sk_buff *skb, struct xfrm_mark *m) |
1086 | { |
1087 | int ret = 0; |
1088 | |
1089 | if (m->v | m->m) { |
1090 | ret = nla_put_u32(skb, attrtype: XFRMA_SET_MARK, value: m->v); |
1091 | if (!ret) |
1092 | ret = nla_put_u32(skb, attrtype: XFRMA_SET_MARK_MASK, value: m->m); |
1093 | } |
1094 | return ret; |
1095 | } |
1096 | |
1097 | /* Don't change this without updating xfrm_sa_len! */ |
1098 | static int (struct xfrm_state *x, |
1099 | struct xfrm_usersa_info *p, |
1100 | struct sk_buff *skb) |
1101 | { |
1102 | int ret = 0; |
1103 | |
1104 | copy_to_user_state(x, p); |
1105 | |
1106 | if (x->props.extra_flags) { |
1107 | ret = nla_put_u32(skb, attrtype: XFRMA_SA_EXTRA_FLAGS, |
1108 | value: x->props.extra_flags); |
1109 | if (ret) |
1110 | goto out; |
1111 | } |
1112 | |
1113 | if (x->coaddr) { |
1114 | ret = nla_put(skb, attrtype: XFRMA_COADDR, attrlen: sizeof(*x->coaddr), data: x->coaddr); |
1115 | if (ret) |
1116 | goto out; |
1117 | } |
1118 | if (x->lastused) { |
1119 | ret = nla_put_u64_64bit(skb, attrtype: XFRMA_LASTUSED, value: x->lastused, |
1120 | padattr: XFRMA_PAD); |
1121 | if (ret) |
1122 | goto out; |
1123 | } |
1124 | if (x->aead) { |
1125 | ret = copy_to_user_aead(aead: x->aead, skb); |
1126 | if (ret) |
1127 | goto out; |
1128 | } |
1129 | if (x->aalg) { |
1130 | ret = copy_to_user_auth(auth: x->aalg, skb); |
1131 | if (ret) |
1132 | goto out; |
1133 | } |
1134 | if (x->ealg) { |
1135 | ret = copy_to_user_ealg(ealg: x->ealg, skb); |
1136 | if (ret) |
1137 | goto out; |
1138 | } |
1139 | if (x->calg) { |
1140 | ret = copy_to_user_calg(calg: x->calg, skb); |
1141 | if (ret) |
1142 | goto out; |
1143 | } |
1144 | if (x->encap) { |
1145 | ret = copy_to_user_encap(ep: x->encap, skb); |
1146 | if (ret) |
1147 | goto out; |
1148 | } |
1149 | if (x->tfcpad) { |
1150 | ret = nla_put_u32(skb, attrtype: XFRMA_TFCPAD, value: x->tfcpad); |
1151 | if (ret) |
1152 | goto out; |
1153 | } |
1154 | ret = xfrm_mark_put(skb, m: &x->mark); |
1155 | if (ret) |
1156 | goto out; |
1157 | |
1158 | ret = xfrm_smark_put(skb, m: &x->props.smark); |
1159 | if (ret) |
1160 | goto out; |
1161 | |
1162 | if (x->replay_esn) |
1163 | ret = nla_put(skb, attrtype: XFRMA_REPLAY_ESN_VAL, |
1164 | attrlen: xfrm_replay_state_esn_len(replay_esn: x->replay_esn), |
1165 | data: x->replay_esn); |
1166 | else |
1167 | ret = nla_put(skb, attrtype: XFRMA_REPLAY_VAL, attrlen: sizeof(x->replay), |
1168 | data: &x->replay); |
1169 | if (ret) |
1170 | goto out; |
1171 | if(x->xso.dev) |
1172 | ret = copy_user_offload(xso: &x->xso, skb); |
1173 | if (ret) |
1174 | goto out; |
1175 | if (x->if_id) { |
1176 | ret = nla_put_u32(skb, attrtype: XFRMA_IF_ID, value: x->if_id); |
1177 | if (ret) |
1178 | goto out; |
1179 | } |
1180 | if (x->security) { |
1181 | ret = copy_sec_ctx(s: x->security, skb); |
1182 | if (ret) |
1183 | goto out; |
1184 | } |
1185 | if (x->mapping_maxage) |
1186 | ret = nla_put_u32(skb, attrtype: XFRMA_MTIMER_THRESH, value: x->mapping_maxage); |
1187 | out: |
1188 | return ret; |
1189 | } |
1190 | |
1191 | static int dump_one_state(struct xfrm_state *x, int count, void *ptr) |
1192 | { |
1193 | struct xfrm_dump_info *sp = ptr; |
1194 | struct sk_buff *in_skb = sp->in_skb; |
1195 | struct sk_buff *skb = sp->out_skb; |
1196 | struct xfrm_translator *xtr; |
1197 | struct xfrm_usersa_info *p; |
1198 | struct nlmsghdr *nlh; |
1199 | int err; |
1200 | |
1201 | nlh = nlmsg_put(skb, NETLINK_CB(in_skb).portid, seq: sp->nlmsg_seq, |
1202 | XFRM_MSG_NEWSA, payload: sizeof(*p), flags: sp->nlmsg_flags); |
1203 | if (nlh == NULL) |
1204 | return -EMSGSIZE; |
1205 | |
1206 | p = nlmsg_data(nlh); |
1207 | |
1208 | err = copy_to_user_state_extra(x, p, skb); |
1209 | if (err) { |
1210 | nlmsg_cancel(skb, nlh); |
1211 | return err; |
1212 | } |
1213 | nlmsg_end(skb, nlh); |
1214 | |
1215 | xtr = xfrm_get_translator(); |
1216 | if (xtr) { |
1217 | err = xtr->alloc_compat(skb, nlh); |
1218 | |
1219 | xfrm_put_translator(xtr); |
1220 | if (err) { |
1221 | nlmsg_cancel(skb, nlh); |
1222 | return err; |
1223 | } |
1224 | } |
1225 | |
1226 | return 0; |
1227 | } |
1228 | |
1229 | static int xfrm_dump_sa_done(struct netlink_callback *cb) |
1230 | { |
1231 | struct xfrm_state_walk *walk = (struct xfrm_state_walk *) &cb->args[1]; |
1232 | struct sock *sk = cb->skb->sk; |
1233 | struct net *net = sock_net(sk); |
1234 | |
1235 | if (cb->args[0]) |
1236 | xfrm_state_walk_done(walk, net); |
1237 | return 0; |
1238 | } |
1239 | |
1240 | static int xfrm_dump_sa(struct sk_buff *skb, struct netlink_callback *cb) |
1241 | { |
1242 | struct net *net = sock_net(sk: skb->sk); |
1243 | struct xfrm_state_walk *walk = (struct xfrm_state_walk *) &cb->args[1]; |
1244 | struct xfrm_dump_info info; |
1245 | |
1246 | BUILD_BUG_ON(sizeof(struct xfrm_state_walk) > |
1247 | sizeof(cb->args) - sizeof(cb->args[0])); |
1248 | |
1249 | info.in_skb = cb->skb; |
1250 | info.out_skb = skb; |
1251 | info.nlmsg_seq = cb->nlh->nlmsg_seq; |
1252 | info.nlmsg_flags = NLM_F_MULTI; |
1253 | |
1254 | if (!cb->args[0]) { |
1255 | struct nlattr *attrs[XFRMA_MAX+1]; |
1256 | struct xfrm_address_filter *filter = NULL; |
1257 | u8 proto = 0; |
1258 | int err; |
1259 | |
1260 | err = nlmsg_parse_deprecated(nlh: cb->nlh, hdrlen: 0, tb: attrs, XFRMA_MAX, |
1261 | policy: xfrma_policy, extack: cb->extack); |
1262 | if (err < 0) |
1263 | return err; |
1264 | |
1265 | if (attrs[XFRMA_ADDRESS_FILTER]) { |
1266 | filter = kmemdup(p: nla_data(nla: attrs[XFRMA_ADDRESS_FILTER]), |
1267 | size: sizeof(*filter), GFP_KERNEL); |
1268 | if (filter == NULL) |
1269 | return -ENOMEM; |
1270 | |
1271 | /* see addr_match(), (prefix length >> 5) << 2 |
1272 | * will be used to compare xfrm_address_t |
1273 | */ |
1274 | if (filter->splen > (sizeof(xfrm_address_t) << 3) || |
1275 | filter->dplen > (sizeof(xfrm_address_t) << 3)) { |
1276 | kfree(objp: filter); |
1277 | return -EINVAL; |
1278 | } |
1279 | } |
1280 | |
1281 | if (attrs[XFRMA_PROTO]) |
1282 | proto = nla_get_u8(nla: attrs[XFRMA_PROTO]); |
1283 | |
1284 | xfrm_state_walk_init(walk, proto, filter); |
1285 | cb->args[0] = 1; |
1286 | } |
1287 | |
1288 | (void) xfrm_state_walk(net, walk, func: dump_one_state, &info); |
1289 | |
1290 | return skb->len; |
1291 | } |
1292 | |
1293 | static struct sk_buff *xfrm_state_netlink(struct sk_buff *in_skb, |
1294 | struct xfrm_state *x, u32 seq) |
1295 | { |
1296 | struct xfrm_dump_info info; |
1297 | struct sk_buff *skb; |
1298 | int err; |
1299 | |
1300 | skb = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_ATOMIC); |
1301 | if (!skb) |
1302 | return ERR_PTR(error: -ENOMEM); |
1303 | |
1304 | info.in_skb = in_skb; |
1305 | info.out_skb = skb; |
1306 | info.nlmsg_seq = seq; |
1307 | info.nlmsg_flags = 0; |
1308 | |
1309 | err = dump_one_state(x, count: 0, ptr: &info); |
1310 | if (err) { |
1311 | kfree_skb(skb); |
1312 | return ERR_PTR(error: err); |
1313 | } |
1314 | |
1315 | return skb; |
1316 | } |
1317 | |
1318 | /* A wrapper for nlmsg_multicast() checking that nlsk is still available. |
1319 | * Must be called with RCU read lock. |
1320 | */ |
1321 | static inline int xfrm_nlmsg_multicast(struct net *net, struct sk_buff *skb, |
1322 | u32 pid, unsigned int group) |
1323 | { |
1324 | struct sock *nlsk = rcu_dereference(net->xfrm.nlsk); |
1325 | struct xfrm_translator *xtr; |
1326 | |
1327 | if (!nlsk) { |
1328 | kfree_skb(skb); |
1329 | return -EPIPE; |
1330 | } |
1331 | |
1332 | xtr = xfrm_get_translator(); |
1333 | if (xtr) { |
1334 | int err = xtr->alloc_compat(skb, nlmsg_hdr(skb)); |
1335 | |
1336 | xfrm_put_translator(xtr); |
1337 | if (err) { |
1338 | kfree_skb(skb); |
1339 | return err; |
1340 | } |
1341 | } |
1342 | |
1343 | return nlmsg_multicast(sk: nlsk, skb, portid: pid, group, GFP_ATOMIC); |
1344 | } |
1345 | |
1346 | static inline unsigned int xfrm_spdinfo_msgsize(void) |
1347 | { |
1348 | return NLMSG_ALIGN(4) |
1349 | + nla_total_size(payload: sizeof(struct xfrmu_spdinfo)) |
1350 | + nla_total_size(payload: sizeof(struct xfrmu_spdhinfo)) |
1351 | + nla_total_size(payload: sizeof(struct xfrmu_spdhthresh)) |
1352 | + nla_total_size(payload: sizeof(struct xfrmu_spdhthresh)); |
1353 | } |
1354 | |
1355 | static int build_spdinfo(struct sk_buff *skb, struct net *net, |
1356 | u32 portid, u32 seq, u32 flags) |
1357 | { |
1358 | struct xfrmk_spdinfo si; |
1359 | struct xfrmu_spdinfo spc; |
1360 | struct xfrmu_spdhinfo sph; |
1361 | struct xfrmu_spdhthresh spt4, spt6; |
1362 | struct nlmsghdr *nlh; |
1363 | int err; |
1364 | u32 *f; |
1365 | unsigned lseq; |
1366 | |
1367 | nlh = nlmsg_put(skb, portid, seq, XFRM_MSG_NEWSPDINFO, payload: sizeof(u32), flags: 0); |
1368 | if (nlh == NULL) /* shouldn't really happen ... */ |
1369 | return -EMSGSIZE; |
1370 | |
1371 | f = nlmsg_data(nlh); |
1372 | *f = flags; |
1373 | xfrm_spd_getinfo(net, si: &si); |
1374 | spc.incnt = si.incnt; |
1375 | spc.outcnt = si.outcnt; |
1376 | spc.fwdcnt = si.fwdcnt; |
1377 | spc.inscnt = si.inscnt; |
1378 | spc.outscnt = si.outscnt; |
1379 | spc.fwdscnt = si.fwdscnt; |
1380 | sph.spdhcnt = si.spdhcnt; |
1381 | sph.spdhmcnt = si.spdhmcnt; |
1382 | |
1383 | do { |
1384 | lseq = read_seqbegin(sl: &net->xfrm.policy_hthresh.lock); |
1385 | |
1386 | spt4.lbits = net->xfrm.policy_hthresh.lbits4; |
1387 | spt4.rbits = net->xfrm.policy_hthresh.rbits4; |
1388 | spt6.lbits = net->xfrm.policy_hthresh.lbits6; |
1389 | spt6.rbits = net->xfrm.policy_hthresh.rbits6; |
1390 | } while (read_seqretry(sl: &net->xfrm.policy_hthresh.lock, start: lseq)); |
1391 | |
1392 | err = nla_put(skb, attrtype: XFRMA_SPD_INFO, attrlen: sizeof(spc), data: &spc); |
1393 | if (!err) |
1394 | err = nla_put(skb, attrtype: XFRMA_SPD_HINFO, attrlen: sizeof(sph), data: &sph); |
1395 | if (!err) |
1396 | err = nla_put(skb, attrtype: XFRMA_SPD_IPV4_HTHRESH, attrlen: sizeof(spt4), data: &spt4); |
1397 | if (!err) |
1398 | err = nla_put(skb, attrtype: XFRMA_SPD_IPV6_HTHRESH, attrlen: sizeof(spt6), data: &spt6); |
1399 | if (err) { |
1400 | nlmsg_cancel(skb, nlh); |
1401 | return err; |
1402 | } |
1403 | |
1404 | nlmsg_end(skb, nlh); |
1405 | return 0; |
1406 | } |
1407 | |
1408 | static int xfrm_set_spdinfo(struct sk_buff *skb, struct nlmsghdr *nlh, |
1409 | struct nlattr **attrs, |
1410 | struct netlink_ext_ack *extack) |
1411 | { |
1412 | struct net *net = sock_net(sk: skb->sk); |
1413 | struct xfrmu_spdhthresh *thresh4 = NULL; |
1414 | struct xfrmu_spdhthresh *thresh6 = NULL; |
1415 | |
1416 | /* selector prefixlen thresholds to hash policies */ |
1417 | if (attrs[XFRMA_SPD_IPV4_HTHRESH]) { |
1418 | struct nlattr *rta = attrs[XFRMA_SPD_IPV4_HTHRESH]; |
1419 | |
1420 | if (nla_len(nla: rta) < sizeof(*thresh4)) { |
1421 | NL_SET_ERR_MSG(extack, "Invalid SPD_IPV4_HTHRESH attribute length" ); |
1422 | return -EINVAL; |
1423 | } |
1424 | thresh4 = nla_data(nla: rta); |
1425 | if (thresh4->lbits > 32 || thresh4->rbits > 32) { |
1426 | NL_SET_ERR_MSG(extack, "Invalid hash threshold (must be <= 32 for IPv4)" ); |
1427 | return -EINVAL; |
1428 | } |
1429 | } |
1430 | if (attrs[XFRMA_SPD_IPV6_HTHRESH]) { |
1431 | struct nlattr *rta = attrs[XFRMA_SPD_IPV6_HTHRESH]; |
1432 | |
1433 | if (nla_len(nla: rta) < sizeof(*thresh6)) { |
1434 | NL_SET_ERR_MSG(extack, "Invalid SPD_IPV6_HTHRESH attribute length" ); |
1435 | return -EINVAL; |
1436 | } |
1437 | thresh6 = nla_data(nla: rta); |
1438 | if (thresh6->lbits > 128 || thresh6->rbits > 128) { |
1439 | NL_SET_ERR_MSG(extack, "Invalid hash threshold (must be <= 128 for IPv6)" ); |
1440 | return -EINVAL; |
1441 | } |
1442 | } |
1443 | |
1444 | if (thresh4 || thresh6) { |
1445 | write_seqlock(sl: &net->xfrm.policy_hthresh.lock); |
1446 | if (thresh4) { |
1447 | net->xfrm.policy_hthresh.lbits4 = thresh4->lbits; |
1448 | net->xfrm.policy_hthresh.rbits4 = thresh4->rbits; |
1449 | } |
1450 | if (thresh6) { |
1451 | net->xfrm.policy_hthresh.lbits6 = thresh6->lbits; |
1452 | net->xfrm.policy_hthresh.rbits6 = thresh6->rbits; |
1453 | } |
1454 | write_sequnlock(sl: &net->xfrm.policy_hthresh.lock); |
1455 | |
1456 | xfrm_policy_hash_rebuild(net); |
1457 | } |
1458 | |
1459 | return 0; |
1460 | } |
1461 | |
1462 | static int xfrm_get_spdinfo(struct sk_buff *skb, struct nlmsghdr *nlh, |
1463 | struct nlattr **attrs, |
1464 | struct netlink_ext_ack *extack) |
1465 | { |
1466 | struct net *net = sock_net(sk: skb->sk); |
1467 | struct sk_buff *r_skb; |
1468 | u32 *flags = nlmsg_data(nlh); |
1469 | u32 sportid = NETLINK_CB(skb).portid; |
1470 | u32 seq = nlh->nlmsg_seq; |
1471 | int err; |
1472 | |
1473 | r_skb = nlmsg_new(payload: xfrm_spdinfo_msgsize(), GFP_ATOMIC); |
1474 | if (r_skb == NULL) |
1475 | return -ENOMEM; |
1476 | |
1477 | err = build_spdinfo(skb: r_skb, net, portid: sportid, seq, flags: *flags); |
1478 | BUG_ON(err < 0); |
1479 | |
1480 | return nlmsg_unicast(sk: net->xfrm.nlsk, skb: r_skb, portid: sportid); |
1481 | } |
1482 | |
1483 | static inline unsigned int xfrm_sadinfo_msgsize(void) |
1484 | { |
1485 | return NLMSG_ALIGN(4) |
1486 | + nla_total_size(payload: sizeof(struct xfrmu_sadhinfo)) |
1487 | + nla_total_size(payload: 4); /* XFRMA_SAD_CNT */ |
1488 | } |
1489 | |
1490 | static int build_sadinfo(struct sk_buff *skb, struct net *net, |
1491 | u32 portid, u32 seq, u32 flags) |
1492 | { |
1493 | struct xfrmk_sadinfo si; |
1494 | struct xfrmu_sadhinfo sh; |
1495 | struct nlmsghdr *nlh; |
1496 | int err; |
1497 | u32 *f; |
1498 | |
1499 | nlh = nlmsg_put(skb, portid, seq, XFRM_MSG_NEWSADINFO, payload: sizeof(u32), flags: 0); |
1500 | if (nlh == NULL) /* shouldn't really happen ... */ |
1501 | return -EMSGSIZE; |
1502 | |
1503 | f = nlmsg_data(nlh); |
1504 | *f = flags; |
1505 | xfrm_sad_getinfo(net, si: &si); |
1506 | |
1507 | sh.sadhmcnt = si.sadhmcnt; |
1508 | sh.sadhcnt = si.sadhcnt; |
1509 | |
1510 | err = nla_put_u32(skb, attrtype: XFRMA_SAD_CNT, value: si.sadcnt); |
1511 | if (!err) |
1512 | err = nla_put(skb, attrtype: XFRMA_SAD_HINFO, attrlen: sizeof(sh), data: &sh); |
1513 | if (err) { |
1514 | nlmsg_cancel(skb, nlh); |
1515 | return err; |
1516 | } |
1517 | |
1518 | nlmsg_end(skb, nlh); |
1519 | return 0; |
1520 | } |
1521 | |
1522 | static int xfrm_get_sadinfo(struct sk_buff *skb, struct nlmsghdr *nlh, |
1523 | struct nlattr **attrs, |
1524 | struct netlink_ext_ack *extack) |
1525 | { |
1526 | struct net *net = sock_net(sk: skb->sk); |
1527 | struct sk_buff *r_skb; |
1528 | u32 *flags = nlmsg_data(nlh); |
1529 | u32 sportid = NETLINK_CB(skb).portid; |
1530 | u32 seq = nlh->nlmsg_seq; |
1531 | int err; |
1532 | |
1533 | r_skb = nlmsg_new(payload: xfrm_sadinfo_msgsize(), GFP_ATOMIC); |
1534 | if (r_skb == NULL) |
1535 | return -ENOMEM; |
1536 | |
1537 | err = build_sadinfo(skb: r_skb, net, portid: sportid, seq, flags: *flags); |
1538 | BUG_ON(err < 0); |
1539 | |
1540 | return nlmsg_unicast(sk: net->xfrm.nlsk, skb: r_skb, portid: sportid); |
1541 | } |
1542 | |
1543 | static int xfrm_get_sa(struct sk_buff *skb, struct nlmsghdr *nlh, |
1544 | struct nlattr **attrs, struct netlink_ext_ack *extack) |
1545 | { |
1546 | struct net *net = sock_net(sk: skb->sk); |
1547 | struct xfrm_usersa_id *p = nlmsg_data(nlh); |
1548 | struct xfrm_state *x; |
1549 | struct sk_buff *resp_skb; |
1550 | int err = -ESRCH; |
1551 | |
1552 | x = xfrm_user_state_lookup(net, p, attrs, errp: &err); |
1553 | if (x == NULL) |
1554 | goto out_noput; |
1555 | |
1556 | resp_skb = xfrm_state_netlink(in_skb: skb, x, seq: nlh->nlmsg_seq); |
1557 | if (IS_ERR(ptr: resp_skb)) { |
1558 | err = PTR_ERR(ptr: resp_skb); |
1559 | } else { |
1560 | err = nlmsg_unicast(sk: net->xfrm.nlsk, skb: resp_skb, NETLINK_CB(skb).portid); |
1561 | } |
1562 | xfrm_state_put(x); |
1563 | out_noput: |
1564 | return err; |
1565 | } |
1566 | |
1567 | static int xfrm_alloc_userspi(struct sk_buff *skb, struct nlmsghdr *nlh, |
1568 | struct nlattr **attrs, |
1569 | struct netlink_ext_ack *extack) |
1570 | { |
1571 | struct net *net = sock_net(sk: skb->sk); |
1572 | struct xfrm_state *x; |
1573 | struct xfrm_userspi_info *p; |
1574 | struct xfrm_translator *xtr; |
1575 | struct sk_buff *resp_skb; |
1576 | xfrm_address_t *daddr; |
1577 | int family; |
1578 | int err; |
1579 | u32 mark; |
1580 | struct xfrm_mark m; |
1581 | u32 if_id = 0; |
1582 | |
1583 | p = nlmsg_data(nlh); |
1584 | err = verify_spi_info(proto: p->info.id.proto, min: p->min, max: p->max, extack); |
1585 | if (err) |
1586 | goto out_noput; |
1587 | |
1588 | family = p->info.family; |
1589 | daddr = &p->info.id.daddr; |
1590 | |
1591 | x = NULL; |
1592 | |
1593 | mark = xfrm_mark_get(attrs, m: &m); |
1594 | |
1595 | if (attrs[XFRMA_IF_ID]) |
1596 | if_id = nla_get_u32(nla: attrs[XFRMA_IF_ID]); |
1597 | |
1598 | if (p->info.seq) { |
1599 | x = xfrm_find_acq_byseq(net, mark, seq: p->info.seq); |
1600 | if (x && !xfrm_addr_equal(a: &x->id.daddr, b: daddr, family)) { |
1601 | xfrm_state_put(x); |
1602 | x = NULL; |
1603 | } |
1604 | } |
1605 | |
1606 | if (!x) |
1607 | x = xfrm_find_acq(net, mark: &m, mode: p->info.mode, reqid: p->info.reqid, |
1608 | if_id, proto: p->info.id.proto, daddr, |
1609 | saddr: &p->info.saddr, create: 1, |
1610 | family); |
1611 | err = -ENOENT; |
1612 | if (!x) { |
1613 | NL_SET_ERR_MSG(extack, "Target ACQUIRE not found" ); |
1614 | goto out_noput; |
1615 | } |
1616 | |
1617 | err = xfrm_alloc_spi(x, minspi: p->min, maxspi: p->max, extack); |
1618 | if (err) |
1619 | goto out; |
1620 | |
1621 | resp_skb = xfrm_state_netlink(in_skb: skb, x, seq: nlh->nlmsg_seq); |
1622 | if (IS_ERR(ptr: resp_skb)) { |
1623 | err = PTR_ERR(ptr: resp_skb); |
1624 | goto out; |
1625 | } |
1626 | |
1627 | xtr = xfrm_get_translator(); |
1628 | if (xtr) { |
1629 | err = xtr->alloc_compat(skb, nlmsg_hdr(skb)); |
1630 | |
1631 | xfrm_put_translator(xtr); |
1632 | if (err) { |
1633 | kfree_skb(skb: resp_skb); |
1634 | goto out; |
1635 | } |
1636 | } |
1637 | |
1638 | err = nlmsg_unicast(sk: net->xfrm.nlsk, skb: resp_skb, NETLINK_CB(skb).portid); |
1639 | |
1640 | out: |
1641 | xfrm_state_put(x); |
1642 | out_noput: |
1643 | return err; |
1644 | } |
1645 | |
1646 | static int verify_policy_dir(u8 dir, struct netlink_ext_ack *extack) |
1647 | { |
1648 | switch (dir) { |
1649 | case XFRM_POLICY_IN: |
1650 | case XFRM_POLICY_OUT: |
1651 | case XFRM_POLICY_FWD: |
1652 | break; |
1653 | |
1654 | default: |
1655 | NL_SET_ERR_MSG(extack, "Invalid policy direction" ); |
1656 | return -EINVAL; |
1657 | } |
1658 | |
1659 | return 0; |
1660 | } |
1661 | |
1662 | static int verify_policy_type(u8 type, struct netlink_ext_ack *extack) |
1663 | { |
1664 | switch (type) { |
1665 | case XFRM_POLICY_TYPE_MAIN: |
1666 | #ifdef CONFIG_XFRM_SUB_POLICY |
1667 | case XFRM_POLICY_TYPE_SUB: |
1668 | #endif |
1669 | break; |
1670 | |
1671 | default: |
1672 | NL_SET_ERR_MSG(extack, "Invalid policy type" ); |
1673 | return -EINVAL; |
1674 | } |
1675 | |
1676 | return 0; |
1677 | } |
1678 | |
1679 | static int verify_newpolicy_info(struct xfrm_userpolicy_info *p, |
1680 | struct netlink_ext_ack *extack) |
1681 | { |
1682 | int ret; |
1683 | |
1684 | switch (p->share) { |
1685 | case XFRM_SHARE_ANY: |
1686 | case XFRM_SHARE_SESSION: |
1687 | case XFRM_SHARE_USER: |
1688 | case XFRM_SHARE_UNIQUE: |
1689 | break; |
1690 | |
1691 | default: |
1692 | NL_SET_ERR_MSG(extack, "Invalid policy share" ); |
1693 | return -EINVAL; |
1694 | } |
1695 | |
1696 | switch (p->action) { |
1697 | case XFRM_POLICY_ALLOW: |
1698 | case XFRM_POLICY_BLOCK: |
1699 | break; |
1700 | |
1701 | default: |
1702 | NL_SET_ERR_MSG(extack, "Invalid policy action" ); |
1703 | return -EINVAL; |
1704 | } |
1705 | |
1706 | switch (p->sel.family) { |
1707 | case AF_INET: |
1708 | if (p->sel.prefixlen_d > 32 || p->sel.prefixlen_s > 32) { |
1709 | NL_SET_ERR_MSG(extack, "Invalid prefix length in selector (must be <= 32 for IPv4)" ); |
1710 | return -EINVAL; |
1711 | } |
1712 | |
1713 | break; |
1714 | |
1715 | case AF_INET6: |
1716 | #if IS_ENABLED(CONFIG_IPV6) |
1717 | if (p->sel.prefixlen_d > 128 || p->sel.prefixlen_s > 128) { |
1718 | NL_SET_ERR_MSG(extack, "Invalid prefix length in selector (must be <= 128 for IPv6)" ); |
1719 | return -EINVAL; |
1720 | } |
1721 | |
1722 | break; |
1723 | #else |
1724 | NL_SET_ERR_MSG(extack, "IPv6 support disabled" ); |
1725 | return -EAFNOSUPPORT; |
1726 | #endif |
1727 | |
1728 | default: |
1729 | NL_SET_ERR_MSG(extack, "Invalid selector family" ); |
1730 | return -EINVAL; |
1731 | } |
1732 | |
1733 | ret = verify_policy_dir(dir: p->dir, extack); |
1734 | if (ret) |
1735 | return ret; |
1736 | if (p->index && (xfrm_policy_id2dir(index: p->index) != p->dir)) { |
1737 | NL_SET_ERR_MSG(extack, "Policy index doesn't match direction" ); |
1738 | return -EINVAL; |
1739 | } |
1740 | |
1741 | return 0; |
1742 | } |
1743 | |
1744 | static int copy_from_user_sec_ctx(struct xfrm_policy *pol, struct nlattr **attrs) |
1745 | { |
1746 | struct nlattr *rt = attrs[XFRMA_SEC_CTX]; |
1747 | struct xfrm_user_sec_ctx *uctx; |
1748 | |
1749 | if (!rt) |
1750 | return 0; |
1751 | |
1752 | uctx = nla_data(nla: rt); |
1753 | return security_xfrm_policy_alloc(ctxp: &pol->security, sec_ctx: uctx, GFP_KERNEL); |
1754 | } |
1755 | |
1756 | static void copy_templates(struct xfrm_policy *xp, struct xfrm_user_tmpl *ut, |
1757 | int nr) |
1758 | { |
1759 | int i; |
1760 | |
1761 | xp->xfrm_nr = nr; |
1762 | for (i = 0; i < nr; i++, ut++) { |
1763 | struct xfrm_tmpl *t = &xp->xfrm_vec[i]; |
1764 | |
1765 | memcpy(&t->id, &ut->id, sizeof(struct xfrm_id)); |
1766 | memcpy(&t->saddr, &ut->saddr, |
1767 | sizeof(xfrm_address_t)); |
1768 | t->reqid = ut->reqid; |
1769 | t->mode = ut->mode; |
1770 | t->share = ut->share; |
1771 | t->optional = ut->optional; |
1772 | t->aalgos = ut->aalgos; |
1773 | t->ealgos = ut->ealgos; |
1774 | t->calgos = ut->calgos; |
1775 | /* If all masks are ~0, then we allow all algorithms. */ |
1776 | t->allalgs = !~(t->aalgos & t->ealgos & t->calgos); |
1777 | t->encap_family = ut->family; |
1778 | } |
1779 | } |
1780 | |
1781 | static int validate_tmpl(int nr, struct xfrm_user_tmpl *ut, u16 family, |
1782 | int dir, struct netlink_ext_ack *extack) |
1783 | { |
1784 | u16 prev_family; |
1785 | int i; |
1786 | |
1787 | if (nr > XFRM_MAX_DEPTH) { |
1788 | NL_SET_ERR_MSG(extack, "Template count must be <= XFRM_MAX_DEPTH (" __stringify(XFRM_MAX_DEPTH) ")" ); |
1789 | return -EINVAL; |
1790 | } |
1791 | |
1792 | prev_family = family; |
1793 | |
1794 | for (i = 0; i < nr; i++) { |
1795 | /* We never validated the ut->family value, so many |
1796 | * applications simply leave it at zero. The check was |
1797 | * never made and ut->family was ignored because all |
1798 | * templates could be assumed to have the same family as |
1799 | * the policy itself. Now that we will have ipv4-in-ipv6 |
1800 | * and ipv6-in-ipv4 tunnels, this is no longer true. |
1801 | */ |
1802 | if (!ut[i].family) |
1803 | ut[i].family = family; |
1804 | |
1805 | switch (ut[i].mode) { |
1806 | case XFRM_MODE_TUNNEL: |
1807 | case XFRM_MODE_BEET: |
1808 | if (ut[i].optional && dir == XFRM_POLICY_OUT) { |
1809 | NL_SET_ERR_MSG(extack, "Mode in optional template not allowed in outbound policy" ); |
1810 | return -EINVAL; |
1811 | } |
1812 | break; |
1813 | default: |
1814 | if (ut[i].family != prev_family) { |
1815 | NL_SET_ERR_MSG(extack, "Mode in template doesn't support a family change" ); |
1816 | return -EINVAL; |
1817 | } |
1818 | break; |
1819 | } |
1820 | if (ut[i].mode >= XFRM_MODE_MAX) { |
1821 | NL_SET_ERR_MSG(extack, "Mode in template must be < XFRM_MODE_MAX (" __stringify(XFRM_MODE_MAX) ")" ); |
1822 | return -EINVAL; |
1823 | } |
1824 | |
1825 | prev_family = ut[i].family; |
1826 | |
1827 | switch (ut[i].family) { |
1828 | case AF_INET: |
1829 | break; |
1830 | #if IS_ENABLED(CONFIG_IPV6) |
1831 | case AF_INET6: |
1832 | break; |
1833 | #endif |
1834 | default: |
1835 | NL_SET_ERR_MSG(extack, "Invalid family in template" ); |
1836 | return -EINVAL; |
1837 | } |
1838 | |
1839 | if (!xfrm_id_proto_valid(proto: ut[i].id.proto)) { |
1840 | NL_SET_ERR_MSG(extack, "Invalid XFRM protocol in template" ); |
1841 | return -EINVAL; |
1842 | } |
1843 | } |
1844 | |
1845 | return 0; |
1846 | } |
1847 | |
1848 | static int copy_from_user_tmpl(struct xfrm_policy *pol, struct nlattr **attrs, |
1849 | int dir, struct netlink_ext_ack *extack) |
1850 | { |
1851 | struct nlattr *rt = attrs[XFRMA_TMPL]; |
1852 | |
1853 | if (!rt) { |
1854 | pol->xfrm_nr = 0; |
1855 | } else { |
1856 | struct xfrm_user_tmpl *utmpl = nla_data(nla: rt); |
1857 | int nr = nla_len(nla: rt) / sizeof(*utmpl); |
1858 | int err; |
1859 | |
1860 | err = validate_tmpl(nr, ut: utmpl, family: pol->family, dir, extack); |
1861 | if (err) |
1862 | return err; |
1863 | |
1864 | copy_templates(xp: pol, ut: utmpl, nr); |
1865 | } |
1866 | return 0; |
1867 | } |
1868 | |
1869 | static int copy_from_user_policy_type(u8 *tp, struct nlattr **attrs, |
1870 | struct netlink_ext_ack *extack) |
1871 | { |
1872 | struct nlattr *rt = attrs[XFRMA_POLICY_TYPE]; |
1873 | struct xfrm_userpolicy_type *upt; |
1874 | u8 type = XFRM_POLICY_TYPE_MAIN; |
1875 | int err; |
1876 | |
1877 | if (rt) { |
1878 | upt = nla_data(nla: rt); |
1879 | type = upt->type; |
1880 | } |
1881 | |
1882 | err = verify_policy_type(type, extack); |
1883 | if (err) |
1884 | return err; |
1885 | |
1886 | *tp = type; |
1887 | return 0; |
1888 | } |
1889 | |
1890 | static void copy_from_user_policy(struct xfrm_policy *xp, struct xfrm_userpolicy_info *p) |
1891 | { |
1892 | xp->priority = p->priority; |
1893 | xp->index = p->index; |
1894 | memcpy(&xp->selector, &p->sel, sizeof(xp->selector)); |
1895 | memcpy(&xp->lft, &p->lft, sizeof(xp->lft)); |
1896 | xp->action = p->action; |
1897 | xp->flags = p->flags; |
1898 | xp->family = p->sel.family; |
1899 | /* XXX xp->share = p->share; */ |
1900 | } |
1901 | |
1902 | static void copy_to_user_policy(struct xfrm_policy *xp, struct xfrm_userpolicy_info *p, int dir) |
1903 | { |
1904 | memset(p, 0, sizeof(*p)); |
1905 | memcpy(&p->sel, &xp->selector, sizeof(p->sel)); |
1906 | memcpy(&p->lft, &xp->lft, sizeof(p->lft)); |
1907 | memcpy(&p->curlft, &xp->curlft, sizeof(p->curlft)); |
1908 | p->priority = xp->priority; |
1909 | p->index = xp->index; |
1910 | p->sel.family = xp->family; |
1911 | p->dir = dir; |
1912 | p->action = xp->action; |
1913 | p->flags = xp->flags; |
1914 | p->share = XFRM_SHARE_ANY; /* XXX xp->share */ |
1915 | } |
1916 | |
1917 | static struct xfrm_policy *xfrm_policy_construct(struct net *net, |
1918 | struct xfrm_userpolicy_info *p, |
1919 | struct nlattr **attrs, |
1920 | int *errp, |
1921 | struct netlink_ext_ack *extack) |
1922 | { |
1923 | struct xfrm_policy *xp = xfrm_policy_alloc(net, GFP_KERNEL); |
1924 | int err; |
1925 | |
1926 | if (!xp) { |
1927 | *errp = -ENOMEM; |
1928 | return NULL; |
1929 | } |
1930 | |
1931 | copy_from_user_policy(xp, p); |
1932 | |
1933 | err = copy_from_user_policy_type(tp: &xp->type, attrs, extack); |
1934 | if (err) |
1935 | goto error; |
1936 | |
1937 | if (!(err = copy_from_user_tmpl(pol: xp, attrs, dir: p->dir, extack))) |
1938 | err = copy_from_user_sec_ctx(pol: xp, attrs); |
1939 | if (err) |
1940 | goto error; |
1941 | |
1942 | xfrm_mark_get(attrs, m: &xp->mark); |
1943 | |
1944 | if (attrs[XFRMA_IF_ID]) |
1945 | xp->if_id = nla_get_u32(nla: attrs[XFRMA_IF_ID]); |
1946 | |
1947 | /* configure the hardware if offload is requested */ |
1948 | if (attrs[XFRMA_OFFLOAD_DEV]) { |
1949 | err = xfrm_dev_policy_add(net, xp, |
1950 | xuo: nla_data(nla: attrs[XFRMA_OFFLOAD_DEV]), |
1951 | dir: p->dir, extack); |
1952 | if (err) |
1953 | goto error; |
1954 | } |
1955 | |
1956 | return xp; |
1957 | error: |
1958 | *errp = err; |
1959 | xp->walk.dead = 1; |
1960 | xfrm_policy_destroy(policy: xp); |
1961 | return NULL; |
1962 | } |
1963 | |
1964 | static int xfrm_add_policy(struct sk_buff *skb, struct nlmsghdr *nlh, |
1965 | struct nlattr **attrs, |
1966 | struct netlink_ext_ack *extack) |
1967 | { |
1968 | struct net *net = sock_net(sk: skb->sk); |
1969 | struct xfrm_userpolicy_info *p = nlmsg_data(nlh); |
1970 | struct xfrm_policy *xp; |
1971 | struct km_event c; |
1972 | int err; |
1973 | int excl; |
1974 | |
1975 | err = verify_newpolicy_info(p, extack); |
1976 | if (err) |
1977 | return err; |
1978 | err = verify_sec_ctx_len(attrs, extack); |
1979 | if (err) |
1980 | return err; |
1981 | |
1982 | xp = xfrm_policy_construct(net, p, attrs, errp: &err, extack); |
1983 | if (!xp) |
1984 | return err; |
1985 | |
1986 | /* shouldn't excl be based on nlh flags?? |
1987 | * Aha! this is anti-netlink really i.e more pfkey derived |
1988 | * in netlink excl is a flag and you wouldn't need |
1989 | * a type XFRM_MSG_UPDPOLICY - JHS */ |
1990 | excl = nlh->nlmsg_type == XFRM_MSG_NEWPOLICY; |
1991 | err = xfrm_policy_insert(dir: p->dir, policy: xp, excl); |
1992 | xfrm_audit_policy_add(xp, result: err ? 0 : 1, task_valid: true); |
1993 | |
1994 | if (err) { |
1995 | xfrm_dev_policy_delete(x: xp); |
1996 | xfrm_dev_policy_free(x: xp); |
1997 | security_xfrm_policy_free(ctx: xp->security); |
1998 | kfree(objp: xp); |
1999 | return err; |
2000 | } |
2001 | |
2002 | c.event = nlh->nlmsg_type; |
2003 | c.seq = nlh->nlmsg_seq; |
2004 | c.portid = nlh->nlmsg_pid; |
2005 | km_policy_notify(xp, dir: p->dir, c: &c); |
2006 | |
2007 | xfrm_pol_put(policy: xp); |
2008 | |
2009 | return 0; |
2010 | } |
2011 | |
2012 | static int copy_to_user_tmpl(struct xfrm_policy *xp, struct sk_buff *skb) |
2013 | { |
2014 | struct xfrm_user_tmpl vec[XFRM_MAX_DEPTH]; |
2015 | int i; |
2016 | |
2017 | if (xp->xfrm_nr == 0) |
2018 | return 0; |
2019 | |
2020 | if (xp->xfrm_nr > XFRM_MAX_DEPTH) |
2021 | return -ENOBUFS; |
2022 | |
2023 | for (i = 0; i < xp->xfrm_nr; i++) { |
2024 | struct xfrm_user_tmpl *up = &vec[i]; |
2025 | struct xfrm_tmpl *kp = &xp->xfrm_vec[i]; |
2026 | |
2027 | memset(up, 0, sizeof(*up)); |
2028 | memcpy(&up->id, &kp->id, sizeof(up->id)); |
2029 | up->family = kp->encap_family; |
2030 | memcpy(&up->saddr, &kp->saddr, sizeof(up->saddr)); |
2031 | up->reqid = kp->reqid; |
2032 | up->mode = kp->mode; |
2033 | up->share = kp->share; |
2034 | up->optional = kp->optional; |
2035 | up->aalgos = kp->aalgos; |
2036 | up->ealgos = kp->ealgos; |
2037 | up->calgos = kp->calgos; |
2038 | } |
2039 | |
2040 | return nla_put(skb, attrtype: XFRMA_TMPL, |
2041 | attrlen: sizeof(struct xfrm_user_tmpl) * xp->xfrm_nr, data: vec); |
2042 | } |
2043 | |
2044 | static inline int copy_to_user_state_sec_ctx(struct xfrm_state *x, struct sk_buff *skb) |
2045 | { |
2046 | if (x->security) { |
2047 | return copy_sec_ctx(s: x->security, skb); |
2048 | } |
2049 | return 0; |
2050 | } |
2051 | |
2052 | static inline int copy_to_user_sec_ctx(struct xfrm_policy *xp, struct sk_buff *skb) |
2053 | { |
2054 | if (xp->security) |
2055 | return copy_sec_ctx(s: xp->security, skb); |
2056 | return 0; |
2057 | } |
2058 | static inline unsigned int userpolicy_type_attrsize(void) |
2059 | { |
2060 | #ifdef CONFIG_XFRM_SUB_POLICY |
2061 | return nla_total_size(payload: sizeof(struct xfrm_userpolicy_type)); |
2062 | #else |
2063 | return 0; |
2064 | #endif |
2065 | } |
2066 | |
2067 | #ifdef CONFIG_XFRM_SUB_POLICY |
2068 | static int copy_to_user_policy_type(u8 type, struct sk_buff *skb) |
2069 | { |
2070 | struct xfrm_userpolicy_type upt; |
2071 | |
2072 | /* Sadly there are two holes in struct xfrm_userpolicy_type */ |
2073 | memset(&upt, 0, sizeof(upt)); |
2074 | upt.type = type; |
2075 | |
2076 | return nla_put(skb, attrtype: XFRMA_POLICY_TYPE, attrlen: sizeof(upt), data: &upt); |
2077 | } |
2078 | |
2079 | #else |
2080 | static inline int copy_to_user_policy_type(u8 type, struct sk_buff *skb) |
2081 | { |
2082 | return 0; |
2083 | } |
2084 | #endif |
2085 | |
2086 | static int dump_one_policy(struct xfrm_policy *xp, int dir, int count, void *ptr) |
2087 | { |
2088 | struct xfrm_dump_info *sp = ptr; |
2089 | struct xfrm_userpolicy_info *p; |
2090 | struct sk_buff *in_skb = sp->in_skb; |
2091 | struct sk_buff *skb = sp->out_skb; |
2092 | struct xfrm_translator *xtr; |
2093 | struct nlmsghdr *nlh; |
2094 | int err; |
2095 | |
2096 | nlh = nlmsg_put(skb, NETLINK_CB(in_skb).portid, seq: sp->nlmsg_seq, |
2097 | XFRM_MSG_NEWPOLICY, payload: sizeof(*p), flags: sp->nlmsg_flags); |
2098 | if (nlh == NULL) |
2099 | return -EMSGSIZE; |
2100 | |
2101 | p = nlmsg_data(nlh); |
2102 | copy_to_user_policy(xp, p, dir); |
2103 | err = copy_to_user_tmpl(xp, skb); |
2104 | if (!err) |
2105 | err = copy_to_user_sec_ctx(xp, skb); |
2106 | if (!err) |
2107 | err = copy_to_user_policy_type(type: xp->type, skb); |
2108 | if (!err) |
2109 | err = xfrm_mark_put(skb, m: &xp->mark); |
2110 | if (!err) |
2111 | err = xfrm_if_id_put(skb, if_id: xp->if_id); |
2112 | if (!err && xp->xdo.dev) |
2113 | err = copy_user_offload(xso: &xp->xdo, skb); |
2114 | if (err) { |
2115 | nlmsg_cancel(skb, nlh); |
2116 | return err; |
2117 | } |
2118 | nlmsg_end(skb, nlh); |
2119 | |
2120 | xtr = xfrm_get_translator(); |
2121 | if (xtr) { |
2122 | err = xtr->alloc_compat(skb, nlh); |
2123 | |
2124 | xfrm_put_translator(xtr); |
2125 | if (err) { |
2126 | nlmsg_cancel(skb, nlh); |
2127 | return err; |
2128 | } |
2129 | } |
2130 | |
2131 | return 0; |
2132 | } |
2133 | |
2134 | static int xfrm_dump_policy_done(struct netlink_callback *cb) |
2135 | { |
2136 | struct xfrm_policy_walk *walk = (struct xfrm_policy_walk *)cb->args; |
2137 | struct net *net = sock_net(sk: cb->skb->sk); |
2138 | |
2139 | xfrm_policy_walk_done(walk, net); |
2140 | return 0; |
2141 | } |
2142 | |
2143 | static int xfrm_dump_policy_start(struct netlink_callback *cb) |
2144 | { |
2145 | struct xfrm_policy_walk *walk = (struct xfrm_policy_walk *)cb->args; |
2146 | |
2147 | BUILD_BUG_ON(sizeof(*walk) > sizeof(cb->args)); |
2148 | |
2149 | xfrm_policy_walk_init(walk, type: XFRM_POLICY_TYPE_ANY); |
2150 | return 0; |
2151 | } |
2152 | |
2153 | static int xfrm_dump_policy(struct sk_buff *skb, struct netlink_callback *cb) |
2154 | { |
2155 | struct net *net = sock_net(sk: skb->sk); |
2156 | struct xfrm_policy_walk *walk = (struct xfrm_policy_walk *)cb->args; |
2157 | struct xfrm_dump_info info; |
2158 | |
2159 | info.in_skb = cb->skb; |
2160 | info.out_skb = skb; |
2161 | info.nlmsg_seq = cb->nlh->nlmsg_seq; |
2162 | info.nlmsg_flags = NLM_F_MULTI; |
2163 | |
2164 | (void) xfrm_policy_walk(net, walk, func: dump_one_policy, &info); |
2165 | |
2166 | return skb->len; |
2167 | } |
2168 | |
2169 | static struct sk_buff *xfrm_policy_netlink(struct sk_buff *in_skb, |
2170 | struct xfrm_policy *xp, |
2171 | int dir, u32 seq) |
2172 | { |
2173 | struct xfrm_dump_info info; |
2174 | struct sk_buff *skb; |
2175 | int err; |
2176 | |
2177 | skb = nlmsg_new(NLMSG_DEFAULT_SIZE, GFP_KERNEL); |
2178 | if (!skb) |
2179 | return ERR_PTR(error: -ENOMEM); |
2180 | |
2181 | info.in_skb = in_skb; |
2182 | info.out_skb = skb; |
2183 | info.nlmsg_seq = seq; |
2184 | info.nlmsg_flags = 0; |
2185 | |
2186 | err = dump_one_policy(xp, dir, count: 0, ptr: &info); |
2187 | if (err) { |
2188 | kfree_skb(skb); |
2189 | return ERR_PTR(error: err); |
2190 | } |
2191 | |
2192 | return skb; |
2193 | } |
2194 | |
2195 | static int xfrm_notify_userpolicy(struct net *net) |
2196 | { |
2197 | struct xfrm_userpolicy_default *up; |
2198 | int len = NLMSG_ALIGN(sizeof(*up)); |
2199 | struct nlmsghdr *nlh; |
2200 | struct sk_buff *skb; |
2201 | int err; |
2202 | |
2203 | skb = nlmsg_new(payload: len, GFP_ATOMIC); |
2204 | if (skb == NULL) |
2205 | return -ENOMEM; |
2206 | |
2207 | nlh = nlmsg_put(skb, portid: 0, seq: 0, XFRM_MSG_GETDEFAULT, payload: sizeof(*up), flags: 0); |
2208 | if (nlh == NULL) { |
2209 | kfree_skb(skb); |
2210 | return -EMSGSIZE; |
2211 | } |
2212 | |
2213 | up = nlmsg_data(nlh); |
2214 | up->in = net->xfrm.policy_default[XFRM_POLICY_IN]; |
2215 | up->fwd = net->xfrm.policy_default[XFRM_POLICY_FWD]; |
2216 | up->out = net->xfrm.policy_default[XFRM_POLICY_OUT]; |
2217 | |
2218 | nlmsg_end(skb, nlh); |
2219 | |
2220 | rcu_read_lock(); |
2221 | err = xfrm_nlmsg_multicast(net, skb, pid: 0, XFRMNLGRP_POLICY); |
2222 | rcu_read_unlock(); |
2223 | |
2224 | return err; |
2225 | } |
2226 | |
2227 | static bool xfrm_userpolicy_is_valid(__u8 policy) |
2228 | { |
2229 | return policy == XFRM_USERPOLICY_BLOCK || |
2230 | policy == XFRM_USERPOLICY_ACCEPT; |
2231 | } |
2232 | |
2233 | static int xfrm_set_default(struct sk_buff *skb, struct nlmsghdr *nlh, |
2234 | struct nlattr **attrs, struct netlink_ext_ack *extack) |
2235 | { |
2236 | struct net *net = sock_net(sk: skb->sk); |
2237 | struct xfrm_userpolicy_default *up = nlmsg_data(nlh); |
2238 | |
2239 | if (xfrm_userpolicy_is_valid(policy: up->in)) |
2240 | net->xfrm.policy_default[XFRM_POLICY_IN] = up->in; |
2241 | |
2242 | if (xfrm_userpolicy_is_valid(policy: up->fwd)) |
2243 | net->xfrm.policy_default[XFRM_POLICY_FWD] = up->fwd; |
2244 | |
2245 | if (xfrm_userpolicy_is_valid(policy: up->out)) |
2246 | net->xfrm.policy_default[XFRM_POLICY_OUT] = up->out; |
2247 | |
2248 | rt_genid_bump_all(net); |
2249 | |
2250 | xfrm_notify_userpolicy(net); |
2251 | return 0; |
2252 | } |
2253 | |
2254 | static int xfrm_get_default(struct sk_buff *skb, struct nlmsghdr *nlh, |
2255 | struct nlattr **attrs, struct netlink_ext_ack *extack) |
2256 | { |
2257 | struct sk_buff *r_skb; |
2258 | struct nlmsghdr *r_nlh; |
2259 | struct net *net = sock_net(sk: skb->sk); |
2260 | struct xfrm_userpolicy_default *r_up; |
2261 | int len = NLMSG_ALIGN(sizeof(struct xfrm_userpolicy_default)); |
2262 | u32 portid = NETLINK_CB(skb).portid; |
2263 | u32 seq = nlh->nlmsg_seq; |
2264 | |
2265 | r_skb = nlmsg_new(payload: len, GFP_ATOMIC); |
2266 | if (!r_skb) |
2267 | return -ENOMEM; |
2268 | |
2269 | r_nlh = nlmsg_put(skb: r_skb, portid, seq, XFRM_MSG_GETDEFAULT, payload: sizeof(*r_up), flags: 0); |
2270 | if (!r_nlh) { |
2271 | kfree_skb(skb: r_skb); |
2272 | return -EMSGSIZE; |
2273 | } |
2274 | |
2275 | r_up = nlmsg_data(nlh: r_nlh); |
2276 | r_up->in = net->xfrm.policy_default[XFRM_POLICY_IN]; |
2277 | r_up->fwd = net->xfrm.policy_default[XFRM_POLICY_FWD]; |
2278 | r_up->out = net->xfrm.policy_default[XFRM_POLICY_OUT]; |
2279 | nlmsg_end(skb: r_skb, nlh: r_nlh); |
2280 | |
2281 | return nlmsg_unicast(sk: net->xfrm.nlsk, skb: r_skb, portid); |
2282 | } |
2283 | |
2284 | static int xfrm_get_policy(struct sk_buff *skb, struct nlmsghdr *nlh, |
2285 | struct nlattr **attrs, |
2286 | struct netlink_ext_ack *extack) |
2287 | { |
2288 | struct net *net = sock_net(sk: skb->sk); |
2289 | struct xfrm_policy *xp; |
2290 | struct xfrm_userpolicy_id *p; |
2291 | u8 type = XFRM_POLICY_TYPE_MAIN; |
2292 | int err; |
2293 | struct km_event c; |
2294 | int delete; |
2295 | struct xfrm_mark m; |
2296 | u32 if_id = 0; |
2297 | |
2298 | p = nlmsg_data(nlh); |
2299 | delete = nlh->nlmsg_type == XFRM_MSG_DELPOLICY; |
2300 | |
2301 | err = copy_from_user_policy_type(tp: &type, attrs, extack); |
2302 | if (err) |
2303 | return err; |
2304 | |
2305 | err = verify_policy_dir(dir: p->dir, extack); |
2306 | if (err) |
2307 | return err; |
2308 | |
2309 | if (attrs[XFRMA_IF_ID]) |
2310 | if_id = nla_get_u32(nla: attrs[XFRMA_IF_ID]); |
2311 | |
2312 | xfrm_mark_get(attrs, m: &m); |
2313 | |
2314 | if (p->index) |
2315 | xp = xfrm_policy_byid(net, mark: &m, if_id, type, dir: p->dir, |
2316 | id: p->index, delete, err: &err); |
2317 | else { |
2318 | struct nlattr *rt = attrs[XFRMA_SEC_CTX]; |
2319 | struct xfrm_sec_ctx *ctx; |
2320 | |
2321 | err = verify_sec_ctx_len(attrs, extack); |
2322 | if (err) |
2323 | return err; |
2324 | |
2325 | ctx = NULL; |
2326 | if (rt) { |
2327 | struct xfrm_user_sec_ctx *uctx = nla_data(nla: rt); |
2328 | |
2329 | err = security_xfrm_policy_alloc(ctxp: &ctx, sec_ctx: uctx, GFP_KERNEL); |
2330 | if (err) |
2331 | return err; |
2332 | } |
2333 | xp = xfrm_policy_bysel_ctx(net, mark: &m, if_id, type, dir: p->dir, |
2334 | sel: &p->sel, ctx, delete, err: &err); |
2335 | security_xfrm_policy_free(ctx); |
2336 | } |
2337 | if (xp == NULL) |
2338 | return -ENOENT; |
2339 | |
2340 | if (!delete) { |
2341 | struct sk_buff *resp_skb; |
2342 | |
2343 | resp_skb = xfrm_policy_netlink(in_skb: skb, xp, dir: p->dir, seq: nlh->nlmsg_seq); |
2344 | if (IS_ERR(ptr: resp_skb)) { |
2345 | err = PTR_ERR(ptr: resp_skb); |
2346 | } else { |
2347 | err = nlmsg_unicast(sk: net->xfrm.nlsk, skb: resp_skb, |
2348 | NETLINK_CB(skb).portid); |
2349 | } |
2350 | } else { |
2351 | xfrm_dev_policy_delete(x: xp); |
2352 | xfrm_audit_policy_delete(xp, result: err ? 0 : 1, task_valid: true); |
2353 | |
2354 | if (err != 0) |
2355 | goto out; |
2356 | |
2357 | c.data.byid = p->index; |
2358 | c.event = nlh->nlmsg_type; |
2359 | c.seq = nlh->nlmsg_seq; |
2360 | c.portid = nlh->nlmsg_pid; |
2361 | km_policy_notify(xp, dir: p->dir, c: &c); |
2362 | } |
2363 | |
2364 | out: |
2365 | xfrm_pol_put(policy: xp); |
2366 | return err; |
2367 | } |
2368 | |
2369 | static int xfrm_flush_sa(struct sk_buff *skb, struct nlmsghdr *nlh, |
2370 | struct nlattr **attrs, |
2371 | struct netlink_ext_ack *extack) |
2372 | { |
2373 | struct net *net = sock_net(sk: skb->sk); |
2374 | struct km_event c; |
2375 | struct xfrm_usersa_flush *p = nlmsg_data(nlh); |
2376 | int err; |
2377 | |
2378 | err = xfrm_state_flush(net, proto: p->proto, task_valid: true, sync: false); |
2379 | if (err) { |
2380 | if (err == -ESRCH) /* empty table */ |
2381 | return 0; |
2382 | return err; |
2383 | } |
2384 | c.data.proto = p->proto; |
2385 | c.event = nlh->nlmsg_type; |
2386 | c.seq = nlh->nlmsg_seq; |
2387 | c.portid = nlh->nlmsg_pid; |
2388 | c.net = net; |
2389 | km_state_notify(NULL, c: &c); |
2390 | |
2391 | return 0; |
2392 | } |
2393 | |
2394 | static inline unsigned int xfrm_aevent_msgsize(struct xfrm_state *x) |
2395 | { |
2396 | unsigned int replay_size = x->replay_esn ? |
2397 | xfrm_replay_state_esn_len(replay_esn: x->replay_esn) : |
2398 | sizeof(struct xfrm_replay_state); |
2399 | |
2400 | return NLMSG_ALIGN(sizeof(struct xfrm_aevent_id)) |
2401 | + nla_total_size(payload: replay_size) |
2402 | + nla_total_size_64bit(payload: sizeof(struct xfrm_lifetime_cur)) |
2403 | + nla_total_size(payload: sizeof(struct xfrm_mark)) |
2404 | + nla_total_size(payload: 4) /* XFRM_AE_RTHR */ |
2405 | + nla_total_size(payload: 4); /* XFRM_AE_ETHR */ |
2406 | } |
2407 | |
2408 | static int build_aevent(struct sk_buff *skb, struct xfrm_state *x, const struct km_event *c) |
2409 | { |
2410 | struct xfrm_aevent_id *id; |
2411 | struct nlmsghdr *nlh; |
2412 | int err; |
2413 | |
2414 | nlh = nlmsg_put(skb, portid: c->portid, seq: c->seq, XFRM_MSG_NEWAE, payload: sizeof(*id), flags: 0); |
2415 | if (nlh == NULL) |
2416 | return -EMSGSIZE; |
2417 | |
2418 | id = nlmsg_data(nlh); |
2419 | memset(&id->sa_id, 0, sizeof(id->sa_id)); |
2420 | memcpy(&id->sa_id.daddr, &x->id.daddr, sizeof(x->id.daddr)); |
2421 | id->sa_id.spi = x->id.spi; |
2422 | id->sa_id.family = x->props.family; |
2423 | id->sa_id.proto = x->id.proto; |
2424 | memcpy(&id->saddr, &x->props.saddr, sizeof(x->props.saddr)); |
2425 | id->reqid = x->props.reqid; |
2426 | id->flags = c->data.aevent; |
2427 | |
2428 | if (x->replay_esn) { |
2429 | err = nla_put(skb, attrtype: XFRMA_REPLAY_ESN_VAL, |
2430 | attrlen: xfrm_replay_state_esn_len(replay_esn: x->replay_esn), |
2431 | data: x->replay_esn); |
2432 | } else { |
2433 | err = nla_put(skb, attrtype: XFRMA_REPLAY_VAL, attrlen: sizeof(x->replay), |
2434 | data: &x->replay); |
2435 | } |
2436 | if (err) |
2437 | goto out_cancel; |
2438 | err = nla_put_64bit(skb, attrtype: XFRMA_LTIME_VAL, attrlen: sizeof(x->curlft), data: &x->curlft, |
2439 | padattr: XFRMA_PAD); |
2440 | if (err) |
2441 | goto out_cancel; |
2442 | |
2443 | if (id->flags & XFRM_AE_RTHR) { |
2444 | err = nla_put_u32(skb, attrtype: XFRMA_REPLAY_THRESH, value: x->replay_maxdiff); |
2445 | if (err) |
2446 | goto out_cancel; |
2447 | } |
2448 | if (id->flags & XFRM_AE_ETHR) { |
2449 | err = nla_put_u32(skb, attrtype: XFRMA_ETIMER_THRESH, |
2450 | value: x->replay_maxage * 10 / HZ); |
2451 | if (err) |
2452 | goto out_cancel; |
2453 | } |
2454 | err = xfrm_mark_put(skb, m: &x->mark); |
2455 | if (err) |
2456 | goto out_cancel; |
2457 | |
2458 | err = xfrm_if_id_put(skb, if_id: x->if_id); |
2459 | if (err) |
2460 | goto out_cancel; |
2461 | |
2462 | nlmsg_end(skb, nlh); |
2463 | return 0; |
2464 | |
2465 | out_cancel: |
2466 | nlmsg_cancel(skb, nlh); |
2467 | return err; |
2468 | } |
2469 | |
2470 | static int xfrm_get_ae(struct sk_buff *skb, struct nlmsghdr *nlh, |
2471 | struct nlattr **attrs, struct netlink_ext_ack *extack) |
2472 | { |
2473 | struct net *net = sock_net(sk: skb->sk); |
2474 | struct xfrm_state *x; |
2475 | struct sk_buff *r_skb; |
2476 | int err; |
2477 | struct km_event c; |
2478 | u32 mark; |
2479 | struct xfrm_mark m; |
2480 | struct xfrm_aevent_id *p = nlmsg_data(nlh); |
2481 | struct xfrm_usersa_id *id = &p->sa_id; |
2482 | |
2483 | mark = xfrm_mark_get(attrs, m: &m); |
2484 | |
2485 | x = xfrm_state_lookup(net, mark, daddr: &id->daddr, spi: id->spi, proto: id->proto, family: id->family); |
2486 | if (x == NULL) |
2487 | return -ESRCH; |
2488 | |
2489 | r_skb = nlmsg_new(payload: xfrm_aevent_msgsize(x), GFP_ATOMIC); |
2490 | if (r_skb == NULL) { |
2491 | xfrm_state_put(x); |
2492 | return -ENOMEM; |
2493 | } |
2494 | |
2495 | /* |
2496 | * XXX: is this lock really needed - none of the other |
2497 | * gets lock (the concern is things getting updated |
2498 | * while we are still reading) - jhs |
2499 | */ |
2500 | spin_lock_bh(lock: &x->lock); |
2501 | c.data.aevent = p->flags; |
2502 | c.seq = nlh->nlmsg_seq; |
2503 | c.portid = nlh->nlmsg_pid; |
2504 | |
2505 | err = build_aevent(skb: r_skb, x, c: &c); |
2506 | BUG_ON(err < 0); |
2507 | |
2508 | err = nlmsg_unicast(sk: net->xfrm.nlsk, skb: r_skb, NETLINK_CB(skb).portid); |
2509 | spin_unlock_bh(lock: &x->lock); |
2510 | xfrm_state_put(x); |
2511 | return err; |
2512 | } |
2513 | |
2514 | static int xfrm_new_ae(struct sk_buff *skb, struct nlmsghdr *nlh, |
2515 | struct nlattr **attrs, struct netlink_ext_ack *extack) |
2516 | { |
2517 | struct net *net = sock_net(sk: skb->sk); |
2518 | struct xfrm_state *x; |
2519 | struct km_event c; |
2520 | int err = -EINVAL; |
2521 | u32 mark = 0; |
2522 | struct xfrm_mark m; |
2523 | struct xfrm_aevent_id *p = nlmsg_data(nlh); |
2524 | struct nlattr *rp = attrs[XFRMA_REPLAY_VAL]; |
2525 | struct nlattr *re = attrs[XFRMA_REPLAY_ESN_VAL]; |
2526 | struct nlattr *lt = attrs[XFRMA_LTIME_VAL]; |
2527 | struct nlattr *et = attrs[XFRMA_ETIMER_THRESH]; |
2528 | struct nlattr *rt = attrs[XFRMA_REPLAY_THRESH]; |
2529 | |
2530 | if (!lt && !rp && !re && !et && !rt) { |
2531 | NL_SET_ERR_MSG(extack, "Missing required attribute for AE" ); |
2532 | return err; |
2533 | } |
2534 | |
2535 | /* pedantic mode - thou shalt sayeth replaceth */ |
2536 | if (!(nlh->nlmsg_flags & NLM_F_REPLACE)) { |
2537 | NL_SET_ERR_MSG(extack, "NLM_F_REPLACE flag is required" ); |
2538 | return err; |
2539 | } |
2540 | |
2541 | mark = xfrm_mark_get(attrs, m: &m); |
2542 | |
2543 | x = xfrm_state_lookup(net, mark, daddr: &p->sa_id.daddr, spi: p->sa_id.spi, proto: p->sa_id.proto, family: p->sa_id.family); |
2544 | if (x == NULL) |
2545 | return -ESRCH; |
2546 | |
2547 | if (x->km.state != XFRM_STATE_VALID) { |
2548 | NL_SET_ERR_MSG(extack, "SA must be in VALID state" ); |
2549 | goto out; |
2550 | } |
2551 | |
2552 | err = xfrm_replay_verify_len(replay_esn: x->replay_esn, rp: re, extack); |
2553 | if (err) |
2554 | goto out; |
2555 | |
2556 | spin_lock_bh(lock: &x->lock); |
2557 | xfrm_update_ae_params(x, attrs, update_esn: 1); |
2558 | spin_unlock_bh(lock: &x->lock); |
2559 | |
2560 | c.event = nlh->nlmsg_type; |
2561 | c.seq = nlh->nlmsg_seq; |
2562 | c.portid = nlh->nlmsg_pid; |
2563 | c.data.aevent = XFRM_AE_CU; |
2564 | km_state_notify(x, c: &c); |
2565 | err = 0; |
2566 | out: |
2567 | xfrm_state_put(x); |
2568 | return err; |
2569 | } |
2570 | |
2571 | static int xfrm_flush_policy(struct sk_buff *skb, struct nlmsghdr *nlh, |
2572 | struct nlattr **attrs, |
2573 | struct netlink_ext_ack *extack) |
2574 | { |
2575 | struct net *net = sock_net(sk: skb->sk); |
2576 | struct km_event c; |
2577 | u8 type = XFRM_POLICY_TYPE_MAIN; |
2578 | int err; |
2579 | |
2580 | err = copy_from_user_policy_type(tp: &type, attrs, extack); |
2581 | if (err) |
2582 | return err; |
2583 | |
2584 | err = xfrm_policy_flush(net, type, task_valid: true); |
2585 | if (err) { |
2586 | if (err == -ESRCH) /* empty table */ |
2587 | return 0; |
2588 | return err; |
2589 | } |
2590 | |
2591 | c.data.type = type; |
2592 | c.event = nlh->nlmsg_type; |
2593 | c.seq = nlh->nlmsg_seq; |
2594 | c.portid = nlh->nlmsg_pid; |
2595 | c.net = net; |
2596 | km_policy_notify(NULL, dir: 0, c: &c); |
2597 | return 0; |
2598 | } |
2599 | |
2600 | static int xfrm_add_pol_expire(struct sk_buff *skb, struct nlmsghdr *nlh, |
2601 | struct nlattr **attrs, |
2602 | struct netlink_ext_ack *extack) |
2603 | { |
2604 | struct net *net = sock_net(sk: skb->sk); |
2605 | struct xfrm_policy *xp; |
2606 | struct xfrm_user_polexpire *up = nlmsg_data(nlh); |
2607 | struct xfrm_userpolicy_info *p = &up->pol; |
2608 | u8 type = XFRM_POLICY_TYPE_MAIN; |
2609 | int err = -ENOENT; |
2610 | struct xfrm_mark m; |
2611 | u32 if_id = 0; |
2612 | |
2613 | err = copy_from_user_policy_type(tp: &type, attrs, extack); |
2614 | if (err) |
2615 | return err; |
2616 | |
2617 | err = verify_policy_dir(dir: p->dir, extack); |
2618 | if (err) |
2619 | return err; |
2620 | |
2621 | if (attrs[XFRMA_IF_ID]) |
2622 | if_id = nla_get_u32(nla: attrs[XFRMA_IF_ID]); |
2623 | |
2624 | xfrm_mark_get(attrs, m: &m); |
2625 | |
2626 | if (p->index) |
2627 | xp = xfrm_policy_byid(net, mark: &m, if_id, type, dir: p->dir, id: p->index, |
2628 | delete: 0, err: &err); |
2629 | else { |
2630 | struct nlattr *rt = attrs[XFRMA_SEC_CTX]; |
2631 | struct xfrm_sec_ctx *ctx; |
2632 | |
2633 | err = verify_sec_ctx_len(attrs, extack); |
2634 | if (err) |
2635 | return err; |
2636 | |
2637 | ctx = NULL; |
2638 | if (rt) { |
2639 | struct xfrm_user_sec_ctx *uctx = nla_data(nla: rt); |
2640 | |
2641 | err = security_xfrm_policy_alloc(ctxp: &ctx, sec_ctx: uctx, GFP_KERNEL); |
2642 | if (err) |
2643 | return err; |
2644 | } |
2645 | xp = xfrm_policy_bysel_ctx(net, mark: &m, if_id, type, dir: p->dir, |
2646 | sel: &p->sel, ctx, delete: 0, err: &err); |
2647 | security_xfrm_policy_free(ctx); |
2648 | } |
2649 | if (xp == NULL) |
2650 | return -ENOENT; |
2651 | |
2652 | if (unlikely(xp->walk.dead)) |
2653 | goto out; |
2654 | |
2655 | err = 0; |
2656 | if (up->hard) { |
2657 | xfrm_policy_delete(pol: xp, dir: p->dir); |
2658 | xfrm_audit_policy_delete(xp, result: 1, task_valid: true); |
2659 | } |
2660 | km_policy_expired(pol: xp, dir: p->dir, hard: up->hard, portid: nlh->nlmsg_pid); |
2661 | |
2662 | out: |
2663 | xfrm_pol_put(policy: xp); |
2664 | return err; |
2665 | } |
2666 | |
2667 | static int xfrm_add_sa_expire(struct sk_buff *skb, struct nlmsghdr *nlh, |
2668 | struct nlattr **attrs, |
2669 | struct netlink_ext_ack *extack) |
2670 | { |
2671 | struct net *net = sock_net(sk: skb->sk); |
2672 | struct xfrm_state *x; |
2673 | int err; |
2674 | struct xfrm_user_expire *ue = nlmsg_data(nlh); |
2675 | struct xfrm_usersa_info *p = &ue->state; |
2676 | struct xfrm_mark m; |
2677 | u32 mark = xfrm_mark_get(attrs, m: &m); |
2678 | |
2679 | x = xfrm_state_lookup(net, mark, daddr: &p->id.daddr, spi: p->id.spi, proto: p->id.proto, family: p->family); |
2680 | |
2681 | err = -ENOENT; |
2682 | if (x == NULL) |
2683 | return err; |
2684 | |
2685 | spin_lock_bh(lock: &x->lock); |
2686 | err = -EINVAL; |
2687 | if (x->km.state != XFRM_STATE_VALID) { |
2688 | NL_SET_ERR_MSG(extack, "SA must be in VALID state" ); |
2689 | goto out; |
2690 | } |
2691 | |
2692 | km_state_expired(x, hard: ue->hard, portid: nlh->nlmsg_pid); |
2693 | |
2694 | if (ue->hard) { |
2695 | __xfrm_state_delete(x); |
2696 | xfrm_audit_state_delete(x, result: 1, task_valid: true); |
2697 | } |
2698 | err = 0; |
2699 | out: |
2700 | spin_unlock_bh(lock: &x->lock); |
2701 | xfrm_state_put(x); |
2702 | return err; |
2703 | } |
2704 | |
2705 | static int xfrm_add_acquire(struct sk_buff *skb, struct nlmsghdr *nlh, |
2706 | struct nlattr **attrs, |
2707 | struct netlink_ext_ack *extack) |
2708 | { |
2709 | struct net *net = sock_net(sk: skb->sk); |
2710 | struct xfrm_policy *xp; |
2711 | struct xfrm_user_tmpl *ut; |
2712 | int i; |
2713 | struct nlattr *rt = attrs[XFRMA_TMPL]; |
2714 | struct xfrm_mark mark; |
2715 | |
2716 | struct xfrm_user_acquire *ua = nlmsg_data(nlh); |
2717 | struct xfrm_state *x = xfrm_state_alloc(net); |
2718 | int err = -ENOMEM; |
2719 | |
2720 | if (!x) |
2721 | goto nomem; |
2722 | |
2723 | xfrm_mark_get(attrs, m: &mark); |
2724 | |
2725 | err = verify_newpolicy_info(p: &ua->policy, extack); |
2726 | if (err) |
2727 | goto free_state; |
2728 | err = verify_sec_ctx_len(attrs, extack); |
2729 | if (err) |
2730 | goto free_state; |
2731 | |
2732 | /* build an XP */ |
2733 | xp = xfrm_policy_construct(net, p: &ua->policy, attrs, errp: &err, extack); |
2734 | if (!xp) |
2735 | goto free_state; |
2736 | |
2737 | memcpy(&x->id, &ua->id, sizeof(ua->id)); |
2738 | memcpy(&x->props.saddr, &ua->saddr, sizeof(ua->saddr)); |
2739 | memcpy(&x->sel, &ua->sel, sizeof(ua->sel)); |
2740 | xp->mark.m = x->mark.m = mark.m; |
2741 | xp->mark.v = x->mark.v = mark.v; |
2742 | ut = nla_data(nla: rt); |
2743 | /* extract the templates and for each call km_key */ |
2744 | for (i = 0; i < xp->xfrm_nr; i++, ut++) { |
2745 | struct xfrm_tmpl *t = &xp->xfrm_vec[i]; |
2746 | memcpy(&x->id, &t->id, sizeof(x->id)); |
2747 | x->props.mode = t->mode; |
2748 | x->props.reqid = t->reqid; |
2749 | x->props.family = ut->family; |
2750 | t->aalgos = ua->aalgos; |
2751 | t->ealgos = ua->ealgos; |
2752 | t->calgos = ua->calgos; |
2753 | err = km_query(x, t, pol: xp); |
2754 | |
2755 | } |
2756 | |
2757 | xfrm_state_free(x); |
2758 | kfree(objp: xp); |
2759 | |
2760 | return 0; |
2761 | |
2762 | free_state: |
2763 | xfrm_state_free(x); |
2764 | nomem: |
2765 | return err; |
2766 | } |
2767 | |
2768 | #ifdef CONFIG_XFRM_MIGRATE |
2769 | static int copy_from_user_migrate(struct xfrm_migrate *ma, |
2770 | struct xfrm_kmaddress *k, |
2771 | struct nlattr **attrs, int *num, |
2772 | struct netlink_ext_ack *extack) |
2773 | { |
2774 | struct nlattr *rt = attrs[XFRMA_MIGRATE]; |
2775 | struct xfrm_user_migrate *um; |
2776 | int i, num_migrate; |
2777 | |
2778 | if (k != NULL) { |
2779 | struct xfrm_user_kmaddress *uk; |
2780 | |
2781 | uk = nla_data(nla: attrs[XFRMA_KMADDRESS]); |
2782 | memcpy(&k->local, &uk->local, sizeof(k->local)); |
2783 | memcpy(&k->remote, &uk->remote, sizeof(k->remote)); |
2784 | k->family = uk->family; |
2785 | k->reserved = uk->reserved; |
2786 | } |
2787 | |
2788 | um = nla_data(nla: rt); |
2789 | num_migrate = nla_len(nla: rt) / sizeof(*um); |
2790 | |
2791 | if (num_migrate <= 0 || num_migrate > XFRM_MAX_DEPTH) { |
2792 | NL_SET_ERR_MSG(extack, "Invalid number of SAs to migrate, must be 0 < num <= XFRM_MAX_DEPTH (6)" ); |
2793 | return -EINVAL; |
2794 | } |
2795 | |
2796 | for (i = 0; i < num_migrate; i++, um++, ma++) { |
2797 | memcpy(&ma->old_daddr, &um->old_daddr, sizeof(ma->old_daddr)); |
2798 | memcpy(&ma->old_saddr, &um->old_saddr, sizeof(ma->old_saddr)); |
2799 | memcpy(&ma->new_daddr, &um->new_daddr, sizeof(ma->new_daddr)); |
2800 | memcpy(&ma->new_saddr, &um->new_saddr, sizeof(ma->new_saddr)); |
2801 | |
2802 | ma->proto = um->proto; |
2803 | ma->mode = um->mode; |
2804 | ma->reqid = um->reqid; |
2805 | |
2806 | ma->old_family = um->old_family; |
2807 | ma->new_family = um->new_family; |
2808 | } |
2809 | |
2810 | *num = i; |
2811 | return 0; |
2812 | } |
2813 | |
2814 | static int xfrm_do_migrate(struct sk_buff *skb, struct nlmsghdr *nlh, |
2815 | struct nlattr **attrs, struct netlink_ext_ack *extack) |
2816 | { |
2817 | struct xfrm_userpolicy_id *pi = nlmsg_data(nlh); |
2818 | struct xfrm_migrate m[XFRM_MAX_DEPTH]; |
2819 | struct xfrm_kmaddress km, *kmp; |
2820 | u8 type; |
2821 | int err; |
2822 | int n = 0; |
2823 | struct net *net = sock_net(sk: skb->sk); |
2824 | struct xfrm_encap_tmpl *encap = NULL; |
2825 | u32 if_id = 0; |
2826 | |
2827 | if (!attrs[XFRMA_MIGRATE]) { |
2828 | NL_SET_ERR_MSG(extack, "Missing required MIGRATE attribute" ); |
2829 | return -EINVAL; |
2830 | } |
2831 | |
2832 | kmp = attrs[XFRMA_KMADDRESS] ? &km : NULL; |
2833 | |
2834 | err = copy_from_user_policy_type(tp: &type, attrs, extack); |
2835 | if (err) |
2836 | return err; |
2837 | |
2838 | err = copy_from_user_migrate(ma: m, k: kmp, attrs, num: &n, extack); |
2839 | if (err) |
2840 | return err; |
2841 | |
2842 | if (!n) |
2843 | return 0; |
2844 | |
2845 | if (attrs[XFRMA_ENCAP]) { |
2846 | encap = kmemdup(p: nla_data(nla: attrs[XFRMA_ENCAP]), |
2847 | size: sizeof(*encap), GFP_KERNEL); |
2848 | if (!encap) |
2849 | return -ENOMEM; |
2850 | } |
2851 | |
2852 | if (attrs[XFRMA_IF_ID]) |
2853 | if_id = nla_get_u32(nla: attrs[XFRMA_IF_ID]); |
2854 | |
2855 | err = xfrm_migrate(sel: &pi->sel, dir: pi->dir, type, m, num_bundles: n, k: kmp, net, encap, |
2856 | if_id, extack); |
2857 | |
2858 | kfree(objp: encap); |
2859 | |
2860 | return err; |
2861 | } |
2862 | #else |
2863 | static int xfrm_do_migrate(struct sk_buff *skb, struct nlmsghdr *nlh, |
2864 | struct nlattr **attrs, struct netlink_ext_ack *extack) |
2865 | { |
2866 | return -ENOPROTOOPT; |
2867 | } |
2868 | #endif |
2869 | |
2870 | #ifdef CONFIG_XFRM_MIGRATE |
2871 | static int copy_to_user_migrate(const struct xfrm_migrate *m, struct sk_buff *skb) |
2872 | { |
2873 | struct xfrm_user_migrate um; |
2874 | |
2875 | memset(&um, 0, sizeof(um)); |
2876 | um.proto = m->proto; |
2877 | um.mode = m->mode; |
2878 | um.reqid = m->reqid; |
2879 | um.old_family = m->old_family; |
2880 | memcpy(&um.old_daddr, &m->old_daddr, sizeof(um.old_daddr)); |
2881 | memcpy(&um.old_saddr, &m->old_saddr, sizeof(um.old_saddr)); |
2882 | um.new_family = m->new_family; |
2883 | memcpy(&um.new_daddr, &m->new_daddr, sizeof(um.new_daddr)); |
2884 | memcpy(&um.new_saddr, &m->new_saddr, sizeof(um.new_saddr)); |
2885 | |
2886 | return nla_put(skb, attrtype: XFRMA_MIGRATE, attrlen: sizeof(um), data: &um); |
2887 | } |
2888 | |
2889 | static int copy_to_user_kmaddress(const struct xfrm_kmaddress *k, struct sk_buff *skb) |
2890 | { |
2891 | struct xfrm_user_kmaddress uk; |
2892 | |
2893 | memset(&uk, 0, sizeof(uk)); |
2894 | uk.family = k->family; |
2895 | uk.reserved = k->reserved; |
2896 | memcpy(&uk.local, &k->local, sizeof(uk.local)); |
2897 | memcpy(&uk.remote, &k->remote, sizeof(uk.remote)); |
2898 | |
2899 | return nla_put(skb, attrtype: XFRMA_KMADDRESS, attrlen: sizeof(uk), data: &uk); |
2900 | } |
2901 | |
2902 | static inline unsigned int xfrm_migrate_msgsize(int num_migrate, int with_kma, |
2903 | int with_encp) |
2904 | { |
2905 | return NLMSG_ALIGN(sizeof(struct xfrm_userpolicy_id)) |
2906 | + (with_kma ? nla_total_size(payload: sizeof(struct xfrm_kmaddress)) : 0) |
2907 | + (with_encp ? nla_total_size(payload: sizeof(struct xfrm_encap_tmpl)) : 0) |
2908 | + nla_total_size(payload: sizeof(struct xfrm_user_migrate) * num_migrate) |
2909 | + userpolicy_type_attrsize(); |
2910 | } |
2911 | |
2912 | static int build_migrate(struct sk_buff *skb, const struct xfrm_migrate *m, |
2913 | int num_migrate, const struct xfrm_kmaddress *k, |
2914 | const struct xfrm_selector *sel, |
2915 | const struct xfrm_encap_tmpl *encap, u8 dir, u8 type) |
2916 | { |
2917 | const struct xfrm_migrate *mp; |
2918 | struct xfrm_userpolicy_id *pol_id; |
2919 | struct nlmsghdr *nlh; |
2920 | int i, err; |
2921 | |
2922 | nlh = nlmsg_put(skb, portid: 0, seq: 0, XFRM_MSG_MIGRATE, payload: sizeof(*pol_id), flags: 0); |
2923 | if (nlh == NULL) |
2924 | return -EMSGSIZE; |
2925 | |
2926 | pol_id = nlmsg_data(nlh); |
2927 | /* copy data from selector, dir, and type to the pol_id */ |
2928 | memset(pol_id, 0, sizeof(*pol_id)); |
2929 | memcpy(&pol_id->sel, sel, sizeof(pol_id->sel)); |
2930 | pol_id->dir = dir; |
2931 | |
2932 | if (k != NULL) { |
2933 | err = copy_to_user_kmaddress(k, skb); |
2934 | if (err) |
2935 | goto out_cancel; |
2936 | } |
2937 | if (encap) { |
2938 | err = nla_put(skb, attrtype: XFRMA_ENCAP, attrlen: sizeof(*encap), data: encap); |
2939 | if (err) |
2940 | goto out_cancel; |
2941 | } |
2942 | err = copy_to_user_policy_type(type, skb); |
2943 | if (err) |
2944 | goto out_cancel; |
2945 | for (i = 0, mp = m ; i < num_migrate; i++, mp++) { |
2946 | err = copy_to_user_migrate(m: mp, skb); |
2947 | if (err) |
2948 | goto out_cancel; |
2949 | } |
2950 | |
2951 | nlmsg_end(skb, nlh); |
2952 | return 0; |
2953 | |
2954 | out_cancel: |
2955 | nlmsg_cancel(skb, nlh); |
2956 | return err; |
2957 | } |
2958 | |
2959 | static int xfrm_send_migrate(const struct xfrm_selector *sel, u8 dir, u8 type, |
2960 | const struct xfrm_migrate *m, int num_migrate, |
2961 | const struct xfrm_kmaddress *k, |
2962 | const struct xfrm_encap_tmpl *encap) |
2963 | { |
2964 | struct net *net = &init_net; |
2965 | struct sk_buff *skb; |
2966 | int err; |
2967 | |
2968 | skb = nlmsg_new(payload: xfrm_migrate_msgsize(num_migrate, with_kma: !!k, with_encp: !!encap), |
2969 | GFP_ATOMIC); |
2970 | if (skb == NULL) |
2971 | return -ENOMEM; |
2972 | |
2973 | /* build migrate */ |
2974 | err = build_migrate(skb, m, num_migrate, k, sel, encap, dir, type); |
2975 | BUG_ON(err < 0); |
2976 | |
2977 | return xfrm_nlmsg_multicast(net, skb, pid: 0, XFRMNLGRP_MIGRATE); |
2978 | } |
2979 | #else |
2980 | static int xfrm_send_migrate(const struct xfrm_selector *sel, u8 dir, u8 type, |
2981 | const struct xfrm_migrate *m, int num_migrate, |
2982 | const struct xfrm_kmaddress *k, |
2983 | const struct xfrm_encap_tmpl *encap) |
2984 | { |
2985 | return -ENOPROTOOPT; |
2986 | } |
2987 | #endif |
2988 | |
2989 | #define XMSGSIZE(type) sizeof(struct type) |
2990 | |
2991 | const int xfrm_msg_min[XFRM_NR_MSGTYPES] = { |
2992 | [XFRM_MSG_NEWSA - XFRM_MSG_BASE] = XMSGSIZE(xfrm_usersa_info), |
2993 | [XFRM_MSG_DELSA - XFRM_MSG_BASE] = XMSGSIZE(xfrm_usersa_id), |
2994 | [XFRM_MSG_GETSA - XFRM_MSG_BASE] = XMSGSIZE(xfrm_usersa_id), |
2995 | [XFRM_MSG_NEWPOLICY - XFRM_MSG_BASE] = XMSGSIZE(xfrm_userpolicy_info), |
2996 | [XFRM_MSG_DELPOLICY - XFRM_MSG_BASE] = XMSGSIZE(xfrm_userpolicy_id), |
2997 | [XFRM_MSG_GETPOLICY - XFRM_MSG_BASE] = XMSGSIZE(xfrm_userpolicy_id), |
2998 | [XFRM_MSG_ALLOCSPI - XFRM_MSG_BASE] = XMSGSIZE(xfrm_userspi_info), |
2999 | [XFRM_MSG_ACQUIRE - XFRM_MSG_BASE] = XMSGSIZE(xfrm_user_acquire), |
3000 | [XFRM_MSG_EXPIRE - XFRM_MSG_BASE] = XMSGSIZE(xfrm_user_expire), |
3001 | [XFRM_MSG_UPDPOLICY - XFRM_MSG_BASE] = XMSGSIZE(xfrm_userpolicy_info), |
3002 | [XFRM_MSG_UPDSA - XFRM_MSG_BASE] = XMSGSIZE(xfrm_usersa_info), |
3003 | [XFRM_MSG_POLEXPIRE - XFRM_MSG_BASE] = XMSGSIZE(xfrm_user_polexpire), |
3004 | [XFRM_MSG_FLUSHSA - XFRM_MSG_BASE] = XMSGSIZE(xfrm_usersa_flush), |
3005 | [XFRM_MSG_FLUSHPOLICY - XFRM_MSG_BASE] = 0, |
3006 | [XFRM_MSG_NEWAE - XFRM_MSG_BASE] = XMSGSIZE(xfrm_aevent_id), |
3007 | [XFRM_MSG_GETAE - XFRM_MSG_BASE] = XMSGSIZE(xfrm_aevent_id), |
3008 | [XFRM_MSG_REPORT - XFRM_MSG_BASE] = XMSGSIZE(xfrm_user_report), |
3009 | [XFRM_MSG_MIGRATE - XFRM_MSG_BASE] = XMSGSIZE(xfrm_userpolicy_id), |
3010 | [XFRM_MSG_GETSADINFO - XFRM_MSG_BASE] = sizeof(u32), |
3011 | [XFRM_MSG_NEWSPDINFO - XFRM_MSG_BASE] = sizeof(u32), |
3012 | [XFRM_MSG_GETSPDINFO - XFRM_MSG_BASE] = sizeof(u32), |
3013 | [XFRM_MSG_SETDEFAULT - XFRM_MSG_BASE] = XMSGSIZE(xfrm_userpolicy_default), |
3014 | [XFRM_MSG_GETDEFAULT - XFRM_MSG_BASE] = XMSGSIZE(xfrm_userpolicy_default), |
3015 | }; |
3016 | EXPORT_SYMBOL_GPL(xfrm_msg_min); |
3017 | |
3018 | #undef XMSGSIZE |
3019 | |
3020 | const struct nla_policy xfrma_policy[XFRMA_MAX+1] = { |
3021 | [XFRMA_SA] = { .len = sizeof(struct xfrm_usersa_info)}, |
3022 | [XFRMA_POLICY] = { .len = sizeof(struct xfrm_userpolicy_info)}, |
3023 | [XFRMA_LASTUSED] = { .type = NLA_U64}, |
3024 | [XFRMA_ALG_AUTH_TRUNC] = { .len = sizeof(struct xfrm_algo_auth)}, |
3025 | [XFRMA_ALG_AEAD] = { .len = sizeof(struct xfrm_algo_aead) }, |
3026 | [XFRMA_ALG_AUTH] = { .len = sizeof(struct xfrm_algo) }, |
3027 | [XFRMA_ALG_CRYPT] = { .len = sizeof(struct xfrm_algo) }, |
3028 | [XFRMA_ALG_COMP] = { .len = sizeof(struct xfrm_algo) }, |
3029 | [XFRMA_ENCAP] = { .len = sizeof(struct xfrm_encap_tmpl) }, |
3030 | [XFRMA_TMPL] = { .len = sizeof(struct xfrm_user_tmpl) }, |
3031 | [XFRMA_SEC_CTX] = { .len = sizeof(struct xfrm_user_sec_ctx) }, |
3032 | [XFRMA_LTIME_VAL] = { .len = sizeof(struct xfrm_lifetime_cur) }, |
3033 | [XFRMA_REPLAY_VAL] = { .len = sizeof(struct xfrm_replay_state) }, |
3034 | [XFRMA_REPLAY_THRESH] = { .type = NLA_U32 }, |
3035 | [XFRMA_ETIMER_THRESH] = { .type = NLA_U32 }, |
3036 | [XFRMA_SRCADDR] = { .len = sizeof(xfrm_address_t) }, |
3037 | [XFRMA_COADDR] = { .len = sizeof(xfrm_address_t) }, |
3038 | [XFRMA_POLICY_TYPE] = { .len = sizeof(struct xfrm_userpolicy_type)}, |
3039 | [XFRMA_MIGRATE] = { .len = sizeof(struct xfrm_user_migrate) }, |
3040 | [XFRMA_KMADDRESS] = { .len = sizeof(struct xfrm_user_kmaddress) }, |
3041 | [XFRMA_MARK] = { .len = sizeof(struct xfrm_mark) }, |
3042 | [XFRMA_TFCPAD] = { .type = NLA_U32 }, |
3043 | [XFRMA_REPLAY_ESN_VAL] = { .len = sizeof(struct xfrm_replay_state_esn) }, |
3044 | [XFRMA_SA_EXTRA_FLAGS] = { .type = NLA_U32 }, |
3045 | [XFRMA_PROTO] = { .type = NLA_U8 }, |
3046 | [XFRMA_ADDRESS_FILTER] = { .len = sizeof(struct xfrm_address_filter) }, |
3047 | [XFRMA_OFFLOAD_DEV] = { .len = sizeof(struct xfrm_user_offload) }, |
3048 | [XFRMA_SET_MARK] = { .type = NLA_U32 }, |
3049 | [XFRMA_SET_MARK_MASK] = { .type = NLA_U32 }, |
3050 | [XFRMA_IF_ID] = { .type = NLA_U32 }, |
3051 | [XFRMA_MTIMER_THRESH] = { .type = NLA_U32 }, |
3052 | }; |
3053 | EXPORT_SYMBOL_GPL(xfrma_policy); |
3054 | |
3055 | static const struct nla_policy xfrma_spd_policy[XFRMA_SPD_MAX+1] = { |
3056 | [XFRMA_SPD_IPV4_HTHRESH] = { .len = sizeof(struct xfrmu_spdhthresh) }, |
3057 | [XFRMA_SPD_IPV6_HTHRESH] = { .len = sizeof(struct xfrmu_spdhthresh) }, |
3058 | }; |
3059 | |
3060 | static const struct xfrm_link { |
3061 | int (*doit)(struct sk_buff *, struct nlmsghdr *, struct nlattr **, |
3062 | struct netlink_ext_ack *); |
3063 | int (*start)(struct netlink_callback *); |
3064 | int (*dump)(struct sk_buff *, struct netlink_callback *); |
3065 | int (*done)(struct netlink_callback *); |
3066 | const struct nla_policy *nla_pol; |
3067 | int nla_max; |
3068 | } xfrm_dispatch[XFRM_NR_MSGTYPES] = { |
3069 | [XFRM_MSG_NEWSA - XFRM_MSG_BASE] = { .doit = xfrm_add_sa }, |
3070 | [XFRM_MSG_DELSA - XFRM_MSG_BASE] = { .doit = xfrm_del_sa }, |
3071 | [XFRM_MSG_GETSA - XFRM_MSG_BASE] = { .doit = xfrm_get_sa, |
3072 | .dump = xfrm_dump_sa, |
3073 | .done = xfrm_dump_sa_done }, |
3074 | [XFRM_MSG_NEWPOLICY - XFRM_MSG_BASE] = { .doit = xfrm_add_policy }, |
3075 | [XFRM_MSG_DELPOLICY - XFRM_MSG_BASE] = { .doit = xfrm_get_policy }, |
3076 | [XFRM_MSG_GETPOLICY - XFRM_MSG_BASE] = { .doit = xfrm_get_policy, |
3077 | .start = xfrm_dump_policy_start, |
3078 | .dump = xfrm_dump_policy, |
3079 | .done = xfrm_dump_policy_done }, |
3080 | [XFRM_MSG_ALLOCSPI - XFRM_MSG_BASE] = { .doit = xfrm_alloc_userspi }, |
3081 | [XFRM_MSG_ACQUIRE - XFRM_MSG_BASE] = { .doit = xfrm_add_acquire }, |
3082 | [XFRM_MSG_EXPIRE - XFRM_MSG_BASE] = { .doit = xfrm_add_sa_expire }, |
3083 | [XFRM_MSG_UPDPOLICY - XFRM_MSG_BASE] = { .doit = xfrm_add_policy }, |
3084 | [XFRM_MSG_UPDSA - XFRM_MSG_BASE] = { .doit = xfrm_add_sa }, |
3085 | [XFRM_MSG_POLEXPIRE - XFRM_MSG_BASE] = { .doit = xfrm_add_pol_expire}, |
3086 | [XFRM_MSG_FLUSHSA - XFRM_MSG_BASE] = { .doit = xfrm_flush_sa }, |
3087 | [XFRM_MSG_FLUSHPOLICY - XFRM_MSG_BASE] = { .doit = xfrm_flush_policy }, |
3088 | [XFRM_MSG_NEWAE - XFRM_MSG_BASE] = { .doit = xfrm_new_ae }, |
3089 | [XFRM_MSG_GETAE - XFRM_MSG_BASE] = { .doit = xfrm_get_ae }, |
3090 | [XFRM_MSG_MIGRATE - XFRM_MSG_BASE] = { .doit = xfrm_do_migrate }, |
3091 | [XFRM_MSG_GETSADINFO - XFRM_MSG_BASE] = { .doit = xfrm_get_sadinfo }, |
3092 | [XFRM_MSG_NEWSPDINFO - XFRM_MSG_BASE] = { .doit = xfrm_set_spdinfo, |
3093 | .nla_pol = xfrma_spd_policy, |
3094 | .nla_max = XFRMA_SPD_MAX }, |
3095 | [XFRM_MSG_GETSPDINFO - XFRM_MSG_BASE] = { .doit = xfrm_get_spdinfo }, |
3096 | [XFRM_MSG_SETDEFAULT - XFRM_MSG_BASE] = { .doit = xfrm_set_default }, |
3097 | [XFRM_MSG_GETDEFAULT - XFRM_MSG_BASE] = { .doit = xfrm_get_default }, |
3098 | }; |
3099 | |
3100 | static int xfrm_user_rcv_msg(struct sk_buff *skb, struct nlmsghdr *nlh, |
3101 | struct netlink_ext_ack *extack) |
3102 | { |
3103 | struct net *net = sock_net(sk: skb->sk); |
3104 | struct nlattr *attrs[XFRMA_MAX+1]; |
3105 | const struct xfrm_link *link; |
3106 | struct nlmsghdr *nlh64 = NULL; |
3107 | int type, err; |
3108 | |
3109 | type = nlh->nlmsg_type; |
3110 | if (type > XFRM_MSG_MAX) |
3111 | return -EINVAL; |
3112 | |
3113 | type -= XFRM_MSG_BASE; |
3114 | link = &xfrm_dispatch[type]; |
3115 | |
3116 | /* All operations require privileges, even GET */ |
3117 | if (!netlink_net_capable(skb, CAP_NET_ADMIN)) |
3118 | return -EPERM; |
3119 | |
3120 | if (in_compat_syscall()) { |
3121 | struct xfrm_translator *xtr = xfrm_get_translator(); |
3122 | |
3123 | if (!xtr) |
3124 | return -EOPNOTSUPP; |
3125 | |
3126 | nlh64 = xtr->rcv_msg_compat(nlh, link->nla_max, |
3127 | link->nla_pol, extack); |
3128 | xfrm_put_translator(xtr); |
3129 | if (IS_ERR(ptr: nlh64)) |
3130 | return PTR_ERR(ptr: nlh64); |
3131 | if (nlh64) |
3132 | nlh = nlh64; |
3133 | } |
3134 | |
3135 | if ((type == (XFRM_MSG_GETSA - XFRM_MSG_BASE) || |
3136 | type == (XFRM_MSG_GETPOLICY - XFRM_MSG_BASE)) && |
3137 | (nlh->nlmsg_flags & NLM_F_DUMP)) { |
3138 | struct netlink_dump_control c = { |
3139 | .start = link->start, |
3140 | .dump = link->dump, |
3141 | .done = link->done, |
3142 | }; |
3143 | |
3144 | if (link->dump == NULL) { |
3145 | err = -EINVAL; |
3146 | goto err; |
3147 | } |
3148 | |
3149 | err = netlink_dump_start(ssk: net->xfrm.nlsk, skb, nlh, control: &c); |
3150 | goto err; |
3151 | } |
3152 | |
3153 | err = nlmsg_parse_deprecated(nlh, hdrlen: xfrm_msg_min[type], tb: attrs, |
3154 | maxtype: link->nla_max ? : XFRMA_MAX, |
3155 | policy: link->nla_pol ? : xfrma_policy, extack); |
3156 | if (err < 0) |
3157 | goto err; |
3158 | |
3159 | if (link->doit == NULL) { |
3160 | err = -EINVAL; |
3161 | goto err; |
3162 | } |
3163 | |
3164 | err = link->doit(skb, nlh, attrs, extack); |
3165 | |
3166 | /* We need to free skb allocated in xfrm_alloc_compat() before |
3167 | * returning from this function, because consume_skb() won't take |
3168 | * care of frag_list since netlink destructor sets |
3169 | * sbk->head to NULL. (see netlink_skb_destructor()) |
3170 | */ |
3171 | if (skb_has_frag_list(skb)) { |
3172 | kfree_skb(skb_shinfo(skb)->frag_list); |
3173 | skb_shinfo(skb)->frag_list = NULL; |
3174 | } |
3175 | |
3176 | err: |
3177 | kvfree(addr: nlh64); |
3178 | return err; |
3179 | } |
3180 | |
3181 | static void xfrm_netlink_rcv(struct sk_buff *skb) |
3182 | { |
3183 | struct net *net = sock_net(sk: skb->sk); |
3184 | |
3185 | mutex_lock(&net->xfrm.xfrm_cfg_mutex); |
3186 | netlink_rcv_skb(skb, cb: &xfrm_user_rcv_msg); |
3187 | mutex_unlock(lock: &net->xfrm.xfrm_cfg_mutex); |
3188 | } |
3189 | |
3190 | static inline unsigned int xfrm_expire_msgsize(void) |
3191 | { |
3192 | return NLMSG_ALIGN(sizeof(struct xfrm_user_expire)) |
3193 | + nla_total_size(payload: sizeof(struct xfrm_mark)); |
3194 | } |
3195 | |
3196 | static int build_expire(struct sk_buff *skb, struct xfrm_state *x, const struct km_event *c) |
3197 | { |
3198 | struct xfrm_user_expire *ue; |
3199 | struct nlmsghdr *nlh; |
3200 | int err; |
3201 | |
3202 | nlh = nlmsg_put(skb, portid: c->portid, seq: 0, XFRM_MSG_EXPIRE, payload: sizeof(*ue), flags: 0); |
3203 | if (nlh == NULL) |
3204 | return -EMSGSIZE; |
3205 | |
3206 | ue = nlmsg_data(nlh); |
3207 | copy_to_user_state(x, p: &ue->state); |
3208 | ue->hard = (c->data.hard != 0) ? 1 : 0; |
3209 | /* clear the padding bytes */ |
3210 | memset_after(ue, 0, hard); |
3211 | |
3212 | err = xfrm_mark_put(skb, m: &x->mark); |
3213 | if (err) |
3214 | return err; |
3215 | |
3216 | err = xfrm_if_id_put(skb, if_id: x->if_id); |
3217 | if (err) |
3218 | return err; |
3219 | |
3220 | nlmsg_end(skb, nlh); |
3221 | return 0; |
3222 | } |
3223 | |
3224 | static int xfrm_exp_state_notify(struct xfrm_state *x, const struct km_event *c) |
3225 | { |
3226 | struct net *net = xs_net(x); |
3227 | struct sk_buff *skb; |
3228 | |
3229 | skb = nlmsg_new(payload: xfrm_expire_msgsize(), GFP_ATOMIC); |
3230 | if (skb == NULL) |
3231 | return -ENOMEM; |
3232 | |
3233 | if (build_expire(skb, x, c) < 0) { |
3234 | kfree_skb(skb); |
3235 | return -EMSGSIZE; |
3236 | } |
3237 | |
3238 | return xfrm_nlmsg_multicast(net, skb, pid: 0, XFRMNLGRP_EXPIRE); |
3239 | } |
3240 | |
3241 | static int xfrm_aevent_state_notify(struct xfrm_state *x, const struct km_event *c) |
3242 | { |
3243 | struct net *net = xs_net(x); |
3244 | struct sk_buff *skb; |
3245 | int err; |
3246 | |
3247 | skb = nlmsg_new(payload: xfrm_aevent_msgsize(x), GFP_ATOMIC); |
3248 | if (skb == NULL) |
3249 | return -ENOMEM; |
3250 | |
3251 | err = build_aevent(skb, x, c); |
3252 | BUG_ON(err < 0); |
3253 | |
3254 | return xfrm_nlmsg_multicast(net, skb, pid: 0, XFRMNLGRP_AEVENTS); |
3255 | } |
3256 | |
3257 | static int xfrm_notify_sa_flush(const struct km_event *c) |
3258 | { |
3259 | struct net *net = c->net; |
3260 | struct xfrm_usersa_flush *p; |
3261 | struct nlmsghdr *nlh; |
3262 | struct sk_buff *skb; |
3263 | int len = NLMSG_ALIGN(sizeof(struct xfrm_usersa_flush)); |
3264 | |
3265 | skb = nlmsg_new(payload: len, GFP_ATOMIC); |
3266 | if (skb == NULL) |
3267 | return -ENOMEM; |
3268 | |
3269 | nlh = nlmsg_put(skb, portid: c->portid, seq: c->seq, XFRM_MSG_FLUSHSA, payload: sizeof(*p), flags: 0); |
3270 | if (nlh == NULL) { |
3271 | kfree_skb(skb); |
3272 | return -EMSGSIZE; |
3273 | } |
3274 | |
3275 | p = nlmsg_data(nlh); |
3276 | p->proto = c->data.proto; |
3277 | |
3278 | nlmsg_end(skb, nlh); |
3279 | |
3280 | return xfrm_nlmsg_multicast(net, skb, pid: 0, XFRMNLGRP_SA); |
3281 | } |
3282 | |
3283 | static inline unsigned int xfrm_sa_len(struct xfrm_state *x) |
3284 | { |
3285 | unsigned int l = 0; |
3286 | if (x->aead) |
3287 | l += nla_total_size(payload: aead_len(alg: x->aead)); |
3288 | if (x->aalg) { |
3289 | l += nla_total_size(payload: sizeof(struct xfrm_algo) + |
3290 | (x->aalg->alg_key_len + 7) / 8); |
3291 | l += nla_total_size(payload: xfrm_alg_auth_len(alg: x->aalg)); |
3292 | } |
3293 | if (x->ealg) |
3294 | l += nla_total_size(payload: xfrm_alg_len(alg: x->ealg)); |
3295 | if (x->calg) |
3296 | l += nla_total_size(payload: sizeof(*x->calg)); |
3297 | if (x->encap) |
3298 | l += nla_total_size(payload: sizeof(*x->encap)); |
3299 | if (x->tfcpad) |
3300 | l += nla_total_size(payload: sizeof(x->tfcpad)); |
3301 | if (x->replay_esn) |
3302 | l += nla_total_size(payload: xfrm_replay_state_esn_len(replay_esn: x->replay_esn)); |
3303 | else |
3304 | l += nla_total_size(payload: sizeof(struct xfrm_replay_state)); |
3305 | if (x->security) |
3306 | l += nla_total_size(payload: sizeof(struct xfrm_user_sec_ctx) + |
3307 | x->security->ctx_len); |
3308 | if (x->coaddr) |
3309 | l += nla_total_size(payload: sizeof(*x->coaddr)); |
3310 | if (x->props.extra_flags) |
3311 | l += nla_total_size(payload: sizeof(x->props.extra_flags)); |
3312 | if (x->xso.dev) |
3313 | l += nla_total_size(payload: sizeof(struct xfrm_user_offload)); |
3314 | if (x->props.smark.v | x->props.smark.m) { |
3315 | l += nla_total_size(payload: sizeof(x->props.smark.v)); |
3316 | l += nla_total_size(payload: sizeof(x->props.smark.m)); |
3317 | } |
3318 | if (x->if_id) |
3319 | l += nla_total_size(payload: sizeof(x->if_id)); |
3320 | |
3321 | /* Must count x->lastused as it may become non-zero behind our back. */ |
3322 | l += nla_total_size_64bit(payload: sizeof(u64)); |
3323 | |
3324 | if (x->mapping_maxage) |
3325 | l += nla_total_size(payload: sizeof(x->mapping_maxage)); |
3326 | |
3327 | return l; |
3328 | } |
3329 | |
3330 | static int xfrm_notify_sa(struct xfrm_state *x, const struct km_event *c) |
3331 | { |
3332 | struct net *net = xs_net(x); |
3333 | struct xfrm_usersa_info *p; |
3334 | struct xfrm_usersa_id *id; |
3335 | struct nlmsghdr *nlh; |
3336 | struct sk_buff *skb; |
3337 | unsigned int len = xfrm_sa_len(x); |
3338 | unsigned int headlen; |
3339 | int err; |
3340 | |
3341 | headlen = sizeof(*p); |
3342 | if (c->event == XFRM_MSG_DELSA) { |
3343 | len += nla_total_size(payload: headlen); |
3344 | headlen = sizeof(*id); |
3345 | len += nla_total_size(payload: sizeof(struct xfrm_mark)); |
3346 | } |
3347 | len += NLMSG_ALIGN(headlen); |
3348 | |
3349 | skb = nlmsg_new(payload: len, GFP_ATOMIC); |
3350 | if (skb == NULL) |
3351 | return -ENOMEM; |
3352 | |
3353 | nlh = nlmsg_put(skb, portid: c->portid, seq: c->seq, type: c->event, payload: headlen, flags: 0); |
3354 | err = -EMSGSIZE; |
3355 | if (nlh == NULL) |
3356 | goto out_free_skb; |
3357 | |
3358 | p = nlmsg_data(nlh); |
3359 | if (c->event == XFRM_MSG_DELSA) { |
3360 | struct nlattr *attr; |
3361 | |
3362 | id = nlmsg_data(nlh); |
3363 | memset(id, 0, sizeof(*id)); |
3364 | memcpy(&id->daddr, &x->id.daddr, sizeof(id->daddr)); |
3365 | id->spi = x->id.spi; |
3366 | id->family = x->props.family; |
3367 | id->proto = x->id.proto; |
3368 | |
3369 | attr = nla_reserve(skb, attrtype: XFRMA_SA, attrlen: sizeof(*p)); |
3370 | err = -EMSGSIZE; |
3371 | if (attr == NULL) |
3372 | goto out_free_skb; |
3373 | |
3374 | p = nla_data(nla: attr); |
3375 | } |
3376 | err = copy_to_user_state_extra(x, p, skb); |
3377 | if (err) |
3378 | goto out_free_skb; |
3379 | |
3380 | nlmsg_end(skb, nlh); |
3381 | |
3382 | return xfrm_nlmsg_multicast(net, skb, pid: 0, XFRMNLGRP_SA); |
3383 | |
3384 | out_free_skb: |
3385 | kfree_skb(skb); |
3386 | return err; |
3387 | } |
3388 | |
3389 | static int xfrm_send_state_notify(struct xfrm_state *x, const struct km_event *c) |
3390 | { |
3391 | |
3392 | switch (c->event) { |
3393 | case XFRM_MSG_EXPIRE: |
3394 | return xfrm_exp_state_notify(x, c); |
3395 | case XFRM_MSG_NEWAE: |
3396 | return xfrm_aevent_state_notify(x, c); |
3397 | case XFRM_MSG_DELSA: |
3398 | case XFRM_MSG_UPDSA: |
3399 | case XFRM_MSG_NEWSA: |
3400 | return xfrm_notify_sa(x, c); |
3401 | case XFRM_MSG_FLUSHSA: |
3402 | return xfrm_notify_sa_flush(c); |
3403 | default: |
3404 | printk(KERN_NOTICE "xfrm_user: Unknown SA event %d\n" , |
3405 | c->event); |
3406 | break; |
3407 | } |
3408 | |
3409 | return 0; |
3410 | |
3411 | } |
3412 | |
3413 | static inline unsigned int xfrm_acquire_msgsize(struct xfrm_state *x, |
3414 | struct xfrm_policy *xp) |
3415 | { |
3416 | return NLMSG_ALIGN(sizeof(struct xfrm_user_acquire)) |
3417 | + nla_total_size(payload: sizeof(struct xfrm_user_tmpl) * xp->xfrm_nr) |
3418 | + nla_total_size(payload: sizeof(struct xfrm_mark)) |
3419 | + nla_total_size(payload: xfrm_user_sec_ctx_size(xfrm_ctx: x->security)) |
3420 | + userpolicy_type_attrsize(); |
3421 | } |
3422 | |
3423 | static int build_acquire(struct sk_buff *skb, struct xfrm_state *x, |
3424 | struct xfrm_tmpl *xt, struct xfrm_policy *xp) |
3425 | { |
3426 | __u32 seq = xfrm_get_acqseq(); |
3427 | struct xfrm_user_acquire *ua; |
3428 | struct nlmsghdr *nlh; |
3429 | int err; |
3430 | |
3431 | nlh = nlmsg_put(skb, portid: 0, seq: 0, XFRM_MSG_ACQUIRE, payload: sizeof(*ua), flags: 0); |
3432 | if (nlh == NULL) |
3433 | return -EMSGSIZE; |
3434 | |
3435 | ua = nlmsg_data(nlh); |
3436 | memcpy(&ua->id, &x->id, sizeof(ua->id)); |
3437 | memcpy(&ua->saddr, &x->props.saddr, sizeof(ua->saddr)); |
3438 | memcpy(&ua->sel, &x->sel, sizeof(ua->sel)); |
3439 | copy_to_user_policy(xp, p: &ua->policy, dir: XFRM_POLICY_OUT); |
3440 | ua->aalgos = xt->aalgos; |
3441 | ua->ealgos = xt->ealgos; |
3442 | ua->calgos = xt->calgos; |
3443 | ua->seq = x->km.seq = seq; |
3444 | |
3445 | err = copy_to_user_tmpl(xp, skb); |
3446 | if (!err) |
3447 | err = copy_to_user_state_sec_ctx(x, skb); |
3448 | if (!err) |
3449 | err = copy_to_user_policy_type(type: xp->type, skb); |
3450 | if (!err) |
3451 | err = xfrm_mark_put(skb, m: &xp->mark); |
3452 | if (!err) |
3453 | err = xfrm_if_id_put(skb, if_id: xp->if_id); |
3454 | if (!err && xp->xdo.dev) |
3455 | err = copy_user_offload(xso: &xp->xdo, skb); |
3456 | if (err) { |
3457 | nlmsg_cancel(skb, nlh); |
3458 | return err; |
3459 | } |
3460 | |
3461 | nlmsg_end(skb, nlh); |
3462 | return 0; |
3463 | } |
3464 | |
3465 | static int xfrm_send_acquire(struct xfrm_state *x, struct xfrm_tmpl *xt, |
3466 | struct xfrm_policy *xp) |
3467 | { |
3468 | struct net *net = xs_net(x); |
3469 | struct sk_buff *skb; |
3470 | int err; |
3471 | |
3472 | skb = nlmsg_new(payload: xfrm_acquire_msgsize(x, xp), GFP_ATOMIC); |
3473 | if (skb == NULL) |
3474 | return -ENOMEM; |
3475 | |
3476 | err = build_acquire(skb, x, xt, xp); |
3477 | BUG_ON(err < 0); |
3478 | |
3479 | return xfrm_nlmsg_multicast(net, skb, pid: 0, XFRMNLGRP_ACQUIRE); |
3480 | } |
3481 | |
3482 | /* User gives us xfrm_user_policy_info followed by an array of 0 |
3483 | * or more templates. |
3484 | */ |
3485 | static struct xfrm_policy *xfrm_compile_policy(struct sock *sk, int opt, |
3486 | u8 *data, int len, int *dir) |
3487 | { |
3488 | struct net *net = sock_net(sk); |
3489 | struct xfrm_userpolicy_info *p = (struct xfrm_userpolicy_info *)data; |
3490 | struct xfrm_user_tmpl *ut = (struct xfrm_user_tmpl *) (p + 1); |
3491 | struct xfrm_policy *xp; |
3492 | int nr; |
3493 | |
3494 | switch (sk->sk_family) { |
3495 | case AF_INET: |
3496 | if (opt != IP_XFRM_POLICY) { |
3497 | *dir = -EOPNOTSUPP; |
3498 | return NULL; |
3499 | } |
3500 | break; |
3501 | #if IS_ENABLED(CONFIG_IPV6) |
3502 | case AF_INET6: |
3503 | if (opt != IPV6_XFRM_POLICY) { |
3504 | *dir = -EOPNOTSUPP; |
3505 | return NULL; |
3506 | } |
3507 | break; |
3508 | #endif |
3509 | default: |
3510 | *dir = -EINVAL; |
3511 | return NULL; |
3512 | } |
3513 | |
3514 | *dir = -EINVAL; |
3515 | |
3516 | if (len < sizeof(*p) || |
3517 | verify_newpolicy_info(p, NULL)) |
3518 | return NULL; |
3519 | |
3520 | nr = ((len - sizeof(*p)) / sizeof(*ut)); |
3521 | if (validate_tmpl(nr, ut, family: p->sel.family, dir: p->dir, NULL)) |
3522 | return NULL; |
3523 | |
3524 | if (p->dir > XFRM_POLICY_OUT) |
3525 | return NULL; |
3526 | |
3527 | xp = xfrm_policy_alloc(net, GFP_ATOMIC); |
3528 | if (xp == NULL) { |
3529 | *dir = -ENOBUFS; |
3530 | return NULL; |
3531 | } |
3532 | |
3533 | copy_from_user_policy(xp, p); |
3534 | xp->type = XFRM_POLICY_TYPE_MAIN; |
3535 | copy_templates(xp, ut, nr); |
3536 | |
3537 | *dir = p->dir; |
3538 | |
3539 | return xp; |
3540 | } |
3541 | |
3542 | static inline unsigned int xfrm_polexpire_msgsize(struct xfrm_policy *xp) |
3543 | { |
3544 | return NLMSG_ALIGN(sizeof(struct xfrm_user_polexpire)) |
3545 | + nla_total_size(payload: sizeof(struct xfrm_user_tmpl) * xp->xfrm_nr) |
3546 | + nla_total_size(payload: xfrm_user_sec_ctx_size(xfrm_ctx: xp->security)) |
3547 | + nla_total_size(payload: sizeof(struct xfrm_mark)) |
3548 | + userpolicy_type_attrsize(); |
3549 | } |
3550 | |
3551 | static int build_polexpire(struct sk_buff *skb, struct xfrm_policy *xp, |
3552 | int dir, const struct km_event *c) |
3553 | { |
3554 | struct xfrm_user_polexpire *upe; |
3555 | int hard = c->data.hard; |
3556 | struct nlmsghdr *nlh; |
3557 | int err; |
3558 | |
3559 | nlh = nlmsg_put(skb, portid: c->portid, seq: 0, XFRM_MSG_POLEXPIRE, payload: sizeof(*upe), flags: 0); |
3560 | if (nlh == NULL) |
3561 | return -EMSGSIZE; |
3562 | |
3563 | upe = nlmsg_data(nlh); |
3564 | copy_to_user_policy(xp, p: &upe->pol, dir); |
3565 | err = copy_to_user_tmpl(xp, skb); |
3566 | if (!err) |
3567 | err = copy_to_user_sec_ctx(xp, skb); |
3568 | if (!err) |
3569 | err = copy_to_user_policy_type(type: xp->type, skb); |
3570 | if (!err) |
3571 | err = xfrm_mark_put(skb, m: &xp->mark); |
3572 | if (!err) |
3573 | err = xfrm_if_id_put(skb, if_id: xp->if_id); |
3574 | if (!err && xp->xdo.dev) |
3575 | err = copy_user_offload(xso: &xp->xdo, skb); |
3576 | if (err) { |
3577 | nlmsg_cancel(skb, nlh); |
3578 | return err; |
3579 | } |
3580 | upe->hard = !!hard; |
3581 | |
3582 | nlmsg_end(skb, nlh); |
3583 | return 0; |
3584 | } |
3585 | |
3586 | static int xfrm_exp_policy_notify(struct xfrm_policy *xp, int dir, const struct km_event *c) |
3587 | { |
3588 | struct net *net = xp_net(xp); |
3589 | struct sk_buff *skb; |
3590 | int err; |
3591 | |
3592 | skb = nlmsg_new(payload: xfrm_polexpire_msgsize(xp), GFP_ATOMIC); |
3593 | if (skb == NULL) |
3594 | return -ENOMEM; |
3595 | |
3596 | err = build_polexpire(skb, xp, dir, c); |
3597 | BUG_ON(err < 0); |
3598 | |
3599 | return xfrm_nlmsg_multicast(net, skb, pid: 0, XFRMNLGRP_EXPIRE); |
3600 | } |
3601 | |
3602 | static int xfrm_notify_policy(struct xfrm_policy *xp, int dir, const struct km_event *c) |
3603 | { |
3604 | unsigned int len = nla_total_size(payload: sizeof(struct xfrm_user_tmpl) * xp->xfrm_nr); |
3605 | struct net *net = xp_net(xp); |
3606 | struct xfrm_userpolicy_info *p; |
3607 | struct xfrm_userpolicy_id *id; |
3608 | struct nlmsghdr *nlh; |
3609 | struct sk_buff *skb; |
3610 | unsigned int headlen; |
3611 | int err; |
3612 | |
3613 | headlen = sizeof(*p); |
3614 | if (c->event == XFRM_MSG_DELPOLICY) { |
3615 | len += nla_total_size(payload: headlen); |
3616 | headlen = sizeof(*id); |
3617 | } |
3618 | len += userpolicy_type_attrsize(); |
3619 | len += nla_total_size(payload: sizeof(struct xfrm_mark)); |
3620 | len += NLMSG_ALIGN(headlen); |
3621 | |
3622 | skb = nlmsg_new(payload: len, GFP_ATOMIC); |
3623 | if (skb == NULL) |
3624 | return -ENOMEM; |
3625 | |
3626 | nlh = nlmsg_put(skb, portid: c->portid, seq: c->seq, type: c->event, payload: headlen, flags: 0); |
3627 | err = -EMSGSIZE; |
3628 | if (nlh == NULL) |
3629 | goto out_free_skb; |
3630 | |
3631 | p = nlmsg_data(nlh); |
3632 | if (c->event == XFRM_MSG_DELPOLICY) { |
3633 | struct nlattr *attr; |
3634 | |
3635 | id = nlmsg_data(nlh); |
3636 | memset(id, 0, sizeof(*id)); |
3637 | id->dir = dir; |
3638 | if (c->data.byid) |
3639 | id->index = xp->index; |
3640 | else |
3641 | memcpy(&id->sel, &xp->selector, sizeof(id->sel)); |
3642 | |
3643 | attr = nla_reserve(skb, attrtype: XFRMA_POLICY, attrlen: sizeof(*p)); |
3644 | err = -EMSGSIZE; |
3645 | if (attr == NULL) |
3646 | goto out_free_skb; |
3647 | |
3648 | p = nla_data(nla: attr); |
3649 | } |
3650 | |
3651 | copy_to_user_policy(xp, p, dir); |
3652 | err = copy_to_user_tmpl(xp, skb); |
3653 | if (!err) |
3654 | err = copy_to_user_policy_type(type: xp->type, skb); |
3655 | if (!err) |
3656 | err = xfrm_mark_put(skb, m: &xp->mark); |
3657 | if (!err) |
3658 | err = xfrm_if_id_put(skb, if_id: xp->if_id); |
3659 | if (!err && xp->xdo.dev) |
3660 | err = copy_user_offload(xso: &xp->xdo, skb); |
3661 | if (err) |
3662 | goto out_free_skb; |
3663 | |
3664 | nlmsg_end(skb, nlh); |
3665 | |
3666 | return xfrm_nlmsg_multicast(net, skb, pid: 0, XFRMNLGRP_POLICY); |
3667 | |
3668 | out_free_skb: |
3669 | kfree_skb(skb); |
3670 | return err; |
3671 | } |
3672 | |
3673 | static int xfrm_notify_policy_flush(const struct km_event *c) |
3674 | { |
3675 | struct net *net = c->net; |
3676 | struct nlmsghdr *nlh; |
3677 | struct sk_buff *skb; |
3678 | int err; |
3679 | |
3680 | skb = nlmsg_new(payload: userpolicy_type_attrsize(), GFP_ATOMIC); |
3681 | if (skb == NULL) |
3682 | return -ENOMEM; |
3683 | |
3684 | nlh = nlmsg_put(skb, portid: c->portid, seq: c->seq, XFRM_MSG_FLUSHPOLICY, payload: 0, flags: 0); |
3685 | err = -EMSGSIZE; |
3686 | if (nlh == NULL) |
3687 | goto out_free_skb; |
3688 | err = copy_to_user_policy_type(type: c->data.type, skb); |
3689 | if (err) |
3690 | goto out_free_skb; |
3691 | |
3692 | nlmsg_end(skb, nlh); |
3693 | |
3694 | return xfrm_nlmsg_multicast(net, skb, pid: 0, XFRMNLGRP_POLICY); |
3695 | |
3696 | out_free_skb: |
3697 | kfree_skb(skb); |
3698 | return err; |
3699 | } |
3700 | |
3701 | static int xfrm_send_policy_notify(struct xfrm_policy *xp, int dir, const struct km_event *c) |
3702 | { |
3703 | |
3704 | switch (c->event) { |
3705 | case XFRM_MSG_NEWPOLICY: |
3706 | case XFRM_MSG_UPDPOLICY: |
3707 | case XFRM_MSG_DELPOLICY: |
3708 | return xfrm_notify_policy(xp, dir, c); |
3709 | case XFRM_MSG_FLUSHPOLICY: |
3710 | return xfrm_notify_policy_flush(c); |
3711 | case XFRM_MSG_POLEXPIRE: |
3712 | return xfrm_exp_policy_notify(xp, dir, c); |
3713 | default: |
3714 | printk(KERN_NOTICE "xfrm_user: Unknown Policy event %d\n" , |
3715 | c->event); |
3716 | } |
3717 | |
3718 | return 0; |
3719 | |
3720 | } |
3721 | |
3722 | static inline unsigned int xfrm_report_msgsize(void) |
3723 | { |
3724 | return NLMSG_ALIGN(sizeof(struct xfrm_user_report)); |
3725 | } |
3726 | |
3727 | static int build_report(struct sk_buff *skb, u8 proto, |
3728 | struct xfrm_selector *sel, xfrm_address_t *addr) |
3729 | { |
3730 | struct xfrm_user_report *ur; |
3731 | struct nlmsghdr *nlh; |
3732 | |
3733 | nlh = nlmsg_put(skb, portid: 0, seq: 0, XFRM_MSG_REPORT, payload: sizeof(*ur), flags: 0); |
3734 | if (nlh == NULL) |
3735 | return -EMSGSIZE; |
3736 | |
3737 | ur = nlmsg_data(nlh); |
3738 | ur->proto = proto; |
3739 | memcpy(&ur->sel, sel, sizeof(ur->sel)); |
3740 | |
3741 | if (addr) { |
3742 | int err = nla_put(skb, attrtype: XFRMA_COADDR, attrlen: sizeof(*addr), data: addr); |
3743 | if (err) { |
3744 | nlmsg_cancel(skb, nlh); |
3745 | return err; |
3746 | } |
3747 | } |
3748 | nlmsg_end(skb, nlh); |
3749 | return 0; |
3750 | } |
3751 | |
3752 | static int xfrm_send_report(struct net *net, u8 proto, |
3753 | struct xfrm_selector *sel, xfrm_address_t *addr) |
3754 | { |
3755 | struct sk_buff *skb; |
3756 | int err; |
3757 | |
3758 | skb = nlmsg_new(payload: xfrm_report_msgsize(), GFP_ATOMIC); |
3759 | if (skb == NULL) |
3760 | return -ENOMEM; |
3761 | |
3762 | err = build_report(skb, proto, sel, addr); |
3763 | BUG_ON(err < 0); |
3764 | |
3765 | return xfrm_nlmsg_multicast(net, skb, pid: 0, XFRMNLGRP_REPORT); |
3766 | } |
3767 | |
3768 | static inline unsigned int xfrm_mapping_msgsize(void) |
3769 | { |
3770 | return NLMSG_ALIGN(sizeof(struct xfrm_user_mapping)); |
3771 | } |
3772 | |
3773 | static int build_mapping(struct sk_buff *skb, struct xfrm_state *x, |
3774 | xfrm_address_t *new_saddr, __be16 new_sport) |
3775 | { |
3776 | struct xfrm_user_mapping *um; |
3777 | struct nlmsghdr *nlh; |
3778 | |
3779 | nlh = nlmsg_put(skb, portid: 0, seq: 0, XFRM_MSG_MAPPING, payload: sizeof(*um), flags: 0); |
3780 | if (nlh == NULL) |
3781 | return -EMSGSIZE; |
3782 | |
3783 | um = nlmsg_data(nlh); |
3784 | |
3785 | memcpy(&um->id.daddr, &x->id.daddr, sizeof(um->id.daddr)); |
3786 | um->id.spi = x->id.spi; |
3787 | um->id.family = x->props.family; |
3788 | um->id.proto = x->id.proto; |
3789 | memcpy(&um->new_saddr, new_saddr, sizeof(um->new_saddr)); |
3790 | memcpy(&um->old_saddr, &x->props.saddr, sizeof(um->old_saddr)); |
3791 | um->new_sport = new_sport; |
3792 | um->old_sport = x->encap->encap_sport; |
3793 | um->reqid = x->props.reqid; |
3794 | |
3795 | nlmsg_end(skb, nlh); |
3796 | return 0; |
3797 | } |
3798 | |
3799 | static int xfrm_send_mapping(struct xfrm_state *x, xfrm_address_t *ipaddr, |
3800 | __be16 sport) |
3801 | { |
3802 | struct net *net = xs_net(x); |
3803 | struct sk_buff *skb; |
3804 | int err; |
3805 | |
3806 | if (x->id.proto != IPPROTO_ESP) |
3807 | return -EINVAL; |
3808 | |
3809 | if (!x->encap) |
3810 | return -EINVAL; |
3811 | |
3812 | skb = nlmsg_new(payload: xfrm_mapping_msgsize(), GFP_ATOMIC); |
3813 | if (skb == NULL) |
3814 | return -ENOMEM; |
3815 | |
3816 | err = build_mapping(skb, x, new_saddr: ipaddr, new_sport: sport); |
3817 | BUG_ON(err < 0); |
3818 | |
3819 | return xfrm_nlmsg_multicast(net, skb, pid: 0, XFRMNLGRP_MAPPING); |
3820 | } |
3821 | |
3822 | static bool xfrm_is_alive(const struct km_event *c) |
3823 | { |
3824 | return (bool)xfrm_acquire_is_on(net: c->net); |
3825 | } |
3826 | |
3827 | static struct xfrm_mgr netlink_mgr = { |
3828 | .notify = xfrm_send_state_notify, |
3829 | .acquire = xfrm_send_acquire, |
3830 | .compile_policy = xfrm_compile_policy, |
3831 | .notify_policy = xfrm_send_policy_notify, |
3832 | .report = xfrm_send_report, |
3833 | .migrate = xfrm_send_migrate, |
3834 | .new_mapping = xfrm_send_mapping, |
3835 | .is_alive = xfrm_is_alive, |
3836 | }; |
3837 | |
3838 | static int __net_init xfrm_user_net_init(struct net *net) |
3839 | { |
3840 | struct sock *nlsk; |
3841 | struct netlink_kernel_cfg cfg = { |
3842 | .groups = XFRMNLGRP_MAX, |
3843 | .input = xfrm_netlink_rcv, |
3844 | }; |
3845 | |
3846 | nlsk = netlink_kernel_create(net, NETLINK_XFRM, cfg: &cfg); |
3847 | if (nlsk == NULL) |
3848 | return -ENOMEM; |
3849 | net->xfrm.nlsk_stash = nlsk; /* Don't set to NULL */ |
3850 | rcu_assign_pointer(net->xfrm.nlsk, nlsk); |
3851 | return 0; |
3852 | } |
3853 | |
3854 | static void __net_exit xfrm_user_net_pre_exit(struct net *net) |
3855 | { |
3856 | RCU_INIT_POINTER(net->xfrm.nlsk, NULL); |
3857 | } |
3858 | |
3859 | static void __net_exit xfrm_user_net_exit(struct list_head *net_exit_list) |
3860 | { |
3861 | struct net *net; |
3862 | |
3863 | list_for_each_entry(net, net_exit_list, exit_list) |
3864 | netlink_kernel_release(sk: net->xfrm.nlsk_stash); |
3865 | } |
3866 | |
3867 | static struct pernet_operations xfrm_user_net_ops = { |
3868 | .init = xfrm_user_net_init, |
3869 | .pre_exit = xfrm_user_net_pre_exit, |
3870 | .exit_batch = xfrm_user_net_exit, |
3871 | }; |
3872 | |
3873 | static int __init xfrm_user_init(void) |
3874 | { |
3875 | int rv; |
3876 | |
3877 | printk(KERN_INFO "Initializing XFRM netlink socket\n" ); |
3878 | |
3879 | rv = register_pernet_subsys(&xfrm_user_net_ops); |
3880 | if (rv < 0) |
3881 | return rv; |
3882 | xfrm_register_km(km: &netlink_mgr); |
3883 | return 0; |
3884 | } |
3885 | |
3886 | static void __exit xfrm_user_exit(void) |
3887 | { |
3888 | xfrm_unregister_km(km: &netlink_mgr); |
3889 | unregister_pernet_subsys(&xfrm_user_net_ops); |
3890 | } |
3891 | |
3892 | module_init(xfrm_user_init); |
3893 | module_exit(xfrm_user_exit); |
3894 | MODULE_DESCRIPTION("XFRM User interface" ); |
3895 | MODULE_LICENSE("GPL" ); |
3896 | MODULE_ALIAS_NET_PF_PROTO(PF_NETLINK, NETLINK_XFRM); |
3897 | |