Lines Matching +full:xo +full:- +full:1
1 // SPDX-License-Identifier: GPL-2.0-only
36 #define ESP_SKB_CB(__skb) ((struct esp_skb_cb *)&((__skb)->cb[0]))
56 ~(crypto_tfm_ctx_alignment() - 1); in esp_alloc_tmp()
77 crypto_aead_alignmask(aead) + 1) : tmp + extralen; in esp_tmp_iv()
93 return (void *)ALIGN((unsigned long)(req + 1) + in esp_req_sg()
101 struct crypto_aead *aead = x->data; in esp_ssg_unref()
107 if (x->props.flags & XFRM_STATE_ESN) in esp_ssg_unref()
115 * Skip the first sg which comes from skb->data. in esp_ssg_unref()
117 if (req->src != req->dst) in esp_ssg_unref()
118 for (sg = sg_next(req->src); sg; sg = sg_next(sg)) in esp_ssg_unref()
132 sock_put(esk->sk); in esp_free_tcp_sk()
138 struct xfrm_encap_tmpl *encap = x->encap; in esp_find_tcp_sk()
144 sk = rcu_dereference(x->encap_sk); in esp_find_tcp_sk()
145 if (sk && sk->sk_state == TCP_ESTABLISHED) in esp_find_tcp_sk()
148 spin_lock_bh(&x->lock); in esp_find_tcp_sk()
149 sport = encap->encap_sport; in esp_find_tcp_sk()
150 dport = encap->encap_dport; in esp_find_tcp_sk()
151 nsk = rcu_dereference_protected(x->encap_sk, in esp_find_tcp_sk()
152 lockdep_is_held(&x->lock)); in esp_find_tcp_sk()
156 spin_unlock_bh(&x->lock); in esp_find_tcp_sk()
157 return ERR_PTR(-ENOMEM); in esp_find_tcp_sk()
159 RCU_INIT_POINTER(x->encap_sk, NULL); in esp_find_tcp_sk()
160 esk->sk = sk; in esp_find_tcp_sk()
161 call_rcu(&esk->rcu, esp_free_tcp_sk); in esp_find_tcp_sk()
163 spin_unlock_bh(&x->lock); in esp_find_tcp_sk()
165 sk = inet_lookup_established(xs_net(x), &tcp_hashinfo, x->id.daddr.a4, in esp_find_tcp_sk()
166 dport, x->props.saddr.a4, sport, 0); in esp_find_tcp_sk()
168 return ERR_PTR(-ENOENT); in esp_find_tcp_sk()
172 return ERR_PTR(-EINVAL); in esp_find_tcp_sk()
175 spin_lock_bh(&x->lock); in esp_find_tcp_sk()
176 nsk = rcu_dereference_protected(x->encap_sk, in esp_find_tcp_sk()
177 lockdep_is_held(&x->lock)); in esp_find_tcp_sk()
178 if (encap->encap_sport != sport || in esp_find_tcp_sk()
179 encap->encap_dport != dport) { in esp_find_tcp_sk()
181 sk = nsk ?: ERR_PTR(-EREMCHG); in esp_find_tcp_sk()
185 rcu_assign_pointer(x->encap_sk, sk); in esp_find_tcp_sk()
187 spin_unlock_bh(&x->lock); in esp_find_tcp_sk()
220 struct xfrm_state *x = dst->xfrm; in esp_output_tcp_encap_cb()
237 return err ?: -EINPROGRESS; in esp_output_tail_tcp()
244 return -EOPNOTSUPP; in esp_output_tail_tcp()
250 struct sk_buff *skb = base->data; in esp_output_done()
251 struct xfrm_offload *xo = xfrm_offload(skb); in esp_output_done() local
255 if (xo && (xo->flags & XFRM_DEV_RESUME)) { in esp_output_done()
258 x = sp->xvec[sp->len - 1]; in esp_output_done()
260 x = skb_dst(skb)->xfrm; in esp_output_done()
263 tmp = ESP_SKB_CB(skb)->tmp; in esp_output_done()
267 if (xo && (xo->flags & XFRM_DEV_RESUME)) { in esp_output_done()
274 skb_push(skb, skb->data - skb_mac_header(skb)); in esp_output_done()
279 x->encap && x->encap->encap_type == TCP_ENCAP_ESPINTCP) in esp_output_done()
289 struct ip_esp_hdr *esph = (void *)(skb->data + offset); in esp_restore_header()
290 void *tmp = ESP_SKB_CB(skb)->tmp; in esp_restore_header()
293 esph->seq_no = esph->spi; in esp_restore_header()
294 esph->spi = *seqhi; in esp_restore_header()
299 void *tmp = ESP_SKB_CB(skb)->tmp; in esp_output_restore_header()
302 esp_restore_header(skb, skb_transport_offset(skb) + extra->esphoff - in esp_output_restore_header()
315 if ((x->props.flags & XFRM_STATE_ESN)) { in esp_output_set_extra()
317 struct xfrm_offload *xo = xfrm_offload(skb); in esp_output_set_extra() local
319 if (xo) in esp_output_set_extra()
320 seqhi = xo->seq.hi; in esp_output_set_extra()
322 seqhi = XFRM_SKB_CB(skb)->seq.output.hi; in esp_output_set_extra()
324 extra->esphoff = (unsigned char *)esph - in esp_output_set_extra()
326 esph = (struct ip_esp_hdr *)((unsigned char *)esph - 4); in esp_output_set_extra()
327 extra->seqhi = esph->spi; in esp_output_set_extra()
328 esph->seq_no = htonl(seqhi); in esp_output_set_extra()
331 esph->spi = x->id.spi; in esp_output_set_extra()
338 struct sk_buff *skb = base->data; in esp_output_done_esn()
354 len = skb->len + esp->tailen - skb_transport_offset(skb); in esp_output_udp_encap()
356 return ERR_PTR(-EMSGSIZE); in esp_output_udp_encap()
358 uh = (struct udphdr *)esp->esph; in esp_output_udp_encap()
359 uh->source = sport; in esp_output_udp_encap()
360 uh->dest = dport; in esp_output_udp_encap()
361 uh->len = htons(len); in esp_output_udp_encap()
362 uh->check = 0; in esp_output_udp_encap()
367 udpdata32 = (__be32 *)(uh + 1); in esp_output_udp_encap()
368 udpdata32[0] = udpdata32[1] = 0; in esp_output_udp_encap()
372 return (struct ip_esp_hdr *)(uh + 1); in esp_output_udp_encap()
380 __be16 *lenp = (void *)esp->esph; in esp_output_tcp_encap()
385 len = skb->len + esp->tailen - skb_transport_offset(skb); in esp_output_tcp_encap()
387 return ERR_PTR(-EMSGSIZE); in esp_output_tcp_encap()
397 esph = (struct ip_esp_hdr *)(lenp + 1); in esp_output_tcp_encap()
406 return ERR_PTR(-EOPNOTSUPP); in esp_output_tcp_encap()
413 struct xfrm_encap_tmpl *encap = x->encap; in esp_output_encap()
418 spin_lock_bh(&x->lock); in esp_output_encap()
419 sport = encap->encap_sport; in esp_output_encap()
420 dport = encap->encap_dport; in esp_output_encap()
421 encap_type = encap->encap_type; in esp_output_encap()
422 spin_unlock_bh(&x->lock); in esp_output_encap()
438 esp->esph = esph; in esp_output_encap()
450 int tailen = esp->tailen; in esp_output_head()
452 /* this is non-NULL only with TCP/UDP Encapsulation */ in esp_output_head()
453 if (x->encap) { in esp_output_head()
461 ALIGN(skb->data_len, L1_CACHE_BYTES) > PAGE_SIZE) in esp_output_head()
466 nfrags = 1; in esp_output_head()
471 } else if ((skb_shinfo(skb)->nr_frags < MAX_SKB_FRAGS) in esp_output_head()
474 struct sock *sk = skb->sk; in esp_output_head()
475 struct page_frag *pfrag = &x->xfrag; in esp_output_head()
477 esp->inplace = false; in esp_output_head()
481 spin_lock_bh(&x->lock); in esp_output_head()
484 spin_unlock_bh(&x->lock); in esp_output_head()
488 page = pfrag->page; in esp_output_head()
491 tail = page_address(page) + pfrag->offset; in esp_output_head()
493 esp_output_fill_trailer(tail, esp->tfclen, esp->plen, esp->proto); in esp_output_head()
495 nfrags = skb_shinfo(skb)->nr_frags; in esp_output_head()
497 __skb_fill_page_desc(skb, nfrags, page, pfrag->offset, in esp_output_head()
499 skb_shinfo(skb)->nr_frags = ++nfrags; in esp_output_head()
501 pfrag->offset = pfrag->offset + allocsize; in esp_output_head()
503 spin_unlock_bh(&x->lock); in esp_output_head()
507 skb->len += tailen; in esp_output_head()
508 skb->data_len += tailen; in esp_output_head()
509 skb->truesize += tailen; in esp_output_head()
511 refcount_add(tailen, &sk->sk_wmem_alloc); in esp_output_head()
518 esph_offset = (unsigned char *)esp->esph - skb_transport_header(skb); in esp_output_head()
524 esp->esph = (struct ip_esp_hdr *)(skb_transport_header(skb) + esph_offset); in esp_output_head()
527 esp_output_fill_trailer(tail, esp->tfclen, esp->plen, esp->proto); in esp_output_head()
549 int err = -ENOMEM; in esp_output_tail()
554 if (x->props.flags & XFRM_STATE_ESN) { in esp_output_tail()
559 aead = x->data; in esp_output_tail()
563 tmp = esp_alloc_tmp(aead, esp->nfrags + 2, extralen); in esp_output_tail()
572 if (esp->inplace) in esp_output_tail()
575 dsg = &sg[esp->nfrags]; in esp_output_tail()
577 esph = esp_output_set_extra(skb, x, esp->esph, extra); in esp_output_tail()
578 esp->esph = esph; in esp_output_tail()
580 sg_init_table(sg, esp->nfrags); in esp_output_tail()
582 (unsigned char *)esph - skb->data, in esp_output_tail()
583 assoclen + ivlen + esp->clen + alen); in esp_output_tail()
587 if (!esp->inplace) { in esp_output_tail()
589 struct page_frag *pfrag = &x->xfrag; in esp_output_tail()
591 allocsize = ALIGN(skb->data_len, L1_CACHE_BYTES); in esp_output_tail()
593 spin_lock_bh(&x->lock); in esp_output_tail()
595 spin_unlock_bh(&x->lock); in esp_output_tail()
599 skb_shinfo(skb)->nr_frags = 1; in esp_output_tail()
601 page = pfrag->page; in esp_output_tail()
604 __skb_fill_page_desc(skb, 0, page, pfrag->offset, skb->data_len); in esp_output_tail()
605 pfrag->offset = pfrag->offset + allocsize; in esp_output_tail()
606 spin_unlock_bh(&x->lock); in esp_output_tail()
608 sg_init_table(dsg, skb_shinfo(skb)->nr_frags + 1); in esp_output_tail()
610 (unsigned char *)esph - skb->data, in esp_output_tail()
611 assoclen + ivlen + esp->clen + alen); in esp_output_tail()
616 if ((x->props.flags & XFRM_STATE_ESN)) in esp_output_tail()
621 aead_request_set_crypt(req, sg, dsg, ivlen + esp->clen, iv); in esp_output_tail()
625 memcpy(iv + ivlen - min(ivlen, 8), (u8 *)&esp->seqno + 8 - min(ivlen, 8), in esp_output_tail()
628 ESP_SKB_CB(skb)->tmp = tmp; in esp_output_tail()
632 case -EINPROGRESS: in esp_output_tail()
635 case -ENOSPC: in esp_output_tail()
640 if ((x->props.flags & XFRM_STATE_ESN)) in esp_output_tail()
647 if (!err && x->encap && x->encap->encap_type == TCP_ENCAP_ESPINTCP) in esp_output_tail()
672 aead = x->data; in esp_output()
676 if (x->tfcpad) { in esp_output()
680 padto = min(x->tfcpad, xfrm_state_mtu(x, dst->child_mtu_cached)); in esp_output()
681 if (skb->len < padto) in esp_output()
682 esp.tfclen = padto - skb->len; in esp_output()
685 esp.clen = ALIGN(skb->len + 2 + esp.tfclen, blksize); in esp_output()
686 esp.plen = esp.clen - skb->len - esp.tfclen; in esp_output()
696 esph->spi = x->id.spi; in esp_output()
698 esph->seq_no = htonl(XFRM_SKB_CB(skb)->seq.output.low); in esp_output()
699 esp.seqno = cpu_to_be64(XFRM_SKB_CB(skb)->seq.output.low + in esp_output()
700 ((u64)XFRM_SKB_CB(skb)->seq.output.hi << 32)); in esp_output()
702 skb_push(skb, -skb_network_offset(skb)); in esp_output()
710 struct xfrm_offload *xo = xfrm_offload(skb); in esp_remove_trailer() local
711 struct crypto_aead *aead = x->data; in esp_remove_trailer()
720 elen = skb->len - hlen; in esp_remove_trailer()
722 if (xo && (xo->flags & XFRM_ESP_NO_TRAILER)) { in esp_remove_trailer()
723 ret = xo->proto; in esp_remove_trailer()
727 if (skb_copy_bits(skb, skb->len - alen - 2, nexthdr, 2)) in esp_remove_trailer()
730 ret = -EINVAL; in esp_remove_trailer()
734 padlen + 2, elen - alen); in esp_remove_trailer()
739 if (skb->ip_summed == CHECKSUM_COMPLETE) { in esp_remove_trailer()
740 csumdiff = skb_checksum(skb, skb->len - trimlen, trimlen, 0); in esp_remove_trailer()
741 skb->csum = csum_block_sub(skb->csum, csumdiff, in esp_remove_trailer()
742 skb->len - trimlen); in esp_remove_trailer()
744 pskb_trim(skb, skb->len - trimlen); in esp_remove_trailer()
746 ret = nexthdr[1]; in esp_remove_trailer()
756 struct xfrm_offload *xo = xfrm_offload(skb); in esp_input_done2() local
757 struct crypto_aead *aead = x->data; in esp_input_done2()
761 if (!xo || (xo && !(xo->flags & CRYPTO_DONE))) in esp_input_done2()
762 kfree(ESP_SKB_CB(skb)->tmp); in esp_input_done2()
772 ihl = iph->ihl * 4; in esp_input_done2()
774 if (x->encap) { in esp_input_done2()
775 struct xfrm_encap_tmpl *encap = x->encap; in esp_input_done2()
780 switch (x->encap->encap_type) { in esp_input_done2()
782 source = th->source; in esp_input_done2()
786 source = uh->source; in esp_input_done2()
789 WARN_ON_ONCE(1); in esp_input_done2()
790 err = -EINVAL; in esp_input_done2()
795 * 1) if the NAT-T peer's IP or port changed then in esp_input_done2()
800 if (iph->saddr != x->props.saddr.a4 || in esp_input_done2()
801 source != encap->encap_sport) { in esp_input_done2()
804 ipaddr.a4 = iph->saddr; in esp_input_done2()
818 * of NAT-T in Transport Mode, or in esp_input_done2()
819 * perform other post-processing fixes in esp_input_done2()
820 * as per draft-ietf-ipsec-udp-encaps-06, in esp_input_done2()
823 if (x->props.mode == XFRM_MODE_TRANSPORT) in esp_input_done2()
824 skb->ip_summed = CHECKSUM_UNNECESSARY; in esp_input_done2()
828 if (x->props.mode == XFRM_MODE_TUNNEL) in esp_input_done2()
831 skb_set_transport_header(skb, -ihl); in esp_input_done2()
835 err = -EINVAL; in esp_input_done2()
844 struct sk_buff *skb = base->data; in esp_input_done()
864 if ((x->props.flags & XFRM_STATE_ESN)) { in esp_input_set_header()
866 *seqhi = esph->spi; in esp_input_set_header()
867 esph->spi = esph->seq_no; in esp_input_set_header()
868 esph->seq_no = XFRM_SKB_CB(skb)->seq.input.hi; in esp_input_set_header()
874 struct sk_buff *skb = base->data; in esp_input_done_esn()
881 * Note: detecting truncated vs. non-truncated authentication data is very
887 struct crypto_aead *aead = x->data; in esp_input()
891 int elen = skb->len - sizeof(struct ip_esp_hdr) - ivlen; in esp_input()
899 int err = -EINVAL; in esp_input()
910 if (x->props.flags & XFRM_STATE_ESN) { in esp_input()
917 nfrags = 1; in esp_input()
921 nfrags = skb_shinfo(skb)->nr_frags; in esp_input()
935 err = -ENOMEM; in esp_input()
940 ESP_SKB_CB(skb)->tmp = tmp; in esp_input()
949 err = skb_to_sgvec(skb, sg, 0, skb->len); in esp_input()
955 skb->ip_summed = CHECKSUM_NONE; in esp_input()
957 if ((x->props.flags & XFRM_STATE_ESN)) in esp_input()
966 if (err == -EINPROGRESS) in esp_input()
969 if ((x->props.flags & XFRM_STATE_ESN)) in esp_input()
980 struct net *net = dev_net(skb->dev); in esp4_err()
981 const struct iphdr *iph = (const struct iphdr *)skb->data; in esp4_err()
982 struct ip_esp_hdr *esph = (struct ip_esp_hdr *)(skb->data+(iph->ihl<<2)); in esp4_err()
985 switch (icmp_hdr(skb)->type) { in esp4_err()
987 if (icmp_hdr(skb)->code != ICMP_FRAG_NEEDED) in esp4_err()
995 x = xfrm_state_lookup(net, skb->mark, (const xfrm_address_t *)&iph->daddr, in esp4_err()
996 esph->spi, IPPROTO_ESP, AF_INET); in esp4_err()
1000 if (icmp_hdr(skb)->type == ICMP_DEST_UNREACH) in esp4_err()
1011 struct crypto_aead *aead = x->data; in esp_destroy()
1025 err = -ENAMETOOLONG; in esp_init_aead()
1027 x->geniv, x->aead->alg_name) >= CRYPTO_MAX_ALG_NAME) in esp_init_aead()
1035 x->data = aead; in esp_init_aead()
1037 err = crypto_aead_setkey(aead, x->aead->alg_key, in esp_init_aead()
1038 (x->aead->alg_key_len + 7) / 8); in esp_init_aead()
1042 err = crypto_aead_setauthsize(aead, x->aead->alg_icv_len / 8); in esp_init_aead()
1061 err = -EINVAL; in esp_init_authenc()
1062 if (!x->ealg) in esp_init_authenc()
1065 err = -ENAMETOOLONG; in esp_init_authenc()
1067 if ((x->props.flags & XFRM_STATE_ESN)) { in esp_init_authenc()
1070 x->geniv ?: "", x->geniv ? "(" : "", in esp_init_authenc()
1071 x->aalg ? x->aalg->alg_name : "digest_null", in esp_init_authenc()
1072 x->ealg->alg_name, in esp_init_authenc()
1073 x->geniv ? ")" : "") >= CRYPTO_MAX_ALG_NAME) in esp_init_authenc()
1078 x->geniv ?: "", x->geniv ? "(" : "", in esp_init_authenc()
1079 x->aalg ? x->aalg->alg_name : "digest_null", in esp_init_authenc()
1080 x->ealg->alg_name, in esp_init_authenc()
1081 x->geniv ? ")" : "") >= CRYPTO_MAX_ALG_NAME) in esp_init_authenc()
1090 x->data = aead; in esp_init_authenc()
1092 keylen = (x->aalg ? (x->aalg->alg_key_len + 7) / 8 : 0) + in esp_init_authenc()
1093 (x->ealg->alg_key_len + 7) / 8 + RTA_SPACE(sizeof(*param)); in esp_init_authenc()
1094 err = -ENOMEM; in esp_init_authenc()
1101 rta->rta_type = CRYPTO_AUTHENC_KEYA_PARAM; in esp_init_authenc()
1102 rta->rta_len = RTA_LENGTH(sizeof(*param)); in esp_init_authenc()
1106 if (x->aalg) { in esp_init_authenc()
1109 memcpy(p, x->aalg->alg_key, (x->aalg->alg_key_len + 7) / 8); in esp_init_authenc()
1110 p += (x->aalg->alg_key_len + 7) / 8; in esp_init_authenc()
1112 aalg_desc = xfrm_aalg_get_byname(x->aalg->alg_name, 0); in esp_init_authenc()
1115 err = -EINVAL; in esp_init_authenc()
1116 if (aalg_desc->uinfo.auth.icv_fullbits / 8 != in esp_init_authenc()
1119 x->aalg->alg_name, in esp_init_authenc()
1121 aalg_desc->uinfo.auth.icv_fullbits / 8); in esp_init_authenc()
1126 aead, x->aalg->alg_trunc_len / 8); in esp_init_authenc()
1131 param->enckeylen = cpu_to_be32((x->ealg->alg_key_len + 7) / 8); in esp_init_authenc()
1132 memcpy(p, x->ealg->alg_key, (x->ealg->alg_key_len + 7) / 8); in esp_init_authenc()
1149 x->data = NULL; in esp_init_state()
1151 if (x->aead) in esp_init_state()
1159 aead = x->data; in esp_init_state()
1161 x->props.header_len = sizeof(struct ip_esp_hdr) + in esp_init_state()
1163 if (x->props.mode == XFRM_MODE_TUNNEL) in esp_init_state()
1164 x->props.header_len += sizeof(struct iphdr); in esp_init_state()
1165 else if (x->props.mode == XFRM_MODE_BEET && x->sel.family != AF_INET6) in esp_init_state()
1166 x->props.header_len += IPV4_BEET_PHMAXLEN; in esp_init_state()
1167 if (x->encap) { in esp_init_state()
1168 struct xfrm_encap_tmpl *encap = x->encap; in esp_init_state()
1170 switch (encap->encap_type) { in esp_init_state()
1172 err = -EINVAL; in esp_init_state()
1175 x->props.header_len += sizeof(struct udphdr); in esp_init_state()
1178 x->props.header_len += sizeof(struct udphdr) + 2 * sizeof(u32); in esp_init_state()
1185 x->props.header_len += 2; in esp_init_state()
1192 x->props.trailer_len = align + 1 + crypto_aead_authsize(aead); in esp_init_state()
1227 return -EAGAIN; in esp4_init()
1232 return -EAGAIN; in esp4_init()