Lines Matching refs:rx_sa

142 		struct macsec_rx_sa *rx_sa;  member
725 struct macsec_rx_sa *rx_sa = macsec_skb_cb(skb)->rx_sa; in macsec_post_decrypt() local
726 struct pcpu_rx_sc_stats *rxsc_stats = this_cpu_ptr(rx_sa->sc->stats); in macsec_post_decrypt()
730 spin_lock(&rx_sa->lock); in macsec_post_decrypt()
731 if (rx_sa->next_pn_halves.lower >= secy->replay_window) in macsec_post_decrypt()
732 lowest_pn = rx_sa->next_pn_halves.lower - secy->replay_window; in macsec_post_decrypt()
739 spin_unlock(&rx_sa->lock); in macsec_post_decrypt()
756 spin_unlock(&rx_sa->lock); in macsec_post_decrypt()
770 this_cpu_inc(rx_sa->stats->InPktsInvalid); in macsec_post_decrypt()
783 this_cpu_inc(rx_sa->stats->InPktsOK); in macsec_post_decrypt()
788 if (pn + 1 > rx_sa->next_pn_halves.lower) { in macsec_post_decrypt()
789 rx_sa->next_pn_halves.lower = pn + 1; in macsec_post_decrypt()
791 !pn_same_half(pn, rx_sa->next_pn_halves.lower)) { in macsec_post_decrypt()
792 rx_sa->next_pn_halves.upper++; in macsec_post_decrypt()
793 rx_sa->next_pn_halves.lower = pn + 1; in macsec_post_decrypt()
796 spin_unlock(&rx_sa->lock); in macsec_post_decrypt()
836 struct macsec_rx_sa *rx_sa = macsec_skb_cb(skb)->rx_sa; in macsec_decrypt_done() local
837 struct macsec_rx_sc *rx_sc = rx_sa->sc; in macsec_decrypt_done()
865 macsec_rxsa_put(rx_sa); in macsec_decrypt_done()
872 struct macsec_rx_sa *rx_sa, in macsec_decrypt() argument
895 req = macsec_alloc_req(rx_sa->key.tfm, &iv, &sg, ret); in macsec_decrypt()
905 pn_t recovered_pn = rx_sa->next_pn_halves; in macsec_decrypt()
908 if (hdr_pn < rx_sa->next_pn_halves.lower && in macsec_decrypt()
909 !pn_same_half(hdr_pn, rx_sa->next_pn_halves.lower)) in macsec_decrypt()
912 macsec_fill_iv_xpn(iv, rx_sa->ssci, recovered_pn.full64, in macsec_decrypt()
913 rx_sa->key.salt); in macsec_decrypt()
1078 struct macsec_rx_sa *rx_sa; in macsec_handle_frame() local
1161 rx_sa = macsec_rxsa_get(rx_sc->sa[macsec_skb_cb(skb)->assoc_num]); in macsec_handle_frame()
1162 if (!rx_sa) { in macsec_handle_frame()
1190 spin_lock(&rx_sa->lock); in macsec_handle_frame()
1191 late = rx_sa->next_pn_halves.lower >= secy->replay_window && in macsec_handle_frame()
1192 hdr_pn < (rx_sa->next_pn_halves.lower - secy->replay_window); in macsec_handle_frame()
1195 late = late && pn_same_half(rx_sa->next_pn_halves.lower, hdr_pn); in macsec_handle_frame()
1196 spin_unlock(&rx_sa->lock); in macsec_handle_frame()
1206 macsec_skb_cb(skb)->rx_sa = rx_sa; in macsec_handle_frame()
1211 skb = macsec_decrypt(skb, dev, rx_sa, sci, secy); in macsec_handle_frame()
1216 macsec_rxsa_put(rx_sa); in macsec_handle_frame()
1232 if (rx_sa) in macsec_handle_frame()
1233 macsec_rxsa_put(rx_sa); in macsec_handle_frame()
1250 macsec_rxsa_put(rx_sa); in macsec_handle_frame()
1331 static int init_rx_sa(struct macsec_rx_sa *rx_sa, char *sak, int key_len, in init_rx_sa() argument
1334 rx_sa->stats = alloc_percpu(struct macsec_rx_sa_stats); in init_rx_sa()
1335 if (!rx_sa->stats) in init_rx_sa()
1338 rx_sa->key.tfm = macsec_alloc_tfm(sak, key_len, icv_len); in init_rx_sa()
1339 if (IS_ERR(rx_sa->key.tfm)) { in init_rx_sa()
1340 free_percpu(rx_sa->stats); in init_rx_sa()
1341 return PTR_ERR(rx_sa->key.tfm); in init_rx_sa()
1344 rx_sa->ssci = MACSEC_UNDEF_SSCI; in init_rx_sa()
1345 rx_sa->active = false; in init_rx_sa()
1346 rx_sa->next_pn = 1; in init_rx_sa()
1347 refcount_set(&rx_sa->refcnt, 1); in init_rx_sa()
1348 spin_lock_init(&rx_sa->lock); in init_rx_sa()
1353 static void clear_rx_sa(struct macsec_rx_sa *rx_sa) in clear_rx_sa() argument
1355 rx_sa->active = false; in clear_rx_sa()
1357 macsec_rxsa_put(rx_sa); in clear_rx_sa()
1582 struct macsec_rx_sa *rx_sa; in get_rxsa_from_nl() local
1595 rx_sa = rtnl_dereference(rx_sc->sa[*assoc_num]); in get_rxsa_from_nl()
1596 if (!rx_sa) in get_rxsa_from_nl()
1600 return rx_sa; in get_rxsa_from_nl()
1719 struct macsec_rx_sa *rx_sa; in macsec_add_rxsa() local
1778 rx_sa = rtnl_dereference(rx_sc->sa[assoc_num]); in macsec_add_rxsa()
1779 if (rx_sa) { in macsec_add_rxsa()
1784 rx_sa = kmalloc(sizeof(*rx_sa), GFP_KERNEL); in macsec_add_rxsa()
1785 if (!rx_sa) { in macsec_add_rxsa()
1790 err = init_rx_sa(rx_sa, nla_data(tb_sa[MACSEC_SA_ATTR_KEY]), in macsec_add_rxsa()
1793 kfree(rx_sa); in macsec_add_rxsa()
1799 spin_lock_bh(&rx_sa->lock); in macsec_add_rxsa()
1800 rx_sa->next_pn = nla_get_u64(tb_sa[MACSEC_SA_ATTR_PN]); in macsec_add_rxsa()
1801 spin_unlock_bh(&rx_sa->lock); in macsec_add_rxsa()
1805 rx_sa->active = !!nla_get_u8(tb_sa[MACSEC_SA_ATTR_ACTIVE]); in macsec_add_rxsa()
1807 rx_sa->sc = rx_sc; in macsec_add_rxsa()
1821 ctx.sa.rx_sa = rx_sa; in macsec_add_rxsa()
1833 rx_sa->ssci = nla_get_ssci(tb_sa[MACSEC_SA_ATTR_SSCI]); in macsec_add_rxsa()
1834 nla_memcpy(rx_sa->key.salt.bytes, tb_sa[MACSEC_SA_ATTR_SALT], in macsec_add_rxsa()
1838 nla_memcpy(rx_sa->key.id, tb_sa[MACSEC_SA_ATTR_KEYID], MACSEC_KEYID_LEN); in macsec_add_rxsa()
1839 rcu_assign_pointer(rx_sc->sa[assoc_num], rx_sa); in macsec_add_rxsa()
1846 macsec_rxsa_put(rx_sa); in macsec_add_rxsa()
2101 struct macsec_rx_sa *rx_sa; in macsec_del_rxsa() local
2117 rx_sa = get_rxsa_from_nl(genl_info_net(info), attrs, tb_rxsc, tb_sa, in macsec_del_rxsa()
2119 if (IS_ERR(rx_sa)) { in macsec_del_rxsa()
2121 return PTR_ERR(rx_sa); in macsec_del_rxsa()
2124 if (rx_sa->active) { in macsec_del_rxsa()
2141 ctx.sa.rx_sa = rx_sa; in macsec_del_rxsa()
2150 clear_rx_sa(rx_sa); in macsec_del_rxsa()
2408 struct macsec_rx_sa *rx_sa; in macsec_upd_rxsa() local
2431 rx_sa = get_rxsa_from_nl(genl_info_net(info), attrs, tb_rxsc, tb_sa, in macsec_upd_rxsa()
2433 if (IS_ERR(rx_sa)) { in macsec_upd_rxsa()
2435 return PTR_ERR(rx_sa); in macsec_upd_rxsa()
2449 spin_lock_bh(&rx_sa->lock); in macsec_upd_rxsa()
2450 prev_pn = rx_sa->next_pn_halves; in macsec_upd_rxsa()
2451 rx_sa->next_pn = nla_get_u64(tb_sa[MACSEC_SA_ATTR_PN]); in macsec_upd_rxsa()
2452 spin_unlock_bh(&rx_sa->lock); in macsec_upd_rxsa()
2455 was_active = rx_sa->active; in macsec_upd_rxsa()
2457 rx_sa->active = nla_get_u8(tb_sa[MACSEC_SA_ATTR_ACTIVE]); in macsec_upd_rxsa()
2471 ctx.sa.rx_sa = rx_sa; in macsec_upd_rxsa()
2484 spin_lock_bh(&rx_sa->lock); in macsec_upd_rxsa()
2485 rx_sa->next_pn_halves = prev_pn; in macsec_upd_rxsa()
2486 spin_unlock_bh(&rx_sa->lock); in macsec_upd_rxsa()
2488 rx_sa->active = was_active; in macsec_upd_rxsa()
2706 struct macsec_rx_sa *rx_sa, in get_rx_sa_stats() argument
2720 ctx.sa.rx_sa = rx_sa; in get_rx_sa_stats()
2731 per_cpu_ptr(rx_sa->stats, cpu); in get_rx_sa_stats()
3182 struct macsec_rx_sa *rx_sa = rtnl_dereference(rx_sc->sa[i]); in dump_secy() local
3187 if (!rx_sa) in dump_secy()
3207 get_rx_sa_stats(dev, rx_sc, i, rx_sa, &rx_sa_stats); in dump_secy()
3218 pn = rx_sa->next_pn; in dump_secy()
3221 pn = rx_sa->next_pn_halves.lower; in dump_secy()
3227 nla_put(skb, MACSEC_SA_ATTR_KEYID, MACSEC_KEYID_LEN, rx_sa->key.id) || in dump_secy()
3228 (secy->xpn && nla_put_ssci(skb, MACSEC_SA_ATTR_SSCI, rx_sa->ssci)) || in dump_secy()
3229 nla_put_u8(skb, MACSEC_SA_ATTR_ACTIVE, rx_sa->active)) { in dump_secy()