Home
last modified time | relevance | path

Searched refs:sock_net (Results 1 – 25 of 242) sorted by relevance

12345678910

/OK3568_Linux_fs/kernel/net/ipv4/
H A Dtcp_fastopen.c335 __NET_INC_STATS(sock_net(sk), in tcp_fastopen_queue_check()
352 return (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen) & flag) || in tcp_fastopen_no_cookie()
367 int tcp_fastopen = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen); in tcp_try_fastopen()
373 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPFASTOPENCOOKIEREQD); in tcp_try_fastopen()
392 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen()
410 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen()
415 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen()
419 NET_INC_STATS(sock_net(sk), in tcp_try_fastopen()
507 struct net *net = sock_net(sk); in tcp_fastopen_active_disable()
509 if (!READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_fastopen_blackhole_timeout)) in tcp_fastopen_active_disable()
[all …]
H A Dtcp_timer.c75 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONTIMEOUT); in tcp_write_err()
126 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPABORTONMEMORY); in tcp_out_of_resources()
130 if (!check_net(sock_net(sk))) { in tcp_out_of_resources()
146 int retries = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_orphan_retries); /* May be zero. */ in tcp_orphan_retries()
162 const struct net *net = sock_net(sk); in tcp_mtu_probing()
235 struct net *net = sock_net(sk); in tcp_write_timeout()
282 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPTIMEOUTREHASH); in tcp_write_timeout()
318 __NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKS); in tcp_delack_timer_handler()
346 __NET_INC_STATS(sock_net(sk), LINUX_MIB_DELAYEDACKLOCKED); in tcp_delack_timer()
383 max_probes = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_retries2); in tcp_probe_timer()
[all …]
H A Dtcp_input.c216 dev = dev_get_by_index_rcu(sock_net(sk), skb->skb_iif); in tcp_gro_dev_warn()
429 min(sndmem, READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_wmem[2]))); in tcp_sndbuf_expand()
464 int window = tcp_win_from_space(sk, READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_rmem[2])) >> 1; in __tcp_grow_window()
529 int tcp_app_win = READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_app_win); in tcp_init_buffer_space()
568 struct net *net = sock_net(sk); in tcp_clamp_window()
720 if (READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_moderate_rcvbuf) && in tcp_rcv_space_adjust()
741 READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_rmem[2])); in tcp_rcv_space_adjust()
906 rate *= sock_net(sk)->ipv4.sysctl_tcp_pacing_ss_ratio; in tcp_update_pacing_rate()
908 rate *= sock_net(sk)->ipv4.sysctl_tcp_pacing_ca_ratio; in tcp_update_pacing_rate()
1038 READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_max_reordering)); in tcp_check_sack_reordering()
[all …]
H A Dsyncookies.c345 if (!READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_syncookies) || in cookie_v4_check()
354 __NET_INC_STATS(sock_net(sk), LINUX_MIB_SYNCOOKIESFAILED); in cookie_v4_check()
358 __NET_INC_STATS(sock_net(sk), LINUX_MIB_SYNCOOKIESRECV); in cookie_v4_check()
362 tcp_parse_options(sock_net(sk), skb, &tcp_opt, 0, NULL); in cookie_v4_check()
365 tsoff = secure_tcp_ts_off(sock_net(sk), in cookie_v4_check()
371 if (!cookie_timestamp_decode(sock_net(sk), &tcp_opt)) in cookie_v4_check()
408 RCU_INIT_POINTER(ireq->ireq_opt, tcp_v4_save_options(sock_net(sk), skb)); in cookie_v4_check()
429 rt = ip_route_output_key(sock_net(sk), &fl4); in cookie_v4_check()
449 ireq->ecn_ok = cookie_ecn_ok(&tcp_opt, sock_net(sk), &rt->dst); in cookie_v4_check()
H A Dinet_hashtables.c51 return inet6_ehashfn(sock_net(sk), in sk_ehashfn()
55 return inet_ehashfn(sock_net(sk), in sk_ehashfn()
109 const int bhash = inet_bhashfn(sock_net(sk), inet_sk(sk)->inet_num, in __inet_put_port()
135 const int bhash = inet_bhashfn(sock_net(sk), port, in __inet_inherit_port()
156 if (net_eq(ib_net(tb), sock_net(sk)) && in __inet_inherit_port()
162 sock_net(sk), head, port, in __inet_inherit_port()
185 hash = ipv6_portaddr_hash(sock_net(sk), in inet_lhash2_bucket_sk()
190 hash = ipv4_portaddr_hash(sock_net(sk), in inet_lhash2_bucket_sk()
238 if (net_eq(sock_net(sk), net) && sk->sk_num == hnum && in compute_score()
448 struct net *net = sock_net(sk); in __inet_check_established()
[all …]
H A Dtcp_minisocks.c258 struct inet_timewait_death_row *tcp_death_row = &sock_net(sk)->ipv4.tcp_death_row; in tcp_time_wait()
336 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPTIMEWAITOVERFLOW); in tcp_time_wait()
553 __TCP_INC_STATS(sock_net(sk), TCP_MIB_PASSIVEOPENS); in tcp_create_openreq_child()
583 tcp_parse_options(sock_net(sk), skb, &tmp_opt, 0, NULL); in tcp_check_req()
625 if (!tcp_oow_rate_limited(sock_net(sk), skb, in tcp_check_req()
715 !tcp_oow_rate_limited(sock_net(sk), skb, in tcp_check_req()
720 __NET_INC_STATS(sock_net(sk), LINUX_MIB_PAWSESTABREJECTED); in tcp_check_req()
739 __TCP_INC_STATS(sock_net(sk), TCP_MIB_ATTEMPTFAILS); in tcp_check_req()
762 __NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPDEFERACCEPTDROP); in tcp_check_req()
789 if (!sock_net(sk)->ipv4.sysctl_tcp_abort_on_overflow) { in tcp_check_req()
[all …]
H A Dtcp_output.c83 NET_ADD_STATS(sock_net(sk), LINUX_MIB_TCPORIGDATASENT, in tcp_event_new_data_sent()
186 NET_ADD_STATS(sock_net(sk), LINUX_MIB_TCPACKCOMPRESSED, in tcp_event_ack_sent()
230 if (sock_net(sk)->ipv4.sysctl_tcp_workaround_signed_windows) in tcp_select_initial_window()
241 space = max_t(u32, space, READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_rmem[2])); in tcp_select_initial_window()
274 NET_INC_STATS(sock_net(sk), in tcp_select_window()
285 sock_net(sk)->ipv4.sysctl_tcp_workaround_signed_windows) in tcp_select_window()
297 NET_INC_STATS(sock_net(sk), in tcp_select_window()
300 NET_INC_STATS(sock_net(sk), LINUX_MIB_TCPFROMZEROWINDOWADV); in tcp_select_window()
324 bool use_ecn = sock_net(sk)->ipv4.sysctl_tcp_ecn == 1 || in tcp_ecn_send_syn()
346 if (sock_net(sk)->ipv4.sysctl_tcp_ecn_fallback) in tcp_ecn_clear_syn()
[all …]
H A Dudp.c142 if (net_eq(sock_net(sk2), net) && in udp_lib_lport_inuse()
179 if (net_eq(sock_net(sk2), net) && in udp_lib_lport_inuse2()
202 struct net *net = sock_net(sk); in udp_reuseport_add_sock()
207 if (net_eq(sock_net(sk2), net) && in udp_reuseport_add_sock()
237 struct net *net = sock_net(sk); in udp_lib_get_port()
323 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in udp_lib_get_port()
349 ipv4_portaddr_hash(sock_net(sk), htonl(INADDR_ANY), snum); in udp_v4_get_port()
351 ipv4_portaddr_hash(sock_net(sk), inet_sk(sk)->inet_rcv_saddr, 0); in udp_v4_get_port()
367 if (!net_eq(sock_net(sk), net) || in compute_score()
582 if (!net_eq(sock_net(sk), net) || in __udp_is_mcast_sock()
[all …]
H A Ddatagram.c42 if (!oif || netif_index_is_l3_master(sock_net(sk), oif)) in __ip4_datagram_connect()
55 IP_INC_STATS(sock_net(sk), IPSTATS_MIB_OUTNOROUTES); in __ip4_datagram_connect()
119 rt = ip_route_output_ports(sock_net(sk), &fl4, sk, daddr, in ip4_datagram_release_cb()
H A Draw_diag.c99 net = sock_net(in_skb->sk); in raw_diag_dump_one()
146 struct net *net = sock_net(skb->sk); in raw_diag_dump()
167 if (!net_eq(sock_net(sk), net)) in raw_diag_dump()
204 struct net *net = sock_net(in_skb->sk); in raw_diag_destroy()
H A Draw.c102 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in raw_hash_sk()
115 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, -1); in raw_unhash_sk()
127 if (net_eq(sock_net(sk), net) && inet->inet_num == num && in __raw_v4_lookup()
350 struct net *net = sock_net(sk); in raw_send_hdrinc()
502 struct net *net = sock_net(sk); in raw_sendmsg()
607 if (!ipc.oif || netif_index_is_l3_master(sock_net(sk), ipc.oif)) in raw_sendmsg()
621 ipc.oif == l3mdev_master_ifindex_by_index(sock_net(sk), in raw_sendmsg()
729 tb_id = l3mdev_fib_table_by_index(sock_net(sk), in raw_bind()
732 chk_addr_ret = inet_addr_type_table(sock_net(sk), addr->sin_addr.s_addr, in raw_bind()
982 if (sock_net(sk) == seq_file_net(seq)) in raw_get_first()
[all …]
/OK3568_Linux_fs/kernel/net/can/
H A Draw.c275 if (!net_eq(dev_net(dev), sock_net(sk))) in raw_notify()
390 dev = dev_get_by_index(sock_net(sk), ro->ifindex); in raw_release()
396 raw_disable_allfilters(sock_net(sk), NULL, sk); in raw_release()
439 dev = dev_get_by_index(sock_net(sk), addr->can_ifindex); in raw_bind()
455 err = raw_enable_allfilters(sock_net(sk), dev, sk); in raw_bind()
461 err = raw_enable_allfilters(sock_net(sk), NULL, sk); in raw_bind()
470 dev = dev_get_by_index(sock_net(sk), in raw_bind()
478 raw_disable_allfilters(sock_net(sk), NULL, sk); in raw_bind()
553 dev = dev_get_by_index(sock_net(sk), ro->ifindex); in raw_setsockopt()
565 err = raw_enable_filters(sock_net(sk), dev, sk, in raw_setsockopt()
[all …]
/OK3568_Linux_fs/kernel/net/ipv6/
H A Dsyncookies.c144 if (!READ_ONCE(sock_net(sk)->ipv4.sysctl_tcp_syncookies) || in cookie_v6_check()
153 __NET_INC_STATS(sock_net(sk), LINUX_MIB_SYNCOOKIESFAILED); in cookie_v6_check()
157 __NET_INC_STATS(sock_net(sk), LINUX_MIB_SYNCOOKIESRECV); in cookie_v6_check()
161 tcp_parse_options(sock_net(sk), skb, &tcp_opt, 0, NULL); in cookie_v6_check()
164 tsoff = secure_tcpv6_ts_off(sock_net(sk), in cookie_v6_check()
170 if (!cookie_timestamp_decode(sock_net(sk), &tcp_opt)) in cookie_v6_check()
240 dst = ip6_dst_lookup_flow(sock_net(sk), sk, &fl6, final_p); in cookie_v6_check()
258 ireq->ecn_ok = cookie_ecn_ok(&tcp_opt, sock_net(sk), dst); in cookie_v6_check()
/OK3568_Linux_fs/kernel/net/vmw_vsock/
H A Ddiag.c59 net = sock_net(skb->sk); in vsock_diag_dump()
79 if (!net_eq(sock_net(sk), net)) in vsock_diag_dump()
113 if (!net_eq(sock_net(sk), net)) in vsock_diag_dump()
144 struct net *net = sock_net(skb->sk); in vsock_diag_handler_dump()
/OK3568_Linux_fs/kernel/net/tls/
H A Dtls_main.c284 TLS_DEC_STATS(sock_net(sk), LINUX_MIB_TLSCURRTXSW); in tls_sk_proto_cleanup()
287 TLS_DEC_STATS(sock_net(sk), LINUX_MIB_TLSCURRTXDEVICE); in tls_sk_proto_cleanup()
292 TLS_DEC_STATS(sock_net(sk), LINUX_MIB_TLSCURRRXSW); in tls_sk_proto_cleanup()
295 TLS_DEC_STATS(sock_net(sk), LINUX_MIB_TLSCURRRXDEVICE); in tls_sk_proto_cleanup()
548 TLS_INC_STATS(sock_net(sk), LINUX_MIB_TLSTXDEVICE); in do_tls_setsockopt_conf()
549 TLS_INC_STATS(sock_net(sk), LINUX_MIB_TLSCURRTXDEVICE); in do_tls_setsockopt_conf()
554 TLS_INC_STATS(sock_net(sk), LINUX_MIB_TLSTXSW); in do_tls_setsockopt_conf()
555 TLS_INC_STATS(sock_net(sk), LINUX_MIB_TLSCURRTXSW); in do_tls_setsockopt_conf()
562 TLS_INC_STATS(sock_net(sk), LINUX_MIB_TLSRXDEVICE); in do_tls_setsockopt_conf()
563 TLS_INC_STATS(sock_net(sk), LINUX_MIB_TLSCURRRXDEVICE); in do_tls_setsockopt_conf()
[all …]
/OK3568_Linux_fs/kernel/net/ieee802154/
H A Dsocket.c137 dev_load(sock_net(sk), ifr.ifr_name); in ieee802154_dev_ioctl()
138 dev = dev_get_by_name(sock_net(sk), ifr.ifr_name); in ieee802154_dev_ioctl()
178 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in raw_hash()
188 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, -1); in raw_unhash()
215 dev = ieee802154_get_dev(sock_net(sk), &addr); in raw_bind()
257 dev = dev_getfirstbyhwtype(sock_net(sk), ARPHRD_IEEE802154); in raw_sendmsg()
259 dev = dev_get_by_index(sock_net(sk), sk->sk_bound_dev_if); in raw_sendmsg()
462 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in dgram_hash()
472 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, -1); in dgram_unhash()
512 dev = ieee802154_get_dev(sock_net(sk), &haddr); in dgram_bind()
[all …]
/OK3568_Linux_fs/kernel/drivers/infiniband/core/
H A Dnldev.c884 device = ib_device_get_by_index(sock_net(skb->sk), index); in nldev_get_doit()
905 return rdma_nl_unicast(sock_net(skb->sk), msg, NETLINK_CB(skb).portid); in nldev_get_doit()
928 device = ib_device_get_by_index(sock_net(skb->sk), index); in nldev_set_doit()
1022 device = ib_device_get_by_index(sock_net(skb->sk), index); in nldev_port_get_doit()
1042 err = fill_port_info(msg, device, port, sock_net(skb->sk)); in nldev_port_get_doit()
1049 return rdma_nl_unicast(sock_net(skb->sk), msg, NETLINK_CB(skb).portid); in nldev_port_get_doit()
1076 device = ib_device_get_by_index(sock_net(skb->sk), ifindex); in nldev_port_get_dumpit()
1102 if (fill_port_info(skb, device, p, sock_net(skb->sk))) { in nldev_port_get_dumpit()
1131 device = ib_device_get_by_index(sock_net(skb->sk), index); in nldev_res_get_doit()
1151 return rdma_nl_unicast(sock_net(skb->sk), msg, NETLINK_CB(skb).portid); in nldev_res_get_doit()
[all …]
/OK3568_Linux_fs/kernel/net/phonet/
H A Dsocket.c79 if (!net_eq(sock_net(sknode), net)) in pn_find_sock_by_sa()
116 if (!net_eq(sock_net(sknode), net)) in pn_deliver_sock_broadcast()
175 if (saddr && phonet_address_lookup(sock_net(sk), saddr)) in pn_socket_bind()
372 dev = dev_get_by_index(sock_net(sk), in pn_socket_ioctl()
375 dev = phonet_device_get(sock_net(sk)); in pn_socket_ioctl()
474 struct net *net = sock_net(sk); in pn_sock_get_port()
529 if (!net_eq(net, sock_net(sknode))) in pn_sock_get_idx()
546 while (sk && !net_eq(net, sock_net(sk))); in pn_sock_get_next()
636 if (!net_eq(sock_net(sk), &init_net)) in pn_sock_bind_res()
/OK3568_Linux_fs/kernel/net/netlink/
H A Ddiag.c94 struct net *net = sock_net(skb->sk); in __netlink_diag_dump()
133 if (!net_eq(sock_net(sk), net)) in __netlink_diag_dump()
159 if (!net_eq(sock_net(sk), net)) in __netlink_diag_dump()
227 struct net *net = sock_net(skb->sk); in netlink_diag_handler_dump()
H A Daf_netlink.c282 if (!net_eq(dev_net(dev), sock_net(sk))) in __netlink_deliver_tap_skb()
337 netlink_deliver_tap(sock_net(dst), skb); in netlink_deliver_tap_kernel()
491 !net_eq(sock_net(&nlk->sk), read_pnet(&x->pnet)); in netlink_compare()
516 netlink_compare_arg_init(&arg, sock_net(sk), nlk_sk(sk)->portid); in __netlink_insert()
765 nlk->netlink_unbind(sock_net(sk), i + 1); in netlink_release()
778 .net = sock_net(sk), in netlink_release()
807 sock_prot_inuse_add(sock_net(sk), &netlink_proto, -1); in netlink_release()
816 struct net *net = sock_net(sk); in netlink_autobind()
913 return netlink_ns_capable(skb, sock_net(skb->sk)->user_ns, cap); in netlink_net_capable()
920 ns_capable(sock_net(sock->sk)->user_ns, CAP_NET_ADMIN); in netlink_allowed()
[all …]
/OK3568_Linux_fs/kernel/net/unix/
H A Ddiag.c202 struct net *net = sock_net(skb->sk); in unix_diag_dump()
217 if (!net_eq(sock_net(sk), net)) in unix_diag_dump()
268 struct net *net = sock_net(in_skb->sk); in unix_diag_get_exact()
277 if (!net_eq(sock_net(sk), net)) in unix_diag_get_exact()
316 struct net *net = sock_net(skb->sk); in unix_diag_handler_dump()
/OK3568_Linux_fs/kernel/net/xfrm/
H A Despintcp.c18 XFRM_INC_STATS(sock_net(sk), LINUX_MIB_XFRMINERROR); in handle_nonesp()
40 skb->dev = dev_get_by_index_rcu(sock_net(sk), skb->skb_iif); in handle_esp()
67 XFRM_INC_STATS(sock_net(strp->sk), LINUX_MIB_XFRMINHDRERROR); in espintcp_rcv()
80 XFRM_INC_STATS(sock_net(strp->sk), LINUX_MIB_XFRMINHDRERROR); in espintcp_rcv()
88 XFRM_INC_STATS(sock_net(strp->sk), LINUX_MIB_XFRMINHDRERROR); in espintcp_rcv()
95 XFRM_INC_STATS(sock_net(strp->sk), LINUX_MIB_XFRMINERROR); in espintcp_rcv()
101 XFRM_INC_STATS(sock_net(strp->sk), LINUX_MIB_XFRMINERROR); in espintcp_rcv()
/OK3568_Linux_fs/kernel/include/net/
H A Dudp.h426 ipv4 ? (IS_UDPLITE(sk) ? sock_net(sk)->mib.udplite_statistics : \
427 sock_net(sk)->mib.udp_statistics) : \
428 (IS_UDPLITE(sk) ? sock_net(sk)->mib.udplite_stats_in6 : \
429 sock_net(sk)->mib.udp_stats_in6); \
434 IS_UDPLITE(sk) ? sock_net(sk)->mib.udplite_statistics : \
435 sock_net(sk)->mib.udp_statistics; \
/OK3568_Linux_fs/kernel/net/tipc/
H A Dsocket.c279 u32 onode = tipc_own_addr(sock_net(sk)); in tipc_sk_respond()
287 tipc_node_xmit_skb(sock_net(sk), skb, dnode, selector); in tipc_sk_respond()
326 u32 self = tipc_own_addr(sock_net(sk)); in tsk_peer_msg()
541 struct net *net = sock_net(sk); in __tipc_shutdown()
739 addr->addr.id.node = tipc_own_addr(sock_net(sk)); in tipc_getname()
826 struct net *net = sock_net(sk); in tipc_sendmcast()
942 struct net *net = sock_net(sk); in tipc_send_group_unicast()
990 struct net *net = sock_net(sk); in tipc_send_group_anycast()
1069 struct net *net = sock_net(sk); in tipc_send_group_bcast()
1146 struct net *net = sock_net(sk); in tipc_send_group_mcast()
[all …]
/OK3568_Linux_fs/kernel/net/mptcp/
H A Dtoken.c257 !net_eq(sock_net(sk), net)) in mptcp_token_get_sock()
264 !net_eq(sock_net(sk), net)) { in mptcp_token_get_sock()
313 if (!net_eq(sock_net(sk), net)) in mptcp_token_iter_next()
322 if (!net_eq(sock_net(sk), net)) { in mptcp_token_iter_next()

12345678910