Lines Matching +full:smc +full:- +full:id

1 // SPDX-License-Identifier: GPL-2.0-only
3 * Shared Memory Communications over RDMA (SMC-R) and RoCE
7 * offers an alternative communication option for TCP-protocol sockets
8 * applicable with RoCE-cards only
11 * - support for alternate links postponed
19 #define KMSG_COMPONENT "smc"
33 #include <net/smc.h>
40 #include "smc.h"
67 struct smc_sock *smc = smc_sk(sk); in smc_set_keepalive() local
69 smc->clcsock->sk->sk_prot->keepalive(smc->clcsock->sk, val); in smc_set_keepalive()
82 struct smc_hashinfo *h = sk->sk_prot->h.smc_hash; in smc_hash_sk()
85 head = &h->ht; in smc_hash_sk()
87 write_lock_bh(&h->lock); in smc_hash_sk()
89 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, 1); in smc_hash_sk()
90 write_unlock_bh(&h->lock); in smc_hash_sk()
98 struct smc_hashinfo *h = sk->sk_prot->h.smc_hash; in smc_unhash_sk()
100 write_lock_bh(&h->lock); in smc_unhash_sk()
102 sock_prot_inuse_add(sock_net(sk), sk->sk_prot, -1); in smc_unhash_sk()
103 write_unlock_bh(&h->lock); in smc_unhash_sk()
108 .name = "SMC",
131 static void smc_restore_fallback_changes(struct smc_sock *smc) in smc_restore_fallback_changes() argument
133 if (smc->clcsock->file) { /* non-accepted sockets have no file yet */ in smc_restore_fallback_changes()
134 smc->clcsock->file->private_data = smc->sk.sk_socket; in smc_restore_fallback_changes()
135 smc->clcsock->file = NULL; in smc_restore_fallback_changes()
139 static int __smc_release(struct smc_sock *smc) in __smc_release() argument
141 struct sock *sk = &smc->sk; in __smc_release()
144 if (!smc->use_fallback) { in __smc_release()
145 rc = smc_close_active(smc); in __smc_release()
147 sk->sk_shutdown |= SHUTDOWN_MASK; in __smc_release()
149 if (sk->sk_state != SMC_CLOSED) { in __smc_release()
150 if (sk->sk_state != SMC_LISTEN && in __smc_release()
151 sk->sk_state != SMC_INIT) in __smc_release()
153 if (sk->sk_state == SMC_LISTEN) { in __smc_release()
155 rc = kernel_sock_shutdown(smc->clcsock, in __smc_release()
158 sk->sk_state = SMC_CLOSED; in __smc_release()
159 sk->sk_state_change(sk); in __smc_release()
161 smc_restore_fallback_changes(smc); in __smc_release()
164 sk->sk_prot->unhash(sk); in __smc_release()
166 if (sk->sk_state == SMC_CLOSED) { in __smc_release()
167 if (smc->clcsock) { in __smc_release()
169 smc_clcsock_release(smc); in __smc_release()
172 if (!smc->use_fallback) in __smc_release()
173 smc_conn_free(&smc->conn); in __smc_release()
181 struct sock *sk = sock->sk; in smc_release()
182 struct smc_sock *smc; in smc_release() local
189 smc = smc_sk(sk); in smc_release()
191 old_state = sk->sk_state; in smc_release()
193 /* cleanup for a dangling non-blocking connect */ in smc_release()
194 if (smc->connect_nonblock && old_state == SMC_INIT) in smc_release()
195 tcp_abort(smc->clcsock->sk, ECONNABORTED); in smc_release()
197 if (cancel_work_sync(&smc->connect_work)) in smc_release()
198 sock_put(&smc->sk); /* sock_hold in smc_connect for passive closing */ in smc_release()
200 if (sk->sk_state == SMC_LISTEN) in smc_release()
208 if (old_state == SMC_INIT && sk->sk_state == SMC_ACTIVE && in smc_release()
209 !smc->use_fallback) in smc_release()
210 smc_close_active_abort(smc); in smc_release()
212 rc = __smc_release(smc); in smc_release()
216 sock->sk = NULL; in smc_release()
227 if (sk->sk_state != SMC_CLOSED) in smc_destruct()
238 struct smc_sock *smc; in smc_sock_alloc() local
248 sk->sk_state = SMC_INIT; in smc_sock_alloc()
249 sk->sk_destruct = smc_destruct; in smc_sock_alloc()
250 sk->sk_protocol = protocol; in smc_sock_alloc()
251 smc = smc_sk(sk); in smc_sock_alloc()
252 INIT_WORK(&smc->tcp_listen_work, smc_tcp_listen_work); in smc_sock_alloc()
253 INIT_WORK(&smc->connect_work, smc_connect_work); in smc_sock_alloc()
254 INIT_DELAYED_WORK(&smc->conn.tx_work, smc_tx_work); in smc_sock_alloc()
255 INIT_LIST_HEAD(&smc->accept_q); in smc_sock_alloc()
256 spin_lock_init(&smc->accept_q_lock); in smc_sock_alloc()
257 spin_lock_init(&smc->conn.send_lock); in smc_sock_alloc()
258 sk->sk_prot->hash(sk); in smc_sock_alloc()
260 mutex_init(&smc->clcsock_release_lock); in smc_sock_alloc()
269 struct sock *sk = sock->sk; in smc_bind()
270 struct smc_sock *smc; in smc_bind() local
273 smc = smc_sk(sk); in smc_bind()
276 rc = -EINVAL; in smc_bind()
280 rc = -EAFNOSUPPORT; in smc_bind()
281 if (addr->sin_family != AF_INET && in smc_bind()
282 addr->sin_family != AF_INET6 && in smc_bind()
283 addr->sin_family != AF_UNSPEC) in smc_bind()
286 if (addr->sin_family == AF_UNSPEC && in smc_bind()
287 addr->sin_addr.s_addr != htonl(INADDR_ANY)) in smc_bind()
293 rc = -EINVAL; in smc_bind()
294 if (sk->sk_state != SMC_INIT || smc->connect_nonblock) in smc_bind()
297 smc->clcsock->sk->sk_reuse = sk->sk_reuse; in smc_bind()
298 rc = kernel_bind(smc->clcsock, uaddr, addr_len); in smc_bind()
310 nsk->sk_type = osk->sk_type; in smc_copy_sock_settings()
311 nsk->sk_sndbuf = osk->sk_sndbuf; in smc_copy_sock_settings()
312 nsk->sk_rcvbuf = osk->sk_rcvbuf; in smc_copy_sock_settings()
313 nsk->sk_sndtimeo = osk->sk_sndtimeo; in smc_copy_sock_settings()
314 nsk->sk_rcvtimeo = osk->sk_rcvtimeo; in smc_copy_sock_settings()
315 nsk->sk_mark = osk->sk_mark; in smc_copy_sock_settings()
316 nsk->sk_priority = osk->sk_priority; in smc_copy_sock_settings()
317 nsk->sk_rcvlowat = osk->sk_rcvlowat; in smc_copy_sock_settings()
318 nsk->sk_bound_dev_if = osk->sk_bound_dev_if; in smc_copy_sock_settings()
319 nsk->sk_err = osk->sk_err; in smc_copy_sock_settings()
321 nsk->sk_flags &= ~mask; in smc_copy_sock_settings()
322 nsk->sk_flags |= osk->sk_flags & mask; in smc_copy_sock_settings()
340 /* copy only relevant settings and flags of SOL_SOCKET level from smc to
341 * clc socket (since smc is not called for these options from net/core)
343 static void smc_copy_sock_settings_to_clc(struct smc_sock *smc) in smc_copy_sock_settings_to_clc() argument
345 smc_copy_sock_settings(smc->clcsock->sk, &smc->sk, SK_FLAGS_SMC_TO_CLC); in smc_copy_sock_settings_to_clc()
352 /* copy only settings and flags relevant for smc from clc to smc socket */
353 static void smc_copy_sock_settings_to_smc(struct smc_sock *smc) in smc_copy_sock_settings_to_smc() argument
355 smc_copy_sock_settings(&smc->sk, smc->clcsock->sk, SK_FLAGS_CLC_TO_SMC); in smc_copy_sock_settings_to_smc()
362 struct smc_link_group *lgr = link->lgr; in smcr_lgr_reg_rmbs()
371 mutex_lock(&lgr->llc_conf_mutex); in smcr_lgr_reg_rmbs()
373 if (!smc_link_active(&lgr->lnk[i])) in smcr_lgr_reg_rmbs()
375 rc = smcr_link_reg_rmb(&lgr->lnk[i], rmb_desc); in smcr_lgr_reg_rmbs()
383 rc = -EFAULT; in smcr_lgr_reg_rmbs()
386 rmb_desc->is_conf_rkey = true; in smcr_lgr_reg_rmbs()
388 mutex_unlock(&lgr->llc_conf_mutex); in smcr_lgr_reg_rmbs()
389 smc_llc_flow_stop(lgr, &lgr->llc_flow_lcl); in smcr_lgr_reg_rmbs()
393 static int smcr_clnt_conf_first_link(struct smc_sock *smc) in smcr_clnt_conf_first_link() argument
395 struct smc_link *link = smc->conn.lnk; in smcr_clnt_conf_first_link()
400 qentry = smc_llc_wait(link->lgr, NULL, SMC_LLC_WAIT_TIME, in smcr_clnt_conf_first_link()
405 rc = smc_clc_wait_msg(smc, &dclc, sizeof(dclc), in smcr_clnt_conf_first_link()
407 return rc == -EAGAIN ? SMC_CLC_DECL_TIMEOUT_CL : rc; in smcr_clnt_conf_first_link()
411 smc_llc_flow_qentry_del(&link->lgr->llc_flow_lcl); in smcr_clnt_conf_first_link()
421 if (smcr_link_reg_rmb(link, smc->conn.rmb_desc)) in smcr_clnt_conf_first_link()
425 smc->conn.rmb_desc->is_conf_rkey = true; in smcr_clnt_conf_first_link()
433 smcr_lgr_set_type(link->lgr, SMC_LGR_SINGLE); in smcr_clnt_conf_first_link()
436 qentry = smc_llc_wait(link->lgr, NULL, SMC_LLC_WAIT_TIME, in smcr_clnt_conf_first_link()
441 rc = smc_clc_wait_msg(smc, &dclc, sizeof(dclc), in smcr_clnt_conf_first_link()
443 if (rc == -EAGAIN) in smcr_clnt_conf_first_link()
447 smc_llc_flow_qentry_clr(&link->lgr->llc_flow_lcl); in smcr_clnt_conf_first_link()
452 static void smcr_conn_save_peer_info(struct smc_sock *smc, in smcr_conn_save_peer_info() argument
455 int bufsize = smc_uncompress_bufsize(clc->r0.rmbe_size); in smcr_conn_save_peer_info()
457 smc->conn.peer_rmbe_idx = clc->r0.rmbe_idx; in smcr_conn_save_peer_info()
458 smc->conn.local_tx_ctrl.token = ntohl(clc->r0.rmbe_alert_token); in smcr_conn_save_peer_info()
459 smc->conn.peer_rmbe_size = bufsize; in smcr_conn_save_peer_info()
460 atomic_set(&smc->conn.peer_rmbe_space, smc->conn.peer_rmbe_size); in smcr_conn_save_peer_info()
461 smc->conn.tx_off = bufsize * (smc->conn.peer_rmbe_idx - 1); in smcr_conn_save_peer_info()
474 static void smcd_conn_save_peer_info(struct smc_sock *smc, in smcd_conn_save_peer_info() argument
477 int bufsize = smc_uncompress_bufsize(clc->d0.dmbe_size); in smcd_conn_save_peer_info()
479 smc->conn.peer_rmbe_idx = clc->d0.dmbe_idx; in smcd_conn_save_peer_info()
480 smc->conn.peer_token = clc->d0.token; in smcd_conn_save_peer_info()
482 smc->conn.peer_rmbe_size = bufsize - sizeof(struct smcd_cdc_msg); in smcd_conn_save_peer_info()
483 atomic_set(&smc->conn.peer_rmbe_space, smc->conn.peer_rmbe_size); in smcd_conn_save_peer_info()
484 smc->conn.tx_off = bufsize * smc->conn.peer_rmbe_idx; in smcd_conn_save_peer_info()
485 if (clc->hdr.version > SMC_V1 && in smcd_conn_save_peer_info()
486 (clc->hdr.typev2 & SMC_FIRST_CONTACT_MASK)) { in smcd_conn_save_peer_info()
493 memcpy(smc->conn.lgr->negotiated_eid, clc_v2->eid, in smcd_conn_save_peer_info()
495 smc->conn.lgr->peer_os = fce->os_type; in smcd_conn_save_peer_info()
496 smc->conn.lgr->peer_smc_release = fce->release; in smcd_conn_save_peer_info()
497 if (smc_isascii(fce->hostname)) in smcd_conn_save_peer_info()
498 memcpy(smc->conn.lgr->peer_hostname, fce->hostname, in smcd_conn_save_peer_info()
503 static void smc_conn_save_peer_info(struct smc_sock *smc, in smc_conn_save_peer_info() argument
506 if (smc->conn.lgr->is_smcd) in smc_conn_save_peer_info()
507 smcd_conn_save_peer_info(smc, clc); in smc_conn_save_peer_info()
509 smcr_conn_save_peer_info(smc, clc); in smc_conn_save_peer_info()
515 link->peer_qpn = ntoh24(clc->r0.qpn); in smc_link_save_peer_info()
516 memcpy(link->peer_gid, clc->r0.lcl.gid, SMC_GID_SIZE); in smc_link_save_peer_info()
517 memcpy(link->peer_mac, clc->r0.lcl.mac, sizeof(link->peer_mac)); in smc_link_save_peer_info()
518 link->peer_psn = ntoh24(clc->r0.psn); in smc_link_save_peer_info()
519 link->peer_mtu = clc->r0.qp_mtu; in smc_link_save_peer_info()
522 static void smc_switch_to_fallback(struct smc_sock *smc) in smc_switch_to_fallback() argument
524 wait_queue_head_t *smc_wait = sk_sleep(&smc->sk); in smc_switch_to_fallback()
525 wait_queue_head_t *clc_wait = sk_sleep(smc->clcsock->sk); in smc_switch_to_fallback()
528 smc->use_fallback = true; in smc_switch_to_fallback()
529 if (smc->sk.sk_socket && smc->sk.sk_socket->file) { in smc_switch_to_fallback()
530 smc->clcsock->file = smc->sk.sk_socket->file; in smc_switch_to_fallback()
531 smc->clcsock->file->private_data = smc->clcsock; in smc_switch_to_fallback()
532 smc->clcsock->wq.fasync_list = in smc_switch_to_fallback()
533 smc->sk.sk_socket->wq.fasync_list; in smc_switch_to_fallback()
536 * smc socket->wq, which should be removed in smc_switch_to_fallback()
537 * to clcsocket->wq during the fallback. in smc_switch_to_fallback()
539 spin_lock_irqsave(&smc_wait->lock, flags); in smc_switch_to_fallback()
540 spin_lock_nested(&clc_wait->lock, SINGLE_DEPTH_NESTING); in smc_switch_to_fallback()
541 list_splice_init(&smc_wait->head, &clc_wait->head); in smc_switch_to_fallback()
542 spin_unlock(&clc_wait->lock); in smc_switch_to_fallback()
543 spin_unlock_irqrestore(&smc_wait->lock, flags); in smc_switch_to_fallback()
548 static int smc_connect_fallback(struct smc_sock *smc, int reason_code) in smc_connect_fallback() argument
550 smc_switch_to_fallback(smc); in smc_connect_fallback()
551 smc->fallback_rsn = reason_code; in smc_connect_fallback()
552 smc_copy_sock_settings_to_clc(smc); in smc_connect_fallback()
553 smc->connect_nonblock = 0; in smc_connect_fallback()
554 if (smc->sk.sk_state == SMC_INIT) in smc_connect_fallback()
555 smc->sk.sk_state = SMC_ACTIVE; in smc_connect_fallback()
560 static int smc_connect_decline_fallback(struct smc_sock *smc, int reason_code, in smc_connect_decline_fallback() argument
566 if (smc->sk.sk_state == SMC_INIT) in smc_connect_decline_fallback()
567 sock_put(&smc->sk); /* passive closing */ in smc_connect_decline_fallback()
571 rc = smc_clc_send_decline(smc, reason_code, version); in smc_connect_decline_fallback()
573 if (smc->sk.sk_state == SMC_INIT) in smc_connect_decline_fallback()
574 sock_put(&smc->sk); /* passive closing */ in smc_connect_decline_fallback()
578 return smc_connect_fallback(smc, reason_code); in smc_connect_decline_fallback()
582 static void smc_connect_abort(struct smc_sock *smc, int local_first) in smc_connect_abort() argument
585 smc_lgr_cleanup_early(&smc->conn); in smc_connect_abort()
587 smc_conn_free(&smc->conn); in smc_connect_abort()
592 static int smc_find_rdma_device(struct smc_sock *smc, struct smc_init_info *ini) in smc_find_rdma_device() argument
598 smc_pnet_find_roce_resource(smc->clcsock->sk, ini); in smc_find_rdma_device()
599 if (!ini->ib_dev) in smc_find_rdma_device()
606 static int smc_find_ism_device(struct smc_sock *smc, struct smc_init_info *ini) in smc_find_ism_device() argument
609 smc_pnet_find_ism_resource(smc->clcsock->sk, ini); in smc_find_ism_device()
610 if (!ini->ism_dev[0]) in smc_find_ism_device()
613 ini->ism_chid[0] = smc_ism_get_chid(ini->ism_dev[0]); in smc_find_ism_device()
621 int i = (!ini->ism_dev[0]) ? 1 : 0; in smc_find_ism_v2_is_unique_chid()
624 if (ini->ism_chid[i] == chid) in smc_find_ism_v2_is_unique_chid()
632 static int smc_find_ism_v2_device_clnt(struct smc_sock *smc, in smc_find_ism_v2_device_clnt() argument
640 if (smcd_indicated(ini->smc_type_v1)) in smc_find_ism_v2_device_clnt()
644 if (smcd->going_away || smcd == ini->ism_dev[0]) in smc_find_ism_v2_device_clnt()
649 if (!smc_pnet_is_pnetid_set(smcd->pnetid) || in smc_find_ism_v2_device_clnt()
650 smc_pnet_is_ndev_pnetid(sock_net(&smc->sk), smcd->pnetid)) { in smc_find_ism_v2_device_clnt()
651 ini->ism_dev[i] = smcd; in smc_find_ism_v2_device_clnt()
652 ini->ism_chid[i] = chid; in smc_find_ism_v2_device_clnt()
653 ini->is_smcd = true; in smc_find_ism_v2_device_clnt()
661 ini->ism_offered_cnt = i - 1; in smc_find_ism_v2_device_clnt()
662 if (!ini->ism_dev[0] && !ini->ism_dev[1]) in smc_find_ism_v2_device_clnt()
663 ini->smcd_version = 0; in smc_find_ism_v2_device_clnt()
668 /* Check for VLAN ID and register it on ISM device just for CLC handshake */
669 static int smc_connect_ism_vlan_setup(struct smc_sock *smc, in smc_connect_ism_vlan_setup() argument
672 if (ini->vlan_id && smc_ism_get_vlan(ini->ism_dev[0], ini->vlan_id)) in smc_connect_ism_vlan_setup()
677 static int smc_find_proposal_devices(struct smc_sock *smc, in smc_find_proposal_devices() argument
683 if (ini->smcd_version & SMC_V1) { in smc_find_proposal_devices()
684 if (smc_find_ism_device(smc, ini) || in smc_find_proposal_devices()
685 smc_connect_ism_vlan_setup(smc, ini)) { in smc_find_proposal_devices()
686 if (ini->smc_type_v1 == SMC_TYPE_B) in smc_find_proposal_devices()
687 ini->smc_type_v1 = SMC_TYPE_R; in smc_find_proposal_devices()
689 ini->smc_type_v1 = SMC_TYPE_N; in smc_find_proposal_devices()
691 if (smc_find_rdma_device(smc, ini)) { in smc_find_proposal_devices()
692 if (ini->smc_type_v1 == SMC_TYPE_B) in smc_find_proposal_devices()
693 ini->smc_type_v1 = SMC_TYPE_D; in smc_find_proposal_devices()
695 ini->smc_type_v1 = SMC_TYPE_N; in smc_find_proposal_devices()
698 if (smc_ism_v2_capable && smc_find_ism_v2_device_clnt(smc, ini)) in smc_find_proposal_devices()
699 ini->smc_type_v2 = SMC_TYPE_N; in smc_find_proposal_devices()
702 if (!smcr_indicated(ini->smc_type_v1) && in smc_find_proposal_devices()
703 ini->smc_type_v1 == SMC_TYPE_N && ini->smc_type_v2 == SMC_TYPE_N) in smc_find_proposal_devices()
709 /* cleanup temporary VLAN ID registration used for CLC handshake. If ISM is
710 * used, the VLAN ID will be registered again during the connection setup.
712 static int smc_connect_ism_vlan_cleanup(struct smc_sock *smc, in smc_connect_ism_vlan_cleanup() argument
715 if (!smcd_indicated(ini->smc_type_v1)) in smc_connect_ism_vlan_cleanup()
717 if (ini->vlan_id && smc_ism_put_vlan(ini->ism_dev[0], ini->vlan_id)) in smc_connect_ism_vlan_cleanup()
728 static int smc_connect_clc(struct smc_sock *smc, in smc_connect_clc() argument
735 rc = smc_clc_send_proposal(smc, ini); in smc_connect_clc()
738 /* receive SMC Accept CLC message */ in smc_connect_clc()
739 return smc_clc_wait_msg(smc, aclc2, SMC_CLC_MAX_ACCEPT_LEN, in smc_connect_clc()
744 static int smc_connect_rdma(struct smc_sock *smc, in smc_connect_rdma() argument
751 ini->is_smcd = false; in smc_connect_rdma()
752 ini->ib_lcl = &aclc->r0.lcl; in smc_connect_rdma()
753 ini->ib_clcqpn = ntoh24(aclc->r0.qpn); in smc_connect_rdma()
754 ini->first_contact_peer = aclc->hdr.typev2 & SMC_FIRST_CONTACT_MASK; in smc_connect_rdma()
757 reason_code = smc_conn_create(smc, ini); in smc_connect_rdma()
763 smc_conn_save_peer_info(smc, aclc); in smc_connect_rdma()
765 if (ini->first_contact_local) { in smc_connect_rdma()
766 link = smc->conn.lnk; in smc_connect_rdma()
771 struct smc_link *l = &smc->conn.lgr->lnk[i]; in smc_connect_rdma()
773 if (l->peer_qpn == ntoh24(aclc->r0.qpn) && in smc_connect_rdma()
774 !memcmp(l->peer_gid, &aclc->r0.lcl.gid, in smc_connect_rdma()
776 !memcmp(l->peer_mac, &aclc->r0.lcl.mac, in smc_connect_rdma()
777 sizeof(l->peer_mac))) { in smc_connect_rdma()
786 smc->conn.lnk = link; in smc_connect_rdma()
790 if (smc_buf_create(smc, false)) { in smc_connect_rdma()
795 if (ini->first_contact_local) in smc_connect_rdma()
798 if (smc_rmb_rtoken_handling(&smc->conn, link, aclc)) { in smc_connect_rdma()
803 smc_close_init(smc); in smc_connect_rdma()
804 smc_rx_init(smc); in smc_connect_rdma()
806 if (ini->first_contact_local) { in smc_connect_rdma()
812 if (smcr_lgr_reg_rmbs(link, smc->conn.rmb_desc)) { in smc_connect_rdma()
817 smc_rmb_sync_sg_for_device(&smc->conn); in smc_connect_rdma()
819 reason_code = smc_clc_send_confirm(smc, ini->first_contact_local, in smc_connect_rdma()
824 smc_tx_init(smc); in smc_connect_rdma()
826 if (ini->first_contact_local) { in smc_connect_rdma()
828 smc_llc_flow_initiate(link->lgr, SMC_LLC_FLOW_ADD_LINK); in smc_connect_rdma()
829 reason_code = smcr_clnt_conf_first_link(smc); in smc_connect_rdma()
830 smc_llc_flow_stop(link->lgr, &link->lgr->llc_flow_lcl); in smc_connect_rdma()
836 smc_copy_sock_settings_to_clc(smc); in smc_connect_rdma()
837 smc->connect_nonblock = 0; in smc_connect_rdma()
838 if (smc->sk.sk_state == SMC_INIT) in smc_connect_rdma()
839 smc->sk.sk_state = SMC_ACTIVE; in smc_connect_rdma()
843 smc_connect_abort(smc, ini->first_contact_local); in smc_connect_rdma()
845 smc->connect_nonblock = 0; in smc_connect_rdma()
859 for (i = 0; i < ini->ism_offered_cnt + 1; i++) { in smc_v2_determine_accepted_chid()
860 if (ini->ism_chid[i] == ntohs(aclc->chid)) { in smc_v2_determine_accepted_chid()
861 ini->ism_selected = i; in smc_v2_determine_accepted_chid()
866 return -EPROTO; in smc_v2_determine_accepted_chid()
870 static int smc_connect_ism(struct smc_sock *smc, in smc_connect_ism() argument
876 ini->is_smcd = true; in smc_connect_ism()
877 ini->first_contact_peer = aclc->hdr.typev2 & SMC_FIRST_CONTACT_MASK; in smc_connect_ism()
879 if (aclc->hdr.version == SMC_V2) { in smc_connect_ism()
887 ini->ism_peer_gid[ini->ism_selected] = aclc->d0.gid; in smc_connect_ism()
889 /* there is only one lgr role for SMC-D; use server lock */ in smc_connect_ism()
891 rc = smc_conn_create(smc, ini); in smc_connect_ism()
898 rc = smc_buf_create(smc, true); in smc_connect_ism()
900 rc = (rc == -ENOSPC) ? SMC_CLC_DECL_MAX_DMB : SMC_CLC_DECL_MEM; in smc_connect_ism()
904 smc_conn_save_peer_info(smc, aclc); in smc_connect_ism()
905 smc_close_init(smc); in smc_connect_ism()
906 smc_rx_init(smc); in smc_connect_ism()
907 smc_tx_init(smc); in smc_connect_ism()
909 rc = smc_clc_send_confirm(smc, ini->first_contact_local, in smc_connect_ism()
910 aclc->hdr.version); in smc_connect_ism()
915 smc_copy_sock_settings_to_clc(smc); in smc_connect_ism()
916 smc->connect_nonblock = 0; in smc_connect_ism()
917 if (smc->sk.sk_state == SMC_INIT) in smc_connect_ism()
918 smc->sk.sk_state = SMC_ACTIVE; in smc_connect_ism()
922 smc_connect_abort(smc, ini->first_contact_local); in smc_connect_ism()
924 smc->connect_nonblock = 0; in smc_connect_ism()
933 if ((aclc->hdr.typev1 == SMC_TYPE_R && in smc_connect_check_aclc()
934 !smcr_indicated(ini->smc_type_v1)) || in smc_connect_check_aclc()
935 (aclc->hdr.typev1 == SMC_TYPE_D && in smc_connect_check_aclc()
936 ((!smcd_indicated(ini->smc_type_v1) && in smc_connect_check_aclc()
937 !smcd_indicated(ini->smc_type_v2)) || in smc_connect_check_aclc()
938 (aclc->hdr.version == SMC_V1 && in smc_connect_check_aclc()
939 !smcd_indicated(ini->smc_type_v1)) || in smc_connect_check_aclc()
940 (aclc->hdr.version == SMC_V2 && in smc_connect_check_aclc()
941 !smcd_indicated(ini->smc_type_v2))))) in smc_connect_check_aclc()
948 static int __smc_connect(struct smc_sock *smc) in __smc_connect() argument
957 if (smc->use_fallback) in __smc_connect()
958 return smc_connect_fallback(smc, smc->fallback_rsn); in __smc_connect()
960 /* if peer has not signalled SMC-capability, fall back */ in __smc_connect()
961 if (!tcp_sk(smc->clcsock->sk)->syn_smc) in __smc_connect()
962 return smc_connect_fallback(smc, SMC_CLC_DECL_PEERNOSMC); in __smc_connect()
964 /* IPSec connections opt out of SMC optimizations */ in __smc_connect()
965 if (using_ipsec(smc)) in __smc_connect()
966 return smc_connect_decline_fallback(smc, SMC_CLC_DECL_IPSEC, in __smc_connect()
971 return smc_connect_decline_fallback(smc, SMC_CLC_DECL_MEM, in __smc_connect()
974 ini->smcd_version = SMC_V1; in __smc_connect()
975 ini->smcd_version |= smc_ism_v2_capable ? SMC_V2 : 0; in __smc_connect()
976 ini->smc_type_v1 = SMC_TYPE_B; in __smc_connect()
977 ini->smc_type_v2 = smc_ism_v2_capable ? SMC_TYPE_D : SMC_TYPE_N; in __smc_connect()
979 /* get vlan id from IP device */ in __smc_connect()
980 if (smc_vlan_by_tcpsk(smc->clcsock, ini)) { in __smc_connect()
981 ini->smcd_version &= ~SMC_V1; in __smc_connect()
982 ini->smc_type_v1 = SMC_TYPE_N; in __smc_connect()
983 if (!ini->smcd_version) { in __smc_connect()
989 rc = smc_find_proposal_devices(smc, ini); in __smc_connect()
1002 rc = smc_connect_clc(smc, aclc2, ini); in __smc_connect()
1006 /* check if smc modes and versions of CLC proposal and accept match */ in __smc_connect()
1008 version = aclc->hdr.version == SMC_V1 ? SMC_V1 : SMC_V2; in __smc_connect()
1009 ini->smcd_version = version; in __smc_connect()
1014 if (aclc->hdr.typev1 == SMC_TYPE_R) in __smc_connect()
1015 rc = smc_connect_rdma(smc, aclc, ini); in __smc_connect()
1016 else if (aclc->hdr.typev1 == SMC_TYPE_D) in __smc_connect()
1017 rc = smc_connect_ism(smc, aclc, ini); in __smc_connect()
1021 smc_connect_ism_vlan_cleanup(smc, ini); in __smc_connect()
1027 smc_connect_ism_vlan_cleanup(smc, ini); in __smc_connect()
1031 return smc_connect_decline_fallback(smc, rc, version); in __smc_connect()
1036 struct smc_sock *smc = container_of(work, struct smc_sock, in smc_connect_work() local
1038 long timeo = smc->sk.sk_sndtimeo; in smc_connect_work()
1043 lock_sock(smc->clcsock->sk); in smc_connect_work()
1044 if (smc->clcsock->sk->sk_err) { in smc_connect_work()
1045 smc->sk.sk_err = smc->clcsock->sk->sk_err; in smc_connect_work()
1046 } else if ((1 << smc->clcsock->sk->sk_state) & in smc_connect_work()
1048 rc = sk_stream_wait_connect(smc->clcsock->sk, &timeo); in smc_connect_work()
1049 if ((rc == -EPIPE) && in smc_connect_work()
1050 ((1 << smc->clcsock->sk->sk_state) & in smc_connect_work()
1054 release_sock(smc->clcsock->sk); in smc_connect_work()
1055 lock_sock(&smc->sk); in smc_connect_work()
1056 if (rc != 0 || smc->sk.sk_err) { in smc_connect_work()
1057 smc->sk.sk_state = SMC_CLOSED; in smc_connect_work()
1058 if (rc == -EPIPE || rc == -EAGAIN) in smc_connect_work()
1059 smc->sk.sk_err = EPIPE; in smc_connect_work()
1060 else if (rc == -ECONNREFUSED) in smc_connect_work()
1061 smc->sk.sk_err = ECONNREFUSED; in smc_connect_work()
1063 smc->sk.sk_err = -sock_intr_errno(timeo); in smc_connect_work()
1064 sock_put(&smc->sk); /* passive closing */ in smc_connect_work()
1068 rc = __smc_connect(smc); in smc_connect_work()
1070 smc->sk.sk_err = -rc; in smc_connect_work()
1073 if (!sock_flag(&smc->sk, SOCK_DEAD)) { in smc_connect_work()
1074 if (smc->sk.sk_err) { in smc_connect_work()
1075 smc->sk.sk_state_change(&smc->sk); in smc_connect_work()
1077 smc->clcsock->sk->sk_write_space(smc->clcsock->sk); in smc_connect_work()
1078 smc->sk.sk_write_space(&smc->sk); in smc_connect_work()
1081 release_sock(&smc->sk); in smc_connect_work()
1087 struct sock *sk = sock->sk; in smc_connect()
1088 struct smc_sock *smc; in smc_connect() local
1089 int rc = -EINVAL; in smc_connect()
1091 smc = smc_sk(sk); in smc_connect()
1093 /* separate smc parameter checking to be safe */ in smc_connect()
1094 if (alen < sizeof(addr->sa_family)) in smc_connect()
1096 if (addr->sa_family != AF_INET && addr->sa_family != AF_INET6) in smc_connect()
1100 switch (sk->sk_state) { in smc_connect()
1104 rc = -EISCONN; in smc_connect()
1111 smc_copy_sock_settings_to_clc(smc); in smc_connect()
1112 tcp_sk(smc->clcsock->sk)->syn_smc = 1; in smc_connect()
1113 if (smc->connect_nonblock) { in smc_connect()
1114 rc = -EALREADY; in smc_connect()
1117 rc = kernel_connect(smc->clcsock, addr, alen, flags); in smc_connect()
1118 if (rc && rc != -EINPROGRESS) in smc_connect()
1121 if (smc->use_fallback) in smc_connect()
1123 sock_hold(&smc->sk); /* sock put in passive closing */ in smc_connect()
1125 if (queue_work(smc_hs_wq, &smc->connect_work)) in smc_connect()
1126 smc->connect_nonblock = 1; in smc_connect()
1127 rc = -EINPROGRESS; in smc_connect()
1129 rc = __smc_connect(smc); in smc_connect()
1145 struct sock *lsk = &lsmc->sk; in smc_clcsock_accept()
1147 int rc = -EINVAL; in smc_clcsock_accept()
1150 new_sk = smc_sock_alloc(sock_net(lsk), NULL, lsk->sk_protocol); in smc_clcsock_accept()
1152 rc = -ENOMEM; in smc_clcsock_accept()
1153 lsk->sk_err = ENOMEM; in smc_clcsock_accept()
1160 mutex_lock(&lsmc->clcsock_release_lock); in smc_clcsock_accept()
1161 if (lsmc->clcsock) in smc_clcsock_accept()
1162 rc = kernel_accept(lsmc->clcsock, &new_clcsock, SOCK_NONBLOCK); in smc_clcsock_accept()
1163 mutex_unlock(&lsmc->clcsock_release_lock); in smc_clcsock_accept()
1165 if (rc < 0 && rc != -EAGAIN) in smc_clcsock_accept()
1166 lsk->sk_err = -rc; in smc_clcsock_accept()
1167 if (rc < 0 || lsk->sk_state == SMC_CLOSED) { in smc_clcsock_accept()
1168 new_sk->sk_prot->unhash(new_sk); in smc_clcsock_accept()
1171 new_sk->sk_state = SMC_CLOSED; in smc_clcsock_accept()
1178 /* new clcsock has inherited the smc listen-specific sk_data_ready in smc_clcsock_accept()
1181 new_clcsock->sk->sk_data_ready = lsmc->clcsk_data_ready; in smc_clcsock_accept()
1182 (*new_smc)->clcsock = new_clcsock; in smc_clcsock_accept()
1195 spin_lock(&par->accept_q_lock); in smc_accept_enqueue()
1196 list_add_tail(&smc_sk(sk)->accept_q, &par->accept_q); in smc_accept_enqueue()
1197 spin_unlock(&par->accept_q_lock); in smc_accept_enqueue()
1204 struct smc_sock *par = smc_sk(sk)->listen_smc; in smc_accept_unlink()
1206 spin_lock(&par->accept_q_lock); in smc_accept_unlink()
1207 list_del_init(&smc_sk(sk)->accept_q); in smc_accept_unlink()
1208 spin_unlock(&par->accept_q_lock); in smc_accept_unlink()
1209 sk_acceptq_removed(&smc_sk(sk)->listen_smc->sk); in smc_accept_unlink()
1222 list_for_each_entry_safe(isk, n, &smc_sk(parent)->accept_q, accept_q) { in smc_accept_dequeue()
1226 if (new_sk->sk_state == SMC_CLOSED) { in smc_accept_dequeue()
1227 new_sk->sk_prot->unhash(new_sk); in smc_accept_dequeue()
1228 if (isk->clcsock) { in smc_accept_dequeue()
1229 sock_release(isk->clcsock); in smc_accept_dequeue()
1230 isk->clcsock = NULL; in smc_accept_dequeue()
1237 if (isk->use_fallback) { in smc_accept_dequeue()
1238 smc_sk(new_sk)->clcsock->file = new_sock->file; in smc_accept_dequeue()
1239 isk->clcsock->file->private_data = isk->clcsock; in smc_accept_dequeue()
1250 struct smc_sock *smc = smc_sk(sk); in smc_close_non_accepted() local
1254 if (!sk->sk_lingertime) in smc_close_non_accepted()
1256 sk->sk_lingertime = SMC_MAX_STREAM_WAIT_TIMEOUT; in smc_close_non_accepted()
1257 __smc_release(smc); in smc_close_non_accepted()
1263 static int smcr_serv_conf_first_link(struct smc_sock *smc) in smcr_serv_conf_first_link() argument
1265 struct smc_link *link = smc->conn.lnk; in smcr_serv_conf_first_link()
1269 if (smcr_link_reg_rmb(link, smc->conn.rmb_desc)) in smcr_serv_conf_first_link()
1278 qentry = smc_llc_wait(link->lgr, link, SMC_LLC_WAIT_TIME, in smcr_serv_conf_first_link()
1283 rc = smc_clc_wait_msg(smc, &dclc, sizeof(dclc), in smcr_serv_conf_first_link()
1285 return rc == -EAGAIN ? SMC_CLC_DECL_TIMEOUT_CL : rc; in smcr_serv_conf_first_link()
1289 smc_llc_flow_qentry_del(&link->lgr->llc_flow_lcl); in smcr_serv_conf_first_link()
1294 smc->conn.rmb_desc->is_conf_rkey = true; in smcr_serv_conf_first_link()
1297 smcr_lgr_set_type(link->lgr, SMC_LGR_SINGLE); in smcr_serv_conf_first_link()
1299 /* initial contact - try to establish second link */ in smcr_serv_conf_first_link()
1307 struct smc_sock *lsmc = new_smc->listen_smc; in smc_listen_out()
1308 struct sock *newsmcsk = &new_smc->sk; in smc_listen_out()
1310 if (lsmc->sk.sk_state == SMC_LISTEN) { in smc_listen_out()
1311 lock_sock_nested(&lsmc->sk, SINGLE_DEPTH_NESTING); in smc_listen_out()
1312 smc_accept_enqueue(&lsmc->sk, newsmcsk); in smc_listen_out()
1313 release_sock(&lsmc->sk); in smc_listen_out()
1319 lsmc->sk.sk_data_ready(&lsmc->sk); in smc_listen_out()
1320 sock_put(&lsmc->sk); /* sock_hold in smc_tcp_listen_work */ in smc_listen_out()
1326 struct sock *newsmcsk = &new_smc->sk; in smc_listen_out_connected()
1328 if (newsmcsk->sk_state == SMC_INIT) in smc_listen_out_connected()
1329 newsmcsk->sk_state = SMC_ACTIVE; in smc_listen_out_connected()
1337 struct sock *newsmcsk = &new_smc->sk; in smc_listen_out_err()
1339 if (newsmcsk->sk_state == SMC_INIT) in smc_listen_out_err()
1340 sock_put(&new_smc->sk); /* passive closing */ in smc_listen_out_err()
1341 newsmcsk->sk_state = SMC_CLOSED; in smc_listen_out_err()
1352 smc_lgr_cleanup_early(&new_smc->conn); in smc_listen_decline()
1354 smc_conn_free(&new_smc->conn); in smc_listen_decline()
1360 new_smc->fallback_rsn = reason_code; in smc_listen_decline()
1378 ini->smc_type_v1 = pclc->hdr.typev1; in smc_listen_v2_check()
1379 ini->smc_type_v2 = pclc->hdr.typev2; in smc_listen_v2_check()
1380 ini->smcd_version = ini->smc_type_v1 != SMC_TYPE_N ? SMC_V1 : 0; in smc_listen_v2_check()
1381 if (pclc->hdr.version > SMC_V1) in smc_listen_v2_check()
1382 ini->smcd_version |= in smc_listen_v2_check()
1383 ini->smc_type_v2 != SMC_TYPE_N ? SMC_V2 : 0; in smc_listen_v2_check()
1385 ini->smcd_version &= ~SMC_V2; in smc_listen_v2_check()
1390 ini->smcd_version &= ~SMC_V2; in smc_listen_v2_check()
1395 ini->smcd_version &= ~SMC_V2; in smc_listen_v2_check()
1398 if (!ini->smcd_version) { in smc_listen_v2_check()
1399 if (pclc->hdr.typev1 == SMC_TYPE_B || in smc_listen_v2_check()
1400 pclc->hdr.typev2 == SMC_TYPE_B) in smc_listen_v2_check()
1402 if (pclc->hdr.typev1 == SMC_TYPE_D || in smc_listen_v2_check()
1403 pclc->hdr.typev2 == SMC_TYPE_D) in smc_listen_v2_check()
1416 struct socket *newclcsock = new_smc->clcsock; in smc_listen_prfx_check()
1418 if (pclc->hdr.typev1 == SMC_TYPE_N) in smc_listen_prfx_check()
1445 /* listen worker: initialize connection and buffers for SMC-D */
1458 if (ini->first_contact_local) in smc_listen_ism_init()
1459 smc_lgr_cleanup_early(&new_smc->conn); in smc_listen_ism_init()
1461 smc_conn_free(&new_smc->conn); in smc_listen_ism_init()
1462 return (rc == -ENOSPC) ? SMC_CLC_DECL_MAX_DMB : in smc_listen_ism_init()
1476 if (smcd == ini->ism_dev[i]) in smc_is_already_selected()
1490 if (smcd->going_away) in smc_check_ism_v2_match()
1496 ini->ism_peer_gid[*matches] = proposed_gid; in smc_check_ism_v2_match()
1497 ini->ism_dev[*matches] = smcd; in smc_check_ism_v2_match()
1516 if (!(ini->smcd_version & SMC_V2) || !smcd_indicated(ini->smc_type_v2)) in smc_find_ism_v2_device_serv()
1523 !smc_v2_ext->hdr.flag.seid) /* no system EID support for SMCD */ in smc_find_ism_v2_device_serv()
1527 if (pclc_smcd->ism.chid) in smc_find_ism_v2_device_serv()
1529 smc_check_ism_v2_match(ini, ntohs(pclc_smcd->ism.chid), in smc_find_ism_v2_device_serv()
1530 ntohll(pclc_smcd->ism.gid), &matches); in smc_find_ism_v2_device_serv()
1531 for (i = 1; i <= smc_v2_ext->hdr.ism_gid_cnt; i++) { in smc_find_ism_v2_device_serv()
1532 /* check for ISM devices matching proposed non-native ISM in smc_find_ism_v2_device_serv()
1536 ntohs(smcd_v2_ext->gidchid[i - 1].chid), in smc_find_ism_v2_device_serv()
1537 ntohll(smcd_v2_ext->gidchid[i - 1].gid), in smc_find_ism_v2_device_serv()
1542 if (ini->ism_dev[0]) { in smc_find_ism_v2_device_serv()
1543 smc_ism_get_system_eid(ini->ism_dev[0], &eid); in smc_find_ism_v2_device_serv()
1544 if (memcmp(eid, smcd_v2_ext->system_eid, SMC_MAX_EID_LEN)) in smc_find_ism_v2_device_serv()
1550 /* separate - outside the smcd_dev_list.lock */ in smc_find_ism_v2_device_serv()
1551 smcd_version = ini->smcd_version; in smc_find_ism_v2_device_serv()
1553 ini->smcd_version = SMC_V2; in smc_find_ism_v2_device_serv()
1554 ini->is_smcd = true; in smc_find_ism_v2_device_serv()
1555 ini->ism_selected = i; in smc_find_ism_v2_device_serv()
1562 ini->smcd_version = smcd_version; /* restore original value */ in smc_find_ism_v2_device_serv()
1565 ini->smcd_version &= ~SMC_V2; in smc_find_ism_v2_device_serv()
1566 ini->ism_dev[0] = NULL; in smc_find_ism_v2_device_serv()
1567 ini->is_smcd = false; in smc_find_ism_v2_device_serv()
1577 if (!(ini->smcd_version & SMC_V1) || !smcd_indicated(ini->smc_type_v1)) in smc_find_ism_v1_device_serv()
1579 ini->is_smcd = true; /* prepare ISM check */ in smc_find_ism_v1_device_serv()
1580 ini->ism_peer_gid[0] = ntohll(pclc_smcd->ism.gid); in smc_find_ism_v1_device_serv()
1583 ini->ism_selected = 0; in smc_find_ism_v1_device_serv()
1588 ini->ism_dev[0] = NULL; in smc_find_ism_v1_device_serv()
1589 ini->is_smcd = false; in smc_find_ism_v1_device_serv()
1595 struct smc_connection *conn = &new_smc->conn; in smc_listen_rdma_reg()
1598 if (smcr_lgr_reg_rmbs(conn->lnk, conn->rmb_desc)) in smc_listen_rdma_reg()
1601 smc_rmb_sync_sg_for_device(&new_smc->conn); in smc_listen_rdma_reg()
1612 if (!smcr_indicated(ini->smc_type_v1)) in smc_find_rdma_v1_device_serv()
1616 ini->ib_lcl = &pclc->lcl; in smc_find_rdma_v1_device_serv()
1620 if (ini->smc_type_v1 == SMC_TYPE_B) in smc_find_rdma_v1_device_serv()
1628 return smc_listen_rdma_reg(new_smc, ini->first_contact_local); in smc_find_rdma_v1_device_serv()
1640 if (ini->ism_dev[0]) in smc_listen_find_device()
1643 if (!(ini->smcd_version & SMC_V1)) in smc_listen_find_device()
1651 /* get vlan id from IP device */ in smc_listen_find_device()
1652 if (smc_vlan_by_tcpsk(new_smc->clcsock, ini)) in smc_listen_find_device()
1657 if (ini->ism_dev[0]) in smc_listen_find_device()
1660 if (pclc->hdr.typev1 == SMC_TYPE_D) in smc_listen_find_device()
1672 struct smc_link *link = new_smc->conn.lnk; in smc_listen_rdma_finish()
1678 if (smc_rmb_rtoken_handling(&new_smc->conn, link, cclc)) in smc_listen_rdma_finish()
1685 smc_llc_flow_initiate(link->lgr, SMC_LLC_FLOW_ADD_LINK); in smc_listen_rdma_finish()
1687 smc_llc_flow_stop(link->lgr, &link->lgr->llc_flow_lcl); in smc_listen_rdma_finish()
1698 struct socket *newclcsock = new_smc->clcsock; in smc_listen_work()
1705 if (new_smc->listen_smc->sk.sk_state != SMC_LISTEN) in smc_listen_work()
1708 if (new_smc->use_fallback) { in smc_listen_work()
1713 /* check if peer is smc capable */ in smc_listen_work()
1714 if (!tcp_sk(newclcsock->sk)->syn_smc) { in smc_listen_work()
1716 new_smc->fallback_rsn = SMC_CLC_DECL_PEERNOSMC; in smc_listen_work()
1721 /* do inband token exchange - in smc_listen_work()
1722 * wait for and receive SMC Proposal CLC message in smc_listen_work()
1734 version = pclc->hdr.version == SMC_V1 ? SMC_V1 : version; in smc_listen_work()
1736 /* IPSec connections opt out of SMC optimizations */ in smc_listen_work()
1763 /* send SMC Accept CLC message */ in smc_listen_work()
1764 rc = smc_clc_send_accept(new_smc, ini->first_contact_local, in smc_listen_work()
1765 ini->smcd_version == SMC_V2 ? SMC_V2 : SMC_V1); in smc_listen_work()
1769 /* SMC-D does not need this lock any more */ in smc_listen_work()
1770 if (ini->is_smcd) in smc_listen_work()
1773 /* receive SMC Confirm CLC message */ in smc_listen_work()
1779 if (!ini->is_smcd) in smc_listen_work()
1785 if (!ini->is_smcd) { in smc_listen_work()
1787 ini->first_contact_local); in smc_listen_work()
1799 smc_listen_decline(new_smc, rc, ini ? ini->first_contact_local : 0, in smc_listen_work()
1810 struct sock *lsk = &lsmc->sk; in smc_tcp_listen_work()
1815 while (lsk->sk_state == SMC_LISTEN) { in smc_tcp_listen_work()
1822 new_smc->listen_smc = lsmc; in smc_tcp_listen_work()
1823 new_smc->use_fallback = lsmc->use_fallback; in smc_tcp_listen_work()
1824 new_smc->fallback_rsn = lsmc->fallback_rsn; in smc_tcp_listen_work()
1826 INIT_WORK(&new_smc->smc_listen_work, smc_listen_work); in smc_tcp_listen_work()
1828 new_smc->sk.sk_sndbuf = lsmc->sk.sk_sndbuf; in smc_tcp_listen_work()
1829 new_smc->sk.sk_rcvbuf = lsmc->sk.sk_rcvbuf; in smc_tcp_listen_work()
1830 sock_hold(&new_smc->sk); /* sock_put in passive closing */ in smc_tcp_listen_work()
1831 if (!queue_work(smc_hs_wq, &new_smc->smc_listen_work)) in smc_tcp_listen_work()
1832 sock_put(&new_smc->sk); in smc_tcp_listen_work()
1837 sock_put(&lsmc->sk); /* sock_hold in smc_clcsock_data_ready() */ in smc_tcp_listen_work()
1845 ((uintptr_t)listen_clcsock->sk_user_data & ~SK_USER_DATA_NOCOPY); in smc_clcsock_data_ready()
1848 lsmc->clcsk_data_ready(listen_clcsock); in smc_clcsock_data_ready()
1849 if (lsmc->sk.sk_state == SMC_LISTEN) { in smc_clcsock_data_ready()
1850 sock_hold(&lsmc->sk); /* sock_put in smc_tcp_listen_work() */ in smc_clcsock_data_ready()
1851 if (!queue_work(smc_hs_wq, &lsmc->tcp_listen_work)) in smc_clcsock_data_ready()
1852 sock_put(&lsmc->sk); in smc_clcsock_data_ready()
1858 struct sock *sk = sock->sk; in smc_listen()
1859 struct smc_sock *smc; in smc_listen() local
1862 smc = smc_sk(sk); in smc_listen()
1865 rc = -EINVAL; in smc_listen()
1866 if ((sk->sk_state != SMC_INIT && sk->sk_state != SMC_LISTEN) || in smc_listen()
1867 smc->connect_nonblock) in smc_listen()
1871 if (sk->sk_state == SMC_LISTEN) { in smc_listen()
1872 sk->sk_max_ack_backlog = backlog; in smc_listen()
1876 * them to the clc socket -- copy smc socket options to clc socket in smc_listen()
1878 smc_copy_sock_settings_to_clc(smc); in smc_listen()
1879 if (!smc->use_fallback) in smc_listen()
1880 tcp_sk(smc->clcsock->sk)->syn_smc = 1; in smc_listen()
1883 * smc-specific sk_data_ready function in smc_listen()
1885 smc->clcsk_data_ready = smc->clcsock->sk->sk_data_ready; in smc_listen()
1886 smc->clcsock->sk->sk_data_ready = smc_clcsock_data_ready; in smc_listen()
1887 smc->clcsock->sk->sk_user_data = in smc_listen()
1888 (void *)((uintptr_t)smc | SK_USER_DATA_NOCOPY); in smc_listen()
1889 rc = kernel_listen(smc->clcsock, backlog); in smc_listen()
1891 smc->clcsock->sk->sk_data_ready = smc->clcsk_data_ready; in smc_listen()
1894 sk->sk_max_ack_backlog = backlog; in smc_listen()
1895 sk->sk_ack_backlog = 0; in smc_listen()
1896 sk->sk_state = SMC_LISTEN; in smc_listen()
1906 struct sock *sk = sock->sk, *nsk; in smc_accept()
1916 if (lsmc->sk.sk_state != SMC_LISTEN) { in smc_accept()
1917 rc = -EINVAL; in smc_accept()
1928 rc = -EAGAIN; in smc_accept()
1950 if (lsmc->sockopt_defer_accept && !(flags & O_NONBLOCK)) { in smc_accept()
1952 timeo = msecs_to_jiffies(lsmc->sockopt_defer_accept * in smc_accept()
1954 if (smc_sk(nsk)->use_fallback) { in smc_accept()
1955 struct sock *clcsk = smc_sk(nsk)->clcsock->sk; in smc_accept()
1958 if (skb_queue_empty(&clcsk->sk_receive_queue)) in smc_accept()
1961 } else if (!atomic_read(&smc_sk(nsk)->conn.bytes_to_rcv)) { in smc_accept()
1976 struct smc_sock *smc; in smc_getname() local
1978 if (peer && (sock->sk->sk_state != SMC_ACTIVE) && in smc_getname()
1979 (sock->sk->sk_state != SMC_APPCLOSEWAIT1)) in smc_getname()
1980 return -ENOTCONN; in smc_getname()
1982 smc = smc_sk(sock->sk); in smc_getname()
1984 return smc->clcsock->ops->getname(smc->clcsock, addr, peer); in smc_getname()
1989 struct sock *sk = sock->sk; in smc_sendmsg()
1990 struct smc_sock *smc; in smc_sendmsg() local
1991 int rc = -EPIPE; in smc_sendmsg()
1993 smc = smc_sk(sk); in smc_sendmsg()
1995 if ((sk->sk_state != SMC_ACTIVE) && in smc_sendmsg()
1996 (sk->sk_state != SMC_APPCLOSEWAIT1) && in smc_sendmsg()
1997 (sk->sk_state != SMC_INIT)) in smc_sendmsg()
2000 if (msg->msg_flags & MSG_FASTOPEN) { in smc_sendmsg()
2001 if (sk->sk_state == SMC_INIT && !smc->connect_nonblock) { in smc_sendmsg()
2002 smc_switch_to_fallback(smc); in smc_sendmsg()
2003 smc->fallback_rsn = SMC_CLC_DECL_OPTUNSUPP; in smc_sendmsg()
2005 rc = -EINVAL; in smc_sendmsg()
2010 if (smc->use_fallback) in smc_sendmsg()
2011 rc = smc->clcsock->ops->sendmsg(smc->clcsock, msg, len); in smc_sendmsg()
2013 rc = smc_tx_sendmsg(smc, msg, len); in smc_sendmsg()
2022 struct sock *sk = sock->sk; in smc_recvmsg()
2023 struct smc_sock *smc; in smc_recvmsg() local
2024 int rc = -ENOTCONN; in smc_recvmsg()
2026 smc = smc_sk(sk); in smc_recvmsg()
2028 if (sk->sk_state == SMC_CLOSED && (sk->sk_shutdown & RCV_SHUTDOWN)) { in smc_recvmsg()
2033 if ((sk->sk_state == SMC_INIT) || in smc_recvmsg()
2034 (sk->sk_state == SMC_LISTEN) || in smc_recvmsg()
2035 (sk->sk_state == SMC_CLOSED)) in smc_recvmsg()
2038 if (sk->sk_state == SMC_PEERFINCLOSEWAIT) { in smc_recvmsg()
2043 if (smc->use_fallback) { in smc_recvmsg()
2044 rc = smc->clcsock->ops->recvmsg(smc->clcsock, msg, len, flags); in smc_recvmsg()
2046 msg->msg_namelen = 0; in smc_recvmsg()
2047 rc = smc_rx_recvmsg(smc, msg, NULL, len, flags); in smc_recvmsg()
2060 spin_lock(&isk->accept_q_lock); in smc_accept_poll()
2061 if (!list_empty(&isk->accept_q)) in smc_accept_poll()
2063 spin_unlock(&isk->accept_q_lock); in smc_accept_poll()
2071 struct sock *sk = sock->sk; in smc_poll()
2072 struct smc_sock *smc; in smc_poll() local
2078 smc = smc_sk(sock->sk); in smc_poll()
2079 if (smc->use_fallback) { in smc_poll()
2081 mask = smc->clcsock->ops->poll(file, smc->clcsock, wait); in smc_poll()
2082 sk->sk_err = smc->clcsock->sk->sk_err; in smc_poll()
2084 if (sk->sk_state != SMC_CLOSED) in smc_poll()
2086 if (sk->sk_err) in smc_poll()
2088 if ((sk->sk_shutdown == SHUTDOWN_MASK) || in smc_poll()
2089 (sk->sk_state == SMC_CLOSED)) in smc_poll()
2091 if (sk->sk_state == SMC_LISTEN) { in smc_poll()
2094 } else if (smc->use_fallback) { /* as result of connect_work()*/ in smc_poll()
2095 mask |= smc->clcsock->ops->poll(file, smc->clcsock, in smc_poll()
2097 sk->sk_err = smc->clcsock->sk->sk_err; in smc_poll()
2099 if ((sk->sk_state != SMC_INIT && in smc_poll()
2100 atomic_read(&smc->conn.sndbuf_space)) || in smc_poll()
2101 sk->sk_shutdown & SEND_SHUTDOWN) { in smc_poll()
2105 set_bit(SOCK_NOSPACE, &sk->sk_socket->flags); in smc_poll()
2107 if (atomic_read(&smc->conn.bytes_to_rcv)) in smc_poll()
2109 if (sk->sk_shutdown & RCV_SHUTDOWN) in smc_poll()
2111 if (sk->sk_state == SMC_APPCLOSEWAIT1) in smc_poll()
2113 if (smc->conn.urg_state == SMC_URG_VALID) in smc_poll()
2123 struct sock *sk = sock->sk; in smc_shutdown()
2125 struct smc_sock *smc; in smc_shutdown() local
2126 int rc = -EINVAL; in smc_shutdown()
2130 smc = smc_sk(sk); in smc_shutdown()
2137 rc = -ENOTCONN; in smc_shutdown()
2138 if ((sk->sk_state != SMC_ACTIVE) && in smc_shutdown()
2139 (sk->sk_state != SMC_PEERCLOSEWAIT1) && in smc_shutdown()
2140 (sk->sk_state != SMC_PEERCLOSEWAIT2) && in smc_shutdown()
2141 (sk->sk_state != SMC_APPCLOSEWAIT1) && in smc_shutdown()
2142 (sk->sk_state != SMC_APPCLOSEWAIT2) && in smc_shutdown()
2143 (sk->sk_state != SMC_APPFINCLOSEWAIT)) in smc_shutdown()
2145 if (smc->use_fallback) { in smc_shutdown()
2146 rc = kernel_sock_shutdown(smc->clcsock, how); in smc_shutdown()
2147 sk->sk_shutdown = smc->clcsock->sk->sk_shutdown; in smc_shutdown()
2148 if (sk->sk_shutdown == SHUTDOWN_MASK) { in smc_shutdown()
2149 sk->sk_state = SMC_CLOSED; in smc_shutdown()
2156 old_state = sk->sk_state; in smc_shutdown()
2157 rc = smc_close_active(smc); in smc_shutdown()
2159 sk->sk_state == SMC_PEERCLOSEWAIT1) in smc_shutdown()
2163 rc = smc_close_shutdown_write(smc); in smc_shutdown()
2170 if (do_shutdown && smc->clcsock) in smc_shutdown()
2171 rc1 = kernel_sock_shutdown(smc->clcsock, how); in smc_shutdown()
2173 sk->sk_shutdown |= how + 1; in smc_shutdown()
2183 struct sock *sk = sock->sk; in smc_setsockopt()
2184 struct smc_sock *smc; in smc_setsockopt() local
2188 return -EOPNOTSUPP; in smc_setsockopt()
2190 smc = smc_sk(sk); in smc_setsockopt()
2195 if (unlikely(!smc->clcsock->ops->setsockopt)) in smc_setsockopt()
2196 rc = -EOPNOTSUPP; in smc_setsockopt()
2198 rc = smc->clcsock->ops->setsockopt(smc->clcsock, level, optname, in smc_setsockopt()
2200 if (smc->clcsock->sk->sk_err) { in smc_setsockopt()
2201 sk->sk_err = smc->clcsock->sk->sk_err; in smc_setsockopt()
2202 sk->sk_error_report(sk); in smc_setsockopt()
2206 return -EINVAL; in smc_setsockopt()
2208 return -EFAULT; in smc_setsockopt()
2211 if (rc || smc->use_fallback) in smc_setsockopt()
2218 /* option not supported by SMC */ in smc_setsockopt()
2219 if (sk->sk_state == SMC_INIT && !smc->connect_nonblock) { in smc_setsockopt()
2220 smc_switch_to_fallback(smc); in smc_setsockopt()
2221 smc->fallback_rsn = SMC_CLC_DECL_OPTUNSUPP; in smc_setsockopt()
2223 rc = -EINVAL; in smc_setsockopt()
2227 if (sk->sk_state != SMC_INIT && in smc_setsockopt()
2228 sk->sk_state != SMC_LISTEN && in smc_setsockopt()
2229 sk->sk_state != SMC_CLOSED) { in smc_setsockopt()
2231 mod_delayed_work(smc->conn.lgr->tx_wq, in smc_setsockopt()
2232 &smc->conn.tx_work, 0); in smc_setsockopt()
2236 if (sk->sk_state != SMC_INIT && in smc_setsockopt()
2237 sk->sk_state != SMC_LISTEN && in smc_setsockopt()
2238 sk->sk_state != SMC_CLOSED) { in smc_setsockopt()
2240 mod_delayed_work(smc->conn.lgr->tx_wq, in smc_setsockopt()
2241 &smc->conn.tx_work, 0); in smc_setsockopt()
2245 smc->sockopt_defer_accept = val; in smc_setsockopt()
2259 struct smc_sock *smc; in smc_getsockopt() local
2261 smc = smc_sk(sock->sk); in smc_getsockopt()
2263 if (unlikely(!smc->clcsock->ops->getsockopt)) in smc_getsockopt()
2264 return -EOPNOTSUPP; in smc_getsockopt()
2265 return smc->clcsock->ops->getsockopt(smc->clcsock, level, optname, in smc_getsockopt()
2274 struct smc_sock *smc; in smc_ioctl() local
2277 smc = smc_sk(sock->sk); in smc_ioctl()
2278 conn = &smc->conn; in smc_ioctl()
2279 lock_sock(&smc->sk); in smc_ioctl()
2280 if (smc->use_fallback) { in smc_ioctl()
2281 if (!smc->clcsock) { in smc_ioctl()
2282 release_sock(&smc->sk); in smc_ioctl()
2283 return -EBADF; in smc_ioctl()
2285 answ = smc->clcsock->ops->ioctl(smc->clcsock, cmd, arg); in smc_ioctl()
2286 release_sock(&smc->sk); in smc_ioctl()
2291 if (smc->sk.sk_state == SMC_LISTEN) { in smc_ioctl()
2292 release_sock(&smc->sk); in smc_ioctl()
2293 return -EINVAL; in smc_ioctl()
2295 if (smc->sk.sk_state == SMC_INIT || in smc_ioctl()
2296 smc->sk.sk_state == SMC_CLOSED) in smc_ioctl()
2299 answ = atomic_read(&smc->conn.bytes_to_rcv); in smc_ioctl()
2303 if (smc->sk.sk_state == SMC_LISTEN) { in smc_ioctl()
2304 release_sock(&smc->sk); in smc_ioctl()
2305 return -EINVAL; in smc_ioctl()
2307 if (smc->sk.sk_state == SMC_INIT || in smc_ioctl()
2308 smc->sk.sk_state == SMC_CLOSED) in smc_ioctl()
2311 answ = smc->conn.sndbuf_desc->len - in smc_ioctl()
2312 atomic_read(&smc->conn.sndbuf_space); in smc_ioctl()
2316 if (smc->sk.sk_state == SMC_LISTEN) { in smc_ioctl()
2317 release_sock(&smc->sk); in smc_ioctl()
2318 return -EINVAL; in smc_ioctl()
2320 if (smc->sk.sk_state == SMC_INIT || in smc_ioctl()
2321 smc->sk.sk_state == SMC_CLOSED) in smc_ioctl()
2324 answ = smc_tx_prepared_sends(&smc->conn); in smc_ioctl()
2327 if (smc->sk.sk_state == SMC_LISTEN) { in smc_ioctl()
2328 release_sock(&smc->sk); in smc_ioctl()
2329 return -EINVAL; in smc_ioctl()
2331 if (smc->sk.sk_state == SMC_INIT || in smc_ioctl()
2332 smc->sk.sk_state == SMC_CLOSED) { in smc_ioctl()
2335 smc_curs_copy(&cons, &conn->local_tx_ctrl.cons, conn); in smc_ioctl()
2336 smc_curs_copy(&urg, &conn->urg_curs, conn); in smc_ioctl()
2337 answ = smc_curs_diff(conn->rmb_desc->len, in smc_ioctl()
2342 release_sock(&smc->sk); in smc_ioctl()
2343 return -ENOIOCTLCMD; in smc_ioctl()
2345 release_sock(&smc->sk); in smc_ioctl()
2353 struct sock *sk = sock->sk; in smc_sendpage()
2354 struct smc_sock *smc; in smc_sendpage() local
2355 int rc = -EPIPE; in smc_sendpage()
2357 smc = smc_sk(sk); in smc_sendpage()
2359 if (sk->sk_state != SMC_ACTIVE) { in smc_sendpage()
2364 if (smc->use_fallback) in smc_sendpage()
2365 rc = kernel_sendpage(smc->clcsock, page, offset, in smc_sendpage()
2375 * to splice in conn->splice_pending, and press 'go'. Delays consumer cursor
2384 struct sock *sk = sock->sk; in smc_splice_read()
2385 struct smc_sock *smc; in smc_splice_read() local
2386 int rc = -ENOTCONN; in smc_splice_read()
2388 smc = smc_sk(sk); in smc_splice_read()
2390 if (sk->sk_state == SMC_CLOSED && (sk->sk_shutdown & RCV_SHUTDOWN)) { in smc_splice_read()
2395 if (sk->sk_state == SMC_INIT || in smc_splice_read()
2396 sk->sk_state == SMC_LISTEN || in smc_splice_read()
2397 sk->sk_state == SMC_CLOSED) in smc_splice_read()
2400 if (sk->sk_state == SMC_PEERFINCLOSEWAIT) { in smc_splice_read()
2405 if (smc->use_fallback) { in smc_splice_read()
2406 rc = smc->clcsock->ops->splice_read(smc->clcsock, ppos, in smc_splice_read()
2410 rc = -ESPIPE; in smc_splice_read()
2417 rc = smc_rx_recvmsg(smc, NULL, pipe, len, flags); in smc_splice_read()
2452 struct smc_sock *smc; in smc_create() local
2456 rc = -ESOCKTNOSUPPORT; in smc_create()
2457 if (sock->type != SOCK_STREAM) in smc_create()
2460 rc = -EPROTONOSUPPORT; in smc_create()
2464 rc = -ENOBUFS; in smc_create()
2465 sock->ops = &smc_sock_ops; in smc_create()
2471 smc = smc_sk(sk); in smc_create()
2472 smc->use_fallback = false; /* assume rdma capability first */ in smc_create()
2473 smc->fallback_rsn = 0; in smc_create()
2475 &smc->clcsock); in smc_create()
2480 smc->sk.sk_sndbuf = max(smc->clcsock->sk->sk_sndbuf, SMC_BUF_MIN_SIZE); in smc_create()
2481 smc->sk.sk_rcvbuf = max(smc->clcsock->sk->sk_rcvbuf, SMC_BUF_MIN_SIZE); in smc_create()
2508 .id = &smc_net_id,
2527 rc = -ENOMEM; in smc_init()
2622 MODULE_DESCRIPTION("smc socket address family");