xref: /OK3568_Linux_fs/kernel/drivers/infiniband/hw/hfi1/opfn.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun // SPDX-License-Identifier: (GPL-2.0 OR BSD-3-Clause)
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun  * Copyright(c) 2018 Intel Corporation.
4*4882a593Smuzhiyun  *
5*4882a593Smuzhiyun  */
6*4882a593Smuzhiyun #include "hfi.h"
7*4882a593Smuzhiyun #include "trace.h"
8*4882a593Smuzhiyun #include "qp.h"
9*4882a593Smuzhiyun #include "opfn.h"
10*4882a593Smuzhiyun 
11*4882a593Smuzhiyun #define IB_BTHE_E                 BIT(IB_BTHE_E_SHIFT)
12*4882a593Smuzhiyun 
13*4882a593Smuzhiyun #define OPFN_CODE(code) BIT((code) - 1)
14*4882a593Smuzhiyun #define OPFN_MASK(code) OPFN_CODE(STL_VERBS_EXTD_##code)
15*4882a593Smuzhiyun 
16*4882a593Smuzhiyun struct hfi1_opfn_type {
17*4882a593Smuzhiyun 	bool (*request)(struct rvt_qp *qp, u64 *data);
18*4882a593Smuzhiyun 	bool (*response)(struct rvt_qp *qp, u64 *data);
19*4882a593Smuzhiyun 	bool (*reply)(struct rvt_qp *qp, u64 data);
20*4882a593Smuzhiyun 	void (*error)(struct rvt_qp *qp);
21*4882a593Smuzhiyun };
22*4882a593Smuzhiyun 
23*4882a593Smuzhiyun static struct hfi1_opfn_type hfi1_opfn_handlers[STL_VERBS_EXTD_MAX] = {
24*4882a593Smuzhiyun 	[STL_VERBS_EXTD_TID_RDMA] = {
25*4882a593Smuzhiyun 		.request = tid_rdma_conn_req,
26*4882a593Smuzhiyun 		.response = tid_rdma_conn_resp,
27*4882a593Smuzhiyun 		.reply = tid_rdma_conn_reply,
28*4882a593Smuzhiyun 		.error = tid_rdma_conn_error,
29*4882a593Smuzhiyun 	},
30*4882a593Smuzhiyun };
31*4882a593Smuzhiyun 
32*4882a593Smuzhiyun static struct workqueue_struct *opfn_wq;
33*4882a593Smuzhiyun 
34*4882a593Smuzhiyun static void opfn_schedule_conn_request(struct rvt_qp *qp);
35*4882a593Smuzhiyun 
hfi1_opfn_extended(u32 bth1)36*4882a593Smuzhiyun static bool hfi1_opfn_extended(u32 bth1)
37*4882a593Smuzhiyun {
38*4882a593Smuzhiyun 	return !!(bth1 & IB_BTHE_E);
39*4882a593Smuzhiyun }
40*4882a593Smuzhiyun 
opfn_conn_request(struct rvt_qp * qp)41*4882a593Smuzhiyun static void opfn_conn_request(struct rvt_qp *qp)
42*4882a593Smuzhiyun {
43*4882a593Smuzhiyun 	struct hfi1_qp_priv *priv = qp->priv;
44*4882a593Smuzhiyun 	struct ib_atomic_wr wr;
45*4882a593Smuzhiyun 	u16 mask, capcode;
46*4882a593Smuzhiyun 	struct hfi1_opfn_type *extd;
47*4882a593Smuzhiyun 	u64 data;
48*4882a593Smuzhiyun 	unsigned long flags;
49*4882a593Smuzhiyun 	int ret = 0;
50*4882a593Smuzhiyun 
51*4882a593Smuzhiyun 	trace_hfi1_opfn_state_conn_request(qp);
52*4882a593Smuzhiyun 	spin_lock_irqsave(&priv->opfn.lock, flags);
53*4882a593Smuzhiyun 	/*
54*4882a593Smuzhiyun 	 * Exit if the extended bit is not set, or if nothing is requested, or
55*4882a593Smuzhiyun 	 * if we have completed all requests, or if a previous request is in
56*4882a593Smuzhiyun 	 * progress
57*4882a593Smuzhiyun 	 */
58*4882a593Smuzhiyun 	if (!priv->opfn.extended || !priv->opfn.requested ||
59*4882a593Smuzhiyun 	    priv->opfn.requested == priv->opfn.completed || priv->opfn.curr)
60*4882a593Smuzhiyun 		goto done;
61*4882a593Smuzhiyun 
62*4882a593Smuzhiyun 	mask = priv->opfn.requested & ~priv->opfn.completed;
63*4882a593Smuzhiyun 	capcode = ilog2(mask & ~(mask - 1)) + 1;
64*4882a593Smuzhiyun 	if (capcode >= STL_VERBS_EXTD_MAX) {
65*4882a593Smuzhiyun 		priv->opfn.completed |= OPFN_CODE(capcode);
66*4882a593Smuzhiyun 		goto done;
67*4882a593Smuzhiyun 	}
68*4882a593Smuzhiyun 
69*4882a593Smuzhiyun 	extd = &hfi1_opfn_handlers[capcode];
70*4882a593Smuzhiyun 	if (!extd || !extd->request || !extd->request(qp, &data)) {
71*4882a593Smuzhiyun 		/*
72*4882a593Smuzhiyun 		 * Either there is no handler for this capability or the request
73*4882a593Smuzhiyun 		 * packet could not be generated. Either way, mark it as done so
74*4882a593Smuzhiyun 		 * we don't keep attempting to complete it.
75*4882a593Smuzhiyun 		 */
76*4882a593Smuzhiyun 		priv->opfn.completed |= OPFN_CODE(capcode);
77*4882a593Smuzhiyun 		goto done;
78*4882a593Smuzhiyun 	}
79*4882a593Smuzhiyun 
80*4882a593Smuzhiyun 	trace_hfi1_opfn_data_conn_request(qp, capcode, data);
81*4882a593Smuzhiyun 	data = (data & ~0xf) | capcode;
82*4882a593Smuzhiyun 
83*4882a593Smuzhiyun 	memset(&wr, 0, sizeof(wr));
84*4882a593Smuzhiyun 	wr.wr.opcode = IB_WR_OPFN;
85*4882a593Smuzhiyun 	wr.remote_addr = HFI1_VERBS_E_ATOMIC_VADDR;
86*4882a593Smuzhiyun 	wr.compare_add = data;
87*4882a593Smuzhiyun 
88*4882a593Smuzhiyun 	priv->opfn.curr = capcode;	/* A new request is now in progress */
89*4882a593Smuzhiyun 	/* Drop opfn.lock before calling ib_post_send() */
90*4882a593Smuzhiyun 	spin_unlock_irqrestore(&priv->opfn.lock, flags);
91*4882a593Smuzhiyun 
92*4882a593Smuzhiyun 	ret = ib_post_send(&qp->ibqp, &wr.wr, NULL);
93*4882a593Smuzhiyun 	if (ret)
94*4882a593Smuzhiyun 		goto err;
95*4882a593Smuzhiyun 	trace_hfi1_opfn_state_conn_request(qp);
96*4882a593Smuzhiyun 	return;
97*4882a593Smuzhiyun err:
98*4882a593Smuzhiyun 	trace_hfi1_msg_opfn_conn_request(qp, "ib_ost_send failed: ret = ",
99*4882a593Smuzhiyun 					 (u64)ret);
100*4882a593Smuzhiyun 	spin_lock_irqsave(&priv->opfn.lock, flags);
101*4882a593Smuzhiyun 	/*
102*4882a593Smuzhiyun 	 * In case of an unexpected error return from ib_post_send
103*4882a593Smuzhiyun 	 * clear opfn.curr and reschedule to try again
104*4882a593Smuzhiyun 	 */
105*4882a593Smuzhiyun 	priv->opfn.curr = STL_VERBS_EXTD_NONE;
106*4882a593Smuzhiyun 	opfn_schedule_conn_request(qp);
107*4882a593Smuzhiyun done:
108*4882a593Smuzhiyun 	spin_unlock_irqrestore(&priv->opfn.lock, flags);
109*4882a593Smuzhiyun }
110*4882a593Smuzhiyun 
opfn_send_conn_request(struct work_struct * work)111*4882a593Smuzhiyun void opfn_send_conn_request(struct work_struct *work)
112*4882a593Smuzhiyun {
113*4882a593Smuzhiyun 	struct hfi1_opfn_data *od;
114*4882a593Smuzhiyun 	struct hfi1_qp_priv *qpriv;
115*4882a593Smuzhiyun 
116*4882a593Smuzhiyun 	od = container_of(work, struct hfi1_opfn_data, opfn_work);
117*4882a593Smuzhiyun 	qpriv = container_of(od, struct hfi1_qp_priv, opfn);
118*4882a593Smuzhiyun 
119*4882a593Smuzhiyun 	opfn_conn_request(qpriv->owner);
120*4882a593Smuzhiyun }
121*4882a593Smuzhiyun 
122*4882a593Smuzhiyun /*
123*4882a593Smuzhiyun  * When QP s_lock is held in the caller, the OPFN request must be scheduled
124*4882a593Smuzhiyun  * to a different workqueue to avoid double locking QP s_lock in call to
125*4882a593Smuzhiyun  * ib_post_send in opfn_conn_request
126*4882a593Smuzhiyun  */
opfn_schedule_conn_request(struct rvt_qp * qp)127*4882a593Smuzhiyun static void opfn_schedule_conn_request(struct rvt_qp *qp)
128*4882a593Smuzhiyun {
129*4882a593Smuzhiyun 	struct hfi1_qp_priv *priv = qp->priv;
130*4882a593Smuzhiyun 
131*4882a593Smuzhiyun 	trace_hfi1_opfn_state_sched_conn_request(qp);
132*4882a593Smuzhiyun 	queue_work(opfn_wq, &priv->opfn.opfn_work);
133*4882a593Smuzhiyun }
134*4882a593Smuzhiyun 
opfn_conn_response(struct rvt_qp * qp,struct rvt_ack_entry * e,struct ib_atomic_eth * ateth)135*4882a593Smuzhiyun void opfn_conn_response(struct rvt_qp *qp, struct rvt_ack_entry *e,
136*4882a593Smuzhiyun 			struct ib_atomic_eth *ateth)
137*4882a593Smuzhiyun {
138*4882a593Smuzhiyun 	struct hfi1_qp_priv *priv = qp->priv;
139*4882a593Smuzhiyun 	u64 data = be64_to_cpu(ateth->compare_data);
140*4882a593Smuzhiyun 	struct hfi1_opfn_type *extd;
141*4882a593Smuzhiyun 	u8 capcode;
142*4882a593Smuzhiyun 	unsigned long flags;
143*4882a593Smuzhiyun 
144*4882a593Smuzhiyun 	trace_hfi1_opfn_state_conn_response(qp);
145*4882a593Smuzhiyun 	capcode = data & 0xf;
146*4882a593Smuzhiyun 	trace_hfi1_opfn_data_conn_response(qp, capcode, data);
147*4882a593Smuzhiyun 	if (!capcode || capcode >= STL_VERBS_EXTD_MAX)
148*4882a593Smuzhiyun 		return;
149*4882a593Smuzhiyun 
150*4882a593Smuzhiyun 	extd = &hfi1_opfn_handlers[capcode];
151*4882a593Smuzhiyun 
152*4882a593Smuzhiyun 	if (!extd || !extd->response) {
153*4882a593Smuzhiyun 		e->atomic_data = capcode;
154*4882a593Smuzhiyun 		return;
155*4882a593Smuzhiyun 	}
156*4882a593Smuzhiyun 
157*4882a593Smuzhiyun 	spin_lock_irqsave(&priv->opfn.lock, flags);
158*4882a593Smuzhiyun 	if (priv->opfn.completed & OPFN_CODE(capcode)) {
159*4882a593Smuzhiyun 		/*
160*4882a593Smuzhiyun 		 * We are receiving a request for a feature that has already
161*4882a593Smuzhiyun 		 * been negotiated. This may mean that the other side has reset
162*4882a593Smuzhiyun 		 */
163*4882a593Smuzhiyun 		priv->opfn.completed &= ~OPFN_CODE(capcode);
164*4882a593Smuzhiyun 		if (extd->error)
165*4882a593Smuzhiyun 			extd->error(qp);
166*4882a593Smuzhiyun 	}
167*4882a593Smuzhiyun 
168*4882a593Smuzhiyun 	if (extd->response(qp, &data))
169*4882a593Smuzhiyun 		priv->opfn.completed |= OPFN_CODE(capcode);
170*4882a593Smuzhiyun 	e->atomic_data = (data & ~0xf) | capcode;
171*4882a593Smuzhiyun 	trace_hfi1_opfn_state_conn_response(qp);
172*4882a593Smuzhiyun 	spin_unlock_irqrestore(&priv->opfn.lock, flags);
173*4882a593Smuzhiyun }
174*4882a593Smuzhiyun 
opfn_conn_reply(struct rvt_qp * qp,u64 data)175*4882a593Smuzhiyun void opfn_conn_reply(struct rvt_qp *qp, u64 data)
176*4882a593Smuzhiyun {
177*4882a593Smuzhiyun 	struct hfi1_qp_priv *priv = qp->priv;
178*4882a593Smuzhiyun 	struct hfi1_opfn_type *extd;
179*4882a593Smuzhiyun 	u8 capcode;
180*4882a593Smuzhiyun 	unsigned long flags;
181*4882a593Smuzhiyun 
182*4882a593Smuzhiyun 	trace_hfi1_opfn_state_conn_reply(qp);
183*4882a593Smuzhiyun 	capcode = data & 0xf;
184*4882a593Smuzhiyun 	trace_hfi1_opfn_data_conn_reply(qp, capcode, data);
185*4882a593Smuzhiyun 	if (!capcode || capcode >= STL_VERBS_EXTD_MAX)
186*4882a593Smuzhiyun 		return;
187*4882a593Smuzhiyun 
188*4882a593Smuzhiyun 	spin_lock_irqsave(&priv->opfn.lock, flags);
189*4882a593Smuzhiyun 	/*
190*4882a593Smuzhiyun 	 * Either there is no previous request or the reply is not for the
191*4882a593Smuzhiyun 	 * current request
192*4882a593Smuzhiyun 	 */
193*4882a593Smuzhiyun 	if (!priv->opfn.curr || capcode != priv->opfn.curr)
194*4882a593Smuzhiyun 		goto done;
195*4882a593Smuzhiyun 
196*4882a593Smuzhiyun 	extd = &hfi1_opfn_handlers[capcode];
197*4882a593Smuzhiyun 
198*4882a593Smuzhiyun 	if (!extd || !extd->reply)
199*4882a593Smuzhiyun 		goto clear;
200*4882a593Smuzhiyun 
201*4882a593Smuzhiyun 	if (extd->reply(qp, data))
202*4882a593Smuzhiyun 		priv->opfn.completed |= OPFN_CODE(capcode);
203*4882a593Smuzhiyun clear:
204*4882a593Smuzhiyun 	/*
205*4882a593Smuzhiyun 	 * Clear opfn.curr to indicate that the previous request is no longer in
206*4882a593Smuzhiyun 	 * progress
207*4882a593Smuzhiyun 	 */
208*4882a593Smuzhiyun 	priv->opfn.curr = STL_VERBS_EXTD_NONE;
209*4882a593Smuzhiyun 	trace_hfi1_opfn_state_conn_reply(qp);
210*4882a593Smuzhiyun done:
211*4882a593Smuzhiyun 	spin_unlock_irqrestore(&priv->opfn.lock, flags);
212*4882a593Smuzhiyun }
213*4882a593Smuzhiyun 
opfn_conn_error(struct rvt_qp * qp)214*4882a593Smuzhiyun void opfn_conn_error(struct rvt_qp *qp)
215*4882a593Smuzhiyun {
216*4882a593Smuzhiyun 	struct hfi1_qp_priv *priv = qp->priv;
217*4882a593Smuzhiyun 	struct hfi1_opfn_type *extd = NULL;
218*4882a593Smuzhiyun 	unsigned long flags;
219*4882a593Smuzhiyun 	u16 capcode;
220*4882a593Smuzhiyun 
221*4882a593Smuzhiyun 	trace_hfi1_opfn_state_conn_error(qp);
222*4882a593Smuzhiyun 	trace_hfi1_msg_opfn_conn_error(qp, "error. qp state ", (u64)qp->state);
223*4882a593Smuzhiyun 	/*
224*4882a593Smuzhiyun 	 * The QP has gone into the Error state. We have to invalidate all
225*4882a593Smuzhiyun 	 * negotiated feature, including the one in progress (if any). The RC
226*4882a593Smuzhiyun 	 * QP handling will clean the WQE for the connection request.
227*4882a593Smuzhiyun 	 */
228*4882a593Smuzhiyun 	spin_lock_irqsave(&priv->opfn.lock, flags);
229*4882a593Smuzhiyun 	while (priv->opfn.completed) {
230*4882a593Smuzhiyun 		capcode = priv->opfn.completed & ~(priv->opfn.completed - 1);
231*4882a593Smuzhiyun 		extd = &hfi1_opfn_handlers[ilog2(capcode) + 1];
232*4882a593Smuzhiyun 		if (extd->error)
233*4882a593Smuzhiyun 			extd->error(qp);
234*4882a593Smuzhiyun 		priv->opfn.completed &= ~OPFN_CODE(capcode);
235*4882a593Smuzhiyun 	}
236*4882a593Smuzhiyun 	priv->opfn.extended = 0;
237*4882a593Smuzhiyun 	priv->opfn.requested = 0;
238*4882a593Smuzhiyun 	priv->opfn.curr = STL_VERBS_EXTD_NONE;
239*4882a593Smuzhiyun 	spin_unlock_irqrestore(&priv->opfn.lock, flags);
240*4882a593Smuzhiyun }
241*4882a593Smuzhiyun 
opfn_qp_init(struct rvt_qp * qp,struct ib_qp_attr * attr,int attr_mask)242*4882a593Smuzhiyun void opfn_qp_init(struct rvt_qp *qp, struct ib_qp_attr *attr, int attr_mask)
243*4882a593Smuzhiyun {
244*4882a593Smuzhiyun 	struct ib_qp *ibqp = &qp->ibqp;
245*4882a593Smuzhiyun 	struct hfi1_qp_priv *priv = qp->priv;
246*4882a593Smuzhiyun 	unsigned long flags;
247*4882a593Smuzhiyun 
248*4882a593Smuzhiyun 	if (attr_mask & IB_QP_RETRY_CNT)
249*4882a593Smuzhiyun 		priv->s_retry = attr->retry_cnt;
250*4882a593Smuzhiyun 
251*4882a593Smuzhiyun 	spin_lock_irqsave(&priv->opfn.lock, flags);
252*4882a593Smuzhiyun 	if (ibqp->qp_type == IB_QPT_RC && HFI1_CAP_IS_KSET(TID_RDMA)) {
253*4882a593Smuzhiyun 		struct tid_rdma_params *local = &priv->tid_rdma.local;
254*4882a593Smuzhiyun 
255*4882a593Smuzhiyun 		if (attr_mask & IB_QP_TIMEOUT)
256*4882a593Smuzhiyun 			priv->tid_retry_timeout_jiffies = qp->timeout_jiffies;
257*4882a593Smuzhiyun 		if (qp->pmtu == enum_to_mtu(OPA_MTU_4096) ||
258*4882a593Smuzhiyun 		    qp->pmtu == enum_to_mtu(OPA_MTU_8192)) {
259*4882a593Smuzhiyun 			tid_rdma_opfn_init(qp, local);
260*4882a593Smuzhiyun 			/*
261*4882a593Smuzhiyun 			 * We only want to set the OPFN requested bit when the
262*4882a593Smuzhiyun 			 * QP transitions to RTS.
263*4882a593Smuzhiyun 			 */
264*4882a593Smuzhiyun 			if (attr_mask & IB_QP_STATE &&
265*4882a593Smuzhiyun 			    attr->qp_state == IB_QPS_RTS) {
266*4882a593Smuzhiyun 				priv->opfn.requested |= OPFN_MASK(TID_RDMA);
267*4882a593Smuzhiyun 				/*
268*4882a593Smuzhiyun 				 * If the QP is transitioning to RTS and the
269*4882a593Smuzhiyun 				 * opfn.completed for TID RDMA has already been
270*4882a593Smuzhiyun 				 * set, the QP is being moved *back* into RTS.
271*4882a593Smuzhiyun 				 * We can now renegotiate the TID RDMA
272*4882a593Smuzhiyun 				 * parameters.
273*4882a593Smuzhiyun 				 */
274*4882a593Smuzhiyun 				if (priv->opfn.completed &
275*4882a593Smuzhiyun 				    OPFN_MASK(TID_RDMA)) {
276*4882a593Smuzhiyun 					priv->opfn.completed &=
277*4882a593Smuzhiyun 						~OPFN_MASK(TID_RDMA);
278*4882a593Smuzhiyun 					/*
279*4882a593Smuzhiyun 					 * Since the opfn.completed bit was
280*4882a593Smuzhiyun 					 * already set, it is safe to assume
281*4882a593Smuzhiyun 					 * that the opfn.extended is also set.
282*4882a593Smuzhiyun 					 */
283*4882a593Smuzhiyun 					opfn_schedule_conn_request(qp);
284*4882a593Smuzhiyun 				}
285*4882a593Smuzhiyun 			}
286*4882a593Smuzhiyun 		} else {
287*4882a593Smuzhiyun 			memset(local, 0, sizeof(*local));
288*4882a593Smuzhiyun 		}
289*4882a593Smuzhiyun 	}
290*4882a593Smuzhiyun 	spin_unlock_irqrestore(&priv->opfn.lock, flags);
291*4882a593Smuzhiyun }
292*4882a593Smuzhiyun 
opfn_trigger_conn_request(struct rvt_qp * qp,u32 bth1)293*4882a593Smuzhiyun void opfn_trigger_conn_request(struct rvt_qp *qp, u32 bth1)
294*4882a593Smuzhiyun {
295*4882a593Smuzhiyun 	struct hfi1_qp_priv *priv = qp->priv;
296*4882a593Smuzhiyun 
297*4882a593Smuzhiyun 	if (!priv->opfn.extended && hfi1_opfn_extended(bth1) &&
298*4882a593Smuzhiyun 	    HFI1_CAP_IS_KSET(OPFN)) {
299*4882a593Smuzhiyun 		priv->opfn.extended = 1;
300*4882a593Smuzhiyun 		if (qp->state == IB_QPS_RTS)
301*4882a593Smuzhiyun 			opfn_conn_request(qp);
302*4882a593Smuzhiyun 	}
303*4882a593Smuzhiyun }
304*4882a593Smuzhiyun 
opfn_init(void)305*4882a593Smuzhiyun int opfn_init(void)
306*4882a593Smuzhiyun {
307*4882a593Smuzhiyun 	opfn_wq = alloc_workqueue("hfi_opfn",
308*4882a593Smuzhiyun 				  WQ_SYSFS | WQ_HIGHPRI | WQ_CPU_INTENSIVE |
309*4882a593Smuzhiyun 				  WQ_MEM_RECLAIM,
310*4882a593Smuzhiyun 				  HFI1_MAX_ACTIVE_WORKQUEUE_ENTRIES);
311*4882a593Smuzhiyun 	if (!opfn_wq)
312*4882a593Smuzhiyun 		return -ENOMEM;
313*4882a593Smuzhiyun 
314*4882a593Smuzhiyun 	return 0;
315*4882a593Smuzhiyun }
316*4882a593Smuzhiyun 
opfn_exit(void)317*4882a593Smuzhiyun void opfn_exit(void)
318*4882a593Smuzhiyun {
319*4882a593Smuzhiyun 	if (opfn_wq) {
320*4882a593Smuzhiyun 		destroy_workqueue(opfn_wq);
321*4882a593Smuzhiyun 		opfn_wq = NULL;
322*4882a593Smuzhiyun 	}
323*4882a593Smuzhiyun }
324