1*4882a593Smuzhiyun /*
2*4882a593Smuzhiyun * Copyright (c) 2004, 2005 Topspin Communications. All rights reserved.
3*4882a593Smuzhiyun * Copyright (c) 2005 Sun Microsystems, Inc. All rights reserved.
4*4882a593Smuzhiyun * Copyright (c) 2005, 2006 Cisco Systems, Inc. All rights reserved.
5*4882a593Smuzhiyun * Copyright (c) 2005 Mellanox Technologies. All rights reserved.
6*4882a593Smuzhiyun * Copyright (c) 2004 Voltaire, Inc. All rights reserved.
7*4882a593Smuzhiyun *
8*4882a593Smuzhiyun * This software is available to you under a choice of one of two
9*4882a593Smuzhiyun * licenses. You may choose to be licensed under the terms of the GNU
10*4882a593Smuzhiyun * General Public License (GPL) Version 2, available from the file
11*4882a593Smuzhiyun * COPYING in the main directory of this source tree, or the
12*4882a593Smuzhiyun * OpenIB.org BSD license below:
13*4882a593Smuzhiyun *
14*4882a593Smuzhiyun * Redistribution and use in source and binary forms, with or
15*4882a593Smuzhiyun * without modification, are permitted provided that the following
16*4882a593Smuzhiyun * conditions are met:
17*4882a593Smuzhiyun *
18*4882a593Smuzhiyun * - Redistributions of source code must retain the above
19*4882a593Smuzhiyun * copyright notice, this list of conditions and the following
20*4882a593Smuzhiyun * disclaimer.
21*4882a593Smuzhiyun *
22*4882a593Smuzhiyun * - Redistributions in binary form must reproduce the above
23*4882a593Smuzhiyun * copyright notice, this list of conditions and the following
24*4882a593Smuzhiyun * disclaimer in the documentation and/or other materials
25*4882a593Smuzhiyun * provided with the distribution.
26*4882a593Smuzhiyun *
27*4882a593Smuzhiyun * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
28*4882a593Smuzhiyun * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
29*4882a593Smuzhiyun * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
30*4882a593Smuzhiyun * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
31*4882a593Smuzhiyun * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
32*4882a593Smuzhiyun * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
33*4882a593Smuzhiyun * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
34*4882a593Smuzhiyun * SOFTWARE.
35*4882a593Smuzhiyun */
36*4882a593Smuzhiyun
37*4882a593Smuzhiyun #include <linux/gfp.h>
38*4882a593Smuzhiyun #include <linux/hardirq.h>
39*4882a593Smuzhiyun #include <linux/sched.h>
40*4882a593Smuzhiyun
41*4882a593Smuzhiyun #include <asm/io.h>
42*4882a593Smuzhiyun
43*4882a593Smuzhiyun #include <rdma/ib_pack.h>
44*4882a593Smuzhiyun
45*4882a593Smuzhiyun #include "mthca_dev.h"
46*4882a593Smuzhiyun #include "mthca_cmd.h"
47*4882a593Smuzhiyun #include "mthca_memfree.h"
48*4882a593Smuzhiyun
49*4882a593Smuzhiyun enum {
50*4882a593Smuzhiyun MTHCA_MAX_DIRECT_CQ_SIZE = 4 * PAGE_SIZE
51*4882a593Smuzhiyun };
52*4882a593Smuzhiyun
53*4882a593Smuzhiyun enum {
54*4882a593Smuzhiyun MTHCA_CQ_ENTRY_SIZE = 0x20
55*4882a593Smuzhiyun };
56*4882a593Smuzhiyun
57*4882a593Smuzhiyun enum {
58*4882a593Smuzhiyun MTHCA_ATOMIC_BYTE_LEN = 8
59*4882a593Smuzhiyun };
60*4882a593Smuzhiyun
61*4882a593Smuzhiyun /*
62*4882a593Smuzhiyun * Must be packed because start is 64 bits but only aligned to 32 bits.
63*4882a593Smuzhiyun */
64*4882a593Smuzhiyun struct mthca_cq_context {
65*4882a593Smuzhiyun __be32 flags;
66*4882a593Smuzhiyun __be64 start;
67*4882a593Smuzhiyun __be32 logsize_usrpage;
68*4882a593Smuzhiyun __be32 error_eqn; /* Tavor only */
69*4882a593Smuzhiyun __be32 comp_eqn;
70*4882a593Smuzhiyun __be32 pd;
71*4882a593Smuzhiyun __be32 lkey;
72*4882a593Smuzhiyun __be32 last_notified_index;
73*4882a593Smuzhiyun __be32 solicit_producer_index;
74*4882a593Smuzhiyun __be32 consumer_index;
75*4882a593Smuzhiyun __be32 producer_index;
76*4882a593Smuzhiyun __be32 cqn;
77*4882a593Smuzhiyun __be32 ci_db; /* Arbel only */
78*4882a593Smuzhiyun __be32 state_db; /* Arbel only */
79*4882a593Smuzhiyun u32 reserved;
80*4882a593Smuzhiyun } __packed;
81*4882a593Smuzhiyun
82*4882a593Smuzhiyun #define MTHCA_CQ_STATUS_OK ( 0 << 28)
83*4882a593Smuzhiyun #define MTHCA_CQ_STATUS_OVERFLOW ( 9 << 28)
84*4882a593Smuzhiyun #define MTHCA_CQ_STATUS_WRITE_FAIL (10 << 28)
85*4882a593Smuzhiyun #define MTHCA_CQ_FLAG_TR ( 1 << 18)
86*4882a593Smuzhiyun #define MTHCA_CQ_FLAG_OI ( 1 << 17)
87*4882a593Smuzhiyun #define MTHCA_CQ_STATE_DISARMED ( 0 << 8)
88*4882a593Smuzhiyun #define MTHCA_CQ_STATE_ARMED ( 1 << 8)
89*4882a593Smuzhiyun #define MTHCA_CQ_STATE_ARMED_SOL ( 4 << 8)
90*4882a593Smuzhiyun #define MTHCA_EQ_STATE_FIRED (10 << 8)
91*4882a593Smuzhiyun
92*4882a593Smuzhiyun enum {
93*4882a593Smuzhiyun MTHCA_ERROR_CQE_OPCODE_MASK = 0xfe
94*4882a593Smuzhiyun };
95*4882a593Smuzhiyun
96*4882a593Smuzhiyun enum {
97*4882a593Smuzhiyun SYNDROME_LOCAL_LENGTH_ERR = 0x01,
98*4882a593Smuzhiyun SYNDROME_LOCAL_QP_OP_ERR = 0x02,
99*4882a593Smuzhiyun SYNDROME_LOCAL_EEC_OP_ERR = 0x03,
100*4882a593Smuzhiyun SYNDROME_LOCAL_PROT_ERR = 0x04,
101*4882a593Smuzhiyun SYNDROME_WR_FLUSH_ERR = 0x05,
102*4882a593Smuzhiyun SYNDROME_MW_BIND_ERR = 0x06,
103*4882a593Smuzhiyun SYNDROME_BAD_RESP_ERR = 0x10,
104*4882a593Smuzhiyun SYNDROME_LOCAL_ACCESS_ERR = 0x11,
105*4882a593Smuzhiyun SYNDROME_REMOTE_INVAL_REQ_ERR = 0x12,
106*4882a593Smuzhiyun SYNDROME_REMOTE_ACCESS_ERR = 0x13,
107*4882a593Smuzhiyun SYNDROME_REMOTE_OP_ERR = 0x14,
108*4882a593Smuzhiyun SYNDROME_RETRY_EXC_ERR = 0x15,
109*4882a593Smuzhiyun SYNDROME_RNR_RETRY_EXC_ERR = 0x16,
110*4882a593Smuzhiyun SYNDROME_LOCAL_RDD_VIOL_ERR = 0x20,
111*4882a593Smuzhiyun SYNDROME_REMOTE_INVAL_RD_REQ_ERR = 0x21,
112*4882a593Smuzhiyun SYNDROME_REMOTE_ABORTED_ERR = 0x22,
113*4882a593Smuzhiyun SYNDROME_INVAL_EECN_ERR = 0x23,
114*4882a593Smuzhiyun SYNDROME_INVAL_EEC_STATE_ERR = 0x24
115*4882a593Smuzhiyun };
116*4882a593Smuzhiyun
117*4882a593Smuzhiyun struct mthca_cqe {
118*4882a593Smuzhiyun __be32 my_qpn;
119*4882a593Smuzhiyun __be32 my_ee;
120*4882a593Smuzhiyun __be32 rqpn;
121*4882a593Smuzhiyun u8 sl_ipok;
122*4882a593Smuzhiyun u8 g_mlpath;
123*4882a593Smuzhiyun __be16 rlid;
124*4882a593Smuzhiyun __be32 imm_etype_pkey_eec;
125*4882a593Smuzhiyun __be32 byte_cnt;
126*4882a593Smuzhiyun __be32 wqe;
127*4882a593Smuzhiyun u8 opcode;
128*4882a593Smuzhiyun u8 is_send;
129*4882a593Smuzhiyun u8 reserved;
130*4882a593Smuzhiyun u8 owner;
131*4882a593Smuzhiyun };
132*4882a593Smuzhiyun
133*4882a593Smuzhiyun struct mthca_err_cqe {
134*4882a593Smuzhiyun __be32 my_qpn;
135*4882a593Smuzhiyun u32 reserved1[3];
136*4882a593Smuzhiyun u8 syndrome;
137*4882a593Smuzhiyun u8 vendor_err;
138*4882a593Smuzhiyun __be16 db_cnt;
139*4882a593Smuzhiyun u32 reserved2;
140*4882a593Smuzhiyun __be32 wqe;
141*4882a593Smuzhiyun u8 opcode;
142*4882a593Smuzhiyun u8 reserved3[2];
143*4882a593Smuzhiyun u8 owner;
144*4882a593Smuzhiyun };
145*4882a593Smuzhiyun
146*4882a593Smuzhiyun #define MTHCA_CQ_ENTRY_OWNER_SW (0 << 7)
147*4882a593Smuzhiyun #define MTHCA_CQ_ENTRY_OWNER_HW (1 << 7)
148*4882a593Smuzhiyun
149*4882a593Smuzhiyun #define MTHCA_TAVOR_CQ_DB_INC_CI (1 << 24)
150*4882a593Smuzhiyun #define MTHCA_TAVOR_CQ_DB_REQ_NOT (2 << 24)
151*4882a593Smuzhiyun #define MTHCA_TAVOR_CQ_DB_REQ_NOT_SOL (3 << 24)
152*4882a593Smuzhiyun #define MTHCA_TAVOR_CQ_DB_SET_CI (4 << 24)
153*4882a593Smuzhiyun #define MTHCA_TAVOR_CQ_DB_REQ_NOT_MULT (5 << 24)
154*4882a593Smuzhiyun
155*4882a593Smuzhiyun #define MTHCA_ARBEL_CQ_DB_REQ_NOT_SOL (1 << 24)
156*4882a593Smuzhiyun #define MTHCA_ARBEL_CQ_DB_REQ_NOT (2 << 24)
157*4882a593Smuzhiyun #define MTHCA_ARBEL_CQ_DB_REQ_NOT_MULT (3 << 24)
158*4882a593Smuzhiyun
get_cqe_from_buf(struct mthca_cq_buf * buf,int entry)159*4882a593Smuzhiyun static inline struct mthca_cqe *get_cqe_from_buf(struct mthca_cq_buf *buf,
160*4882a593Smuzhiyun int entry)
161*4882a593Smuzhiyun {
162*4882a593Smuzhiyun if (buf->is_direct)
163*4882a593Smuzhiyun return buf->queue.direct.buf + (entry * MTHCA_CQ_ENTRY_SIZE);
164*4882a593Smuzhiyun else
165*4882a593Smuzhiyun return buf->queue.page_list[entry * MTHCA_CQ_ENTRY_SIZE / PAGE_SIZE].buf
166*4882a593Smuzhiyun + (entry * MTHCA_CQ_ENTRY_SIZE) % PAGE_SIZE;
167*4882a593Smuzhiyun }
168*4882a593Smuzhiyun
get_cqe(struct mthca_cq * cq,int entry)169*4882a593Smuzhiyun static inline struct mthca_cqe *get_cqe(struct mthca_cq *cq, int entry)
170*4882a593Smuzhiyun {
171*4882a593Smuzhiyun return get_cqe_from_buf(&cq->buf, entry);
172*4882a593Smuzhiyun }
173*4882a593Smuzhiyun
cqe_sw(struct mthca_cqe * cqe)174*4882a593Smuzhiyun static inline struct mthca_cqe *cqe_sw(struct mthca_cqe *cqe)
175*4882a593Smuzhiyun {
176*4882a593Smuzhiyun return MTHCA_CQ_ENTRY_OWNER_HW & cqe->owner ? NULL : cqe;
177*4882a593Smuzhiyun }
178*4882a593Smuzhiyun
next_cqe_sw(struct mthca_cq * cq)179*4882a593Smuzhiyun static inline struct mthca_cqe *next_cqe_sw(struct mthca_cq *cq)
180*4882a593Smuzhiyun {
181*4882a593Smuzhiyun return cqe_sw(get_cqe(cq, cq->cons_index & cq->ibcq.cqe));
182*4882a593Smuzhiyun }
183*4882a593Smuzhiyun
set_cqe_hw(struct mthca_cqe * cqe)184*4882a593Smuzhiyun static inline void set_cqe_hw(struct mthca_cqe *cqe)
185*4882a593Smuzhiyun {
186*4882a593Smuzhiyun cqe->owner = MTHCA_CQ_ENTRY_OWNER_HW;
187*4882a593Smuzhiyun }
188*4882a593Smuzhiyun
dump_cqe(struct mthca_dev * dev,void * cqe_ptr)189*4882a593Smuzhiyun static void dump_cqe(struct mthca_dev *dev, void *cqe_ptr)
190*4882a593Smuzhiyun {
191*4882a593Smuzhiyun __be32 *cqe = cqe_ptr;
192*4882a593Smuzhiyun
193*4882a593Smuzhiyun (void) cqe; /* avoid warning if mthca_dbg compiled away... */
194*4882a593Smuzhiyun mthca_dbg(dev, "CQE contents %08x %08x %08x %08x %08x %08x %08x %08x\n",
195*4882a593Smuzhiyun be32_to_cpu(cqe[0]), be32_to_cpu(cqe[1]), be32_to_cpu(cqe[2]),
196*4882a593Smuzhiyun be32_to_cpu(cqe[3]), be32_to_cpu(cqe[4]), be32_to_cpu(cqe[5]),
197*4882a593Smuzhiyun be32_to_cpu(cqe[6]), be32_to_cpu(cqe[7]));
198*4882a593Smuzhiyun }
199*4882a593Smuzhiyun
200*4882a593Smuzhiyun /*
201*4882a593Smuzhiyun * incr is ignored in native Arbel (mem-free) mode, so cq->cons_index
202*4882a593Smuzhiyun * should be correct before calling update_cons_index().
203*4882a593Smuzhiyun */
update_cons_index(struct mthca_dev * dev,struct mthca_cq * cq,int incr)204*4882a593Smuzhiyun static inline void update_cons_index(struct mthca_dev *dev, struct mthca_cq *cq,
205*4882a593Smuzhiyun int incr)
206*4882a593Smuzhiyun {
207*4882a593Smuzhiyun if (mthca_is_memfree(dev)) {
208*4882a593Smuzhiyun *cq->set_ci_db = cpu_to_be32(cq->cons_index);
209*4882a593Smuzhiyun wmb();
210*4882a593Smuzhiyun } else {
211*4882a593Smuzhiyun mthca_write64(MTHCA_TAVOR_CQ_DB_INC_CI | cq->cqn, incr - 1,
212*4882a593Smuzhiyun dev->kar + MTHCA_CQ_DOORBELL,
213*4882a593Smuzhiyun MTHCA_GET_DOORBELL_LOCK(&dev->doorbell_lock));
214*4882a593Smuzhiyun }
215*4882a593Smuzhiyun }
216*4882a593Smuzhiyun
mthca_cq_completion(struct mthca_dev * dev,u32 cqn)217*4882a593Smuzhiyun void mthca_cq_completion(struct mthca_dev *dev, u32 cqn)
218*4882a593Smuzhiyun {
219*4882a593Smuzhiyun struct mthca_cq *cq;
220*4882a593Smuzhiyun
221*4882a593Smuzhiyun cq = mthca_array_get(&dev->cq_table.cq, cqn & (dev->limits.num_cqs - 1));
222*4882a593Smuzhiyun
223*4882a593Smuzhiyun if (!cq) {
224*4882a593Smuzhiyun mthca_warn(dev, "Completion event for bogus CQ %08x\n", cqn);
225*4882a593Smuzhiyun return;
226*4882a593Smuzhiyun }
227*4882a593Smuzhiyun
228*4882a593Smuzhiyun ++cq->arm_sn;
229*4882a593Smuzhiyun
230*4882a593Smuzhiyun cq->ibcq.comp_handler(&cq->ibcq, cq->ibcq.cq_context);
231*4882a593Smuzhiyun }
232*4882a593Smuzhiyun
mthca_cq_event(struct mthca_dev * dev,u32 cqn,enum ib_event_type event_type)233*4882a593Smuzhiyun void mthca_cq_event(struct mthca_dev *dev, u32 cqn,
234*4882a593Smuzhiyun enum ib_event_type event_type)
235*4882a593Smuzhiyun {
236*4882a593Smuzhiyun struct mthca_cq *cq;
237*4882a593Smuzhiyun struct ib_event event;
238*4882a593Smuzhiyun
239*4882a593Smuzhiyun spin_lock(&dev->cq_table.lock);
240*4882a593Smuzhiyun
241*4882a593Smuzhiyun cq = mthca_array_get(&dev->cq_table.cq, cqn & (dev->limits.num_cqs - 1));
242*4882a593Smuzhiyun if (cq)
243*4882a593Smuzhiyun ++cq->refcount;
244*4882a593Smuzhiyun
245*4882a593Smuzhiyun spin_unlock(&dev->cq_table.lock);
246*4882a593Smuzhiyun
247*4882a593Smuzhiyun if (!cq) {
248*4882a593Smuzhiyun mthca_warn(dev, "Async event for bogus CQ %08x\n", cqn);
249*4882a593Smuzhiyun return;
250*4882a593Smuzhiyun }
251*4882a593Smuzhiyun
252*4882a593Smuzhiyun event.device = &dev->ib_dev;
253*4882a593Smuzhiyun event.event = event_type;
254*4882a593Smuzhiyun event.element.cq = &cq->ibcq;
255*4882a593Smuzhiyun if (cq->ibcq.event_handler)
256*4882a593Smuzhiyun cq->ibcq.event_handler(&event, cq->ibcq.cq_context);
257*4882a593Smuzhiyun
258*4882a593Smuzhiyun spin_lock(&dev->cq_table.lock);
259*4882a593Smuzhiyun if (!--cq->refcount)
260*4882a593Smuzhiyun wake_up(&cq->wait);
261*4882a593Smuzhiyun spin_unlock(&dev->cq_table.lock);
262*4882a593Smuzhiyun }
263*4882a593Smuzhiyun
is_recv_cqe(struct mthca_cqe * cqe)264*4882a593Smuzhiyun static inline int is_recv_cqe(struct mthca_cqe *cqe)
265*4882a593Smuzhiyun {
266*4882a593Smuzhiyun if ((cqe->opcode & MTHCA_ERROR_CQE_OPCODE_MASK) ==
267*4882a593Smuzhiyun MTHCA_ERROR_CQE_OPCODE_MASK)
268*4882a593Smuzhiyun return !(cqe->opcode & 0x01);
269*4882a593Smuzhiyun else
270*4882a593Smuzhiyun return !(cqe->is_send & 0x80);
271*4882a593Smuzhiyun }
272*4882a593Smuzhiyun
mthca_cq_clean(struct mthca_dev * dev,struct mthca_cq * cq,u32 qpn,struct mthca_srq * srq)273*4882a593Smuzhiyun void mthca_cq_clean(struct mthca_dev *dev, struct mthca_cq *cq, u32 qpn,
274*4882a593Smuzhiyun struct mthca_srq *srq)
275*4882a593Smuzhiyun {
276*4882a593Smuzhiyun struct mthca_cqe *cqe;
277*4882a593Smuzhiyun u32 prod_index;
278*4882a593Smuzhiyun int i, nfreed = 0;
279*4882a593Smuzhiyun
280*4882a593Smuzhiyun spin_lock_irq(&cq->lock);
281*4882a593Smuzhiyun
282*4882a593Smuzhiyun /*
283*4882a593Smuzhiyun * First we need to find the current producer index, so we
284*4882a593Smuzhiyun * know where to start cleaning from. It doesn't matter if HW
285*4882a593Smuzhiyun * adds new entries after this loop -- the QP we're worried
286*4882a593Smuzhiyun * about is already in RESET, so the new entries won't come
287*4882a593Smuzhiyun * from our QP and therefore don't need to be checked.
288*4882a593Smuzhiyun */
289*4882a593Smuzhiyun for (prod_index = cq->cons_index;
290*4882a593Smuzhiyun cqe_sw(get_cqe(cq, prod_index & cq->ibcq.cqe));
291*4882a593Smuzhiyun ++prod_index)
292*4882a593Smuzhiyun if (prod_index == cq->cons_index + cq->ibcq.cqe)
293*4882a593Smuzhiyun break;
294*4882a593Smuzhiyun
295*4882a593Smuzhiyun if (0)
296*4882a593Smuzhiyun mthca_dbg(dev, "Cleaning QPN %06x from CQN %06x; ci %d, pi %d\n",
297*4882a593Smuzhiyun qpn, cq->cqn, cq->cons_index, prod_index);
298*4882a593Smuzhiyun
299*4882a593Smuzhiyun /*
300*4882a593Smuzhiyun * Now sweep backwards through the CQ, removing CQ entries
301*4882a593Smuzhiyun * that match our QP by copying older entries on top of them.
302*4882a593Smuzhiyun */
303*4882a593Smuzhiyun while ((int) --prod_index - (int) cq->cons_index >= 0) {
304*4882a593Smuzhiyun cqe = get_cqe(cq, prod_index & cq->ibcq.cqe);
305*4882a593Smuzhiyun if (cqe->my_qpn == cpu_to_be32(qpn)) {
306*4882a593Smuzhiyun if (srq && is_recv_cqe(cqe))
307*4882a593Smuzhiyun mthca_free_srq_wqe(srq, be32_to_cpu(cqe->wqe));
308*4882a593Smuzhiyun ++nfreed;
309*4882a593Smuzhiyun } else if (nfreed)
310*4882a593Smuzhiyun memcpy(get_cqe(cq, (prod_index + nfreed) & cq->ibcq.cqe),
311*4882a593Smuzhiyun cqe, MTHCA_CQ_ENTRY_SIZE);
312*4882a593Smuzhiyun }
313*4882a593Smuzhiyun
314*4882a593Smuzhiyun if (nfreed) {
315*4882a593Smuzhiyun for (i = 0; i < nfreed; ++i)
316*4882a593Smuzhiyun set_cqe_hw(get_cqe(cq, (cq->cons_index + i) & cq->ibcq.cqe));
317*4882a593Smuzhiyun wmb();
318*4882a593Smuzhiyun cq->cons_index += nfreed;
319*4882a593Smuzhiyun update_cons_index(dev, cq, nfreed);
320*4882a593Smuzhiyun }
321*4882a593Smuzhiyun
322*4882a593Smuzhiyun spin_unlock_irq(&cq->lock);
323*4882a593Smuzhiyun }
324*4882a593Smuzhiyun
mthca_cq_resize_copy_cqes(struct mthca_cq * cq)325*4882a593Smuzhiyun void mthca_cq_resize_copy_cqes(struct mthca_cq *cq)
326*4882a593Smuzhiyun {
327*4882a593Smuzhiyun int i;
328*4882a593Smuzhiyun
329*4882a593Smuzhiyun /*
330*4882a593Smuzhiyun * In Tavor mode, the hardware keeps the consumer and producer
331*4882a593Smuzhiyun * indices mod the CQ size. Since we might be making the CQ
332*4882a593Smuzhiyun * bigger, we need to deal with the case where the producer
333*4882a593Smuzhiyun * index wrapped around before the CQ was resized.
334*4882a593Smuzhiyun */
335*4882a593Smuzhiyun if (!mthca_is_memfree(to_mdev(cq->ibcq.device)) &&
336*4882a593Smuzhiyun cq->ibcq.cqe < cq->resize_buf->cqe) {
337*4882a593Smuzhiyun cq->cons_index &= cq->ibcq.cqe;
338*4882a593Smuzhiyun if (cqe_sw(get_cqe(cq, cq->ibcq.cqe)))
339*4882a593Smuzhiyun cq->cons_index -= cq->ibcq.cqe + 1;
340*4882a593Smuzhiyun }
341*4882a593Smuzhiyun
342*4882a593Smuzhiyun for (i = cq->cons_index; cqe_sw(get_cqe(cq, i & cq->ibcq.cqe)); ++i)
343*4882a593Smuzhiyun memcpy(get_cqe_from_buf(&cq->resize_buf->buf,
344*4882a593Smuzhiyun i & cq->resize_buf->cqe),
345*4882a593Smuzhiyun get_cqe(cq, i & cq->ibcq.cqe), MTHCA_CQ_ENTRY_SIZE);
346*4882a593Smuzhiyun }
347*4882a593Smuzhiyun
mthca_alloc_cq_buf(struct mthca_dev * dev,struct mthca_cq_buf * buf,int nent)348*4882a593Smuzhiyun int mthca_alloc_cq_buf(struct mthca_dev *dev, struct mthca_cq_buf *buf, int nent)
349*4882a593Smuzhiyun {
350*4882a593Smuzhiyun int ret;
351*4882a593Smuzhiyun int i;
352*4882a593Smuzhiyun
353*4882a593Smuzhiyun ret = mthca_buf_alloc(dev, nent * MTHCA_CQ_ENTRY_SIZE,
354*4882a593Smuzhiyun MTHCA_MAX_DIRECT_CQ_SIZE,
355*4882a593Smuzhiyun &buf->queue, &buf->is_direct,
356*4882a593Smuzhiyun &dev->driver_pd, 1, &buf->mr);
357*4882a593Smuzhiyun if (ret)
358*4882a593Smuzhiyun return ret;
359*4882a593Smuzhiyun
360*4882a593Smuzhiyun for (i = 0; i < nent; ++i)
361*4882a593Smuzhiyun set_cqe_hw(get_cqe_from_buf(buf, i));
362*4882a593Smuzhiyun
363*4882a593Smuzhiyun return 0;
364*4882a593Smuzhiyun }
365*4882a593Smuzhiyun
mthca_free_cq_buf(struct mthca_dev * dev,struct mthca_cq_buf * buf,int cqe)366*4882a593Smuzhiyun void mthca_free_cq_buf(struct mthca_dev *dev, struct mthca_cq_buf *buf, int cqe)
367*4882a593Smuzhiyun {
368*4882a593Smuzhiyun mthca_buf_free(dev, (cqe + 1) * MTHCA_CQ_ENTRY_SIZE, &buf->queue,
369*4882a593Smuzhiyun buf->is_direct, &buf->mr);
370*4882a593Smuzhiyun }
371*4882a593Smuzhiyun
handle_error_cqe(struct mthca_dev * dev,struct mthca_cq * cq,struct mthca_qp * qp,int wqe_index,int is_send,struct mthca_err_cqe * cqe,struct ib_wc * entry,int * free_cqe)372*4882a593Smuzhiyun static void handle_error_cqe(struct mthca_dev *dev, struct mthca_cq *cq,
373*4882a593Smuzhiyun struct mthca_qp *qp, int wqe_index, int is_send,
374*4882a593Smuzhiyun struct mthca_err_cqe *cqe,
375*4882a593Smuzhiyun struct ib_wc *entry, int *free_cqe)
376*4882a593Smuzhiyun {
377*4882a593Smuzhiyun int dbd;
378*4882a593Smuzhiyun __be32 new_wqe;
379*4882a593Smuzhiyun
380*4882a593Smuzhiyun if (cqe->syndrome == SYNDROME_LOCAL_QP_OP_ERR) {
381*4882a593Smuzhiyun mthca_dbg(dev, "local QP operation err "
382*4882a593Smuzhiyun "(QPN %06x, WQE @ %08x, CQN %06x, index %d)\n",
383*4882a593Smuzhiyun be32_to_cpu(cqe->my_qpn), be32_to_cpu(cqe->wqe),
384*4882a593Smuzhiyun cq->cqn, cq->cons_index);
385*4882a593Smuzhiyun dump_cqe(dev, cqe);
386*4882a593Smuzhiyun }
387*4882a593Smuzhiyun
388*4882a593Smuzhiyun /*
389*4882a593Smuzhiyun * For completions in error, only work request ID, status, vendor error
390*4882a593Smuzhiyun * (and freed resource count for RD) have to be set.
391*4882a593Smuzhiyun */
392*4882a593Smuzhiyun switch (cqe->syndrome) {
393*4882a593Smuzhiyun case SYNDROME_LOCAL_LENGTH_ERR:
394*4882a593Smuzhiyun entry->status = IB_WC_LOC_LEN_ERR;
395*4882a593Smuzhiyun break;
396*4882a593Smuzhiyun case SYNDROME_LOCAL_QP_OP_ERR:
397*4882a593Smuzhiyun entry->status = IB_WC_LOC_QP_OP_ERR;
398*4882a593Smuzhiyun break;
399*4882a593Smuzhiyun case SYNDROME_LOCAL_EEC_OP_ERR:
400*4882a593Smuzhiyun entry->status = IB_WC_LOC_EEC_OP_ERR;
401*4882a593Smuzhiyun break;
402*4882a593Smuzhiyun case SYNDROME_LOCAL_PROT_ERR:
403*4882a593Smuzhiyun entry->status = IB_WC_LOC_PROT_ERR;
404*4882a593Smuzhiyun break;
405*4882a593Smuzhiyun case SYNDROME_WR_FLUSH_ERR:
406*4882a593Smuzhiyun entry->status = IB_WC_WR_FLUSH_ERR;
407*4882a593Smuzhiyun break;
408*4882a593Smuzhiyun case SYNDROME_MW_BIND_ERR:
409*4882a593Smuzhiyun entry->status = IB_WC_MW_BIND_ERR;
410*4882a593Smuzhiyun break;
411*4882a593Smuzhiyun case SYNDROME_BAD_RESP_ERR:
412*4882a593Smuzhiyun entry->status = IB_WC_BAD_RESP_ERR;
413*4882a593Smuzhiyun break;
414*4882a593Smuzhiyun case SYNDROME_LOCAL_ACCESS_ERR:
415*4882a593Smuzhiyun entry->status = IB_WC_LOC_ACCESS_ERR;
416*4882a593Smuzhiyun break;
417*4882a593Smuzhiyun case SYNDROME_REMOTE_INVAL_REQ_ERR:
418*4882a593Smuzhiyun entry->status = IB_WC_REM_INV_REQ_ERR;
419*4882a593Smuzhiyun break;
420*4882a593Smuzhiyun case SYNDROME_REMOTE_ACCESS_ERR:
421*4882a593Smuzhiyun entry->status = IB_WC_REM_ACCESS_ERR;
422*4882a593Smuzhiyun break;
423*4882a593Smuzhiyun case SYNDROME_REMOTE_OP_ERR:
424*4882a593Smuzhiyun entry->status = IB_WC_REM_OP_ERR;
425*4882a593Smuzhiyun break;
426*4882a593Smuzhiyun case SYNDROME_RETRY_EXC_ERR:
427*4882a593Smuzhiyun entry->status = IB_WC_RETRY_EXC_ERR;
428*4882a593Smuzhiyun break;
429*4882a593Smuzhiyun case SYNDROME_RNR_RETRY_EXC_ERR:
430*4882a593Smuzhiyun entry->status = IB_WC_RNR_RETRY_EXC_ERR;
431*4882a593Smuzhiyun break;
432*4882a593Smuzhiyun case SYNDROME_LOCAL_RDD_VIOL_ERR:
433*4882a593Smuzhiyun entry->status = IB_WC_LOC_RDD_VIOL_ERR;
434*4882a593Smuzhiyun break;
435*4882a593Smuzhiyun case SYNDROME_REMOTE_INVAL_RD_REQ_ERR:
436*4882a593Smuzhiyun entry->status = IB_WC_REM_INV_RD_REQ_ERR;
437*4882a593Smuzhiyun break;
438*4882a593Smuzhiyun case SYNDROME_REMOTE_ABORTED_ERR:
439*4882a593Smuzhiyun entry->status = IB_WC_REM_ABORT_ERR;
440*4882a593Smuzhiyun break;
441*4882a593Smuzhiyun case SYNDROME_INVAL_EECN_ERR:
442*4882a593Smuzhiyun entry->status = IB_WC_INV_EECN_ERR;
443*4882a593Smuzhiyun break;
444*4882a593Smuzhiyun case SYNDROME_INVAL_EEC_STATE_ERR:
445*4882a593Smuzhiyun entry->status = IB_WC_INV_EEC_STATE_ERR;
446*4882a593Smuzhiyun break;
447*4882a593Smuzhiyun default:
448*4882a593Smuzhiyun entry->status = IB_WC_GENERAL_ERR;
449*4882a593Smuzhiyun break;
450*4882a593Smuzhiyun }
451*4882a593Smuzhiyun
452*4882a593Smuzhiyun entry->vendor_err = cqe->vendor_err;
453*4882a593Smuzhiyun
454*4882a593Smuzhiyun /*
455*4882a593Smuzhiyun * Mem-free HCAs always generate one CQE per WQE, even in the
456*4882a593Smuzhiyun * error case, so we don't have to check the doorbell count, etc.
457*4882a593Smuzhiyun */
458*4882a593Smuzhiyun if (mthca_is_memfree(dev))
459*4882a593Smuzhiyun return;
460*4882a593Smuzhiyun
461*4882a593Smuzhiyun mthca_free_err_wqe(dev, qp, is_send, wqe_index, &dbd, &new_wqe);
462*4882a593Smuzhiyun
463*4882a593Smuzhiyun /*
464*4882a593Smuzhiyun * If we're at the end of the WQE chain, or we've used up our
465*4882a593Smuzhiyun * doorbell count, free the CQE. Otherwise just update it for
466*4882a593Smuzhiyun * the next poll operation.
467*4882a593Smuzhiyun */
468*4882a593Smuzhiyun if (!(new_wqe & cpu_to_be32(0x3f)) || (!cqe->db_cnt && dbd))
469*4882a593Smuzhiyun return;
470*4882a593Smuzhiyun
471*4882a593Smuzhiyun be16_add_cpu(&cqe->db_cnt, -dbd);
472*4882a593Smuzhiyun cqe->wqe = new_wqe;
473*4882a593Smuzhiyun cqe->syndrome = SYNDROME_WR_FLUSH_ERR;
474*4882a593Smuzhiyun
475*4882a593Smuzhiyun *free_cqe = 0;
476*4882a593Smuzhiyun }
477*4882a593Smuzhiyun
mthca_poll_one(struct mthca_dev * dev,struct mthca_cq * cq,struct mthca_qp ** cur_qp,int * freed,struct ib_wc * entry)478*4882a593Smuzhiyun static inline int mthca_poll_one(struct mthca_dev *dev,
479*4882a593Smuzhiyun struct mthca_cq *cq,
480*4882a593Smuzhiyun struct mthca_qp **cur_qp,
481*4882a593Smuzhiyun int *freed,
482*4882a593Smuzhiyun struct ib_wc *entry)
483*4882a593Smuzhiyun {
484*4882a593Smuzhiyun struct mthca_wq *wq;
485*4882a593Smuzhiyun struct mthca_cqe *cqe;
486*4882a593Smuzhiyun int wqe_index;
487*4882a593Smuzhiyun int is_error;
488*4882a593Smuzhiyun int is_send;
489*4882a593Smuzhiyun int free_cqe = 1;
490*4882a593Smuzhiyun int err = 0;
491*4882a593Smuzhiyun u16 checksum;
492*4882a593Smuzhiyun
493*4882a593Smuzhiyun cqe = next_cqe_sw(cq);
494*4882a593Smuzhiyun if (!cqe)
495*4882a593Smuzhiyun return -EAGAIN;
496*4882a593Smuzhiyun
497*4882a593Smuzhiyun /*
498*4882a593Smuzhiyun * Make sure we read CQ entry contents after we've checked the
499*4882a593Smuzhiyun * ownership bit.
500*4882a593Smuzhiyun */
501*4882a593Smuzhiyun rmb();
502*4882a593Smuzhiyun
503*4882a593Smuzhiyun if (0) {
504*4882a593Smuzhiyun mthca_dbg(dev, "%x/%d: CQE -> QPN %06x, WQE @ %08x\n",
505*4882a593Smuzhiyun cq->cqn, cq->cons_index, be32_to_cpu(cqe->my_qpn),
506*4882a593Smuzhiyun be32_to_cpu(cqe->wqe));
507*4882a593Smuzhiyun dump_cqe(dev, cqe);
508*4882a593Smuzhiyun }
509*4882a593Smuzhiyun
510*4882a593Smuzhiyun is_error = (cqe->opcode & MTHCA_ERROR_CQE_OPCODE_MASK) ==
511*4882a593Smuzhiyun MTHCA_ERROR_CQE_OPCODE_MASK;
512*4882a593Smuzhiyun is_send = is_error ? cqe->opcode & 0x01 : cqe->is_send & 0x80;
513*4882a593Smuzhiyun
514*4882a593Smuzhiyun if (!*cur_qp || be32_to_cpu(cqe->my_qpn) != (*cur_qp)->qpn) {
515*4882a593Smuzhiyun /*
516*4882a593Smuzhiyun * We do not have to take the QP table lock here,
517*4882a593Smuzhiyun * because CQs will be locked while QPs are removed
518*4882a593Smuzhiyun * from the table.
519*4882a593Smuzhiyun */
520*4882a593Smuzhiyun *cur_qp = mthca_array_get(&dev->qp_table.qp,
521*4882a593Smuzhiyun be32_to_cpu(cqe->my_qpn) &
522*4882a593Smuzhiyun (dev->limits.num_qps - 1));
523*4882a593Smuzhiyun if (!*cur_qp) {
524*4882a593Smuzhiyun mthca_warn(dev, "CQ entry for unknown QP %06x\n",
525*4882a593Smuzhiyun be32_to_cpu(cqe->my_qpn) & 0xffffff);
526*4882a593Smuzhiyun err = -EINVAL;
527*4882a593Smuzhiyun goto out;
528*4882a593Smuzhiyun }
529*4882a593Smuzhiyun }
530*4882a593Smuzhiyun
531*4882a593Smuzhiyun entry->qp = &(*cur_qp)->ibqp;
532*4882a593Smuzhiyun
533*4882a593Smuzhiyun if (is_send) {
534*4882a593Smuzhiyun wq = &(*cur_qp)->sq;
535*4882a593Smuzhiyun wqe_index = ((be32_to_cpu(cqe->wqe) - (*cur_qp)->send_wqe_offset)
536*4882a593Smuzhiyun >> wq->wqe_shift);
537*4882a593Smuzhiyun entry->wr_id = (*cur_qp)->wrid[wqe_index +
538*4882a593Smuzhiyun (*cur_qp)->rq.max];
539*4882a593Smuzhiyun } else if ((*cur_qp)->ibqp.srq) {
540*4882a593Smuzhiyun struct mthca_srq *srq = to_msrq((*cur_qp)->ibqp.srq);
541*4882a593Smuzhiyun u32 wqe = be32_to_cpu(cqe->wqe);
542*4882a593Smuzhiyun wq = NULL;
543*4882a593Smuzhiyun wqe_index = wqe >> srq->wqe_shift;
544*4882a593Smuzhiyun entry->wr_id = srq->wrid[wqe_index];
545*4882a593Smuzhiyun mthca_free_srq_wqe(srq, wqe);
546*4882a593Smuzhiyun } else {
547*4882a593Smuzhiyun s32 wqe;
548*4882a593Smuzhiyun wq = &(*cur_qp)->rq;
549*4882a593Smuzhiyun wqe = be32_to_cpu(cqe->wqe);
550*4882a593Smuzhiyun wqe_index = wqe >> wq->wqe_shift;
551*4882a593Smuzhiyun /*
552*4882a593Smuzhiyun * WQE addr == base - 1 might be reported in receive completion
553*4882a593Smuzhiyun * with error instead of (rq size - 1) by Sinai FW 1.0.800 and
554*4882a593Smuzhiyun * Arbel FW 5.1.400. This bug should be fixed in later FW revs.
555*4882a593Smuzhiyun */
556*4882a593Smuzhiyun if (unlikely(wqe_index < 0))
557*4882a593Smuzhiyun wqe_index = wq->max - 1;
558*4882a593Smuzhiyun entry->wr_id = (*cur_qp)->wrid[wqe_index];
559*4882a593Smuzhiyun }
560*4882a593Smuzhiyun
561*4882a593Smuzhiyun if (wq) {
562*4882a593Smuzhiyun if (wq->last_comp < wqe_index)
563*4882a593Smuzhiyun wq->tail += wqe_index - wq->last_comp;
564*4882a593Smuzhiyun else
565*4882a593Smuzhiyun wq->tail += wqe_index + wq->max - wq->last_comp;
566*4882a593Smuzhiyun
567*4882a593Smuzhiyun wq->last_comp = wqe_index;
568*4882a593Smuzhiyun }
569*4882a593Smuzhiyun
570*4882a593Smuzhiyun if (is_error) {
571*4882a593Smuzhiyun handle_error_cqe(dev, cq, *cur_qp, wqe_index, is_send,
572*4882a593Smuzhiyun (struct mthca_err_cqe *) cqe,
573*4882a593Smuzhiyun entry, &free_cqe);
574*4882a593Smuzhiyun goto out;
575*4882a593Smuzhiyun }
576*4882a593Smuzhiyun
577*4882a593Smuzhiyun if (is_send) {
578*4882a593Smuzhiyun entry->wc_flags = 0;
579*4882a593Smuzhiyun switch (cqe->opcode) {
580*4882a593Smuzhiyun case MTHCA_OPCODE_RDMA_WRITE:
581*4882a593Smuzhiyun entry->opcode = IB_WC_RDMA_WRITE;
582*4882a593Smuzhiyun break;
583*4882a593Smuzhiyun case MTHCA_OPCODE_RDMA_WRITE_IMM:
584*4882a593Smuzhiyun entry->opcode = IB_WC_RDMA_WRITE;
585*4882a593Smuzhiyun entry->wc_flags |= IB_WC_WITH_IMM;
586*4882a593Smuzhiyun break;
587*4882a593Smuzhiyun case MTHCA_OPCODE_SEND:
588*4882a593Smuzhiyun entry->opcode = IB_WC_SEND;
589*4882a593Smuzhiyun break;
590*4882a593Smuzhiyun case MTHCA_OPCODE_SEND_IMM:
591*4882a593Smuzhiyun entry->opcode = IB_WC_SEND;
592*4882a593Smuzhiyun entry->wc_flags |= IB_WC_WITH_IMM;
593*4882a593Smuzhiyun break;
594*4882a593Smuzhiyun case MTHCA_OPCODE_RDMA_READ:
595*4882a593Smuzhiyun entry->opcode = IB_WC_RDMA_READ;
596*4882a593Smuzhiyun entry->byte_len = be32_to_cpu(cqe->byte_cnt);
597*4882a593Smuzhiyun break;
598*4882a593Smuzhiyun case MTHCA_OPCODE_ATOMIC_CS:
599*4882a593Smuzhiyun entry->opcode = IB_WC_COMP_SWAP;
600*4882a593Smuzhiyun entry->byte_len = MTHCA_ATOMIC_BYTE_LEN;
601*4882a593Smuzhiyun break;
602*4882a593Smuzhiyun case MTHCA_OPCODE_ATOMIC_FA:
603*4882a593Smuzhiyun entry->opcode = IB_WC_FETCH_ADD;
604*4882a593Smuzhiyun entry->byte_len = MTHCA_ATOMIC_BYTE_LEN;
605*4882a593Smuzhiyun break;
606*4882a593Smuzhiyun default:
607*4882a593Smuzhiyun entry->opcode = 0xFF;
608*4882a593Smuzhiyun break;
609*4882a593Smuzhiyun }
610*4882a593Smuzhiyun } else {
611*4882a593Smuzhiyun entry->byte_len = be32_to_cpu(cqe->byte_cnt);
612*4882a593Smuzhiyun switch (cqe->opcode & 0x1f) {
613*4882a593Smuzhiyun case IB_OPCODE_SEND_LAST_WITH_IMMEDIATE:
614*4882a593Smuzhiyun case IB_OPCODE_SEND_ONLY_WITH_IMMEDIATE:
615*4882a593Smuzhiyun entry->wc_flags = IB_WC_WITH_IMM;
616*4882a593Smuzhiyun entry->ex.imm_data = cqe->imm_etype_pkey_eec;
617*4882a593Smuzhiyun entry->opcode = IB_WC_RECV;
618*4882a593Smuzhiyun break;
619*4882a593Smuzhiyun case IB_OPCODE_RDMA_WRITE_LAST_WITH_IMMEDIATE:
620*4882a593Smuzhiyun case IB_OPCODE_RDMA_WRITE_ONLY_WITH_IMMEDIATE:
621*4882a593Smuzhiyun entry->wc_flags = IB_WC_WITH_IMM;
622*4882a593Smuzhiyun entry->ex.imm_data = cqe->imm_etype_pkey_eec;
623*4882a593Smuzhiyun entry->opcode = IB_WC_RECV_RDMA_WITH_IMM;
624*4882a593Smuzhiyun break;
625*4882a593Smuzhiyun default:
626*4882a593Smuzhiyun entry->wc_flags = 0;
627*4882a593Smuzhiyun entry->opcode = IB_WC_RECV;
628*4882a593Smuzhiyun break;
629*4882a593Smuzhiyun }
630*4882a593Smuzhiyun entry->slid = be16_to_cpu(cqe->rlid);
631*4882a593Smuzhiyun entry->sl = cqe->sl_ipok >> 4;
632*4882a593Smuzhiyun entry->src_qp = be32_to_cpu(cqe->rqpn) & 0xffffff;
633*4882a593Smuzhiyun entry->dlid_path_bits = cqe->g_mlpath & 0x7f;
634*4882a593Smuzhiyun entry->pkey_index = be32_to_cpu(cqe->imm_etype_pkey_eec) >> 16;
635*4882a593Smuzhiyun entry->wc_flags |= cqe->g_mlpath & 0x80 ? IB_WC_GRH : 0;
636*4882a593Smuzhiyun checksum = (be32_to_cpu(cqe->rqpn) >> 24) |
637*4882a593Smuzhiyun ((be32_to_cpu(cqe->my_ee) >> 16) & 0xff00);
638*4882a593Smuzhiyun entry->wc_flags |= (cqe->sl_ipok & 1 && checksum == 0xffff) ?
639*4882a593Smuzhiyun IB_WC_IP_CSUM_OK : 0;
640*4882a593Smuzhiyun }
641*4882a593Smuzhiyun
642*4882a593Smuzhiyun entry->status = IB_WC_SUCCESS;
643*4882a593Smuzhiyun
644*4882a593Smuzhiyun out:
645*4882a593Smuzhiyun if (likely(free_cqe)) {
646*4882a593Smuzhiyun set_cqe_hw(cqe);
647*4882a593Smuzhiyun ++(*freed);
648*4882a593Smuzhiyun ++cq->cons_index;
649*4882a593Smuzhiyun }
650*4882a593Smuzhiyun
651*4882a593Smuzhiyun return err;
652*4882a593Smuzhiyun }
653*4882a593Smuzhiyun
mthca_poll_cq(struct ib_cq * ibcq,int num_entries,struct ib_wc * entry)654*4882a593Smuzhiyun int mthca_poll_cq(struct ib_cq *ibcq, int num_entries,
655*4882a593Smuzhiyun struct ib_wc *entry)
656*4882a593Smuzhiyun {
657*4882a593Smuzhiyun struct mthca_dev *dev = to_mdev(ibcq->device);
658*4882a593Smuzhiyun struct mthca_cq *cq = to_mcq(ibcq);
659*4882a593Smuzhiyun struct mthca_qp *qp = NULL;
660*4882a593Smuzhiyun unsigned long flags;
661*4882a593Smuzhiyun int err = 0;
662*4882a593Smuzhiyun int freed = 0;
663*4882a593Smuzhiyun int npolled;
664*4882a593Smuzhiyun
665*4882a593Smuzhiyun spin_lock_irqsave(&cq->lock, flags);
666*4882a593Smuzhiyun
667*4882a593Smuzhiyun npolled = 0;
668*4882a593Smuzhiyun repoll:
669*4882a593Smuzhiyun while (npolled < num_entries) {
670*4882a593Smuzhiyun err = mthca_poll_one(dev, cq, &qp,
671*4882a593Smuzhiyun &freed, entry + npolled);
672*4882a593Smuzhiyun if (err)
673*4882a593Smuzhiyun break;
674*4882a593Smuzhiyun ++npolled;
675*4882a593Smuzhiyun }
676*4882a593Smuzhiyun
677*4882a593Smuzhiyun if (freed) {
678*4882a593Smuzhiyun wmb();
679*4882a593Smuzhiyun update_cons_index(dev, cq, freed);
680*4882a593Smuzhiyun }
681*4882a593Smuzhiyun
682*4882a593Smuzhiyun /*
683*4882a593Smuzhiyun * If a CQ resize is in progress and we discovered that the
684*4882a593Smuzhiyun * old buffer is empty, then peek in the new buffer, and if
685*4882a593Smuzhiyun * it's not empty, switch to the new buffer and continue
686*4882a593Smuzhiyun * polling there.
687*4882a593Smuzhiyun */
688*4882a593Smuzhiyun if (unlikely(err == -EAGAIN && cq->resize_buf &&
689*4882a593Smuzhiyun cq->resize_buf->state == CQ_RESIZE_READY)) {
690*4882a593Smuzhiyun /*
691*4882a593Smuzhiyun * In Tavor mode, the hardware keeps the producer
692*4882a593Smuzhiyun * index modulo the CQ size. Since we might be making
693*4882a593Smuzhiyun * the CQ bigger, we need to mask our consumer index
694*4882a593Smuzhiyun * using the size of the old CQ buffer before looking
695*4882a593Smuzhiyun * in the new CQ buffer.
696*4882a593Smuzhiyun */
697*4882a593Smuzhiyun if (!mthca_is_memfree(dev))
698*4882a593Smuzhiyun cq->cons_index &= cq->ibcq.cqe;
699*4882a593Smuzhiyun
700*4882a593Smuzhiyun if (cqe_sw(get_cqe_from_buf(&cq->resize_buf->buf,
701*4882a593Smuzhiyun cq->cons_index & cq->resize_buf->cqe))) {
702*4882a593Smuzhiyun struct mthca_cq_buf tbuf;
703*4882a593Smuzhiyun int tcqe;
704*4882a593Smuzhiyun
705*4882a593Smuzhiyun tbuf = cq->buf;
706*4882a593Smuzhiyun tcqe = cq->ibcq.cqe;
707*4882a593Smuzhiyun cq->buf = cq->resize_buf->buf;
708*4882a593Smuzhiyun cq->ibcq.cqe = cq->resize_buf->cqe;
709*4882a593Smuzhiyun
710*4882a593Smuzhiyun cq->resize_buf->buf = tbuf;
711*4882a593Smuzhiyun cq->resize_buf->cqe = tcqe;
712*4882a593Smuzhiyun cq->resize_buf->state = CQ_RESIZE_SWAPPED;
713*4882a593Smuzhiyun
714*4882a593Smuzhiyun goto repoll;
715*4882a593Smuzhiyun }
716*4882a593Smuzhiyun }
717*4882a593Smuzhiyun
718*4882a593Smuzhiyun spin_unlock_irqrestore(&cq->lock, flags);
719*4882a593Smuzhiyun
720*4882a593Smuzhiyun return err == 0 || err == -EAGAIN ? npolled : err;
721*4882a593Smuzhiyun }
722*4882a593Smuzhiyun
mthca_tavor_arm_cq(struct ib_cq * cq,enum ib_cq_notify_flags flags)723*4882a593Smuzhiyun int mthca_tavor_arm_cq(struct ib_cq *cq, enum ib_cq_notify_flags flags)
724*4882a593Smuzhiyun {
725*4882a593Smuzhiyun u32 dbhi = ((flags & IB_CQ_SOLICITED_MASK) == IB_CQ_SOLICITED ?
726*4882a593Smuzhiyun MTHCA_TAVOR_CQ_DB_REQ_NOT_SOL :
727*4882a593Smuzhiyun MTHCA_TAVOR_CQ_DB_REQ_NOT) |
728*4882a593Smuzhiyun to_mcq(cq)->cqn;
729*4882a593Smuzhiyun
730*4882a593Smuzhiyun mthca_write64(dbhi, 0xffffffff, to_mdev(cq->device)->kar + MTHCA_CQ_DOORBELL,
731*4882a593Smuzhiyun MTHCA_GET_DOORBELL_LOCK(&to_mdev(cq->device)->doorbell_lock));
732*4882a593Smuzhiyun
733*4882a593Smuzhiyun return 0;
734*4882a593Smuzhiyun }
735*4882a593Smuzhiyun
mthca_arbel_arm_cq(struct ib_cq * ibcq,enum ib_cq_notify_flags flags)736*4882a593Smuzhiyun int mthca_arbel_arm_cq(struct ib_cq *ibcq, enum ib_cq_notify_flags flags)
737*4882a593Smuzhiyun {
738*4882a593Smuzhiyun struct mthca_cq *cq = to_mcq(ibcq);
739*4882a593Smuzhiyun __be32 db_rec[2];
740*4882a593Smuzhiyun u32 dbhi;
741*4882a593Smuzhiyun u32 sn = cq->arm_sn & 3;
742*4882a593Smuzhiyun
743*4882a593Smuzhiyun db_rec[0] = cpu_to_be32(cq->cons_index);
744*4882a593Smuzhiyun db_rec[1] = cpu_to_be32((cq->cqn << 8) | (2 << 5) | (sn << 3) |
745*4882a593Smuzhiyun ((flags & IB_CQ_SOLICITED_MASK) ==
746*4882a593Smuzhiyun IB_CQ_SOLICITED ? 1 : 2));
747*4882a593Smuzhiyun
748*4882a593Smuzhiyun mthca_write_db_rec(db_rec, cq->arm_db);
749*4882a593Smuzhiyun
750*4882a593Smuzhiyun /*
751*4882a593Smuzhiyun * Make sure that the doorbell record in host memory is
752*4882a593Smuzhiyun * written before ringing the doorbell via PCI MMIO.
753*4882a593Smuzhiyun */
754*4882a593Smuzhiyun wmb();
755*4882a593Smuzhiyun
756*4882a593Smuzhiyun dbhi = (sn << 28) |
757*4882a593Smuzhiyun ((flags & IB_CQ_SOLICITED_MASK) == IB_CQ_SOLICITED ?
758*4882a593Smuzhiyun MTHCA_ARBEL_CQ_DB_REQ_NOT_SOL :
759*4882a593Smuzhiyun MTHCA_ARBEL_CQ_DB_REQ_NOT) | cq->cqn;
760*4882a593Smuzhiyun
761*4882a593Smuzhiyun mthca_write64(dbhi, cq->cons_index,
762*4882a593Smuzhiyun to_mdev(ibcq->device)->kar + MTHCA_CQ_DOORBELL,
763*4882a593Smuzhiyun MTHCA_GET_DOORBELL_LOCK(&to_mdev(ibcq->device)->doorbell_lock));
764*4882a593Smuzhiyun
765*4882a593Smuzhiyun return 0;
766*4882a593Smuzhiyun }
767*4882a593Smuzhiyun
mthca_init_cq(struct mthca_dev * dev,int nent,struct mthca_ucontext * ctx,u32 pdn,struct mthca_cq * cq)768*4882a593Smuzhiyun int mthca_init_cq(struct mthca_dev *dev, int nent,
769*4882a593Smuzhiyun struct mthca_ucontext *ctx, u32 pdn,
770*4882a593Smuzhiyun struct mthca_cq *cq)
771*4882a593Smuzhiyun {
772*4882a593Smuzhiyun struct mthca_mailbox *mailbox;
773*4882a593Smuzhiyun struct mthca_cq_context *cq_context;
774*4882a593Smuzhiyun int err = -ENOMEM;
775*4882a593Smuzhiyun
776*4882a593Smuzhiyun cq->ibcq.cqe = nent - 1;
777*4882a593Smuzhiyun cq->is_kernel = !ctx;
778*4882a593Smuzhiyun
779*4882a593Smuzhiyun cq->cqn = mthca_alloc(&dev->cq_table.alloc);
780*4882a593Smuzhiyun if (cq->cqn == -1)
781*4882a593Smuzhiyun return -ENOMEM;
782*4882a593Smuzhiyun
783*4882a593Smuzhiyun if (mthca_is_memfree(dev)) {
784*4882a593Smuzhiyun err = mthca_table_get(dev, dev->cq_table.table, cq->cqn);
785*4882a593Smuzhiyun if (err)
786*4882a593Smuzhiyun goto err_out;
787*4882a593Smuzhiyun
788*4882a593Smuzhiyun if (cq->is_kernel) {
789*4882a593Smuzhiyun cq->arm_sn = 1;
790*4882a593Smuzhiyun
791*4882a593Smuzhiyun err = -ENOMEM;
792*4882a593Smuzhiyun
793*4882a593Smuzhiyun cq->set_ci_db_index = mthca_alloc_db(dev, MTHCA_DB_TYPE_CQ_SET_CI,
794*4882a593Smuzhiyun cq->cqn, &cq->set_ci_db);
795*4882a593Smuzhiyun if (cq->set_ci_db_index < 0)
796*4882a593Smuzhiyun goto err_out_icm;
797*4882a593Smuzhiyun
798*4882a593Smuzhiyun cq->arm_db_index = mthca_alloc_db(dev, MTHCA_DB_TYPE_CQ_ARM,
799*4882a593Smuzhiyun cq->cqn, &cq->arm_db);
800*4882a593Smuzhiyun if (cq->arm_db_index < 0)
801*4882a593Smuzhiyun goto err_out_ci;
802*4882a593Smuzhiyun }
803*4882a593Smuzhiyun }
804*4882a593Smuzhiyun
805*4882a593Smuzhiyun mailbox = mthca_alloc_mailbox(dev, GFP_KERNEL);
806*4882a593Smuzhiyun if (IS_ERR(mailbox)) {
807*4882a593Smuzhiyun err = PTR_ERR(mailbox);
808*4882a593Smuzhiyun goto err_out_arm;
809*4882a593Smuzhiyun }
810*4882a593Smuzhiyun
811*4882a593Smuzhiyun cq_context = mailbox->buf;
812*4882a593Smuzhiyun
813*4882a593Smuzhiyun if (cq->is_kernel) {
814*4882a593Smuzhiyun err = mthca_alloc_cq_buf(dev, &cq->buf, nent);
815*4882a593Smuzhiyun if (err)
816*4882a593Smuzhiyun goto err_out_mailbox;
817*4882a593Smuzhiyun }
818*4882a593Smuzhiyun
819*4882a593Smuzhiyun spin_lock_init(&cq->lock);
820*4882a593Smuzhiyun cq->refcount = 1;
821*4882a593Smuzhiyun init_waitqueue_head(&cq->wait);
822*4882a593Smuzhiyun mutex_init(&cq->mutex);
823*4882a593Smuzhiyun
824*4882a593Smuzhiyun memset(cq_context, 0, sizeof *cq_context);
825*4882a593Smuzhiyun cq_context->flags = cpu_to_be32(MTHCA_CQ_STATUS_OK |
826*4882a593Smuzhiyun MTHCA_CQ_STATE_DISARMED |
827*4882a593Smuzhiyun MTHCA_CQ_FLAG_TR);
828*4882a593Smuzhiyun cq_context->logsize_usrpage = cpu_to_be32((ffs(nent) - 1) << 24);
829*4882a593Smuzhiyun if (ctx)
830*4882a593Smuzhiyun cq_context->logsize_usrpage |= cpu_to_be32(ctx->uar.index);
831*4882a593Smuzhiyun else
832*4882a593Smuzhiyun cq_context->logsize_usrpage |= cpu_to_be32(dev->driver_uar.index);
833*4882a593Smuzhiyun cq_context->error_eqn = cpu_to_be32(dev->eq_table.eq[MTHCA_EQ_ASYNC].eqn);
834*4882a593Smuzhiyun cq_context->comp_eqn = cpu_to_be32(dev->eq_table.eq[MTHCA_EQ_COMP].eqn);
835*4882a593Smuzhiyun cq_context->pd = cpu_to_be32(pdn);
836*4882a593Smuzhiyun cq_context->lkey = cpu_to_be32(cq->buf.mr.ibmr.lkey);
837*4882a593Smuzhiyun cq_context->cqn = cpu_to_be32(cq->cqn);
838*4882a593Smuzhiyun
839*4882a593Smuzhiyun if (mthca_is_memfree(dev)) {
840*4882a593Smuzhiyun cq_context->ci_db = cpu_to_be32(cq->set_ci_db_index);
841*4882a593Smuzhiyun cq_context->state_db = cpu_to_be32(cq->arm_db_index);
842*4882a593Smuzhiyun }
843*4882a593Smuzhiyun
844*4882a593Smuzhiyun err = mthca_SW2HW_CQ(dev, mailbox, cq->cqn);
845*4882a593Smuzhiyun if (err) {
846*4882a593Smuzhiyun mthca_warn(dev, "SW2HW_CQ failed (%d)\n", err);
847*4882a593Smuzhiyun goto err_out_free_mr;
848*4882a593Smuzhiyun }
849*4882a593Smuzhiyun
850*4882a593Smuzhiyun spin_lock_irq(&dev->cq_table.lock);
851*4882a593Smuzhiyun err = mthca_array_set(&dev->cq_table.cq,
852*4882a593Smuzhiyun cq->cqn & (dev->limits.num_cqs - 1), cq);
853*4882a593Smuzhiyun if (err) {
854*4882a593Smuzhiyun spin_unlock_irq(&dev->cq_table.lock);
855*4882a593Smuzhiyun goto err_out_free_mr;
856*4882a593Smuzhiyun }
857*4882a593Smuzhiyun spin_unlock_irq(&dev->cq_table.lock);
858*4882a593Smuzhiyun
859*4882a593Smuzhiyun cq->cons_index = 0;
860*4882a593Smuzhiyun
861*4882a593Smuzhiyun mthca_free_mailbox(dev, mailbox);
862*4882a593Smuzhiyun
863*4882a593Smuzhiyun return 0;
864*4882a593Smuzhiyun
865*4882a593Smuzhiyun err_out_free_mr:
866*4882a593Smuzhiyun if (cq->is_kernel)
867*4882a593Smuzhiyun mthca_free_cq_buf(dev, &cq->buf, cq->ibcq.cqe);
868*4882a593Smuzhiyun
869*4882a593Smuzhiyun err_out_mailbox:
870*4882a593Smuzhiyun mthca_free_mailbox(dev, mailbox);
871*4882a593Smuzhiyun
872*4882a593Smuzhiyun err_out_arm:
873*4882a593Smuzhiyun if (cq->is_kernel && mthca_is_memfree(dev))
874*4882a593Smuzhiyun mthca_free_db(dev, MTHCA_DB_TYPE_CQ_ARM, cq->arm_db_index);
875*4882a593Smuzhiyun
876*4882a593Smuzhiyun err_out_ci:
877*4882a593Smuzhiyun if (cq->is_kernel && mthca_is_memfree(dev))
878*4882a593Smuzhiyun mthca_free_db(dev, MTHCA_DB_TYPE_CQ_SET_CI, cq->set_ci_db_index);
879*4882a593Smuzhiyun
880*4882a593Smuzhiyun err_out_icm:
881*4882a593Smuzhiyun mthca_table_put(dev, dev->cq_table.table, cq->cqn);
882*4882a593Smuzhiyun
883*4882a593Smuzhiyun err_out:
884*4882a593Smuzhiyun mthca_free(&dev->cq_table.alloc, cq->cqn);
885*4882a593Smuzhiyun
886*4882a593Smuzhiyun return err;
887*4882a593Smuzhiyun }
888*4882a593Smuzhiyun
get_cq_refcount(struct mthca_dev * dev,struct mthca_cq * cq)889*4882a593Smuzhiyun static inline int get_cq_refcount(struct mthca_dev *dev, struct mthca_cq *cq)
890*4882a593Smuzhiyun {
891*4882a593Smuzhiyun int c;
892*4882a593Smuzhiyun
893*4882a593Smuzhiyun spin_lock_irq(&dev->cq_table.lock);
894*4882a593Smuzhiyun c = cq->refcount;
895*4882a593Smuzhiyun spin_unlock_irq(&dev->cq_table.lock);
896*4882a593Smuzhiyun
897*4882a593Smuzhiyun return c;
898*4882a593Smuzhiyun }
899*4882a593Smuzhiyun
mthca_free_cq(struct mthca_dev * dev,struct mthca_cq * cq)900*4882a593Smuzhiyun void mthca_free_cq(struct mthca_dev *dev,
901*4882a593Smuzhiyun struct mthca_cq *cq)
902*4882a593Smuzhiyun {
903*4882a593Smuzhiyun struct mthca_mailbox *mailbox;
904*4882a593Smuzhiyun int err;
905*4882a593Smuzhiyun
906*4882a593Smuzhiyun mailbox = mthca_alloc_mailbox(dev, GFP_KERNEL);
907*4882a593Smuzhiyun if (IS_ERR(mailbox)) {
908*4882a593Smuzhiyun mthca_warn(dev, "No memory for mailbox to free CQ.\n");
909*4882a593Smuzhiyun return;
910*4882a593Smuzhiyun }
911*4882a593Smuzhiyun
912*4882a593Smuzhiyun err = mthca_HW2SW_CQ(dev, mailbox, cq->cqn);
913*4882a593Smuzhiyun if (err)
914*4882a593Smuzhiyun mthca_warn(dev, "HW2SW_CQ failed (%d)\n", err);
915*4882a593Smuzhiyun
916*4882a593Smuzhiyun if (0) {
917*4882a593Smuzhiyun __be32 *ctx = mailbox->buf;
918*4882a593Smuzhiyun int j;
919*4882a593Smuzhiyun
920*4882a593Smuzhiyun printk(KERN_ERR "context for CQN %x (cons index %x, next sw %d)\n",
921*4882a593Smuzhiyun cq->cqn, cq->cons_index,
922*4882a593Smuzhiyun cq->is_kernel ? !!next_cqe_sw(cq) : 0);
923*4882a593Smuzhiyun for (j = 0; j < 16; ++j)
924*4882a593Smuzhiyun printk(KERN_ERR "[%2x] %08x\n", j * 4, be32_to_cpu(ctx[j]));
925*4882a593Smuzhiyun }
926*4882a593Smuzhiyun
927*4882a593Smuzhiyun spin_lock_irq(&dev->cq_table.lock);
928*4882a593Smuzhiyun mthca_array_clear(&dev->cq_table.cq,
929*4882a593Smuzhiyun cq->cqn & (dev->limits.num_cqs - 1));
930*4882a593Smuzhiyun --cq->refcount;
931*4882a593Smuzhiyun spin_unlock_irq(&dev->cq_table.lock);
932*4882a593Smuzhiyun
933*4882a593Smuzhiyun if (dev->mthca_flags & MTHCA_FLAG_MSI_X)
934*4882a593Smuzhiyun synchronize_irq(dev->eq_table.eq[MTHCA_EQ_COMP].msi_x_vector);
935*4882a593Smuzhiyun else
936*4882a593Smuzhiyun synchronize_irq(dev->pdev->irq);
937*4882a593Smuzhiyun
938*4882a593Smuzhiyun wait_event(cq->wait, !get_cq_refcount(dev, cq));
939*4882a593Smuzhiyun
940*4882a593Smuzhiyun if (cq->is_kernel) {
941*4882a593Smuzhiyun mthca_free_cq_buf(dev, &cq->buf, cq->ibcq.cqe);
942*4882a593Smuzhiyun if (mthca_is_memfree(dev)) {
943*4882a593Smuzhiyun mthca_free_db(dev, MTHCA_DB_TYPE_CQ_ARM, cq->arm_db_index);
944*4882a593Smuzhiyun mthca_free_db(dev, MTHCA_DB_TYPE_CQ_SET_CI, cq->set_ci_db_index);
945*4882a593Smuzhiyun }
946*4882a593Smuzhiyun }
947*4882a593Smuzhiyun
948*4882a593Smuzhiyun mthca_table_put(dev, dev->cq_table.table, cq->cqn);
949*4882a593Smuzhiyun mthca_free(&dev->cq_table.alloc, cq->cqn);
950*4882a593Smuzhiyun mthca_free_mailbox(dev, mailbox);
951*4882a593Smuzhiyun }
952*4882a593Smuzhiyun
mthca_init_cq_table(struct mthca_dev * dev)953*4882a593Smuzhiyun int mthca_init_cq_table(struct mthca_dev *dev)
954*4882a593Smuzhiyun {
955*4882a593Smuzhiyun int err;
956*4882a593Smuzhiyun
957*4882a593Smuzhiyun spin_lock_init(&dev->cq_table.lock);
958*4882a593Smuzhiyun
959*4882a593Smuzhiyun err = mthca_alloc_init(&dev->cq_table.alloc,
960*4882a593Smuzhiyun dev->limits.num_cqs,
961*4882a593Smuzhiyun (1 << 24) - 1,
962*4882a593Smuzhiyun dev->limits.reserved_cqs);
963*4882a593Smuzhiyun if (err)
964*4882a593Smuzhiyun return err;
965*4882a593Smuzhiyun
966*4882a593Smuzhiyun err = mthca_array_init(&dev->cq_table.cq,
967*4882a593Smuzhiyun dev->limits.num_cqs);
968*4882a593Smuzhiyun if (err)
969*4882a593Smuzhiyun mthca_alloc_cleanup(&dev->cq_table.alloc);
970*4882a593Smuzhiyun
971*4882a593Smuzhiyun return err;
972*4882a593Smuzhiyun }
973*4882a593Smuzhiyun
mthca_cleanup_cq_table(struct mthca_dev * dev)974*4882a593Smuzhiyun void mthca_cleanup_cq_table(struct mthca_dev *dev)
975*4882a593Smuzhiyun {
976*4882a593Smuzhiyun mthca_array_cleanup(&dev->cq_table.cq, dev->limits.num_cqs);
977*4882a593Smuzhiyun mthca_alloc_cleanup(&dev->cq_table.alloc);
978*4882a593Smuzhiyun }
979