1*4882a593Smuzhiyun // SPDX-License-Identifier: (GPL-2.0 OR BSD-3-Clause)
2*4882a593Smuzhiyun
3*4882a593Smuzhiyun #include <linux/irq.h>
4*4882a593Smuzhiyun #include <linux/module.h>
5*4882a593Smuzhiyun #include <linux/ntb.h>
6*4882a593Smuzhiyun #include <linux/msi.h>
7*4882a593Smuzhiyun #include <linux/pci.h>
8*4882a593Smuzhiyun
9*4882a593Smuzhiyun struct ntb_msi {
10*4882a593Smuzhiyun u64 base_addr;
11*4882a593Smuzhiyun u64 end_addr;
12*4882a593Smuzhiyun
13*4882a593Smuzhiyun void (*desc_changed)(void *ctx);
14*4882a593Smuzhiyun
15*4882a593Smuzhiyun u32 __iomem *peer_mws[];
16*4882a593Smuzhiyun };
17*4882a593Smuzhiyun
18*4882a593Smuzhiyun /**
19*4882a593Smuzhiyun * ntb_msi_init() - Initialize the MSI context
20*4882a593Smuzhiyun * @ntb: NTB device context
21*4882a593Smuzhiyun *
22*4882a593Smuzhiyun * This function must be called before any other ntb_msi function.
23*4882a593Smuzhiyun * It initializes the context for MSI operations and maps
24*4882a593Smuzhiyun * the peer memory windows.
25*4882a593Smuzhiyun *
26*4882a593Smuzhiyun * This function reserves the last N outbound memory windows (where N
27*4882a593Smuzhiyun * is the number of peers).
28*4882a593Smuzhiyun *
29*4882a593Smuzhiyun * Return: Zero on success, otherwise a negative error number.
30*4882a593Smuzhiyun */
ntb_msi_init(struct ntb_dev * ntb,void (* desc_changed)(void * ctx))31*4882a593Smuzhiyun int ntb_msi_init(struct ntb_dev *ntb,
32*4882a593Smuzhiyun void (*desc_changed)(void *ctx))
33*4882a593Smuzhiyun {
34*4882a593Smuzhiyun phys_addr_t mw_phys_addr;
35*4882a593Smuzhiyun resource_size_t mw_size;
36*4882a593Smuzhiyun size_t struct_size;
37*4882a593Smuzhiyun int peer_widx;
38*4882a593Smuzhiyun int peers;
39*4882a593Smuzhiyun int ret;
40*4882a593Smuzhiyun int i;
41*4882a593Smuzhiyun
42*4882a593Smuzhiyun peers = ntb_peer_port_count(ntb);
43*4882a593Smuzhiyun if (peers <= 0)
44*4882a593Smuzhiyun return -EINVAL;
45*4882a593Smuzhiyun
46*4882a593Smuzhiyun struct_size = sizeof(*ntb->msi) + sizeof(*ntb->msi->peer_mws) * peers;
47*4882a593Smuzhiyun
48*4882a593Smuzhiyun ntb->msi = devm_kzalloc(&ntb->dev, struct_size, GFP_KERNEL);
49*4882a593Smuzhiyun if (!ntb->msi)
50*4882a593Smuzhiyun return -ENOMEM;
51*4882a593Smuzhiyun
52*4882a593Smuzhiyun ntb->msi->desc_changed = desc_changed;
53*4882a593Smuzhiyun
54*4882a593Smuzhiyun for (i = 0; i < peers; i++) {
55*4882a593Smuzhiyun peer_widx = ntb_peer_mw_count(ntb) - 1 - i;
56*4882a593Smuzhiyun
57*4882a593Smuzhiyun ret = ntb_peer_mw_get_addr(ntb, peer_widx, &mw_phys_addr,
58*4882a593Smuzhiyun &mw_size);
59*4882a593Smuzhiyun if (ret)
60*4882a593Smuzhiyun goto unroll;
61*4882a593Smuzhiyun
62*4882a593Smuzhiyun ntb->msi->peer_mws[i] = devm_ioremap(&ntb->dev, mw_phys_addr,
63*4882a593Smuzhiyun mw_size);
64*4882a593Smuzhiyun if (!ntb->msi->peer_mws[i]) {
65*4882a593Smuzhiyun ret = -EFAULT;
66*4882a593Smuzhiyun goto unroll;
67*4882a593Smuzhiyun }
68*4882a593Smuzhiyun }
69*4882a593Smuzhiyun
70*4882a593Smuzhiyun return 0;
71*4882a593Smuzhiyun
72*4882a593Smuzhiyun unroll:
73*4882a593Smuzhiyun for (i = 0; i < peers; i++)
74*4882a593Smuzhiyun if (ntb->msi->peer_mws[i])
75*4882a593Smuzhiyun devm_iounmap(&ntb->dev, ntb->msi->peer_mws[i]);
76*4882a593Smuzhiyun
77*4882a593Smuzhiyun devm_kfree(&ntb->dev, ntb->msi);
78*4882a593Smuzhiyun ntb->msi = NULL;
79*4882a593Smuzhiyun return ret;
80*4882a593Smuzhiyun }
81*4882a593Smuzhiyun EXPORT_SYMBOL(ntb_msi_init);
82*4882a593Smuzhiyun
83*4882a593Smuzhiyun /**
84*4882a593Smuzhiyun * ntb_msi_setup_mws() - Initialize the MSI inbound memory windows
85*4882a593Smuzhiyun * @ntb: NTB device context
86*4882a593Smuzhiyun *
87*4882a593Smuzhiyun * This function sets up the required inbound memory windows. It should be
88*4882a593Smuzhiyun * called from a work function after a link up event.
89*4882a593Smuzhiyun *
90*4882a593Smuzhiyun * Over the entire network, this function will reserves the last N
91*4882a593Smuzhiyun * inbound memory windows for each peer (where N is the number of peers).
92*4882a593Smuzhiyun *
93*4882a593Smuzhiyun * ntb_msi_init() must be called before this function.
94*4882a593Smuzhiyun *
95*4882a593Smuzhiyun * Return: Zero on success, otherwise a negative error number.
96*4882a593Smuzhiyun */
ntb_msi_setup_mws(struct ntb_dev * ntb)97*4882a593Smuzhiyun int ntb_msi_setup_mws(struct ntb_dev *ntb)
98*4882a593Smuzhiyun {
99*4882a593Smuzhiyun struct msi_desc *desc;
100*4882a593Smuzhiyun u64 addr;
101*4882a593Smuzhiyun int peer, peer_widx;
102*4882a593Smuzhiyun resource_size_t addr_align, size_align, size_max;
103*4882a593Smuzhiyun resource_size_t mw_size = SZ_32K;
104*4882a593Smuzhiyun resource_size_t mw_min_size = mw_size;
105*4882a593Smuzhiyun int i;
106*4882a593Smuzhiyun int ret;
107*4882a593Smuzhiyun
108*4882a593Smuzhiyun if (!ntb->msi)
109*4882a593Smuzhiyun return -EINVAL;
110*4882a593Smuzhiyun
111*4882a593Smuzhiyun desc = first_msi_entry(&ntb->pdev->dev);
112*4882a593Smuzhiyun addr = desc->msg.address_lo + ((uint64_t)desc->msg.address_hi << 32);
113*4882a593Smuzhiyun
114*4882a593Smuzhiyun for (peer = 0; peer < ntb_peer_port_count(ntb); peer++) {
115*4882a593Smuzhiyun peer_widx = ntb_peer_highest_mw_idx(ntb, peer);
116*4882a593Smuzhiyun if (peer_widx < 0)
117*4882a593Smuzhiyun return peer_widx;
118*4882a593Smuzhiyun
119*4882a593Smuzhiyun ret = ntb_mw_get_align(ntb, peer, peer_widx, &addr_align,
120*4882a593Smuzhiyun NULL, NULL);
121*4882a593Smuzhiyun if (ret)
122*4882a593Smuzhiyun return ret;
123*4882a593Smuzhiyun
124*4882a593Smuzhiyun addr &= ~(addr_align - 1);
125*4882a593Smuzhiyun }
126*4882a593Smuzhiyun
127*4882a593Smuzhiyun for (peer = 0; peer < ntb_peer_port_count(ntb); peer++) {
128*4882a593Smuzhiyun peer_widx = ntb_peer_highest_mw_idx(ntb, peer);
129*4882a593Smuzhiyun if (peer_widx < 0) {
130*4882a593Smuzhiyun ret = peer_widx;
131*4882a593Smuzhiyun goto error_out;
132*4882a593Smuzhiyun }
133*4882a593Smuzhiyun
134*4882a593Smuzhiyun ret = ntb_mw_get_align(ntb, peer, peer_widx, NULL,
135*4882a593Smuzhiyun &size_align, &size_max);
136*4882a593Smuzhiyun if (ret)
137*4882a593Smuzhiyun goto error_out;
138*4882a593Smuzhiyun
139*4882a593Smuzhiyun mw_size = round_up(mw_size, size_align);
140*4882a593Smuzhiyun mw_size = max(mw_size, size_max);
141*4882a593Smuzhiyun if (mw_size < mw_min_size)
142*4882a593Smuzhiyun mw_min_size = mw_size;
143*4882a593Smuzhiyun
144*4882a593Smuzhiyun ret = ntb_mw_set_trans(ntb, peer, peer_widx,
145*4882a593Smuzhiyun addr, mw_size);
146*4882a593Smuzhiyun if (ret)
147*4882a593Smuzhiyun goto error_out;
148*4882a593Smuzhiyun }
149*4882a593Smuzhiyun
150*4882a593Smuzhiyun ntb->msi->base_addr = addr;
151*4882a593Smuzhiyun ntb->msi->end_addr = addr + mw_min_size;
152*4882a593Smuzhiyun
153*4882a593Smuzhiyun return 0;
154*4882a593Smuzhiyun
155*4882a593Smuzhiyun error_out:
156*4882a593Smuzhiyun for (i = 0; i < peer; i++) {
157*4882a593Smuzhiyun peer_widx = ntb_peer_highest_mw_idx(ntb, peer);
158*4882a593Smuzhiyun if (peer_widx < 0)
159*4882a593Smuzhiyun continue;
160*4882a593Smuzhiyun
161*4882a593Smuzhiyun ntb_mw_clear_trans(ntb, i, peer_widx);
162*4882a593Smuzhiyun }
163*4882a593Smuzhiyun
164*4882a593Smuzhiyun return ret;
165*4882a593Smuzhiyun }
166*4882a593Smuzhiyun EXPORT_SYMBOL(ntb_msi_setup_mws);
167*4882a593Smuzhiyun
168*4882a593Smuzhiyun /**
169*4882a593Smuzhiyun * ntb_msi_clear_mws() - Clear all inbound memory windows
170*4882a593Smuzhiyun * @ntb: NTB device context
171*4882a593Smuzhiyun *
172*4882a593Smuzhiyun * This function tears down the resources used by ntb_msi_setup_mws().
173*4882a593Smuzhiyun */
ntb_msi_clear_mws(struct ntb_dev * ntb)174*4882a593Smuzhiyun void ntb_msi_clear_mws(struct ntb_dev *ntb)
175*4882a593Smuzhiyun {
176*4882a593Smuzhiyun int peer;
177*4882a593Smuzhiyun int peer_widx;
178*4882a593Smuzhiyun
179*4882a593Smuzhiyun for (peer = 0; peer < ntb_peer_port_count(ntb); peer++) {
180*4882a593Smuzhiyun peer_widx = ntb_peer_highest_mw_idx(ntb, peer);
181*4882a593Smuzhiyun if (peer_widx < 0)
182*4882a593Smuzhiyun continue;
183*4882a593Smuzhiyun
184*4882a593Smuzhiyun ntb_mw_clear_trans(ntb, peer, peer_widx);
185*4882a593Smuzhiyun }
186*4882a593Smuzhiyun }
187*4882a593Smuzhiyun EXPORT_SYMBOL(ntb_msi_clear_mws);
188*4882a593Smuzhiyun
189*4882a593Smuzhiyun struct ntb_msi_devres {
190*4882a593Smuzhiyun struct ntb_dev *ntb;
191*4882a593Smuzhiyun struct msi_desc *entry;
192*4882a593Smuzhiyun struct ntb_msi_desc *msi_desc;
193*4882a593Smuzhiyun };
194*4882a593Smuzhiyun
ntb_msi_set_desc(struct ntb_dev * ntb,struct msi_desc * entry,struct ntb_msi_desc * msi_desc)195*4882a593Smuzhiyun static int ntb_msi_set_desc(struct ntb_dev *ntb, struct msi_desc *entry,
196*4882a593Smuzhiyun struct ntb_msi_desc *msi_desc)
197*4882a593Smuzhiyun {
198*4882a593Smuzhiyun u64 addr;
199*4882a593Smuzhiyun
200*4882a593Smuzhiyun addr = entry->msg.address_lo +
201*4882a593Smuzhiyun ((uint64_t)entry->msg.address_hi << 32);
202*4882a593Smuzhiyun
203*4882a593Smuzhiyun if (addr < ntb->msi->base_addr || addr >= ntb->msi->end_addr) {
204*4882a593Smuzhiyun dev_warn_once(&ntb->dev,
205*4882a593Smuzhiyun "IRQ %d: MSI Address not within the memory window (%llx, [%llx %llx])\n",
206*4882a593Smuzhiyun entry->irq, addr, ntb->msi->base_addr,
207*4882a593Smuzhiyun ntb->msi->end_addr);
208*4882a593Smuzhiyun return -EFAULT;
209*4882a593Smuzhiyun }
210*4882a593Smuzhiyun
211*4882a593Smuzhiyun msi_desc->addr_offset = addr - ntb->msi->base_addr;
212*4882a593Smuzhiyun msi_desc->data = entry->msg.data;
213*4882a593Smuzhiyun
214*4882a593Smuzhiyun return 0;
215*4882a593Smuzhiyun }
216*4882a593Smuzhiyun
ntb_msi_write_msg(struct msi_desc * entry,void * data)217*4882a593Smuzhiyun static void ntb_msi_write_msg(struct msi_desc *entry, void *data)
218*4882a593Smuzhiyun {
219*4882a593Smuzhiyun struct ntb_msi_devres *dr = data;
220*4882a593Smuzhiyun
221*4882a593Smuzhiyun WARN_ON(ntb_msi_set_desc(dr->ntb, entry, dr->msi_desc));
222*4882a593Smuzhiyun
223*4882a593Smuzhiyun if (dr->ntb->msi->desc_changed)
224*4882a593Smuzhiyun dr->ntb->msi->desc_changed(dr->ntb->ctx);
225*4882a593Smuzhiyun }
226*4882a593Smuzhiyun
ntbm_msi_callback_release(struct device * dev,void * res)227*4882a593Smuzhiyun static void ntbm_msi_callback_release(struct device *dev, void *res)
228*4882a593Smuzhiyun {
229*4882a593Smuzhiyun struct ntb_msi_devres *dr = res;
230*4882a593Smuzhiyun
231*4882a593Smuzhiyun dr->entry->write_msi_msg = NULL;
232*4882a593Smuzhiyun dr->entry->write_msi_msg_data = NULL;
233*4882a593Smuzhiyun }
234*4882a593Smuzhiyun
ntbm_msi_setup_callback(struct ntb_dev * ntb,struct msi_desc * entry,struct ntb_msi_desc * msi_desc)235*4882a593Smuzhiyun static int ntbm_msi_setup_callback(struct ntb_dev *ntb, struct msi_desc *entry,
236*4882a593Smuzhiyun struct ntb_msi_desc *msi_desc)
237*4882a593Smuzhiyun {
238*4882a593Smuzhiyun struct ntb_msi_devres *dr;
239*4882a593Smuzhiyun
240*4882a593Smuzhiyun dr = devres_alloc(ntbm_msi_callback_release,
241*4882a593Smuzhiyun sizeof(struct ntb_msi_devres), GFP_KERNEL);
242*4882a593Smuzhiyun if (!dr)
243*4882a593Smuzhiyun return -ENOMEM;
244*4882a593Smuzhiyun
245*4882a593Smuzhiyun dr->ntb = ntb;
246*4882a593Smuzhiyun dr->entry = entry;
247*4882a593Smuzhiyun dr->msi_desc = msi_desc;
248*4882a593Smuzhiyun
249*4882a593Smuzhiyun devres_add(&ntb->dev, dr);
250*4882a593Smuzhiyun
251*4882a593Smuzhiyun dr->entry->write_msi_msg = ntb_msi_write_msg;
252*4882a593Smuzhiyun dr->entry->write_msi_msg_data = dr;
253*4882a593Smuzhiyun
254*4882a593Smuzhiyun return 0;
255*4882a593Smuzhiyun }
256*4882a593Smuzhiyun
257*4882a593Smuzhiyun /**
258*4882a593Smuzhiyun * ntbm_msi_request_threaded_irq() - allocate an MSI interrupt
259*4882a593Smuzhiyun * @ntb: NTB device context
260*4882a593Smuzhiyun * @handler: Function to be called when the IRQ occurs
261*4882a593Smuzhiyun * @thread_fn: Function to be called in a threaded interrupt context. NULL
262*4882a593Smuzhiyun * for clients which handle everything in @handler
263*4882a593Smuzhiyun * @devname: An ascii name for the claiming device, dev_name(dev) if NULL
264*4882a593Smuzhiyun * @dev_id: A cookie passed back to the handler function
265*4882a593Smuzhiyun *
266*4882a593Smuzhiyun * This function assigns an interrupt handler to an unused
267*4882a593Smuzhiyun * MSI interrupt and returns the descriptor used to trigger
268*4882a593Smuzhiyun * it. The descriptor can then be sent to a peer to trigger
269*4882a593Smuzhiyun * the interrupt.
270*4882a593Smuzhiyun *
271*4882a593Smuzhiyun * The interrupt resource is managed with devres so it will
272*4882a593Smuzhiyun * be automatically freed when the NTB device is torn down.
273*4882a593Smuzhiyun *
274*4882a593Smuzhiyun * If an IRQ allocated with this function needs to be freed
275*4882a593Smuzhiyun * separately, ntbm_free_irq() must be used.
276*4882a593Smuzhiyun *
277*4882a593Smuzhiyun * Return: IRQ number assigned on success, otherwise a negative error number.
278*4882a593Smuzhiyun */
ntbm_msi_request_threaded_irq(struct ntb_dev * ntb,irq_handler_t handler,irq_handler_t thread_fn,const char * name,void * dev_id,struct ntb_msi_desc * msi_desc)279*4882a593Smuzhiyun int ntbm_msi_request_threaded_irq(struct ntb_dev *ntb, irq_handler_t handler,
280*4882a593Smuzhiyun irq_handler_t thread_fn,
281*4882a593Smuzhiyun const char *name, void *dev_id,
282*4882a593Smuzhiyun struct ntb_msi_desc *msi_desc)
283*4882a593Smuzhiyun {
284*4882a593Smuzhiyun struct msi_desc *entry;
285*4882a593Smuzhiyun struct irq_desc *desc;
286*4882a593Smuzhiyun int ret;
287*4882a593Smuzhiyun
288*4882a593Smuzhiyun if (!ntb->msi)
289*4882a593Smuzhiyun return -EINVAL;
290*4882a593Smuzhiyun
291*4882a593Smuzhiyun for_each_pci_msi_entry(entry, ntb->pdev) {
292*4882a593Smuzhiyun desc = irq_to_desc(entry->irq);
293*4882a593Smuzhiyun if (desc->action)
294*4882a593Smuzhiyun continue;
295*4882a593Smuzhiyun
296*4882a593Smuzhiyun ret = devm_request_threaded_irq(&ntb->dev, entry->irq, handler,
297*4882a593Smuzhiyun thread_fn, 0, name, dev_id);
298*4882a593Smuzhiyun if (ret)
299*4882a593Smuzhiyun continue;
300*4882a593Smuzhiyun
301*4882a593Smuzhiyun if (ntb_msi_set_desc(ntb, entry, msi_desc)) {
302*4882a593Smuzhiyun devm_free_irq(&ntb->dev, entry->irq, dev_id);
303*4882a593Smuzhiyun continue;
304*4882a593Smuzhiyun }
305*4882a593Smuzhiyun
306*4882a593Smuzhiyun ret = ntbm_msi_setup_callback(ntb, entry, msi_desc);
307*4882a593Smuzhiyun if (ret) {
308*4882a593Smuzhiyun devm_free_irq(&ntb->dev, entry->irq, dev_id);
309*4882a593Smuzhiyun return ret;
310*4882a593Smuzhiyun }
311*4882a593Smuzhiyun
312*4882a593Smuzhiyun
313*4882a593Smuzhiyun return entry->irq;
314*4882a593Smuzhiyun }
315*4882a593Smuzhiyun
316*4882a593Smuzhiyun return -ENODEV;
317*4882a593Smuzhiyun }
318*4882a593Smuzhiyun EXPORT_SYMBOL(ntbm_msi_request_threaded_irq);
319*4882a593Smuzhiyun
ntbm_msi_callback_match(struct device * dev,void * res,void * data)320*4882a593Smuzhiyun static int ntbm_msi_callback_match(struct device *dev, void *res, void *data)
321*4882a593Smuzhiyun {
322*4882a593Smuzhiyun struct ntb_dev *ntb = dev_ntb(dev);
323*4882a593Smuzhiyun struct ntb_msi_devres *dr = res;
324*4882a593Smuzhiyun
325*4882a593Smuzhiyun return dr->ntb == ntb && dr->entry == data;
326*4882a593Smuzhiyun }
327*4882a593Smuzhiyun
328*4882a593Smuzhiyun /**
329*4882a593Smuzhiyun * ntbm_msi_free_irq() - free an interrupt
330*4882a593Smuzhiyun * @ntb: NTB device context
331*4882a593Smuzhiyun * @irq: Interrupt line to free
332*4882a593Smuzhiyun * @dev_id: Device identity to free
333*4882a593Smuzhiyun *
334*4882a593Smuzhiyun * This function should be used to manually free IRQs allocated with
335*4882a593Smuzhiyun * ntbm_request_[threaded_]irq().
336*4882a593Smuzhiyun */
ntbm_msi_free_irq(struct ntb_dev * ntb,unsigned int irq,void * dev_id)337*4882a593Smuzhiyun void ntbm_msi_free_irq(struct ntb_dev *ntb, unsigned int irq, void *dev_id)
338*4882a593Smuzhiyun {
339*4882a593Smuzhiyun struct msi_desc *entry = irq_get_msi_desc(irq);
340*4882a593Smuzhiyun
341*4882a593Smuzhiyun entry->write_msi_msg = NULL;
342*4882a593Smuzhiyun entry->write_msi_msg_data = NULL;
343*4882a593Smuzhiyun
344*4882a593Smuzhiyun WARN_ON(devres_destroy(&ntb->dev, ntbm_msi_callback_release,
345*4882a593Smuzhiyun ntbm_msi_callback_match, entry));
346*4882a593Smuzhiyun
347*4882a593Smuzhiyun devm_free_irq(&ntb->dev, irq, dev_id);
348*4882a593Smuzhiyun }
349*4882a593Smuzhiyun EXPORT_SYMBOL(ntbm_msi_free_irq);
350*4882a593Smuzhiyun
351*4882a593Smuzhiyun /**
352*4882a593Smuzhiyun * ntb_msi_peer_trigger() - Trigger an interrupt handler on a peer
353*4882a593Smuzhiyun * @ntb: NTB device context
354*4882a593Smuzhiyun * @peer: Peer index
355*4882a593Smuzhiyun * @desc: MSI descriptor data which triggers the interrupt
356*4882a593Smuzhiyun *
357*4882a593Smuzhiyun * This function triggers an interrupt on a peer. It requires
358*4882a593Smuzhiyun * the descriptor structure to have been passed from that peer
359*4882a593Smuzhiyun * by some other means.
360*4882a593Smuzhiyun *
361*4882a593Smuzhiyun * Return: Zero on success, otherwise a negative error number.
362*4882a593Smuzhiyun */
ntb_msi_peer_trigger(struct ntb_dev * ntb,int peer,struct ntb_msi_desc * desc)363*4882a593Smuzhiyun int ntb_msi_peer_trigger(struct ntb_dev *ntb, int peer,
364*4882a593Smuzhiyun struct ntb_msi_desc *desc)
365*4882a593Smuzhiyun {
366*4882a593Smuzhiyun int idx;
367*4882a593Smuzhiyun
368*4882a593Smuzhiyun if (!ntb->msi)
369*4882a593Smuzhiyun return -EINVAL;
370*4882a593Smuzhiyun
371*4882a593Smuzhiyun idx = desc->addr_offset / sizeof(*ntb->msi->peer_mws[peer]);
372*4882a593Smuzhiyun
373*4882a593Smuzhiyun iowrite32(desc->data, &ntb->msi->peer_mws[peer][idx]);
374*4882a593Smuzhiyun
375*4882a593Smuzhiyun return 0;
376*4882a593Smuzhiyun }
377*4882a593Smuzhiyun EXPORT_SYMBOL(ntb_msi_peer_trigger);
378*4882a593Smuzhiyun
379*4882a593Smuzhiyun /**
380*4882a593Smuzhiyun * ntb_msi_peer_addr() - Get the DMA address to trigger a peer's MSI interrupt
381*4882a593Smuzhiyun * @ntb: NTB device context
382*4882a593Smuzhiyun * @peer: Peer index
383*4882a593Smuzhiyun * @desc: MSI descriptor data which triggers the interrupt
384*4882a593Smuzhiyun * @msi_addr: Physical address to trigger the interrupt
385*4882a593Smuzhiyun *
386*4882a593Smuzhiyun * This function allows using DMA engines to trigger an interrupt
387*4882a593Smuzhiyun * (for example, trigger an interrupt to process the data after
388*4882a593Smuzhiyun * sending it). To trigger the interrupt, write @desc.data to the address
389*4882a593Smuzhiyun * returned in @msi_addr
390*4882a593Smuzhiyun *
391*4882a593Smuzhiyun * Return: Zero on success, otherwise a negative error number.
392*4882a593Smuzhiyun */
ntb_msi_peer_addr(struct ntb_dev * ntb,int peer,struct ntb_msi_desc * desc,phys_addr_t * msi_addr)393*4882a593Smuzhiyun int ntb_msi_peer_addr(struct ntb_dev *ntb, int peer,
394*4882a593Smuzhiyun struct ntb_msi_desc *desc,
395*4882a593Smuzhiyun phys_addr_t *msi_addr)
396*4882a593Smuzhiyun {
397*4882a593Smuzhiyun int peer_widx = ntb_peer_mw_count(ntb) - 1 - peer;
398*4882a593Smuzhiyun phys_addr_t mw_phys_addr;
399*4882a593Smuzhiyun int ret;
400*4882a593Smuzhiyun
401*4882a593Smuzhiyun ret = ntb_peer_mw_get_addr(ntb, peer_widx, &mw_phys_addr, NULL);
402*4882a593Smuzhiyun if (ret)
403*4882a593Smuzhiyun return ret;
404*4882a593Smuzhiyun
405*4882a593Smuzhiyun if (msi_addr)
406*4882a593Smuzhiyun *msi_addr = mw_phys_addr + desc->addr_offset;
407*4882a593Smuzhiyun
408*4882a593Smuzhiyun return 0;
409*4882a593Smuzhiyun }
410*4882a593Smuzhiyun EXPORT_SYMBOL(ntb_msi_peer_addr);
411