1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun * Copyright (C) Rockchip Electronics Co.Ltd
4*4882a593Smuzhiyun * Author: Felix Zeng <felix.zeng@rock-chips.com>
5*4882a593Smuzhiyun */
6*4882a593Smuzhiyun
7*4882a593Smuzhiyun #include <linux/version.h>
8*4882a593Smuzhiyun #include <linux/rk-dma-heap.h>
9*4882a593Smuzhiyun
10*4882a593Smuzhiyun #if KERNEL_VERSION(5, 10, 0) <= LINUX_VERSION_CODE
11*4882a593Smuzhiyun #include <linux/dma-map-ops.h>
12*4882a593Smuzhiyun #endif
13*4882a593Smuzhiyun
14*4882a593Smuzhiyun #include "rknpu_drv.h"
15*4882a593Smuzhiyun #include "rknpu_ioctl.h"
16*4882a593Smuzhiyun #include "rknpu_mem.h"
17*4882a593Smuzhiyun
rknpu_mem_create_ioctl(struct rknpu_device * rknpu_dev,unsigned long data,struct file * file)18*4882a593Smuzhiyun int rknpu_mem_create_ioctl(struct rknpu_device *rknpu_dev, unsigned long data,
19*4882a593Smuzhiyun struct file *file)
20*4882a593Smuzhiyun {
21*4882a593Smuzhiyun struct rknpu_mem_create args;
22*4882a593Smuzhiyun int ret = -EINVAL;
23*4882a593Smuzhiyun struct dma_buf_attachment *attachment;
24*4882a593Smuzhiyun struct sg_table *table;
25*4882a593Smuzhiyun struct scatterlist *sgl;
26*4882a593Smuzhiyun dma_addr_t phys;
27*4882a593Smuzhiyun struct dma_buf *dmabuf;
28*4882a593Smuzhiyun struct page **pages;
29*4882a593Smuzhiyun struct page *page;
30*4882a593Smuzhiyun struct rknpu_mem_object *rknpu_obj = NULL;
31*4882a593Smuzhiyun struct rknpu_session *session = NULL;
32*4882a593Smuzhiyun int i, fd;
33*4882a593Smuzhiyun unsigned int length, page_count;
34*4882a593Smuzhiyun
35*4882a593Smuzhiyun if (unlikely(copy_from_user(&args, (struct rknpu_mem_create *)data,
36*4882a593Smuzhiyun sizeof(struct rknpu_mem_create)))) {
37*4882a593Smuzhiyun LOG_ERROR("%s: copy_from_user failed\n", __func__);
38*4882a593Smuzhiyun ret = -EFAULT;
39*4882a593Smuzhiyun return ret;
40*4882a593Smuzhiyun }
41*4882a593Smuzhiyun
42*4882a593Smuzhiyun if (args.flags & RKNPU_MEM_NON_CONTIGUOUS) {
43*4882a593Smuzhiyun LOG_ERROR("%s: malloc iommu memory unsupported in current!\n",
44*4882a593Smuzhiyun __func__);
45*4882a593Smuzhiyun ret = -EINVAL;
46*4882a593Smuzhiyun return ret;
47*4882a593Smuzhiyun }
48*4882a593Smuzhiyun
49*4882a593Smuzhiyun rknpu_obj = kzalloc(sizeof(*rknpu_obj), GFP_KERNEL);
50*4882a593Smuzhiyun if (!rknpu_obj)
51*4882a593Smuzhiyun return -ENOMEM;
52*4882a593Smuzhiyun
53*4882a593Smuzhiyun if (args.handle > 0) {
54*4882a593Smuzhiyun fd = args.handle;
55*4882a593Smuzhiyun
56*4882a593Smuzhiyun dmabuf = dma_buf_get(fd);
57*4882a593Smuzhiyun if (IS_ERR(dmabuf)) {
58*4882a593Smuzhiyun ret = PTR_ERR(dmabuf);
59*4882a593Smuzhiyun goto err_free_obj;
60*4882a593Smuzhiyun }
61*4882a593Smuzhiyun
62*4882a593Smuzhiyun rknpu_obj->dmabuf = dmabuf;
63*4882a593Smuzhiyun rknpu_obj->owner = 0;
64*4882a593Smuzhiyun } else {
65*4882a593Smuzhiyun /* Start test kernel alloc/free dma buf */
66*4882a593Smuzhiyun dmabuf = rk_dma_heap_buffer_alloc(rknpu_dev->heap, args.size,
67*4882a593Smuzhiyun O_CLOEXEC | O_RDWR, 0x0,
68*4882a593Smuzhiyun dev_name(rknpu_dev->dev));
69*4882a593Smuzhiyun if (IS_ERR(dmabuf)) {
70*4882a593Smuzhiyun LOG_ERROR("dmabuf alloc failed, args.size = %llu\n",
71*4882a593Smuzhiyun args.size);
72*4882a593Smuzhiyun ret = PTR_ERR(dmabuf);
73*4882a593Smuzhiyun goto err_free_obj;
74*4882a593Smuzhiyun }
75*4882a593Smuzhiyun
76*4882a593Smuzhiyun rknpu_obj->dmabuf = dmabuf;
77*4882a593Smuzhiyun rknpu_obj->owner = 1;
78*4882a593Smuzhiyun
79*4882a593Smuzhiyun fd = dma_buf_fd(dmabuf, O_CLOEXEC | O_RDWR);
80*4882a593Smuzhiyun if (fd < 0) {
81*4882a593Smuzhiyun LOG_ERROR("dmabuf fd get failed\n");
82*4882a593Smuzhiyun ret = -EFAULT;
83*4882a593Smuzhiyun goto err_free_dma_buf;
84*4882a593Smuzhiyun }
85*4882a593Smuzhiyun }
86*4882a593Smuzhiyun
87*4882a593Smuzhiyun attachment = dma_buf_attach(dmabuf, rknpu_dev->dev);
88*4882a593Smuzhiyun if (IS_ERR(attachment)) {
89*4882a593Smuzhiyun LOG_ERROR("dma_buf_attach failed\n");
90*4882a593Smuzhiyun ret = PTR_ERR(attachment);
91*4882a593Smuzhiyun goto err_free_dma_buf;
92*4882a593Smuzhiyun }
93*4882a593Smuzhiyun
94*4882a593Smuzhiyun table = dma_buf_map_attachment(attachment, DMA_BIDIRECTIONAL);
95*4882a593Smuzhiyun if (IS_ERR(table)) {
96*4882a593Smuzhiyun LOG_ERROR("dma_buf_attach failed\n");
97*4882a593Smuzhiyun dma_buf_detach(dmabuf, attachment);
98*4882a593Smuzhiyun ret = PTR_ERR(table);
99*4882a593Smuzhiyun goto err_free_dma_buf;
100*4882a593Smuzhiyun }
101*4882a593Smuzhiyun
102*4882a593Smuzhiyun for_each_sgtable_sg(table, sgl, i) {
103*4882a593Smuzhiyun phys = sg_dma_address(sgl);
104*4882a593Smuzhiyun page = sg_page(sgl);
105*4882a593Smuzhiyun length = sg_dma_len(sgl);
106*4882a593Smuzhiyun LOG_DEBUG("%s, %d, phys: %pad, length: %u\n", __func__,
107*4882a593Smuzhiyun __LINE__, &phys, length);
108*4882a593Smuzhiyun }
109*4882a593Smuzhiyun
110*4882a593Smuzhiyun page_count = length >> PAGE_SHIFT;
111*4882a593Smuzhiyun pages = kmalloc_array(page_count, sizeof(struct page), GFP_KERNEL);
112*4882a593Smuzhiyun if (!pages) {
113*4882a593Smuzhiyun LOG_ERROR("alloc pages failed\n");
114*4882a593Smuzhiyun ret = -ENOMEM;
115*4882a593Smuzhiyun goto err_detach_dma_buf;
116*4882a593Smuzhiyun }
117*4882a593Smuzhiyun
118*4882a593Smuzhiyun for (i = 0; i < page_count; i++)
119*4882a593Smuzhiyun pages[i] = &page[i];
120*4882a593Smuzhiyun
121*4882a593Smuzhiyun rknpu_obj->kv_addr = vmap(pages, page_count, VM_MAP, PAGE_KERNEL);
122*4882a593Smuzhiyun if (!rknpu_obj->kv_addr) {
123*4882a593Smuzhiyun LOG_ERROR("vmap pages addr failed\n");
124*4882a593Smuzhiyun ret = -ENOMEM;
125*4882a593Smuzhiyun goto err_free_pages;
126*4882a593Smuzhiyun }
127*4882a593Smuzhiyun
128*4882a593Smuzhiyun rknpu_obj->size = PAGE_ALIGN(args.size);
129*4882a593Smuzhiyun rknpu_obj->dma_addr = phys;
130*4882a593Smuzhiyun rknpu_obj->sgt = table;
131*4882a593Smuzhiyun
132*4882a593Smuzhiyun args.size = rknpu_obj->size;
133*4882a593Smuzhiyun args.obj_addr = (__u64)(uintptr_t)rknpu_obj;
134*4882a593Smuzhiyun args.dma_addr = rknpu_obj->dma_addr;
135*4882a593Smuzhiyun args.handle = fd;
136*4882a593Smuzhiyun
137*4882a593Smuzhiyun LOG_DEBUG(
138*4882a593Smuzhiyun "args.handle: %d, args.size: %lld, rknpu_obj: %#llx, rknpu_obj->dma_addr: %#llx\n",
139*4882a593Smuzhiyun args.handle, args.size, (__u64)(uintptr_t)rknpu_obj,
140*4882a593Smuzhiyun (__u64)rknpu_obj->dma_addr);
141*4882a593Smuzhiyun
142*4882a593Smuzhiyun if (unlikely(copy_to_user((struct rknpu_mem_create *)data, &args,
143*4882a593Smuzhiyun sizeof(struct rknpu_mem_create)))) {
144*4882a593Smuzhiyun LOG_ERROR("%s: copy_to_user failed\n", __func__);
145*4882a593Smuzhiyun ret = -EFAULT;
146*4882a593Smuzhiyun goto err_unmap_kv_addr;
147*4882a593Smuzhiyun }
148*4882a593Smuzhiyun
149*4882a593Smuzhiyun kfree(pages);
150*4882a593Smuzhiyun dma_buf_unmap_attachment(attachment, table, DMA_BIDIRECTIONAL);
151*4882a593Smuzhiyun dma_buf_detach(dmabuf, attachment);
152*4882a593Smuzhiyun
153*4882a593Smuzhiyun spin_lock(&rknpu_dev->lock);
154*4882a593Smuzhiyun
155*4882a593Smuzhiyun session = file->private_data;
156*4882a593Smuzhiyun if (!session) {
157*4882a593Smuzhiyun spin_unlock(&rknpu_dev->lock);
158*4882a593Smuzhiyun ret = -EFAULT;
159*4882a593Smuzhiyun goto err_unmap_kv_addr;
160*4882a593Smuzhiyun }
161*4882a593Smuzhiyun list_add_tail(&rknpu_obj->head, &session->list);
162*4882a593Smuzhiyun
163*4882a593Smuzhiyun spin_unlock(&rknpu_dev->lock);
164*4882a593Smuzhiyun
165*4882a593Smuzhiyun return 0;
166*4882a593Smuzhiyun
167*4882a593Smuzhiyun err_unmap_kv_addr:
168*4882a593Smuzhiyun vunmap(rknpu_obj->kv_addr);
169*4882a593Smuzhiyun rknpu_obj->kv_addr = NULL;
170*4882a593Smuzhiyun
171*4882a593Smuzhiyun err_free_pages:
172*4882a593Smuzhiyun kfree(pages);
173*4882a593Smuzhiyun
174*4882a593Smuzhiyun err_detach_dma_buf:
175*4882a593Smuzhiyun dma_buf_unmap_attachment(attachment, table, DMA_BIDIRECTIONAL);
176*4882a593Smuzhiyun dma_buf_detach(dmabuf, attachment);
177*4882a593Smuzhiyun
178*4882a593Smuzhiyun err_free_dma_buf:
179*4882a593Smuzhiyun if (rknpu_obj->owner)
180*4882a593Smuzhiyun rk_dma_heap_buffer_free(dmabuf);
181*4882a593Smuzhiyun else
182*4882a593Smuzhiyun dma_buf_put(dmabuf);
183*4882a593Smuzhiyun
184*4882a593Smuzhiyun err_free_obj:
185*4882a593Smuzhiyun kfree(rknpu_obj);
186*4882a593Smuzhiyun
187*4882a593Smuzhiyun return ret;
188*4882a593Smuzhiyun }
189*4882a593Smuzhiyun
rknpu_mem_destroy_ioctl(struct rknpu_device * rknpu_dev,unsigned long data,struct file * file)190*4882a593Smuzhiyun int rknpu_mem_destroy_ioctl(struct rknpu_device *rknpu_dev, unsigned long data,
191*4882a593Smuzhiyun struct file *file)
192*4882a593Smuzhiyun {
193*4882a593Smuzhiyun struct rknpu_mem_object *rknpu_obj, *entry, *q;
194*4882a593Smuzhiyun struct rknpu_session *session = NULL;
195*4882a593Smuzhiyun struct rknpu_mem_destroy args;
196*4882a593Smuzhiyun int ret = -EFAULT;
197*4882a593Smuzhiyun
198*4882a593Smuzhiyun if (unlikely(copy_from_user(&args, (struct rknpu_mem_destroy *)data,
199*4882a593Smuzhiyun sizeof(struct rknpu_mem_destroy)))) {
200*4882a593Smuzhiyun LOG_ERROR("%s: copy_from_user failed\n", __func__);
201*4882a593Smuzhiyun ret = -EFAULT;
202*4882a593Smuzhiyun return ret;
203*4882a593Smuzhiyun }
204*4882a593Smuzhiyun
205*4882a593Smuzhiyun if (!kern_addr_valid(args.obj_addr)) {
206*4882a593Smuzhiyun LOG_ERROR("%s: invalid obj_addr: %#llx\n", __func__,
207*4882a593Smuzhiyun (__u64)(uintptr_t)args.obj_addr);
208*4882a593Smuzhiyun ret = -EINVAL;
209*4882a593Smuzhiyun return ret;
210*4882a593Smuzhiyun }
211*4882a593Smuzhiyun
212*4882a593Smuzhiyun rknpu_obj = (struct rknpu_mem_object *)(uintptr_t)args.obj_addr;
213*4882a593Smuzhiyun LOG_DEBUG(
214*4882a593Smuzhiyun "free args.handle: %d, rknpu_obj: %#llx, rknpu_obj->dma_addr: %#llx\n",
215*4882a593Smuzhiyun args.handle, (__u64)(uintptr_t)rknpu_obj,
216*4882a593Smuzhiyun (__u64)rknpu_obj->dma_addr);
217*4882a593Smuzhiyun
218*4882a593Smuzhiyun spin_lock(&rknpu_dev->lock);
219*4882a593Smuzhiyun session = file->private_data;
220*4882a593Smuzhiyun if (!session) {
221*4882a593Smuzhiyun spin_unlock(&rknpu_dev->lock);
222*4882a593Smuzhiyun ret = -EFAULT;
223*4882a593Smuzhiyun return ret;
224*4882a593Smuzhiyun }
225*4882a593Smuzhiyun list_for_each_entry_safe(entry, q, &session->list, head) {
226*4882a593Smuzhiyun if (entry == rknpu_obj) {
227*4882a593Smuzhiyun list_del(&entry->head);
228*4882a593Smuzhiyun break;
229*4882a593Smuzhiyun }
230*4882a593Smuzhiyun }
231*4882a593Smuzhiyun spin_unlock(&rknpu_dev->lock);
232*4882a593Smuzhiyun
233*4882a593Smuzhiyun if (rknpu_obj == entry) {
234*4882a593Smuzhiyun vunmap(rknpu_obj->kv_addr);
235*4882a593Smuzhiyun rknpu_obj->kv_addr = NULL;
236*4882a593Smuzhiyun
237*4882a593Smuzhiyun if (!rknpu_obj->owner)
238*4882a593Smuzhiyun dma_buf_put(rknpu_obj->dmabuf);
239*4882a593Smuzhiyun
240*4882a593Smuzhiyun kfree(rknpu_obj);
241*4882a593Smuzhiyun }
242*4882a593Smuzhiyun
243*4882a593Smuzhiyun return 0;
244*4882a593Smuzhiyun }
245*4882a593Smuzhiyun
246*4882a593Smuzhiyun /*
247*4882a593Smuzhiyun * begin cpu access => for_cpu = true
248*4882a593Smuzhiyun * end cpu access => for_cpu = false
249*4882a593Smuzhiyun */
rknpu_dma_buf_sync(struct rknpu_device * rknpu_dev,struct rknpu_mem_object * rknpu_obj,u32 offset,u32 length,enum dma_data_direction dir,bool for_cpu)250*4882a593Smuzhiyun static void __maybe_unused rknpu_dma_buf_sync(
251*4882a593Smuzhiyun struct rknpu_device *rknpu_dev, struct rknpu_mem_object *rknpu_obj,
252*4882a593Smuzhiyun u32 offset, u32 length, enum dma_data_direction dir, bool for_cpu)
253*4882a593Smuzhiyun {
254*4882a593Smuzhiyun struct device *dev = rknpu_dev->dev;
255*4882a593Smuzhiyun struct sg_table *sgt = rknpu_obj->sgt;
256*4882a593Smuzhiyun struct scatterlist *sg = sgt->sgl;
257*4882a593Smuzhiyun dma_addr_t sg_dma_addr = sg_dma_address(sg);
258*4882a593Smuzhiyun unsigned int len = 0;
259*4882a593Smuzhiyun int i;
260*4882a593Smuzhiyun
261*4882a593Smuzhiyun for_each_sgtable_sg(sgt, sg, i) {
262*4882a593Smuzhiyun unsigned int sg_offset, sg_left, size = 0;
263*4882a593Smuzhiyun
264*4882a593Smuzhiyun len += sg->length;
265*4882a593Smuzhiyun if (len <= offset) {
266*4882a593Smuzhiyun sg_dma_addr += sg->length;
267*4882a593Smuzhiyun continue;
268*4882a593Smuzhiyun }
269*4882a593Smuzhiyun
270*4882a593Smuzhiyun sg_left = len - offset;
271*4882a593Smuzhiyun sg_offset = sg->length - sg_left;
272*4882a593Smuzhiyun
273*4882a593Smuzhiyun size = (length < sg_left) ? length : sg_left;
274*4882a593Smuzhiyun
275*4882a593Smuzhiyun if (for_cpu)
276*4882a593Smuzhiyun dma_sync_single_range_for_cpu(dev, sg_dma_addr,
277*4882a593Smuzhiyun sg_offset, size, dir);
278*4882a593Smuzhiyun else
279*4882a593Smuzhiyun dma_sync_single_range_for_device(dev, sg_dma_addr,
280*4882a593Smuzhiyun sg_offset, size, dir);
281*4882a593Smuzhiyun
282*4882a593Smuzhiyun offset += size;
283*4882a593Smuzhiyun length -= size;
284*4882a593Smuzhiyun sg_dma_addr += sg->length;
285*4882a593Smuzhiyun
286*4882a593Smuzhiyun if (length == 0)
287*4882a593Smuzhiyun break;
288*4882a593Smuzhiyun }
289*4882a593Smuzhiyun }
290*4882a593Smuzhiyun
rknpu_mem_sync_ioctl(struct rknpu_device * rknpu_dev,unsigned long data)291*4882a593Smuzhiyun int rknpu_mem_sync_ioctl(struct rknpu_device *rknpu_dev, unsigned long data)
292*4882a593Smuzhiyun {
293*4882a593Smuzhiyun struct rknpu_mem_object *rknpu_obj = NULL;
294*4882a593Smuzhiyun struct rknpu_mem_sync args;
295*4882a593Smuzhiyun struct dma_buf *dmabuf;
296*4882a593Smuzhiyun int ret = -EFAULT;
297*4882a593Smuzhiyun
298*4882a593Smuzhiyun if (unlikely(copy_from_user(&args, (struct rknpu_mem_sync *)data,
299*4882a593Smuzhiyun sizeof(struct rknpu_mem_sync)))) {
300*4882a593Smuzhiyun LOG_ERROR("%s: copy_from_user failed\n", __func__);
301*4882a593Smuzhiyun ret = -EFAULT;
302*4882a593Smuzhiyun return ret;
303*4882a593Smuzhiyun }
304*4882a593Smuzhiyun
305*4882a593Smuzhiyun if (!kern_addr_valid(args.obj_addr)) {
306*4882a593Smuzhiyun LOG_ERROR("%s: invalid obj_addr: %#llx\n", __func__,
307*4882a593Smuzhiyun (__u64)(uintptr_t)args.obj_addr);
308*4882a593Smuzhiyun ret = -EINVAL;
309*4882a593Smuzhiyun return ret;
310*4882a593Smuzhiyun }
311*4882a593Smuzhiyun
312*4882a593Smuzhiyun rknpu_obj = (struct rknpu_mem_object *)(uintptr_t)args.obj_addr;
313*4882a593Smuzhiyun dmabuf = rknpu_obj->dmabuf;
314*4882a593Smuzhiyun
315*4882a593Smuzhiyun #ifndef CONFIG_DMABUF_PARTIAL
316*4882a593Smuzhiyun if (args.flags & RKNPU_MEM_SYNC_TO_DEVICE) {
317*4882a593Smuzhiyun rknpu_dma_buf_sync(rknpu_dev, rknpu_obj, args.offset, args.size,
318*4882a593Smuzhiyun DMA_TO_DEVICE, false);
319*4882a593Smuzhiyun }
320*4882a593Smuzhiyun if (args.flags & RKNPU_MEM_SYNC_FROM_DEVICE) {
321*4882a593Smuzhiyun rknpu_dma_buf_sync(rknpu_dev, rknpu_obj, args.offset, args.size,
322*4882a593Smuzhiyun DMA_FROM_DEVICE, true);
323*4882a593Smuzhiyun }
324*4882a593Smuzhiyun #else
325*4882a593Smuzhiyun if (args.flags & RKNPU_MEM_SYNC_TO_DEVICE) {
326*4882a593Smuzhiyun dmabuf->ops->end_cpu_access_partial(dmabuf, DMA_TO_DEVICE,
327*4882a593Smuzhiyun args.offset, args.size);
328*4882a593Smuzhiyun }
329*4882a593Smuzhiyun if (args.flags & RKNPU_MEM_SYNC_FROM_DEVICE) {
330*4882a593Smuzhiyun dmabuf->ops->begin_cpu_access_partial(dmabuf, DMA_FROM_DEVICE,
331*4882a593Smuzhiyun args.offset, args.size);
332*4882a593Smuzhiyun }
333*4882a593Smuzhiyun #endif
334*4882a593Smuzhiyun
335*4882a593Smuzhiyun return 0;
336*4882a593Smuzhiyun }
337