xref: /OK3568_Linux_fs/kernel/drivers/rknpu/rknpu_mem.c (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun  * Copyright (C) Rockchip Electronics Co.Ltd
4*4882a593Smuzhiyun  * Author: Felix Zeng <felix.zeng@rock-chips.com>
5*4882a593Smuzhiyun  */
6*4882a593Smuzhiyun 
7*4882a593Smuzhiyun #include <linux/version.h>
8*4882a593Smuzhiyun #include <linux/rk-dma-heap.h>
9*4882a593Smuzhiyun 
10*4882a593Smuzhiyun #if KERNEL_VERSION(5, 10, 0) <= LINUX_VERSION_CODE
11*4882a593Smuzhiyun #include <linux/dma-map-ops.h>
12*4882a593Smuzhiyun #endif
13*4882a593Smuzhiyun 
14*4882a593Smuzhiyun #include "rknpu_drv.h"
15*4882a593Smuzhiyun #include "rknpu_ioctl.h"
16*4882a593Smuzhiyun #include "rknpu_mem.h"
17*4882a593Smuzhiyun 
rknpu_mem_create_ioctl(struct rknpu_device * rknpu_dev,unsigned long data,struct file * file)18*4882a593Smuzhiyun int rknpu_mem_create_ioctl(struct rknpu_device *rknpu_dev, unsigned long data,
19*4882a593Smuzhiyun 			   struct file *file)
20*4882a593Smuzhiyun {
21*4882a593Smuzhiyun 	struct rknpu_mem_create args;
22*4882a593Smuzhiyun 	int ret = -EINVAL;
23*4882a593Smuzhiyun 	struct dma_buf_attachment *attachment;
24*4882a593Smuzhiyun 	struct sg_table *table;
25*4882a593Smuzhiyun 	struct scatterlist *sgl;
26*4882a593Smuzhiyun 	dma_addr_t phys;
27*4882a593Smuzhiyun 	struct dma_buf *dmabuf;
28*4882a593Smuzhiyun 	struct page **pages;
29*4882a593Smuzhiyun 	struct page *page;
30*4882a593Smuzhiyun 	struct rknpu_mem_object *rknpu_obj = NULL;
31*4882a593Smuzhiyun 	struct rknpu_session *session = NULL;
32*4882a593Smuzhiyun 	int i, fd;
33*4882a593Smuzhiyun 	unsigned int length, page_count;
34*4882a593Smuzhiyun 
35*4882a593Smuzhiyun 	if (unlikely(copy_from_user(&args, (struct rknpu_mem_create *)data,
36*4882a593Smuzhiyun 				    sizeof(struct rknpu_mem_create)))) {
37*4882a593Smuzhiyun 		LOG_ERROR("%s: copy_from_user failed\n", __func__);
38*4882a593Smuzhiyun 		ret = -EFAULT;
39*4882a593Smuzhiyun 		return ret;
40*4882a593Smuzhiyun 	}
41*4882a593Smuzhiyun 
42*4882a593Smuzhiyun 	if (args.flags & RKNPU_MEM_NON_CONTIGUOUS) {
43*4882a593Smuzhiyun 		LOG_ERROR("%s: malloc iommu memory unsupported in current!\n",
44*4882a593Smuzhiyun 			  __func__);
45*4882a593Smuzhiyun 		ret = -EINVAL;
46*4882a593Smuzhiyun 		return ret;
47*4882a593Smuzhiyun 	}
48*4882a593Smuzhiyun 
49*4882a593Smuzhiyun 	rknpu_obj = kzalloc(sizeof(*rknpu_obj), GFP_KERNEL);
50*4882a593Smuzhiyun 	if (!rknpu_obj)
51*4882a593Smuzhiyun 		return -ENOMEM;
52*4882a593Smuzhiyun 
53*4882a593Smuzhiyun 	if (args.handle > 0) {
54*4882a593Smuzhiyun 		fd = args.handle;
55*4882a593Smuzhiyun 
56*4882a593Smuzhiyun 		dmabuf = dma_buf_get(fd);
57*4882a593Smuzhiyun 		if (IS_ERR(dmabuf)) {
58*4882a593Smuzhiyun 			ret = PTR_ERR(dmabuf);
59*4882a593Smuzhiyun 			goto err_free_obj;
60*4882a593Smuzhiyun 		}
61*4882a593Smuzhiyun 
62*4882a593Smuzhiyun 		rknpu_obj->dmabuf = dmabuf;
63*4882a593Smuzhiyun 		rknpu_obj->owner = 0;
64*4882a593Smuzhiyun 	} else {
65*4882a593Smuzhiyun 		/* Start test kernel alloc/free dma buf */
66*4882a593Smuzhiyun 		dmabuf = rk_dma_heap_buffer_alloc(rknpu_dev->heap, args.size,
67*4882a593Smuzhiyun 						  O_CLOEXEC | O_RDWR, 0x0,
68*4882a593Smuzhiyun 						  dev_name(rknpu_dev->dev));
69*4882a593Smuzhiyun 		if (IS_ERR(dmabuf)) {
70*4882a593Smuzhiyun 			LOG_ERROR("dmabuf alloc failed, args.size = %llu\n",
71*4882a593Smuzhiyun 				  args.size);
72*4882a593Smuzhiyun 			ret = PTR_ERR(dmabuf);
73*4882a593Smuzhiyun 			goto err_free_obj;
74*4882a593Smuzhiyun 		}
75*4882a593Smuzhiyun 
76*4882a593Smuzhiyun 		rknpu_obj->dmabuf = dmabuf;
77*4882a593Smuzhiyun 		rknpu_obj->owner = 1;
78*4882a593Smuzhiyun 
79*4882a593Smuzhiyun 		fd = dma_buf_fd(dmabuf, O_CLOEXEC | O_RDWR);
80*4882a593Smuzhiyun 		if (fd < 0) {
81*4882a593Smuzhiyun 			LOG_ERROR("dmabuf fd get failed\n");
82*4882a593Smuzhiyun 			ret = -EFAULT;
83*4882a593Smuzhiyun 			goto err_free_dma_buf;
84*4882a593Smuzhiyun 		}
85*4882a593Smuzhiyun 	}
86*4882a593Smuzhiyun 
87*4882a593Smuzhiyun 	attachment = dma_buf_attach(dmabuf, rknpu_dev->dev);
88*4882a593Smuzhiyun 	if (IS_ERR(attachment)) {
89*4882a593Smuzhiyun 		LOG_ERROR("dma_buf_attach failed\n");
90*4882a593Smuzhiyun 		ret = PTR_ERR(attachment);
91*4882a593Smuzhiyun 		goto err_free_dma_buf;
92*4882a593Smuzhiyun 	}
93*4882a593Smuzhiyun 
94*4882a593Smuzhiyun 	table = dma_buf_map_attachment(attachment, DMA_BIDIRECTIONAL);
95*4882a593Smuzhiyun 	if (IS_ERR(table)) {
96*4882a593Smuzhiyun 		LOG_ERROR("dma_buf_attach failed\n");
97*4882a593Smuzhiyun 		dma_buf_detach(dmabuf, attachment);
98*4882a593Smuzhiyun 		ret = PTR_ERR(table);
99*4882a593Smuzhiyun 		goto err_free_dma_buf;
100*4882a593Smuzhiyun 	}
101*4882a593Smuzhiyun 
102*4882a593Smuzhiyun 	for_each_sgtable_sg(table, sgl, i) {
103*4882a593Smuzhiyun 		phys = sg_dma_address(sgl);
104*4882a593Smuzhiyun 		page = sg_page(sgl);
105*4882a593Smuzhiyun 		length = sg_dma_len(sgl);
106*4882a593Smuzhiyun 		LOG_DEBUG("%s, %d, phys: %pad, length: %u\n", __func__,
107*4882a593Smuzhiyun 			  __LINE__, &phys, length);
108*4882a593Smuzhiyun 	}
109*4882a593Smuzhiyun 
110*4882a593Smuzhiyun 	page_count = length >> PAGE_SHIFT;
111*4882a593Smuzhiyun 	pages = kmalloc_array(page_count, sizeof(struct page), GFP_KERNEL);
112*4882a593Smuzhiyun 	if (!pages) {
113*4882a593Smuzhiyun 		LOG_ERROR("alloc pages failed\n");
114*4882a593Smuzhiyun 		ret = -ENOMEM;
115*4882a593Smuzhiyun 		goto err_detach_dma_buf;
116*4882a593Smuzhiyun 	}
117*4882a593Smuzhiyun 
118*4882a593Smuzhiyun 	for (i = 0; i < page_count; i++)
119*4882a593Smuzhiyun 		pages[i] = &page[i];
120*4882a593Smuzhiyun 
121*4882a593Smuzhiyun 	rknpu_obj->kv_addr = vmap(pages, page_count, VM_MAP, PAGE_KERNEL);
122*4882a593Smuzhiyun 	if (!rknpu_obj->kv_addr) {
123*4882a593Smuzhiyun 		LOG_ERROR("vmap pages addr failed\n");
124*4882a593Smuzhiyun 		ret = -ENOMEM;
125*4882a593Smuzhiyun 		goto err_free_pages;
126*4882a593Smuzhiyun 	}
127*4882a593Smuzhiyun 
128*4882a593Smuzhiyun 	rknpu_obj->size = PAGE_ALIGN(args.size);
129*4882a593Smuzhiyun 	rknpu_obj->dma_addr = phys;
130*4882a593Smuzhiyun 	rknpu_obj->sgt = table;
131*4882a593Smuzhiyun 
132*4882a593Smuzhiyun 	args.size = rknpu_obj->size;
133*4882a593Smuzhiyun 	args.obj_addr = (__u64)(uintptr_t)rknpu_obj;
134*4882a593Smuzhiyun 	args.dma_addr = rknpu_obj->dma_addr;
135*4882a593Smuzhiyun 	args.handle = fd;
136*4882a593Smuzhiyun 
137*4882a593Smuzhiyun 	LOG_DEBUG(
138*4882a593Smuzhiyun 		"args.handle: %d, args.size: %lld, rknpu_obj: %#llx, rknpu_obj->dma_addr: %#llx\n",
139*4882a593Smuzhiyun 		args.handle, args.size, (__u64)(uintptr_t)rknpu_obj,
140*4882a593Smuzhiyun 		(__u64)rknpu_obj->dma_addr);
141*4882a593Smuzhiyun 
142*4882a593Smuzhiyun 	if (unlikely(copy_to_user((struct rknpu_mem_create *)data, &args,
143*4882a593Smuzhiyun 				  sizeof(struct rknpu_mem_create)))) {
144*4882a593Smuzhiyun 		LOG_ERROR("%s: copy_to_user failed\n", __func__);
145*4882a593Smuzhiyun 		ret = -EFAULT;
146*4882a593Smuzhiyun 		goto err_unmap_kv_addr;
147*4882a593Smuzhiyun 	}
148*4882a593Smuzhiyun 
149*4882a593Smuzhiyun 	kfree(pages);
150*4882a593Smuzhiyun 	dma_buf_unmap_attachment(attachment, table, DMA_BIDIRECTIONAL);
151*4882a593Smuzhiyun 	dma_buf_detach(dmabuf, attachment);
152*4882a593Smuzhiyun 
153*4882a593Smuzhiyun 	spin_lock(&rknpu_dev->lock);
154*4882a593Smuzhiyun 
155*4882a593Smuzhiyun 	session = file->private_data;
156*4882a593Smuzhiyun 	if (!session) {
157*4882a593Smuzhiyun 		spin_unlock(&rknpu_dev->lock);
158*4882a593Smuzhiyun 		ret = -EFAULT;
159*4882a593Smuzhiyun 		goto err_unmap_kv_addr;
160*4882a593Smuzhiyun 	}
161*4882a593Smuzhiyun 	list_add_tail(&rknpu_obj->head, &session->list);
162*4882a593Smuzhiyun 
163*4882a593Smuzhiyun 	spin_unlock(&rknpu_dev->lock);
164*4882a593Smuzhiyun 
165*4882a593Smuzhiyun 	return 0;
166*4882a593Smuzhiyun 
167*4882a593Smuzhiyun err_unmap_kv_addr:
168*4882a593Smuzhiyun 	vunmap(rknpu_obj->kv_addr);
169*4882a593Smuzhiyun 	rknpu_obj->kv_addr = NULL;
170*4882a593Smuzhiyun 
171*4882a593Smuzhiyun err_free_pages:
172*4882a593Smuzhiyun 	kfree(pages);
173*4882a593Smuzhiyun 
174*4882a593Smuzhiyun err_detach_dma_buf:
175*4882a593Smuzhiyun 	dma_buf_unmap_attachment(attachment, table, DMA_BIDIRECTIONAL);
176*4882a593Smuzhiyun 	dma_buf_detach(dmabuf, attachment);
177*4882a593Smuzhiyun 
178*4882a593Smuzhiyun err_free_dma_buf:
179*4882a593Smuzhiyun 	if (rknpu_obj->owner)
180*4882a593Smuzhiyun 		rk_dma_heap_buffer_free(dmabuf);
181*4882a593Smuzhiyun 	else
182*4882a593Smuzhiyun 		dma_buf_put(dmabuf);
183*4882a593Smuzhiyun 
184*4882a593Smuzhiyun err_free_obj:
185*4882a593Smuzhiyun 	kfree(rknpu_obj);
186*4882a593Smuzhiyun 
187*4882a593Smuzhiyun 	return ret;
188*4882a593Smuzhiyun }
189*4882a593Smuzhiyun 
rknpu_mem_destroy_ioctl(struct rknpu_device * rknpu_dev,unsigned long data,struct file * file)190*4882a593Smuzhiyun int rknpu_mem_destroy_ioctl(struct rknpu_device *rknpu_dev, unsigned long data,
191*4882a593Smuzhiyun 			    struct file *file)
192*4882a593Smuzhiyun {
193*4882a593Smuzhiyun 	struct rknpu_mem_object *rknpu_obj, *entry, *q;
194*4882a593Smuzhiyun 	struct rknpu_session *session = NULL;
195*4882a593Smuzhiyun 	struct rknpu_mem_destroy args;
196*4882a593Smuzhiyun 	int ret = -EFAULT;
197*4882a593Smuzhiyun 
198*4882a593Smuzhiyun 	if (unlikely(copy_from_user(&args, (struct rknpu_mem_destroy *)data,
199*4882a593Smuzhiyun 				    sizeof(struct rknpu_mem_destroy)))) {
200*4882a593Smuzhiyun 		LOG_ERROR("%s: copy_from_user failed\n", __func__);
201*4882a593Smuzhiyun 		ret = -EFAULT;
202*4882a593Smuzhiyun 		return ret;
203*4882a593Smuzhiyun 	}
204*4882a593Smuzhiyun 
205*4882a593Smuzhiyun 	if (!kern_addr_valid(args.obj_addr)) {
206*4882a593Smuzhiyun 		LOG_ERROR("%s: invalid obj_addr: %#llx\n", __func__,
207*4882a593Smuzhiyun 			  (__u64)(uintptr_t)args.obj_addr);
208*4882a593Smuzhiyun 		ret = -EINVAL;
209*4882a593Smuzhiyun 		return ret;
210*4882a593Smuzhiyun 	}
211*4882a593Smuzhiyun 
212*4882a593Smuzhiyun 	rknpu_obj = (struct rknpu_mem_object *)(uintptr_t)args.obj_addr;
213*4882a593Smuzhiyun 	LOG_DEBUG(
214*4882a593Smuzhiyun 		"free args.handle: %d, rknpu_obj: %#llx, rknpu_obj->dma_addr: %#llx\n",
215*4882a593Smuzhiyun 		args.handle, (__u64)(uintptr_t)rknpu_obj,
216*4882a593Smuzhiyun 		(__u64)rknpu_obj->dma_addr);
217*4882a593Smuzhiyun 
218*4882a593Smuzhiyun 	spin_lock(&rknpu_dev->lock);
219*4882a593Smuzhiyun 	session = file->private_data;
220*4882a593Smuzhiyun 	if (!session) {
221*4882a593Smuzhiyun 		spin_unlock(&rknpu_dev->lock);
222*4882a593Smuzhiyun 		ret = -EFAULT;
223*4882a593Smuzhiyun 		return ret;
224*4882a593Smuzhiyun 	}
225*4882a593Smuzhiyun 	list_for_each_entry_safe(entry, q, &session->list, head) {
226*4882a593Smuzhiyun 		if (entry == rknpu_obj) {
227*4882a593Smuzhiyun 			list_del(&entry->head);
228*4882a593Smuzhiyun 			break;
229*4882a593Smuzhiyun 		}
230*4882a593Smuzhiyun 	}
231*4882a593Smuzhiyun 	spin_unlock(&rknpu_dev->lock);
232*4882a593Smuzhiyun 
233*4882a593Smuzhiyun 	if (rknpu_obj == entry) {
234*4882a593Smuzhiyun 		vunmap(rknpu_obj->kv_addr);
235*4882a593Smuzhiyun 		rknpu_obj->kv_addr = NULL;
236*4882a593Smuzhiyun 
237*4882a593Smuzhiyun 		if (!rknpu_obj->owner)
238*4882a593Smuzhiyun 			dma_buf_put(rknpu_obj->dmabuf);
239*4882a593Smuzhiyun 
240*4882a593Smuzhiyun 		kfree(rknpu_obj);
241*4882a593Smuzhiyun 	}
242*4882a593Smuzhiyun 
243*4882a593Smuzhiyun 	return 0;
244*4882a593Smuzhiyun }
245*4882a593Smuzhiyun 
246*4882a593Smuzhiyun /*
247*4882a593Smuzhiyun  * begin cpu access => for_cpu = true
248*4882a593Smuzhiyun  * end cpu access => for_cpu = false
249*4882a593Smuzhiyun  */
rknpu_dma_buf_sync(struct rknpu_device * rknpu_dev,struct rknpu_mem_object * rknpu_obj,u32 offset,u32 length,enum dma_data_direction dir,bool for_cpu)250*4882a593Smuzhiyun static void __maybe_unused rknpu_dma_buf_sync(
251*4882a593Smuzhiyun 	struct rknpu_device *rknpu_dev, struct rknpu_mem_object *rknpu_obj,
252*4882a593Smuzhiyun 	u32 offset, u32 length, enum dma_data_direction dir, bool for_cpu)
253*4882a593Smuzhiyun {
254*4882a593Smuzhiyun 	struct device *dev = rknpu_dev->dev;
255*4882a593Smuzhiyun 	struct sg_table *sgt = rknpu_obj->sgt;
256*4882a593Smuzhiyun 	struct scatterlist *sg = sgt->sgl;
257*4882a593Smuzhiyun 	dma_addr_t sg_dma_addr = sg_dma_address(sg);
258*4882a593Smuzhiyun 	unsigned int len = 0;
259*4882a593Smuzhiyun 	int i;
260*4882a593Smuzhiyun 
261*4882a593Smuzhiyun 	for_each_sgtable_sg(sgt, sg, i) {
262*4882a593Smuzhiyun 		unsigned int sg_offset, sg_left, size = 0;
263*4882a593Smuzhiyun 
264*4882a593Smuzhiyun 		len += sg->length;
265*4882a593Smuzhiyun 		if (len <= offset) {
266*4882a593Smuzhiyun 			sg_dma_addr += sg->length;
267*4882a593Smuzhiyun 			continue;
268*4882a593Smuzhiyun 		}
269*4882a593Smuzhiyun 
270*4882a593Smuzhiyun 		sg_left = len - offset;
271*4882a593Smuzhiyun 		sg_offset = sg->length - sg_left;
272*4882a593Smuzhiyun 
273*4882a593Smuzhiyun 		size = (length < sg_left) ? length : sg_left;
274*4882a593Smuzhiyun 
275*4882a593Smuzhiyun 		if (for_cpu)
276*4882a593Smuzhiyun 			dma_sync_single_range_for_cpu(dev, sg_dma_addr,
277*4882a593Smuzhiyun 						      sg_offset, size, dir);
278*4882a593Smuzhiyun 		else
279*4882a593Smuzhiyun 			dma_sync_single_range_for_device(dev, sg_dma_addr,
280*4882a593Smuzhiyun 							 sg_offset, size, dir);
281*4882a593Smuzhiyun 
282*4882a593Smuzhiyun 		offset += size;
283*4882a593Smuzhiyun 		length -= size;
284*4882a593Smuzhiyun 		sg_dma_addr += sg->length;
285*4882a593Smuzhiyun 
286*4882a593Smuzhiyun 		if (length == 0)
287*4882a593Smuzhiyun 			break;
288*4882a593Smuzhiyun 	}
289*4882a593Smuzhiyun }
290*4882a593Smuzhiyun 
rknpu_mem_sync_ioctl(struct rknpu_device * rknpu_dev,unsigned long data)291*4882a593Smuzhiyun int rknpu_mem_sync_ioctl(struct rknpu_device *rknpu_dev, unsigned long data)
292*4882a593Smuzhiyun {
293*4882a593Smuzhiyun 	struct rknpu_mem_object *rknpu_obj = NULL;
294*4882a593Smuzhiyun 	struct rknpu_mem_sync args;
295*4882a593Smuzhiyun 	struct dma_buf *dmabuf;
296*4882a593Smuzhiyun 	int ret = -EFAULT;
297*4882a593Smuzhiyun 
298*4882a593Smuzhiyun 	if (unlikely(copy_from_user(&args, (struct rknpu_mem_sync *)data,
299*4882a593Smuzhiyun 				    sizeof(struct rknpu_mem_sync)))) {
300*4882a593Smuzhiyun 		LOG_ERROR("%s: copy_from_user failed\n", __func__);
301*4882a593Smuzhiyun 		ret = -EFAULT;
302*4882a593Smuzhiyun 		return ret;
303*4882a593Smuzhiyun 	}
304*4882a593Smuzhiyun 
305*4882a593Smuzhiyun 	if (!kern_addr_valid(args.obj_addr)) {
306*4882a593Smuzhiyun 		LOG_ERROR("%s: invalid obj_addr: %#llx\n", __func__,
307*4882a593Smuzhiyun 			  (__u64)(uintptr_t)args.obj_addr);
308*4882a593Smuzhiyun 		ret = -EINVAL;
309*4882a593Smuzhiyun 		return ret;
310*4882a593Smuzhiyun 	}
311*4882a593Smuzhiyun 
312*4882a593Smuzhiyun 	rknpu_obj = (struct rknpu_mem_object *)(uintptr_t)args.obj_addr;
313*4882a593Smuzhiyun 	dmabuf = rknpu_obj->dmabuf;
314*4882a593Smuzhiyun 
315*4882a593Smuzhiyun #ifndef CONFIG_DMABUF_PARTIAL
316*4882a593Smuzhiyun 	if (args.flags & RKNPU_MEM_SYNC_TO_DEVICE) {
317*4882a593Smuzhiyun 		rknpu_dma_buf_sync(rknpu_dev, rknpu_obj, args.offset, args.size,
318*4882a593Smuzhiyun 				   DMA_TO_DEVICE, false);
319*4882a593Smuzhiyun 	}
320*4882a593Smuzhiyun 	if (args.flags & RKNPU_MEM_SYNC_FROM_DEVICE) {
321*4882a593Smuzhiyun 		rknpu_dma_buf_sync(rknpu_dev, rknpu_obj, args.offset, args.size,
322*4882a593Smuzhiyun 				   DMA_FROM_DEVICE, true);
323*4882a593Smuzhiyun 	}
324*4882a593Smuzhiyun #else
325*4882a593Smuzhiyun 	if (args.flags & RKNPU_MEM_SYNC_TO_DEVICE) {
326*4882a593Smuzhiyun 		dmabuf->ops->end_cpu_access_partial(dmabuf, DMA_TO_DEVICE,
327*4882a593Smuzhiyun 						    args.offset, args.size);
328*4882a593Smuzhiyun 	}
329*4882a593Smuzhiyun 	if (args.flags & RKNPU_MEM_SYNC_FROM_DEVICE) {
330*4882a593Smuzhiyun 		dmabuf->ops->begin_cpu_access_partial(dmabuf, DMA_FROM_DEVICE,
331*4882a593Smuzhiyun 						      args.offset, args.size);
332*4882a593Smuzhiyun 	}
333*4882a593Smuzhiyun #endif
334*4882a593Smuzhiyun 
335*4882a593Smuzhiyun 	return 0;
336*4882a593Smuzhiyun }
337