1*4882a593Smuzhiyun /* SPDX-License-Identifier: GPL-2.0 */
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun * Copyright (C) Rockchip Electronics Co., Ltd.
4*4882a593Smuzhiyun *
5*4882a593Smuzhiyun * Author: Huang Lee <Putin.li@rock-chips.com>
6*4882a593Smuzhiyun */
7*4882a593Smuzhiyun
8*4882a593Smuzhiyun #ifndef __LINUX_RVE_DRV_H_
9*4882a593Smuzhiyun #define __LINUX_RVE_DRV_H_
10*4882a593Smuzhiyun
11*4882a593Smuzhiyun #include <linux/clk.h>
12*4882a593Smuzhiyun #include <linux/completion.h>
13*4882a593Smuzhiyun #include <linux/debugfs.h>
14*4882a593Smuzhiyun #include <linux/delay.h>
15*4882a593Smuzhiyun #include <linux/device.h>
16*4882a593Smuzhiyun #include <linux/dma-buf-cache.h>
17*4882a593Smuzhiyun #include <linux/dma-mapping.h>
18*4882a593Smuzhiyun #include <linux/err.h>
19*4882a593Smuzhiyun #include <linux/fb.h>
20*4882a593Smuzhiyun #include <linux/fs.h>
21*4882a593Smuzhiyun #include <linux/init.h>
22*4882a593Smuzhiyun #include <linux/interrupt.h>
23*4882a593Smuzhiyun #include <linux/io.h>
24*4882a593Smuzhiyun #include <linux/irq.h>
25*4882a593Smuzhiyun #include <linux/kernel.h>
26*4882a593Smuzhiyun #include <linux/kref.h>
27*4882a593Smuzhiyun #include <linux/miscdevice.h>
28*4882a593Smuzhiyun #include <linux/module.h>
29*4882a593Smuzhiyun #include <linux/mutex.h>
30*4882a593Smuzhiyun #include <linux/of_device.h>
31*4882a593Smuzhiyun #include <linux/platform_device.h>
32*4882a593Smuzhiyun #include <linux/poll.h>
33*4882a593Smuzhiyun #include <linux/regulator/consumer.h>
34*4882a593Smuzhiyun #include <linux/scatterlist.h>
35*4882a593Smuzhiyun #include <linux/sched.h>
36*4882a593Smuzhiyun #include <linux/slab.h>
37*4882a593Smuzhiyun #include <linux/spinlock.h>
38*4882a593Smuzhiyun #include <linux/syscalls.h>
39*4882a593Smuzhiyun #include <linux/time.h>
40*4882a593Smuzhiyun #include <linux/timer.h>
41*4882a593Smuzhiyun #include <linux/uaccess.h>
42*4882a593Smuzhiyun #include <linux/version.h>
43*4882a593Smuzhiyun #include <linux/wait.h>
44*4882a593Smuzhiyun #include <linux/wakelock.h>
45*4882a593Smuzhiyun #include <linux/pm_runtime.h>
46*4882a593Smuzhiyun #include <linux/sched/mm.h>
47*4882a593Smuzhiyun
48*4882a593Smuzhiyun #include <asm/cacheflush.h>
49*4882a593Smuzhiyun
50*4882a593Smuzhiyun #include <linux/iommu.h>
51*4882a593Smuzhiyun #include <linux/iova.h>
52*4882a593Smuzhiyun #include <linux/dma-iommu.h>
53*4882a593Smuzhiyun #include <linux/dma-map-ops.h>
54*4882a593Smuzhiyun #include <linux/hrtimer.h>
55*4882a593Smuzhiyun
56*4882a593Smuzhiyun #include "rve_debugger.h"
57*4882a593Smuzhiyun #include "rve.h"
58*4882a593Smuzhiyun
59*4882a593Smuzhiyun /* sample interval: 1000ms */
60*4882a593Smuzhiyun #define RVE_LOAD_INTERVAL 1000000000
61*4882a593Smuzhiyun
62*4882a593Smuzhiyun /* Driver information */
63*4882a593Smuzhiyun #define DRIVER_DESC "RVE Device Driver"
64*4882a593Smuzhiyun #define DRIVER_NAME "rve"
65*4882a593Smuzhiyun
66*4882a593Smuzhiyun #define STR_HELPER(x) #x
67*4882a593Smuzhiyun #define STR(x) STR_HELPER(x)
68*4882a593Smuzhiyun
69*4882a593Smuzhiyun #define RVE_MAJOR_VERSION_MASK (0x0000FF00)
70*4882a593Smuzhiyun #define RVE_MINOR_VERSION_MASK (0x000000FF)
71*4882a593Smuzhiyun #define RVE_PROD_NUM_MASK (0xFFFF0000)
72*4882a593Smuzhiyun
73*4882a593Smuzhiyun #define DRIVER_MAJOR_VERSION 1
74*4882a593Smuzhiyun #define DRIVER_MINOR_VERSION 0
75*4882a593Smuzhiyun #define DRIVER_REVISION_VERSION 4
76*4882a593Smuzhiyun
77*4882a593Smuzhiyun #define DRIVER_VERSION (STR(DRIVER_MAJOR_VERSION) "." STR(DRIVER_MINOR_VERSION) \
78*4882a593Smuzhiyun "." STR(DRIVER_REVISION_VERSION))
79*4882a593Smuzhiyun
80*4882a593Smuzhiyun /* time limit */
81*4882a593Smuzhiyun #define RVE_ASYNC_TIMEOUT_DELAY 500
82*4882a593Smuzhiyun #define RVE_SYNC_TIMEOUT_DELAY HZ
83*4882a593Smuzhiyun #define RVE_RESET_TIMEOUT 10000
84*4882a593Smuzhiyun
85*4882a593Smuzhiyun #define RVE_BUFFER_POOL_MAX_SIZE 64
86*4882a593Smuzhiyun #define RVE_MAX_SCHEDULER 1
87*4882a593Smuzhiyun
88*4882a593Smuzhiyun #define RVE_MAX_BUS_CLK 10
89*4882a593Smuzhiyun #define RVE_MAX_PID_INFO 10
90*4882a593Smuzhiyun
91*4882a593Smuzhiyun extern struct rve_drvdata_t *rve_drvdata;
92*4882a593Smuzhiyun
93*4882a593Smuzhiyun enum {
94*4882a593Smuzhiyun RVE_SCHEDULER_CORE0 = 1,
95*4882a593Smuzhiyun RVE_NONE_CORE = 0,
96*4882a593Smuzhiyun };
97*4882a593Smuzhiyun
98*4882a593Smuzhiyun enum {
99*4882a593Smuzhiyun RVE_CMD_SLAVE = 1,
100*4882a593Smuzhiyun RVE_CMD_MASTER = 2,
101*4882a593Smuzhiyun };
102*4882a593Smuzhiyun
103*4882a593Smuzhiyun struct rve_fence_context {
104*4882a593Smuzhiyun unsigned int context;
105*4882a593Smuzhiyun unsigned int seqno;
106*4882a593Smuzhiyun spinlock_t spinlock;
107*4882a593Smuzhiyun };
108*4882a593Smuzhiyun
109*4882a593Smuzhiyun struct rve_fence_waiter {
110*4882a593Smuzhiyun /* Base sync driver waiter structure */
111*4882a593Smuzhiyun struct dma_fence_cb waiter;
112*4882a593Smuzhiyun
113*4882a593Smuzhiyun struct rve_job *job;
114*4882a593Smuzhiyun };
115*4882a593Smuzhiyun
116*4882a593Smuzhiyun struct rve_scheduler_t;
117*4882a593Smuzhiyun struct rve_internal_ctx_t;
118*4882a593Smuzhiyun
119*4882a593Smuzhiyun struct rve_session {
120*4882a593Smuzhiyun int id;
121*4882a593Smuzhiyun
122*4882a593Smuzhiyun pid_t tgid;
123*4882a593Smuzhiyun };
124*4882a593Smuzhiyun
125*4882a593Smuzhiyun struct rve_job {
126*4882a593Smuzhiyun struct list_head head;
127*4882a593Smuzhiyun struct rve_scheduler_t *scheduler;
128*4882a593Smuzhiyun struct rve_session *session;
129*4882a593Smuzhiyun
130*4882a593Smuzhiyun struct rve_cmd_reg_array_t *regcmd_data;
131*4882a593Smuzhiyun
132*4882a593Smuzhiyun struct rve_internal_ctx_t *ctx;
133*4882a593Smuzhiyun
134*4882a593Smuzhiyun /* for rve virtual_address */
135*4882a593Smuzhiyun struct mm_struct *mm;
136*4882a593Smuzhiyun
137*4882a593Smuzhiyun struct dma_fence *out_fence;
138*4882a593Smuzhiyun struct dma_fence *in_fence;
139*4882a593Smuzhiyun spinlock_t fence_lock;
140*4882a593Smuzhiyun ktime_t timestamp;
141*4882a593Smuzhiyun ktime_t hw_running_time;
142*4882a593Smuzhiyun ktime_t hw_recoder_time;
143*4882a593Smuzhiyun unsigned int flags;
144*4882a593Smuzhiyun
145*4882a593Smuzhiyun int priority;
146*4882a593Smuzhiyun int core;
147*4882a593Smuzhiyun int ret;
148*4882a593Smuzhiyun pid_t pid;
149*4882a593Smuzhiyun };
150*4882a593Smuzhiyun
151*4882a593Smuzhiyun struct rve_backend_ops {
152*4882a593Smuzhiyun int (*get_version)(struct rve_scheduler_t *scheduler);
153*4882a593Smuzhiyun int (*set_reg)(struct rve_job *job, struct rve_scheduler_t *scheduler);
154*4882a593Smuzhiyun int (*init_reg)(struct rve_job *job);
155*4882a593Smuzhiyun void (*soft_reset)(struct rve_scheduler_t *scheduler);
156*4882a593Smuzhiyun };
157*4882a593Smuzhiyun
158*4882a593Smuzhiyun struct rve_timer {
159*4882a593Smuzhiyun u32 busy_time;
160*4882a593Smuzhiyun u32 busy_time_record;
161*4882a593Smuzhiyun };
162*4882a593Smuzhiyun
163*4882a593Smuzhiyun struct rve_sche_pid_info_t {
164*4882a593Smuzhiyun pid_t pid;
165*4882a593Smuzhiyun /* hw total use time, per hrtimer */
166*4882a593Smuzhiyun u32 hw_time_total;
167*4882a593Smuzhiyun
168*4882a593Smuzhiyun uint32_t last_job_rd_bandwidth;
169*4882a593Smuzhiyun uint32_t last_job_wr_bandwidth;
170*4882a593Smuzhiyun uint32_t last_job_cycle_cnt;
171*4882a593Smuzhiyun };
172*4882a593Smuzhiyun
173*4882a593Smuzhiyun struct rve_sche_session_info_t {
174*4882a593Smuzhiyun struct rve_sche_pid_info_t pid_info[RVE_MAX_PID_INFO];
175*4882a593Smuzhiyun
176*4882a593Smuzhiyun int pd_refcount;
177*4882a593Smuzhiyun
178*4882a593Smuzhiyun /* the bandwidth of total read bytes, per hrtimer */
179*4882a593Smuzhiyun uint32_t rd_bandwidth;
180*4882a593Smuzhiyun /* the bandwidth of total write bytes, per hrtimer */
181*4882a593Smuzhiyun uint32_t wr_bandwidth;
182*4882a593Smuzhiyun /* the total running cycle of current frame, per hrtimer */
183*4882a593Smuzhiyun uint32_t cycle_cnt;
184*4882a593Smuzhiyun /* total interrupt count */
185*4882a593Smuzhiyun uint64_t total_int_cnt;
186*4882a593Smuzhiyun };
187*4882a593Smuzhiyun
188*4882a593Smuzhiyun struct rve_scheduler_t {
189*4882a593Smuzhiyun struct device *dev;
190*4882a593Smuzhiyun void __iomem *rve_base;
191*4882a593Smuzhiyun
192*4882a593Smuzhiyun struct clk *clks[RVE_MAX_BUS_CLK];
193*4882a593Smuzhiyun int num_clks;
194*4882a593Smuzhiyun
195*4882a593Smuzhiyun struct rve_job *running_job;
196*4882a593Smuzhiyun struct list_head todo_list;
197*4882a593Smuzhiyun spinlock_t irq_lock;
198*4882a593Smuzhiyun wait_queue_head_t job_done_wq;
199*4882a593Smuzhiyun const struct rve_backend_ops *ops;
200*4882a593Smuzhiyun const struct rve_hw_data *data;
201*4882a593Smuzhiyun int job_count;
202*4882a593Smuzhiyun int irq;
203*4882a593Smuzhiyun struct rve_version_t version;
204*4882a593Smuzhiyun int core;
205*4882a593Smuzhiyun
206*4882a593Smuzhiyun struct rve_timer timer;
207*4882a593Smuzhiyun
208*4882a593Smuzhiyun struct rve_sche_session_info_t session;
209*4882a593Smuzhiyun };
210*4882a593Smuzhiyun
211*4882a593Smuzhiyun struct rve_cmd_reg_array_t {
212*4882a593Smuzhiyun uint32_t cmd_reg[58];
213*4882a593Smuzhiyun };
214*4882a593Smuzhiyun
215*4882a593Smuzhiyun struct rve_ctx_debug_info_t {
216*4882a593Smuzhiyun pid_t pid;
217*4882a593Smuzhiyun u32 timestamp;
218*4882a593Smuzhiyun /* hw total use time, per hrtimer */
219*4882a593Smuzhiyun u32 hw_time_total;
220*4882a593Smuzhiyun /* last job use time, per hrtimer*/
221*4882a593Smuzhiyun u32 last_job_use_time;
222*4882a593Smuzhiyun /* last job hardware use time, per hrtimer*/
223*4882a593Smuzhiyun u32 last_job_hw_use_time;
224*4882a593Smuzhiyun /* the most time-consuming job, per hrtimer */
225*4882a593Smuzhiyun u32 max_cost_time_per_sec;
226*4882a593Smuzhiyun };
227*4882a593Smuzhiyun
228*4882a593Smuzhiyun struct rve_internal_ctx_t {
229*4882a593Smuzhiyun struct rve_scheduler_t *scheduler;
230*4882a593Smuzhiyun struct rve_session *session;
231*4882a593Smuzhiyun
232*4882a593Smuzhiyun struct rve_cmd_reg_array_t *regcmd_data;
233*4882a593Smuzhiyun uint32_t cmd_num;
234*4882a593Smuzhiyun
235*4882a593Smuzhiyun uint32_t sync_mode;
236*4882a593Smuzhiyun int flags;
237*4882a593Smuzhiyun int id;
238*4882a593Smuzhiyun
239*4882a593Smuzhiyun uint32_t running_job_count;
240*4882a593Smuzhiyun uint32_t finished_job_count;
241*4882a593Smuzhiyun bool is_running;
242*4882a593Smuzhiyun
243*4882a593Smuzhiyun uint32_t disable_auto_cancel;
244*4882a593Smuzhiyun
245*4882a593Smuzhiyun int priority;
246*4882a593Smuzhiyun int32_t out_fence_fd;
247*4882a593Smuzhiyun int32_t in_fence_fd;
248*4882a593Smuzhiyun
249*4882a593Smuzhiyun struct dma_fence *out_fence;
250*4882a593Smuzhiyun
251*4882a593Smuzhiyun spinlock_t lock;
252*4882a593Smuzhiyun struct kref refcount;
253*4882a593Smuzhiyun
254*4882a593Smuzhiyun /* debug info */
255*4882a593Smuzhiyun struct rve_ctx_debug_info_t debug_info;
256*4882a593Smuzhiyun
257*4882a593Smuzhiyun /* TODO: add some common work */
258*4882a593Smuzhiyun };
259*4882a593Smuzhiyun
260*4882a593Smuzhiyun struct rve_pending_ctx_manager {
261*4882a593Smuzhiyun spinlock_t lock;
262*4882a593Smuzhiyun
263*4882a593Smuzhiyun /*
264*4882a593Smuzhiyun * @ctx_id_idr:
265*4882a593Smuzhiyun *
266*4882a593Smuzhiyun * Mapping of ctx id to object pointers. Used by the GEM
267*4882a593Smuzhiyun * subsystem. Protected by @lock.
268*4882a593Smuzhiyun */
269*4882a593Smuzhiyun struct idr ctx_id_idr;
270*4882a593Smuzhiyun
271*4882a593Smuzhiyun int ctx_count;
272*4882a593Smuzhiyun };
273*4882a593Smuzhiyun
274*4882a593Smuzhiyun struct rve_session_manager {
275*4882a593Smuzhiyun struct mutex lock;
276*4882a593Smuzhiyun
277*4882a593Smuzhiyun struct idr ctx_id_idr;
278*4882a593Smuzhiyun
279*4882a593Smuzhiyun int session_cnt;
280*4882a593Smuzhiyun };
281*4882a593Smuzhiyun
282*4882a593Smuzhiyun struct rve_drvdata_t {
283*4882a593Smuzhiyun struct rve_fence_context *fence_ctx;
284*4882a593Smuzhiyun
285*4882a593Smuzhiyun /* used by rve2's mmu lock */
286*4882a593Smuzhiyun struct mutex lock;
287*4882a593Smuzhiyun
288*4882a593Smuzhiyun struct rve_scheduler_t *scheduler[RVE_MAX_SCHEDULER];
289*4882a593Smuzhiyun int num_of_scheduler;
290*4882a593Smuzhiyun
291*4882a593Smuzhiyun struct delayed_work power_off_work;
292*4882a593Smuzhiyun struct wake_lock wake_lock;
293*4882a593Smuzhiyun
294*4882a593Smuzhiyun struct rve_mm *mm;
295*4882a593Smuzhiyun
296*4882a593Smuzhiyun /* rve_job pending manager, import by RVE_IOC_START_CONFIG */
297*4882a593Smuzhiyun struct rve_pending_ctx_manager *pend_ctx_manager;
298*4882a593Smuzhiyun
299*4882a593Smuzhiyun struct rve_session_manager *session_manager;
300*4882a593Smuzhiyun
301*4882a593Smuzhiyun #ifdef CONFIG_ROCKCHIP_RVE_DEBUGGER
302*4882a593Smuzhiyun struct rve_debugger *debugger;
303*4882a593Smuzhiyun #endif
304*4882a593Smuzhiyun };
305*4882a593Smuzhiyun
306*4882a593Smuzhiyun struct rve_irqs_data_t {
307*4882a593Smuzhiyun const char *name;
308*4882a593Smuzhiyun irqreturn_t (*irq_hdl)(int irq, void *ctx);
309*4882a593Smuzhiyun irqreturn_t (*irq_thread)(int irq, void *ctx);
310*4882a593Smuzhiyun };
311*4882a593Smuzhiyun
312*4882a593Smuzhiyun struct rve_match_data_t {
313*4882a593Smuzhiyun const char * const *clks;
314*4882a593Smuzhiyun int num_clks;
315*4882a593Smuzhiyun const struct rve_irqs_data_t *irqs;
316*4882a593Smuzhiyun int num_irqs;
317*4882a593Smuzhiyun };
318*4882a593Smuzhiyun
rve_read(int offset,struct rve_scheduler_t * scheduler)319*4882a593Smuzhiyun static inline int rve_read(int offset, struct rve_scheduler_t *scheduler)
320*4882a593Smuzhiyun {
321*4882a593Smuzhiyun return readl(scheduler->rve_base + offset);
322*4882a593Smuzhiyun }
323*4882a593Smuzhiyun
rve_write(int value,int offset,struct rve_scheduler_t * scheduler)324*4882a593Smuzhiyun static inline void rve_write(int value, int offset, struct rve_scheduler_t *scheduler)
325*4882a593Smuzhiyun {
326*4882a593Smuzhiyun writel(value, scheduler->rve_base + offset);
327*4882a593Smuzhiyun }
328*4882a593Smuzhiyun
329*4882a593Smuzhiyun int rve_power_enable(struct rve_scheduler_t *scheduler);
330*4882a593Smuzhiyun int rve_power_disable(struct rve_scheduler_t *scheduler);
331*4882a593Smuzhiyun
332*4882a593Smuzhiyun #endif /* __LINUX_RVE_FENCE_H_ */
333