1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0
2*4882a593Smuzhiyun /* Copyright (C) 2019-2020 ARM Limited or its affiliates. */
3*4882a593Smuzhiyun
4*4882a593Smuzhiyun #include <linux/kernel.h>
5*4882a593Smuzhiyun #include <linux/module.h>
6*4882a593Smuzhiyun #include <linux/clk.h>
7*4882a593Smuzhiyun #include <linux/hw_random.h>
8*4882a593Smuzhiyun #include <linux/io.h>
9*4882a593Smuzhiyun #include <linux/platform_device.h>
10*4882a593Smuzhiyun #include <linux/pm_runtime.h>
11*4882a593Smuzhiyun #include <linux/interrupt.h>
12*4882a593Smuzhiyun #include <linux/irqreturn.h>
13*4882a593Smuzhiyun #include <linux/workqueue.h>
14*4882a593Smuzhiyun #include <linux/circ_buf.h>
15*4882a593Smuzhiyun #include <linux/completion.h>
16*4882a593Smuzhiyun #include <linux/of.h>
17*4882a593Smuzhiyun #include <linux/bitfield.h>
18*4882a593Smuzhiyun #include <linux/fips.h>
19*4882a593Smuzhiyun
20*4882a593Smuzhiyun #include "cctrng.h"
21*4882a593Smuzhiyun
22*4882a593Smuzhiyun #define CC_REG_LOW(name) (name ## _BIT_SHIFT)
23*4882a593Smuzhiyun #define CC_REG_HIGH(name) (CC_REG_LOW(name) + name ## _BIT_SIZE - 1)
24*4882a593Smuzhiyun #define CC_GENMASK(name) GENMASK(CC_REG_HIGH(name), CC_REG_LOW(name))
25*4882a593Smuzhiyun
26*4882a593Smuzhiyun #define CC_REG_FLD_GET(reg_name, fld_name, reg_val) \
27*4882a593Smuzhiyun (FIELD_GET(CC_GENMASK(CC_ ## reg_name ## _ ## fld_name), reg_val))
28*4882a593Smuzhiyun
29*4882a593Smuzhiyun #define CC_HW_RESET_LOOP_COUNT 10
30*4882a593Smuzhiyun #define CC_TRNG_SUSPEND_TIMEOUT 3000
31*4882a593Smuzhiyun
32*4882a593Smuzhiyun /* data circular buffer in words must be:
33*4882a593Smuzhiyun * - of a power-of-2 size (limitation of circ_buf.h macros)
34*4882a593Smuzhiyun * - at least 6, the size generated in the EHR according to HW implementation
35*4882a593Smuzhiyun */
36*4882a593Smuzhiyun #define CCTRNG_DATA_BUF_WORDS 32
37*4882a593Smuzhiyun
38*4882a593Smuzhiyun /* The timeout for the TRNG operation should be calculated with the formula:
39*4882a593Smuzhiyun * Timeout = EHR_NUM * VN_COEFF * EHR_LENGTH * SAMPLE_CNT * SCALE_VALUE
40*4882a593Smuzhiyun * while:
41*4882a593Smuzhiyun * - SAMPLE_CNT is input value from the characterisation process
42*4882a593Smuzhiyun * - all the rest are constants
43*4882a593Smuzhiyun */
44*4882a593Smuzhiyun #define EHR_NUM 1
45*4882a593Smuzhiyun #define VN_COEFF 4
46*4882a593Smuzhiyun #define EHR_LENGTH CC_TRNG_EHR_IN_BITS
47*4882a593Smuzhiyun #define SCALE_VALUE 2
48*4882a593Smuzhiyun #define CCTRNG_TIMEOUT(smpl_cnt) \
49*4882a593Smuzhiyun (EHR_NUM * VN_COEFF * EHR_LENGTH * smpl_cnt * SCALE_VALUE)
50*4882a593Smuzhiyun
51*4882a593Smuzhiyun struct cctrng_drvdata {
52*4882a593Smuzhiyun struct platform_device *pdev;
53*4882a593Smuzhiyun void __iomem *cc_base;
54*4882a593Smuzhiyun struct clk *clk;
55*4882a593Smuzhiyun struct hwrng rng;
56*4882a593Smuzhiyun u32 active_rosc;
57*4882a593Smuzhiyun /* Sampling interval for each ring oscillator:
58*4882a593Smuzhiyun * count of ring oscillator cycles between consecutive bits sampling.
59*4882a593Smuzhiyun * Value of 0 indicates non-valid rosc
60*4882a593Smuzhiyun */
61*4882a593Smuzhiyun u32 smpl_ratio[CC_TRNG_NUM_OF_ROSCS];
62*4882a593Smuzhiyun
63*4882a593Smuzhiyun u32 data_buf[CCTRNG_DATA_BUF_WORDS];
64*4882a593Smuzhiyun struct circ_buf circ;
65*4882a593Smuzhiyun struct work_struct compwork;
66*4882a593Smuzhiyun struct work_struct startwork;
67*4882a593Smuzhiyun
68*4882a593Smuzhiyun /* pending_hw - 1 when HW is pending, 0 when it is idle */
69*4882a593Smuzhiyun atomic_t pending_hw;
70*4882a593Smuzhiyun
71*4882a593Smuzhiyun /* protects against multiple concurrent consumers of data_buf */
72*4882a593Smuzhiyun spinlock_t read_lock;
73*4882a593Smuzhiyun };
74*4882a593Smuzhiyun
75*4882a593Smuzhiyun
76*4882a593Smuzhiyun /* functions for write/read CC registers */
cc_iowrite(struct cctrng_drvdata * drvdata,u32 reg,u32 val)77*4882a593Smuzhiyun static inline void cc_iowrite(struct cctrng_drvdata *drvdata, u32 reg, u32 val)
78*4882a593Smuzhiyun {
79*4882a593Smuzhiyun iowrite32(val, (drvdata->cc_base + reg));
80*4882a593Smuzhiyun }
cc_ioread(struct cctrng_drvdata * drvdata,u32 reg)81*4882a593Smuzhiyun static inline u32 cc_ioread(struct cctrng_drvdata *drvdata, u32 reg)
82*4882a593Smuzhiyun {
83*4882a593Smuzhiyun return ioread32(drvdata->cc_base + reg);
84*4882a593Smuzhiyun }
85*4882a593Smuzhiyun
86*4882a593Smuzhiyun
cc_trng_pm_get(struct device * dev)87*4882a593Smuzhiyun static int cc_trng_pm_get(struct device *dev)
88*4882a593Smuzhiyun {
89*4882a593Smuzhiyun int rc = 0;
90*4882a593Smuzhiyun
91*4882a593Smuzhiyun rc = pm_runtime_get_sync(dev);
92*4882a593Smuzhiyun
93*4882a593Smuzhiyun /* pm_runtime_get_sync() can return 1 as a valid return code */
94*4882a593Smuzhiyun return (rc == 1 ? 0 : rc);
95*4882a593Smuzhiyun }
96*4882a593Smuzhiyun
cc_trng_pm_put_suspend(struct device * dev)97*4882a593Smuzhiyun static void cc_trng_pm_put_suspend(struct device *dev)
98*4882a593Smuzhiyun {
99*4882a593Smuzhiyun int rc = 0;
100*4882a593Smuzhiyun
101*4882a593Smuzhiyun pm_runtime_mark_last_busy(dev);
102*4882a593Smuzhiyun rc = pm_runtime_put_autosuspend(dev);
103*4882a593Smuzhiyun if (rc)
104*4882a593Smuzhiyun dev_err(dev, "pm_runtime_put_autosuspend returned %x\n", rc);
105*4882a593Smuzhiyun }
106*4882a593Smuzhiyun
cc_trng_pm_init(struct cctrng_drvdata * drvdata)107*4882a593Smuzhiyun static int cc_trng_pm_init(struct cctrng_drvdata *drvdata)
108*4882a593Smuzhiyun {
109*4882a593Smuzhiyun struct device *dev = &(drvdata->pdev->dev);
110*4882a593Smuzhiyun
111*4882a593Smuzhiyun /* must be before the enabling to avoid redundant suspending */
112*4882a593Smuzhiyun pm_runtime_set_autosuspend_delay(dev, CC_TRNG_SUSPEND_TIMEOUT);
113*4882a593Smuzhiyun pm_runtime_use_autosuspend(dev);
114*4882a593Smuzhiyun /* set us as active - note we won't do PM ops until cc_trng_pm_go()! */
115*4882a593Smuzhiyun return pm_runtime_set_active(dev);
116*4882a593Smuzhiyun }
117*4882a593Smuzhiyun
cc_trng_pm_go(struct cctrng_drvdata * drvdata)118*4882a593Smuzhiyun static void cc_trng_pm_go(struct cctrng_drvdata *drvdata)
119*4882a593Smuzhiyun {
120*4882a593Smuzhiyun struct device *dev = &(drvdata->pdev->dev);
121*4882a593Smuzhiyun
122*4882a593Smuzhiyun /* enable the PM module*/
123*4882a593Smuzhiyun pm_runtime_enable(dev);
124*4882a593Smuzhiyun }
125*4882a593Smuzhiyun
cc_trng_pm_fini(struct cctrng_drvdata * drvdata)126*4882a593Smuzhiyun static void cc_trng_pm_fini(struct cctrng_drvdata *drvdata)
127*4882a593Smuzhiyun {
128*4882a593Smuzhiyun struct device *dev = &(drvdata->pdev->dev);
129*4882a593Smuzhiyun
130*4882a593Smuzhiyun pm_runtime_disable(dev);
131*4882a593Smuzhiyun }
132*4882a593Smuzhiyun
133*4882a593Smuzhiyun
cc_trng_parse_sampling_ratio(struct cctrng_drvdata * drvdata)134*4882a593Smuzhiyun static inline int cc_trng_parse_sampling_ratio(struct cctrng_drvdata *drvdata)
135*4882a593Smuzhiyun {
136*4882a593Smuzhiyun struct device *dev = &(drvdata->pdev->dev);
137*4882a593Smuzhiyun struct device_node *np = drvdata->pdev->dev.of_node;
138*4882a593Smuzhiyun int rc;
139*4882a593Smuzhiyun int i;
140*4882a593Smuzhiyun /* ret will be set to 0 if at least one rosc has (sampling ratio > 0) */
141*4882a593Smuzhiyun int ret = -EINVAL;
142*4882a593Smuzhiyun
143*4882a593Smuzhiyun rc = of_property_read_u32_array(np, "arm,rosc-ratio",
144*4882a593Smuzhiyun drvdata->smpl_ratio,
145*4882a593Smuzhiyun CC_TRNG_NUM_OF_ROSCS);
146*4882a593Smuzhiyun if (rc) {
147*4882a593Smuzhiyun /* arm,rosc-ratio was not found in device tree */
148*4882a593Smuzhiyun return rc;
149*4882a593Smuzhiyun }
150*4882a593Smuzhiyun
151*4882a593Smuzhiyun /* verify that at least one rosc has (sampling ratio > 0) */
152*4882a593Smuzhiyun for (i = 0; i < CC_TRNG_NUM_OF_ROSCS; ++i) {
153*4882a593Smuzhiyun dev_dbg(dev, "rosc %d sampling ratio %u",
154*4882a593Smuzhiyun i, drvdata->smpl_ratio[i]);
155*4882a593Smuzhiyun
156*4882a593Smuzhiyun if (drvdata->smpl_ratio[i] > 0)
157*4882a593Smuzhiyun ret = 0;
158*4882a593Smuzhiyun }
159*4882a593Smuzhiyun
160*4882a593Smuzhiyun return ret;
161*4882a593Smuzhiyun }
162*4882a593Smuzhiyun
cc_trng_change_rosc(struct cctrng_drvdata * drvdata)163*4882a593Smuzhiyun static int cc_trng_change_rosc(struct cctrng_drvdata *drvdata)
164*4882a593Smuzhiyun {
165*4882a593Smuzhiyun struct device *dev = &(drvdata->pdev->dev);
166*4882a593Smuzhiyun
167*4882a593Smuzhiyun dev_dbg(dev, "cctrng change rosc (was %d)\n", drvdata->active_rosc);
168*4882a593Smuzhiyun drvdata->active_rosc += 1;
169*4882a593Smuzhiyun
170*4882a593Smuzhiyun while (drvdata->active_rosc < CC_TRNG_NUM_OF_ROSCS) {
171*4882a593Smuzhiyun if (drvdata->smpl_ratio[drvdata->active_rosc] > 0)
172*4882a593Smuzhiyun return 0;
173*4882a593Smuzhiyun
174*4882a593Smuzhiyun drvdata->active_rosc += 1;
175*4882a593Smuzhiyun }
176*4882a593Smuzhiyun return -EINVAL;
177*4882a593Smuzhiyun }
178*4882a593Smuzhiyun
179*4882a593Smuzhiyun
cc_trng_enable_rnd_source(struct cctrng_drvdata * drvdata)180*4882a593Smuzhiyun static void cc_trng_enable_rnd_source(struct cctrng_drvdata *drvdata)
181*4882a593Smuzhiyun {
182*4882a593Smuzhiyun u32 max_cycles;
183*4882a593Smuzhiyun
184*4882a593Smuzhiyun /* Set watchdog threshold to maximal allowed time (in CPU cycles) */
185*4882a593Smuzhiyun max_cycles = CCTRNG_TIMEOUT(drvdata->smpl_ratio[drvdata->active_rosc]);
186*4882a593Smuzhiyun cc_iowrite(drvdata, CC_RNG_WATCHDOG_VAL_REG_OFFSET, max_cycles);
187*4882a593Smuzhiyun
188*4882a593Smuzhiyun /* enable the RND source */
189*4882a593Smuzhiyun cc_iowrite(drvdata, CC_RND_SOURCE_ENABLE_REG_OFFSET, 0x1);
190*4882a593Smuzhiyun
191*4882a593Smuzhiyun /* unmask RNG interrupts */
192*4882a593Smuzhiyun cc_iowrite(drvdata, CC_RNG_IMR_REG_OFFSET, (u32)~CC_RNG_INT_MASK);
193*4882a593Smuzhiyun }
194*4882a593Smuzhiyun
195*4882a593Smuzhiyun
196*4882a593Smuzhiyun /* increase circular data buffer index (head/tail) */
circ_idx_inc(int * idx,int bytes)197*4882a593Smuzhiyun static inline void circ_idx_inc(int *idx, int bytes)
198*4882a593Smuzhiyun {
199*4882a593Smuzhiyun *idx += (bytes + 3) >> 2;
200*4882a593Smuzhiyun *idx &= (CCTRNG_DATA_BUF_WORDS - 1);
201*4882a593Smuzhiyun }
202*4882a593Smuzhiyun
circ_buf_space(struct cctrng_drvdata * drvdata)203*4882a593Smuzhiyun static inline size_t circ_buf_space(struct cctrng_drvdata *drvdata)
204*4882a593Smuzhiyun {
205*4882a593Smuzhiyun return CIRC_SPACE(drvdata->circ.head,
206*4882a593Smuzhiyun drvdata->circ.tail, CCTRNG_DATA_BUF_WORDS);
207*4882a593Smuzhiyun
208*4882a593Smuzhiyun }
209*4882a593Smuzhiyun
cctrng_read(struct hwrng * rng,void * data,size_t max,bool wait)210*4882a593Smuzhiyun static int cctrng_read(struct hwrng *rng, void *data, size_t max, bool wait)
211*4882a593Smuzhiyun {
212*4882a593Smuzhiyun /* current implementation ignores "wait" */
213*4882a593Smuzhiyun
214*4882a593Smuzhiyun struct cctrng_drvdata *drvdata = (struct cctrng_drvdata *)rng->priv;
215*4882a593Smuzhiyun struct device *dev = &(drvdata->pdev->dev);
216*4882a593Smuzhiyun u32 *buf = (u32 *)drvdata->circ.buf;
217*4882a593Smuzhiyun size_t copied = 0;
218*4882a593Smuzhiyun size_t cnt_w;
219*4882a593Smuzhiyun size_t size;
220*4882a593Smuzhiyun size_t left;
221*4882a593Smuzhiyun
222*4882a593Smuzhiyun if (!spin_trylock(&drvdata->read_lock)) {
223*4882a593Smuzhiyun /* concurrent consumers from data_buf cannot be served */
224*4882a593Smuzhiyun dev_dbg_ratelimited(dev, "unable to hold lock\n");
225*4882a593Smuzhiyun return 0;
226*4882a593Smuzhiyun }
227*4882a593Smuzhiyun
228*4882a593Smuzhiyun /* copy till end of data buffer (without wrap back) */
229*4882a593Smuzhiyun cnt_w = CIRC_CNT_TO_END(drvdata->circ.head,
230*4882a593Smuzhiyun drvdata->circ.tail, CCTRNG_DATA_BUF_WORDS);
231*4882a593Smuzhiyun size = min((cnt_w<<2), max);
232*4882a593Smuzhiyun memcpy(data, &(buf[drvdata->circ.tail]), size);
233*4882a593Smuzhiyun copied = size;
234*4882a593Smuzhiyun circ_idx_inc(&drvdata->circ.tail, size);
235*4882a593Smuzhiyun /* copy rest of data in data buffer */
236*4882a593Smuzhiyun left = max - copied;
237*4882a593Smuzhiyun if (left > 0) {
238*4882a593Smuzhiyun cnt_w = CIRC_CNT(drvdata->circ.head,
239*4882a593Smuzhiyun drvdata->circ.tail, CCTRNG_DATA_BUF_WORDS);
240*4882a593Smuzhiyun size = min((cnt_w<<2), left);
241*4882a593Smuzhiyun memcpy(data, &(buf[drvdata->circ.tail]), size);
242*4882a593Smuzhiyun copied += size;
243*4882a593Smuzhiyun circ_idx_inc(&drvdata->circ.tail, size);
244*4882a593Smuzhiyun }
245*4882a593Smuzhiyun
246*4882a593Smuzhiyun spin_unlock(&drvdata->read_lock);
247*4882a593Smuzhiyun
248*4882a593Smuzhiyun if (circ_buf_space(drvdata) >= CC_TRNG_EHR_IN_WORDS) {
249*4882a593Smuzhiyun if (atomic_cmpxchg(&drvdata->pending_hw, 0, 1) == 0) {
250*4882a593Smuzhiyun /* re-check space in buffer to avoid potential race */
251*4882a593Smuzhiyun if (circ_buf_space(drvdata) >= CC_TRNG_EHR_IN_WORDS) {
252*4882a593Smuzhiyun /* increment device's usage counter */
253*4882a593Smuzhiyun int rc = cc_trng_pm_get(dev);
254*4882a593Smuzhiyun
255*4882a593Smuzhiyun if (rc) {
256*4882a593Smuzhiyun dev_err(dev,
257*4882a593Smuzhiyun "cc_trng_pm_get returned %x\n",
258*4882a593Smuzhiyun rc);
259*4882a593Smuzhiyun return rc;
260*4882a593Smuzhiyun }
261*4882a593Smuzhiyun
262*4882a593Smuzhiyun /* schedule execution of deferred work handler
263*4882a593Smuzhiyun * for filling of data buffer
264*4882a593Smuzhiyun */
265*4882a593Smuzhiyun schedule_work(&drvdata->startwork);
266*4882a593Smuzhiyun } else {
267*4882a593Smuzhiyun atomic_set(&drvdata->pending_hw, 0);
268*4882a593Smuzhiyun }
269*4882a593Smuzhiyun }
270*4882a593Smuzhiyun }
271*4882a593Smuzhiyun
272*4882a593Smuzhiyun return copied;
273*4882a593Smuzhiyun }
274*4882a593Smuzhiyun
cc_trng_hw_trigger(struct cctrng_drvdata * drvdata)275*4882a593Smuzhiyun static void cc_trng_hw_trigger(struct cctrng_drvdata *drvdata)
276*4882a593Smuzhiyun {
277*4882a593Smuzhiyun u32 tmp_smpl_cnt = 0;
278*4882a593Smuzhiyun struct device *dev = &(drvdata->pdev->dev);
279*4882a593Smuzhiyun
280*4882a593Smuzhiyun dev_dbg(dev, "cctrng hw trigger.\n");
281*4882a593Smuzhiyun
282*4882a593Smuzhiyun /* enable the HW RND clock */
283*4882a593Smuzhiyun cc_iowrite(drvdata, CC_RNG_CLK_ENABLE_REG_OFFSET, 0x1);
284*4882a593Smuzhiyun
285*4882a593Smuzhiyun /* do software reset */
286*4882a593Smuzhiyun cc_iowrite(drvdata, CC_RNG_SW_RESET_REG_OFFSET, 0x1);
287*4882a593Smuzhiyun /* in order to verify that the reset has completed,
288*4882a593Smuzhiyun * the sample count need to be verified
289*4882a593Smuzhiyun */
290*4882a593Smuzhiyun do {
291*4882a593Smuzhiyun /* enable the HW RND clock */
292*4882a593Smuzhiyun cc_iowrite(drvdata, CC_RNG_CLK_ENABLE_REG_OFFSET, 0x1);
293*4882a593Smuzhiyun
294*4882a593Smuzhiyun /* set sampling ratio (rng_clocks) between consecutive bits */
295*4882a593Smuzhiyun cc_iowrite(drvdata, CC_SAMPLE_CNT1_REG_OFFSET,
296*4882a593Smuzhiyun drvdata->smpl_ratio[drvdata->active_rosc]);
297*4882a593Smuzhiyun
298*4882a593Smuzhiyun /* read the sampling ratio */
299*4882a593Smuzhiyun tmp_smpl_cnt = cc_ioread(drvdata, CC_SAMPLE_CNT1_REG_OFFSET);
300*4882a593Smuzhiyun
301*4882a593Smuzhiyun } while (tmp_smpl_cnt != drvdata->smpl_ratio[drvdata->active_rosc]);
302*4882a593Smuzhiyun
303*4882a593Smuzhiyun /* disable the RND source for setting new parameters in HW */
304*4882a593Smuzhiyun cc_iowrite(drvdata, CC_RND_SOURCE_ENABLE_REG_OFFSET, 0);
305*4882a593Smuzhiyun
306*4882a593Smuzhiyun cc_iowrite(drvdata, CC_RNG_ICR_REG_OFFSET, 0xFFFFFFFF);
307*4882a593Smuzhiyun
308*4882a593Smuzhiyun cc_iowrite(drvdata, CC_TRNG_CONFIG_REG_OFFSET, drvdata->active_rosc);
309*4882a593Smuzhiyun
310*4882a593Smuzhiyun /* Debug Control register: set to 0 - no bypasses */
311*4882a593Smuzhiyun cc_iowrite(drvdata, CC_TRNG_DEBUG_CONTROL_REG_OFFSET, 0);
312*4882a593Smuzhiyun
313*4882a593Smuzhiyun cc_trng_enable_rnd_source(drvdata);
314*4882a593Smuzhiyun }
315*4882a593Smuzhiyun
cc_trng_compwork_handler(struct work_struct * w)316*4882a593Smuzhiyun static void cc_trng_compwork_handler(struct work_struct *w)
317*4882a593Smuzhiyun {
318*4882a593Smuzhiyun u32 isr = 0;
319*4882a593Smuzhiyun u32 ehr_valid = 0;
320*4882a593Smuzhiyun struct cctrng_drvdata *drvdata =
321*4882a593Smuzhiyun container_of(w, struct cctrng_drvdata, compwork);
322*4882a593Smuzhiyun struct device *dev = &(drvdata->pdev->dev);
323*4882a593Smuzhiyun int i;
324*4882a593Smuzhiyun
325*4882a593Smuzhiyun /* stop DMA and the RNG source */
326*4882a593Smuzhiyun cc_iowrite(drvdata, CC_RNG_DMA_ENABLE_REG_OFFSET, 0);
327*4882a593Smuzhiyun cc_iowrite(drvdata, CC_RND_SOURCE_ENABLE_REG_OFFSET, 0);
328*4882a593Smuzhiyun
329*4882a593Smuzhiyun /* read RNG_ISR and check for errors */
330*4882a593Smuzhiyun isr = cc_ioread(drvdata, CC_RNG_ISR_REG_OFFSET);
331*4882a593Smuzhiyun ehr_valid = CC_REG_FLD_GET(RNG_ISR, EHR_VALID, isr);
332*4882a593Smuzhiyun dev_dbg(dev, "Got RNG_ISR=0x%08X (EHR_VALID=%u)\n", isr, ehr_valid);
333*4882a593Smuzhiyun
334*4882a593Smuzhiyun if (fips_enabled && CC_REG_FLD_GET(RNG_ISR, CRNGT_ERR, isr)) {
335*4882a593Smuzhiyun fips_fail_notify();
336*4882a593Smuzhiyun /* FIPS error is fatal */
337*4882a593Smuzhiyun panic("Got HW CRNGT error while fips is enabled!\n");
338*4882a593Smuzhiyun }
339*4882a593Smuzhiyun
340*4882a593Smuzhiyun /* Clear all pending RNG interrupts */
341*4882a593Smuzhiyun cc_iowrite(drvdata, CC_RNG_ICR_REG_OFFSET, isr);
342*4882a593Smuzhiyun
343*4882a593Smuzhiyun
344*4882a593Smuzhiyun if (!ehr_valid) {
345*4882a593Smuzhiyun /* in case of AUTOCORR/TIMEOUT error, try the next ROSC */
346*4882a593Smuzhiyun if (CC_REG_FLD_GET(RNG_ISR, AUTOCORR_ERR, isr) ||
347*4882a593Smuzhiyun CC_REG_FLD_GET(RNG_ISR, WATCHDOG, isr)) {
348*4882a593Smuzhiyun dev_dbg(dev, "cctrng autocorr/timeout error.\n");
349*4882a593Smuzhiyun goto next_rosc;
350*4882a593Smuzhiyun }
351*4882a593Smuzhiyun
352*4882a593Smuzhiyun /* in case of VN error, ignore it */
353*4882a593Smuzhiyun }
354*4882a593Smuzhiyun
355*4882a593Smuzhiyun /* read EHR data from registers */
356*4882a593Smuzhiyun for (i = 0; i < CC_TRNG_EHR_IN_WORDS; i++) {
357*4882a593Smuzhiyun /* calc word ptr in data_buf */
358*4882a593Smuzhiyun u32 *buf = (u32 *)drvdata->circ.buf;
359*4882a593Smuzhiyun
360*4882a593Smuzhiyun buf[drvdata->circ.head] = cc_ioread(drvdata,
361*4882a593Smuzhiyun CC_EHR_DATA_0_REG_OFFSET + (i*sizeof(u32)));
362*4882a593Smuzhiyun
363*4882a593Smuzhiyun /* EHR_DATA registers are cleared on read. In case 0 value was
364*4882a593Smuzhiyun * returned, restart the entropy collection.
365*4882a593Smuzhiyun */
366*4882a593Smuzhiyun if (buf[drvdata->circ.head] == 0) {
367*4882a593Smuzhiyun dev_dbg(dev, "Got 0 value in EHR. active_rosc %u\n",
368*4882a593Smuzhiyun drvdata->active_rosc);
369*4882a593Smuzhiyun goto next_rosc;
370*4882a593Smuzhiyun }
371*4882a593Smuzhiyun
372*4882a593Smuzhiyun circ_idx_inc(&drvdata->circ.head, 1<<2);
373*4882a593Smuzhiyun }
374*4882a593Smuzhiyun
375*4882a593Smuzhiyun atomic_set(&drvdata->pending_hw, 0);
376*4882a593Smuzhiyun
377*4882a593Smuzhiyun /* continue to fill data buffer if needed */
378*4882a593Smuzhiyun if (circ_buf_space(drvdata) >= CC_TRNG_EHR_IN_WORDS) {
379*4882a593Smuzhiyun if (atomic_cmpxchg(&drvdata->pending_hw, 0, 1) == 0) {
380*4882a593Smuzhiyun /* Re-enable rnd source */
381*4882a593Smuzhiyun cc_trng_enable_rnd_source(drvdata);
382*4882a593Smuzhiyun return;
383*4882a593Smuzhiyun }
384*4882a593Smuzhiyun }
385*4882a593Smuzhiyun
386*4882a593Smuzhiyun cc_trng_pm_put_suspend(dev);
387*4882a593Smuzhiyun
388*4882a593Smuzhiyun dev_dbg(dev, "compwork handler done\n");
389*4882a593Smuzhiyun return;
390*4882a593Smuzhiyun
391*4882a593Smuzhiyun next_rosc:
392*4882a593Smuzhiyun if ((circ_buf_space(drvdata) >= CC_TRNG_EHR_IN_WORDS) &&
393*4882a593Smuzhiyun (cc_trng_change_rosc(drvdata) == 0)) {
394*4882a593Smuzhiyun /* trigger trng hw with next rosc */
395*4882a593Smuzhiyun cc_trng_hw_trigger(drvdata);
396*4882a593Smuzhiyun } else {
397*4882a593Smuzhiyun atomic_set(&drvdata->pending_hw, 0);
398*4882a593Smuzhiyun cc_trng_pm_put_suspend(dev);
399*4882a593Smuzhiyun }
400*4882a593Smuzhiyun }
401*4882a593Smuzhiyun
cc_isr(int irq,void * dev_id)402*4882a593Smuzhiyun static irqreturn_t cc_isr(int irq, void *dev_id)
403*4882a593Smuzhiyun {
404*4882a593Smuzhiyun struct cctrng_drvdata *drvdata = (struct cctrng_drvdata *)dev_id;
405*4882a593Smuzhiyun struct device *dev = &(drvdata->pdev->dev);
406*4882a593Smuzhiyun u32 irr;
407*4882a593Smuzhiyun
408*4882a593Smuzhiyun /* if driver suspended return, probably shared interrupt */
409*4882a593Smuzhiyun if (pm_runtime_suspended(dev))
410*4882a593Smuzhiyun return IRQ_NONE;
411*4882a593Smuzhiyun
412*4882a593Smuzhiyun /* read the interrupt status */
413*4882a593Smuzhiyun irr = cc_ioread(drvdata, CC_HOST_RGF_IRR_REG_OFFSET);
414*4882a593Smuzhiyun dev_dbg(dev, "Got IRR=0x%08X\n", irr);
415*4882a593Smuzhiyun
416*4882a593Smuzhiyun if (irr == 0) /* Probably shared interrupt line */
417*4882a593Smuzhiyun return IRQ_NONE;
418*4882a593Smuzhiyun
419*4882a593Smuzhiyun /* clear interrupt - must be before processing events */
420*4882a593Smuzhiyun cc_iowrite(drvdata, CC_HOST_RGF_ICR_REG_OFFSET, irr);
421*4882a593Smuzhiyun
422*4882a593Smuzhiyun /* RNG interrupt - most probable */
423*4882a593Smuzhiyun if (irr & CC_HOST_RNG_IRQ_MASK) {
424*4882a593Smuzhiyun /* Mask RNG interrupts - will be unmasked in deferred work */
425*4882a593Smuzhiyun cc_iowrite(drvdata, CC_RNG_IMR_REG_OFFSET, 0xFFFFFFFF);
426*4882a593Smuzhiyun
427*4882a593Smuzhiyun /* We clear RNG interrupt here,
428*4882a593Smuzhiyun * to avoid it from firing as we'll unmask RNG interrupts.
429*4882a593Smuzhiyun */
430*4882a593Smuzhiyun cc_iowrite(drvdata, CC_HOST_RGF_ICR_REG_OFFSET,
431*4882a593Smuzhiyun CC_HOST_RNG_IRQ_MASK);
432*4882a593Smuzhiyun
433*4882a593Smuzhiyun irr &= ~CC_HOST_RNG_IRQ_MASK;
434*4882a593Smuzhiyun
435*4882a593Smuzhiyun /* schedule execution of deferred work handler */
436*4882a593Smuzhiyun schedule_work(&drvdata->compwork);
437*4882a593Smuzhiyun }
438*4882a593Smuzhiyun
439*4882a593Smuzhiyun if (irr) {
440*4882a593Smuzhiyun dev_dbg_ratelimited(dev,
441*4882a593Smuzhiyun "IRR includes unknown cause bits (0x%08X)\n",
442*4882a593Smuzhiyun irr);
443*4882a593Smuzhiyun /* Just warning */
444*4882a593Smuzhiyun }
445*4882a593Smuzhiyun
446*4882a593Smuzhiyun return IRQ_HANDLED;
447*4882a593Smuzhiyun }
448*4882a593Smuzhiyun
cc_trng_startwork_handler(struct work_struct * w)449*4882a593Smuzhiyun static void cc_trng_startwork_handler(struct work_struct *w)
450*4882a593Smuzhiyun {
451*4882a593Smuzhiyun struct cctrng_drvdata *drvdata =
452*4882a593Smuzhiyun container_of(w, struct cctrng_drvdata, startwork);
453*4882a593Smuzhiyun
454*4882a593Smuzhiyun drvdata->active_rosc = 0;
455*4882a593Smuzhiyun cc_trng_hw_trigger(drvdata);
456*4882a593Smuzhiyun }
457*4882a593Smuzhiyun
458*4882a593Smuzhiyun
cc_trng_clk_init(struct cctrng_drvdata * drvdata)459*4882a593Smuzhiyun static int cc_trng_clk_init(struct cctrng_drvdata *drvdata)
460*4882a593Smuzhiyun {
461*4882a593Smuzhiyun struct clk *clk;
462*4882a593Smuzhiyun struct device *dev = &(drvdata->pdev->dev);
463*4882a593Smuzhiyun int rc = 0;
464*4882a593Smuzhiyun
465*4882a593Smuzhiyun clk = devm_clk_get_optional(dev, NULL);
466*4882a593Smuzhiyun if (IS_ERR(clk))
467*4882a593Smuzhiyun return dev_err_probe(dev, PTR_ERR(clk),
468*4882a593Smuzhiyun "Error getting clock\n");
469*4882a593Smuzhiyun
470*4882a593Smuzhiyun drvdata->clk = clk;
471*4882a593Smuzhiyun
472*4882a593Smuzhiyun rc = clk_prepare_enable(drvdata->clk);
473*4882a593Smuzhiyun if (rc) {
474*4882a593Smuzhiyun dev_err(dev, "Failed to enable clock\n");
475*4882a593Smuzhiyun return rc;
476*4882a593Smuzhiyun }
477*4882a593Smuzhiyun
478*4882a593Smuzhiyun return 0;
479*4882a593Smuzhiyun }
480*4882a593Smuzhiyun
cc_trng_clk_fini(struct cctrng_drvdata * drvdata)481*4882a593Smuzhiyun static void cc_trng_clk_fini(struct cctrng_drvdata *drvdata)
482*4882a593Smuzhiyun {
483*4882a593Smuzhiyun clk_disable_unprepare(drvdata->clk);
484*4882a593Smuzhiyun }
485*4882a593Smuzhiyun
486*4882a593Smuzhiyun
cctrng_probe(struct platform_device * pdev)487*4882a593Smuzhiyun static int cctrng_probe(struct platform_device *pdev)
488*4882a593Smuzhiyun {
489*4882a593Smuzhiyun struct resource *req_mem_cc_regs = NULL;
490*4882a593Smuzhiyun struct cctrng_drvdata *drvdata;
491*4882a593Smuzhiyun struct device *dev = &pdev->dev;
492*4882a593Smuzhiyun int rc = 0;
493*4882a593Smuzhiyun u32 val;
494*4882a593Smuzhiyun int irq;
495*4882a593Smuzhiyun
496*4882a593Smuzhiyun drvdata = devm_kzalloc(dev, sizeof(*drvdata), GFP_KERNEL);
497*4882a593Smuzhiyun if (!drvdata)
498*4882a593Smuzhiyun return -ENOMEM;
499*4882a593Smuzhiyun
500*4882a593Smuzhiyun drvdata->rng.name = devm_kstrdup(dev, dev_name(dev), GFP_KERNEL);
501*4882a593Smuzhiyun if (!drvdata->rng.name)
502*4882a593Smuzhiyun return -ENOMEM;
503*4882a593Smuzhiyun
504*4882a593Smuzhiyun drvdata->rng.read = cctrng_read;
505*4882a593Smuzhiyun drvdata->rng.priv = (unsigned long)drvdata;
506*4882a593Smuzhiyun drvdata->rng.quality = CC_TRNG_QUALITY;
507*4882a593Smuzhiyun
508*4882a593Smuzhiyun platform_set_drvdata(pdev, drvdata);
509*4882a593Smuzhiyun drvdata->pdev = pdev;
510*4882a593Smuzhiyun
511*4882a593Smuzhiyun drvdata->circ.buf = (char *)drvdata->data_buf;
512*4882a593Smuzhiyun
513*4882a593Smuzhiyun /* Get device resources */
514*4882a593Smuzhiyun /* First CC registers space */
515*4882a593Smuzhiyun req_mem_cc_regs = platform_get_resource(pdev, IORESOURCE_MEM, 0);
516*4882a593Smuzhiyun /* Map registers space */
517*4882a593Smuzhiyun drvdata->cc_base = devm_ioremap_resource(dev, req_mem_cc_regs);
518*4882a593Smuzhiyun if (IS_ERR(drvdata->cc_base)) {
519*4882a593Smuzhiyun dev_err(dev, "Failed to ioremap registers");
520*4882a593Smuzhiyun return PTR_ERR(drvdata->cc_base);
521*4882a593Smuzhiyun }
522*4882a593Smuzhiyun
523*4882a593Smuzhiyun dev_dbg(dev, "Got MEM resource (%s): %pR\n", req_mem_cc_regs->name,
524*4882a593Smuzhiyun req_mem_cc_regs);
525*4882a593Smuzhiyun dev_dbg(dev, "CC registers mapped from %pa to 0x%p\n",
526*4882a593Smuzhiyun &req_mem_cc_regs->start, drvdata->cc_base);
527*4882a593Smuzhiyun
528*4882a593Smuzhiyun /* Then IRQ */
529*4882a593Smuzhiyun irq = platform_get_irq(pdev, 0);
530*4882a593Smuzhiyun if (irq < 0) {
531*4882a593Smuzhiyun dev_err(dev, "Failed getting IRQ resource\n");
532*4882a593Smuzhiyun return irq;
533*4882a593Smuzhiyun }
534*4882a593Smuzhiyun
535*4882a593Smuzhiyun /* parse sampling rate from device tree */
536*4882a593Smuzhiyun rc = cc_trng_parse_sampling_ratio(drvdata);
537*4882a593Smuzhiyun if (rc) {
538*4882a593Smuzhiyun dev_err(dev, "Failed to get legal sampling ratio for rosc\n");
539*4882a593Smuzhiyun return rc;
540*4882a593Smuzhiyun }
541*4882a593Smuzhiyun
542*4882a593Smuzhiyun rc = cc_trng_clk_init(drvdata);
543*4882a593Smuzhiyun if (rc) {
544*4882a593Smuzhiyun dev_err(dev, "cc_trng_clk_init failed\n");
545*4882a593Smuzhiyun return rc;
546*4882a593Smuzhiyun }
547*4882a593Smuzhiyun
548*4882a593Smuzhiyun INIT_WORK(&drvdata->compwork, cc_trng_compwork_handler);
549*4882a593Smuzhiyun INIT_WORK(&drvdata->startwork, cc_trng_startwork_handler);
550*4882a593Smuzhiyun spin_lock_init(&drvdata->read_lock);
551*4882a593Smuzhiyun
552*4882a593Smuzhiyun /* register the driver isr function */
553*4882a593Smuzhiyun rc = devm_request_irq(dev, irq, cc_isr, IRQF_SHARED, "cctrng", drvdata);
554*4882a593Smuzhiyun if (rc) {
555*4882a593Smuzhiyun dev_err(dev, "Could not register to interrupt %d\n", irq);
556*4882a593Smuzhiyun goto post_clk_err;
557*4882a593Smuzhiyun }
558*4882a593Smuzhiyun dev_dbg(dev, "Registered to IRQ: %d\n", irq);
559*4882a593Smuzhiyun
560*4882a593Smuzhiyun /* Clear all pending interrupts */
561*4882a593Smuzhiyun val = cc_ioread(drvdata, CC_HOST_RGF_IRR_REG_OFFSET);
562*4882a593Smuzhiyun dev_dbg(dev, "IRR=0x%08X\n", val);
563*4882a593Smuzhiyun cc_iowrite(drvdata, CC_HOST_RGF_ICR_REG_OFFSET, val);
564*4882a593Smuzhiyun
565*4882a593Smuzhiyun /* unmask HOST RNG interrupt */
566*4882a593Smuzhiyun cc_iowrite(drvdata, CC_HOST_RGF_IMR_REG_OFFSET,
567*4882a593Smuzhiyun cc_ioread(drvdata, CC_HOST_RGF_IMR_REG_OFFSET) &
568*4882a593Smuzhiyun ~CC_HOST_RNG_IRQ_MASK);
569*4882a593Smuzhiyun
570*4882a593Smuzhiyun /* init PM */
571*4882a593Smuzhiyun rc = cc_trng_pm_init(drvdata);
572*4882a593Smuzhiyun if (rc) {
573*4882a593Smuzhiyun dev_err(dev, "cc_trng_pm_init failed\n");
574*4882a593Smuzhiyun goto post_clk_err;
575*4882a593Smuzhiyun }
576*4882a593Smuzhiyun
577*4882a593Smuzhiyun /* increment device's usage counter */
578*4882a593Smuzhiyun rc = cc_trng_pm_get(dev);
579*4882a593Smuzhiyun if (rc) {
580*4882a593Smuzhiyun dev_err(dev, "cc_trng_pm_get returned %x\n", rc);
581*4882a593Smuzhiyun goto post_pm_err;
582*4882a593Smuzhiyun }
583*4882a593Smuzhiyun
584*4882a593Smuzhiyun /* set pending_hw to verify that HW won't be triggered from read */
585*4882a593Smuzhiyun atomic_set(&drvdata->pending_hw, 1);
586*4882a593Smuzhiyun
587*4882a593Smuzhiyun /* registration of the hwrng device */
588*4882a593Smuzhiyun rc = hwrng_register(&drvdata->rng);
589*4882a593Smuzhiyun if (rc) {
590*4882a593Smuzhiyun dev_err(dev, "Could not register hwrng device.\n");
591*4882a593Smuzhiyun goto post_pm_err;
592*4882a593Smuzhiyun }
593*4882a593Smuzhiyun
594*4882a593Smuzhiyun /* trigger HW to start generate data */
595*4882a593Smuzhiyun drvdata->active_rosc = 0;
596*4882a593Smuzhiyun cc_trng_hw_trigger(drvdata);
597*4882a593Smuzhiyun
598*4882a593Smuzhiyun /* All set, we can allow auto-suspend */
599*4882a593Smuzhiyun cc_trng_pm_go(drvdata);
600*4882a593Smuzhiyun
601*4882a593Smuzhiyun dev_info(dev, "ARM cctrng device initialized\n");
602*4882a593Smuzhiyun
603*4882a593Smuzhiyun return 0;
604*4882a593Smuzhiyun
605*4882a593Smuzhiyun post_pm_err:
606*4882a593Smuzhiyun cc_trng_pm_fini(drvdata);
607*4882a593Smuzhiyun
608*4882a593Smuzhiyun post_clk_err:
609*4882a593Smuzhiyun cc_trng_clk_fini(drvdata);
610*4882a593Smuzhiyun
611*4882a593Smuzhiyun return rc;
612*4882a593Smuzhiyun }
613*4882a593Smuzhiyun
cctrng_remove(struct platform_device * pdev)614*4882a593Smuzhiyun static int cctrng_remove(struct platform_device *pdev)
615*4882a593Smuzhiyun {
616*4882a593Smuzhiyun struct cctrng_drvdata *drvdata = platform_get_drvdata(pdev);
617*4882a593Smuzhiyun struct device *dev = &pdev->dev;
618*4882a593Smuzhiyun
619*4882a593Smuzhiyun dev_dbg(dev, "Releasing cctrng resources...\n");
620*4882a593Smuzhiyun
621*4882a593Smuzhiyun hwrng_unregister(&drvdata->rng);
622*4882a593Smuzhiyun
623*4882a593Smuzhiyun cc_trng_pm_fini(drvdata);
624*4882a593Smuzhiyun
625*4882a593Smuzhiyun cc_trng_clk_fini(drvdata);
626*4882a593Smuzhiyun
627*4882a593Smuzhiyun dev_info(dev, "ARM cctrng device terminated\n");
628*4882a593Smuzhiyun
629*4882a593Smuzhiyun return 0;
630*4882a593Smuzhiyun }
631*4882a593Smuzhiyun
cctrng_suspend(struct device * dev)632*4882a593Smuzhiyun static int __maybe_unused cctrng_suspend(struct device *dev)
633*4882a593Smuzhiyun {
634*4882a593Smuzhiyun struct cctrng_drvdata *drvdata = dev_get_drvdata(dev);
635*4882a593Smuzhiyun
636*4882a593Smuzhiyun dev_dbg(dev, "set HOST_POWER_DOWN_EN\n");
637*4882a593Smuzhiyun cc_iowrite(drvdata, CC_HOST_POWER_DOWN_EN_REG_OFFSET,
638*4882a593Smuzhiyun POWER_DOWN_ENABLE);
639*4882a593Smuzhiyun
640*4882a593Smuzhiyun clk_disable_unprepare(drvdata->clk);
641*4882a593Smuzhiyun
642*4882a593Smuzhiyun return 0;
643*4882a593Smuzhiyun }
644*4882a593Smuzhiyun
cctrng_wait_for_reset_completion(struct cctrng_drvdata * drvdata)645*4882a593Smuzhiyun static bool cctrng_wait_for_reset_completion(struct cctrng_drvdata *drvdata)
646*4882a593Smuzhiyun {
647*4882a593Smuzhiyun unsigned int val;
648*4882a593Smuzhiyun unsigned int i;
649*4882a593Smuzhiyun
650*4882a593Smuzhiyun for (i = 0; i < CC_HW_RESET_LOOP_COUNT; i++) {
651*4882a593Smuzhiyun /* in cc7x3 NVM_IS_IDLE indicates that CC reset is
652*4882a593Smuzhiyun * completed and device is fully functional
653*4882a593Smuzhiyun */
654*4882a593Smuzhiyun val = cc_ioread(drvdata, CC_NVM_IS_IDLE_REG_OFFSET);
655*4882a593Smuzhiyun if (val & BIT(CC_NVM_IS_IDLE_VALUE_BIT_SHIFT)) {
656*4882a593Smuzhiyun /* hw indicate reset completed */
657*4882a593Smuzhiyun return true;
658*4882a593Smuzhiyun }
659*4882a593Smuzhiyun /* allow scheduling other process on the processor */
660*4882a593Smuzhiyun schedule();
661*4882a593Smuzhiyun }
662*4882a593Smuzhiyun /* reset not completed */
663*4882a593Smuzhiyun return false;
664*4882a593Smuzhiyun }
665*4882a593Smuzhiyun
cctrng_resume(struct device * dev)666*4882a593Smuzhiyun static int __maybe_unused cctrng_resume(struct device *dev)
667*4882a593Smuzhiyun {
668*4882a593Smuzhiyun struct cctrng_drvdata *drvdata = dev_get_drvdata(dev);
669*4882a593Smuzhiyun int rc;
670*4882a593Smuzhiyun
671*4882a593Smuzhiyun dev_dbg(dev, "unset HOST_POWER_DOWN_EN\n");
672*4882a593Smuzhiyun /* Enables the device source clk */
673*4882a593Smuzhiyun rc = clk_prepare_enable(drvdata->clk);
674*4882a593Smuzhiyun if (rc) {
675*4882a593Smuzhiyun dev_err(dev, "failed getting clock back on. We're toast.\n");
676*4882a593Smuzhiyun return rc;
677*4882a593Smuzhiyun }
678*4882a593Smuzhiyun
679*4882a593Smuzhiyun /* wait for Cryptocell reset completion */
680*4882a593Smuzhiyun if (!cctrng_wait_for_reset_completion(drvdata)) {
681*4882a593Smuzhiyun dev_err(dev, "Cryptocell reset not completed");
682*4882a593Smuzhiyun return -EBUSY;
683*4882a593Smuzhiyun }
684*4882a593Smuzhiyun
685*4882a593Smuzhiyun /* unmask HOST RNG interrupt */
686*4882a593Smuzhiyun cc_iowrite(drvdata, CC_HOST_RGF_IMR_REG_OFFSET,
687*4882a593Smuzhiyun cc_ioread(drvdata, CC_HOST_RGF_IMR_REG_OFFSET) &
688*4882a593Smuzhiyun ~CC_HOST_RNG_IRQ_MASK);
689*4882a593Smuzhiyun
690*4882a593Smuzhiyun cc_iowrite(drvdata, CC_HOST_POWER_DOWN_EN_REG_OFFSET,
691*4882a593Smuzhiyun POWER_DOWN_DISABLE);
692*4882a593Smuzhiyun
693*4882a593Smuzhiyun return 0;
694*4882a593Smuzhiyun }
695*4882a593Smuzhiyun
696*4882a593Smuzhiyun static UNIVERSAL_DEV_PM_OPS(cctrng_pm, cctrng_suspend, cctrng_resume, NULL);
697*4882a593Smuzhiyun
698*4882a593Smuzhiyun static const struct of_device_id arm_cctrng_dt_match[] = {
699*4882a593Smuzhiyun { .compatible = "arm,cryptocell-713-trng", },
700*4882a593Smuzhiyun { .compatible = "arm,cryptocell-703-trng", },
701*4882a593Smuzhiyun {},
702*4882a593Smuzhiyun };
703*4882a593Smuzhiyun MODULE_DEVICE_TABLE(of, arm_cctrng_dt_match);
704*4882a593Smuzhiyun
705*4882a593Smuzhiyun static struct platform_driver cctrng_driver = {
706*4882a593Smuzhiyun .driver = {
707*4882a593Smuzhiyun .name = "cctrng",
708*4882a593Smuzhiyun .of_match_table = arm_cctrng_dt_match,
709*4882a593Smuzhiyun .pm = &cctrng_pm,
710*4882a593Smuzhiyun },
711*4882a593Smuzhiyun .probe = cctrng_probe,
712*4882a593Smuzhiyun .remove = cctrng_remove,
713*4882a593Smuzhiyun };
714*4882a593Smuzhiyun
cctrng_mod_init(void)715*4882a593Smuzhiyun static int __init cctrng_mod_init(void)
716*4882a593Smuzhiyun {
717*4882a593Smuzhiyun /* Compile time assertion checks */
718*4882a593Smuzhiyun BUILD_BUG_ON(CCTRNG_DATA_BUF_WORDS < 6);
719*4882a593Smuzhiyun BUILD_BUG_ON((CCTRNG_DATA_BUF_WORDS & (CCTRNG_DATA_BUF_WORDS-1)) != 0);
720*4882a593Smuzhiyun
721*4882a593Smuzhiyun return platform_driver_register(&cctrng_driver);
722*4882a593Smuzhiyun }
723*4882a593Smuzhiyun module_init(cctrng_mod_init);
724*4882a593Smuzhiyun
cctrng_mod_exit(void)725*4882a593Smuzhiyun static void __exit cctrng_mod_exit(void)
726*4882a593Smuzhiyun {
727*4882a593Smuzhiyun platform_driver_unregister(&cctrng_driver);
728*4882a593Smuzhiyun }
729*4882a593Smuzhiyun module_exit(cctrng_mod_exit);
730*4882a593Smuzhiyun
731*4882a593Smuzhiyun /* Module description */
732*4882a593Smuzhiyun MODULE_DESCRIPTION("ARM CryptoCell TRNG Driver");
733*4882a593Smuzhiyun MODULE_AUTHOR("ARM");
734*4882a593Smuzhiyun MODULE_LICENSE("GPL v2");
735