1*4882a593Smuzhiyun /* SPDX-License-Identifier: GPL-2.0+ */
2*4882a593Smuzhiyun /*
3*4882a593Smuzhiyun * Copyright (c) 2001-2002 by David Brownell
4*4882a593Smuzhiyun */
5*4882a593Smuzhiyun
6*4882a593Smuzhiyun #ifndef __LINUX_EHCI_HCD_H
7*4882a593Smuzhiyun #define __LINUX_EHCI_HCD_H
8*4882a593Smuzhiyun
9*4882a593Smuzhiyun /* definitions used for the EHCI driver */
10*4882a593Smuzhiyun
11*4882a593Smuzhiyun /*
12*4882a593Smuzhiyun * __hc32 and __hc16 are "Host Controller" types, they may be equivalent to
13*4882a593Smuzhiyun * __leXX (normally) or __beXX (given EHCI_BIG_ENDIAN_DESC), depending on
14*4882a593Smuzhiyun * the host controller implementation.
15*4882a593Smuzhiyun *
16*4882a593Smuzhiyun * To facilitate the strongest possible byte-order checking from "sparse"
17*4882a593Smuzhiyun * and so on, we use __leXX unless that's not practical.
18*4882a593Smuzhiyun */
19*4882a593Smuzhiyun #ifdef CONFIG_USB_EHCI_BIG_ENDIAN_DESC
20*4882a593Smuzhiyun typedef __u32 __bitwise __hc32;
21*4882a593Smuzhiyun typedef __u16 __bitwise __hc16;
22*4882a593Smuzhiyun #else
23*4882a593Smuzhiyun #define __hc32 __le32
24*4882a593Smuzhiyun #define __hc16 __le16
25*4882a593Smuzhiyun #endif
26*4882a593Smuzhiyun
27*4882a593Smuzhiyun /* statistics can be kept for tuning/monitoring */
28*4882a593Smuzhiyun #ifdef CONFIG_DYNAMIC_DEBUG
29*4882a593Smuzhiyun #define EHCI_STATS
30*4882a593Smuzhiyun #endif
31*4882a593Smuzhiyun
32*4882a593Smuzhiyun struct ehci_stats {
33*4882a593Smuzhiyun /* irq usage */
34*4882a593Smuzhiyun unsigned long normal;
35*4882a593Smuzhiyun unsigned long error;
36*4882a593Smuzhiyun unsigned long iaa;
37*4882a593Smuzhiyun unsigned long lost_iaa;
38*4882a593Smuzhiyun
39*4882a593Smuzhiyun /* termination of urbs from core */
40*4882a593Smuzhiyun unsigned long complete;
41*4882a593Smuzhiyun unsigned long unlink;
42*4882a593Smuzhiyun };
43*4882a593Smuzhiyun
44*4882a593Smuzhiyun /*
45*4882a593Smuzhiyun * Scheduling and budgeting information for periodic transfers, for both
46*4882a593Smuzhiyun * high-speed devices and full/low-speed devices lying behind a TT.
47*4882a593Smuzhiyun */
48*4882a593Smuzhiyun struct ehci_per_sched {
49*4882a593Smuzhiyun struct usb_device *udev; /* access to the TT */
50*4882a593Smuzhiyun struct usb_host_endpoint *ep;
51*4882a593Smuzhiyun struct list_head ps_list; /* node on ehci_tt's ps_list */
52*4882a593Smuzhiyun u16 tt_usecs; /* time on the FS/LS bus */
53*4882a593Smuzhiyun u16 cs_mask; /* C-mask and S-mask bytes */
54*4882a593Smuzhiyun u16 period; /* actual period in frames */
55*4882a593Smuzhiyun u16 phase; /* actual phase, frame part */
56*4882a593Smuzhiyun u8 bw_phase; /* same, for bandwidth
57*4882a593Smuzhiyun reservation */
58*4882a593Smuzhiyun u8 phase_uf; /* uframe part of the phase */
59*4882a593Smuzhiyun u8 usecs, c_usecs; /* times on the HS bus */
60*4882a593Smuzhiyun u8 bw_uperiod; /* period in microframes, for
61*4882a593Smuzhiyun bandwidth reservation */
62*4882a593Smuzhiyun u8 bw_period; /* same, in frames */
63*4882a593Smuzhiyun };
64*4882a593Smuzhiyun #define NO_FRAME 29999 /* frame not assigned yet */
65*4882a593Smuzhiyun
66*4882a593Smuzhiyun /* ehci_hcd->lock guards shared data against other CPUs:
67*4882a593Smuzhiyun * ehci_hcd: async, unlink, periodic (and shadow), ...
68*4882a593Smuzhiyun * usb_host_endpoint: hcpriv
69*4882a593Smuzhiyun * ehci_qh: qh_next, qtd_list
70*4882a593Smuzhiyun * ehci_qtd: qtd_list
71*4882a593Smuzhiyun *
72*4882a593Smuzhiyun * Also, hold this lock when talking to HC registers or
73*4882a593Smuzhiyun * when updating hw_* fields in shared qh/qtd/... structures.
74*4882a593Smuzhiyun */
75*4882a593Smuzhiyun
76*4882a593Smuzhiyun #define EHCI_MAX_ROOT_PORTS 15 /* see HCS_N_PORTS */
77*4882a593Smuzhiyun
78*4882a593Smuzhiyun /*
79*4882a593Smuzhiyun * ehci_rh_state values of EHCI_RH_RUNNING or above mean that the
80*4882a593Smuzhiyun * controller may be doing DMA. Lower values mean there's no DMA.
81*4882a593Smuzhiyun */
82*4882a593Smuzhiyun enum ehci_rh_state {
83*4882a593Smuzhiyun EHCI_RH_HALTED,
84*4882a593Smuzhiyun EHCI_RH_SUSPENDED,
85*4882a593Smuzhiyun EHCI_RH_RUNNING,
86*4882a593Smuzhiyun EHCI_RH_STOPPING
87*4882a593Smuzhiyun };
88*4882a593Smuzhiyun
89*4882a593Smuzhiyun /*
90*4882a593Smuzhiyun * Timer events, ordered by increasing delay length.
91*4882a593Smuzhiyun * Always update event_delays_ns[] and event_handlers[] (defined in
92*4882a593Smuzhiyun * ehci-timer.c) in parallel with this list.
93*4882a593Smuzhiyun */
94*4882a593Smuzhiyun enum ehci_hrtimer_event {
95*4882a593Smuzhiyun EHCI_HRTIMER_POLL_ASS, /* Poll for async schedule off */
96*4882a593Smuzhiyun EHCI_HRTIMER_POLL_PSS, /* Poll for periodic schedule off */
97*4882a593Smuzhiyun EHCI_HRTIMER_POLL_DEAD, /* Wait for dead controller to stop */
98*4882a593Smuzhiyun EHCI_HRTIMER_UNLINK_INTR, /* Wait for interrupt QH unlink */
99*4882a593Smuzhiyun EHCI_HRTIMER_FREE_ITDS, /* Wait for unused iTDs and siTDs */
100*4882a593Smuzhiyun EHCI_HRTIMER_ACTIVE_UNLINK, /* Wait while unlinking an active QH */
101*4882a593Smuzhiyun EHCI_HRTIMER_START_UNLINK_INTR, /* Unlink empty interrupt QHs */
102*4882a593Smuzhiyun EHCI_HRTIMER_ASYNC_UNLINKS, /* Unlink empty async QHs */
103*4882a593Smuzhiyun EHCI_HRTIMER_IAA_WATCHDOG, /* Handle lost IAA interrupts */
104*4882a593Smuzhiyun EHCI_HRTIMER_DISABLE_PERIODIC, /* Wait to disable periodic sched */
105*4882a593Smuzhiyun EHCI_HRTIMER_DISABLE_ASYNC, /* Wait to disable async sched */
106*4882a593Smuzhiyun EHCI_HRTIMER_IO_WATCHDOG, /* Check for missing IRQs */
107*4882a593Smuzhiyun EHCI_HRTIMER_NUM_EVENTS /* Must come last */
108*4882a593Smuzhiyun };
109*4882a593Smuzhiyun #define EHCI_HRTIMER_NO_EVENT 99
110*4882a593Smuzhiyun
111*4882a593Smuzhiyun struct ehci_hcd { /* one per controller */
112*4882a593Smuzhiyun /* timing support */
113*4882a593Smuzhiyun enum ehci_hrtimer_event next_hrtimer_event;
114*4882a593Smuzhiyun unsigned enabled_hrtimer_events;
115*4882a593Smuzhiyun ktime_t hr_timeouts[EHCI_HRTIMER_NUM_EVENTS];
116*4882a593Smuzhiyun struct hrtimer hrtimer;
117*4882a593Smuzhiyun
118*4882a593Smuzhiyun int PSS_poll_count;
119*4882a593Smuzhiyun int ASS_poll_count;
120*4882a593Smuzhiyun int died_poll_count;
121*4882a593Smuzhiyun
122*4882a593Smuzhiyun /* glue to PCI and HCD framework */
123*4882a593Smuzhiyun struct ehci_caps __iomem *caps;
124*4882a593Smuzhiyun struct ehci_regs __iomem *regs;
125*4882a593Smuzhiyun struct ehci_dbg_port __iomem *debug;
126*4882a593Smuzhiyun
127*4882a593Smuzhiyun __u32 hcs_params; /* cached register copy */
128*4882a593Smuzhiyun spinlock_t lock;
129*4882a593Smuzhiyun enum ehci_rh_state rh_state;
130*4882a593Smuzhiyun
131*4882a593Smuzhiyun /* general schedule support */
132*4882a593Smuzhiyun bool scanning:1;
133*4882a593Smuzhiyun bool need_rescan:1;
134*4882a593Smuzhiyun bool intr_unlinking:1;
135*4882a593Smuzhiyun bool iaa_in_progress:1;
136*4882a593Smuzhiyun bool async_unlinking:1;
137*4882a593Smuzhiyun bool shutdown:1;
138*4882a593Smuzhiyun struct ehci_qh *qh_scan_next;
139*4882a593Smuzhiyun
140*4882a593Smuzhiyun /* async schedule support */
141*4882a593Smuzhiyun struct ehci_qh *async;
142*4882a593Smuzhiyun struct ehci_qh *dummy; /* For AMD quirk use */
143*4882a593Smuzhiyun struct list_head async_unlink;
144*4882a593Smuzhiyun struct list_head async_idle;
145*4882a593Smuzhiyun unsigned async_unlink_cycle;
146*4882a593Smuzhiyun unsigned async_count; /* async activity count */
147*4882a593Smuzhiyun __hc32 old_current; /* Test for QH becoming */
148*4882a593Smuzhiyun __hc32 old_token; /* inactive during unlink */
149*4882a593Smuzhiyun
150*4882a593Smuzhiyun /* periodic schedule support */
151*4882a593Smuzhiyun #define DEFAULT_I_TDPS 1024 /* some HCs can do less */
152*4882a593Smuzhiyun unsigned periodic_size;
153*4882a593Smuzhiyun __hc32 *periodic; /* hw periodic table */
154*4882a593Smuzhiyun dma_addr_t periodic_dma;
155*4882a593Smuzhiyun struct list_head intr_qh_list;
156*4882a593Smuzhiyun unsigned i_thresh; /* uframes HC might cache */
157*4882a593Smuzhiyun
158*4882a593Smuzhiyun union ehci_shadow *pshadow; /* mirror hw periodic table */
159*4882a593Smuzhiyun struct list_head intr_unlink_wait;
160*4882a593Smuzhiyun struct list_head intr_unlink;
161*4882a593Smuzhiyun unsigned intr_unlink_wait_cycle;
162*4882a593Smuzhiyun unsigned intr_unlink_cycle;
163*4882a593Smuzhiyun unsigned now_frame; /* frame from HC hardware */
164*4882a593Smuzhiyun unsigned last_iso_frame; /* last frame scanned for iso */
165*4882a593Smuzhiyun unsigned intr_count; /* intr activity count */
166*4882a593Smuzhiyun unsigned isoc_count; /* isoc activity count */
167*4882a593Smuzhiyun unsigned periodic_count; /* periodic activity count */
168*4882a593Smuzhiyun unsigned uframe_periodic_max; /* max periodic time per uframe */
169*4882a593Smuzhiyun
170*4882a593Smuzhiyun
171*4882a593Smuzhiyun /* list of itds & sitds completed while now_frame was still active */
172*4882a593Smuzhiyun struct list_head cached_itd_list;
173*4882a593Smuzhiyun struct ehci_itd *last_itd_to_free;
174*4882a593Smuzhiyun struct list_head cached_sitd_list;
175*4882a593Smuzhiyun struct ehci_sitd *last_sitd_to_free;
176*4882a593Smuzhiyun
177*4882a593Smuzhiyun /* per root hub port */
178*4882a593Smuzhiyun unsigned long reset_done[EHCI_MAX_ROOT_PORTS];
179*4882a593Smuzhiyun
180*4882a593Smuzhiyun /* bit vectors (one bit per port) */
181*4882a593Smuzhiyun unsigned long bus_suspended; /* which ports were
182*4882a593Smuzhiyun already suspended at the start of a bus suspend */
183*4882a593Smuzhiyun unsigned long companion_ports; /* which ports are
184*4882a593Smuzhiyun dedicated to the companion controller */
185*4882a593Smuzhiyun unsigned long owned_ports; /* which ports are
186*4882a593Smuzhiyun owned by the companion during a bus suspend */
187*4882a593Smuzhiyun unsigned long port_c_suspend; /* which ports have
188*4882a593Smuzhiyun the change-suspend feature turned on */
189*4882a593Smuzhiyun unsigned long suspended_ports; /* which ports are
190*4882a593Smuzhiyun suspended */
191*4882a593Smuzhiyun unsigned long resuming_ports; /* which ports have
192*4882a593Smuzhiyun started to resume */
193*4882a593Smuzhiyun
194*4882a593Smuzhiyun /* per-HC memory pools (could be per-bus, but ...) */
195*4882a593Smuzhiyun struct dma_pool *qh_pool; /* qh per active urb */
196*4882a593Smuzhiyun struct dma_pool *qtd_pool; /* one or more per qh */
197*4882a593Smuzhiyun struct dma_pool *itd_pool; /* itd per iso urb */
198*4882a593Smuzhiyun struct dma_pool *sitd_pool; /* sitd per split iso urb */
199*4882a593Smuzhiyun
200*4882a593Smuzhiyun unsigned random_frame;
201*4882a593Smuzhiyun unsigned long next_statechange;
202*4882a593Smuzhiyun ktime_t last_periodic_enable;
203*4882a593Smuzhiyun u32 command;
204*4882a593Smuzhiyun
205*4882a593Smuzhiyun /* SILICON QUIRKS */
206*4882a593Smuzhiyun unsigned no_selective_suspend:1;
207*4882a593Smuzhiyun unsigned has_fsl_port_bug:1; /* FreeScale */
208*4882a593Smuzhiyun unsigned has_fsl_hs_errata:1; /* Freescale HS quirk */
209*4882a593Smuzhiyun unsigned has_fsl_susp_errata:1; /* NXP SUSP quirk */
210*4882a593Smuzhiyun unsigned big_endian_mmio:1;
211*4882a593Smuzhiyun unsigned big_endian_desc:1;
212*4882a593Smuzhiyun unsigned big_endian_capbase:1;
213*4882a593Smuzhiyun unsigned has_amcc_usb23:1;
214*4882a593Smuzhiyun unsigned need_io_watchdog:1;
215*4882a593Smuzhiyun unsigned amd_pll_fix:1;
216*4882a593Smuzhiyun unsigned use_dummy_qh:1; /* AMD Frame List table quirk*/
217*4882a593Smuzhiyun unsigned has_synopsys_hc_bug:1; /* Synopsys HC */
218*4882a593Smuzhiyun unsigned frame_index_bug:1; /* MosChip (AKA NetMos) */
219*4882a593Smuzhiyun unsigned need_oc_pp_cycle:1; /* MPC834X port power */
220*4882a593Smuzhiyun unsigned imx28_write_fix:1; /* For Freescale i.MX28 */
221*4882a593Smuzhiyun /*
222*4882a593Smuzhiyun * __GENKSYMS__ test is an abi workaround for commit
223*4882a593Smuzhiyun * 7f2d73788d90 ("usb: ehci: handshake CMD_RUN * instead of STS_HALT")
224*4882a593Smuzhiyun */
225*4882a593Smuzhiyun #ifndef __GENKSYMS__
226*4882a593Smuzhiyun unsigned is_aspeed:1;
227*4882a593Smuzhiyun #endif
228*4882a593Smuzhiyun
229*4882a593Smuzhiyun /* required for usb32 quirk */
230*4882a593Smuzhiyun #define OHCI_CTRL_HCFS (3 << 6)
231*4882a593Smuzhiyun #define OHCI_USB_OPER (2 << 6)
232*4882a593Smuzhiyun #define OHCI_USB_SUSPEND (3 << 6)
233*4882a593Smuzhiyun
234*4882a593Smuzhiyun #define OHCI_HCCTRL_OFFSET 0x4
235*4882a593Smuzhiyun #define OHCI_HCCTRL_LEN 0x4
236*4882a593Smuzhiyun __hc32 *ohci_hcctrl_reg;
237*4882a593Smuzhiyun unsigned has_hostpc:1;
238*4882a593Smuzhiyun unsigned has_tdi_phy_lpm:1;
239*4882a593Smuzhiyun unsigned has_ppcd:1; /* support per-port change bits */
240*4882a593Smuzhiyun u8 sbrn; /* packed release number */
241*4882a593Smuzhiyun
242*4882a593Smuzhiyun /* irq statistics */
243*4882a593Smuzhiyun #ifdef EHCI_STATS
244*4882a593Smuzhiyun struct ehci_stats stats;
245*4882a593Smuzhiyun # define INCR(x) ((x)++)
246*4882a593Smuzhiyun #else
247*4882a593Smuzhiyun # define INCR(x) do {} while (0)
248*4882a593Smuzhiyun #endif
249*4882a593Smuzhiyun
250*4882a593Smuzhiyun /* debug files */
251*4882a593Smuzhiyun #ifdef CONFIG_DYNAMIC_DEBUG
252*4882a593Smuzhiyun struct dentry *debug_dir;
253*4882a593Smuzhiyun #endif
254*4882a593Smuzhiyun
255*4882a593Smuzhiyun /* bandwidth usage */
256*4882a593Smuzhiyun #define EHCI_BANDWIDTH_SIZE 64
257*4882a593Smuzhiyun #define EHCI_BANDWIDTH_FRAMES (EHCI_BANDWIDTH_SIZE >> 3)
258*4882a593Smuzhiyun u8 bandwidth[EHCI_BANDWIDTH_SIZE];
259*4882a593Smuzhiyun /* us allocated per uframe */
260*4882a593Smuzhiyun u8 tt_budget[EHCI_BANDWIDTH_SIZE];
261*4882a593Smuzhiyun /* us budgeted per uframe */
262*4882a593Smuzhiyun struct list_head tt_list;
263*4882a593Smuzhiyun
264*4882a593Smuzhiyun /* platform-specific data -- must come last */
265*4882a593Smuzhiyun unsigned long priv[] __aligned(sizeof(s64));
266*4882a593Smuzhiyun };
267*4882a593Smuzhiyun
268*4882a593Smuzhiyun /* convert between an HCD pointer and the corresponding EHCI_HCD */
hcd_to_ehci(struct usb_hcd * hcd)269*4882a593Smuzhiyun static inline struct ehci_hcd *hcd_to_ehci(struct usb_hcd *hcd)
270*4882a593Smuzhiyun {
271*4882a593Smuzhiyun return (struct ehci_hcd *) (hcd->hcd_priv);
272*4882a593Smuzhiyun }
ehci_to_hcd(struct ehci_hcd * ehci)273*4882a593Smuzhiyun static inline struct usb_hcd *ehci_to_hcd(struct ehci_hcd *ehci)
274*4882a593Smuzhiyun {
275*4882a593Smuzhiyun return container_of((void *) ehci, struct usb_hcd, hcd_priv);
276*4882a593Smuzhiyun }
277*4882a593Smuzhiyun
278*4882a593Smuzhiyun /*-------------------------------------------------------------------------*/
279*4882a593Smuzhiyun
280*4882a593Smuzhiyun #include <linux/usb/ehci_def.h>
281*4882a593Smuzhiyun
282*4882a593Smuzhiyun /*-------------------------------------------------------------------------*/
283*4882a593Smuzhiyun
284*4882a593Smuzhiyun #define QTD_NEXT(ehci, dma) cpu_to_hc32(ehci, (u32)dma)
285*4882a593Smuzhiyun
286*4882a593Smuzhiyun /*
287*4882a593Smuzhiyun * EHCI Specification 0.95 Section 3.5
288*4882a593Smuzhiyun * QTD: describe data transfer components (buffer, direction, ...)
289*4882a593Smuzhiyun * See Fig 3-6 "Queue Element Transfer Descriptor Block Diagram".
290*4882a593Smuzhiyun *
291*4882a593Smuzhiyun * These are associated only with "QH" (Queue Head) structures,
292*4882a593Smuzhiyun * used with control, bulk, and interrupt transfers.
293*4882a593Smuzhiyun */
294*4882a593Smuzhiyun struct ehci_qtd {
295*4882a593Smuzhiyun /* first part defined by EHCI spec */
296*4882a593Smuzhiyun __hc32 hw_next; /* see EHCI 3.5.1 */
297*4882a593Smuzhiyun __hc32 hw_alt_next; /* see EHCI 3.5.2 */
298*4882a593Smuzhiyun __hc32 hw_token; /* see EHCI 3.5.3 */
299*4882a593Smuzhiyun #define QTD_TOGGLE (1 << 31) /* data toggle */
300*4882a593Smuzhiyun #define QTD_LENGTH(tok) (((tok)>>16) & 0x7fff)
301*4882a593Smuzhiyun #define QTD_IOC (1 << 15) /* interrupt on complete */
302*4882a593Smuzhiyun #define QTD_CERR(tok) (((tok)>>10) & 0x3)
303*4882a593Smuzhiyun #define QTD_PID(tok) (((tok)>>8) & 0x3)
304*4882a593Smuzhiyun #define QTD_STS_ACTIVE (1 << 7) /* HC may execute this */
305*4882a593Smuzhiyun #define QTD_STS_HALT (1 << 6) /* halted on error */
306*4882a593Smuzhiyun #define QTD_STS_DBE (1 << 5) /* data buffer error (in HC) */
307*4882a593Smuzhiyun #define QTD_STS_BABBLE (1 << 4) /* device was babbling (qtd halted) */
308*4882a593Smuzhiyun #define QTD_STS_XACT (1 << 3) /* device gave illegal response */
309*4882a593Smuzhiyun #define QTD_STS_MMF (1 << 2) /* incomplete split transaction */
310*4882a593Smuzhiyun #define QTD_STS_STS (1 << 1) /* split transaction state */
311*4882a593Smuzhiyun #define QTD_STS_PING (1 << 0) /* issue PING? */
312*4882a593Smuzhiyun
313*4882a593Smuzhiyun #define ACTIVE_BIT(ehci) cpu_to_hc32(ehci, QTD_STS_ACTIVE)
314*4882a593Smuzhiyun #define HALT_BIT(ehci) cpu_to_hc32(ehci, QTD_STS_HALT)
315*4882a593Smuzhiyun #define STATUS_BIT(ehci) cpu_to_hc32(ehci, QTD_STS_STS)
316*4882a593Smuzhiyun
317*4882a593Smuzhiyun __hc32 hw_buf[5]; /* see EHCI 3.5.4 */
318*4882a593Smuzhiyun __hc32 hw_buf_hi[5]; /* Appendix B */
319*4882a593Smuzhiyun
320*4882a593Smuzhiyun /* the rest is HCD-private */
321*4882a593Smuzhiyun dma_addr_t qtd_dma; /* qtd address */
322*4882a593Smuzhiyun struct list_head qtd_list; /* sw qtd list */
323*4882a593Smuzhiyun struct urb *urb; /* qtd's urb */
324*4882a593Smuzhiyun size_t length; /* length of buffer */
325*4882a593Smuzhiyun } __aligned(32);
326*4882a593Smuzhiyun
327*4882a593Smuzhiyun /* mask NakCnt+T in qh->hw_alt_next */
328*4882a593Smuzhiyun #define QTD_MASK(ehci) cpu_to_hc32(ehci, ~0x1f)
329*4882a593Smuzhiyun
330*4882a593Smuzhiyun #define IS_SHORT_READ(token) (QTD_LENGTH(token) != 0 && QTD_PID(token) == 1)
331*4882a593Smuzhiyun
332*4882a593Smuzhiyun /*-------------------------------------------------------------------------*/
333*4882a593Smuzhiyun
334*4882a593Smuzhiyun /* type tag from {qh,itd,sitd,fstn}->hw_next */
335*4882a593Smuzhiyun #define Q_NEXT_TYPE(ehci, dma) ((dma) & cpu_to_hc32(ehci, 3 << 1))
336*4882a593Smuzhiyun
337*4882a593Smuzhiyun /*
338*4882a593Smuzhiyun * Now the following defines are not converted using the
339*4882a593Smuzhiyun * cpu_to_le32() macro anymore, since we have to support
340*4882a593Smuzhiyun * "dynamic" switching between be and le support, so that the driver
341*4882a593Smuzhiyun * can be used on one system with SoC EHCI controller using big-endian
342*4882a593Smuzhiyun * descriptors as well as a normal little-endian PCI EHCI controller.
343*4882a593Smuzhiyun */
344*4882a593Smuzhiyun /* values for that type tag */
345*4882a593Smuzhiyun #define Q_TYPE_ITD (0 << 1)
346*4882a593Smuzhiyun #define Q_TYPE_QH (1 << 1)
347*4882a593Smuzhiyun #define Q_TYPE_SITD (2 << 1)
348*4882a593Smuzhiyun #define Q_TYPE_FSTN (3 << 1)
349*4882a593Smuzhiyun
350*4882a593Smuzhiyun /* next async queue entry, or pointer to interrupt/periodic QH */
351*4882a593Smuzhiyun #define QH_NEXT(ehci, dma) \
352*4882a593Smuzhiyun (cpu_to_hc32(ehci, (((u32) dma) & ~0x01f) | Q_TYPE_QH))
353*4882a593Smuzhiyun
354*4882a593Smuzhiyun /* for periodic/async schedules and qtd lists, mark end of list */
355*4882a593Smuzhiyun #define EHCI_LIST_END(ehci) cpu_to_hc32(ehci, 1) /* "null pointer" to hw */
356*4882a593Smuzhiyun
357*4882a593Smuzhiyun /*
358*4882a593Smuzhiyun * Entries in periodic shadow table are pointers to one of four kinds
359*4882a593Smuzhiyun * of data structure. That's dictated by the hardware; a type tag is
360*4882a593Smuzhiyun * encoded in the low bits of the hardware's periodic schedule. Use
361*4882a593Smuzhiyun * Q_NEXT_TYPE to get the tag.
362*4882a593Smuzhiyun *
363*4882a593Smuzhiyun * For entries in the async schedule, the type tag always says "qh".
364*4882a593Smuzhiyun */
365*4882a593Smuzhiyun union ehci_shadow {
366*4882a593Smuzhiyun struct ehci_qh *qh; /* Q_TYPE_QH */
367*4882a593Smuzhiyun struct ehci_itd *itd; /* Q_TYPE_ITD */
368*4882a593Smuzhiyun struct ehci_sitd *sitd; /* Q_TYPE_SITD */
369*4882a593Smuzhiyun struct ehci_fstn *fstn; /* Q_TYPE_FSTN */
370*4882a593Smuzhiyun __hc32 *hw_next; /* (all types) */
371*4882a593Smuzhiyun void *ptr;
372*4882a593Smuzhiyun };
373*4882a593Smuzhiyun
374*4882a593Smuzhiyun /*-------------------------------------------------------------------------*/
375*4882a593Smuzhiyun
376*4882a593Smuzhiyun /*
377*4882a593Smuzhiyun * EHCI Specification 0.95 Section 3.6
378*4882a593Smuzhiyun * QH: describes control/bulk/interrupt endpoints
379*4882a593Smuzhiyun * See Fig 3-7 "Queue Head Structure Layout".
380*4882a593Smuzhiyun *
381*4882a593Smuzhiyun * These appear in both the async and (for interrupt) periodic schedules.
382*4882a593Smuzhiyun */
383*4882a593Smuzhiyun
384*4882a593Smuzhiyun /* first part defined by EHCI spec */
385*4882a593Smuzhiyun struct ehci_qh_hw {
386*4882a593Smuzhiyun __hc32 hw_next; /* see EHCI 3.6.1 */
387*4882a593Smuzhiyun __hc32 hw_info1; /* see EHCI 3.6.2 */
388*4882a593Smuzhiyun #define QH_CONTROL_EP (1 << 27) /* FS/LS control endpoint */
389*4882a593Smuzhiyun #define QH_HEAD (1 << 15) /* Head of async reclamation list */
390*4882a593Smuzhiyun #define QH_TOGGLE_CTL (1 << 14) /* Data toggle control */
391*4882a593Smuzhiyun #define QH_HIGH_SPEED (2 << 12) /* Endpoint speed */
392*4882a593Smuzhiyun #define QH_LOW_SPEED (1 << 12)
393*4882a593Smuzhiyun #define QH_FULL_SPEED (0 << 12)
394*4882a593Smuzhiyun #define QH_INACTIVATE (1 << 7) /* Inactivate on next transaction */
395*4882a593Smuzhiyun __hc32 hw_info2; /* see EHCI 3.6.2 */
396*4882a593Smuzhiyun #define QH_SMASK 0x000000ff
397*4882a593Smuzhiyun #define QH_CMASK 0x0000ff00
398*4882a593Smuzhiyun #define QH_HUBADDR 0x007f0000
399*4882a593Smuzhiyun #define QH_HUBPORT 0x3f800000
400*4882a593Smuzhiyun #define QH_MULT 0xc0000000
401*4882a593Smuzhiyun __hc32 hw_current; /* qtd list - see EHCI 3.6.4 */
402*4882a593Smuzhiyun
403*4882a593Smuzhiyun /* qtd overlay (hardware parts of a struct ehci_qtd) */
404*4882a593Smuzhiyun __hc32 hw_qtd_next;
405*4882a593Smuzhiyun __hc32 hw_alt_next;
406*4882a593Smuzhiyun __hc32 hw_token;
407*4882a593Smuzhiyun __hc32 hw_buf[5];
408*4882a593Smuzhiyun __hc32 hw_buf_hi[5];
409*4882a593Smuzhiyun } __aligned(32);
410*4882a593Smuzhiyun
411*4882a593Smuzhiyun struct ehci_qh {
412*4882a593Smuzhiyun struct ehci_qh_hw *hw; /* Must come first */
413*4882a593Smuzhiyun /* the rest is HCD-private */
414*4882a593Smuzhiyun dma_addr_t qh_dma; /* address of qh */
415*4882a593Smuzhiyun union ehci_shadow qh_next; /* ptr to qh; or periodic */
416*4882a593Smuzhiyun struct list_head qtd_list; /* sw qtd list */
417*4882a593Smuzhiyun struct list_head intr_node; /* list of intr QHs */
418*4882a593Smuzhiyun struct ehci_qtd *dummy;
419*4882a593Smuzhiyun struct list_head unlink_node;
420*4882a593Smuzhiyun struct ehci_per_sched ps; /* scheduling info */
421*4882a593Smuzhiyun
422*4882a593Smuzhiyun unsigned unlink_cycle;
423*4882a593Smuzhiyun
424*4882a593Smuzhiyun u8 qh_state;
425*4882a593Smuzhiyun #define QH_STATE_LINKED 1 /* HC sees this */
426*4882a593Smuzhiyun #define QH_STATE_UNLINK 2 /* HC may still see this */
427*4882a593Smuzhiyun #define QH_STATE_IDLE 3 /* HC doesn't see this */
428*4882a593Smuzhiyun #define QH_STATE_UNLINK_WAIT 4 /* LINKED and on unlink q */
429*4882a593Smuzhiyun #define QH_STATE_COMPLETING 5 /* don't touch token.HALT */
430*4882a593Smuzhiyun
431*4882a593Smuzhiyun u8 xacterrs; /* XactErr retry counter */
432*4882a593Smuzhiyun #define QH_XACTERR_MAX 32 /* XactErr retry limit */
433*4882a593Smuzhiyun
434*4882a593Smuzhiyun u8 unlink_reason;
435*4882a593Smuzhiyun #define QH_UNLINK_HALTED 0x01 /* Halt flag is set */
436*4882a593Smuzhiyun #define QH_UNLINK_SHORT_READ 0x02 /* Recover from a short read */
437*4882a593Smuzhiyun #define QH_UNLINK_DUMMY_OVERLAY 0x04 /* QH overlayed the dummy TD */
438*4882a593Smuzhiyun #define QH_UNLINK_SHUTDOWN 0x08 /* The HC isn't running */
439*4882a593Smuzhiyun #define QH_UNLINK_QUEUE_EMPTY 0x10 /* Reached end of the queue */
440*4882a593Smuzhiyun #define QH_UNLINK_REQUESTED 0x20 /* Disable, reset, or dequeue */
441*4882a593Smuzhiyun
442*4882a593Smuzhiyun u8 gap_uf; /* uframes split/csplit gap */
443*4882a593Smuzhiyun
444*4882a593Smuzhiyun unsigned is_out:1; /* bulk or intr OUT */
445*4882a593Smuzhiyun unsigned clearing_tt:1; /* Clear-TT-Buf in progress */
446*4882a593Smuzhiyun unsigned dequeue_during_giveback:1;
447*4882a593Smuzhiyun unsigned should_be_inactive:1;
448*4882a593Smuzhiyun };
449*4882a593Smuzhiyun
450*4882a593Smuzhiyun /*-------------------------------------------------------------------------*/
451*4882a593Smuzhiyun
452*4882a593Smuzhiyun /* description of one iso transaction (up to 3 KB data if highspeed) */
453*4882a593Smuzhiyun struct ehci_iso_packet {
454*4882a593Smuzhiyun /* These will be copied to iTD when scheduling */
455*4882a593Smuzhiyun u64 bufp; /* itd->hw_bufp{,_hi}[pg] |= */
456*4882a593Smuzhiyun __hc32 transaction; /* itd->hw_transaction[i] |= */
457*4882a593Smuzhiyun u8 cross; /* buf crosses pages */
458*4882a593Smuzhiyun /* for full speed OUT splits */
459*4882a593Smuzhiyun u32 buf1;
460*4882a593Smuzhiyun };
461*4882a593Smuzhiyun
462*4882a593Smuzhiyun /* temporary schedule data for packets from iso urbs (both speeds)
463*4882a593Smuzhiyun * each packet is one logical usb transaction to the device (not TT),
464*4882a593Smuzhiyun * beginning at stream->next_uframe
465*4882a593Smuzhiyun */
466*4882a593Smuzhiyun struct ehci_iso_sched {
467*4882a593Smuzhiyun struct list_head td_list;
468*4882a593Smuzhiyun unsigned span;
469*4882a593Smuzhiyun unsigned first_packet;
470*4882a593Smuzhiyun struct ehci_iso_packet packet[];
471*4882a593Smuzhiyun };
472*4882a593Smuzhiyun
473*4882a593Smuzhiyun /*
474*4882a593Smuzhiyun * ehci_iso_stream - groups all (s)itds for this endpoint.
475*4882a593Smuzhiyun * acts like a qh would, if EHCI had them for ISO.
476*4882a593Smuzhiyun */
477*4882a593Smuzhiyun struct ehci_iso_stream {
478*4882a593Smuzhiyun /* first field matches ehci_hq, but is NULL */
479*4882a593Smuzhiyun struct ehci_qh_hw *hw;
480*4882a593Smuzhiyun
481*4882a593Smuzhiyun u8 bEndpointAddress;
482*4882a593Smuzhiyun u8 highspeed;
483*4882a593Smuzhiyun struct list_head td_list; /* queued itds/sitds */
484*4882a593Smuzhiyun struct list_head free_list; /* list of unused itds/sitds */
485*4882a593Smuzhiyun
486*4882a593Smuzhiyun /* output of (re)scheduling */
487*4882a593Smuzhiyun struct ehci_per_sched ps; /* scheduling info */
488*4882a593Smuzhiyun unsigned next_uframe;
489*4882a593Smuzhiyun __hc32 splits;
490*4882a593Smuzhiyun
491*4882a593Smuzhiyun /* the rest is derived from the endpoint descriptor,
492*4882a593Smuzhiyun * including the extra info for hw_bufp[0..2]
493*4882a593Smuzhiyun */
494*4882a593Smuzhiyun u16 uperiod; /* period in uframes */
495*4882a593Smuzhiyun u16 maxp;
496*4882a593Smuzhiyun unsigned bandwidth;
497*4882a593Smuzhiyun
498*4882a593Smuzhiyun /* This is used to initialize iTD's hw_bufp fields */
499*4882a593Smuzhiyun __hc32 buf0;
500*4882a593Smuzhiyun __hc32 buf1;
501*4882a593Smuzhiyun __hc32 buf2;
502*4882a593Smuzhiyun
503*4882a593Smuzhiyun /* this is used to initialize sITD's tt info */
504*4882a593Smuzhiyun __hc32 address;
505*4882a593Smuzhiyun };
506*4882a593Smuzhiyun
507*4882a593Smuzhiyun /*-------------------------------------------------------------------------*/
508*4882a593Smuzhiyun
509*4882a593Smuzhiyun /*
510*4882a593Smuzhiyun * EHCI Specification 0.95 Section 3.3
511*4882a593Smuzhiyun * Fig 3-4 "Isochronous Transaction Descriptor (iTD)"
512*4882a593Smuzhiyun *
513*4882a593Smuzhiyun * Schedule records for high speed iso xfers
514*4882a593Smuzhiyun */
515*4882a593Smuzhiyun struct ehci_itd {
516*4882a593Smuzhiyun /* first part defined by EHCI spec */
517*4882a593Smuzhiyun __hc32 hw_next; /* see EHCI 3.3.1 */
518*4882a593Smuzhiyun __hc32 hw_transaction[8]; /* see EHCI 3.3.2 */
519*4882a593Smuzhiyun #define EHCI_ISOC_ACTIVE (1<<31) /* activate transfer this slot */
520*4882a593Smuzhiyun #define EHCI_ISOC_BUF_ERR (1<<30) /* Data buffer error */
521*4882a593Smuzhiyun #define EHCI_ISOC_BABBLE (1<<29) /* babble detected */
522*4882a593Smuzhiyun #define EHCI_ISOC_XACTERR (1<<28) /* XactErr - transaction error */
523*4882a593Smuzhiyun #define EHCI_ITD_LENGTH(tok) (((tok)>>16) & 0x0fff)
524*4882a593Smuzhiyun #define EHCI_ITD_IOC (1 << 15) /* interrupt on complete */
525*4882a593Smuzhiyun
526*4882a593Smuzhiyun #define ITD_ACTIVE(ehci) cpu_to_hc32(ehci, EHCI_ISOC_ACTIVE)
527*4882a593Smuzhiyun
528*4882a593Smuzhiyun __hc32 hw_bufp[7]; /* see EHCI 3.3.3 */
529*4882a593Smuzhiyun __hc32 hw_bufp_hi[7]; /* Appendix B */
530*4882a593Smuzhiyun
531*4882a593Smuzhiyun /* the rest is HCD-private */
532*4882a593Smuzhiyun dma_addr_t itd_dma; /* for this itd */
533*4882a593Smuzhiyun union ehci_shadow itd_next; /* ptr to periodic q entry */
534*4882a593Smuzhiyun
535*4882a593Smuzhiyun struct urb *urb;
536*4882a593Smuzhiyun struct ehci_iso_stream *stream; /* endpoint's queue */
537*4882a593Smuzhiyun struct list_head itd_list; /* list of stream's itds */
538*4882a593Smuzhiyun
539*4882a593Smuzhiyun /* any/all hw_transactions here may be used by that urb */
540*4882a593Smuzhiyun unsigned frame; /* where scheduled */
541*4882a593Smuzhiyun unsigned pg;
542*4882a593Smuzhiyun unsigned index[8]; /* in urb->iso_frame_desc */
543*4882a593Smuzhiyun } __aligned(32);
544*4882a593Smuzhiyun
545*4882a593Smuzhiyun /*-------------------------------------------------------------------------*/
546*4882a593Smuzhiyun
547*4882a593Smuzhiyun /*
548*4882a593Smuzhiyun * EHCI Specification 0.95 Section 3.4
549*4882a593Smuzhiyun * siTD, aka split-transaction isochronous Transfer Descriptor
550*4882a593Smuzhiyun * ... describe full speed iso xfers through TT in hubs
551*4882a593Smuzhiyun * see Figure 3-5 "Split-transaction Isochronous Transaction Descriptor (siTD)
552*4882a593Smuzhiyun */
553*4882a593Smuzhiyun struct ehci_sitd {
554*4882a593Smuzhiyun /* first part defined by EHCI spec */
555*4882a593Smuzhiyun __hc32 hw_next;
556*4882a593Smuzhiyun /* uses bit field macros above - see EHCI 0.95 Table 3-8 */
557*4882a593Smuzhiyun __hc32 hw_fullspeed_ep; /* EHCI table 3-9 */
558*4882a593Smuzhiyun __hc32 hw_uframe; /* EHCI table 3-10 */
559*4882a593Smuzhiyun __hc32 hw_results; /* EHCI table 3-11 */
560*4882a593Smuzhiyun #define SITD_IOC (1 << 31) /* interrupt on completion */
561*4882a593Smuzhiyun #define SITD_PAGE (1 << 30) /* buffer 0/1 */
562*4882a593Smuzhiyun #define SITD_LENGTH(x) (((x) >> 16) & 0x3ff)
563*4882a593Smuzhiyun #define SITD_STS_ACTIVE (1 << 7) /* HC may execute this */
564*4882a593Smuzhiyun #define SITD_STS_ERR (1 << 6) /* error from TT */
565*4882a593Smuzhiyun #define SITD_STS_DBE (1 << 5) /* data buffer error (in HC) */
566*4882a593Smuzhiyun #define SITD_STS_BABBLE (1 << 4) /* device was babbling */
567*4882a593Smuzhiyun #define SITD_STS_XACT (1 << 3) /* illegal IN response */
568*4882a593Smuzhiyun #define SITD_STS_MMF (1 << 2) /* incomplete split transaction */
569*4882a593Smuzhiyun #define SITD_STS_STS (1 << 1) /* split transaction state */
570*4882a593Smuzhiyun
571*4882a593Smuzhiyun #define SITD_ACTIVE(ehci) cpu_to_hc32(ehci, SITD_STS_ACTIVE)
572*4882a593Smuzhiyun
573*4882a593Smuzhiyun __hc32 hw_buf[2]; /* EHCI table 3-12 */
574*4882a593Smuzhiyun __hc32 hw_backpointer; /* EHCI table 3-13 */
575*4882a593Smuzhiyun __hc32 hw_buf_hi[2]; /* Appendix B */
576*4882a593Smuzhiyun
577*4882a593Smuzhiyun /* the rest is HCD-private */
578*4882a593Smuzhiyun dma_addr_t sitd_dma;
579*4882a593Smuzhiyun union ehci_shadow sitd_next; /* ptr to periodic q entry */
580*4882a593Smuzhiyun
581*4882a593Smuzhiyun struct urb *urb;
582*4882a593Smuzhiyun struct ehci_iso_stream *stream; /* endpoint's queue */
583*4882a593Smuzhiyun struct list_head sitd_list; /* list of stream's sitds */
584*4882a593Smuzhiyun unsigned frame;
585*4882a593Smuzhiyun unsigned index;
586*4882a593Smuzhiyun } __aligned(32);
587*4882a593Smuzhiyun
588*4882a593Smuzhiyun /*-------------------------------------------------------------------------*/
589*4882a593Smuzhiyun
590*4882a593Smuzhiyun /*
591*4882a593Smuzhiyun * EHCI Specification 0.96 Section 3.7
592*4882a593Smuzhiyun * Periodic Frame Span Traversal Node (FSTN)
593*4882a593Smuzhiyun *
594*4882a593Smuzhiyun * Manages split interrupt transactions (using TT) that span frame boundaries
595*4882a593Smuzhiyun * into uframes 0/1; see 4.12.2.2. In those uframes, a "save place" FSTN
596*4882a593Smuzhiyun * makes the HC jump (back) to a QH to scan for fs/ls QH completions until
597*4882a593Smuzhiyun * it hits a "restore" FSTN; then it returns to finish other uframe 0/1 work.
598*4882a593Smuzhiyun */
599*4882a593Smuzhiyun struct ehci_fstn {
600*4882a593Smuzhiyun __hc32 hw_next; /* any periodic q entry */
601*4882a593Smuzhiyun __hc32 hw_prev; /* qh or EHCI_LIST_END */
602*4882a593Smuzhiyun
603*4882a593Smuzhiyun /* the rest is HCD-private */
604*4882a593Smuzhiyun dma_addr_t fstn_dma;
605*4882a593Smuzhiyun union ehci_shadow fstn_next; /* ptr to periodic q entry */
606*4882a593Smuzhiyun } __aligned(32);
607*4882a593Smuzhiyun
608*4882a593Smuzhiyun /*-------------------------------------------------------------------------*/
609*4882a593Smuzhiyun
610*4882a593Smuzhiyun /*
611*4882a593Smuzhiyun * USB-2.0 Specification Sections 11.14 and 11.18
612*4882a593Smuzhiyun * Scheduling and budgeting split transactions using TTs
613*4882a593Smuzhiyun *
614*4882a593Smuzhiyun * A hub can have a single TT for all its ports, or multiple TTs (one for each
615*4882a593Smuzhiyun * port). The bandwidth and budgeting information for the full/low-speed bus
616*4882a593Smuzhiyun * below each TT is self-contained and independent of the other TTs or the
617*4882a593Smuzhiyun * high-speed bus.
618*4882a593Smuzhiyun *
619*4882a593Smuzhiyun * "Bandwidth" refers to the number of microseconds on the FS/LS bus allocated
620*4882a593Smuzhiyun * to an interrupt or isochronous endpoint for each frame. "Budget" refers to
621*4882a593Smuzhiyun * the best-case estimate of the number of full-speed bytes allocated to an
622*4882a593Smuzhiyun * endpoint for each microframe within an allocated frame.
623*4882a593Smuzhiyun *
624*4882a593Smuzhiyun * Removal of an endpoint invalidates a TT's budget. Instead of trying to
625*4882a593Smuzhiyun * keep an up-to-date record, we recompute the budget when it is needed.
626*4882a593Smuzhiyun */
627*4882a593Smuzhiyun
628*4882a593Smuzhiyun struct ehci_tt {
629*4882a593Smuzhiyun u16 bandwidth[EHCI_BANDWIDTH_FRAMES];
630*4882a593Smuzhiyun
631*4882a593Smuzhiyun struct list_head tt_list; /* List of all ehci_tt's */
632*4882a593Smuzhiyun struct list_head ps_list; /* Items using this TT */
633*4882a593Smuzhiyun struct usb_tt *usb_tt;
634*4882a593Smuzhiyun int tt_port; /* TT port number */
635*4882a593Smuzhiyun };
636*4882a593Smuzhiyun
637*4882a593Smuzhiyun /*-------------------------------------------------------------------------*/
638*4882a593Smuzhiyun
639*4882a593Smuzhiyun /* Prepare the PORTSC wakeup flags during controller suspend/resume */
640*4882a593Smuzhiyun
641*4882a593Smuzhiyun #define ehci_prepare_ports_for_controller_suspend(ehci, do_wakeup) \
642*4882a593Smuzhiyun ehci_adjust_port_wakeup_flags(ehci, true, do_wakeup)
643*4882a593Smuzhiyun
644*4882a593Smuzhiyun #define ehci_prepare_ports_for_controller_resume(ehci) \
645*4882a593Smuzhiyun ehci_adjust_port_wakeup_flags(ehci, false, false)
646*4882a593Smuzhiyun
647*4882a593Smuzhiyun /*-------------------------------------------------------------------------*/
648*4882a593Smuzhiyun
649*4882a593Smuzhiyun #ifdef CONFIG_USB_EHCI_ROOT_HUB_TT
650*4882a593Smuzhiyun
651*4882a593Smuzhiyun /*
652*4882a593Smuzhiyun * Some EHCI controllers have a Transaction Translator built into the
653*4882a593Smuzhiyun * root hub. This is a non-standard feature. Each controller will need
654*4882a593Smuzhiyun * to add code to the following inline functions, and call them as
655*4882a593Smuzhiyun * needed (mostly in root hub code).
656*4882a593Smuzhiyun */
657*4882a593Smuzhiyun
658*4882a593Smuzhiyun #define ehci_is_TDI(e) (ehci_to_hcd(e)->has_tt)
659*4882a593Smuzhiyun
660*4882a593Smuzhiyun /* Returns the speed of a device attached to a port on the root hub. */
661*4882a593Smuzhiyun static inline unsigned int
ehci_port_speed(struct ehci_hcd * ehci,unsigned int portsc)662*4882a593Smuzhiyun ehci_port_speed(struct ehci_hcd *ehci, unsigned int portsc)
663*4882a593Smuzhiyun {
664*4882a593Smuzhiyun if (ehci_is_TDI(ehci)) {
665*4882a593Smuzhiyun switch ((portsc >> (ehci->has_hostpc ? 25 : 26)) & 3) {
666*4882a593Smuzhiyun case 0:
667*4882a593Smuzhiyun return 0;
668*4882a593Smuzhiyun case 1:
669*4882a593Smuzhiyun return USB_PORT_STAT_LOW_SPEED;
670*4882a593Smuzhiyun case 2:
671*4882a593Smuzhiyun default:
672*4882a593Smuzhiyun return USB_PORT_STAT_HIGH_SPEED;
673*4882a593Smuzhiyun }
674*4882a593Smuzhiyun }
675*4882a593Smuzhiyun return USB_PORT_STAT_HIGH_SPEED;
676*4882a593Smuzhiyun }
677*4882a593Smuzhiyun
678*4882a593Smuzhiyun #else
679*4882a593Smuzhiyun
680*4882a593Smuzhiyun #define ehci_is_TDI(e) (0)
681*4882a593Smuzhiyun
682*4882a593Smuzhiyun #define ehci_port_speed(ehci, portsc) USB_PORT_STAT_HIGH_SPEED
683*4882a593Smuzhiyun #endif
684*4882a593Smuzhiyun
685*4882a593Smuzhiyun /*-------------------------------------------------------------------------*/
686*4882a593Smuzhiyun
687*4882a593Smuzhiyun #ifdef CONFIG_PPC_83xx
688*4882a593Smuzhiyun /* Some Freescale processors have an erratum in which the TT
689*4882a593Smuzhiyun * port number in the queue head was 0..N-1 instead of 1..N.
690*4882a593Smuzhiyun */
691*4882a593Smuzhiyun #define ehci_has_fsl_portno_bug(e) ((e)->has_fsl_port_bug)
692*4882a593Smuzhiyun #else
693*4882a593Smuzhiyun #define ehci_has_fsl_portno_bug(e) (0)
694*4882a593Smuzhiyun #endif
695*4882a593Smuzhiyun
696*4882a593Smuzhiyun #define PORTSC_FSL_PFSC 24 /* Port Force Full-Speed Connect */
697*4882a593Smuzhiyun
698*4882a593Smuzhiyun #if defined(CONFIG_PPC_85xx)
699*4882a593Smuzhiyun /* Some Freescale processors have an erratum (USB A-005275) in which
700*4882a593Smuzhiyun * incoming packets get corrupted in HS mode
701*4882a593Smuzhiyun */
702*4882a593Smuzhiyun #define ehci_has_fsl_hs_errata(e) ((e)->has_fsl_hs_errata)
703*4882a593Smuzhiyun #else
704*4882a593Smuzhiyun #define ehci_has_fsl_hs_errata(e) (0)
705*4882a593Smuzhiyun #endif
706*4882a593Smuzhiyun
707*4882a593Smuzhiyun /*
708*4882a593Smuzhiyun * Some Freescale/NXP processors have an erratum (USB A-005697)
709*4882a593Smuzhiyun * in which we need to wait for 10ms for bus to enter suspend mode
710*4882a593Smuzhiyun * after setting SUSP bit.
711*4882a593Smuzhiyun */
712*4882a593Smuzhiyun #define ehci_has_fsl_susp_errata(e) ((e)->has_fsl_susp_errata)
713*4882a593Smuzhiyun
714*4882a593Smuzhiyun /*
715*4882a593Smuzhiyun * While most USB host controllers implement their registers in
716*4882a593Smuzhiyun * little-endian format, a minority (celleb companion chip) implement
717*4882a593Smuzhiyun * them in big endian format.
718*4882a593Smuzhiyun *
719*4882a593Smuzhiyun * This attempts to support either format at compile time without a
720*4882a593Smuzhiyun * runtime penalty, or both formats with the additional overhead
721*4882a593Smuzhiyun * of checking a flag bit.
722*4882a593Smuzhiyun *
723*4882a593Smuzhiyun * ehci_big_endian_capbase is a special quirk for controllers that
724*4882a593Smuzhiyun * implement the HC capability registers as separate registers and not
725*4882a593Smuzhiyun * as fields of a 32-bit register.
726*4882a593Smuzhiyun */
727*4882a593Smuzhiyun
728*4882a593Smuzhiyun #ifdef CONFIG_USB_EHCI_BIG_ENDIAN_MMIO
729*4882a593Smuzhiyun #define ehci_big_endian_mmio(e) ((e)->big_endian_mmio)
730*4882a593Smuzhiyun #define ehci_big_endian_capbase(e) ((e)->big_endian_capbase)
731*4882a593Smuzhiyun #else
732*4882a593Smuzhiyun #define ehci_big_endian_mmio(e) 0
733*4882a593Smuzhiyun #define ehci_big_endian_capbase(e) 0
734*4882a593Smuzhiyun #endif
735*4882a593Smuzhiyun
736*4882a593Smuzhiyun /*
737*4882a593Smuzhiyun * Big-endian read/write functions are arch-specific.
738*4882a593Smuzhiyun * Other arches can be added if/when they're needed.
739*4882a593Smuzhiyun */
740*4882a593Smuzhiyun #if defined(CONFIG_ARM) && defined(CONFIG_ARCH_IXP4XX)
741*4882a593Smuzhiyun #define readl_be(addr) __raw_readl((__force unsigned *)addr)
742*4882a593Smuzhiyun #define writel_be(val, addr) __raw_writel(val, (__force unsigned *)addr)
743*4882a593Smuzhiyun #endif
744*4882a593Smuzhiyun
ehci_readl(const struct ehci_hcd * ehci,__u32 __iomem * regs)745*4882a593Smuzhiyun static inline unsigned int ehci_readl(const struct ehci_hcd *ehci,
746*4882a593Smuzhiyun __u32 __iomem *regs)
747*4882a593Smuzhiyun {
748*4882a593Smuzhiyun #ifdef CONFIG_USB_EHCI_BIG_ENDIAN_MMIO
749*4882a593Smuzhiyun return ehci_big_endian_mmio(ehci) ?
750*4882a593Smuzhiyun readl_be(regs) :
751*4882a593Smuzhiyun readl(regs);
752*4882a593Smuzhiyun #else
753*4882a593Smuzhiyun return readl(regs);
754*4882a593Smuzhiyun #endif
755*4882a593Smuzhiyun }
756*4882a593Smuzhiyun
757*4882a593Smuzhiyun #ifdef CONFIG_SOC_IMX28
imx28_ehci_writel(const unsigned int val,volatile __u32 __iomem * addr)758*4882a593Smuzhiyun static inline void imx28_ehci_writel(const unsigned int val,
759*4882a593Smuzhiyun volatile __u32 __iomem *addr)
760*4882a593Smuzhiyun {
761*4882a593Smuzhiyun __asm__ ("swp %0, %0, [%1]" : : "r"(val), "r"(addr));
762*4882a593Smuzhiyun }
763*4882a593Smuzhiyun #else
imx28_ehci_writel(const unsigned int val,volatile __u32 __iomem * addr)764*4882a593Smuzhiyun static inline void imx28_ehci_writel(const unsigned int val,
765*4882a593Smuzhiyun volatile __u32 __iomem *addr)
766*4882a593Smuzhiyun {
767*4882a593Smuzhiyun }
768*4882a593Smuzhiyun #endif
ehci_writel(const struct ehci_hcd * ehci,const unsigned int val,__u32 __iomem * regs)769*4882a593Smuzhiyun static inline void ehci_writel(const struct ehci_hcd *ehci,
770*4882a593Smuzhiyun const unsigned int val, __u32 __iomem *regs)
771*4882a593Smuzhiyun {
772*4882a593Smuzhiyun #ifdef CONFIG_USB_EHCI_BIG_ENDIAN_MMIO
773*4882a593Smuzhiyun ehci_big_endian_mmio(ehci) ?
774*4882a593Smuzhiyun writel_be(val, regs) :
775*4882a593Smuzhiyun writel(val, regs);
776*4882a593Smuzhiyun #else
777*4882a593Smuzhiyun if (ehci->imx28_write_fix)
778*4882a593Smuzhiyun imx28_ehci_writel(val, regs);
779*4882a593Smuzhiyun else
780*4882a593Smuzhiyun writel(val, regs);
781*4882a593Smuzhiyun #endif
782*4882a593Smuzhiyun }
783*4882a593Smuzhiyun
784*4882a593Smuzhiyun /*
785*4882a593Smuzhiyun * On certain ppc-44x SoC there is a HW issue, that could only worked around with
786*4882a593Smuzhiyun * explicit suspend/operate of OHCI. This function hereby makes sense only on that arch.
787*4882a593Smuzhiyun * Other common bits are dependent on has_amcc_usb23 quirk flag.
788*4882a593Smuzhiyun */
789*4882a593Smuzhiyun #ifdef CONFIG_44x
set_ohci_hcfs(struct ehci_hcd * ehci,int operational)790*4882a593Smuzhiyun static inline void set_ohci_hcfs(struct ehci_hcd *ehci, int operational)
791*4882a593Smuzhiyun {
792*4882a593Smuzhiyun u32 hc_control;
793*4882a593Smuzhiyun
794*4882a593Smuzhiyun hc_control = (readl_be(ehci->ohci_hcctrl_reg) & ~OHCI_CTRL_HCFS);
795*4882a593Smuzhiyun if (operational)
796*4882a593Smuzhiyun hc_control |= OHCI_USB_OPER;
797*4882a593Smuzhiyun else
798*4882a593Smuzhiyun hc_control |= OHCI_USB_SUSPEND;
799*4882a593Smuzhiyun
800*4882a593Smuzhiyun writel_be(hc_control, ehci->ohci_hcctrl_reg);
801*4882a593Smuzhiyun (void) readl_be(ehci->ohci_hcctrl_reg);
802*4882a593Smuzhiyun }
803*4882a593Smuzhiyun #else
set_ohci_hcfs(struct ehci_hcd * ehci,int operational)804*4882a593Smuzhiyun static inline void set_ohci_hcfs(struct ehci_hcd *ehci, int operational)
805*4882a593Smuzhiyun { }
806*4882a593Smuzhiyun #endif
807*4882a593Smuzhiyun
808*4882a593Smuzhiyun /*-------------------------------------------------------------------------*/
809*4882a593Smuzhiyun
810*4882a593Smuzhiyun /*
811*4882a593Smuzhiyun * The AMCC 440EPx not only implements its EHCI registers in big-endian
812*4882a593Smuzhiyun * format, but also its DMA data structures (descriptors).
813*4882a593Smuzhiyun *
814*4882a593Smuzhiyun * EHCI controllers accessed through PCI work normally (little-endian
815*4882a593Smuzhiyun * everywhere), so we won't bother supporting a BE-only mode for now.
816*4882a593Smuzhiyun */
817*4882a593Smuzhiyun #ifdef CONFIG_USB_EHCI_BIG_ENDIAN_DESC
818*4882a593Smuzhiyun #define ehci_big_endian_desc(e) ((e)->big_endian_desc)
819*4882a593Smuzhiyun
820*4882a593Smuzhiyun /* cpu to ehci */
cpu_to_hc32(const struct ehci_hcd * ehci,const u32 x)821*4882a593Smuzhiyun static inline __hc32 cpu_to_hc32(const struct ehci_hcd *ehci, const u32 x)
822*4882a593Smuzhiyun {
823*4882a593Smuzhiyun return ehci_big_endian_desc(ehci)
824*4882a593Smuzhiyun ? (__force __hc32)cpu_to_be32(x)
825*4882a593Smuzhiyun : (__force __hc32)cpu_to_le32(x);
826*4882a593Smuzhiyun }
827*4882a593Smuzhiyun
828*4882a593Smuzhiyun /* ehci to cpu */
hc32_to_cpu(const struct ehci_hcd * ehci,const __hc32 x)829*4882a593Smuzhiyun static inline u32 hc32_to_cpu(const struct ehci_hcd *ehci, const __hc32 x)
830*4882a593Smuzhiyun {
831*4882a593Smuzhiyun return ehci_big_endian_desc(ehci)
832*4882a593Smuzhiyun ? be32_to_cpu((__force __be32)x)
833*4882a593Smuzhiyun : le32_to_cpu((__force __le32)x);
834*4882a593Smuzhiyun }
835*4882a593Smuzhiyun
hc32_to_cpup(const struct ehci_hcd * ehci,const __hc32 * x)836*4882a593Smuzhiyun static inline u32 hc32_to_cpup(const struct ehci_hcd *ehci, const __hc32 *x)
837*4882a593Smuzhiyun {
838*4882a593Smuzhiyun return ehci_big_endian_desc(ehci)
839*4882a593Smuzhiyun ? be32_to_cpup((__force __be32 *)x)
840*4882a593Smuzhiyun : le32_to_cpup((__force __le32 *)x);
841*4882a593Smuzhiyun }
842*4882a593Smuzhiyun
843*4882a593Smuzhiyun #else
844*4882a593Smuzhiyun
845*4882a593Smuzhiyun /* cpu to ehci */
cpu_to_hc32(const struct ehci_hcd * ehci,const u32 x)846*4882a593Smuzhiyun static inline __hc32 cpu_to_hc32(const struct ehci_hcd *ehci, const u32 x)
847*4882a593Smuzhiyun {
848*4882a593Smuzhiyun return cpu_to_le32(x);
849*4882a593Smuzhiyun }
850*4882a593Smuzhiyun
851*4882a593Smuzhiyun /* ehci to cpu */
hc32_to_cpu(const struct ehci_hcd * ehci,const __hc32 x)852*4882a593Smuzhiyun static inline u32 hc32_to_cpu(const struct ehci_hcd *ehci, const __hc32 x)
853*4882a593Smuzhiyun {
854*4882a593Smuzhiyun return le32_to_cpu(x);
855*4882a593Smuzhiyun }
856*4882a593Smuzhiyun
hc32_to_cpup(const struct ehci_hcd * ehci,const __hc32 * x)857*4882a593Smuzhiyun static inline u32 hc32_to_cpup(const struct ehci_hcd *ehci, const __hc32 *x)
858*4882a593Smuzhiyun {
859*4882a593Smuzhiyun return le32_to_cpup(x);
860*4882a593Smuzhiyun }
861*4882a593Smuzhiyun
862*4882a593Smuzhiyun #endif
863*4882a593Smuzhiyun
864*4882a593Smuzhiyun /*-------------------------------------------------------------------------*/
865*4882a593Smuzhiyun
866*4882a593Smuzhiyun #define ehci_dbg(ehci, fmt, args...) \
867*4882a593Smuzhiyun dev_dbg(ehci_to_hcd(ehci)->self.controller, fmt, ## args)
868*4882a593Smuzhiyun #define ehci_err(ehci, fmt, args...) \
869*4882a593Smuzhiyun dev_err(ehci_to_hcd(ehci)->self.controller, fmt, ## args)
870*4882a593Smuzhiyun #define ehci_info(ehci, fmt, args...) \
871*4882a593Smuzhiyun dev_info(ehci_to_hcd(ehci)->self.controller, fmt, ## args)
872*4882a593Smuzhiyun #define ehci_warn(ehci, fmt, args...) \
873*4882a593Smuzhiyun dev_warn(ehci_to_hcd(ehci)->self.controller, fmt, ## args)
874*4882a593Smuzhiyun
875*4882a593Smuzhiyun /*-------------------------------------------------------------------------*/
876*4882a593Smuzhiyun
877*4882a593Smuzhiyun /* Declarations of things exported for use by ehci platform drivers */
878*4882a593Smuzhiyun
879*4882a593Smuzhiyun struct ehci_driver_overrides {
880*4882a593Smuzhiyun size_t extra_priv_size;
881*4882a593Smuzhiyun int (*reset)(struct usb_hcd *hcd);
882*4882a593Smuzhiyun int (*port_power)(struct usb_hcd *hcd,
883*4882a593Smuzhiyun int portnum, bool enable);
884*4882a593Smuzhiyun };
885*4882a593Smuzhiyun
886*4882a593Smuzhiyun extern void ehci_init_driver(struct hc_driver *drv,
887*4882a593Smuzhiyun const struct ehci_driver_overrides *over);
888*4882a593Smuzhiyun extern int ehci_setup(struct usb_hcd *hcd);
889*4882a593Smuzhiyun extern int ehci_handshake(struct ehci_hcd *ehci, void __iomem *ptr,
890*4882a593Smuzhiyun u32 mask, u32 done, int usec);
891*4882a593Smuzhiyun extern int ehci_reset(struct ehci_hcd *ehci);
892*4882a593Smuzhiyun
893*4882a593Smuzhiyun extern int ehci_suspend(struct usb_hcd *hcd, bool do_wakeup);
894*4882a593Smuzhiyun extern int ehci_resume(struct usb_hcd *hcd, bool force_reset);
895*4882a593Smuzhiyun extern void ehci_adjust_port_wakeup_flags(struct ehci_hcd *ehci,
896*4882a593Smuzhiyun bool suspending, bool do_wakeup);
897*4882a593Smuzhiyun
898*4882a593Smuzhiyun extern int ehci_hub_control(struct usb_hcd *hcd, u16 typeReq, u16 wValue,
899*4882a593Smuzhiyun u16 wIndex, char *buf, u16 wLength);
900*4882a593Smuzhiyun
901*4882a593Smuzhiyun #endif /* __LINUX_EHCI_HCD_H */
902