xref: /OK3568_Linux_fs/external/rkwifibt/drivers/rtl8852be/include/osdep_service_bsd.h (revision 4882a59341e53eb6f0b4789bf948001014eff981)
1*4882a593Smuzhiyun /******************************************************************************
2*4882a593Smuzhiyun  *
3*4882a593Smuzhiyun  * Copyright(c) 2007 - 2019 Realtek Corporation.
4*4882a593Smuzhiyun  *
5*4882a593Smuzhiyun  * This program is free software; you can redistribute it and/or modify it
6*4882a593Smuzhiyun  * under the terms of version 2 of the GNU General Public License as
7*4882a593Smuzhiyun  * published by the Free Software Foundation.
8*4882a593Smuzhiyun  *
9*4882a593Smuzhiyun  * This program is distributed in the hope that it will be useful, but WITHOUT
10*4882a593Smuzhiyun  * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
11*4882a593Smuzhiyun  * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
12*4882a593Smuzhiyun  * more details.
13*4882a593Smuzhiyun  *
14*4882a593Smuzhiyun  *****************************************************************************/
15*4882a593Smuzhiyun #ifndef __OSDEP_BSD_SERVICE_H_
16*4882a593Smuzhiyun #define __OSDEP_BSD_SERVICE_H_
17*4882a593Smuzhiyun 
18*4882a593Smuzhiyun 
19*4882a593Smuzhiyun #include <sys/cdefs.h>
20*4882a593Smuzhiyun #include <sys/types.h>
21*4882a593Smuzhiyun #include <sys/systm.h>
22*4882a593Smuzhiyun #include <sys/param.h>
23*4882a593Smuzhiyun #include <sys/sockio.h>
24*4882a593Smuzhiyun #include <sys/sysctl.h>
25*4882a593Smuzhiyun #include <sys/lock.h>
26*4882a593Smuzhiyun #include <sys/mutex.h>
27*4882a593Smuzhiyun #include <sys/mbuf.h>
28*4882a593Smuzhiyun #include <sys/kernel.h>
29*4882a593Smuzhiyun #include <sys/socket.h>
30*4882a593Smuzhiyun #include <sys/systm.h>
31*4882a593Smuzhiyun #include <sys/malloc.h>
32*4882a593Smuzhiyun #include <sys/module.h>
33*4882a593Smuzhiyun #include <sys/bus.h>
34*4882a593Smuzhiyun #include <sys/endian.h>
35*4882a593Smuzhiyun #include <sys/kdb.h>
36*4882a593Smuzhiyun #include <sys/kthread.h>
37*4882a593Smuzhiyun #include <sys/malloc.h>
38*4882a593Smuzhiyun #include <sys/time.h>
39*4882a593Smuzhiyun #include <machine/atomic.h>
40*4882a593Smuzhiyun #include <machine/bus.h>
41*4882a593Smuzhiyun #include <machine/resource.h>
42*4882a593Smuzhiyun #include <sys/rman.h>
43*4882a593Smuzhiyun 
44*4882a593Smuzhiyun #include <net/bpf.h>
45*4882a593Smuzhiyun #include <net/if.h>
46*4882a593Smuzhiyun #include <net/if_arp.h>
47*4882a593Smuzhiyun #include <net/ethernet.h>
48*4882a593Smuzhiyun #include <net/if_dl.h>
49*4882a593Smuzhiyun #include <net/if_media.h>
50*4882a593Smuzhiyun #include <net/if_types.h>
51*4882a593Smuzhiyun #include <net/route.h>
52*4882a593Smuzhiyun 
53*4882a593Smuzhiyun 
54*4882a593Smuzhiyun #include <netinet/in.h>
55*4882a593Smuzhiyun #include <netinet/in_systm.h>
56*4882a593Smuzhiyun #include <netinet/in_var.h>
57*4882a593Smuzhiyun #include <netinet/if_ether.h>
58*4882a593Smuzhiyun #include <if_ether.h>
59*4882a593Smuzhiyun 
60*4882a593Smuzhiyun #include <net80211/ieee80211_var.h>
61*4882a593Smuzhiyun #include <net80211/ieee80211_regdomain.h>
62*4882a593Smuzhiyun #include <net80211/ieee80211_radiotap.h>
63*4882a593Smuzhiyun #include <net80211/ieee80211_ratectl.h>
64*4882a593Smuzhiyun 
65*4882a593Smuzhiyun #include <dev/usb/usb.h>
66*4882a593Smuzhiyun #include <dev/usb/usbdi.h>
67*4882a593Smuzhiyun #include "usbdevs.h"
68*4882a593Smuzhiyun 
69*4882a593Smuzhiyun #define	USB_DEBUG_VAR rum_debug
70*4882a593Smuzhiyun #include <dev/usb/usb_debug.h>
71*4882a593Smuzhiyun 
72*4882a593Smuzhiyun #if 1 //Baron porting from linux, it's all temp solution, needs to check again
73*4882a593Smuzhiyun #include <sys/sema.h>
74*4882a593Smuzhiyun #include <sys/pcpu.h> /* XXX for PCPU_GET */
75*4882a593Smuzhiyun //	typedef struct 	semaphore _sema;
76*4882a593Smuzhiyun 	typedef struct 	sema _sema;
77*4882a593Smuzhiyun //	typedef	spinlock_t	_lock;
78*4882a593Smuzhiyun 	typedef	struct mtx	_lock;
79*4882a593Smuzhiyun 	typedef struct mtx 		_mutex;
80*4882a593Smuzhiyun 	typedef struct rtw_timer_list _timer;
81*4882a593Smuzhiyun 	struct list_head {
82*4882a593Smuzhiyun 	struct list_head *next, *prev;
83*4882a593Smuzhiyun 	};
84*4882a593Smuzhiyun 	struct	__queue	{
85*4882a593Smuzhiyun 		struct	list_head	queue;
86*4882a593Smuzhiyun 		_lock	lock;
87*4882a593Smuzhiyun 	};
88*4882a593Smuzhiyun 
89*4882a593Smuzhiyun 	typedef struct mbuf	_buffer;
90*4882a593Smuzhiyun 
91*4882a593Smuzhiyun 	typedef struct	__queue	_queue;
92*4882a593Smuzhiyun 	typedef struct	list_head	_list;
93*4882a593Smuzhiyun 
94*4882a593Smuzhiyun 	typedef	struct	ifnet * _nic_hdl;
95*4882a593Smuzhiyun 
96*4882a593Smuzhiyun 	typedef pid_t		_thread_hdl_;
97*4882a593Smuzhiyun //	typedef struct thread		_thread_hdl_;
98*4882a593Smuzhiyun 	typedef void		thread_return;
99*4882a593Smuzhiyun 	typedef void*	thread_context;
100*4882a593Smuzhiyun 
101*4882a593Smuzhiyun 	typedef void timer_hdl_return;
102*4882a593Smuzhiyun 	typedef void* timer_hdl_context;
103*4882a593Smuzhiyun 	typedef struct work_struct _workitem;
104*4882a593Smuzhiyun 	typedef struct task _tasklet;
105*4882a593Smuzhiyun 
106*4882a593Smuzhiyun #define   KERNEL_VERSION(a,b,c) (((a) << 16) + ((b) << 8) + (c))
107*4882a593Smuzhiyun /* emulate a modern version */
108*4882a593Smuzhiyun #define LINUX_VERSION_CODE KERNEL_VERSION(2, 6, 35)
109*4882a593Smuzhiyun 
110*4882a593Smuzhiyun #define WIRELESS_EXT -1
111*4882a593Smuzhiyun #define HZ hz
112*4882a593Smuzhiyun 
113*4882a593Smuzhiyun //#define IFT_RTW	0xf9 //ifnet allocate type for RTW
114*4882a593Smuzhiyun #define free_netdev if_free
115*4882a593Smuzhiyun #define LIST_CONTAINOR(ptr, type, member) \
116*4882a593Smuzhiyun         ((type *)((char *)(ptr)-(SIZE_T)(&((type *)0)->member)))
117*4882a593Smuzhiyun #define container_of(p,t,n) (t*)((p)-&(((t*)0)->n))
118*4882a593Smuzhiyun 
119*4882a593Smuzhiyun /*lock - spinlock*/
_rtw_spinlock_init(_lock * plock)120*4882a593Smuzhiyun static inline void _rtw_spinlock_init(_lock *plock)
121*4882a593Smuzhiyun {
122*4882a593Smuzhiyun 	mtx_init(plock, "", NULL, MTX_DEF | MTX_RECURSE);
123*4882a593Smuzhiyun }
124*4882a593Smuzhiyun 
_rtw_spinlock_free(_lock * plock)125*4882a593Smuzhiyun static inline void _rtw_spinlock_free(_lock *plock)
126*4882a593Smuzhiyun {
127*4882a593Smuzhiyun 	mtx_destroy(plock);
128*4882a593Smuzhiyun }
129*4882a593Smuzhiyun 
_rtw_spinlock(_lock * plock)130*4882a593Smuzhiyun static inline void _rtw_spinlock(_lock	*plock)
131*4882a593Smuzhiyun {
132*4882a593Smuzhiyun 	mtx_lock(plock);
133*4882a593Smuzhiyun }
134*4882a593Smuzhiyun 
_rtw_spinunlock(_lock * plock)135*4882a593Smuzhiyun static inline void _rtw_spinunlock(_lock *plock)
136*4882a593Smuzhiyun {
137*4882a593Smuzhiyun 	mtx_unlock(plock);
138*4882a593Smuzhiyun }
139*4882a593Smuzhiyun 
140*4882a593Smuzhiyun #if 0
_rtw_spinlock_ex(_lock * plock)141*4882a593Smuzhiyun static inline void _rtw_spinlock_ex(_lock *plock)
142*4882a593Smuzhiyun {
143*4882a593Smuzhiyun 	mtx_lock(plock);
144*4882a593Smuzhiyun }
145*4882a593Smuzhiyun 
_rtw_spinunlock_ex(_lock * plock)146*4882a593Smuzhiyun static inline void _rtw_spinunlock_ex(_lock *plock)
147*4882a593Smuzhiyun {
148*4882a593Smuzhiyun 	mtx_unlock(plock);
149*4882a593Smuzhiyun }
150*4882a593Smuzhiyun #endif
151*4882a593Smuzhiyun 
_rtw_spinlock_irq(_lock * plock,unsigned long * flags)152*4882a593Smuzhiyun __inline static void _rtw_spinlock_irq(_lock *plock, unsigned long *flags)
153*4882a593Smuzhiyun {
154*4882a593Smuzhiyun 	mtx_lock(plock); /*{local_irq_save((x)); mtx_lock_spin((lock));}*/
155*4882a593Smuzhiyun }
156*4882a593Smuzhiyun 
_rtw_spinunlock_irq(_lock * plock,unsigned long * flags)157*4882a593Smuzhiyun __inline static void _rtw_spinunlock_irq(_lock *plock, unsigned long *flags)
158*4882a593Smuzhiyun {
159*4882a593Smuzhiyun 	mtx_unlock(plock);
160*4882a593Smuzhiyun }
161*4882a593Smuzhiyun 
_rtw_spinlock_bh(_lock * plock)162*4882a593Smuzhiyun __inline static void _rtw_spinlock_bh(_lock *plock)
163*4882a593Smuzhiyun {
164*4882a593Smuzhiyun 	mtx_lock(plock);/*{local_irq_save((x)); mtx_lock_spin((lock));}*/
165*4882a593Smuzhiyun }
166*4882a593Smuzhiyun 
_rtw_spinunlock_bh(_lock * plock)167*4882a593Smuzhiyun __inline static void _rtw_spinunlock_bh(_lock *plock)
168*4882a593Smuzhiyun {
169*4882a593Smuzhiyun 	mtx_unlock(plock);
170*4882a593Smuzhiyun }
171*4882a593Smuzhiyun 
172*4882a593Smuzhiyun /*lock - semaphore*/
_rtw_init_sema(_sema * sema,int init_val)173*4882a593Smuzhiyun static inline void _rtw_init_sema(_sema *sema, int init_val)
174*4882a593Smuzhiyun {
175*4882a593Smuzhiyun 	sema_init(sema, init_val, "rtw_drv");
176*4882a593Smuzhiyun }
177*4882a593Smuzhiyun 
_rtw_free_sema(_sema * sema)178*4882a593Smuzhiyun static inline void _rtw_free_sema(_sema *sema)
179*4882a593Smuzhiyun {
180*4882a593Smuzhiyun 	sema_destroy(sema);
181*4882a593Smuzhiyun }
182*4882a593Smuzhiyun 
_rtw_up_sema(_sema * sema)183*4882a593Smuzhiyun static inline void _rtw_up_sema(_sema *sema)
184*4882a593Smuzhiyun {
185*4882a593Smuzhiyun 	sema_post(sema);
186*4882a593Smuzhiyun }
187*4882a593Smuzhiyun 
_rtw_down_sema(_sema * sema)188*4882a593Smuzhiyun static inline u32 _rtw_down_sema(_sema *sema)
189*4882a593Smuzhiyun {
190*4882a593Smuzhiyun 	sema_wait(sema);
191*4882a593Smuzhiyun 	return  _SUCCESS;
192*4882a593Smuzhiyun }
193*4882a593Smuzhiyun 
194*4882a593Smuzhiyun /*lock - mutex*/
_rtw_mutex_init(_mutex * pmutex)195*4882a593Smuzhiyun static inline void _rtw_mutex_init(_mutex *pmutex)
196*4882a593Smuzhiyun {
197*4882a593Smuzhiyun 	mtx_init(pmutex, "", NULL, MTX_DEF | MTX_RECURSE);
198*4882a593Smuzhiyun }
199*4882a593Smuzhiyun 
_rtw_mutex_free(_mutex * pmutex)200*4882a593Smuzhiyun static inline void _rtw_mutex_free(_mutex *pmutex)
201*4882a593Smuzhiyun {
202*4882a593Smuzhiyun 	sema_destroy(pmutex);
203*4882a593Smuzhiyun }
204*4882a593Smuzhiyun 
_rtw_mutex_lock_interruptible(_mutex * pmutex)205*4882a593Smuzhiyun __inline static void _rtw_mutex_lock_interruptible(_mutex *pmutex)
206*4882a593Smuzhiyun {
207*4882a593Smuzhiyun 
208*4882a593Smuzhiyun 	mtx_lock(pmutex);
209*4882a593Smuzhiyun 
210*4882a593Smuzhiyun }
_rtw_mutex_lock(_mutex * pmutex)211*4882a593Smuzhiyun __inline static void _rtw_mutex_lock(_mutex *pmutex)
212*4882a593Smuzhiyun {
213*4882a593Smuzhiyun 
214*4882a593Smuzhiyun 	mtx_lock(pmutex);
215*4882a593Smuzhiyun 
216*4882a593Smuzhiyun }
217*4882a593Smuzhiyun 
_rtw_mutex_unlock(_mutex * pmutex)218*4882a593Smuzhiyun __inline static void _rtw_mutex_unlock(_mutex *pmutex)
219*4882a593Smuzhiyun {
220*4882a593Smuzhiyun 	mtx_unlock(pmutex);
221*4882a593Smuzhiyun }
222*4882a593Smuzhiyun 
_rtw_vmalloc(u32 sz)223*4882a593Smuzhiyun static inline void *_rtw_vmalloc(u32 sz)
224*4882a593Smuzhiyun {
225*4882a593Smuzhiyun 	void *pbuf;
226*4882a593Smuzhiyun 	pbuf = malloc(sz, M_DEVBUF, M_NOWAIT);
227*4882a593Smuzhiyun 
228*4882a593Smuzhiyun 	return pbuf;
229*4882a593Smuzhiyun }
230*4882a593Smuzhiyun 
_rtw_zvmalloc(u32 sz)231*4882a593Smuzhiyun static inline void *_rtw_zvmalloc(u32 sz)
232*4882a593Smuzhiyun {
233*4882a593Smuzhiyun 	void *pbuf;
234*4882a593Smuzhiyun 
235*4882a593Smuzhiyun 	pbuf = malloc(sz, M_DEVBUF, M_ZERO | M_NOWAIT);
236*4882a593Smuzhiyun 	return pbuf;
237*4882a593Smuzhiyun }
238*4882a593Smuzhiyun 
_rtw_vmfree(void * pbuf,u32 sz)239*4882a593Smuzhiyun static inline void _rtw_vmfree(void *pbuf, u32 sz)
240*4882a593Smuzhiyun {
241*4882a593Smuzhiyun 	free(pbuf, M_DEVBUF);
242*4882a593Smuzhiyun }
243*4882a593Smuzhiyun 
_rtw_malloc(u32 sz)244*4882a593Smuzhiyun static inline void *_rtw_malloc(u32 sz)
245*4882a593Smuzhiyun {
246*4882a593Smuzhiyun 	void *pbuf = NULL;
247*4882a593Smuzhiyun 
248*4882a593Smuzhiyun 	pbuf = malloc(sz, M_DEVBUF, M_NOWAIT);
249*4882a593Smuzhiyun 	return pbuf;
250*4882a593Smuzhiyun }
_rtw_zmalloc(u32 sz)251*4882a593Smuzhiyun static inline void *_rtw_zmalloc(u32 sz)
252*4882a593Smuzhiyun {
253*4882a593Smuzhiyun 	return malloc(sz, M_DEVBUF, M_ZERO | M_NOWAIT);
254*4882a593Smuzhiyun }
_rtw_mfree(void * pbuf,u32 sz)255*4882a593Smuzhiyun static inline void _rtw_mfree(void *pbuf, u32 sz)
256*4882a593Smuzhiyun {
257*4882a593Smuzhiyun 	free(pbuf, M_DEVBUF);
258*4882a593Smuzhiyun }
259*4882a593Smuzhiyun #ifdef CONFIG_USB_HCI
_rtw_usb_buffer_alloc(struct usb_device * dev,size_t size,dma_addr_t * dma)260*4882a593Smuzhiyun static inline void *_rtw_usb_buffer_alloc(struct usb_device *dev, size_t size, dma_addr_t *dma)
261*4882a593Smuzhiyun {
262*4882a593Smuzhiyun 	return malloc(size, M_USBDEV, M_NOWAIT | M_ZERO);
263*4882a593Smuzhiyun }
_rtw_usb_buffer_free(struct usb_device * dev,size_t size,void * addr,dma_addr_t dma)264*4882a593Smuzhiyun static inline void _rtw_usb_buffer_free(struct usb_device *dev, size_t size, void *addr, dma_addr_t dma)
265*4882a593Smuzhiyun {
266*4882a593Smuzhiyun 	free(addr, M_USBDEV);
267*4882a593Smuzhiyun }
268*4882a593Smuzhiyun #endif /* CONFIG_USB_HCI */
269*4882a593Smuzhiyun 
270*4882a593Smuzhiyun struct sk_buff *_rtw_skb_alloc(u32 sz);
271*4882a593Smuzhiyun void _rtw_skb_free(struct sk_buff *skb);
_rtw_skb_copy(const struct sk_buff * skb)272*4882a593Smuzhiyun static inline struct sk_buff *_rtw_skb_copy(const struct sk_buff *skb)
273*4882a593Smuzhiyun {
274*4882a593Smuzhiyun 	return NULL;
275*4882a593Smuzhiyun }
276*4882a593Smuzhiyun 
_rtw_skb_clone(struct sk_buff * skb)277*4882a593Smuzhiyun static inline struct sk_buff *_rtw_skb_clone(struct sk_buff *skb)
278*4882a593Smuzhiyun {
279*4882a593Smuzhiyun 	return skb_clone(skb);
280*4882a593Smuzhiyun }
281*4882a593Smuzhiyun 
_rtw_netif_rx(_nic_hdl ndev,struct sk_buff * skb)282*4882a593Smuzhiyun static inline int _rtw_netif_rx(_nic_hdl ndev, struct sk_buff *skb)
283*4882a593Smuzhiyun {
284*4882a593Smuzhiyun 	return (*ndev->if_input)(ndev, skb);
285*4882a593Smuzhiyun }
286*4882a593Smuzhiyun 
287*4882a593Smuzhiyun #ifdef CONFIG_RTW_NAPI
_rtw_netif_receive_skb(_nic_hdl ndev,struct sk_buff * skb)288*4882a593Smuzhiyun static inline int _rtw_netif_receive_skb(_nic_hdl ndev, struct sk_buff *skb)
289*4882a593Smuzhiyun {
290*4882a593Smuzhiyun 	rtw_warn_on(1);
291*4882a593Smuzhiyun 	return -1;
292*4882a593Smuzhiyun }
293*4882a593Smuzhiyun 
294*4882a593Smuzhiyun #ifdef CONFIG_RTW_GRO
_rtw_napi_gro_receive(struct napi_struct * napi,struct sk_buff * skb)295*4882a593Smuzhiyun static inline gro_result_t _rtw_napi_gro_receive(struct napi_struct *napi, struct sk_buff *skb)
296*4882a593Smuzhiyun {
297*4882a593Smuzhiyun 	rtw_warn_on(1);
298*4882a593Smuzhiyun 	return -1;
299*4882a593Smuzhiyun }
300*4882a593Smuzhiyun #endif /* CONFIG_RTW_GRO */
301*4882a593Smuzhiyun #endif /* CONFIG_RTW_NAPI */
302*4882a593Smuzhiyun 
303*4882a593Smuzhiyun 
304*4882a593Smuzhiyun /*
305*4882a593Smuzhiyun  * Linux timers are emulated using FreeBSD callout functions
306*4882a593Smuzhiyun  * (and taskqueue functionality).
307*4882a593Smuzhiyun  *
308*4882a593Smuzhiyun  * Currently no timer stats functionality.
309*4882a593Smuzhiyun  *
310*4882a593Smuzhiyun  * See (linux_compat) processes.c
311*4882a593Smuzhiyun  *
312*4882a593Smuzhiyun  */
313*4882a593Smuzhiyun struct rtw_timer_list {
314*4882a593Smuzhiyun 	struct callout callout;
315*4882a593Smuzhiyun 	void (*function)(void *);
316*4882a593Smuzhiyun 	void *arg;
317*4882a593Smuzhiyun };
318*4882a593Smuzhiyun 
319*4882a593Smuzhiyun struct workqueue_struct;
320*4882a593Smuzhiyun struct work_struct;
321*4882a593Smuzhiyun typedef void (*work_func_t)(struct work_struct *work);
322*4882a593Smuzhiyun /* Values for the state of an item of work (work_struct) */
323*4882a593Smuzhiyun typedef enum work_state {
324*4882a593Smuzhiyun         WORK_STATE_UNSET = 0,
325*4882a593Smuzhiyun         WORK_STATE_CALLOUT_PENDING = 1,
326*4882a593Smuzhiyun         WORK_STATE_TASK_PENDING = 2,
327*4882a593Smuzhiyun         WORK_STATE_WORK_CANCELLED = 3
328*4882a593Smuzhiyun } work_state_t;
329*4882a593Smuzhiyun 
330*4882a593Smuzhiyun struct work_struct {
331*4882a593Smuzhiyun         struct task task; /* FreeBSD task */
332*4882a593Smuzhiyun         work_state_t state; /* the pending or otherwise state of work. */
333*4882a593Smuzhiyun         work_func_t func;
334*4882a593Smuzhiyun };
335*4882a593Smuzhiyun 
336*4882a593Smuzhiyun //modify private structure to match freebsd
337*4882a593Smuzhiyun #define BITS_PER_LONG 32
338*4882a593Smuzhiyun union ktime {
339*4882a593Smuzhiyun 	s64	tv64;
340*4882a593Smuzhiyun #if BITS_PER_LONG != 64 && !defined(CONFIG_KTIME_SCALAR)
341*4882a593Smuzhiyun 	struct {
342*4882a593Smuzhiyun #ifdef __BIG_ENDIAN
343*4882a593Smuzhiyun 	s32	sec, nsec;
344*4882a593Smuzhiyun #else
345*4882a593Smuzhiyun 	s32	nsec, sec;
346*4882a593Smuzhiyun #endif
347*4882a593Smuzhiyun 	} tv;
348*4882a593Smuzhiyun #endif
349*4882a593Smuzhiyun };
350*4882a593Smuzhiyun #define kmemcheck_bitfield_begin(name)
351*4882a593Smuzhiyun #define kmemcheck_bitfield_end(name)
352*4882a593Smuzhiyun #define CHECKSUM_NONE 0
353*4882a593Smuzhiyun typedef unsigned char *sk_buff_data_t;
354*4882a593Smuzhiyun typedef union ktime ktime_t;		/* Kill this */
355*4882a593Smuzhiyun 
356*4882a593Smuzhiyun void rtw_mtx_lock(_lock *plock);
357*4882a593Smuzhiyun 
358*4882a593Smuzhiyun void rtw_mtx_unlock(_lock *plock);
359*4882a593Smuzhiyun 
360*4882a593Smuzhiyun /**
361*4882a593Smuzhiyun  *	struct sk_buff - socket buffer
362*4882a593Smuzhiyun  *	@next: Next buffer in list
363*4882a593Smuzhiyun  *	@prev: Previous buffer in list
364*4882a593Smuzhiyun  *	@sk: Socket we are owned by
365*4882a593Smuzhiyun  *	@tstamp: Time we arrived
366*4882a593Smuzhiyun  *	@dev: Device we arrived on/are leaving by
367*4882a593Smuzhiyun  *	@transport_header: Transport layer header
368*4882a593Smuzhiyun  *	@network_header: Network layer header
369*4882a593Smuzhiyun  *	@mac_header: Link layer header
370*4882a593Smuzhiyun  *	@_skb_refdst: destination entry (with norefcount bit)
371*4882a593Smuzhiyun  *	@sp: the security path, used for xfrm
372*4882a593Smuzhiyun  *	@cb: Control buffer. Free for use by every layer. Put private vars here
373*4882a593Smuzhiyun  *	@len: Length of actual data
374*4882a593Smuzhiyun  *	@data_len: Data length
375*4882a593Smuzhiyun  *	@mac_len: Length of link layer header
376*4882a593Smuzhiyun  *	@hdr_len: writable header length of cloned skb
377*4882a593Smuzhiyun  *	@csum: Checksum (must include start/offset pair)
378*4882a593Smuzhiyun  *	@csum_start: Offset from skb->head where checksumming should start
379*4882a593Smuzhiyun  *	@csum_offset: Offset from csum_start where checksum should be stored
380*4882a593Smuzhiyun  *	@local_df: allow local fragmentation
381*4882a593Smuzhiyun  *	@cloned: Head may be cloned (check refcnt to be sure)
382*4882a593Smuzhiyun  *	@nohdr: Payload reference only, must not modify header
383*4882a593Smuzhiyun  *	@pkt_type: Packet class
384*4882a593Smuzhiyun  *	@fclone: skbuff clone status
385*4882a593Smuzhiyun  *	@ip_summed: Driver fed us an IP checksum
386*4882a593Smuzhiyun  *	@priority: Packet queueing priority
387*4882a593Smuzhiyun  *	@users: User count - see {datagram,tcp}.c
388*4882a593Smuzhiyun  *	@protocol: Packet protocol from driver
389*4882a593Smuzhiyun  *	@truesize: Buffer size
390*4882a593Smuzhiyun  *	@head: Head of buffer
391*4882a593Smuzhiyun  *	@data: Data head pointer
392*4882a593Smuzhiyun  *	@tail: Tail pointer
393*4882a593Smuzhiyun  *	@end: End pointer
394*4882a593Smuzhiyun  *	@destructor: Destruct function
395*4882a593Smuzhiyun  *	@mark: Generic packet mark
396*4882a593Smuzhiyun  *	@nfct: Associated connection, if any
397*4882a593Smuzhiyun  *	@ipvs_property: skbuff is owned by ipvs
398*4882a593Smuzhiyun  *	@peeked: this packet has been seen already, so stats have been
399*4882a593Smuzhiyun  *		done for it, don't do them again
400*4882a593Smuzhiyun  *	@nf_trace: netfilter packet trace flag
401*4882a593Smuzhiyun  *	@nfctinfo: Relationship of this skb to the connection
402*4882a593Smuzhiyun  *	@nfct_reasm: netfilter conntrack re-assembly pointer
403*4882a593Smuzhiyun  *	@nf_bridge: Saved data about a bridged frame - see br_netfilter.c
404*4882a593Smuzhiyun  *	@skb_iif: ifindex of device we arrived on
405*4882a593Smuzhiyun  *	@rxhash: the packet hash computed on receive
406*4882a593Smuzhiyun  *	@queue_mapping: Queue mapping for multiqueue devices
407*4882a593Smuzhiyun  *	@tc_index: Traffic control index
408*4882a593Smuzhiyun  *	@tc_verd: traffic control verdict
409*4882a593Smuzhiyun  *	@ndisc_nodetype: router type (from link layer)
410*4882a593Smuzhiyun  *	@dma_cookie: a cookie to one of several possible DMA operations
411*4882a593Smuzhiyun  *		done by skb DMA functions
412*4882a593Smuzhiyun  *	@secmark: security marking
413*4882a593Smuzhiyun  *	@vlan_tci: vlan tag control information
414*4882a593Smuzhiyun  */
415*4882a593Smuzhiyun 
416*4882a593Smuzhiyun struct sk_buff {
417*4882a593Smuzhiyun 	/* These two members must be first. */
418*4882a593Smuzhiyun 	struct sk_buff		*next;
419*4882a593Smuzhiyun 	struct sk_buff		*prev;
420*4882a593Smuzhiyun 
421*4882a593Smuzhiyun 	ktime_t			tstamp;
422*4882a593Smuzhiyun 
423*4882a593Smuzhiyun 	struct sock		*sk;
424*4882a593Smuzhiyun 	//struct net_device	*dev;
425*4882a593Smuzhiyun 	struct ifnet *dev;
426*4882a593Smuzhiyun 
427*4882a593Smuzhiyun 	/*
428*4882a593Smuzhiyun 	 * This is the control buffer. It is free to use for every
429*4882a593Smuzhiyun 	 * layer. Please put your private variables there. If you
430*4882a593Smuzhiyun 	 * want to keep them across layers you have to do a skb_clone()
431*4882a593Smuzhiyun 	 * first. This is owned by whoever has the skb queued ATM.
432*4882a593Smuzhiyun 	 */
433*4882a593Smuzhiyun 	char			cb[48] __aligned(8);
434*4882a593Smuzhiyun 
435*4882a593Smuzhiyun 	unsigned long		_skb_refdst;
436*4882a593Smuzhiyun #ifdef CONFIG_XFRM
437*4882a593Smuzhiyun 	struct	sec_path	*sp;
438*4882a593Smuzhiyun #endif
439*4882a593Smuzhiyun 	unsigned int		len,
440*4882a593Smuzhiyun 				data_len;
441*4882a593Smuzhiyun 	u16			mac_len,
442*4882a593Smuzhiyun 				hdr_len;
443*4882a593Smuzhiyun 	union {
444*4882a593Smuzhiyun 		u32		csum;
445*4882a593Smuzhiyun 		struct {
446*4882a593Smuzhiyun 			u16	csum_start;
447*4882a593Smuzhiyun 			u16	csum_offset;
448*4882a593Smuzhiyun 		}smbol2;
449*4882a593Smuzhiyun 	}smbol1;
450*4882a593Smuzhiyun 	u32			priority;
451*4882a593Smuzhiyun 	kmemcheck_bitfield_begin(flags1);
452*4882a593Smuzhiyun 	u8			local_df:1,
453*4882a593Smuzhiyun 				cloned:1,
454*4882a593Smuzhiyun 				ip_summed:2,
455*4882a593Smuzhiyun 				nohdr:1,
456*4882a593Smuzhiyun 				nfctinfo:3;
457*4882a593Smuzhiyun 	u8			pkt_type:3,
458*4882a593Smuzhiyun 				fclone:2,
459*4882a593Smuzhiyun 				ipvs_property:1,
460*4882a593Smuzhiyun 				peeked:1,
461*4882a593Smuzhiyun 				nf_trace:1;
462*4882a593Smuzhiyun 	kmemcheck_bitfield_end(flags1);
463*4882a593Smuzhiyun 	u16			protocol;
464*4882a593Smuzhiyun 
465*4882a593Smuzhiyun 	void			(*destructor)(struct sk_buff *skb);
466*4882a593Smuzhiyun #if defined(CONFIG_NF_CONNTRACK) || defined(CONFIG_NF_CONNTRACK_MODULE)
467*4882a593Smuzhiyun 	struct nf_conntrack	*nfct;
468*4882a593Smuzhiyun 	struct sk_buff		*nfct_reasm;
469*4882a593Smuzhiyun #endif
470*4882a593Smuzhiyun #ifdef CONFIG_BRIDGE_NETFILTER
471*4882a593Smuzhiyun 	struct nf_bridge_info	*nf_bridge;
472*4882a593Smuzhiyun #endif
473*4882a593Smuzhiyun 
474*4882a593Smuzhiyun 	int			skb_iif;
475*4882a593Smuzhiyun #ifdef CONFIG_NET_SCHED
476*4882a593Smuzhiyun 	u16			tc_index;	/* traffic control index */
477*4882a593Smuzhiyun #ifdef CONFIG_NET_CLS_ACT
478*4882a593Smuzhiyun 	u16			tc_verd;	/* traffic control verdict */
479*4882a593Smuzhiyun #endif
480*4882a593Smuzhiyun #endif
481*4882a593Smuzhiyun 
482*4882a593Smuzhiyun 	u32			rxhash;
483*4882a593Smuzhiyun 
484*4882a593Smuzhiyun 	kmemcheck_bitfield_begin(flags2);
485*4882a593Smuzhiyun 	u16			queue_mapping:16;
486*4882a593Smuzhiyun #ifdef CONFIG_IPV6_NDISC_NODETYPE
487*4882a593Smuzhiyun 	u8			ndisc_nodetype:2,
488*4882a593Smuzhiyun 				deliver_no_wcard:1;
489*4882a593Smuzhiyun #else
490*4882a593Smuzhiyun 	u8			deliver_no_wcard:1;
491*4882a593Smuzhiyun #endif
492*4882a593Smuzhiyun 	kmemcheck_bitfield_end(flags2);
493*4882a593Smuzhiyun 
494*4882a593Smuzhiyun 	/* 0/14 bit hole */
495*4882a593Smuzhiyun 
496*4882a593Smuzhiyun #ifdef CONFIG_NET_DMA
497*4882a593Smuzhiyun 	dma_cookie_t		dma_cookie;
498*4882a593Smuzhiyun #endif
499*4882a593Smuzhiyun #ifdef CONFIG_NETWORK_SECMARK
500*4882a593Smuzhiyun 	u32			secmark;
501*4882a593Smuzhiyun #endif
502*4882a593Smuzhiyun 	union {
503*4882a593Smuzhiyun 		u32		mark;
504*4882a593Smuzhiyun 		u32		dropcount;
505*4882a593Smuzhiyun 	}symbol3;
506*4882a593Smuzhiyun 
507*4882a593Smuzhiyun 	u16			vlan_tci;
508*4882a593Smuzhiyun 
509*4882a593Smuzhiyun 	sk_buff_data_t		transport_header;
510*4882a593Smuzhiyun 	sk_buff_data_t		network_header;
511*4882a593Smuzhiyun 	sk_buff_data_t		mac_header;
512*4882a593Smuzhiyun 	/* These elements must be at the end, see alloc_skb() for details.  */
513*4882a593Smuzhiyun 	sk_buff_data_t		tail;
514*4882a593Smuzhiyun 	sk_buff_data_t		end;
515*4882a593Smuzhiyun 	unsigned char		*head,
516*4882a593Smuzhiyun 				*data;
517*4882a593Smuzhiyun 	unsigned int		truesize;
518*4882a593Smuzhiyun 	ATOMIC_T		users;
519*4882a593Smuzhiyun };
520*4882a593Smuzhiyun struct sk_buff_head {
521*4882a593Smuzhiyun 	/* These two members must be first. */
522*4882a593Smuzhiyun 	struct sk_buff	*next;
523*4882a593Smuzhiyun 	struct sk_buff	*prev;
524*4882a593Smuzhiyun 
525*4882a593Smuzhiyun 	u32		qlen;
526*4882a593Smuzhiyun 	_lock	lock;
527*4882a593Smuzhiyun };
528*4882a593Smuzhiyun #define skb_tail_pointer(skb)	skb->tail
skb_put(struct sk_buff * skb,unsigned int len)529*4882a593Smuzhiyun static inline unsigned char *skb_put(struct sk_buff *skb, unsigned int len)
530*4882a593Smuzhiyun {
531*4882a593Smuzhiyun 	unsigned char *tmp = skb_tail_pointer(skb);
532*4882a593Smuzhiyun 	//SKB_LINEAR_ASSERT(skb);
533*4882a593Smuzhiyun 	skb->tail += len;
534*4882a593Smuzhiyun 	skb->len  += len;
535*4882a593Smuzhiyun 	return tmp;
536*4882a593Smuzhiyun }
537*4882a593Smuzhiyun 
__skb_pull(struct sk_buff * skb,unsigned int len)538*4882a593Smuzhiyun static inline unsigned char *__skb_pull(struct sk_buff *skb, unsigned int len)
539*4882a593Smuzhiyun {
540*4882a593Smuzhiyun 	skb->len -= len;
541*4882a593Smuzhiyun 	if(skb->len < skb->data_len)
542*4882a593Smuzhiyun 		printf("%s(),%d,error!\n",__FUNCTION__,__LINE__);
543*4882a593Smuzhiyun 	return skb->data += len;
544*4882a593Smuzhiyun }
skb_pull(struct sk_buff * skb,unsigned int len)545*4882a593Smuzhiyun static inline unsigned char *skb_pull(struct sk_buff *skb, unsigned int len)
546*4882a593Smuzhiyun {
547*4882a593Smuzhiyun 	return __skb_pull(skb, len);
548*4882a593Smuzhiyun }
skb_queue_len(const struct sk_buff_head * list_)549*4882a593Smuzhiyun static inline u32 skb_queue_len(const struct sk_buff_head *list_)
550*4882a593Smuzhiyun {
551*4882a593Smuzhiyun 	return list_->qlen;
552*4882a593Smuzhiyun }
__skb_insert(struct sk_buff * newsk,struct sk_buff * prev,struct sk_buff * next,struct sk_buff_head * list)553*4882a593Smuzhiyun static inline void __skb_insert(struct sk_buff *newsk,
554*4882a593Smuzhiyun 				struct sk_buff *prev, struct sk_buff *next,
555*4882a593Smuzhiyun 				struct sk_buff_head *list)
556*4882a593Smuzhiyun {
557*4882a593Smuzhiyun 	newsk->next = next;
558*4882a593Smuzhiyun 	newsk->prev = prev;
559*4882a593Smuzhiyun 	next->prev  = prev->next = newsk;
560*4882a593Smuzhiyun 	list->qlen++;
561*4882a593Smuzhiyun }
__skb_queue_before(struct sk_buff_head * list,struct sk_buff * next,struct sk_buff * newsk)562*4882a593Smuzhiyun static inline void __skb_queue_before(struct sk_buff_head *list,
563*4882a593Smuzhiyun 				      struct sk_buff *next,
564*4882a593Smuzhiyun 				      struct sk_buff *newsk)
565*4882a593Smuzhiyun {
566*4882a593Smuzhiyun 	__skb_insert(newsk, next->prev, next, list);
567*4882a593Smuzhiyun }
skb_queue_tail(struct sk_buff_head * list,struct sk_buff * newsk)568*4882a593Smuzhiyun static inline void skb_queue_tail(struct sk_buff_head *list,
569*4882a593Smuzhiyun 				   struct sk_buff *newsk)
570*4882a593Smuzhiyun {
571*4882a593Smuzhiyun 	mtx_lock(&list->lock);
572*4882a593Smuzhiyun 	__skb_queue_before(list, (struct sk_buff *)list, newsk);
573*4882a593Smuzhiyun 	mtx_unlock(&list->lock);
574*4882a593Smuzhiyun }
skb_peek(struct sk_buff_head * list_)575*4882a593Smuzhiyun static inline struct sk_buff *skb_peek(struct sk_buff_head *list_)
576*4882a593Smuzhiyun {
577*4882a593Smuzhiyun 	struct sk_buff *list = ((struct sk_buff *)list_)->next;
578*4882a593Smuzhiyun 	if (list == (struct sk_buff *)list_)
579*4882a593Smuzhiyun 		list = NULL;
580*4882a593Smuzhiyun 	return list;
581*4882a593Smuzhiyun }
__skb_unlink(struct sk_buff * skb,struct sk_buff_head * list)582*4882a593Smuzhiyun static inline void __skb_unlink(struct sk_buff *skb, struct sk_buff_head *list)
583*4882a593Smuzhiyun {
584*4882a593Smuzhiyun 	struct sk_buff *next, *prev;
585*4882a593Smuzhiyun 
586*4882a593Smuzhiyun 	list->qlen--;
587*4882a593Smuzhiyun 	next	   = skb->next;
588*4882a593Smuzhiyun 	prev	   = skb->prev;
589*4882a593Smuzhiyun 	skb->next  = skb->prev = NULL;
590*4882a593Smuzhiyun 	next->prev = prev;
591*4882a593Smuzhiyun 	prev->next = next;
592*4882a593Smuzhiyun }
593*4882a593Smuzhiyun 
skb_dequeue(struct sk_buff_head * list)594*4882a593Smuzhiyun static inline struct sk_buff *skb_dequeue(struct sk_buff_head *list)
595*4882a593Smuzhiyun {
596*4882a593Smuzhiyun 	mtx_lock(&list->lock);
597*4882a593Smuzhiyun 
598*4882a593Smuzhiyun 	struct sk_buff *skb = skb_peek(list);
599*4882a593Smuzhiyun 	if (skb)
600*4882a593Smuzhiyun 		__skb_unlink(skb, list);
601*4882a593Smuzhiyun 
602*4882a593Smuzhiyun 	mtx_unlock(&list->lock);
603*4882a593Smuzhiyun 
604*4882a593Smuzhiyun 	return skb;
605*4882a593Smuzhiyun }
skb_reserve(struct sk_buff * skb,int len)606*4882a593Smuzhiyun static inline void skb_reserve(struct sk_buff *skb, int len)
607*4882a593Smuzhiyun {
608*4882a593Smuzhiyun 	skb->data += len;
609*4882a593Smuzhiyun 	skb->tail += len;
610*4882a593Smuzhiyun }
__skb_queue_head_init(struct sk_buff_head * list)611*4882a593Smuzhiyun static inline void __skb_queue_head_init(struct sk_buff_head *list)
612*4882a593Smuzhiyun {
613*4882a593Smuzhiyun 	list->prev = list->next = (struct sk_buff *)list;
614*4882a593Smuzhiyun 	list->qlen = 0;
615*4882a593Smuzhiyun }
616*4882a593Smuzhiyun /*
617*4882a593Smuzhiyun  * This function creates a split out lock class for each invocation;
618*4882a593Smuzhiyun  * this is needed for now since a whole lot of users of the skb-queue
619*4882a593Smuzhiyun  * infrastructure in drivers have different locking usage (in hardirq)
620*4882a593Smuzhiyun  * than the networking core (in softirq only). In the long run either the
621*4882a593Smuzhiyun  * network layer or drivers should need annotation to consolidate the
622*4882a593Smuzhiyun  * main types of usage into 3 classes.
623*4882a593Smuzhiyun  */
skb_queue_head_init(struct sk_buff_head * list)624*4882a593Smuzhiyun static inline void skb_queue_head_init(struct sk_buff_head *list)
625*4882a593Smuzhiyun {
626*4882a593Smuzhiyun 	_rtw_spinlock_init(&list->lock);
627*4882a593Smuzhiyun 	__skb_queue_head_init(list);
628*4882a593Smuzhiyun }
rtw_skb_data(struct sk_buff * pkt)629*4882a593Smuzhiyun static inline u8 *rtw_skb_data(struct sk_buff *pkt)
630*4882a593Smuzhiyun {
631*4882a593Smuzhiyun 	return pkt->data;
632*4882a593Smuzhiyun }
633*4882a593Smuzhiyun 
rtw_skb_len(struct sk_buff * pkt)634*4882a593Smuzhiyun static inline u32 rtw_skb_len(struct sk_buff *pkt)
635*4882a593Smuzhiyun {
636*4882a593Smuzhiyun 	return pkt->len;
637*4882a593Smuzhiyun }
638*4882a593Smuzhiyun 
639*4882a593Smuzhiyun unsigned long copy_from_user(void *to, const void *from, unsigned long n);
640*4882a593Smuzhiyun unsigned long copy_to_user(void *to, const void *from, unsigned long n);
641*4882a593Smuzhiyun struct sk_buff * dev_alloc_skb(unsigned int size);
642*4882a593Smuzhiyun struct sk_buff *skb_clone(const struct sk_buff *skb);
643*4882a593Smuzhiyun void dev_kfree_skb_any(struct sk_buff *skb);
644*4882a593Smuzhiyun #endif //Baron porting from linux, it's all temp solution, needs to check again
645*4882a593Smuzhiyun 
646*4882a593Smuzhiyun 
647*4882a593Smuzhiyun #if 1 // kenny add Linux compatibility code for Linux USB driver
648*4882a593Smuzhiyun #include <dev/usb/usb_compat_linux.h>
649*4882a593Smuzhiyun 
650*4882a593Smuzhiyun #define __init		// __attribute ((constructor))
651*4882a593Smuzhiyun #define __exit		// __attribute ((destructor))
652*4882a593Smuzhiyun 
653*4882a593Smuzhiyun /*
654*4882a593Smuzhiyun  * Definitions for module_init and module_exit macros.
655*4882a593Smuzhiyun  *
656*4882a593Smuzhiyun  * These macros will use the SYSINIT framework to call a specified
657*4882a593Smuzhiyun  * function (with no arguments) on module loading or unloading.
658*4882a593Smuzhiyun  *
659*4882a593Smuzhiyun  */
660*4882a593Smuzhiyun 
661*4882a593Smuzhiyun void module_init_exit_wrapper(void *arg);
662*4882a593Smuzhiyun 
663*4882a593Smuzhiyun #define module_init(initfn)                             \
664*4882a593Smuzhiyun         SYSINIT(mod_init_ ## initfn,                    \
665*4882a593Smuzhiyun                 SI_SUB_KLD, SI_ORDER_FIRST,             \
666*4882a593Smuzhiyun                 module_init_exit_wrapper, initfn)
667*4882a593Smuzhiyun 
668*4882a593Smuzhiyun #define module_exit(exitfn)                             \
669*4882a593Smuzhiyun         SYSUNINIT(mod_exit_ ## exitfn,                  \
670*4882a593Smuzhiyun                   SI_SUB_KLD, SI_ORDER_ANY,             \
671*4882a593Smuzhiyun                   module_init_exit_wrapper, exitfn)
672*4882a593Smuzhiyun 
673*4882a593Smuzhiyun /*
674*4882a593Smuzhiyun  * The usb_register and usb_deregister functions are used to register
675*4882a593Smuzhiyun  * usb drivers with the usb subsystem.
676*4882a593Smuzhiyun  */
677*4882a593Smuzhiyun int usb_register(struct usb_driver *driver);
678*4882a593Smuzhiyun int usb_deregister(struct usb_driver *driver);
679*4882a593Smuzhiyun 
680*4882a593Smuzhiyun /*
681*4882a593Smuzhiyun  * usb_get_dev and usb_put_dev - increment/decrement the reference count
682*4882a593Smuzhiyun  * of the usb device structure.
683*4882a593Smuzhiyun  *
684*4882a593Smuzhiyun  * Original body of usb_get_dev:
685*4882a593Smuzhiyun  *
686*4882a593Smuzhiyun  *       if (dev)
687*4882a593Smuzhiyun  *               get_device(&dev->dev);
688*4882a593Smuzhiyun  *       return dev;
689*4882a593Smuzhiyun  *
690*4882a593Smuzhiyun  * Reference counts are not currently used in this compatibility
691*4882a593Smuzhiyun  * layer. So these functions will do nothing.
692*4882a593Smuzhiyun  */
693*4882a593Smuzhiyun static inline struct usb_device *
usb_get_dev(struct usb_device * dev)694*4882a593Smuzhiyun usb_get_dev(struct usb_device *dev)
695*4882a593Smuzhiyun {
696*4882a593Smuzhiyun         return dev;
697*4882a593Smuzhiyun }
698*4882a593Smuzhiyun 
699*4882a593Smuzhiyun static inline void
usb_put_dev(struct usb_device * dev)700*4882a593Smuzhiyun usb_put_dev(struct usb_device *dev)
701*4882a593Smuzhiyun {
702*4882a593Smuzhiyun         return;
703*4882a593Smuzhiyun }
704*4882a593Smuzhiyun 
705*4882a593Smuzhiyun 
706*4882a593Smuzhiyun // rtw_usb_compat_linux
707*4882a593Smuzhiyun int rtw_usb_submit_urb(struct urb *urb, uint16_t mem_flags);
708*4882a593Smuzhiyun int rtw_usb_unlink_urb(struct urb *urb);
709*4882a593Smuzhiyun int rtw_usb_clear_halt(struct usb_device *dev, struct usb_host_endpoint *uhe);
710*4882a593Smuzhiyun int rtw_usb_control_msg(struct usb_device *dev, struct usb_host_endpoint *uhe,
711*4882a593Smuzhiyun     uint8_t request, uint8_t requesttype,
712*4882a593Smuzhiyun     uint16_t value, uint16_t index, void *data,
713*4882a593Smuzhiyun     uint16_t size, usb_timeout_t timeout);
714*4882a593Smuzhiyun int rtw_usb_set_interface(struct usb_device *dev, uint8_t iface_no, uint8_t alt_index);
715*4882a593Smuzhiyun int rtw_usb_setup_endpoint(struct usb_device *dev,
716*4882a593Smuzhiyun     struct usb_host_endpoint *uhe, usb_size_t bufsize);
717*4882a593Smuzhiyun struct urb *rtw_usb_alloc_urb(uint16_t iso_packets, uint16_t mem_flags);
718*4882a593Smuzhiyun struct usb_host_endpoint *rtw_usb_find_host_endpoint(struct usb_device *dev, uint8_t type, uint8_t ep);
719*4882a593Smuzhiyun struct usb_host_interface *rtw_usb_altnum_to_altsetting(const struct usb_interface *intf, uint8_t alt_index);
720*4882a593Smuzhiyun struct usb_interface *rtw_usb_ifnum_to_if(struct usb_device *dev, uint8_t iface_no);
721*4882a593Smuzhiyun void *rtw_usb_get_intfdata(struct usb_interface *intf);
722*4882a593Smuzhiyun void rtw_usb_linux_register(void *arg);
723*4882a593Smuzhiyun void rtw_usb_linux_deregister(void *arg);
724*4882a593Smuzhiyun void rtw_usb_linux_free_device(struct usb_device *dev);
725*4882a593Smuzhiyun void rtw_usb_free_urb(struct urb *urb);
726*4882a593Smuzhiyun void rtw_usb_init_urb(struct urb *urb);
727*4882a593Smuzhiyun void rtw_usb_kill_urb(struct urb *urb);
728*4882a593Smuzhiyun void rtw_usb_set_intfdata(struct usb_interface *intf, void *data);
729*4882a593Smuzhiyun void rtw_usb_fill_bulk_urb(struct urb *urb, struct usb_device *udev,
730*4882a593Smuzhiyun     struct usb_host_endpoint *uhe, void *buf,
731*4882a593Smuzhiyun     int length, usb_complete_t callback, void *arg);
732*4882a593Smuzhiyun int rtw_usb_bulk_msg(struct usb_device *udev, struct usb_host_endpoint *uhe,
733*4882a593Smuzhiyun     void *data, int len, uint16_t *pactlen, usb_timeout_t timeout);
734*4882a593Smuzhiyun void *usb_get_intfdata(struct usb_interface *intf);
735*4882a593Smuzhiyun int usb_linux_init_endpoints(struct usb_device *udev);
736*4882a593Smuzhiyun 
737*4882a593Smuzhiyun 
738*4882a593Smuzhiyun 
739*4882a593Smuzhiyun typedef struct urb *  PURB;
740*4882a593Smuzhiyun 
741*4882a593Smuzhiyun typedef unsigned gfp_t;
742*4882a593Smuzhiyun #define __GFP_WAIT      ((gfp_t)0x10u)  /* Can wait and reschedule? */
743*4882a593Smuzhiyun #define __GFP_HIGH      ((gfp_t)0x20u)  /* Should access emergency pools? */
744*4882a593Smuzhiyun #define __GFP_IO        ((gfp_t)0x40u)  /* Can start physical IO? */
745*4882a593Smuzhiyun #define __GFP_FS        ((gfp_t)0x80u)  /* Can call down to low-level FS? */
746*4882a593Smuzhiyun #define __GFP_COLD      ((gfp_t)0x100u) /* Cache-cold page required */
747*4882a593Smuzhiyun #define __GFP_NOWARN    ((gfp_t)0x200u) /* Suppress page allocation failure warning */
748*4882a593Smuzhiyun #define __GFP_REPEAT    ((gfp_t)0x400u) /* Retry the allocation.  Might fail */
749*4882a593Smuzhiyun #define __GFP_NOFAIL    ((gfp_t)0x800u) /* Retry for ever.  Cannot fail */
750*4882a593Smuzhiyun #define __GFP_NORETRY   ((gfp_t)0x1000u)/* Do not retry.  Might fail */
751*4882a593Smuzhiyun #define __GFP_NO_GROW   ((gfp_t)0x2000u)/* Slab internal usage */
752*4882a593Smuzhiyun #define __GFP_COMP      ((gfp_t)0x4000u)/* Add compound page metadata */
753*4882a593Smuzhiyun #define __GFP_ZERO      ((gfp_t)0x8000u)/* Return zeroed page on success */
754*4882a593Smuzhiyun #define __GFP_NOMEMALLOC ((gfp_t)0x10000u) /* Don't use emergency reserves */
755*4882a593Smuzhiyun #define __GFP_HARDWALL   ((gfp_t)0x20000u) /* Enforce hardwall cpuset memory allocs */
756*4882a593Smuzhiyun 
757*4882a593Smuzhiyun /* This equals 0, but use constants in case they ever change */
758*4882a593Smuzhiyun #define GFP_NOWAIT      (GFP_ATOMIC & ~__GFP_HIGH)
759*4882a593Smuzhiyun /* GFP_ATOMIC means both !wait (__GFP_WAIT not set) and use emergency pool */
760*4882a593Smuzhiyun #define GFP_ATOMIC      (__GFP_HIGH)
761*4882a593Smuzhiyun #define GFP_NOIO        (__GFP_WAIT)
762*4882a593Smuzhiyun #define GFP_NOFS        (__GFP_WAIT | __GFP_IO)
763*4882a593Smuzhiyun #define GFP_KERNEL      (__GFP_WAIT | __GFP_IO | __GFP_FS)
764*4882a593Smuzhiyun #define GFP_USER        (__GFP_WAIT | __GFP_IO | __GFP_FS | __GFP_HARDWALL)
765*4882a593Smuzhiyun #define GFP_HIGHUSER    (__GFP_WAIT | __GFP_IO | __GFP_FS | __GFP_HARDWALL | \
766*4882a593Smuzhiyun                          __GFP_HIGHMEM)
767*4882a593Smuzhiyun 
768*4882a593Smuzhiyun 
769*4882a593Smuzhiyun #endif // kenny add Linux compatibility code for Linux USB
770*4882a593Smuzhiyun 
get_next(_list * list)771*4882a593Smuzhiyun __inline static _list *get_next(_list	*list)
772*4882a593Smuzhiyun {
773*4882a593Smuzhiyun 	return list->next;
774*4882a593Smuzhiyun }
775*4882a593Smuzhiyun 
get_list_head(_queue * queue)776*4882a593Smuzhiyun __inline static _list	*get_list_head(_queue	*queue)
777*4882a593Smuzhiyun {
778*4882a593Smuzhiyun 	return (&(queue->queue));
779*4882a593Smuzhiyun }
780*4882a593Smuzhiyun 
781*4882a593Smuzhiyun 
782*4882a593Smuzhiyun #define LIST_CONTAINOR(ptr, type, member) \
783*4882a593Smuzhiyun         ((type *)((char *)(ptr)-(SIZE_T)(&((type *)0)->member)))
784*4882a593Smuzhiyun 
785*4882a593Smuzhiyun 
__list_del(struct list_head * prev,struct list_head * next)786*4882a593Smuzhiyun static inline void __list_del(struct list_head * prev, struct list_head * next)
787*4882a593Smuzhiyun {
788*4882a593Smuzhiyun 	next->prev = prev;
789*4882a593Smuzhiyun 	prev->next = next;
790*4882a593Smuzhiyun }
INIT_LIST_HEAD(struct list_head * list)791*4882a593Smuzhiyun static inline void INIT_LIST_HEAD(struct list_head *list)
792*4882a593Smuzhiyun {
793*4882a593Smuzhiyun 	list->next = list;
794*4882a593Smuzhiyun 	list->prev = list;
795*4882a593Smuzhiyun }
rtw_list_delete(_list * plist)796*4882a593Smuzhiyun __inline static void rtw_list_delete(_list *plist)
797*4882a593Smuzhiyun {
798*4882a593Smuzhiyun 	__list_del(plist->prev, plist->next);
799*4882a593Smuzhiyun 	INIT_LIST_HEAD(plist);
800*4882a593Smuzhiyun }
801*4882a593Smuzhiyun 
timer_hdl(void * ctx)802*4882a593Smuzhiyun static inline void timer_hdl(void *ctx)
803*4882a593Smuzhiyun {
804*4882a593Smuzhiyun 	_timer *timer = (_timer *)ctx;
805*4882a593Smuzhiyun 
806*4882a593Smuzhiyun 	rtw_mtx_lock(NULL);
807*4882a593Smuzhiyun 	if (callout_pending(&timer->callout)) {
808*4882a593Smuzhiyun 		/* callout was reset */
809*4882a593Smuzhiyun 		rtw_mtx_unlock(NULL);
810*4882a593Smuzhiyun 		return;
811*4882a593Smuzhiyun 	}
812*4882a593Smuzhiyun 
813*4882a593Smuzhiyun 	if (!callout_active(&timer->callout)) {
814*4882a593Smuzhiyun 		/* callout was stopped */
815*4882a593Smuzhiyun 		rtw_mtx_unlock(NULL);
816*4882a593Smuzhiyun 		return;
817*4882a593Smuzhiyun 	}
818*4882a593Smuzhiyun 
819*4882a593Smuzhiyun 	callout_deactivate(&timer->callout);
820*4882a593Smuzhiyun 
821*4882a593Smuzhiyun 	timer->function(timer->arg);
822*4882a593Smuzhiyun 
823*4882a593Smuzhiyun 	rtw_mtx_unlock(NULL);
824*4882a593Smuzhiyun }
825*4882a593Smuzhiyun 
_init_timer(_timer * ptimer,void * pfunc,void * cntx)826*4882a593Smuzhiyun static inline void _init_timer(_timer *ptimer, void *pfunc, void *cntx)
827*4882a593Smuzhiyun {
828*4882a593Smuzhiyun 	ptimer->function = pfunc;
829*4882a593Smuzhiyun 	ptimer->arg = cntx;
830*4882a593Smuzhiyun 	callout_init(&ptimer->callout, CALLOUT_MPSAFE);
831*4882a593Smuzhiyun }
832*4882a593Smuzhiyun 
_set_timer(_timer * ptimer,u32 delay_time)833*4882a593Smuzhiyun __inline static void _set_timer(_timer *ptimer,u32 delay_time)
834*4882a593Smuzhiyun {
835*4882a593Smuzhiyun 	if (ptimer->function && ptimer->arg) {
836*4882a593Smuzhiyun 		rtw_mtx_lock(NULL);
837*4882a593Smuzhiyun 		callout_reset(&ptimer->callout, delay_time, timer_hdl, ptimer);
838*4882a593Smuzhiyun 		rtw_mtx_unlock(NULL);
839*4882a593Smuzhiyun 	}
840*4882a593Smuzhiyun }
841*4882a593Smuzhiyun 
_cancel_timer(_timer * ptimer,u8 * bcancelled)842*4882a593Smuzhiyun __inline static void _cancel_timer(_timer *ptimer,u8 *bcancelled)
843*4882a593Smuzhiyun {
844*4882a593Smuzhiyun 	rtw_mtx_lock(NULL);
845*4882a593Smuzhiyun 	callout_drain(&ptimer->callout);
846*4882a593Smuzhiyun 	rtw_mtx_unlock(NULL);
847*4882a593Smuzhiyun 	*bcancelled = 1; /* assume an pending timer to be canceled */
848*4882a593Smuzhiyun }
849*4882a593Smuzhiyun 
_init_workitem(_workitem * pwork,void * pfunc,PVOID cntx)850*4882a593Smuzhiyun __inline static void _init_workitem(_workitem *pwork, void *pfunc, PVOID cntx)
851*4882a593Smuzhiyun {
852*4882a593Smuzhiyun 	printf("%s Not implement yet! \n",__FUNCTION__);
853*4882a593Smuzhiyun }
854*4882a593Smuzhiyun 
_set_workitem(_workitem * pwork)855*4882a593Smuzhiyun __inline static void _set_workitem(_workitem *pwork)
856*4882a593Smuzhiyun {
857*4882a593Smuzhiyun 	printf("%s Not implement yet! \n",__FUNCTION__);
858*4882a593Smuzhiyun //	schedule_work(pwork);
859*4882a593Smuzhiyun }
860*4882a593Smuzhiyun 
861*4882a593Smuzhiyun //
862*4882a593Smuzhiyun // Global Mutex: can only be used at PASSIVE level.
863*4882a593Smuzhiyun //
864*4882a593Smuzhiyun 
865*4882a593Smuzhiyun #define ACQUIRE_GLOBAL_MUTEX(_MutexCounter)                              \
866*4882a593Smuzhiyun {                                                               \
867*4882a593Smuzhiyun }
868*4882a593Smuzhiyun 
869*4882a593Smuzhiyun #define RELEASE_GLOBAL_MUTEX(_MutexCounter)                              \
870*4882a593Smuzhiyun {                                                               \
871*4882a593Smuzhiyun }
872*4882a593Smuzhiyun 
873*4882a593Smuzhiyun /* Atomic integer operations */
874*4882a593Smuzhiyun #define ATOMIC_T atomic_t
875*4882a593Smuzhiyun 
876*4882a593Smuzhiyun #define ATOMIC_INIT(i)  { (i) }
ATOMIC_SET(ATOMIC_T * v,int i)877*4882a593Smuzhiyun static inline void ATOMIC_SET(ATOMIC_T *v, int i)
878*4882a593Smuzhiyun {
879*4882a593Smuzhiyun 	atomic_set_int(v, i);
880*4882a593Smuzhiyun }
881*4882a593Smuzhiyun 
ATOMIC_READ(ATOMIC_T * v)882*4882a593Smuzhiyun static inline int ATOMIC_READ(ATOMIC_T *v)
883*4882a593Smuzhiyun {
884*4882a593Smuzhiyun 	return atomic_load_acq_32(v);
885*4882a593Smuzhiyun }
886*4882a593Smuzhiyun 
ATOMIC_ADD(ATOMIC_T * v,int i)887*4882a593Smuzhiyun static inline void ATOMIC_ADD(ATOMIC_T *v, int i)
888*4882a593Smuzhiyun {
889*4882a593Smuzhiyun 	atomic_add_int(v, i);
890*4882a593Smuzhiyun }
ATOMIC_SUB(ATOMIC_T * v,int i)891*4882a593Smuzhiyun static inline void ATOMIC_SUB(ATOMIC_T *v, int i)
892*4882a593Smuzhiyun {
893*4882a593Smuzhiyun 	atomic_subtract_int(v, i);
894*4882a593Smuzhiyun }
895*4882a593Smuzhiyun 
ATOMIC_INC(ATOMIC_T * v)896*4882a593Smuzhiyun static inline void ATOMIC_INC(ATOMIC_T *v)
897*4882a593Smuzhiyun {
898*4882a593Smuzhiyun 	atomic_add_int(v, 1);
899*4882a593Smuzhiyun }
900*4882a593Smuzhiyun 
ATOMIC_DEC(ATOMIC_T * v)901*4882a593Smuzhiyun static inline void ATOMIC_DEC(ATOMIC_T *v)
902*4882a593Smuzhiyun {
903*4882a593Smuzhiyun 	atomic_subtract_int(v, 1);
904*4882a593Smuzhiyun }
905*4882a593Smuzhiyun 
ATOMIC_ADD_RETURN(ATOMIC_T * v,int i)906*4882a593Smuzhiyun static inline int ATOMIC_ADD_RETURN(ATOMIC_T *v, int i)
907*4882a593Smuzhiyun {
908*4882a593Smuzhiyun 	atomic_add_int(v, i);
909*4882a593Smuzhiyun 	return atomic_load_acq_32(v);
910*4882a593Smuzhiyun }
911*4882a593Smuzhiyun 
ATOMIC_SUB_RETURN(ATOMIC_T * v,int i)912*4882a593Smuzhiyun static inline int ATOMIC_SUB_RETURN(ATOMIC_T *v, int i)
913*4882a593Smuzhiyun {
914*4882a593Smuzhiyun 	atomic_subtract_int(v, i);
915*4882a593Smuzhiyun 	return atomic_load_acq_32(v);
916*4882a593Smuzhiyun }
917*4882a593Smuzhiyun 
ATOMIC_INC_RETURN(ATOMIC_T * v)918*4882a593Smuzhiyun static inline int ATOMIC_INC_RETURN(ATOMIC_T *v)
919*4882a593Smuzhiyun {
920*4882a593Smuzhiyun 	atomic_add_int(v, 1);
921*4882a593Smuzhiyun 	return atomic_load_acq_32(v);
922*4882a593Smuzhiyun }
923*4882a593Smuzhiyun 
ATOMIC_DEC_RETURN(ATOMIC_T * v)924*4882a593Smuzhiyun static inline int ATOMIC_DEC_RETURN(ATOMIC_T *v)
925*4882a593Smuzhiyun {
926*4882a593Smuzhiyun 	atomic_subtract_int(v, 1);
927*4882a593Smuzhiyun 	return atomic_load_acq_32(v);
928*4882a593Smuzhiyun }
929*4882a593Smuzhiyun 
ATOMIC_INC_UNLESS(ATOMIC_T * v,int u)930*4882a593Smuzhiyun static inline bool ATOMIC_INC_UNLESS(ATOMIC_T *v, int u)
931*4882a593Smuzhiyun {
932*4882a593Smuzhiyun 	#error "TBD\n"
933*4882a593Smuzhiyun }
934*4882a593Smuzhiyun 
935*4882a593Smuzhiyun /*task*/
936*4882a593Smuzhiyun typedef void (*task_fn_t)(void *context, int pending);
937*4882a593Smuzhiyun #if 0 /*taskqueue -- asynchronous task execution*/
938*4882a593Smuzhiyun 
939*4882a593Smuzhiyun TASK_INIT(struct task *task, int priority, task_fn_t func,
940*4882a593Smuzhiyun 	 void *context);
941*4882a593Smuzhiyun 
942*4882a593Smuzhiyun TASK_INITIALIZER(int priority, task_fn_t func, void *context);
943*4882a593Smuzhiyun 
944*4882a593Smuzhiyun TASKQUEUE_DECLARE(name);
945*4882a593Smuzhiyun 
946*4882a593Smuzhiyun TASKQUEUE_DEFINE(name, taskqueue_enqueue_fn enqueue, void *context,
947*4882a593Smuzhiyun 	 init);
948*4882a593Smuzhiyun 
949*4882a593Smuzhiyun #endif
rtw_tasklet_init(_tasklet * t,task_fn_t func,unsigned long data)950*4882a593Smuzhiyun static inline void rtw_tasklet_init(_tasklet *t,task_fn_t func,
951*4882a593Smuzhiyun 							unsigned long data)
952*4882a593Smuzhiyun {
953*4882a593Smuzhiyun 	TASK_INIT(t, 0, func, padapter);
954*4882a593Smuzhiyun }
955*4882a593Smuzhiyun static inline void rtw_tasklet_kill(_tasklet *t)
956*4882a593Smuzhiyun 
957*4882a593Smuzhiyun 
958*4882a593Smuzhiyun }
959*4882a593Smuzhiyun 
960*4882a593Smuzhiyun static inline void rtw_tasklet_schedule(_tasklet *t)
961*4882a593Smuzhiyun {
962*4882a593Smuzhiyun 
963*4882a593Smuzhiyun }
964*4882a593Smuzhiyun static inline void rtw_tasklet_hi_schedule(_tasklet *t)
965*4882a593Smuzhiyun {
966*4882a593Smuzhiyun 
967*4882a593Smuzhiyun }
968*4882a593Smuzhiyun 
969*4882a593Smuzhiyun 
970*4882a593Smuzhiyun /*thread*/
971*4882a593Smuzhiyun static inline void rtw_thread_enter(char *name)
972*4882a593Smuzhiyun {
973*4882a593Smuzhiyun 	printf("%s", "RTKTHREAD_enter");
974*4882a593Smuzhiyun }
975*4882a593Smuzhiyun 
976*4882a593Smuzhiyun static inline void rtw_thread_exit(_completion *comp)
977*4882a593Smuzhiyun {
978*4882a593Smuzhiyun 	printf("%s", "RTKTHREAD_exit");
979*4882a593Smuzhiyun }
980*4882a593Smuzhiyun 
981*4882a593Smuzhiyun #include <sys/unistd.h>		/* for RFHIGHPID */
982*4882a593Smuzhiyun static inline _thread_hdl_ rtw_thread_start(int (*threadfn)(void *data),
983*4882a593Smuzhiyun 			void *data, const char namefmt[])
984*4882a593Smuzhiyun {
985*4882a593Smuzhiyun 	_thread_hdl_ _rtw_thread = NULL;
986*4882a593Smuzhiyun 	struct proc *p;
987*4882a593Smuzhiyun 	struct thread *td;
988*4882a593Smuzhiyun 
989*4882a593Smuzhiyun 	_rtw_thread = kproc_kthread_add(mp_xmit_packet_thread, data,
990*4882a593Smuzhiyun 			&p, &td, RFHIGHPID, 0, namefmt, namefmt);
991*4882a593Smuzhiyun 
992*4882a593Smuzhiyun 	if (_rtw_thread < 0)
993*4882a593Smuzhiyun 		_rtw_thread = NULL;
994*4882a593Smuzhiyun 	return _rtw_thread;
995*4882a593Smuzhiyun }
996*4882a593Smuzhiyun 
997*4882a593Smuzhiyun static inline bool rtw_thread_stop(_thread_hdl_ th)
998*4882a593Smuzhiyun {
999*4882a593Smuzhiyun 	return _FALSE;
1000*4882a593Smuzhiyun }
1001*4882a593Smuzhiyun static inline void rtw_thread_wait_stop(void)
1002*4882a593Smuzhiyun {
1003*4882a593Smuzhiyun 
1004*4882a593Smuzhiyun }
1005*4882a593Smuzhiyun __inline static void flush_signals_thread(void)
1006*4882a593Smuzhiyun {
1007*4882a593Smuzhiyun 
1008*4882a593Smuzhiyun }
1009*4882a593Smuzhiyun 
1010*4882a593Smuzhiyun #define rtw_dump_stack(void) do {} while (0)
1011*4882a593Smuzhiyun #define rtw_bug_on(condition) do {} while (0)
1012*4882a593Smuzhiyun #define rtw_warn_on(condition) do {} while (0)
1013*4882a593Smuzhiyun #define rtw_sprintf(buf, size, format, arg...) do {} while (0)
1014*4882a593Smuzhiyun 
1015*4882a593Smuzhiyun #define rtw_netdev_priv(netdev) (((struct ifnet *)netdev)->if_softc)
1016*4882a593Smuzhiyun #define rtw_free_netdev(netdev) if_free((netdev))
1017*4882a593Smuzhiyun 
1018*4882a593Smuzhiyun #define RTW_DIV_ROUND_UP(n, d)	(((n) + (d - 1)) / d)
1019*4882a593Smuzhiyun 
1020*4882a593Smuzhiyun #define NDEV_FMT "%s"
1021*4882a593Smuzhiyun #define NDEV_ARG(ndev) ""
1022*4882a593Smuzhiyun #define ADPT_FMT "%s"
1023*4882a593Smuzhiyun #define ADPT_ARG(adapter) ""
1024*4882a593Smuzhiyun #define FUNC_NDEV_FMT "%s"
1025*4882a593Smuzhiyun #define FUNC_NDEV_ARG(ndev) __func__
1026*4882a593Smuzhiyun #define FUNC_ADPT_FMT "%s"
1027*4882a593Smuzhiyun #define FUNC_ADPT_ARG(adapter) __func__
1028*4882a593Smuzhiyun 
1029*4882a593Smuzhiyun #define STRUCT_PACKED
1030*4882a593Smuzhiyun 
1031*4882a593Smuzhiyun #endif
1032*4882a593Smuzhiyun 
1033