1*4882a593Smuzhiyun // SPDX-License-Identifier: GPL-2.0-or-later
2*4882a593Smuzhiyun /******************************************************************************
3*4882a593Smuzhiyun *
4*4882a593Smuzhiyun * (C)Copyright 1998,1999 SysKonnect,
5*4882a593Smuzhiyun * a business unit of Schneider & Koch & Co. Datensysteme GmbH.
6*4882a593Smuzhiyun *
7*4882a593Smuzhiyun * See the file "skfddi.c" for further information.
8*4882a593Smuzhiyun *
9*4882a593Smuzhiyun * The information in this file is provided "AS IS" without warranty.
10*4882a593Smuzhiyun *
11*4882a593Smuzhiyun ******************************************************************************/
12*4882a593Smuzhiyun
13*4882a593Smuzhiyun #define HWMTM
14*4882a593Smuzhiyun
15*4882a593Smuzhiyun #ifndef FDDI
16*4882a593Smuzhiyun #define FDDI
17*4882a593Smuzhiyun #endif
18*4882a593Smuzhiyun
19*4882a593Smuzhiyun #include "h/types.h"
20*4882a593Smuzhiyun #include "h/fddi.h"
21*4882a593Smuzhiyun #include "h/smc.h"
22*4882a593Smuzhiyun #include "h/supern_2.h"
23*4882a593Smuzhiyun #include "h/skfbiinc.h"
24*4882a593Smuzhiyun
25*4882a593Smuzhiyun /*
26*4882a593Smuzhiyun -------------------------------------------------------------
27*4882a593Smuzhiyun DOCUMENTATION
28*4882a593Smuzhiyun -------------------------------------------------------------
29*4882a593Smuzhiyun BEGIN_MANUAL_ENTRY(DOCUMENTATION)
30*4882a593Smuzhiyun
31*4882a593Smuzhiyun T B D
32*4882a593Smuzhiyun
33*4882a593Smuzhiyun END_MANUAL_ENTRY
34*4882a593Smuzhiyun */
35*4882a593Smuzhiyun /*
36*4882a593Smuzhiyun -------------------------------------------------------------
37*4882a593Smuzhiyun LOCAL VARIABLES:
38*4882a593Smuzhiyun -------------------------------------------------------------
39*4882a593Smuzhiyun */
40*4882a593Smuzhiyun #ifdef COMMON_MB_POOL
41*4882a593Smuzhiyun static SMbuf *mb_start = 0 ;
42*4882a593Smuzhiyun static SMbuf *mb_free = 0 ;
43*4882a593Smuzhiyun static int mb_init = FALSE ;
44*4882a593Smuzhiyun static int call_count = 0 ;
45*4882a593Smuzhiyun #endif
46*4882a593Smuzhiyun
47*4882a593Smuzhiyun /*
48*4882a593Smuzhiyun -------------------------------------------------------------
49*4882a593Smuzhiyun EXTERNE VARIABLES:
50*4882a593Smuzhiyun -------------------------------------------------------------
51*4882a593Smuzhiyun */
52*4882a593Smuzhiyun
53*4882a593Smuzhiyun #ifdef DEBUG
54*4882a593Smuzhiyun #ifndef DEBUG_BRD
55*4882a593Smuzhiyun extern struct smt_debug debug ;
56*4882a593Smuzhiyun #endif
57*4882a593Smuzhiyun #endif
58*4882a593Smuzhiyun
59*4882a593Smuzhiyun #ifdef NDIS_OS2
60*4882a593Smuzhiyun extern u_char offDepth ;
61*4882a593Smuzhiyun extern u_char force_irq_pending ;
62*4882a593Smuzhiyun #endif
63*4882a593Smuzhiyun
64*4882a593Smuzhiyun /*
65*4882a593Smuzhiyun -------------------------------------------------------------
66*4882a593Smuzhiyun LOCAL FUNCTIONS:
67*4882a593Smuzhiyun -------------------------------------------------------------
68*4882a593Smuzhiyun */
69*4882a593Smuzhiyun
70*4882a593Smuzhiyun static void queue_llc_rx(struct s_smc *smc, SMbuf *mb);
71*4882a593Smuzhiyun static void smt_to_llc(struct s_smc *smc, SMbuf *mb);
72*4882a593Smuzhiyun static void init_txd_ring(struct s_smc *smc);
73*4882a593Smuzhiyun static void init_rxd_ring(struct s_smc *smc);
74*4882a593Smuzhiyun static void queue_txd_mb(struct s_smc *smc, SMbuf *mb);
75*4882a593Smuzhiyun static u_long init_descr_ring(struct s_smc *smc, union s_fp_descr volatile *start,
76*4882a593Smuzhiyun int count);
77*4882a593Smuzhiyun static u_long repair_txd_ring(struct s_smc *smc, struct s_smt_tx_queue *queue);
78*4882a593Smuzhiyun static u_long repair_rxd_ring(struct s_smc *smc, struct s_smt_rx_queue *queue);
79*4882a593Smuzhiyun static SMbuf* get_llc_rx(struct s_smc *smc);
80*4882a593Smuzhiyun static SMbuf* get_txd_mb(struct s_smc *smc);
81*4882a593Smuzhiyun static void mac_drv_clear_txd(struct s_smc *smc);
82*4882a593Smuzhiyun
83*4882a593Smuzhiyun /*
84*4882a593Smuzhiyun -------------------------------------------------------------
85*4882a593Smuzhiyun EXTERNAL FUNCTIONS:
86*4882a593Smuzhiyun -------------------------------------------------------------
87*4882a593Smuzhiyun */
88*4882a593Smuzhiyun /* The external SMT functions are listed in cmtdef.h */
89*4882a593Smuzhiyun
90*4882a593Smuzhiyun extern void* mac_drv_get_space(struct s_smc *smc, unsigned int size);
91*4882a593Smuzhiyun extern void* mac_drv_get_desc_mem(struct s_smc *smc, unsigned int size);
92*4882a593Smuzhiyun extern void mac_drv_fill_rxd(struct s_smc *smc);
93*4882a593Smuzhiyun extern void mac_drv_tx_complete(struct s_smc *smc,
94*4882a593Smuzhiyun volatile struct s_smt_fp_txd *txd);
95*4882a593Smuzhiyun extern void mac_drv_rx_complete(struct s_smc *smc,
96*4882a593Smuzhiyun volatile struct s_smt_fp_rxd *rxd,
97*4882a593Smuzhiyun int frag_count, int len);
98*4882a593Smuzhiyun extern void mac_drv_requeue_rxd(struct s_smc *smc,
99*4882a593Smuzhiyun volatile struct s_smt_fp_rxd *rxd,
100*4882a593Smuzhiyun int frag_count);
101*4882a593Smuzhiyun extern void mac_drv_clear_rxd(struct s_smc *smc,
102*4882a593Smuzhiyun volatile struct s_smt_fp_rxd *rxd, int frag_count);
103*4882a593Smuzhiyun
104*4882a593Smuzhiyun #ifdef USE_OS_CPY
105*4882a593Smuzhiyun extern void hwm_cpy_rxd2mb(void);
106*4882a593Smuzhiyun extern void hwm_cpy_txd2mb(void);
107*4882a593Smuzhiyun #endif
108*4882a593Smuzhiyun
109*4882a593Smuzhiyun #ifdef ALL_RX_COMPLETE
110*4882a593Smuzhiyun extern void mac_drv_all_receives_complete(void);
111*4882a593Smuzhiyun #endif
112*4882a593Smuzhiyun
113*4882a593Smuzhiyun extern u_long mac_drv_virt2phys(struct s_smc *smc, void *virt);
114*4882a593Smuzhiyun extern u_long dma_master(struct s_smc *smc, void *virt, int len, int flag);
115*4882a593Smuzhiyun
116*4882a593Smuzhiyun #ifdef NDIS_OS2
117*4882a593Smuzhiyun extern void post_proc(void);
118*4882a593Smuzhiyun #else
119*4882a593Smuzhiyun extern void dma_complete(struct s_smc *smc, volatile union s_fp_descr *descr,
120*4882a593Smuzhiyun int flag);
121*4882a593Smuzhiyun #endif
122*4882a593Smuzhiyun
123*4882a593Smuzhiyun extern int mac_drv_rx_init(struct s_smc *smc, int len, int fc, char *look_ahead,
124*4882a593Smuzhiyun int la_len);
125*4882a593Smuzhiyun
126*4882a593Smuzhiyun /*
127*4882a593Smuzhiyun -------------------------------------------------------------
128*4882a593Smuzhiyun PUBLIC FUNCTIONS:
129*4882a593Smuzhiyun -------------------------------------------------------------
130*4882a593Smuzhiyun */
131*4882a593Smuzhiyun void process_receive(struct s_smc *smc);
132*4882a593Smuzhiyun void fddi_isr(struct s_smc *smc);
133*4882a593Smuzhiyun void smt_free_mbuf(struct s_smc *smc, SMbuf *mb);
134*4882a593Smuzhiyun void init_driver_fplus(struct s_smc *smc);
135*4882a593Smuzhiyun void mac_drv_rx_mode(struct s_smc *smc, int mode);
136*4882a593Smuzhiyun void init_fddi_driver(struct s_smc *smc, u_char *mac_addr);
137*4882a593Smuzhiyun void mac_drv_clear_tx_queue(struct s_smc *smc);
138*4882a593Smuzhiyun void mac_drv_clear_rx_queue(struct s_smc *smc);
139*4882a593Smuzhiyun void hwm_tx_frag(struct s_smc *smc, char far *virt, u_long phys, int len,
140*4882a593Smuzhiyun int frame_status);
141*4882a593Smuzhiyun void hwm_rx_frag(struct s_smc *smc, char far *virt, u_long phys, int len,
142*4882a593Smuzhiyun int frame_status);
143*4882a593Smuzhiyun
144*4882a593Smuzhiyun int mac_drv_init(struct s_smc *smc);
145*4882a593Smuzhiyun int hwm_tx_init(struct s_smc *smc, u_char fc, int frag_count, int frame_len,
146*4882a593Smuzhiyun int frame_status);
147*4882a593Smuzhiyun
148*4882a593Smuzhiyun u_int mac_drv_check_space(void);
149*4882a593Smuzhiyun
150*4882a593Smuzhiyun SMbuf* smt_get_mbuf(struct s_smc *smc);
151*4882a593Smuzhiyun
152*4882a593Smuzhiyun #ifdef DEBUG
153*4882a593Smuzhiyun void mac_drv_debug_lev(struct s_smc *smc, int flag, int lev);
154*4882a593Smuzhiyun #endif
155*4882a593Smuzhiyun
156*4882a593Smuzhiyun /*
157*4882a593Smuzhiyun -------------------------------------------------------------
158*4882a593Smuzhiyun MACROS:
159*4882a593Smuzhiyun -------------------------------------------------------------
160*4882a593Smuzhiyun */
161*4882a593Smuzhiyun #ifndef UNUSED
162*4882a593Smuzhiyun #ifdef lint
163*4882a593Smuzhiyun #define UNUSED(x) (x) = (x)
164*4882a593Smuzhiyun #else
165*4882a593Smuzhiyun #define UNUSED(x)
166*4882a593Smuzhiyun #endif
167*4882a593Smuzhiyun #endif
168*4882a593Smuzhiyun
169*4882a593Smuzhiyun #ifdef USE_CAN_ADDR
170*4882a593Smuzhiyun #define MA smc->hw.fddi_canon_addr.a
171*4882a593Smuzhiyun #define GROUP_ADDR_BIT 0x01
172*4882a593Smuzhiyun #else
173*4882a593Smuzhiyun #define MA smc->hw.fddi_home_addr.a
174*4882a593Smuzhiyun #define GROUP_ADDR_BIT 0x80
175*4882a593Smuzhiyun #endif
176*4882a593Smuzhiyun
177*4882a593Smuzhiyun #define RXD_TXD_COUNT (HWM_ASYNC_TXD_COUNT+HWM_SYNC_TXD_COUNT+\
178*4882a593Smuzhiyun SMT_R1_RXD_COUNT+SMT_R2_RXD_COUNT)
179*4882a593Smuzhiyun
180*4882a593Smuzhiyun #ifdef MB_OUTSIDE_SMC
181*4882a593Smuzhiyun #define EXT_VIRT_MEM ((RXD_TXD_COUNT+1)*sizeof(struct s_smt_fp_txd) +\
182*4882a593Smuzhiyun MAX_MBUF*sizeof(SMbuf))
183*4882a593Smuzhiyun #define EXT_VIRT_MEM_2 ((RXD_TXD_COUNT+1)*sizeof(struct s_smt_fp_txd))
184*4882a593Smuzhiyun #else
185*4882a593Smuzhiyun #define EXT_VIRT_MEM ((RXD_TXD_COUNT+1)*sizeof(struct s_smt_fp_txd))
186*4882a593Smuzhiyun #endif
187*4882a593Smuzhiyun
188*4882a593Smuzhiyun /*
189*4882a593Smuzhiyun * define critical read for 16 Bit drivers
190*4882a593Smuzhiyun */
191*4882a593Smuzhiyun #if defined(NDIS_OS2) || defined(ODI2)
192*4882a593Smuzhiyun #define CR_READ(var) ((var) & 0xffff0000 | ((var) & 0xffff))
193*4882a593Smuzhiyun #else
194*4882a593Smuzhiyun #define CR_READ(var) (__le32)(var)
195*4882a593Smuzhiyun #endif
196*4882a593Smuzhiyun
197*4882a593Smuzhiyun #define IMASK_SLOW (IS_PLINT1 | IS_PLINT2 | IS_TIMINT | IS_TOKEN | \
198*4882a593Smuzhiyun IS_MINTR1 | IS_MINTR2 | IS_MINTR3 | IS_R1_P | \
199*4882a593Smuzhiyun IS_R1_C | IS_XA_C | IS_XS_C)
200*4882a593Smuzhiyun
201*4882a593Smuzhiyun /*
202*4882a593Smuzhiyun -------------------------------------------------------------
203*4882a593Smuzhiyun INIT- AND SMT FUNCTIONS:
204*4882a593Smuzhiyun -------------------------------------------------------------
205*4882a593Smuzhiyun */
206*4882a593Smuzhiyun
207*4882a593Smuzhiyun
208*4882a593Smuzhiyun /*
209*4882a593Smuzhiyun * BEGIN_MANUAL_ENTRY(mac_drv_check_space)
210*4882a593Smuzhiyun * u_int mac_drv_check_space()
211*4882a593Smuzhiyun *
212*4882a593Smuzhiyun * function DOWNCALL (drvsr.c)
213*4882a593Smuzhiyun * This function calculates the needed non virtual
214*4882a593Smuzhiyun * memory for MBufs, RxD and TxD descriptors etc.
215*4882a593Smuzhiyun * needed by the driver.
216*4882a593Smuzhiyun *
217*4882a593Smuzhiyun * return u_int memory in bytes
218*4882a593Smuzhiyun *
219*4882a593Smuzhiyun * END_MANUAL_ENTRY
220*4882a593Smuzhiyun */
mac_drv_check_space(void)221*4882a593Smuzhiyun u_int mac_drv_check_space(void)
222*4882a593Smuzhiyun {
223*4882a593Smuzhiyun #ifdef MB_OUTSIDE_SMC
224*4882a593Smuzhiyun #ifdef COMMON_MB_POOL
225*4882a593Smuzhiyun call_count++ ;
226*4882a593Smuzhiyun if (call_count == 1) {
227*4882a593Smuzhiyun return EXT_VIRT_MEM;
228*4882a593Smuzhiyun }
229*4882a593Smuzhiyun else {
230*4882a593Smuzhiyun return EXT_VIRT_MEM_2;
231*4882a593Smuzhiyun }
232*4882a593Smuzhiyun #else
233*4882a593Smuzhiyun return EXT_VIRT_MEM;
234*4882a593Smuzhiyun #endif
235*4882a593Smuzhiyun #else
236*4882a593Smuzhiyun return 0;
237*4882a593Smuzhiyun #endif
238*4882a593Smuzhiyun }
239*4882a593Smuzhiyun
240*4882a593Smuzhiyun /*
241*4882a593Smuzhiyun * BEGIN_MANUAL_ENTRY(mac_drv_init)
242*4882a593Smuzhiyun * void mac_drv_init(smc)
243*4882a593Smuzhiyun *
244*4882a593Smuzhiyun * function DOWNCALL (drvsr.c)
245*4882a593Smuzhiyun * In this function the hardware module allocates it's
246*4882a593Smuzhiyun * memory.
247*4882a593Smuzhiyun * The operating system dependent module should call
248*4882a593Smuzhiyun * mac_drv_init once, after the adatper is detected.
249*4882a593Smuzhiyun * END_MANUAL_ENTRY
250*4882a593Smuzhiyun */
mac_drv_init(struct s_smc * smc)251*4882a593Smuzhiyun int mac_drv_init(struct s_smc *smc)
252*4882a593Smuzhiyun {
253*4882a593Smuzhiyun if (sizeof(struct s_smt_fp_rxd) % 16) {
254*4882a593Smuzhiyun SMT_PANIC(smc,HWM_E0001,HWM_E0001_MSG) ;
255*4882a593Smuzhiyun }
256*4882a593Smuzhiyun if (sizeof(struct s_smt_fp_txd) % 16) {
257*4882a593Smuzhiyun SMT_PANIC(smc,HWM_E0002,HWM_E0002_MSG) ;
258*4882a593Smuzhiyun }
259*4882a593Smuzhiyun
260*4882a593Smuzhiyun /*
261*4882a593Smuzhiyun * get the required memory for the RxDs and TxDs
262*4882a593Smuzhiyun */
263*4882a593Smuzhiyun if (!(smc->os.hwm.descr_p = (union s_fp_descr volatile *)
264*4882a593Smuzhiyun mac_drv_get_desc_mem(smc,(u_int)
265*4882a593Smuzhiyun (RXD_TXD_COUNT+1)*sizeof(struct s_smt_fp_txd)))) {
266*4882a593Smuzhiyun return 1; /* no space the hwm modul can't work */
267*4882a593Smuzhiyun }
268*4882a593Smuzhiyun
269*4882a593Smuzhiyun /*
270*4882a593Smuzhiyun * get the memory for the SMT MBufs
271*4882a593Smuzhiyun */
272*4882a593Smuzhiyun #ifndef MB_OUTSIDE_SMC
273*4882a593Smuzhiyun smc->os.hwm.mbuf_pool.mb_start=(SMbuf *)(&smc->os.hwm.mbuf_pool.mb[0]) ;
274*4882a593Smuzhiyun #else
275*4882a593Smuzhiyun #ifndef COMMON_MB_POOL
276*4882a593Smuzhiyun if (!(smc->os.hwm.mbuf_pool.mb_start = (SMbuf *) mac_drv_get_space(smc,
277*4882a593Smuzhiyun MAX_MBUF*sizeof(SMbuf)))) {
278*4882a593Smuzhiyun return 1; /* no space the hwm modul can't work */
279*4882a593Smuzhiyun }
280*4882a593Smuzhiyun #else
281*4882a593Smuzhiyun if (!mb_start) {
282*4882a593Smuzhiyun if (!(mb_start = (SMbuf *) mac_drv_get_space(smc,
283*4882a593Smuzhiyun MAX_MBUF*sizeof(SMbuf)))) {
284*4882a593Smuzhiyun return 1; /* no space the hwm modul can't work */
285*4882a593Smuzhiyun }
286*4882a593Smuzhiyun }
287*4882a593Smuzhiyun #endif
288*4882a593Smuzhiyun #endif
289*4882a593Smuzhiyun return 0;
290*4882a593Smuzhiyun }
291*4882a593Smuzhiyun
292*4882a593Smuzhiyun /*
293*4882a593Smuzhiyun * BEGIN_MANUAL_ENTRY(init_driver_fplus)
294*4882a593Smuzhiyun * init_driver_fplus(smc)
295*4882a593Smuzhiyun *
296*4882a593Smuzhiyun * Sets hardware modul specific values for the mode register 2
297*4882a593Smuzhiyun * (e.g. the byte alignment for the received frames, the position of the
298*4882a593Smuzhiyun * least significant byte etc.)
299*4882a593Smuzhiyun * END_MANUAL_ENTRY
300*4882a593Smuzhiyun */
init_driver_fplus(struct s_smc * smc)301*4882a593Smuzhiyun void init_driver_fplus(struct s_smc *smc)
302*4882a593Smuzhiyun {
303*4882a593Smuzhiyun smc->hw.fp.mdr2init = FM_LSB | FM_BMMODE | FM_ENNPRQ | FM_ENHSRQ | 3 ;
304*4882a593Smuzhiyun
305*4882a593Smuzhiyun #ifdef PCI
306*4882a593Smuzhiyun smc->hw.fp.mdr2init |= FM_CHKPAR | FM_PARITY ;
307*4882a593Smuzhiyun #endif
308*4882a593Smuzhiyun smc->hw.fp.mdr3init = FM_MENRQAUNLCK | FM_MENRS ;
309*4882a593Smuzhiyun
310*4882a593Smuzhiyun #ifdef USE_CAN_ADDR
311*4882a593Smuzhiyun /* enable address bit swapping */
312*4882a593Smuzhiyun smc->hw.fp.frselreg_init = FM_ENXMTADSWAP | FM_ENRCVADSWAP ;
313*4882a593Smuzhiyun #endif
314*4882a593Smuzhiyun }
315*4882a593Smuzhiyun
init_descr_ring(struct s_smc * smc,union s_fp_descr volatile * start,int count)316*4882a593Smuzhiyun static u_long init_descr_ring(struct s_smc *smc,
317*4882a593Smuzhiyun union s_fp_descr volatile *start,
318*4882a593Smuzhiyun int count)
319*4882a593Smuzhiyun {
320*4882a593Smuzhiyun int i ;
321*4882a593Smuzhiyun union s_fp_descr volatile *d1 ;
322*4882a593Smuzhiyun union s_fp_descr volatile *d2 ;
323*4882a593Smuzhiyun u_long phys ;
324*4882a593Smuzhiyun
325*4882a593Smuzhiyun DB_GEN(3, "descr ring starts at = %p", start);
326*4882a593Smuzhiyun for (i=count-1, d1=start; i ; i--) {
327*4882a593Smuzhiyun d2 = d1 ;
328*4882a593Smuzhiyun d1++ ; /* descr is owned by the host */
329*4882a593Smuzhiyun d2->r.rxd_rbctrl = cpu_to_le32(BMU_CHECK) ;
330*4882a593Smuzhiyun d2->r.rxd_next = &d1->r ;
331*4882a593Smuzhiyun phys = mac_drv_virt2phys(smc,(void *)d1) ;
332*4882a593Smuzhiyun d2->r.rxd_nrdadr = cpu_to_le32(phys) ;
333*4882a593Smuzhiyun }
334*4882a593Smuzhiyun DB_GEN(3, "descr ring ends at = %p", d1);
335*4882a593Smuzhiyun d1->r.rxd_rbctrl = cpu_to_le32(BMU_CHECK) ;
336*4882a593Smuzhiyun d1->r.rxd_next = &start->r ;
337*4882a593Smuzhiyun phys = mac_drv_virt2phys(smc,(void *)start) ;
338*4882a593Smuzhiyun d1->r.rxd_nrdadr = cpu_to_le32(phys) ;
339*4882a593Smuzhiyun
340*4882a593Smuzhiyun for (i=count, d1=start; i ; i--) {
341*4882a593Smuzhiyun DRV_BUF_FLUSH(&d1->r,DDI_DMA_SYNC_FORDEV) ;
342*4882a593Smuzhiyun d1++;
343*4882a593Smuzhiyun }
344*4882a593Smuzhiyun return phys;
345*4882a593Smuzhiyun }
346*4882a593Smuzhiyun
init_txd_ring(struct s_smc * smc)347*4882a593Smuzhiyun static void init_txd_ring(struct s_smc *smc)
348*4882a593Smuzhiyun {
349*4882a593Smuzhiyun struct s_smt_fp_txd volatile *ds ;
350*4882a593Smuzhiyun struct s_smt_tx_queue *queue ;
351*4882a593Smuzhiyun u_long phys ;
352*4882a593Smuzhiyun
353*4882a593Smuzhiyun /*
354*4882a593Smuzhiyun * initialize the transmit descriptors
355*4882a593Smuzhiyun */
356*4882a593Smuzhiyun ds = (struct s_smt_fp_txd volatile *) ((char *)smc->os.hwm.descr_p +
357*4882a593Smuzhiyun SMT_R1_RXD_COUNT*sizeof(struct s_smt_fp_rxd)) ;
358*4882a593Smuzhiyun queue = smc->hw.fp.tx[QUEUE_A0] ;
359*4882a593Smuzhiyun DB_GEN(3, "Init async TxD ring, %d TxDs", HWM_ASYNC_TXD_COUNT);
360*4882a593Smuzhiyun (void)init_descr_ring(smc,(union s_fp_descr volatile *)ds,
361*4882a593Smuzhiyun HWM_ASYNC_TXD_COUNT) ;
362*4882a593Smuzhiyun phys = le32_to_cpu(ds->txd_ntdadr) ;
363*4882a593Smuzhiyun ds++ ;
364*4882a593Smuzhiyun queue->tx_curr_put = queue->tx_curr_get = ds ;
365*4882a593Smuzhiyun ds-- ;
366*4882a593Smuzhiyun queue->tx_free = HWM_ASYNC_TXD_COUNT ;
367*4882a593Smuzhiyun queue->tx_used = 0 ;
368*4882a593Smuzhiyun outpd(ADDR(B5_XA_DA),phys) ;
369*4882a593Smuzhiyun
370*4882a593Smuzhiyun ds = (struct s_smt_fp_txd volatile *) ((char *)ds +
371*4882a593Smuzhiyun HWM_ASYNC_TXD_COUNT*sizeof(struct s_smt_fp_txd)) ;
372*4882a593Smuzhiyun queue = smc->hw.fp.tx[QUEUE_S] ;
373*4882a593Smuzhiyun DB_GEN(3, "Init sync TxD ring, %d TxDs", HWM_SYNC_TXD_COUNT);
374*4882a593Smuzhiyun (void)init_descr_ring(smc,(union s_fp_descr volatile *)ds,
375*4882a593Smuzhiyun HWM_SYNC_TXD_COUNT) ;
376*4882a593Smuzhiyun phys = le32_to_cpu(ds->txd_ntdadr) ;
377*4882a593Smuzhiyun ds++ ;
378*4882a593Smuzhiyun queue->tx_curr_put = queue->tx_curr_get = ds ;
379*4882a593Smuzhiyun queue->tx_free = HWM_SYNC_TXD_COUNT ;
380*4882a593Smuzhiyun queue->tx_used = 0 ;
381*4882a593Smuzhiyun outpd(ADDR(B5_XS_DA),phys) ;
382*4882a593Smuzhiyun }
383*4882a593Smuzhiyun
init_rxd_ring(struct s_smc * smc)384*4882a593Smuzhiyun static void init_rxd_ring(struct s_smc *smc)
385*4882a593Smuzhiyun {
386*4882a593Smuzhiyun struct s_smt_fp_rxd volatile *ds ;
387*4882a593Smuzhiyun struct s_smt_rx_queue *queue ;
388*4882a593Smuzhiyun u_long phys ;
389*4882a593Smuzhiyun
390*4882a593Smuzhiyun /*
391*4882a593Smuzhiyun * initialize the receive descriptors
392*4882a593Smuzhiyun */
393*4882a593Smuzhiyun ds = (struct s_smt_fp_rxd volatile *) smc->os.hwm.descr_p ;
394*4882a593Smuzhiyun queue = smc->hw.fp.rx[QUEUE_R1] ;
395*4882a593Smuzhiyun DB_GEN(3, "Init RxD ring, %d RxDs", SMT_R1_RXD_COUNT);
396*4882a593Smuzhiyun (void)init_descr_ring(smc,(union s_fp_descr volatile *)ds,
397*4882a593Smuzhiyun SMT_R1_RXD_COUNT) ;
398*4882a593Smuzhiyun phys = le32_to_cpu(ds->rxd_nrdadr) ;
399*4882a593Smuzhiyun ds++ ;
400*4882a593Smuzhiyun queue->rx_curr_put = queue->rx_curr_get = ds ;
401*4882a593Smuzhiyun queue->rx_free = SMT_R1_RXD_COUNT ;
402*4882a593Smuzhiyun queue->rx_used = 0 ;
403*4882a593Smuzhiyun outpd(ADDR(B4_R1_DA),phys) ;
404*4882a593Smuzhiyun }
405*4882a593Smuzhiyun
406*4882a593Smuzhiyun /*
407*4882a593Smuzhiyun * BEGIN_MANUAL_ENTRY(init_fddi_driver)
408*4882a593Smuzhiyun * void init_fddi_driver(smc,mac_addr)
409*4882a593Smuzhiyun *
410*4882a593Smuzhiyun * initializes the driver and it's variables
411*4882a593Smuzhiyun *
412*4882a593Smuzhiyun * END_MANUAL_ENTRY
413*4882a593Smuzhiyun */
init_fddi_driver(struct s_smc * smc,u_char * mac_addr)414*4882a593Smuzhiyun void init_fddi_driver(struct s_smc *smc, u_char *mac_addr)
415*4882a593Smuzhiyun {
416*4882a593Smuzhiyun SMbuf *mb ;
417*4882a593Smuzhiyun int i ;
418*4882a593Smuzhiyun
419*4882a593Smuzhiyun init_board(smc,mac_addr) ;
420*4882a593Smuzhiyun (void)init_fplus(smc) ;
421*4882a593Smuzhiyun
422*4882a593Smuzhiyun /*
423*4882a593Smuzhiyun * initialize the SMbufs for the SMT
424*4882a593Smuzhiyun */
425*4882a593Smuzhiyun #ifndef COMMON_MB_POOL
426*4882a593Smuzhiyun mb = smc->os.hwm.mbuf_pool.mb_start ;
427*4882a593Smuzhiyun smc->os.hwm.mbuf_pool.mb_free = (SMbuf *)NULL ;
428*4882a593Smuzhiyun for (i = 0; i < MAX_MBUF; i++) {
429*4882a593Smuzhiyun mb->sm_use_count = 1 ;
430*4882a593Smuzhiyun smt_free_mbuf(smc,mb) ;
431*4882a593Smuzhiyun mb++ ;
432*4882a593Smuzhiyun }
433*4882a593Smuzhiyun #else
434*4882a593Smuzhiyun mb = mb_start ;
435*4882a593Smuzhiyun if (!mb_init) {
436*4882a593Smuzhiyun mb_free = 0 ;
437*4882a593Smuzhiyun for (i = 0; i < MAX_MBUF; i++) {
438*4882a593Smuzhiyun mb->sm_use_count = 1 ;
439*4882a593Smuzhiyun smt_free_mbuf(smc,mb) ;
440*4882a593Smuzhiyun mb++ ;
441*4882a593Smuzhiyun }
442*4882a593Smuzhiyun mb_init = TRUE ;
443*4882a593Smuzhiyun }
444*4882a593Smuzhiyun #endif
445*4882a593Smuzhiyun
446*4882a593Smuzhiyun /*
447*4882a593Smuzhiyun * initialize the other variables
448*4882a593Smuzhiyun */
449*4882a593Smuzhiyun smc->os.hwm.llc_rx_pipe = smc->os.hwm.llc_rx_tail = (SMbuf *)NULL ;
450*4882a593Smuzhiyun smc->os.hwm.txd_tx_pipe = smc->os.hwm.txd_tx_tail = NULL ;
451*4882a593Smuzhiyun smc->os.hwm.pass_SMT = smc->os.hwm.pass_NSA = smc->os.hwm.pass_DB = 0 ;
452*4882a593Smuzhiyun smc->os.hwm.pass_llc_promisc = TRUE ;
453*4882a593Smuzhiyun smc->os.hwm.queued_rx_frames = smc->os.hwm.queued_txd_mb = 0 ;
454*4882a593Smuzhiyun smc->os.hwm.detec_count = 0 ;
455*4882a593Smuzhiyun smc->os.hwm.rx_break = 0 ;
456*4882a593Smuzhiyun smc->os.hwm.rx_len_error = 0 ;
457*4882a593Smuzhiyun smc->os.hwm.isr_flag = FALSE ;
458*4882a593Smuzhiyun
459*4882a593Smuzhiyun /*
460*4882a593Smuzhiyun * make sure that the start pointer is 16 byte aligned
461*4882a593Smuzhiyun */
462*4882a593Smuzhiyun i = 16 - ((long)smc->os.hwm.descr_p & 0xf) ;
463*4882a593Smuzhiyun if (i != 16) {
464*4882a593Smuzhiyun DB_GEN(3, "i = %d", i);
465*4882a593Smuzhiyun smc->os.hwm.descr_p = (union s_fp_descr volatile *)
466*4882a593Smuzhiyun ((char *)smc->os.hwm.descr_p+i) ;
467*4882a593Smuzhiyun }
468*4882a593Smuzhiyun DB_GEN(3, "pt to descr area = %p", smc->os.hwm.descr_p);
469*4882a593Smuzhiyun
470*4882a593Smuzhiyun init_txd_ring(smc) ;
471*4882a593Smuzhiyun init_rxd_ring(smc) ;
472*4882a593Smuzhiyun mac_drv_fill_rxd(smc) ;
473*4882a593Smuzhiyun
474*4882a593Smuzhiyun init_plc(smc) ;
475*4882a593Smuzhiyun }
476*4882a593Smuzhiyun
477*4882a593Smuzhiyun
smt_get_mbuf(struct s_smc * smc)478*4882a593Smuzhiyun SMbuf *smt_get_mbuf(struct s_smc *smc)
479*4882a593Smuzhiyun {
480*4882a593Smuzhiyun register SMbuf *mb ;
481*4882a593Smuzhiyun
482*4882a593Smuzhiyun #ifndef COMMON_MB_POOL
483*4882a593Smuzhiyun mb = smc->os.hwm.mbuf_pool.mb_free ;
484*4882a593Smuzhiyun #else
485*4882a593Smuzhiyun mb = mb_free ;
486*4882a593Smuzhiyun #endif
487*4882a593Smuzhiyun if (mb) {
488*4882a593Smuzhiyun #ifndef COMMON_MB_POOL
489*4882a593Smuzhiyun smc->os.hwm.mbuf_pool.mb_free = mb->sm_next ;
490*4882a593Smuzhiyun #else
491*4882a593Smuzhiyun mb_free = mb->sm_next ;
492*4882a593Smuzhiyun #endif
493*4882a593Smuzhiyun mb->sm_off = 8 ;
494*4882a593Smuzhiyun mb->sm_use_count = 1 ;
495*4882a593Smuzhiyun }
496*4882a593Smuzhiyun DB_GEN(3, "get SMbuf: mb = %p", mb);
497*4882a593Smuzhiyun return mb; /* May be NULL */
498*4882a593Smuzhiyun }
499*4882a593Smuzhiyun
smt_free_mbuf(struct s_smc * smc,SMbuf * mb)500*4882a593Smuzhiyun void smt_free_mbuf(struct s_smc *smc, SMbuf *mb)
501*4882a593Smuzhiyun {
502*4882a593Smuzhiyun
503*4882a593Smuzhiyun if (mb) {
504*4882a593Smuzhiyun mb->sm_use_count-- ;
505*4882a593Smuzhiyun DB_GEN(3, "free_mbuf: sm_use_count = %d", mb->sm_use_count);
506*4882a593Smuzhiyun /*
507*4882a593Smuzhiyun * If the use_count is != zero the MBuf is queued
508*4882a593Smuzhiyun * more than once and must not queued into the
509*4882a593Smuzhiyun * free MBuf queue
510*4882a593Smuzhiyun */
511*4882a593Smuzhiyun if (!mb->sm_use_count) {
512*4882a593Smuzhiyun DB_GEN(3, "free SMbuf: mb = %p", mb);
513*4882a593Smuzhiyun #ifndef COMMON_MB_POOL
514*4882a593Smuzhiyun mb->sm_next = smc->os.hwm.mbuf_pool.mb_free ;
515*4882a593Smuzhiyun smc->os.hwm.mbuf_pool.mb_free = mb ;
516*4882a593Smuzhiyun #else
517*4882a593Smuzhiyun mb->sm_next = mb_free ;
518*4882a593Smuzhiyun mb_free = mb ;
519*4882a593Smuzhiyun #endif
520*4882a593Smuzhiyun }
521*4882a593Smuzhiyun }
522*4882a593Smuzhiyun else
523*4882a593Smuzhiyun SMT_PANIC(smc,HWM_E0003,HWM_E0003_MSG) ;
524*4882a593Smuzhiyun }
525*4882a593Smuzhiyun
526*4882a593Smuzhiyun
527*4882a593Smuzhiyun /*
528*4882a593Smuzhiyun * BEGIN_MANUAL_ENTRY(mac_drv_repair_descr)
529*4882a593Smuzhiyun * void mac_drv_repair_descr(smc)
530*4882a593Smuzhiyun *
531*4882a593Smuzhiyun * function called from SMT (HWM / hwmtm.c)
532*4882a593Smuzhiyun * The BMU is idle when this function is called.
533*4882a593Smuzhiyun * Mac_drv_repair_descr sets up the physical address
534*4882a593Smuzhiyun * for all receive and transmit queues where the BMU
535*4882a593Smuzhiyun * should continue.
536*4882a593Smuzhiyun * It may be that the BMU was reseted during a fragmented
537*4882a593Smuzhiyun * transfer. In this case there are some fragments which will
538*4882a593Smuzhiyun * never completed by the BMU. The OWN bit of this fragments
539*4882a593Smuzhiyun * must be switched to be owned by the host.
540*4882a593Smuzhiyun *
541*4882a593Smuzhiyun * Give a start command to the receive BMU.
542*4882a593Smuzhiyun * Start the transmit BMUs if transmit frames pending.
543*4882a593Smuzhiyun *
544*4882a593Smuzhiyun * END_MANUAL_ENTRY
545*4882a593Smuzhiyun */
mac_drv_repair_descr(struct s_smc * smc)546*4882a593Smuzhiyun void mac_drv_repair_descr(struct s_smc *smc)
547*4882a593Smuzhiyun {
548*4882a593Smuzhiyun u_long phys ;
549*4882a593Smuzhiyun
550*4882a593Smuzhiyun if (smc->hw.hw_state != STOPPED) {
551*4882a593Smuzhiyun SK_BREAK() ;
552*4882a593Smuzhiyun SMT_PANIC(smc,HWM_E0013,HWM_E0013_MSG) ;
553*4882a593Smuzhiyun return ;
554*4882a593Smuzhiyun }
555*4882a593Smuzhiyun
556*4882a593Smuzhiyun /*
557*4882a593Smuzhiyun * repair tx queues: don't start
558*4882a593Smuzhiyun */
559*4882a593Smuzhiyun phys = repair_txd_ring(smc,smc->hw.fp.tx[QUEUE_A0]) ;
560*4882a593Smuzhiyun outpd(ADDR(B5_XA_DA),phys) ;
561*4882a593Smuzhiyun if (smc->hw.fp.tx_q[QUEUE_A0].tx_used) {
562*4882a593Smuzhiyun outpd(ADDR(B0_XA_CSR),CSR_START) ;
563*4882a593Smuzhiyun }
564*4882a593Smuzhiyun phys = repair_txd_ring(smc,smc->hw.fp.tx[QUEUE_S]) ;
565*4882a593Smuzhiyun outpd(ADDR(B5_XS_DA),phys) ;
566*4882a593Smuzhiyun if (smc->hw.fp.tx_q[QUEUE_S].tx_used) {
567*4882a593Smuzhiyun outpd(ADDR(B0_XS_CSR),CSR_START) ;
568*4882a593Smuzhiyun }
569*4882a593Smuzhiyun
570*4882a593Smuzhiyun /*
571*4882a593Smuzhiyun * repair rx queues
572*4882a593Smuzhiyun */
573*4882a593Smuzhiyun phys = repair_rxd_ring(smc,smc->hw.fp.rx[QUEUE_R1]) ;
574*4882a593Smuzhiyun outpd(ADDR(B4_R1_DA),phys) ;
575*4882a593Smuzhiyun outpd(ADDR(B0_R1_CSR),CSR_START) ;
576*4882a593Smuzhiyun }
577*4882a593Smuzhiyun
repair_txd_ring(struct s_smc * smc,struct s_smt_tx_queue * queue)578*4882a593Smuzhiyun static u_long repair_txd_ring(struct s_smc *smc, struct s_smt_tx_queue *queue)
579*4882a593Smuzhiyun {
580*4882a593Smuzhiyun int i ;
581*4882a593Smuzhiyun int tx_used ;
582*4882a593Smuzhiyun u_long phys ;
583*4882a593Smuzhiyun u_long tbctrl ;
584*4882a593Smuzhiyun struct s_smt_fp_txd volatile *t ;
585*4882a593Smuzhiyun
586*4882a593Smuzhiyun SK_UNUSED(smc) ;
587*4882a593Smuzhiyun
588*4882a593Smuzhiyun t = queue->tx_curr_get ;
589*4882a593Smuzhiyun tx_used = queue->tx_used ;
590*4882a593Smuzhiyun for (i = tx_used+queue->tx_free-1 ; i ; i-- ) {
591*4882a593Smuzhiyun t = t->txd_next ;
592*4882a593Smuzhiyun }
593*4882a593Smuzhiyun phys = le32_to_cpu(t->txd_ntdadr) ;
594*4882a593Smuzhiyun
595*4882a593Smuzhiyun t = queue->tx_curr_get ;
596*4882a593Smuzhiyun while (tx_used) {
597*4882a593Smuzhiyun DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORCPU) ;
598*4882a593Smuzhiyun tbctrl = le32_to_cpu(t->txd_tbctrl) ;
599*4882a593Smuzhiyun
600*4882a593Smuzhiyun if (tbctrl & BMU_OWN) {
601*4882a593Smuzhiyun if (tbctrl & BMU_STF) {
602*4882a593Smuzhiyun break ; /* exit the loop */
603*4882a593Smuzhiyun }
604*4882a593Smuzhiyun else {
605*4882a593Smuzhiyun /*
606*4882a593Smuzhiyun * repair the descriptor
607*4882a593Smuzhiyun */
608*4882a593Smuzhiyun t->txd_tbctrl &= ~cpu_to_le32(BMU_OWN) ;
609*4882a593Smuzhiyun }
610*4882a593Smuzhiyun }
611*4882a593Smuzhiyun phys = le32_to_cpu(t->txd_ntdadr) ;
612*4882a593Smuzhiyun DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORDEV) ;
613*4882a593Smuzhiyun t = t->txd_next ;
614*4882a593Smuzhiyun tx_used-- ;
615*4882a593Smuzhiyun }
616*4882a593Smuzhiyun return phys;
617*4882a593Smuzhiyun }
618*4882a593Smuzhiyun
619*4882a593Smuzhiyun /*
620*4882a593Smuzhiyun * Repairs the receive descriptor ring and returns the physical address
621*4882a593Smuzhiyun * where the BMU should continue working.
622*4882a593Smuzhiyun *
623*4882a593Smuzhiyun * o The physical address where the BMU was stopped has to be
624*4882a593Smuzhiyun * determined. This is the next RxD after rx_curr_get with an OWN
625*4882a593Smuzhiyun * bit set.
626*4882a593Smuzhiyun * o The BMU should start working at beginning of the next frame.
627*4882a593Smuzhiyun * RxDs with an OWN bit set but with a reset STF bit should be
628*4882a593Smuzhiyun * skipped and owned by the driver (OWN = 0).
629*4882a593Smuzhiyun */
repair_rxd_ring(struct s_smc * smc,struct s_smt_rx_queue * queue)630*4882a593Smuzhiyun static u_long repair_rxd_ring(struct s_smc *smc, struct s_smt_rx_queue *queue)
631*4882a593Smuzhiyun {
632*4882a593Smuzhiyun int i ;
633*4882a593Smuzhiyun int rx_used ;
634*4882a593Smuzhiyun u_long phys ;
635*4882a593Smuzhiyun u_long rbctrl ;
636*4882a593Smuzhiyun struct s_smt_fp_rxd volatile *r ;
637*4882a593Smuzhiyun
638*4882a593Smuzhiyun SK_UNUSED(smc) ;
639*4882a593Smuzhiyun
640*4882a593Smuzhiyun r = queue->rx_curr_get ;
641*4882a593Smuzhiyun rx_used = queue->rx_used ;
642*4882a593Smuzhiyun for (i = SMT_R1_RXD_COUNT-1 ; i ; i-- ) {
643*4882a593Smuzhiyun r = r->rxd_next ;
644*4882a593Smuzhiyun }
645*4882a593Smuzhiyun phys = le32_to_cpu(r->rxd_nrdadr) ;
646*4882a593Smuzhiyun
647*4882a593Smuzhiyun r = queue->rx_curr_get ;
648*4882a593Smuzhiyun while (rx_used) {
649*4882a593Smuzhiyun DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORCPU) ;
650*4882a593Smuzhiyun rbctrl = le32_to_cpu(r->rxd_rbctrl) ;
651*4882a593Smuzhiyun
652*4882a593Smuzhiyun if (rbctrl & BMU_OWN) {
653*4882a593Smuzhiyun if (rbctrl & BMU_STF) {
654*4882a593Smuzhiyun break ; /* exit the loop */
655*4882a593Smuzhiyun }
656*4882a593Smuzhiyun else {
657*4882a593Smuzhiyun /*
658*4882a593Smuzhiyun * repair the descriptor
659*4882a593Smuzhiyun */
660*4882a593Smuzhiyun r->rxd_rbctrl &= ~cpu_to_le32(BMU_OWN) ;
661*4882a593Smuzhiyun }
662*4882a593Smuzhiyun }
663*4882a593Smuzhiyun phys = le32_to_cpu(r->rxd_nrdadr) ;
664*4882a593Smuzhiyun DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORDEV) ;
665*4882a593Smuzhiyun r = r->rxd_next ;
666*4882a593Smuzhiyun rx_used-- ;
667*4882a593Smuzhiyun }
668*4882a593Smuzhiyun return phys;
669*4882a593Smuzhiyun }
670*4882a593Smuzhiyun
671*4882a593Smuzhiyun
672*4882a593Smuzhiyun /*
673*4882a593Smuzhiyun -------------------------------------------------------------
674*4882a593Smuzhiyun INTERRUPT SERVICE ROUTINE:
675*4882a593Smuzhiyun -------------------------------------------------------------
676*4882a593Smuzhiyun */
677*4882a593Smuzhiyun
678*4882a593Smuzhiyun /*
679*4882a593Smuzhiyun * BEGIN_MANUAL_ENTRY(fddi_isr)
680*4882a593Smuzhiyun * void fddi_isr(smc)
681*4882a593Smuzhiyun *
682*4882a593Smuzhiyun * function DOWNCALL (drvsr.c)
683*4882a593Smuzhiyun * interrupt service routine, handles the interrupt requests
684*4882a593Smuzhiyun * generated by the FDDI adapter.
685*4882a593Smuzhiyun *
686*4882a593Smuzhiyun * NOTE: The operating system dependent module must guarantee that the
687*4882a593Smuzhiyun * interrupts of the adapter are disabled when it calls fddi_isr.
688*4882a593Smuzhiyun *
689*4882a593Smuzhiyun * About the USE_BREAK_ISR mechanismn:
690*4882a593Smuzhiyun *
691*4882a593Smuzhiyun * The main requirement of this mechanismn is to force an timer IRQ when
692*4882a593Smuzhiyun * leaving process_receive() with leave_isr set. process_receive() may
693*4882a593Smuzhiyun * be called at any time from anywhere!
694*4882a593Smuzhiyun * To be sure we don't miss such event we set 'force_irq' per default.
695*4882a593Smuzhiyun * We have to force and Timer IRQ if 'smc->os.hwm.leave_isr' AND
696*4882a593Smuzhiyun * 'force_irq' are set. 'force_irq' may be reset if a receive complete
697*4882a593Smuzhiyun * IRQ is pending.
698*4882a593Smuzhiyun *
699*4882a593Smuzhiyun * END_MANUAL_ENTRY
700*4882a593Smuzhiyun */
fddi_isr(struct s_smc * smc)701*4882a593Smuzhiyun void fddi_isr(struct s_smc *smc)
702*4882a593Smuzhiyun {
703*4882a593Smuzhiyun u_long is ; /* ISR source */
704*4882a593Smuzhiyun u_short stu, stl ;
705*4882a593Smuzhiyun SMbuf *mb ;
706*4882a593Smuzhiyun
707*4882a593Smuzhiyun #ifdef USE_BREAK_ISR
708*4882a593Smuzhiyun int force_irq ;
709*4882a593Smuzhiyun #endif
710*4882a593Smuzhiyun
711*4882a593Smuzhiyun #ifdef ODI2
712*4882a593Smuzhiyun if (smc->os.hwm.rx_break) {
713*4882a593Smuzhiyun mac_drv_fill_rxd(smc) ;
714*4882a593Smuzhiyun if (smc->hw.fp.rx_q[QUEUE_R1].rx_used > 0) {
715*4882a593Smuzhiyun smc->os.hwm.rx_break = 0 ;
716*4882a593Smuzhiyun process_receive(smc) ;
717*4882a593Smuzhiyun }
718*4882a593Smuzhiyun else {
719*4882a593Smuzhiyun smc->os.hwm.detec_count = 0 ;
720*4882a593Smuzhiyun smt_force_irq(smc) ;
721*4882a593Smuzhiyun }
722*4882a593Smuzhiyun }
723*4882a593Smuzhiyun #endif
724*4882a593Smuzhiyun smc->os.hwm.isr_flag = TRUE ;
725*4882a593Smuzhiyun
726*4882a593Smuzhiyun #ifdef USE_BREAK_ISR
727*4882a593Smuzhiyun force_irq = TRUE ;
728*4882a593Smuzhiyun if (smc->os.hwm.leave_isr) {
729*4882a593Smuzhiyun smc->os.hwm.leave_isr = FALSE ;
730*4882a593Smuzhiyun process_receive(smc) ;
731*4882a593Smuzhiyun }
732*4882a593Smuzhiyun #endif
733*4882a593Smuzhiyun
734*4882a593Smuzhiyun while ((is = GET_ISR() & ISR_MASK)) {
735*4882a593Smuzhiyun NDD_TRACE("CH0B",is,0,0) ;
736*4882a593Smuzhiyun DB_GEN(7, "ISA = 0x%lx", is);
737*4882a593Smuzhiyun
738*4882a593Smuzhiyun if (is & IMASK_SLOW) {
739*4882a593Smuzhiyun NDD_TRACE("CH1b",is,0,0) ;
740*4882a593Smuzhiyun if (is & IS_PLINT1) { /* PLC1 */
741*4882a593Smuzhiyun plc1_irq(smc) ;
742*4882a593Smuzhiyun }
743*4882a593Smuzhiyun if (is & IS_PLINT2) { /* PLC2 */
744*4882a593Smuzhiyun plc2_irq(smc) ;
745*4882a593Smuzhiyun }
746*4882a593Smuzhiyun if (is & IS_MINTR1) { /* FORMAC+ STU1(U/L) */
747*4882a593Smuzhiyun stu = inpw(FM_A(FM_ST1U)) ;
748*4882a593Smuzhiyun stl = inpw(FM_A(FM_ST1L)) ;
749*4882a593Smuzhiyun DB_GEN(6, "Slow transmit complete");
750*4882a593Smuzhiyun mac1_irq(smc,stu,stl) ;
751*4882a593Smuzhiyun }
752*4882a593Smuzhiyun if (is & IS_MINTR2) { /* FORMAC+ STU2(U/L) */
753*4882a593Smuzhiyun stu= inpw(FM_A(FM_ST2U)) ;
754*4882a593Smuzhiyun stl= inpw(FM_A(FM_ST2L)) ;
755*4882a593Smuzhiyun DB_GEN(6, "Slow receive complete");
756*4882a593Smuzhiyun DB_GEN(7, "stl = %x : stu = %x", stl, stu);
757*4882a593Smuzhiyun mac2_irq(smc,stu,stl) ;
758*4882a593Smuzhiyun }
759*4882a593Smuzhiyun if (is & IS_MINTR3) { /* FORMAC+ STU3(U/L) */
760*4882a593Smuzhiyun stu= inpw(FM_A(FM_ST3U)) ;
761*4882a593Smuzhiyun stl= inpw(FM_A(FM_ST3L)) ;
762*4882a593Smuzhiyun DB_GEN(6, "FORMAC Mode Register 3");
763*4882a593Smuzhiyun mac3_irq(smc,stu,stl) ;
764*4882a593Smuzhiyun }
765*4882a593Smuzhiyun if (is & IS_TIMINT) { /* Timer 82C54-2 */
766*4882a593Smuzhiyun timer_irq(smc) ;
767*4882a593Smuzhiyun #ifdef NDIS_OS2
768*4882a593Smuzhiyun force_irq_pending = 0 ;
769*4882a593Smuzhiyun #endif
770*4882a593Smuzhiyun /*
771*4882a593Smuzhiyun * out of RxD detection
772*4882a593Smuzhiyun */
773*4882a593Smuzhiyun if (++smc->os.hwm.detec_count > 4) {
774*4882a593Smuzhiyun /*
775*4882a593Smuzhiyun * check out of RxD condition
776*4882a593Smuzhiyun */
777*4882a593Smuzhiyun process_receive(smc) ;
778*4882a593Smuzhiyun }
779*4882a593Smuzhiyun }
780*4882a593Smuzhiyun if (is & IS_TOKEN) { /* Restricted Token Monitor */
781*4882a593Smuzhiyun rtm_irq(smc) ;
782*4882a593Smuzhiyun }
783*4882a593Smuzhiyun if (is & IS_R1_P) { /* Parity error rx queue 1 */
784*4882a593Smuzhiyun /* clear IRQ */
785*4882a593Smuzhiyun outpd(ADDR(B4_R1_CSR),CSR_IRQ_CL_P) ;
786*4882a593Smuzhiyun SMT_PANIC(smc,HWM_E0004,HWM_E0004_MSG) ;
787*4882a593Smuzhiyun }
788*4882a593Smuzhiyun if (is & IS_R1_C) { /* Encoding error rx queue 1 */
789*4882a593Smuzhiyun /* clear IRQ */
790*4882a593Smuzhiyun outpd(ADDR(B4_R1_CSR),CSR_IRQ_CL_C) ;
791*4882a593Smuzhiyun SMT_PANIC(smc,HWM_E0005,HWM_E0005_MSG) ;
792*4882a593Smuzhiyun }
793*4882a593Smuzhiyun if (is & IS_XA_C) { /* Encoding error async tx q */
794*4882a593Smuzhiyun /* clear IRQ */
795*4882a593Smuzhiyun outpd(ADDR(B5_XA_CSR),CSR_IRQ_CL_C) ;
796*4882a593Smuzhiyun SMT_PANIC(smc,HWM_E0006,HWM_E0006_MSG) ;
797*4882a593Smuzhiyun }
798*4882a593Smuzhiyun if (is & IS_XS_C) { /* Encoding error sync tx q */
799*4882a593Smuzhiyun /* clear IRQ */
800*4882a593Smuzhiyun outpd(ADDR(B5_XS_CSR),CSR_IRQ_CL_C) ;
801*4882a593Smuzhiyun SMT_PANIC(smc,HWM_E0007,HWM_E0007_MSG) ;
802*4882a593Smuzhiyun }
803*4882a593Smuzhiyun }
804*4882a593Smuzhiyun
805*4882a593Smuzhiyun /*
806*4882a593Smuzhiyun * Fast Tx complete Async/Sync Queue (BMU service)
807*4882a593Smuzhiyun */
808*4882a593Smuzhiyun if (is & (IS_XS_F|IS_XA_F)) {
809*4882a593Smuzhiyun DB_GEN(6, "Fast tx complete queue");
810*4882a593Smuzhiyun /*
811*4882a593Smuzhiyun * clear IRQ, Note: no IRQ is lost, because
812*4882a593Smuzhiyun * we always service both queues
813*4882a593Smuzhiyun */
814*4882a593Smuzhiyun outpd(ADDR(B5_XS_CSR),CSR_IRQ_CL_F) ;
815*4882a593Smuzhiyun outpd(ADDR(B5_XA_CSR),CSR_IRQ_CL_F) ;
816*4882a593Smuzhiyun mac_drv_clear_txd(smc) ;
817*4882a593Smuzhiyun llc_restart_tx(smc) ;
818*4882a593Smuzhiyun }
819*4882a593Smuzhiyun
820*4882a593Smuzhiyun /*
821*4882a593Smuzhiyun * Fast Rx Complete (BMU service)
822*4882a593Smuzhiyun */
823*4882a593Smuzhiyun if (is & IS_R1_F) {
824*4882a593Smuzhiyun DB_GEN(6, "Fast receive complete");
825*4882a593Smuzhiyun /* clear IRQ */
826*4882a593Smuzhiyun #ifndef USE_BREAK_ISR
827*4882a593Smuzhiyun outpd(ADDR(B4_R1_CSR),CSR_IRQ_CL_F) ;
828*4882a593Smuzhiyun process_receive(smc) ;
829*4882a593Smuzhiyun #else
830*4882a593Smuzhiyun process_receive(smc) ;
831*4882a593Smuzhiyun if (smc->os.hwm.leave_isr) {
832*4882a593Smuzhiyun force_irq = FALSE ;
833*4882a593Smuzhiyun } else {
834*4882a593Smuzhiyun outpd(ADDR(B4_R1_CSR),CSR_IRQ_CL_F) ;
835*4882a593Smuzhiyun process_receive(smc) ;
836*4882a593Smuzhiyun }
837*4882a593Smuzhiyun #endif
838*4882a593Smuzhiyun }
839*4882a593Smuzhiyun
840*4882a593Smuzhiyun #ifndef NDIS_OS2
841*4882a593Smuzhiyun while ((mb = get_llc_rx(smc))) {
842*4882a593Smuzhiyun smt_to_llc(smc,mb) ;
843*4882a593Smuzhiyun }
844*4882a593Smuzhiyun #else
845*4882a593Smuzhiyun if (offDepth)
846*4882a593Smuzhiyun post_proc() ;
847*4882a593Smuzhiyun
848*4882a593Smuzhiyun while (!offDepth && (mb = get_llc_rx(smc))) {
849*4882a593Smuzhiyun smt_to_llc(smc,mb) ;
850*4882a593Smuzhiyun }
851*4882a593Smuzhiyun
852*4882a593Smuzhiyun if (!offDepth && smc->os.hwm.rx_break) {
853*4882a593Smuzhiyun process_receive(smc) ;
854*4882a593Smuzhiyun }
855*4882a593Smuzhiyun #endif
856*4882a593Smuzhiyun if (smc->q.ev_get != smc->q.ev_put) {
857*4882a593Smuzhiyun NDD_TRACE("CH2a",0,0,0) ;
858*4882a593Smuzhiyun ev_dispatcher(smc) ;
859*4882a593Smuzhiyun }
860*4882a593Smuzhiyun #ifdef NDIS_OS2
861*4882a593Smuzhiyun post_proc() ;
862*4882a593Smuzhiyun if (offDepth) { /* leave fddi_isr because */
863*4882a593Smuzhiyun break ; /* indications not allowed */
864*4882a593Smuzhiyun }
865*4882a593Smuzhiyun #endif
866*4882a593Smuzhiyun #ifdef USE_BREAK_ISR
867*4882a593Smuzhiyun if (smc->os.hwm.leave_isr) {
868*4882a593Smuzhiyun break ; /* leave fddi_isr */
869*4882a593Smuzhiyun }
870*4882a593Smuzhiyun #endif
871*4882a593Smuzhiyun
872*4882a593Smuzhiyun /* NOTE: when the isr is left, no rx is pending */
873*4882a593Smuzhiyun } /* end of interrupt source polling loop */
874*4882a593Smuzhiyun
875*4882a593Smuzhiyun #ifdef USE_BREAK_ISR
876*4882a593Smuzhiyun if (smc->os.hwm.leave_isr && force_irq) {
877*4882a593Smuzhiyun smt_force_irq(smc) ;
878*4882a593Smuzhiyun }
879*4882a593Smuzhiyun #endif
880*4882a593Smuzhiyun smc->os.hwm.isr_flag = FALSE ;
881*4882a593Smuzhiyun NDD_TRACE("CH0E",0,0,0) ;
882*4882a593Smuzhiyun }
883*4882a593Smuzhiyun
884*4882a593Smuzhiyun
885*4882a593Smuzhiyun /*
886*4882a593Smuzhiyun -------------------------------------------------------------
887*4882a593Smuzhiyun RECEIVE FUNCTIONS:
888*4882a593Smuzhiyun -------------------------------------------------------------
889*4882a593Smuzhiyun */
890*4882a593Smuzhiyun
891*4882a593Smuzhiyun #ifndef NDIS_OS2
892*4882a593Smuzhiyun /*
893*4882a593Smuzhiyun * BEGIN_MANUAL_ENTRY(mac_drv_rx_mode)
894*4882a593Smuzhiyun * void mac_drv_rx_mode(smc,mode)
895*4882a593Smuzhiyun *
896*4882a593Smuzhiyun * function DOWNCALL (fplus.c)
897*4882a593Smuzhiyun * Corresponding to the parameter mode, the operating system
898*4882a593Smuzhiyun * dependent module can activate several receive modes.
899*4882a593Smuzhiyun *
900*4882a593Smuzhiyun * para mode = 1: RX_ENABLE_ALLMULTI enable all multicasts
901*4882a593Smuzhiyun * = 2: RX_DISABLE_ALLMULTI disable "enable all multicasts"
902*4882a593Smuzhiyun * = 3: RX_ENABLE_PROMISC enable promiscuous
903*4882a593Smuzhiyun * = 4: RX_DISABLE_PROMISC disable promiscuous
904*4882a593Smuzhiyun * = 5: RX_ENABLE_NSA enable rec. of all NSA frames
905*4882a593Smuzhiyun * (disabled after 'driver reset' & 'set station address')
906*4882a593Smuzhiyun * = 6: RX_DISABLE_NSA disable rec. of all NSA frames
907*4882a593Smuzhiyun *
908*4882a593Smuzhiyun * = 21: RX_ENABLE_PASS_SMT ( see description )
909*4882a593Smuzhiyun * = 22: RX_DISABLE_PASS_SMT ( " " )
910*4882a593Smuzhiyun * = 23: RX_ENABLE_PASS_NSA ( " " )
911*4882a593Smuzhiyun * = 24: RX_DISABLE_PASS_NSA ( " " )
912*4882a593Smuzhiyun * = 25: RX_ENABLE_PASS_DB ( " " )
913*4882a593Smuzhiyun * = 26: RX_DISABLE_PASS_DB ( " " )
914*4882a593Smuzhiyun * = 27: RX_DISABLE_PASS_ALL ( " " )
915*4882a593Smuzhiyun * = 28: RX_DISABLE_LLC_PROMISC ( " " )
916*4882a593Smuzhiyun * = 29: RX_ENABLE_LLC_PROMISC ( " " )
917*4882a593Smuzhiyun *
918*4882a593Smuzhiyun *
919*4882a593Smuzhiyun * RX_ENABLE_PASS_SMT / RX_DISABLE_PASS_SMT
920*4882a593Smuzhiyun *
921*4882a593Smuzhiyun * If the operating system dependent module activates the
922*4882a593Smuzhiyun * mode RX_ENABLE_PASS_SMT, the hardware module
923*4882a593Smuzhiyun * duplicates all SMT frames with the frame control
924*4882a593Smuzhiyun * FC_SMT_INFO and passes them to the LLC receive channel
925*4882a593Smuzhiyun * by calling mac_drv_rx_init.
926*4882a593Smuzhiyun * The SMT Frames which are sent by the local SMT and the NSA
927*4882a593Smuzhiyun * frames whose A- and C-Indicator is not set are also duplicated
928*4882a593Smuzhiyun * and passed.
929*4882a593Smuzhiyun * The receive mode RX_DISABLE_PASS_SMT disables the passing
930*4882a593Smuzhiyun * of SMT frames.
931*4882a593Smuzhiyun *
932*4882a593Smuzhiyun * RX_ENABLE_PASS_NSA / RX_DISABLE_PASS_NSA
933*4882a593Smuzhiyun *
934*4882a593Smuzhiyun * If the operating system dependent module activates the
935*4882a593Smuzhiyun * mode RX_ENABLE_PASS_NSA, the hardware module
936*4882a593Smuzhiyun * duplicates all NSA frames with frame control FC_SMT_NSA
937*4882a593Smuzhiyun * and a set A-Indicator and passed them to the LLC
938*4882a593Smuzhiyun * receive channel by calling mac_drv_rx_init.
939*4882a593Smuzhiyun * All NSA Frames which are sent by the local SMT
940*4882a593Smuzhiyun * are also duplicated and passed.
941*4882a593Smuzhiyun * The receive mode RX_DISABLE_PASS_NSA disables the passing
942*4882a593Smuzhiyun * of NSA frames with the A- or C-Indicator set.
943*4882a593Smuzhiyun *
944*4882a593Smuzhiyun * NOTE: For fear that the hardware module receives NSA frames with
945*4882a593Smuzhiyun * a reset A-Indicator, the operating system dependent module
946*4882a593Smuzhiyun * has to call mac_drv_rx_mode with the mode RX_ENABLE_NSA
947*4882a593Smuzhiyun * before activate the RX_ENABLE_PASS_NSA mode and after every
948*4882a593Smuzhiyun * 'driver reset' and 'set station address'.
949*4882a593Smuzhiyun *
950*4882a593Smuzhiyun * RX_ENABLE_PASS_DB / RX_DISABLE_PASS_DB
951*4882a593Smuzhiyun *
952*4882a593Smuzhiyun * If the operating system dependent module activates the
953*4882a593Smuzhiyun * mode RX_ENABLE_PASS_DB, direct BEACON frames
954*4882a593Smuzhiyun * (FC_BEACON frame control) are passed to the LLC receive
955*4882a593Smuzhiyun * channel by mac_drv_rx_init.
956*4882a593Smuzhiyun * The receive mode RX_DISABLE_PASS_DB disables the passing
957*4882a593Smuzhiyun * of direct BEACON frames.
958*4882a593Smuzhiyun *
959*4882a593Smuzhiyun * RX_DISABLE_PASS_ALL
960*4882a593Smuzhiyun *
961*4882a593Smuzhiyun * Disables all special receives modes. It is equal to
962*4882a593Smuzhiyun * call mac_drv_set_rx_mode successively with the
963*4882a593Smuzhiyun * parameters RX_DISABLE_NSA, RX_DISABLE_PASS_SMT,
964*4882a593Smuzhiyun * RX_DISABLE_PASS_NSA and RX_DISABLE_PASS_DB.
965*4882a593Smuzhiyun *
966*4882a593Smuzhiyun * RX_ENABLE_LLC_PROMISC
967*4882a593Smuzhiyun *
968*4882a593Smuzhiyun * (default) all received LLC frames and all SMT/NSA/DBEACON
969*4882a593Smuzhiyun * frames depending on the attitude of the flags
970*4882a593Smuzhiyun * PASS_SMT/PASS_NSA/PASS_DBEACON will be delivered to the
971*4882a593Smuzhiyun * LLC layer
972*4882a593Smuzhiyun *
973*4882a593Smuzhiyun * RX_DISABLE_LLC_PROMISC
974*4882a593Smuzhiyun *
975*4882a593Smuzhiyun * all received SMT/NSA/DBEACON frames depending on the
976*4882a593Smuzhiyun * attitude of the flags PASS_SMT/PASS_NSA/PASS_DBEACON
977*4882a593Smuzhiyun * will be delivered to the LLC layer.
978*4882a593Smuzhiyun * all received LLC frames with a directed address, Multicast
979*4882a593Smuzhiyun * or Broadcast address will be delivered to the LLC
980*4882a593Smuzhiyun * layer too.
981*4882a593Smuzhiyun *
982*4882a593Smuzhiyun * END_MANUAL_ENTRY
983*4882a593Smuzhiyun */
mac_drv_rx_mode(struct s_smc * smc,int mode)984*4882a593Smuzhiyun void mac_drv_rx_mode(struct s_smc *smc, int mode)
985*4882a593Smuzhiyun {
986*4882a593Smuzhiyun switch(mode) {
987*4882a593Smuzhiyun case RX_ENABLE_PASS_SMT:
988*4882a593Smuzhiyun smc->os.hwm.pass_SMT = TRUE ;
989*4882a593Smuzhiyun break ;
990*4882a593Smuzhiyun case RX_DISABLE_PASS_SMT:
991*4882a593Smuzhiyun smc->os.hwm.pass_SMT = FALSE ;
992*4882a593Smuzhiyun break ;
993*4882a593Smuzhiyun case RX_ENABLE_PASS_NSA:
994*4882a593Smuzhiyun smc->os.hwm.pass_NSA = TRUE ;
995*4882a593Smuzhiyun break ;
996*4882a593Smuzhiyun case RX_DISABLE_PASS_NSA:
997*4882a593Smuzhiyun smc->os.hwm.pass_NSA = FALSE ;
998*4882a593Smuzhiyun break ;
999*4882a593Smuzhiyun case RX_ENABLE_PASS_DB:
1000*4882a593Smuzhiyun smc->os.hwm.pass_DB = TRUE ;
1001*4882a593Smuzhiyun break ;
1002*4882a593Smuzhiyun case RX_DISABLE_PASS_DB:
1003*4882a593Smuzhiyun smc->os.hwm.pass_DB = FALSE ;
1004*4882a593Smuzhiyun break ;
1005*4882a593Smuzhiyun case RX_DISABLE_PASS_ALL:
1006*4882a593Smuzhiyun smc->os.hwm.pass_SMT = smc->os.hwm.pass_NSA = FALSE ;
1007*4882a593Smuzhiyun smc->os.hwm.pass_DB = FALSE ;
1008*4882a593Smuzhiyun smc->os.hwm.pass_llc_promisc = TRUE ;
1009*4882a593Smuzhiyun mac_set_rx_mode(smc,RX_DISABLE_NSA) ;
1010*4882a593Smuzhiyun break ;
1011*4882a593Smuzhiyun case RX_DISABLE_LLC_PROMISC:
1012*4882a593Smuzhiyun smc->os.hwm.pass_llc_promisc = FALSE ;
1013*4882a593Smuzhiyun break ;
1014*4882a593Smuzhiyun case RX_ENABLE_LLC_PROMISC:
1015*4882a593Smuzhiyun smc->os.hwm.pass_llc_promisc = TRUE ;
1016*4882a593Smuzhiyun break ;
1017*4882a593Smuzhiyun case RX_ENABLE_ALLMULTI:
1018*4882a593Smuzhiyun case RX_DISABLE_ALLMULTI:
1019*4882a593Smuzhiyun case RX_ENABLE_PROMISC:
1020*4882a593Smuzhiyun case RX_DISABLE_PROMISC:
1021*4882a593Smuzhiyun case RX_ENABLE_NSA:
1022*4882a593Smuzhiyun case RX_DISABLE_NSA:
1023*4882a593Smuzhiyun default:
1024*4882a593Smuzhiyun mac_set_rx_mode(smc,mode) ;
1025*4882a593Smuzhiyun break ;
1026*4882a593Smuzhiyun }
1027*4882a593Smuzhiyun }
1028*4882a593Smuzhiyun #endif /* ifndef NDIS_OS2 */
1029*4882a593Smuzhiyun
1030*4882a593Smuzhiyun /*
1031*4882a593Smuzhiyun * process receive queue
1032*4882a593Smuzhiyun */
process_receive(struct s_smc * smc)1033*4882a593Smuzhiyun void process_receive(struct s_smc *smc)
1034*4882a593Smuzhiyun {
1035*4882a593Smuzhiyun int i ;
1036*4882a593Smuzhiyun int n ;
1037*4882a593Smuzhiyun int frag_count ; /* number of RxDs of the curr rx buf */
1038*4882a593Smuzhiyun int used_frags ; /* number of RxDs of the curr frame */
1039*4882a593Smuzhiyun struct s_smt_rx_queue *queue ; /* points to the queue ctl struct */
1040*4882a593Smuzhiyun struct s_smt_fp_rxd volatile *r ; /* rxd pointer */
1041*4882a593Smuzhiyun struct s_smt_fp_rxd volatile *rxd ; /* first rxd of rx frame */
1042*4882a593Smuzhiyun u_long rbctrl ; /* receive buffer control word */
1043*4882a593Smuzhiyun u_long rfsw ; /* receive frame status word */
1044*4882a593Smuzhiyun u_short rx_used ;
1045*4882a593Smuzhiyun u_char far *virt ;
1046*4882a593Smuzhiyun char far *data ;
1047*4882a593Smuzhiyun SMbuf *mb ;
1048*4882a593Smuzhiyun u_char fc ; /* Frame control */
1049*4882a593Smuzhiyun int len ; /* Frame length */
1050*4882a593Smuzhiyun
1051*4882a593Smuzhiyun smc->os.hwm.detec_count = 0 ;
1052*4882a593Smuzhiyun queue = smc->hw.fp.rx[QUEUE_R1] ;
1053*4882a593Smuzhiyun NDD_TRACE("RHxB",0,0,0) ;
1054*4882a593Smuzhiyun for ( ; ; ) {
1055*4882a593Smuzhiyun r = queue->rx_curr_get ;
1056*4882a593Smuzhiyun rx_used = queue->rx_used ;
1057*4882a593Smuzhiyun frag_count = 0 ;
1058*4882a593Smuzhiyun
1059*4882a593Smuzhiyun #ifdef USE_BREAK_ISR
1060*4882a593Smuzhiyun if (smc->os.hwm.leave_isr) {
1061*4882a593Smuzhiyun goto rx_end ;
1062*4882a593Smuzhiyun }
1063*4882a593Smuzhiyun #endif
1064*4882a593Smuzhiyun #ifdef NDIS_OS2
1065*4882a593Smuzhiyun if (offDepth) {
1066*4882a593Smuzhiyun smc->os.hwm.rx_break = 1 ;
1067*4882a593Smuzhiyun goto rx_end ;
1068*4882a593Smuzhiyun }
1069*4882a593Smuzhiyun smc->os.hwm.rx_break = 0 ;
1070*4882a593Smuzhiyun #endif
1071*4882a593Smuzhiyun #ifdef ODI2
1072*4882a593Smuzhiyun if (smc->os.hwm.rx_break) {
1073*4882a593Smuzhiyun goto rx_end ;
1074*4882a593Smuzhiyun }
1075*4882a593Smuzhiyun #endif
1076*4882a593Smuzhiyun n = 0 ;
1077*4882a593Smuzhiyun do {
1078*4882a593Smuzhiyun DB_RX(5, "Check RxD %p for OWN and EOF", r);
1079*4882a593Smuzhiyun DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORCPU) ;
1080*4882a593Smuzhiyun rbctrl = le32_to_cpu(CR_READ(r->rxd_rbctrl));
1081*4882a593Smuzhiyun
1082*4882a593Smuzhiyun if (rbctrl & BMU_OWN) {
1083*4882a593Smuzhiyun NDD_TRACE("RHxE",r,rfsw,rbctrl) ;
1084*4882a593Smuzhiyun DB_RX(4, "End of RxDs");
1085*4882a593Smuzhiyun goto rx_end ;
1086*4882a593Smuzhiyun }
1087*4882a593Smuzhiyun /*
1088*4882a593Smuzhiyun * out of RxD detection
1089*4882a593Smuzhiyun */
1090*4882a593Smuzhiyun if (!rx_used) {
1091*4882a593Smuzhiyun SK_BREAK() ;
1092*4882a593Smuzhiyun SMT_PANIC(smc,HWM_E0009,HWM_E0009_MSG) ;
1093*4882a593Smuzhiyun /* Either we don't have an RxD or all
1094*4882a593Smuzhiyun * RxDs are filled. Therefore it's allowed
1095*4882a593Smuzhiyun * for to set the STOPPED flag */
1096*4882a593Smuzhiyun smc->hw.hw_state = STOPPED ;
1097*4882a593Smuzhiyun mac_drv_clear_rx_queue(smc) ;
1098*4882a593Smuzhiyun smc->hw.hw_state = STARTED ;
1099*4882a593Smuzhiyun mac_drv_fill_rxd(smc) ;
1100*4882a593Smuzhiyun smc->os.hwm.detec_count = 0 ;
1101*4882a593Smuzhiyun goto rx_end ;
1102*4882a593Smuzhiyun }
1103*4882a593Smuzhiyun rfsw = le32_to_cpu(r->rxd_rfsw) ;
1104*4882a593Smuzhiyun if ((rbctrl & BMU_STF) != ((rbctrl & BMU_ST_BUF) <<5)) {
1105*4882a593Smuzhiyun /*
1106*4882a593Smuzhiyun * The BMU_STF bit is deleted, 1 frame is
1107*4882a593Smuzhiyun * placed into more than 1 rx buffer
1108*4882a593Smuzhiyun *
1109*4882a593Smuzhiyun * skip frame by setting the rx len to 0
1110*4882a593Smuzhiyun *
1111*4882a593Smuzhiyun * if fragment count == 0
1112*4882a593Smuzhiyun * The missing STF bit belongs to the
1113*4882a593Smuzhiyun * current frame, search for the
1114*4882a593Smuzhiyun * EOF bit to complete the frame
1115*4882a593Smuzhiyun * else
1116*4882a593Smuzhiyun * the fragment belongs to the next frame,
1117*4882a593Smuzhiyun * exit the loop and process the frame
1118*4882a593Smuzhiyun */
1119*4882a593Smuzhiyun SK_BREAK() ;
1120*4882a593Smuzhiyun rfsw = 0 ;
1121*4882a593Smuzhiyun if (frag_count) {
1122*4882a593Smuzhiyun break ;
1123*4882a593Smuzhiyun }
1124*4882a593Smuzhiyun }
1125*4882a593Smuzhiyun n += rbctrl & 0xffff ;
1126*4882a593Smuzhiyun r = r->rxd_next ;
1127*4882a593Smuzhiyun frag_count++ ;
1128*4882a593Smuzhiyun rx_used-- ;
1129*4882a593Smuzhiyun } while (!(rbctrl & BMU_EOF)) ;
1130*4882a593Smuzhiyun used_frags = frag_count ;
1131*4882a593Smuzhiyun DB_RX(5, "EOF set in RxD, used_frags = %d", used_frags);
1132*4882a593Smuzhiyun
1133*4882a593Smuzhiyun /* may be next 2 DRV_BUF_FLUSH() can be skipped, because */
1134*4882a593Smuzhiyun /* BMU_ST_BUF will not be changed by the ASIC */
1135*4882a593Smuzhiyun DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORCPU) ;
1136*4882a593Smuzhiyun while (rx_used && !(r->rxd_rbctrl & cpu_to_le32(BMU_ST_BUF))) {
1137*4882a593Smuzhiyun DB_RX(5, "Check STF bit in %p", r);
1138*4882a593Smuzhiyun r = r->rxd_next ;
1139*4882a593Smuzhiyun DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORCPU) ;
1140*4882a593Smuzhiyun frag_count++ ;
1141*4882a593Smuzhiyun rx_used-- ;
1142*4882a593Smuzhiyun }
1143*4882a593Smuzhiyun DB_RX(5, "STF bit found");
1144*4882a593Smuzhiyun
1145*4882a593Smuzhiyun /*
1146*4882a593Smuzhiyun * The received frame is finished for the process receive
1147*4882a593Smuzhiyun */
1148*4882a593Smuzhiyun rxd = queue->rx_curr_get ;
1149*4882a593Smuzhiyun queue->rx_curr_get = r ;
1150*4882a593Smuzhiyun queue->rx_free += frag_count ;
1151*4882a593Smuzhiyun queue->rx_used = rx_used ;
1152*4882a593Smuzhiyun
1153*4882a593Smuzhiyun /*
1154*4882a593Smuzhiyun * ASIC Errata no. 7 (STF - Bit Bug)
1155*4882a593Smuzhiyun */
1156*4882a593Smuzhiyun rxd->rxd_rbctrl &= cpu_to_le32(~BMU_STF) ;
1157*4882a593Smuzhiyun
1158*4882a593Smuzhiyun for (r=rxd, i=frag_count ; i ; r=r->rxd_next, i--){
1159*4882a593Smuzhiyun DB_RX(5, "dma_complete for RxD %p", r);
1160*4882a593Smuzhiyun dma_complete(smc,(union s_fp_descr volatile *)r,DMA_WR);
1161*4882a593Smuzhiyun }
1162*4882a593Smuzhiyun smc->hw.fp.err_stats.err_valid++ ;
1163*4882a593Smuzhiyun smc->mib.m[MAC0].fddiMACCopied_Ct++ ;
1164*4882a593Smuzhiyun
1165*4882a593Smuzhiyun /* the length of the data including the FC */
1166*4882a593Smuzhiyun len = (rfsw & RD_LENGTH) - 4 ;
1167*4882a593Smuzhiyun
1168*4882a593Smuzhiyun DB_RX(4, "frame length = %d", len);
1169*4882a593Smuzhiyun /*
1170*4882a593Smuzhiyun * check the frame_length and all error flags
1171*4882a593Smuzhiyun */
1172*4882a593Smuzhiyun if (rfsw & (RX_MSRABT|RX_FS_E|RX_FS_CRC|RX_FS_IMPL)){
1173*4882a593Smuzhiyun if (rfsw & RD_S_MSRABT) {
1174*4882a593Smuzhiyun DB_RX(2, "Frame aborted by the FORMAC");
1175*4882a593Smuzhiyun smc->hw.fp.err_stats.err_abort++ ;
1176*4882a593Smuzhiyun }
1177*4882a593Smuzhiyun /*
1178*4882a593Smuzhiyun * check frame status
1179*4882a593Smuzhiyun */
1180*4882a593Smuzhiyun if (rfsw & RD_S_SEAC2) {
1181*4882a593Smuzhiyun DB_RX(2, "E-Indicator set");
1182*4882a593Smuzhiyun smc->hw.fp.err_stats.err_e_indicator++ ;
1183*4882a593Smuzhiyun }
1184*4882a593Smuzhiyun if (rfsw & RD_S_SFRMERR) {
1185*4882a593Smuzhiyun DB_RX(2, "CRC error");
1186*4882a593Smuzhiyun smc->hw.fp.err_stats.err_crc++ ;
1187*4882a593Smuzhiyun }
1188*4882a593Smuzhiyun if (rfsw & RX_FS_IMPL) {
1189*4882a593Smuzhiyun DB_RX(2, "Implementer frame");
1190*4882a593Smuzhiyun smc->hw.fp.err_stats.err_imp_frame++ ;
1191*4882a593Smuzhiyun }
1192*4882a593Smuzhiyun goto abort_frame ;
1193*4882a593Smuzhiyun }
1194*4882a593Smuzhiyun if (len > FDDI_RAW_MTU-4) {
1195*4882a593Smuzhiyun DB_RX(2, "Frame too long error");
1196*4882a593Smuzhiyun smc->hw.fp.err_stats.err_too_long++ ;
1197*4882a593Smuzhiyun goto abort_frame ;
1198*4882a593Smuzhiyun }
1199*4882a593Smuzhiyun /*
1200*4882a593Smuzhiyun * SUPERNET 3 Bug: FORMAC delivers status words
1201*4882a593Smuzhiyun * of aborted frames to the BMU
1202*4882a593Smuzhiyun */
1203*4882a593Smuzhiyun if (len <= 4) {
1204*4882a593Smuzhiyun DB_RX(2, "Frame length = 0");
1205*4882a593Smuzhiyun goto abort_frame ;
1206*4882a593Smuzhiyun }
1207*4882a593Smuzhiyun
1208*4882a593Smuzhiyun if (len != (n-4)) {
1209*4882a593Smuzhiyun DB_RX(4, "BMU: rx len differs: [%d:%d]", len, n);
1210*4882a593Smuzhiyun smc->os.hwm.rx_len_error++ ;
1211*4882a593Smuzhiyun goto abort_frame ;
1212*4882a593Smuzhiyun }
1213*4882a593Smuzhiyun
1214*4882a593Smuzhiyun /*
1215*4882a593Smuzhiyun * Check SA == MA
1216*4882a593Smuzhiyun */
1217*4882a593Smuzhiyun virt = (u_char far *) rxd->rxd_virt ;
1218*4882a593Smuzhiyun DB_RX(2, "FC = %x", *virt);
1219*4882a593Smuzhiyun if (virt[12] == MA[5] &&
1220*4882a593Smuzhiyun virt[11] == MA[4] &&
1221*4882a593Smuzhiyun virt[10] == MA[3] &&
1222*4882a593Smuzhiyun virt[9] == MA[2] &&
1223*4882a593Smuzhiyun virt[8] == MA[1] &&
1224*4882a593Smuzhiyun (virt[7] & ~GROUP_ADDR_BIT) == MA[0]) {
1225*4882a593Smuzhiyun goto abort_frame ;
1226*4882a593Smuzhiyun }
1227*4882a593Smuzhiyun
1228*4882a593Smuzhiyun /*
1229*4882a593Smuzhiyun * test if LLC frame
1230*4882a593Smuzhiyun */
1231*4882a593Smuzhiyun if (rfsw & RX_FS_LLC) {
1232*4882a593Smuzhiyun /*
1233*4882a593Smuzhiyun * if pass_llc_promisc is disable
1234*4882a593Smuzhiyun * if DA != Multicast or Broadcast or DA!=MA
1235*4882a593Smuzhiyun * abort the frame
1236*4882a593Smuzhiyun */
1237*4882a593Smuzhiyun if (!smc->os.hwm.pass_llc_promisc) {
1238*4882a593Smuzhiyun if(!(virt[1] & GROUP_ADDR_BIT)) {
1239*4882a593Smuzhiyun if (virt[6] != MA[5] ||
1240*4882a593Smuzhiyun virt[5] != MA[4] ||
1241*4882a593Smuzhiyun virt[4] != MA[3] ||
1242*4882a593Smuzhiyun virt[3] != MA[2] ||
1243*4882a593Smuzhiyun virt[2] != MA[1] ||
1244*4882a593Smuzhiyun virt[1] != MA[0]) {
1245*4882a593Smuzhiyun DB_RX(2, "DA != MA and not multi- or broadcast");
1246*4882a593Smuzhiyun goto abort_frame ;
1247*4882a593Smuzhiyun }
1248*4882a593Smuzhiyun }
1249*4882a593Smuzhiyun }
1250*4882a593Smuzhiyun
1251*4882a593Smuzhiyun /*
1252*4882a593Smuzhiyun * LLC frame received
1253*4882a593Smuzhiyun */
1254*4882a593Smuzhiyun DB_RX(4, "LLC - receive");
1255*4882a593Smuzhiyun mac_drv_rx_complete(smc,rxd,frag_count,len) ;
1256*4882a593Smuzhiyun }
1257*4882a593Smuzhiyun else {
1258*4882a593Smuzhiyun if (!(mb = smt_get_mbuf(smc))) {
1259*4882a593Smuzhiyun smc->hw.fp.err_stats.err_no_buf++ ;
1260*4882a593Smuzhiyun DB_RX(4, "No SMbuf; receive terminated");
1261*4882a593Smuzhiyun goto abort_frame ;
1262*4882a593Smuzhiyun }
1263*4882a593Smuzhiyun data = smtod(mb,char *) - 1 ;
1264*4882a593Smuzhiyun
1265*4882a593Smuzhiyun /*
1266*4882a593Smuzhiyun * copy the frame into a SMT_MBuf
1267*4882a593Smuzhiyun */
1268*4882a593Smuzhiyun #ifdef USE_OS_CPY
1269*4882a593Smuzhiyun hwm_cpy_rxd2mb(rxd,data,len) ;
1270*4882a593Smuzhiyun #else
1271*4882a593Smuzhiyun for (r=rxd, i=used_frags ; i ; r=r->rxd_next, i--){
1272*4882a593Smuzhiyun n = le32_to_cpu(r->rxd_rbctrl) & RD_LENGTH ;
1273*4882a593Smuzhiyun DB_RX(6, "cp SMT frame to mb: len = %d", n);
1274*4882a593Smuzhiyun memcpy(data,r->rxd_virt,n) ;
1275*4882a593Smuzhiyun data += n ;
1276*4882a593Smuzhiyun }
1277*4882a593Smuzhiyun data = smtod(mb,char *) - 1 ;
1278*4882a593Smuzhiyun #endif
1279*4882a593Smuzhiyun fc = *(char *)mb->sm_data = *data ;
1280*4882a593Smuzhiyun mb->sm_len = len - 1 ; /* len - fc */
1281*4882a593Smuzhiyun data++ ;
1282*4882a593Smuzhiyun
1283*4882a593Smuzhiyun /*
1284*4882a593Smuzhiyun * SMT frame received
1285*4882a593Smuzhiyun */
1286*4882a593Smuzhiyun switch(fc) {
1287*4882a593Smuzhiyun case FC_SMT_INFO :
1288*4882a593Smuzhiyun smc->hw.fp.err_stats.err_smt_frame++ ;
1289*4882a593Smuzhiyun DB_RX(5, "SMT frame received");
1290*4882a593Smuzhiyun
1291*4882a593Smuzhiyun if (smc->os.hwm.pass_SMT) {
1292*4882a593Smuzhiyun DB_RX(5, "pass SMT frame");
1293*4882a593Smuzhiyun mac_drv_rx_complete(smc, rxd,
1294*4882a593Smuzhiyun frag_count,len) ;
1295*4882a593Smuzhiyun }
1296*4882a593Smuzhiyun else {
1297*4882a593Smuzhiyun DB_RX(5, "requeue RxD");
1298*4882a593Smuzhiyun mac_drv_requeue_rxd(smc,rxd,frag_count);
1299*4882a593Smuzhiyun }
1300*4882a593Smuzhiyun
1301*4882a593Smuzhiyun smt_received_pack(smc,mb,(int)(rfsw>>25)) ;
1302*4882a593Smuzhiyun break ;
1303*4882a593Smuzhiyun case FC_SMT_NSA :
1304*4882a593Smuzhiyun smc->hw.fp.err_stats.err_smt_frame++ ;
1305*4882a593Smuzhiyun DB_RX(5, "SMT frame received");
1306*4882a593Smuzhiyun
1307*4882a593Smuzhiyun /* if pass_NSA set pass the NSA frame or */
1308*4882a593Smuzhiyun /* pass_SMT set and the A-Indicator */
1309*4882a593Smuzhiyun /* is not set, pass the NSA frame */
1310*4882a593Smuzhiyun if (smc->os.hwm.pass_NSA ||
1311*4882a593Smuzhiyun (smc->os.hwm.pass_SMT &&
1312*4882a593Smuzhiyun !(rfsw & A_INDIC))) {
1313*4882a593Smuzhiyun DB_RX(5, "pass SMT frame");
1314*4882a593Smuzhiyun mac_drv_rx_complete(smc, rxd,
1315*4882a593Smuzhiyun frag_count,len) ;
1316*4882a593Smuzhiyun }
1317*4882a593Smuzhiyun else {
1318*4882a593Smuzhiyun DB_RX(5, "requeue RxD");
1319*4882a593Smuzhiyun mac_drv_requeue_rxd(smc,rxd,frag_count);
1320*4882a593Smuzhiyun }
1321*4882a593Smuzhiyun
1322*4882a593Smuzhiyun smt_received_pack(smc,mb,(int)(rfsw>>25)) ;
1323*4882a593Smuzhiyun break ;
1324*4882a593Smuzhiyun case FC_BEACON :
1325*4882a593Smuzhiyun if (smc->os.hwm.pass_DB) {
1326*4882a593Smuzhiyun DB_RX(5, "pass DB frame");
1327*4882a593Smuzhiyun mac_drv_rx_complete(smc, rxd,
1328*4882a593Smuzhiyun frag_count,len) ;
1329*4882a593Smuzhiyun }
1330*4882a593Smuzhiyun else {
1331*4882a593Smuzhiyun DB_RX(5, "requeue RxD");
1332*4882a593Smuzhiyun mac_drv_requeue_rxd(smc,rxd,frag_count);
1333*4882a593Smuzhiyun }
1334*4882a593Smuzhiyun smt_free_mbuf(smc,mb) ;
1335*4882a593Smuzhiyun break ;
1336*4882a593Smuzhiyun default :
1337*4882a593Smuzhiyun /*
1338*4882a593Smuzhiyun * unknown FC abort the frame
1339*4882a593Smuzhiyun */
1340*4882a593Smuzhiyun DB_RX(2, "unknown FC error");
1341*4882a593Smuzhiyun smt_free_mbuf(smc,mb) ;
1342*4882a593Smuzhiyun DB_RX(5, "requeue RxD");
1343*4882a593Smuzhiyun mac_drv_requeue_rxd(smc,rxd,frag_count) ;
1344*4882a593Smuzhiyun if ((fc & 0xf0) == FC_MAC)
1345*4882a593Smuzhiyun smc->hw.fp.err_stats.err_mac_frame++ ;
1346*4882a593Smuzhiyun else
1347*4882a593Smuzhiyun smc->hw.fp.err_stats.err_imp_frame++ ;
1348*4882a593Smuzhiyun
1349*4882a593Smuzhiyun break ;
1350*4882a593Smuzhiyun }
1351*4882a593Smuzhiyun }
1352*4882a593Smuzhiyun
1353*4882a593Smuzhiyun DB_RX(3, "next RxD is %p", queue->rx_curr_get);
1354*4882a593Smuzhiyun NDD_TRACE("RHx1",queue->rx_curr_get,0,0) ;
1355*4882a593Smuzhiyun
1356*4882a593Smuzhiyun continue ;
1357*4882a593Smuzhiyun /*--------------------------------------------------------------------*/
1358*4882a593Smuzhiyun abort_frame:
1359*4882a593Smuzhiyun DB_RX(5, "requeue RxD");
1360*4882a593Smuzhiyun mac_drv_requeue_rxd(smc,rxd,frag_count) ;
1361*4882a593Smuzhiyun
1362*4882a593Smuzhiyun DB_RX(3, "next RxD is %p", queue->rx_curr_get);
1363*4882a593Smuzhiyun NDD_TRACE("RHx2",queue->rx_curr_get,0,0) ;
1364*4882a593Smuzhiyun }
1365*4882a593Smuzhiyun rx_end:
1366*4882a593Smuzhiyun #ifdef ALL_RX_COMPLETE
1367*4882a593Smuzhiyun mac_drv_all_receives_complete(smc) ;
1368*4882a593Smuzhiyun #endif
1369*4882a593Smuzhiyun return ; /* lint bug: needs return detect end of function */
1370*4882a593Smuzhiyun }
1371*4882a593Smuzhiyun
smt_to_llc(struct s_smc * smc,SMbuf * mb)1372*4882a593Smuzhiyun static void smt_to_llc(struct s_smc *smc, SMbuf *mb)
1373*4882a593Smuzhiyun {
1374*4882a593Smuzhiyun u_char fc ;
1375*4882a593Smuzhiyun
1376*4882a593Smuzhiyun DB_RX(4, "send a queued frame to the llc layer");
1377*4882a593Smuzhiyun smc->os.hwm.r.len = mb->sm_len ;
1378*4882a593Smuzhiyun smc->os.hwm.r.mb_pos = smtod(mb,char *) ;
1379*4882a593Smuzhiyun fc = *smc->os.hwm.r.mb_pos ;
1380*4882a593Smuzhiyun (void)mac_drv_rx_init(smc,(int)mb->sm_len,(int)fc,
1381*4882a593Smuzhiyun smc->os.hwm.r.mb_pos,(int)mb->sm_len) ;
1382*4882a593Smuzhiyun smt_free_mbuf(smc,mb) ;
1383*4882a593Smuzhiyun }
1384*4882a593Smuzhiyun
1385*4882a593Smuzhiyun /*
1386*4882a593Smuzhiyun * BEGIN_MANUAL_ENTRY(hwm_rx_frag)
1387*4882a593Smuzhiyun * void hwm_rx_frag(smc,virt,phys,len,frame_status)
1388*4882a593Smuzhiyun *
1389*4882a593Smuzhiyun * function MACRO (hardware module, hwmtm.h)
1390*4882a593Smuzhiyun * This function calls dma_master for preparing the
1391*4882a593Smuzhiyun * system hardware for the DMA transfer and initializes
1392*4882a593Smuzhiyun * the current RxD with the length and the physical and
1393*4882a593Smuzhiyun * virtual address of the fragment. Furthermore, it sets the
1394*4882a593Smuzhiyun * STF and EOF bits depending on the frame status byte,
1395*4882a593Smuzhiyun * switches the OWN flag of the RxD, so that it is owned by the
1396*4882a593Smuzhiyun * adapter and issues an rx_start.
1397*4882a593Smuzhiyun *
1398*4882a593Smuzhiyun * para virt virtual pointer to the fragment
1399*4882a593Smuzhiyun * len the length of the fragment
1400*4882a593Smuzhiyun * frame_status status of the frame, see design description
1401*4882a593Smuzhiyun *
1402*4882a593Smuzhiyun * NOTE: It is possible to call this function with a fragment length
1403*4882a593Smuzhiyun * of zero.
1404*4882a593Smuzhiyun *
1405*4882a593Smuzhiyun * END_MANUAL_ENTRY
1406*4882a593Smuzhiyun */
hwm_rx_frag(struct s_smc * smc,char far * virt,u_long phys,int len,int frame_status)1407*4882a593Smuzhiyun void hwm_rx_frag(struct s_smc *smc, char far *virt, u_long phys, int len,
1408*4882a593Smuzhiyun int frame_status)
1409*4882a593Smuzhiyun {
1410*4882a593Smuzhiyun struct s_smt_fp_rxd volatile *r ;
1411*4882a593Smuzhiyun __le32 rbctrl;
1412*4882a593Smuzhiyun
1413*4882a593Smuzhiyun NDD_TRACE("RHfB",virt,len,frame_status) ;
1414*4882a593Smuzhiyun DB_RX(2, "hwm_rx_frag: len = %d, frame_status = %x", len, frame_status);
1415*4882a593Smuzhiyun r = smc->hw.fp.rx_q[QUEUE_R1].rx_curr_put ;
1416*4882a593Smuzhiyun r->rxd_virt = virt ;
1417*4882a593Smuzhiyun r->rxd_rbadr = cpu_to_le32(phys) ;
1418*4882a593Smuzhiyun rbctrl = cpu_to_le32( (((__u32)frame_status &
1419*4882a593Smuzhiyun (FIRST_FRAG|LAST_FRAG))<<26) |
1420*4882a593Smuzhiyun (((u_long) frame_status & FIRST_FRAG) << 21) |
1421*4882a593Smuzhiyun BMU_OWN | BMU_CHECK | BMU_EN_IRQ_EOF | len) ;
1422*4882a593Smuzhiyun r->rxd_rbctrl = rbctrl ;
1423*4882a593Smuzhiyun
1424*4882a593Smuzhiyun DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORDEV) ;
1425*4882a593Smuzhiyun outpd(ADDR(B0_R1_CSR),CSR_START) ;
1426*4882a593Smuzhiyun smc->hw.fp.rx_q[QUEUE_R1].rx_free-- ;
1427*4882a593Smuzhiyun smc->hw.fp.rx_q[QUEUE_R1].rx_used++ ;
1428*4882a593Smuzhiyun smc->hw.fp.rx_q[QUEUE_R1].rx_curr_put = r->rxd_next ;
1429*4882a593Smuzhiyun NDD_TRACE("RHfE",r,le32_to_cpu(r->rxd_rbadr),0) ;
1430*4882a593Smuzhiyun }
1431*4882a593Smuzhiyun
1432*4882a593Smuzhiyun /*
1433*4882a593Smuzhiyun * BEGINN_MANUAL_ENTRY(mac_drv_clear_rx_queue)
1434*4882a593Smuzhiyun *
1435*4882a593Smuzhiyun * void mac_drv_clear_rx_queue(smc)
1436*4882a593Smuzhiyun * struct s_smc *smc ;
1437*4882a593Smuzhiyun *
1438*4882a593Smuzhiyun * function DOWNCALL (hardware module, hwmtm.c)
1439*4882a593Smuzhiyun * mac_drv_clear_rx_queue is called by the OS-specific module
1440*4882a593Smuzhiyun * after it has issued a card_stop.
1441*4882a593Smuzhiyun * In this case, the frames in the receive queue are obsolete and
1442*4882a593Smuzhiyun * should be removed. For removing mac_drv_clear_rx_queue
1443*4882a593Smuzhiyun * calls dma_master for each RxD and mac_drv_clear_rxd for each
1444*4882a593Smuzhiyun * receive buffer.
1445*4882a593Smuzhiyun *
1446*4882a593Smuzhiyun * NOTE: calling sequence card_stop:
1447*4882a593Smuzhiyun * CLI_FBI(), card_stop(),
1448*4882a593Smuzhiyun * mac_drv_clear_tx_queue(), mac_drv_clear_rx_queue(),
1449*4882a593Smuzhiyun *
1450*4882a593Smuzhiyun * NOTE: The caller is responsible that the BMUs are idle
1451*4882a593Smuzhiyun * when this function is called.
1452*4882a593Smuzhiyun *
1453*4882a593Smuzhiyun * END_MANUAL_ENTRY
1454*4882a593Smuzhiyun */
mac_drv_clear_rx_queue(struct s_smc * smc)1455*4882a593Smuzhiyun void mac_drv_clear_rx_queue(struct s_smc *smc)
1456*4882a593Smuzhiyun {
1457*4882a593Smuzhiyun struct s_smt_fp_rxd volatile *r ;
1458*4882a593Smuzhiyun struct s_smt_fp_rxd volatile *next_rxd ;
1459*4882a593Smuzhiyun struct s_smt_rx_queue *queue ;
1460*4882a593Smuzhiyun int frag_count ;
1461*4882a593Smuzhiyun int i ;
1462*4882a593Smuzhiyun
1463*4882a593Smuzhiyun if (smc->hw.hw_state != STOPPED) {
1464*4882a593Smuzhiyun SK_BREAK() ;
1465*4882a593Smuzhiyun SMT_PANIC(smc,HWM_E0012,HWM_E0012_MSG) ;
1466*4882a593Smuzhiyun return ;
1467*4882a593Smuzhiyun }
1468*4882a593Smuzhiyun
1469*4882a593Smuzhiyun queue = smc->hw.fp.rx[QUEUE_R1] ;
1470*4882a593Smuzhiyun DB_RX(5, "clear_rx_queue");
1471*4882a593Smuzhiyun
1472*4882a593Smuzhiyun /*
1473*4882a593Smuzhiyun * dma_complete and mac_drv_clear_rxd for all RxDs / receive buffers
1474*4882a593Smuzhiyun */
1475*4882a593Smuzhiyun r = queue->rx_curr_get ;
1476*4882a593Smuzhiyun while (queue->rx_used) {
1477*4882a593Smuzhiyun DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORCPU) ;
1478*4882a593Smuzhiyun DB_RX(5, "switch OWN bit of RxD 0x%p", r);
1479*4882a593Smuzhiyun r->rxd_rbctrl &= ~cpu_to_le32(BMU_OWN) ;
1480*4882a593Smuzhiyun frag_count = 1 ;
1481*4882a593Smuzhiyun DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORDEV) ;
1482*4882a593Smuzhiyun r = r->rxd_next ;
1483*4882a593Smuzhiyun DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORCPU) ;
1484*4882a593Smuzhiyun while (r != queue->rx_curr_put &&
1485*4882a593Smuzhiyun !(r->rxd_rbctrl & cpu_to_le32(BMU_ST_BUF))) {
1486*4882a593Smuzhiyun DB_RX(5, "Check STF bit in %p", r);
1487*4882a593Smuzhiyun r->rxd_rbctrl &= ~cpu_to_le32(BMU_OWN) ;
1488*4882a593Smuzhiyun DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORDEV) ;
1489*4882a593Smuzhiyun r = r->rxd_next ;
1490*4882a593Smuzhiyun DRV_BUF_FLUSH(r,DDI_DMA_SYNC_FORCPU) ;
1491*4882a593Smuzhiyun frag_count++ ;
1492*4882a593Smuzhiyun }
1493*4882a593Smuzhiyun DB_RX(5, "STF bit found");
1494*4882a593Smuzhiyun next_rxd = r ;
1495*4882a593Smuzhiyun
1496*4882a593Smuzhiyun for (r=queue->rx_curr_get,i=frag_count; i ; r=r->rxd_next,i--){
1497*4882a593Smuzhiyun DB_RX(5, "dma_complete for RxD %p", r);
1498*4882a593Smuzhiyun dma_complete(smc,(union s_fp_descr volatile *)r,DMA_WR);
1499*4882a593Smuzhiyun }
1500*4882a593Smuzhiyun
1501*4882a593Smuzhiyun DB_RX(5, "mac_drv_clear_rxd: RxD %p frag_count %d",
1502*4882a593Smuzhiyun queue->rx_curr_get, frag_count);
1503*4882a593Smuzhiyun mac_drv_clear_rxd(smc,queue->rx_curr_get,frag_count) ;
1504*4882a593Smuzhiyun
1505*4882a593Smuzhiyun queue->rx_curr_get = next_rxd ;
1506*4882a593Smuzhiyun queue->rx_used -= frag_count ;
1507*4882a593Smuzhiyun queue->rx_free += frag_count ;
1508*4882a593Smuzhiyun }
1509*4882a593Smuzhiyun }
1510*4882a593Smuzhiyun
1511*4882a593Smuzhiyun
1512*4882a593Smuzhiyun /*
1513*4882a593Smuzhiyun -------------------------------------------------------------
1514*4882a593Smuzhiyun SEND FUNCTIONS:
1515*4882a593Smuzhiyun -------------------------------------------------------------
1516*4882a593Smuzhiyun */
1517*4882a593Smuzhiyun
1518*4882a593Smuzhiyun /*
1519*4882a593Smuzhiyun * BEGIN_MANUAL_ENTRY(hwm_tx_init)
1520*4882a593Smuzhiyun * int hwm_tx_init(smc,fc,frag_count,frame_len,frame_status)
1521*4882a593Smuzhiyun *
1522*4882a593Smuzhiyun * function DOWN_CALL (hardware module, hwmtm.c)
1523*4882a593Smuzhiyun * hwm_tx_init checks if the frame can be sent through the
1524*4882a593Smuzhiyun * corresponding send queue.
1525*4882a593Smuzhiyun *
1526*4882a593Smuzhiyun * para fc the frame control. To determine through which
1527*4882a593Smuzhiyun * send queue the frame should be transmitted.
1528*4882a593Smuzhiyun * 0x50 - 0x57: asynchronous LLC frame
1529*4882a593Smuzhiyun * 0xD0 - 0xD7: synchronous LLC frame
1530*4882a593Smuzhiyun * 0x41, 0x4F: SMT frame to the network
1531*4882a593Smuzhiyun * 0x42: SMT frame to the network and to the local SMT
1532*4882a593Smuzhiyun * 0x43: SMT frame to the local SMT
1533*4882a593Smuzhiyun * frag_count count of the fragments for this frame
1534*4882a593Smuzhiyun * frame_len length of the frame
1535*4882a593Smuzhiyun * frame_status status of the frame, the send queue bit is already
1536*4882a593Smuzhiyun * specified
1537*4882a593Smuzhiyun *
1538*4882a593Smuzhiyun * return frame_status
1539*4882a593Smuzhiyun *
1540*4882a593Smuzhiyun * END_MANUAL_ENTRY
1541*4882a593Smuzhiyun */
hwm_tx_init(struct s_smc * smc,u_char fc,int frag_count,int frame_len,int frame_status)1542*4882a593Smuzhiyun int hwm_tx_init(struct s_smc *smc, u_char fc, int frag_count, int frame_len,
1543*4882a593Smuzhiyun int frame_status)
1544*4882a593Smuzhiyun {
1545*4882a593Smuzhiyun NDD_TRACE("THiB",fc,frag_count,frame_len) ;
1546*4882a593Smuzhiyun smc->os.hwm.tx_p = smc->hw.fp.tx[frame_status & QUEUE_A0] ;
1547*4882a593Smuzhiyun smc->os.hwm.tx_descr = TX_DESCRIPTOR | (((u_long)(frame_len-1)&3)<<27) ;
1548*4882a593Smuzhiyun smc->os.hwm.tx_len = frame_len ;
1549*4882a593Smuzhiyun DB_TX(3, "hwm_tx_init: fc = %x, len = %d", fc, frame_len);
1550*4882a593Smuzhiyun if ((fc & ~(FC_SYNC_BIT|FC_LLC_PRIOR)) == FC_ASYNC_LLC) {
1551*4882a593Smuzhiyun frame_status |= LAN_TX ;
1552*4882a593Smuzhiyun }
1553*4882a593Smuzhiyun else {
1554*4882a593Smuzhiyun switch (fc) {
1555*4882a593Smuzhiyun case FC_SMT_INFO :
1556*4882a593Smuzhiyun case FC_SMT_NSA :
1557*4882a593Smuzhiyun frame_status |= LAN_TX ;
1558*4882a593Smuzhiyun break ;
1559*4882a593Smuzhiyun case FC_SMT_LOC :
1560*4882a593Smuzhiyun frame_status |= LOC_TX ;
1561*4882a593Smuzhiyun break ;
1562*4882a593Smuzhiyun case FC_SMT_LAN_LOC :
1563*4882a593Smuzhiyun frame_status |= LAN_TX | LOC_TX ;
1564*4882a593Smuzhiyun break ;
1565*4882a593Smuzhiyun default :
1566*4882a593Smuzhiyun SMT_PANIC(smc,HWM_E0010,HWM_E0010_MSG) ;
1567*4882a593Smuzhiyun }
1568*4882a593Smuzhiyun }
1569*4882a593Smuzhiyun if (!smc->hw.mac_ring_is_up) {
1570*4882a593Smuzhiyun frame_status &= ~LAN_TX ;
1571*4882a593Smuzhiyun frame_status |= RING_DOWN ;
1572*4882a593Smuzhiyun DB_TX(2, "Ring is down: terminate LAN_TX");
1573*4882a593Smuzhiyun }
1574*4882a593Smuzhiyun if (frag_count > smc->os.hwm.tx_p->tx_free) {
1575*4882a593Smuzhiyun #ifndef NDIS_OS2
1576*4882a593Smuzhiyun mac_drv_clear_txd(smc) ;
1577*4882a593Smuzhiyun if (frag_count > smc->os.hwm.tx_p->tx_free) {
1578*4882a593Smuzhiyun DB_TX(2, "Out of TxDs, terminate LAN_TX");
1579*4882a593Smuzhiyun frame_status &= ~LAN_TX ;
1580*4882a593Smuzhiyun frame_status |= OUT_OF_TXD ;
1581*4882a593Smuzhiyun }
1582*4882a593Smuzhiyun #else
1583*4882a593Smuzhiyun DB_TX(2, "Out of TxDs, terminate LAN_TX");
1584*4882a593Smuzhiyun frame_status &= ~LAN_TX ;
1585*4882a593Smuzhiyun frame_status |= OUT_OF_TXD ;
1586*4882a593Smuzhiyun #endif
1587*4882a593Smuzhiyun }
1588*4882a593Smuzhiyun DB_TX(3, "frame_status = %x", frame_status);
1589*4882a593Smuzhiyun NDD_TRACE("THiE",frame_status,smc->os.hwm.tx_p->tx_free,0) ;
1590*4882a593Smuzhiyun return frame_status;
1591*4882a593Smuzhiyun }
1592*4882a593Smuzhiyun
1593*4882a593Smuzhiyun /*
1594*4882a593Smuzhiyun * BEGIN_MANUAL_ENTRY(hwm_tx_frag)
1595*4882a593Smuzhiyun * void hwm_tx_frag(smc,virt,phys,len,frame_status)
1596*4882a593Smuzhiyun *
1597*4882a593Smuzhiyun * function DOWNCALL (hardware module, hwmtm.c)
1598*4882a593Smuzhiyun * If the frame should be sent to the LAN, this function calls
1599*4882a593Smuzhiyun * dma_master, fills the current TxD with the virtual and the
1600*4882a593Smuzhiyun * physical address, sets the STF and EOF bits dependent on
1601*4882a593Smuzhiyun * the frame status, and requests the BMU to start the
1602*4882a593Smuzhiyun * transmit.
1603*4882a593Smuzhiyun * If the frame should be sent to the local SMT, an SMT_MBuf
1604*4882a593Smuzhiyun * is allocated if the FIRST_FRAG bit is set in the frame_status.
1605*4882a593Smuzhiyun * The fragment of the frame is copied into the SMT MBuf.
1606*4882a593Smuzhiyun * The function smt_received_pack is called if the LAST_FRAG
1607*4882a593Smuzhiyun * bit is set in the frame_status word.
1608*4882a593Smuzhiyun *
1609*4882a593Smuzhiyun * para virt virtual pointer to the fragment
1610*4882a593Smuzhiyun * len the length of the fragment
1611*4882a593Smuzhiyun * frame_status status of the frame, see design description
1612*4882a593Smuzhiyun *
1613*4882a593Smuzhiyun * return nothing returned, no parameter is modified
1614*4882a593Smuzhiyun *
1615*4882a593Smuzhiyun * NOTE: It is possible to invoke this macro with a fragment length
1616*4882a593Smuzhiyun * of zero.
1617*4882a593Smuzhiyun *
1618*4882a593Smuzhiyun * END_MANUAL_ENTRY
1619*4882a593Smuzhiyun */
hwm_tx_frag(struct s_smc * smc,char far * virt,u_long phys,int len,int frame_status)1620*4882a593Smuzhiyun void hwm_tx_frag(struct s_smc *smc, char far *virt, u_long phys, int len,
1621*4882a593Smuzhiyun int frame_status)
1622*4882a593Smuzhiyun {
1623*4882a593Smuzhiyun struct s_smt_fp_txd volatile *t ;
1624*4882a593Smuzhiyun struct s_smt_tx_queue *queue ;
1625*4882a593Smuzhiyun __le32 tbctrl ;
1626*4882a593Smuzhiyun
1627*4882a593Smuzhiyun queue = smc->os.hwm.tx_p ;
1628*4882a593Smuzhiyun
1629*4882a593Smuzhiyun NDD_TRACE("THfB",virt,len,frame_status) ;
1630*4882a593Smuzhiyun /* Bug fix: AF / May 31 1999 (#missing)
1631*4882a593Smuzhiyun * snmpinfo problem reported by IBM is caused by invalid
1632*4882a593Smuzhiyun * t-pointer (txd) if LAN_TX is not set but LOC_TX only.
1633*4882a593Smuzhiyun * Set: t = queue->tx_curr_put here !
1634*4882a593Smuzhiyun */
1635*4882a593Smuzhiyun t = queue->tx_curr_put ;
1636*4882a593Smuzhiyun
1637*4882a593Smuzhiyun DB_TX(2, "hwm_tx_frag: len = %d, frame_status = %x", len, frame_status);
1638*4882a593Smuzhiyun if (frame_status & LAN_TX) {
1639*4882a593Smuzhiyun /* '*t' is already defined */
1640*4882a593Smuzhiyun DB_TX(3, "LAN_TX: TxD = %p, virt = %p", t, virt);
1641*4882a593Smuzhiyun t->txd_virt = virt ;
1642*4882a593Smuzhiyun t->txd_txdscr = cpu_to_le32(smc->os.hwm.tx_descr) ;
1643*4882a593Smuzhiyun t->txd_tbadr = cpu_to_le32(phys) ;
1644*4882a593Smuzhiyun tbctrl = cpu_to_le32((((__u32)frame_status &
1645*4882a593Smuzhiyun (FIRST_FRAG|LAST_FRAG|EN_IRQ_EOF))<< 26) |
1646*4882a593Smuzhiyun BMU_OWN|BMU_CHECK |len) ;
1647*4882a593Smuzhiyun t->txd_tbctrl = tbctrl ;
1648*4882a593Smuzhiyun
1649*4882a593Smuzhiyun #ifndef AIX
1650*4882a593Smuzhiyun DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORDEV) ;
1651*4882a593Smuzhiyun outpd(queue->tx_bmu_ctl,CSR_START) ;
1652*4882a593Smuzhiyun #else /* ifndef AIX */
1653*4882a593Smuzhiyun DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORDEV) ;
1654*4882a593Smuzhiyun if (frame_status & QUEUE_A0) {
1655*4882a593Smuzhiyun outpd(ADDR(B0_XA_CSR),CSR_START) ;
1656*4882a593Smuzhiyun }
1657*4882a593Smuzhiyun else {
1658*4882a593Smuzhiyun outpd(ADDR(B0_XS_CSR),CSR_START) ;
1659*4882a593Smuzhiyun }
1660*4882a593Smuzhiyun #endif
1661*4882a593Smuzhiyun queue->tx_free-- ;
1662*4882a593Smuzhiyun queue->tx_used++ ;
1663*4882a593Smuzhiyun queue->tx_curr_put = t->txd_next ;
1664*4882a593Smuzhiyun if (frame_status & LAST_FRAG) {
1665*4882a593Smuzhiyun smc->mib.m[MAC0].fddiMACTransmit_Ct++ ;
1666*4882a593Smuzhiyun }
1667*4882a593Smuzhiyun }
1668*4882a593Smuzhiyun if (frame_status & LOC_TX) {
1669*4882a593Smuzhiyun DB_TX(3, "LOC_TX:");
1670*4882a593Smuzhiyun if (frame_status & FIRST_FRAG) {
1671*4882a593Smuzhiyun if(!(smc->os.hwm.tx_mb = smt_get_mbuf(smc))) {
1672*4882a593Smuzhiyun smc->hw.fp.err_stats.err_no_buf++ ;
1673*4882a593Smuzhiyun DB_TX(4, "No SMbuf; transmit terminated");
1674*4882a593Smuzhiyun }
1675*4882a593Smuzhiyun else {
1676*4882a593Smuzhiyun smc->os.hwm.tx_data =
1677*4882a593Smuzhiyun smtod(smc->os.hwm.tx_mb,char *) - 1 ;
1678*4882a593Smuzhiyun #ifdef USE_OS_CPY
1679*4882a593Smuzhiyun #ifdef PASS_1ST_TXD_2_TX_COMP
1680*4882a593Smuzhiyun hwm_cpy_txd2mb(t,smc->os.hwm.tx_data,
1681*4882a593Smuzhiyun smc->os.hwm.tx_len) ;
1682*4882a593Smuzhiyun #endif
1683*4882a593Smuzhiyun #endif
1684*4882a593Smuzhiyun }
1685*4882a593Smuzhiyun }
1686*4882a593Smuzhiyun if (smc->os.hwm.tx_mb) {
1687*4882a593Smuzhiyun #ifndef USE_OS_CPY
1688*4882a593Smuzhiyun DB_TX(3, "copy fragment into MBuf");
1689*4882a593Smuzhiyun memcpy(smc->os.hwm.tx_data,virt,len) ;
1690*4882a593Smuzhiyun smc->os.hwm.tx_data += len ;
1691*4882a593Smuzhiyun #endif
1692*4882a593Smuzhiyun if (frame_status & LAST_FRAG) {
1693*4882a593Smuzhiyun #ifdef USE_OS_CPY
1694*4882a593Smuzhiyun #ifndef PASS_1ST_TXD_2_TX_COMP
1695*4882a593Smuzhiyun /*
1696*4882a593Smuzhiyun * hwm_cpy_txd2mb(txd,data,len) copies 'len'
1697*4882a593Smuzhiyun * bytes from the virtual pointer in 'rxd'
1698*4882a593Smuzhiyun * to 'data'. The virtual pointer of the
1699*4882a593Smuzhiyun * os-specific tx-buffer should be written
1700*4882a593Smuzhiyun * in the LAST txd.
1701*4882a593Smuzhiyun */
1702*4882a593Smuzhiyun hwm_cpy_txd2mb(t,smc->os.hwm.tx_data,
1703*4882a593Smuzhiyun smc->os.hwm.tx_len) ;
1704*4882a593Smuzhiyun #endif /* nPASS_1ST_TXD_2_TX_COMP */
1705*4882a593Smuzhiyun #endif /* USE_OS_CPY */
1706*4882a593Smuzhiyun smc->os.hwm.tx_data =
1707*4882a593Smuzhiyun smtod(smc->os.hwm.tx_mb,char *) - 1 ;
1708*4882a593Smuzhiyun *(char *)smc->os.hwm.tx_mb->sm_data =
1709*4882a593Smuzhiyun *smc->os.hwm.tx_data ;
1710*4882a593Smuzhiyun smc->os.hwm.tx_data++ ;
1711*4882a593Smuzhiyun smc->os.hwm.tx_mb->sm_len =
1712*4882a593Smuzhiyun smc->os.hwm.tx_len - 1 ;
1713*4882a593Smuzhiyun DB_TX(3, "pass LLC frame to SMT");
1714*4882a593Smuzhiyun smt_received_pack(smc,smc->os.hwm.tx_mb,
1715*4882a593Smuzhiyun RD_FS_LOCAL) ;
1716*4882a593Smuzhiyun }
1717*4882a593Smuzhiyun }
1718*4882a593Smuzhiyun }
1719*4882a593Smuzhiyun NDD_TRACE("THfE",t,queue->tx_free,0) ;
1720*4882a593Smuzhiyun }
1721*4882a593Smuzhiyun
1722*4882a593Smuzhiyun
1723*4882a593Smuzhiyun /*
1724*4882a593Smuzhiyun * queues a receive for later send
1725*4882a593Smuzhiyun */
queue_llc_rx(struct s_smc * smc,SMbuf * mb)1726*4882a593Smuzhiyun static void queue_llc_rx(struct s_smc *smc, SMbuf *mb)
1727*4882a593Smuzhiyun {
1728*4882a593Smuzhiyun DB_GEN(4, "queue_llc_rx: mb = %p", mb);
1729*4882a593Smuzhiyun smc->os.hwm.queued_rx_frames++ ;
1730*4882a593Smuzhiyun mb->sm_next = (SMbuf *)NULL ;
1731*4882a593Smuzhiyun if (smc->os.hwm.llc_rx_pipe == NULL) {
1732*4882a593Smuzhiyun smc->os.hwm.llc_rx_pipe = mb ;
1733*4882a593Smuzhiyun }
1734*4882a593Smuzhiyun else {
1735*4882a593Smuzhiyun smc->os.hwm.llc_rx_tail->sm_next = mb ;
1736*4882a593Smuzhiyun }
1737*4882a593Smuzhiyun smc->os.hwm.llc_rx_tail = mb ;
1738*4882a593Smuzhiyun
1739*4882a593Smuzhiyun /*
1740*4882a593Smuzhiyun * force an timer IRQ to receive the data
1741*4882a593Smuzhiyun */
1742*4882a593Smuzhiyun if (!smc->os.hwm.isr_flag) {
1743*4882a593Smuzhiyun smt_force_irq(smc) ;
1744*4882a593Smuzhiyun }
1745*4882a593Smuzhiyun }
1746*4882a593Smuzhiyun
1747*4882a593Smuzhiyun /*
1748*4882a593Smuzhiyun * get a SMbuf from the llc_rx_queue
1749*4882a593Smuzhiyun */
get_llc_rx(struct s_smc * smc)1750*4882a593Smuzhiyun static SMbuf *get_llc_rx(struct s_smc *smc)
1751*4882a593Smuzhiyun {
1752*4882a593Smuzhiyun SMbuf *mb ;
1753*4882a593Smuzhiyun
1754*4882a593Smuzhiyun if ((mb = smc->os.hwm.llc_rx_pipe)) {
1755*4882a593Smuzhiyun smc->os.hwm.queued_rx_frames-- ;
1756*4882a593Smuzhiyun smc->os.hwm.llc_rx_pipe = mb->sm_next ;
1757*4882a593Smuzhiyun }
1758*4882a593Smuzhiyun DB_GEN(4, "get_llc_rx: mb = 0x%p", mb);
1759*4882a593Smuzhiyun return mb;
1760*4882a593Smuzhiyun }
1761*4882a593Smuzhiyun
1762*4882a593Smuzhiyun /*
1763*4882a593Smuzhiyun * queues a transmit SMT MBuf during the time were the MBuf is
1764*4882a593Smuzhiyun * queued the TxD ring
1765*4882a593Smuzhiyun */
queue_txd_mb(struct s_smc * smc,SMbuf * mb)1766*4882a593Smuzhiyun static void queue_txd_mb(struct s_smc *smc, SMbuf *mb)
1767*4882a593Smuzhiyun {
1768*4882a593Smuzhiyun DB_GEN(4, "_rx: queue_txd_mb = %p", mb);
1769*4882a593Smuzhiyun smc->os.hwm.queued_txd_mb++ ;
1770*4882a593Smuzhiyun mb->sm_next = (SMbuf *)NULL ;
1771*4882a593Smuzhiyun if (smc->os.hwm.txd_tx_pipe == NULL) {
1772*4882a593Smuzhiyun smc->os.hwm.txd_tx_pipe = mb ;
1773*4882a593Smuzhiyun }
1774*4882a593Smuzhiyun else {
1775*4882a593Smuzhiyun smc->os.hwm.txd_tx_tail->sm_next = mb ;
1776*4882a593Smuzhiyun }
1777*4882a593Smuzhiyun smc->os.hwm.txd_tx_tail = mb ;
1778*4882a593Smuzhiyun }
1779*4882a593Smuzhiyun
1780*4882a593Smuzhiyun /*
1781*4882a593Smuzhiyun * get a SMbuf from the txd_tx_queue
1782*4882a593Smuzhiyun */
get_txd_mb(struct s_smc * smc)1783*4882a593Smuzhiyun static SMbuf *get_txd_mb(struct s_smc *smc)
1784*4882a593Smuzhiyun {
1785*4882a593Smuzhiyun SMbuf *mb ;
1786*4882a593Smuzhiyun
1787*4882a593Smuzhiyun if ((mb = smc->os.hwm.txd_tx_pipe)) {
1788*4882a593Smuzhiyun smc->os.hwm.queued_txd_mb-- ;
1789*4882a593Smuzhiyun smc->os.hwm.txd_tx_pipe = mb->sm_next ;
1790*4882a593Smuzhiyun }
1791*4882a593Smuzhiyun DB_GEN(4, "get_txd_mb: mb = 0x%p", mb);
1792*4882a593Smuzhiyun return mb;
1793*4882a593Smuzhiyun }
1794*4882a593Smuzhiyun
1795*4882a593Smuzhiyun /*
1796*4882a593Smuzhiyun * SMT Send function
1797*4882a593Smuzhiyun */
smt_send_mbuf(struct s_smc * smc,SMbuf * mb,int fc)1798*4882a593Smuzhiyun void smt_send_mbuf(struct s_smc *smc, SMbuf *mb, int fc)
1799*4882a593Smuzhiyun {
1800*4882a593Smuzhiyun char far *data ;
1801*4882a593Smuzhiyun int len ;
1802*4882a593Smuzhiyun int n ;
1803*4882a593Smuzhiyun int i ;
1804*4882a593Smuzhiyun int frag_count ;
1805*4882a593Smuzhiyun int frame_status ;
1806*4882a593Smuzhiyun SK_LOC_DECL(char far,*virt[3]) ;
1807*4882a593Smuzhiyun int frag_len[3] ;
1808*4882a593Smuzhiyun struct s_smt_tx_queue *queue ;
1809*4882a593Smuzhiyun struct s_smt_fp_txd volatile *t ;
1810*4882a593Smuzhiyun u_long phys ;
1811*4882a593Smuzhiyun __le32 tbctrl;
1812*4882a593Smuzhiyun
1813*4882a593Smuzhiyun NDD_TRACE("THSB",mb,fc,0) ;
1814*4882a593Smuzhiyun DB_TX(4, "smt_send_mbuf: mb = 0x%p, fc = 0x%x", mb, fc);
1815*4882a593Smuzhiyun
1816*4882a593Smuzhiyun mb->sm_off-- ; /* set to fc */
1817*4882a593Smuzhiyun mb->sm_len++ ; /* + fc */
1818*4882a593Smuzhiyun data = smtod(mb,char *) ;
1819*4882a593Smuzhiyun *data = fc ;
1820*4882a593Smuzhiyun if (fc == FC_SMT_LOC)
1821*4882a593Smuzhiyun *data = FC_SMT_INFO ;
1822*4882a593Smuzhiyun
1823*4882a593Smuzhiyun /*
1824*4882a593Smuzhiyun * determine the frag count and the virt addresses of the frags
1825*4882a593Smuzhiyun */
1826*4882a593Smuzhiyun frag_count = 0 ;
1827*4882a593Smuzhiyun len = mb->sm_len ;
1828*4882a593Smuzhiyun while (len) {
1829*4882a593Smuzhiyun n = SMT_PAGESIZE - ((long)data & (SMT_PAGESIZE-1)) ;
1830*4882a593Smuzhiyun if (n >= len) {
1831*4882a593Smuzhiyun n = len ;
1832*4882a593Smuzhiyun }
1833*4882a593Smuzhiyun DB_TX(5, "frag: virt/len = 0x%p/%d", data, n);
1834*4882a593Smuzhiyun virt[frag_count] = data ;
1835*4882a593Smuzhiyun frag_len[frag_count] = n ;
1836*4882a593Smuzhiyun frag_count++ ;
1837*4882a593Smuzhiyun len -= n ;
1838*4882a593Smuzhiyun data += n ;
1839*4882a593Smuzhiyun }
1840*4882a593Smuzhiyun
1841*4882a593Smuzhiyun /*
1842*4882a593Smuzhiyun * determine the frame status
1843*4882a593Smuzhiyun */
1844*4882a593Smuzhiyun queue = smc->hw.fp.tx[QUEUE_A0] ;
1845*4882a593Smuzhiyun if (fc == FC_BEACON || fc == FC_SMT_LOC) {
1846*4882a593Smuzhiyun frame_status = LOC_TX ;
1847*4882a593Smuzhiyun }
1848*4882a593Smuzhiyun else {
1849*4882a593Smuzhiyun frame_status = LAN_TX ;
1850*4882a593Smuzhiyun if ((smc->os.hwm.pass_NSA &&(fc == FC_SMT_NSA)) ||
1851*4882a593Smuzhiyun (smc->os.hwm.pass_SMT &&(fc == FC_SMT_INFO)))
1852*4882a593Smuzhiyun frame_status |= LOC_TX ;
1853*4882a593Smuzhiyun }
1854*4882a593Smuzhiyun
1855*4882a593Smuzhiyun if (!smc->hw.mac_ring_is_up || frag_count > queue->tx_free) {
1856*4882a593Smuzhiyun frame_status &= ~LAN_TX;
1857*4882a593Smuzhiyun if (frame_status) {
1858*4882a593Smuzhiyun DB_TX(2, "Ring is down: terminate LAN_TX");
1859*4882a593Smuzhiyun }
1860*4882a593Smuzhiyun else {
1861*4882a593Smuzhiyun DB_TX(2, "Ring is down: terminate transmission");
1862*4882a593Smuzhiyun smt_free_mbuf(smc,mb) ;
1863*4882a593Smuzhiyun return ;
1864*4882a593Smuzhiyun }
1865*4882a593Smuzhiyun }
1866*4882a593Smuzhiyun DB_TX(5, "frame_status = 0x%x", frame_status);
1867*4882a593Smuzhiyun
1868*4882a593Smuzhiyun if ((frame_status & LAN_TX) && (frame_status & LOC_TX)) {
1869*4882a593Smuzhiyun mb->sm_use_count = 2 ;
1870*4882a593Smuzhiyun }
1871*4882a593Smuzhiyun
1872*4882a593Smuzhiyun if (frame_status & LAN_TX) {
1873*4882a593Smuzhiyun t = queue->tx_curr_put ;
1874*4882a593Smuzhiyun frame_status |= FIRST_FRAG ;
1875*4882a593Smuzhiyun for (i = 0; i < frag_count; i++) {
1876*4882a593Smuzhiyun DB_TX(5, "init TxD = 0x%p", t);
1877*4882a593Smuzhiyun if (i == frag_count-1) {
1878*4882a593Smuzhiyun frame_status |= LAST_FRAG ;
1879*4882a593Smuzhiyun t->txd_txdscr = cpu_to_le32(TX_DESCRIPTOR |
1880*4882a593Smuzhiyun (((__u32)(mb->sm_len-1)&3) << 27)) ;
1881*4882a593Smuzhiyun }
1882*4882a593Smuzhiyun t->txd_virt = virt[i] ;
1883*4882a593Smuzhiyun phys = dma_master(smc, (void far *)virt[i],
1884*4882a593Smuzhiyun frag_len[i], DMA_RD|SMT_BUF) ;
1885*4882a593Smuzhiyun t->txd_tbadr = cpu_to_le32(phys) ;
1886*4882a593Smuzhiyun tbctrl = cpu_to_le32((((__u32)frame_status &
1887*4882a593Smuzhiyun (FIRST_FRAG|LAST_FRAG)) << 26) |
1888*4882a593Smuzhiyun BMU_OWN | BMU_CHECK | BMU_SMT_TX |frag_len[i]) ;
1889*4882a593Smuzhiyun t->txd_tbctrl = tbctrl ;
1890*4882a593Smuzhiyun #ifndef AIX
1891*4882a593Smuzhiyun DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORDEV) ;
1892*4882a593Smuzhiyun outpd(queue->tx_bmu_ctl,CSR_START) ;
1893*4882a593Smuzhiyun #else
1894*4882a593Smuzhiyun DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORDEV) ;
1895*4882a593Smuzhiyun outpd(ADDR(B0_XA_CSR),CSR_START) ;
1896*4882a593Smuzhiyun #endif
1897*4882a593Smuzhiyun frame_status &= ~FIRST_FRAG ;
1898*4882a593Smuzhiyun queue->tx_curr_put = t = t->txd_next ;
1899*4882a593Smuzhiyun queue->tx_free-- ;
1900*4882a593Smuzhiyun queue->tx_used++ ;
1901*4882a593Smuzhiyun }
1902*4882a593Smuzhiyun smc->mib.m[MAC0].fddiMACTransmit_Ct++ ;
1903*4882a593Smuzhiyun queue_txd_mb(smc,mb) ;
1904*4882a593Smuzhiyun }
1905*4882a593Smuzhiyun
1906*4882a593Smuzhiyun if (frame_status & LOC_TX) {
1907*4882a593Smuzhiyun DB_TX(5, "pass Mbuf to LLC queue");
1908*4882a593Smuzhiyun queue_llc_rx(smc,mb) ;
1909*4882a593Smuzhiyun }
1910*4882a593Smuzhiyun
1911*4882a593Smuzhiyun /*
1912*4882a593Smuzhiyun * We need to unqueue the free SMT_MBUFs here, because it may
1913*4882a593Smuzhiyun * be that the SMT want's to send more than 1 frame for one down call
1914*4882a593Smuzhiyun */
1915*4882a593Smuzhiyun mac_drv_clear_txd(smc) ;
1916*4882a593Smuzhiyun NDD_TRACE("THSE",t,queue->tx_free,frag_count) ;
1917*4882a593Smuzhiyun }
1918*4882a593Smuzhiyun
1919*4882a593Smuzhiyun /* BEGIN_MANUAL_ENTRY(mac_drv_clear_txd)
1920*4882a593Smuzhiyun * void mac_drv_clear_txd(smc)
1921*4882a593Smuzhiyun *
1922*4882a593Smuzhiyun * function DOWNCALL (hardware module, hwmtm.c)
1923*4882a593Smuzhiyun * mac_drv_clear_txd searches in both send queues for TxD's
1924*4882a593Smuzhiyun * which were finished by the adapter. It calls dma_complete
1925*4882a593Smuzhiyun * for each TxD. If the last fragment of an LLC frame is
1926*4882a593Smuzhiyun * reached, it calls mac_drv_tx_complete to release the
1927*4882a593Smuzhiyun * send buffer.
1928*4882a593Smuzhiyun *
1929*4882a593Smuzhiyun * return nothing
1930*4882a593Smuzhiyun *
1931*4882a593Smuzhiyun * END_MANUAL_ENTRY
1932*4882a593Smuzhiyun */
mac_drv_clear_txd(struct s_smc * smc)1933*4882a593Smuzhiyun static void mac_drv_clear_txd(struct s_smc *smc)
1934*4882a593Smuzhiyun {
1935*4882a593Smuzhiyun struct s_smt_tx_queue *queue ;
1936*4882a593Smuzhiyun struct s_smt_fp_txd volatile *t1 ;
1937*4882a593Smuzhiyun struct s_smt_fp_txd volatile *t2 = NULL ;
1938*4882a593Smuzhiyun SMbuf *mb ;
1939*4882a593Smuzhiyun u_long tbctrl ;
1940*4882a593Smuzhiyun int i ;
1941*4882a593Smuzhiyun int frag_count ;
1942*4882a593Smuzhiyun int n ;
1943*4882a593Smuzhiyun
1944*4882a593Smuzhiyun NDD_TRACE("THcB",0,0,0) ;
1945*4882a593Smuzhiyun for (i = QUEUE_S; i <= QUEUE_A0; i++) {
1946*4882a593Smuzhiyun queue = smc->hw.fp.tx[i] ;
1947*4882a593Smuzhiyun t1 = queue->tx_curr_get ;
1948*4882a593Smuzhiyun DB_TX(5, "clear_txd: QUEUE = %d (0=sync/1=async)", i);
1949*4882a593Smuzhiyun
1950*4882a593Smuzhiyun for ( ; ; ) {
1951*4882a593Smuzhiyun frag_count = 0 ;
1952*4882a593Smuzhiyun
1953*4882a593Smuzhiyun do {
1954*4882a593Smuzhiyun DRV_BUF_FLUSH(t1,DDI_DMA_SYNC_FORCPU) ;
1955*4882a593Smuzhiyun DB_TX(5, "check OWN/EOF bit of TxD 0x%p", t1);
1956*4882a593Smuzhiyun tbctrl = le32_to_cpu(CR_READ(t1->txd_tbctrl));
1957*4882a593Smuzhiyun
1958*4882a593Smuzhiyun if (tbctrl & BMU_OWN || !queue->tx_used){
1959*4882a593Smuzhiyun DB_TX(4, "End of TxDs queue %d", i);
1960*4882a593Smuzhiyun goto free_next_queue ; /* next queue */
1961*4882a593Smuzhiyun }
1962*4882a593Smuzhiyun t1 = t1->txd_next ;
1963*4882a593Smuzhiyun frag_count++ ;
1964*4882a593Smuzhiyun } while (!(tbctrl & BMU_EOF)) ;
1965*4882a593Smuzhiyun
1966*4882a593Smuzhiyun t1 = queue->tx_curr_get ;
1967*4882a593Smuzhiyun for (n = frag_count; n; n--) {
1968*4882a593Smuzhiyun tbctrl = le32_to_cpu(t1->txd_tbctrl) ;
1969*4882a593Smuzhiyun dma_complete(smc,
1970*4882a593Smuzhiyun (union s_fp_descr volatile *) t1,
1971*4882a593Smuzhiyun (int) (DMA_RD |
1972*4882a593Smuzhiyun ((tbctrl & BMU_SMT_TX) >> 18))) ;
1973*4882a593Smuzhiyun t2 = t1 ;
1974*4882a593Smuzhiyun t1 = t1->txd_next ;
1975*4882a593Smuzhiyun }
1976*4882a593Smuzhiyun
1977*4882a593Smuzhiyun if (tbctrl & BMU_SMT_TX) {
1978*4882a593Smuzhiyun mb = get_txd_mb(smc) ;
1979*4882a593Smuzhiyun smt_free_mbuf(smc,mb) ;
1980*4882a593Smuzhiyun }
1981*4882a593Smuzhiyun else {
1982*4882a593Smuzhiyun #ifndef PASS_1ST_TXD_2_TX_COMP
1983*4882a593Smuzhiyun DB_TX(4, "mac_drv_tx_comp for TxD 0x%p", t2);
1984*4882a593Smuzhiyun mac_drv_tx_complete(smc,t2) ;
1985*4882a593Smuzhiyun #else
1986*4882a593Smuzhiyun DB_TX(4, "mac_drv_tx_comp for TxD 0x%x",
1987*4882a593Smuzhiyun queue->tx_curr_get);
1988*4882a593Smuzhiyun mac_drv_tx_complete(smc,queue->tx_curr_get) ;
1989*4882a593Smuzhiyun #endif
1990*4882a593Smuzhiyun }
1991*4882a593Smuzhiyun queue->tx_curr_get = t1 ;
1992*4882a593Smuzhiyun queue->tx_free += frag_count ;
1993*4882a593Smuzhiyun queue->tx_used -= frag_count ;
1994*4882a593Smuzhiyun }
1995*4882a593Smuzhiyun free_next_queue: ;
1996*4882a593Smuzhiyun }
1997*4882a593Smuzhiyun NDD_TRACE("THcE",0,0,0) ;
1998*4882a593Smuzhiyun }
1999*4882a593Smuzhiyun
2000*4882a593Smuzhiyun /*
2001*4882a593Smuzhiyun * BEGINN_MANUAL_ENTRY(mac_drv_clear_tx_queue)
2002*4882a593Smuzhiyun *
2003*4882a593Smuzhiyun * void mac_drv_clear_tx_queue(smc)
2004*4882a593Smuzhiyun * struct s_smc *smc ;
2005*4882a593Smuzhiyun *
2006*4882a593Smuzhiyun * function DOWNCALL (hardware module, hwmtm.c)
2007*4882a593Smuzhiyun * mac_drv_clear_tx_queue is called from the SMT when
2008*4882a593Smuzhiyun * the RMT state machine has entered the ISOLATE state.
2009*4882a593Smuzhiyun * This function is also called by the os-specific module
2010*4882a593Smuzhiyun * after it has called the function card_stop().
2011*4882a593Smuzhiyun * In this case, the frames in the send queues are obsolete and
2012*4882a593Smuzhiyun * should be removed.
2013*4882a593Smuzhiyun *
2014*4882a593Smuzhiyun * note calling sequence:
2015*4882a593Smuzhiyun * CLI_FBI(), card_stop(),
2016*4882a593Smuzhiyun * mac_drv_clear_tx_queue(), mac_drv_clear_rx_queue(),
2017*4882a593Smuzhiyun *
2018*4882a593Smuzhiyun * NOTE: The caller is responsible that the BMUs are idle
2019*4882a593Smuzhiyun * when this function is called.
2020*4882a593Smuzhiyun *
2021*4882a593Smuzhiyun * END_MANUAL_ENTRY
2022*4882a593Smuzhiyun */
mac_drv_clear_tx_queue(struct s_smc * smc)2023*4882a593Smuzhiyun void mac_drv_clear_tx_queue(struct s_smc *smc)
2024*4882a593Smuzhiyun {
2025*4882a593Smuzhiyun struct s_smt_fp_txd volatile *t ;
2026*4882a593Smuzhiyun struct s_smt_tx_queue *queue ;
2027*4882a593Smuzhiyun int tx_used ;
2028*4882a593Smuzhiyun int i ;
2029*4882a593Smuzhiyun
2030*4882a593Smuzhiyun if (smc->hw.hw_state != STOPPED) {
2031*4882a593Smuzhiyun SK_BREAK() ;
2032*4882a593Smuzhiyun SMT_PANIC(smc,HWM_E0011,HWM_E0011_MSG) ;
2033*4882a593Smuzhiyun return ;
2034*4882a593Smuzhiyun }
2035*4882a593Smuzhiyun
2036*4882a593Smuzhiyun for (i = QUEUE_S; i <= QUEUE_A0; i++) {
2037*4882a593Smuzhiyun queue = smc->hw.fp.tx[i] ;
2038*4882a593Smuzhiyun DB_TX(5, "clear_tx_queue: QUEUE = %d (0=sync/1=async)", i);
2039*4882a593Smuzhiyun
2040*4882a593Smuzhiyun /*
2041*4882a593Smuzhiyun * switch the OWN bit of all pending frames to the host
2042*4882a593Smuzhiyun */
2043*4882a593Smuzhiyun t = queue->tx_curr_get ;
2044*4882a593Smuzhiyun tx_used = queue->tx_used ;
2045*4882a593Smuzhiyun while (tx_used) {
2046*4882a593Smuzhiyun DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORCPU) ;
2047*4882a593Smuzhiyun DB_TX(5, "switch OWN bit of TxD 0x%p", t);
2048*4882a593Smuzhiyun t->txd_tbctrl &= ~cpu_to_le32(BMU_OWN) ;
2049*4882a593Smuzhiyun DRV_BUF_FLUSH(t,DDI_DMA_SYNC_FORDEV) ;
2050*4882a593Smuzhiyun t = t->txd_next ;
2051*4882a593Smuzhiyun tx_used-- ;
2052*4882a593Smuzhiyun }
2053*4882a593Smuzhiyun }
2054*4882a593Smuzhiyun
2055*4882a593Smuzhiyun /*
2056*4882a593Smuzhiyun * release all TxD's for both send queues
2057*4882a593Smuzhiyun */
2058*4882a593Smuzhiyun mac_drv_clear_txd(smc) ;
2059*4882a593Smuzhiyun
2060*4882a593Smuzhiyun for (i = QUEUE_S; i <= QUEUE_A0; i++) {
2061*4882a593Smuzhiyun queue = smc->hw.fp.tx[i] ;
2062*4882a593Smuzhiyun t = queue->tx_curr_get ;
2063*4882a593Smuzhiyun
2064*4882a593Smuzhiyun /*
2065*4882a593Smuzhiyun * write the phys pointer of the NEXT descriptor into the
2066*4882a593Smuzhiyun * BMU's current address descriptor pointer and set
2067*4882a593Smuzhiyun * tx_curr_get and tx_curr_put to this position
2068*4882a593Smuzhiyun */
2069*4882a593Smuzhiyun if (i == QUEUE_S) {
2070*4882a593Smuzhiyun outpd(ADDR(B5_XS_DA),le32_to_cpu(t->txd_ntdadr)) ;
2071*4882a593Smuzhiyun }
2072*4882a593Smuzhiyun else {
2073*4882a593Smuzhiyun outpd(ADDR(B5_XA_DA),le32_to_cpu(t->txd_ntdadr)) ;
2074*4882a593Smuzhiyun }
2075*4882a593Smuzhiyun
2076*4882a593Smuzhiyun queue->tx_curr_put = queue->tx_curr_get->txd_next ;
2077*4882a593Smuzhiyun queue->tx_curr_get = queue->tx_curr_put ;
2078*4882a593Smuzhiyun }
2079*4882a593Smuzhiyun }
2080*4882a593Smuzhiyun
2081*4882a593Smuzhiyun
2082*4882a593Smuzhiyun /*
2083*4882a593Smuzhiyun -------------------------------------------------------------
2084*4882a593Smuzhiyun TEST FUNCTIONS:
2085*4882a593Smuzhiyun -------------------------------------------------------------
2086*4882a593Smuzhiyun */
2087*4882a593Smuzhiyun
2088*4882a593Smuzhiyun #ifdef DEBUG
2089*4882a593Smuzhiyun /*
2090*4882a593Smuzhiyun * BEGIN_MANUAL_ENTRY(mac_drv_debug_lev)
2091*4882a593Smuzhiyun * void mac_drv_debug_lev(smc,flag,lev)
2092*4882a593Smuzhiyun *
2093*4882a593Smuzhiyun * function DOWNCALL (drvsr.c)
2094*4882a593Smuzhiyun * To get a special debug info the user can assign a debug level
2095*4882a593Smuzhiyun * to any debug flag.
2096*4882a593Smuzhiyun *
2097*4882a593Smuzhiyun * para flag debug flag, possible values are:
2098*4882a593Smuzhiyun * = 0: reset all debug flags (the defined level is
2099*4882a593Smuzhiyun * ignored)
2100*4882a593Smuzhiyun * = 1: debug.d_smtf
2101*4882a593Smuzhiyun * = 2: debug.d_smt
2102*4882a593Smuzhiyun * = 3: debug.d_ecm
2103*4882a593Smuzhiyun * = 4: debug.d_rmt
2104*4882a593Smuzhiyun * = 5: debug.d_cfm
2105*4882a593Smuzhiyun * = 6: debug.d_pcm
2106*4882a593Smuzhiyun *
2107*4882a593Smuzhiyun * = 10: debug.d_os.hwm_rx (hardware module receive path)
2108*4882a593Smuzhiyun * = 11: debug.d_os.hwm_tx(hardware module transmit path)
2109*4882a593Smuzhiyun * = 12: debug.d_os.hwm_gen(hardware module general flag)
2110*4882a593Smuzhiyun *
2111*4882a593Smuzhiyun * lev debug level
2112*4882a593Smuzhiyun *
2113*4882a593Smuzhiyun * END_MANUAL_ENTRY
2114*4882a593Smuzhiyun */
mac_drv_debug_lev(struct s_smc * smc,int flag,int lev)2115*4882a593Smuzhiyun void mac_drv_debug_lev(struct s_smc *smc, int flag, int lev)
2116*4882a593Smuzhiyun {
2117*4882a593Smuzhiyun switch(flag) {
2118*4882a593Smuzhiyun case (int)NULL:
2119*4882a593Smuzhiyun DB_P.d_smtf = DB_P.d_smt = DB_P.d_ecm = DB_P.d_rmt = 0 ;
2120*4882a593Smuzhiyun DB_P.d_cfm = 0 ;
2121*4882a593Smuzhiyun DB_P.d_os.hwm_rx = DB_P.d_os.hwm_tx = DB_P.d_os.hwm_gen = 0 ;
2122*4882a593Smuzhiyun #ifdef SBA
2123*4882a593Smuzhiyun DB_P.d_sba = 0 ;
2124*4882a593Smuzhiyun #endif
2125*4882a593Smuzhiyun #ifdef ESS
2126*4882a593Smuzhiyun DB_P.d_ess = 0 ;
2127*4882a593Smuzhiyun #endif
2128*4882a593Smuzhiyun break ;
2129*4882a593Smuzhiyun case DEBUG_SMTF:
2130*4882a593Smuzhiyun DB_P.d_smtf = lev ;
2131*4882a593Smuzhiyun break ;
2132*4882a593Smuzhiyun case DEBUG_SMT:
2133*4882a593Smuzhiyun DB_P.d_smt = lev ;
2134*4882a593Smuzhiyun break ;
2135*4882a593Smuzhiyun case DEBUG_ECM:
2136*4882a593Smuzhiyun DB_P.d_ecm = lev ;
2137*4882a593Smuzhiyun break ;
2138*4882a593Smuzhiyun case DEBUG_RMT:
2139*4882a593Smuzhiyun DB_P.d_rmt = lev ;
2140*4882a593Smuzhiyun break ;
2141*4882a593Smuzhiyun case DEBUG_CFM:
2142*4882a593Smuzhiyun DB_P.d_cfm = lev ;
2143*4882a593Smuzhiyun break ;
2144*4882a593Smuzhiyun case DEBUG_PCM:
2145*4882a593Smuzhiyun DB_P.d_pcm = lev ;
2146*4882a593Smuzhiyun break ;
2147*4882a593Smuzhiyun case DEBUG_SBA:
2148*4882a593Smuzhiyun #ifdef SBA
2149*4882a593Smuzhiyun DB_P.d_sba = lev ;
2150*4882a593Smuzhiyun #endif
2151*4882a593Smuzhiyun break ;
2152*4882a593Smuzhiyun case DEBUG_ESS:
2153*4882a593Smuzhiyun #ifdef ESS
2154*4882a593Smuzhiyun DB_P.d_ess = lev ;
2155*4882a593Smuzhiyun #endif
2156*4882a593Smuzhiyun break ;
2157*4882a593Smuzhiyun case DB_HWM_RX:
2158*4882a593Smuzhiyun DB_P.d_os.hwm_rx = lev ;
2159*4882a593Smuzhiyun break ;
2160*4882a593Smuzhiyun case DB_HWM_TX:
2161*4882a593Smuzhiyun DB_P.d_os.hwm_tx = lev ;
2162*4882a593Smuzhiyun break ;
2163*4882a593Smuzhiyun case DB_HWM_GEN:
2164*4882a593Smuzhiyun DB_P.d_os.hwm_gen = lev ;
2165*4882a593Smuzhiyun break ;
2166*4882a593Smuzhiyun default:
2167*4882a593Smuzhiyun break ;
2168*4882a593Smuzhiyun }
2169*4882a593Smuzhiyun }
2170*4882a593Smuzhiyun #endif
2171