xref: /openbmc/linux/drivers/atm/iphase.c (revision a09d2831)
1 /******************************************************************************
2          iphase.c: Device driver for Interphase ATM PCI adapter cards
3                     Author: Peter Wang  <pwang@iphase.com>
4 		   Some fixes: Arnaldo Carvalho de Melo <acme@conectiva.com.br>
5                    Interphase Corporation  <www.iphase.com>
6                                Version: 1.0
7 *******************************************************************************
8 
9       This software may be used and distributed according to the terms
10       of the GNU General Public License (GPL), incorporated herein by reference.
11       Drivers based on this skeleton fall under the GPL and must retain
12       the authorship (implicit copyright) notice.
13 
14       This program is distributed in the hope that it will be useful, but
15       WITHOUT ANY WARRANTY; without even the implied warranty of
16       MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
17       General Public License for more details.
18 
19       Modified from an incomplete driver for Interphase 5575 1KVC 1M card which
20       was originally written by Monalisa Agrawal at UNH. Now this driver
21       supports a variety of varients of Interphase ATM PCI (i)Chip adapter
22       card family (See www.iphase.com/products/ClassSheet.cfm?ClassID=ATM)
23       in terms of PHY type, the size of control memory and the size of
24       packet memory. The followings are the change log and history:
25 
26           Bugfix the Mona's UBR driver.
27           Modify the basic memory allocation and dma logic.
28           Port the driver to the latest kernel from 2.0.46.
29           Complete the ABR logic of the driver, and added the ABR work-
30               around for the hardware anormalies.
31           Add the CBR support.
32 	  Add the flow control logic to the driver to allow rate-limit VC.
33           Add 4K VC support to the board with 512K control memory.
34           Add the support of all the variants of the Interphase ATM PCI
35           (i)Chip adapter cards including x575 (155M OC3 and UTP155), x525
36           (25M UTP25) and x531 (DS3 and E3).
37           Add SMP support.
38 
39       Support and updates available at: ftp://ftp.iphase.com/pub/atm
40 
41 *******************************************************************************/
42 
43 #include <linux/module.h>
44 #include <linux/kernel.h>
45 #include <linux/mm.h>
46 #include <linux/pci.h>
47 #include <linux/errno.h>
48 #include <linux/atm.h>
49 #include <linux/atmdev.h>
50 #include <linux/sonet.h>
51 #include <linux/skbuff.h>
52 #include <linux/time.h>
53 #include <linux/delay.h>
54 #include <linux/uio.h>
55 #include <linux/init.h>
56 #include <linux/wait.h>
57 #include <asm/system.h>
58 #include <asm/io.h>
59 #include <asm/atomic.h>
60 #include <asm/uaccess.h>
61 #include <asm/string.h>
62 #include <asm/byteorder.h>
63 #include <linux/vmalloc.h>
64 #include <linux/jiffies.h>
65 #include "iphase.h"
66 #include "suni.h"
67 #define swap_byte_order(x) (((x & 0xff) << 8) | ((x & 0xff00) >> 8))
68 
69 #define PRIV(dev) ((struct suni_priv *) dev->phy_data)
70 
71 static unsigned char ia_phy_get(struct atm_dev *dev, unsigned long addr);
72 static void desc_dbg(IADEV *iadev);
73 
74 static IADEV *ia_dev[8];
75 static struct atm_dev *_ia_dev[8];
76 static int iadev_count;
77 static void ia_led_timer(unsigned long arg);
78 static DEFINE_TIMER(ia_timer, ia_led_timer, 0, 0);
79 static int IA_TX_BUF = DFL_TX_BUFFERS, IA_TX_BUF_SZ = DFL_TX_BUF_SZ;
80 static int IA_RX_BUF = DFL_RX_BUFFERS, IA_RX_BUF_SZ = DFL_RX_BUF_SZ;
81 static uint IADebugFlag = /* IF_IADBG_ERR | IF_IADBG_CBR| IF_IADBG_INIT_ADAPTER
82             |IF_IADBG_ABR | IF_IADBG_EVENT*/ 0;
83 
84 module_param(IA_TX_BUF, int, 0);
85 module_param(IA_TX_BUF_SZ, int, 0);
86 module_param(IA_RX_BUF, int, 0);
87 module_param(IA_RX_BUF_SZ, int, 0);
88 module_param(IADebugFlag, uint, 0644);
89 
90 MODULE_LICENSE("GPL");
91 
92 /**************************** IA_LIB **********************************/
93 
94 static void ia_init_rtn_q (IARTN_Q *que)
95 {
96    que->next = NULL;
97    que->tail = NULL;
98 }
99 
100 static void ia_enque_head_rtn_q (IARTN_Q *que, IARTN_Q * data)
101 {
102    data->next = NULL;
103    if (que->next == NULL)
104       que->next = que->tail = data;
105    else {
106       data->next = que->next;
107       que->next = data;
108    }
109    return;
110 }
111 
112 static int ia_enque_rtn_q (IARTN_Q *que, struct desc_tbl_t data) {
113    IARTN_Q *entry = kmalloc(sizeof(*entry), GFP_ATOMIC);
114    if (!entry) return -1;
115    entry->data = data;
116    entry->next = NULL;
117    if (que->next == NULL)
118       que->next = que->tail = entry;
119    else {
120       que->tail->next = entry;
121       que->tail = que->tail->next;
122    }
123    return 1;
124 }
125 
126 static IARTN_Q * ia_deque_rtn_q (IARTN_Q *que) {
127    IARTN_Q *tmpdata;
128    if (que->next == NULL)
129       return NULL;
130    tmpdata = que->next;
131    if ( que->next == que->tail)
132       que->next = que->tail = NULL;
133    else
134       que->next = que->next->next;
135    return tmpdata;
136 }
137 
138 static void ia_hack_tcq(IADEV *dev) {
139 
140   u_short 		desc1;
141   u_short		tcq_wr;
142   struct ia_vcc         *iavcc_r = NULL;
143 
144   tcq_wr = readl(dev->seg_reg+TCQ_WR_PTR) & 0xffff;
145   while (dev->host_tcq_wr != tcq_wr) {
146      desc1 = *(u_short *)(dev->seg_ram + dev->host_tcq_wr);
147      if (!desc1) ;
148      else if (!dev->desc_tbl[desc1 -1].timestamp) {
149         IF_ABR(printk(" Desc %d is reset at %ld\n", desc1 -1, jiffies);)
150         *(u_short *) (dev->seg_ram + dev->host_tcq_wr) = 0;
151      }
152      else if (dev->desc_tbl[desc1 -1].timestamp) {
153         if (!(iavcc_r = dev->desc_tbl[desc1 -1].iavcc)) {
154            printk("IA: Fatal err in get_desc\n");
155            continue;
156         }
157         iavcc_r->vc_desc_cnt--;
158         dev->desc_tbl[desc1 -1].timestamp = 0;
159         IF_EVENT(printk("ia_hack: return_q skb = 0x%p desc = %d\n",
160                                    dev->desc_tbl[desc1 -1].txskb, desc1);)
161         if (iavcc_r->pcr < dev->rate_limit) {
162            IA_SKB_STATE (dev->desc_tbl[desc1-1].txskb) |= IA_TX_DONE;
163            if (ia_enque_rtn_q(&dev->tx_return_q, dev->desc_tbl[desc1 -1]) < 0)
164               printk("ia_hack_tcq: No memory available\n");
165         }
166         dev->desc_tbl[desc1 -1].iavcc = NULL;
167         dev->desc_tbl[desc1 -1].txskb = NULL;
168      }
169      dev->host_tcq_wr += 2;
170      if (dev->host_tcq_wr > dev->ffL.tcq_ed)
171         dev->host_tcq_wr = dev->ffL.tcq_st;
172   }
173 } /* ia_hack_tcq */
174 
175 static u16 get_desc (IADEV *dev, struct ia_vcc *iavcc) {
176   u_short 		desc_num, i;
177   struct sk_buff        *skb;
178   struct ia_vcc         *iavcc_r = NULL;
179   unsigned long delta;
180   static unsigned long timer = 0;
181   int ltimeout;
182 
183   ia_hack_tcq (dev);
184   if((time_after(jiffies,timer+50)) || ((dev->ffL.tcq_rd==dev->host_tcq_wr))) {
185      timer = jiffies;
186      i=0;
187      while (i < dev->num_tx_desc) {
188         if (!dev->desc_tbl[i].timestamp) {
189            i++;
190            continue;
191         }
192         ltimeout = dev->desc_tbl[i].iavcc->ltimeout;
193         delta = jiffies - dev->desc_tbl[i].timestamp;
194         if (delta >= ltimeout) {
195            IF_ABR(printk("RECOVER run!! desc_tbl %d = %d  delta = %ld, time = %ld\n", i,dev->desc_tbl[i].timestamp, delta, jiffies);)
196            if (dev->ffL.tcq_rd == dev->ffL.tcq_st)
197               dev->ffL.tcq_rd =  dev->ffL.tcq_ed;
198            else
199               dev->ffL.tcq_rd -= 2;
200            *(u_short *)(dev->seg_ram + dev->ffL.tcq_rd) = i+1;
201            if (!(skb = dev->desc_tbl[i].txskb) ||
202                           !(iavcc_r = dev->desc_tbl[i].iavcc))
203               printk("Fatal err, desc table vcc or skb is NULL\n");
204            else
205               iavcc_r->vc_desc_cnt--;
206            dev->desc_tbl[i].timestamp = 0;
207            dev->desc_tbl[i].iavcc = NULL;
208            dev->desc_tbl[i].txskb = NULL;
209         }
210         i++;
211      } /* while */
212   }
213   if (dev->ffL.tcq_rd == dev->host_tcq_wr)
214      return 0xFFFF;
215 
216   /* Get the next available descriptor number from TCQ */
217   desc_num = *(u_short *)(dev->seg_ram + dev->ffL.tcq_rd);
218 
219   while (!desc_num || (dev->desc_tbl[desc_num -1]).timestamp) {
220      dev->ffL.tcq_rd += 2;
221      if (dev->ffL.tcq_rd > dev->ffL.tcq_ed)
222      dev->ffL.tcq_rd = dev->ffL.tcq_st;
223      if (dev->ffL.tcq_rd == dev->host_tcq_wr)
224         return 0xFFFF;
225      desc_num = *(u_short *)(dev->seg_ram + dev->ffL.tcq_rd);
226   }
227 
228   /* get system time */
229   dev->desc_tbl[desc_num -1].timestamp = jiffies;
230   return desc_num;
231 }
232 
233 static void clear_lockup (struct atm_vcc *vcc, IADEV *dev) {
234   u_char          	foundLockUp;
235   vcstatus_t		*vcstatus;
236   u_short               *shd_tbl;
237   u_short               tempCellSlot, tempFract;
238   struct main_vc *abr_vc = (struct main_vc *)dev->MAIN_VC_TABLE_ADDR;
239   struct ext_vc *eabr_vc = (struct ext_vc *)dev->EXT_VC_TABLE_ADDR;
240   u_int  i;
241 
242   if (vcc->qos.txtp.traffic_class == ATM_ABR) {
243      vcstatus = (vcstatus_t *) &(dev->testTable[vcc->vci]->vc_status);
244      vcstatus->cnt++;
245      foundLockUp = 0;
246      if( vcstatus->cnt == 0x05 ) {
247         abr_vc += vcc->vci;
248 	eabr_vc += vcc->vci;
249 	if( eabr_vc->last_desc ) {
250 	   if( (abr_vc->status & 0x07) == ABR_STATE /* 0x2 */ ) {
251               /* Wait for 10 Micro sec */
252               udelay(10);
253 	      if ((eabr_vc->last_desc)&&((abr_vc->status & 0x07)==ABR_STATE))
254 		 foundLockUp = 1;
255            }
256 	   else {
257 	      tempCellSlot = abr_vc->last_cell_slot;
258               tempFract    = abr_vc->fraction;
259               if((tempCellSlot == dev->testTable[vcc->vci]->lastTime)
260                          && (tempFract == dev->testTable[vcc->vci]->fract))
261 	         foundLockUp = 1;
262               dev->testTable[vcc->vci]->lastTime = tempCellSlot;
263               dev->testTable[vcc->vci]->fract = tempFract;
264 	   }
265         } /* last descriptor */
266         vcstatus->cnt = 0;
267      } /* vcstatus->cnt */
268 
269      if (foundLockUp) {
270         IF_ABR(printk("LOCK UP found\n");)
271 	writew(0xFFFD, dev->seg_reg+MODE_REG_0);
272         /* Wait for 10 Micro sec */
273         udelay(10);
274         abr_vc->status &= 0xFFF8;
275         abr_vc->status |= 0x0001;  /* state is idle */
276 	shd_tbl = (u_short *)dev->ABR_SCHED_TABLE_ADDR;
277 	for( i = 0; ((i < dev->num_vc) && (shd_tbl[i])); i++ );
278 	if (i < dev->num_vc)
279            shd_tbl[i] = vcc->vci;
280         else
281            IF_ERR(printk("ABR Seg. may not continue on VC %x\n",vcc->vci);)
282         writew(T_ONLINE, dev->seg_reg+MODE_REG_0);
283         writew(~(TRANSMIT_DONE|TCQ_NOT_EMPTY), dev->seg_reg+SEG_MASK_REG);
284         writew(TRANSMIT_DONE, dev->seg_reg+SEG_INTR_STATUS_REG);
285 	vcstatus->cnt = 0;
286      } /* foundLockUp */
287 
288   } /* if an ABR VC */
289 
290 
291 }
292 
293 /*
294 ** Conversion of 24-bit cellrate (cells/sec) to 16-bit floating point format.
295 **
296 **  +----+----+------------------+-------------------------------+
297 **  |  R | NZ |  5-bit exponent  |        9-bit mantissa         |
298 **  +----+----+------------------+-------------------------------+
299 **
300 **    R = reserved (written as 0)
301 **    NZ = 0 if 0 cells/sec; 1 otherwise
302 **
303 **    if NZ = 1, rate = 1.mmmmmmmmm x 2^(eeeee) cells/sec
304 */
305 static u16
306 cellrate_to_float(u32 cr)
307 {
308 
309 #define	NZ 		0x4000
310 #define	M_BITS		9		/* Number of bits in mantissa */
311 #define	E_BITS		5		/* Number of bits in exponent */
312 #define	M_MASK		0x1ff
313 #define	E_MASK		0x1f
314   u16   flot;
315   u32	tmp = cr & 0x00ffffff;
316   int 	i   = 0;
317   if (cr == 0)
318      return 0;
319   while (tmp != 1) {
320      tmp >>= 1;
321      i++;
322   }
323   if (i == M_BITS)
324      flot = NZ | (i << M_BITS) | (cr & M_MASK);
325   else if (i < M_BITS)
326      flot = NZ | (i << M_BITS) | ((cr << (M_BITS - i)) & M_MASK);
327   else
328      flot = NZ | (i << M_BITS) | ((cr >> (i - M_BITS)) & M_MASK);
329   return flot;
330 }
331 
332 #if 0
333 /*
334 ** Conversion of 16-bit floating point format to 24-bit cellrate (cells/sec).
335 */
336 static u32
337 float_to_cellrate(u16 rate)
338 {
339   u32   exp, mantissa, cps;
340   if ((rate & NZ) == 0)
341      return 0;
342   exp = (rate >> M_BITS) & E_MASK;
343   mantissa = rate & M_MASK;
344   if (exp == 0)
345      return 1;
346   cps = (1 << M_BITS) | mantissa;
347   if (exp == M_BITS)
348      cps = cps;
349   else if (exp > M_BITS)
350      cps <<= (exp - M_BITS);
351   else
352      cps >>= (M_BITS - exp);
353   return cps;
354 }
355 #endif
356 
357 static void init_abr_vc (IADEV *dev, srv_cls_param_t *srv_p) {
358   srv_p->class_type = ATM_ABR;
359   srv_p->pcr        = dev->LineRate;
360   srv_p->mcr        = 0;
361   srv_p->icr        = 0x055cb7;
362   srv_p->tbe        = 0xffffff;
363   srv_p->frtt       = 0x3a;
364   srv_p->rif        = 0xf;
365   srv_p->rdf        = 0xb;
366   srv_p->nrm        = 0x4;
367   srv_p->trm        = 0x7;
368   srv_p->cdf        = 0x3;
369   srv_p->adtf       = 50;
370 }
371 
372 static int
373 ia_open_abr_vc(IADEV *dev, srv_cls_param_t *srv_p,
374                                                 struct atm_vcc *vcc, u8 flag)
375 {
376   f_vc_abr_entry  *f_abr_vc;
377   r_vc_abr_entry  *r_abr_vc;
378   u32		icr;
379   u8		trm, nrm, crm;
380   u16		adtf, air, *ptr16;
381   f_abr_vc =(f_vc_abr_entry *)dev->MAIN_VC_TABLE_ADDR;
382   f_abr_vc += vcc->vci;
383   switch (flag) {
384      case 1: /* FFRED initialization */
385 #if 0  /* sanity check */
386        if (srv_p->pcr == 0)
387           return INVALID_PCR;
388        if (srv_p->pcr > dev->LineRate)
389           srv_p->pcr = dev->LineRate;
390        if ((srv_p->mcr + dev->sum_mcr) > dev->LineRate)
391 	  return MCR_UNAVAILABLE;
392        if (srv_p->mcr > srv_p->pcr)
393 	  return INVALID_MCR;
394        if (!(srv_p->icr))
395 	  srv_p->icr = srv_p->pcr;
396        if ((srv_p->icr < srv_p->mcr) || (srv_p->icr > srv_p->pcr))
397 	  return INVALID_ICR;
398        if ((srv_p->tbe < MIN_TBE) || (srv_p->tbe > MAX_TBE))
399 	  return INVALID_TBE;
400        if ((srv_p->frtt < MIN_FRTT) || (srv_p->frtt > MAX_FRTT))
401 	  return INVALID_FRTT;
402        if (srv_p->nrm > MAX_NRM)
403 	  return INVALID_NRM;
404        if (srv_p->trm > MAX_TRM)
405 	  return INVALID_TRM;
406        if (srv_p->adtf > MAX_ADTF)
407           return INVALID_ADTF;
408        else if (srv_p->adtf == 0)
409 	  srv_p->adtf = 1;
410        if (srv_p->cdf > MAX_CDF)
411 	  return INVALID_CDF;
412        if (srv_p->rif > MAX_RIF)
413 	  return INVALID_RIF;
414        if (srv_p->rdf > MAX_RDF)
415 	  return INVALID_RDF;
416 #endif
417        memset ((caddr_t)f_abr_vc, 0, sizeof(*f_abr_vc));
418        f_abr_vc->f_vc_type = ABR;
419        nrm = 2 << srv_p->nrm;     /* (2 ** (srv_p->nrm +1)) */
420 			          /* i.e 2**n = 2 << (n-1) */
421        f_abr_vc->f_nrm = nrm << 8 | nrm;
422        trm = 100000/(2 << (16 - srv_p->trm));
423        if ( trm == 0) trm = 1;
424        f_abr_vc->f_nrmexp =(((srv_p->nrm +1) & 0x0f) << 12)|(MRM << 8) | trm;
425        crm = srv_p->tbe / nrm;
426        if (crm == 0) crm = 1;
427        f_abr_vc->f_crm = crm & 0xff;
428        f_abr_vc->f_pcr = cellrate_to_float(srv_p->pcr);
429        icr = min( srv_p->icr, (srv_p->tbe > srv_p->frtt) ?
430 				((srv_p->tbe/srv_p->frtt)*1000000) :
431 				(1000000/(srv_p->frtt/srv_p->tbe)));
432        f_abr_vc->f_icr = cellrate_to_float(icr);
433        adtf = (10000 * srv_p->adtf)/8192;
434        if (adtf == 0) adtf = 1;
435        f_abr_vc->f_cdf = ((7 - srv_p->cdf) << 12 | adtf) & 0xfff;
436        f_abr_vc->f_mcr = cellrate_to_float(srv_p->mcr);
437        f_abr_vc->f_acr = f_abr_vc->f_icr;
438        f_abr_vc->f_status = 0x0042;
439        break;
440     case 0: /* RFRED initialization */
441        ptr16 = (u_short *)(dev->reass_ram + REASS_TABLE*dev->memSize);
442        *(ptr16 + vcc->vci) = NO_AAL5_PKT | REASS_ABR;
443        r_abr_vc = (r_vc_abr_entry*)(dev->reass_ram+ABR_VC_TABLE*dev->memSize);
444        r_abr_vc += vcc->vci;
445        r_abr_vc->r_status_rdf = (15 - srv_p->rdf) & 0x000f;
446        air = srv_p->pcr << (15 - srv_p->rif);
447        if (air == 0) air = 1;
448        r_abr_vc->r_air = cellrate_to_float(air);
449        dev->testTable[vcc->vci]->vc_status = VC_ACTIVE | VC_ABR;
450        dev->sum_mcr	   += srv_p->mcr;
451        dev->n_abr++;
452        break;
453     default:
454        break;
455   }
456   return	0;
457 }
458 static int ia_cbr_setup (IADEV *dev, struct atm_vcc *vcc) {
459    u32 rateLow=0, rateHigh, rate;
460    int entries;
461    struct ia_vcc *ia_vcc;
462 
463    int   idealSlot =0, testSlot, toBeAssigned, inc;
464    u32   spacing;
465    u16  *SchedTbl, *TstSchedTbl;
466    u16  cbrVC, vcIndex;
467    u32   fracSlot    = 0;
468    u32   sp_mod      = 0;
469    u32   sp_mod2     = 0;
470 
471    /* IpAdjustTrafficParams */
472    if (vcc->qos.txtp.max_pcr <= 0) {
473       IF_ERR(printk("PCR for CBR not defined\n");)
474       return -1;
475    }
476    rate = vcc->qos.txtp.max_pcr;
477    entries = rate / dev->Granularity;
478    IF_CBR(printk("CBR: CBR entries=0x%x for rate=0x%x & Gran=0x%x\n",
479                                 entries, rate, dev->Granularity);)
480    if (entries < 1)
481       IF_CBR(printk("CBR: Bandwidth smaller than granularity of CBR table\n");)
482    rateLow  =  entries * dev->Granularity;
483    rateHigh = (entries + 1) * dev->Granularity;
484    if (3*(rate - rateLow) > (rateHigh - rate))
485       entries++;
486    if (entries > dev->CbrRemEntries) {
487       IF_CBR(printk("CBR: Not enough bandwidth to support this PCR.\n");)
488       IF_CBR(printk("Entries = 0x%x, CbrRemEntries = 0x%x.\n",
489                                        entries, dev->CbrRemEntries);)
490       return -EBUSY;
491    }
492 
493    ia_vcc = INPH_IA_VCC(vcc);
494    ia_vcc->NumCbrEntry = entries;
495    dev->sum_mcr += entries * dev->Granularity;
496    /* IaFFrednInsertCbrSched */
497    // Starting at an arbitrary location, place the entries into the table
498    // as smoothly as possible
499    cbrVC   = 0;
500    spacing = dev->CbrTotEntries / entries;
501    sp_mod  = dev->CbrTotEntries % entries; // get modulo
502    toBeAssigned = entries;
503    fracSlot = 0;
504    vcIndex  = vcc->vci;
505    IF_CBR(printk("Vci=0x%x,Spacing=0x%x,Sp_mod=0x%x\n",vcIndex,spacing,sp_mod);)
506    while (toBeAssigned)
507    {
508       // If this is the first time, start the table loading for this connection
509       // as close to entryPoint as possible.
510       if (toBeAssigned == entries)
511       {
512          idealSlot = dev->CbrEntryPt;
513          dev->CbrEntryPt += 2;    // Adding 2 helps to prevent clumping
514          if (dev->CbrEntryPt >= dev->CbrTotEntries)
515             dev->CbrEntryPt -= dev->CbrTotEntries;// Wrap if necessary
516       } else {
517          idealSlot += (u32)(spacing + fracSlot); // Point to the next location
518          // in the table that would be  smoothest
519          fracSlot = ((sp_mod + sp_mod2) / entries);  // get new integer part
520          sp_mod2  = ((sp_mod + sp_mod2) % entries);  // calc new fractional part
521       }
522       if (idealSlot >= (int)dev->CbrTotEntries)
523          idealSlot -= dev->CbrTotEntries;
524       // Continuously check around this ideal value until a null
525       // location is encountered.
526       SchedTbl = (u16*)(dev->seg_ram+CBR_SCHED_TABLE*dev->memSize);
527       inc = 0;
528       testSlot = idealSlot;
529       TstSchedTbl = (u16*)(SchedTbl+testSlot);  //set index and read in value
530       IF_CBR(printk("CBR Testslot 0x%x AT Location 0x%p, NumToAssign=%d\n",
531                                 testSlot, TstSchedTbl,toBeAssigned);)
532       memcpy((caddr_t)&cbrVC,(caddr_t)TstSchedTbl,sizeof(cbrVC));
533       while (cbrVC)  // If another VC at this location, we have to keep looking
534       {
535           inc++;
536           testSlot = idealSlot - inc;
537           if (testSlot < 0) { // Wrap if necessary
538              testSlot += dev->CbrTotEntries;
539              IF_CBR(printk("Testslot Wrap. STable Start=0x%p,Testslot=%d\n",
540                                                        SchedTbl,testSlot);)
541           }
542           TstSchedTbl = (u16 *)(SchedTbl + testSlot);  // set table index
543           memcpy((caddr_t)&cbrVC,(caddr_t)TstSchedTbl,sizeof(cbrVC));
544           if (!cbrVC)
545              break;
546           testSlot = idealSlot + inc;
547           if (testSlot >= (int)dev->CbrTotEntries) { // Wrap if necessary
548              testSlot -= dev->CbrTotEntries;
549              IF_CBR(printk("TotCbrEntries=%d",dev->CbrTotEntries);)
550              IF_CBR(printk(" Testslot=0x%x ToBeAssgned=%d\n",
551                                             testSlot, toBeAssigned);)
552           }
553           // set table index and read in value
554           TstSchedTbl = (u16*)(SchedTbl + testSlot);
555           IF_CBR(printk("Reading CBR Tbl from 0x%p, CbrVal=0x%x Iteration %d\n",
556                           TstSchedTbl,cbrVC,inc);)
557           memcpy((caddr_t)&cbrVC,(caddr_t)TstSchedTbl,sizeof(cbrVC));
558        } /* while */
559        // Move this VCI number into this location of the CBR Sched table.
560        memcpy((caddr_t)TstSchedTbl, (caddr_t)&vcIndex, sizeof(*TstSchedTbl));
561        dev->CbrRemEntries--;
562        toBeAssigned--;
563    } /* while */
564 
565    /* IaFFrednCbrEnable */
566    dev->NumEnabledCBR++;
567    if (dev->NumEnabledCBR == 1) {
568        writew((CBR_EN | UBR_EN | ABR_EN | (0x23 << 2)), dev->seg_reg+STPARMS);
569        IF_CBR(printk("CBR is enabled\n");)
570    }
571    return 0;
572 }
573 static void ia_cbrVc_close (struct atm_vcc *vcc) {
574    IADEV *iadev;
575    u16 *SchedTbl, NullVci = 0;
576    u32 i, NumFound;
577 
578    iadev = INPH_IA_DEV(vcc->dev);
579    iadev->NumEnabledCBR--;
580    SchedTbl = (u16*)(iadev->seg_ram+CBR_SCHED_TABLE*iadev->memSize);
581    if (iadev->NumEnabledCBR == 0) {
582       writew((UBR_EN | ABR_EN | (0x23 << 2)), iadev->seg_reg+STPARMS);
583       IF_CBR (printk("CBR support disabled\n");)
584    }
585    NumFound = 0;
586    for (i=0; i < iadev->CbrTotEntries; i++)
587    {
588       if (*SchedTbl == vcc->vci) {
589          iadev->CbrRemEntries++;
590          *SchedTbl = NullVci;
591          IF_CBR(NumFound++;)
592       }
593       SchedTbl++;
594    }
595    IF_CBR(printk("Exit ia_cbrVc_close, NumRemoved=%d\n",NumFound);)
596 }
597 
598 static int ia_avail_descs(IADEV *iadev) {
599    int tmp = 0;
600    ia_hack_tcq(iadev);
601    if (iadev->host_tcq_wr >= iadev->ffL.tcq_rd)
602       tmp = (iadev->host_tcq_wr - iadev->ffL.tcq_rd) / 2;
603    else
604       tmp = (iadev->ffL.tcq_ed - iadev->ffL.tcq_rd + 2 + iadev->host_tcq_wr -
605                    iadev->ffL.tcq_st) / 2;
606    return tmp;
607 }
608 
609 static int ia_pkt_tx (struct atm_vcc *vcc, struct sk_buff *skb);
610 
611 static int ia_que_tx (IADEV *iadev) {
612    struct sk_buff *skb;
613    int num_desc;
614    struct atm_vcc *vcc;
615    struct ia_vcc *iavcc;
616    num_desc = ia_avail_descs(iadev);
617 
618    while (num_desc && (skb = skb_dequeue(&iadev->tx_backlog))) {
619       if (!(vcc = ATM_SKB(skb)->vcc)) {
620          dev_kfree_skb_any(skb);
621          printk("ia_que_tx: Null vcc\n");
622          break;
623       }
624       if (!test_bit(ATM_VF_READY,&vcc->flags)) {
625          dev_kfree_skb_any(skb);
626          printk("Free the SKB on closed vci %d \n", vcc->vci);
627          break;
628       }
629       iavcc = INPH_IA_VCC(vcc);
630       if (ia_pkt_tx (vcc, skb)) {
631          skb_queue_head(&iadev->tx_backlog, skb);
632       }
633       num_desc--;
634    }
635    return 0;
636 }
637 
638 static void ia_tx_poll (IADEV *iadev) {
639    struct atm_vcc *vcc = NULL;
640    struct sk_buff *skb = NULL, *skb1 = NULL;
641    struct ia_vcc *iavcc;
642    IARTN_Q *  rtne;
643 
644    ia_hack_tcq(iadev);
645    while ( (rtne = ia_deque_rtn_q(&iadev->tx_return_q))) {
646        skb = rtne->data.txskb;
647        if (!skb) {
648            printk("ia_tx_poll: skb is null\n");
649            goto out;
650        }
651        vcc = ATM_SKB(skb)->vcc;
652        if (!vcc) {
653            printk("ia_tx_poll: vcc is null\n");
654            dev_kfree_skb_any(skb);
655 	   goto out;
656        }
657 
658        iavcc = INPH_IA_VCC(vcc);
659        if (!iavcc) {
660            printk("ia_tx_poll: iavcc is null\n");
661            dev_kfree_skb_any(skb);
662 	   goto out;
663        }
664 
665        skb1 = skb_dequeue(&iavcc->txing_skb);
666        while (skb1 && (skb1 != skb)) {
667           if (!(IA_SKB_STATE(skb1) & IA_TX_DONE)) {
668              printk("IA_tx_intr: Vci %d lost pkt!!!\n", vcc->vci);
669           }
670           IF_ERR(printk("Release the SKB not match\n");)
671           if ((vcc->pop) && (skb1->len != 0))
672           {
673              vcc->pop(vcc, skb1);
674              IF_EVENT(printk("Tansmit Done - skb 0x%lx return\n",
675                                                           (long)skb1);)
676           }
677           else
678              dev_kfree_skb_any(skb1);
679           skb1 = skb_dequeue(&iavcc->txing_skb);
680        }
681        if (!skb1) {
682           IF_EVENT(printk("IA: Vci %d - skb not found requed\n",vcc->vci);)
683           ia_enque_head_rtn_q (&iadev->tx_return_q, rtne);
684           break;
685        }
686        if ((vcc->pop) && (skb->len != 0))
687        {
688           vcc->pop(vcc, skb);
689           IF_EVENT(printk("Tx Done - skb 0x%lx return\n",(long)skb);)
690        }
691        else
692           dev_kfree_skb_any(skb);
693        kfree(rtne);
694     }
695     ia_que_tx(iadev);
696 out:
697     return;
698 }
699 #if 0
700 static void ia_eeprom_put (IADEV *iadev, u32 addr, u_short val)
701 {
702         u32	t;
703 	int	i;
704 	/*
705 	 * Issue a command to enable writes to the NOVRAM
706 	 */
707 	NVRAM_CMD (EXTEND + EWEN);
708 	NVRAM_CLR_CE;
709 	/*
710 	 * issue the write command
711 	 */
712 	NVRAM_CMD(IAWRITE + addr);
713 	/*
714 	 * Send the data, starting with D15, then D14, and so on for 16 bits
715 	 */
716 	for (i=15; i>=0; i--) {
717 		NVRAM_CLKOUT (val & 0x8000);
718 		val <<= 1;
719 	}
720 	NVRAM_CLR_CE;
721 	CFG_OR(NVCE);
722 	t = readl(iadev->reg+IPHASE5575_EEPROM_ACCESS);
723 	while (!(t & NVDO))
724 		t = readl(iadev->reg+IPHASE5575_EEPROM_ACCESS);
725 
726 	NVRAM_CLR_CE;
727 	/*
728 	 * disable writes again
729 	 */
730 	NVRAM_CMD(EXTEND + EWDS)
731 	NVRAM_CLR_CE;
732 	CFG_AND(~NVDI);
733 }
734 #endif
735 
736 static u16 ia_eeprom_get (IADEV *iadev, u32 addr)
737 {
738 	u_short	val;
739         u32	t;
740 	int	i;
741 	/*
742 	 * Read the first bit that was clocked with the falling edge of the
743 	 * the last command data clock
744 	 */
745 	NVRAM_CMD(IAREAD + addr);
746 	/*
747 	 * Now read the rest of the bits, the next bit read is D14, then D13,
748 	 * and so on.
749 	 */
750 	val = 0;
751 	for (i=15; i>=0; i--) {
752 		NVRAM_CLKIN(t);
753 		val |= (t << i);
754 	}
755 	NVRAM_CLR_CE;
756 	CFG_AND(~NVDI);
757 	return val;
758 }
759 
760 static void ia_hw_type(IADEV *iadev) {
761    u_short memType = ia_eeprom_get(iadev, 25);
762    iadev->memType = memType;
763    if ((memType & MEM_SIZE_MASK) == MEM_SIZE_1M) {
764       iadev->num_tx_desc = IA_TX_BUF;
765       iadev->tx_buf_sz = IA_TX_BUF_SZ;
766       iadev->num_rx_desc = IA_RX_BUF;
767       iadev->rx_buf_sz = IA_RX_BUF_SZ;
768    } else if ((memType & MEM_SIZE_MASK) == MEM_SIZE_512K) {
769       if (IA_TX_BUF == DFL_TX_BUFFERS)
770         iadev->num_tx_desc = IA_TX_BUF / 2;
771       else
772         iadev->num_tx_desc = IA_TX_BUF;
773       iadev->tx_buf_sz = IA_TX_BUF_SZ;
774       if (IA_RX_BUF == DFL_RX_BUFFERS)
775         iadev->num_rx_desc = IA_RX_BUF / 2;
776       else
777         iadev->num_rx_desc = IA_RX_BUF;
778       iadev->rx_buf_sz = IA_RX_BUF_SZ;
779    }
780    else {
781       if (IA_TX_BUF == DFL_TX_BUFFERS)
782         iadev->num_tx_desc = IA_TX_BUF / 8;
783       else
784         iadev->num_tx_desc = IA_TX_BUF;
785       iadev->tx_buf_sz = IA_TX_BUF_SZ;
786       if (IA_RX_BUF == DFL_RX_BUFFERS)
787         iadev->num_rx_desc = IA_RX_BUF / 8;
788       else
789         iadev->num_rx_desc = IA_RX_BUF;
790       iadev->rx_buf_sz = IA_RX_BUF_SZ;
791    }
792    iadev->rx_pkt_ram = TX_PACKET_RAM + (iadev->num_tx_desc * iadev->tx_buf_sz);
793    IF_INIT(printk("BUF: tx=%d,sz=%d rx=%d sz= %d rx_pkt_ram=%d\n",
794          iadev->num_tx_desc, iadev->tx_buf_sz, iadev->num_rx_desc,
795          iadev->rx_buf_sz, iadev->rx_pkt_ram);)
796 
797 #if 0
798    if ((memType & FE_MASK) == FE_SINGLE_MODE) {
799       iadev->phy_type = PHY_OC3C_S;
800    else if ((memType & FE_MASK) == FE_UTP_OPTION)
801       iadev->phy_type = PHY_UTP155;
802    else
803      iadev->phy_type = PHY_OC3C_M;
804 #endif
805 
806    iadev->phy_type = memType & FE_MASK;
807    IF_INIT(printk("memType = 0x%x iadev->phy_type = 0x%x\n",
808                                          memType,iadev->phy_type);)
809    if (iadev->phy_type == FE_25MBIT_PHY)
810       iadev->LineRate = (u32)(((25600000/8)*26)/(27*53));
811    else if (iadev->phy_type == FE_DS3_PHY)
812       iadev->LineRate = (u32)(((44736000/8)*26)/(27*53));
813    else if (iadev->phy_type == FE_E3_PHY)
814       iadev->LineRate = (u32)(((34368000/8)*26)/(27*53));
815    else
816        iadev->LineRate = (u32)(ATM_OC3_PCR);
817    IF_INIT(printk("iadev->LineRate = %d \n", iadev->LineRate);)
818 
819 }
820 
821 static void IaFrontEndIntr(IADEV *iadev) {
822   volatile IA_SUNI *suni;
823   volatile ia_mb25_t *mb25;
824   volatile suni_pm7345_t *suni_pm7345;
825   u32 intr_status;
826   u_int frmr_intr;
827 
828   if(iadev->phy_type & FE_25MBIT_PHY) {
829      mb25 = (ia_mb25_t*)iadev->phy;
830      iadev->carrier_detect =  Boolean(mb25->mb25_intr_status & MB25_IS_GSB);
831   } else if (iadev->phy_type & FE_DS3_PHY) {
832      suni_pm7345 = (suni_pm7345_t *)iadev->phy;
833      /* clear FRMR interrupts */
834      frmr_intr   = suni_pm7345->suni_ds3_frm_intr_stat;
835      iadev->carrier_detect =
836            Boolean(!(suni_pm7345->suni_ds3_frm_stat & SUNI_DS3_LOSV));
837   } else if (iadev->phy_type & FE_E3_PHY ) {
838      suni_pm7345 = (suni_pm7345_t *)iadev->phy;
839      frmr_intr   = suni_pm7345->suni_e3_frm_maint_intr_ind;
840      iadev->carrier_detect =
841            Boolean(!(suni_pm7345->suni_e3_frm_fram_intr_ind_stat&SUNI_E3_LOS));
842   }
843   else {
844      suni = (IA_SUNI *)iadev->phy;
845      intr_status = suni->suni_rsop_status & 0xff;
846      iadev->carrier_detect = Boolean(!(suni->suni_rsop_status & SUNI_LOSV));
847   }
848   if (iadev->carrier_detect)
849     printk("IA: SUNI carrier detected\n");
850   else
851     printk("IA: SUNI carrier lost signal\n");
852   return;
853 }
854 
855 static void ia_mb25_init (IADEV *iadev)
856 {
857    volatile ia_mb25_t  *mb25 = (ia_mb25_t*)iadev->phy;
858 #if 0
859    mb25->mb25_master_ctrl = MB25_MC_DRIC | MB25_MC_DREC | MB25_MC_ENABLED;
860 #endif
861    mb25->mb25_master_ctrl = MB25_MC_DRIC | MB25_MC_DREC;
862    mb25->mb25_diag_control = 0;
863    /*
864     * Initialize carrier detect state
865     */
866    iadev->carrier_detect =  Boolean(mb25->mb25_intr_status & MB25_IS_GSB);
867    return;
868 }
869 
870 static void ia_suni_pm7345_init (IADEV *iadev)
871 {
872    volatile suni_pm7345_t *suni_pm7345 = (suni_pm7345_t *)iadev->phy;
873    if (iadev->phy_type & FE_DS3_PHY)
874    {
875       iadev->carrier_detect =
876           Boolean(!(suni_pm7345->suni_ds3_frm_stat & SUNI_DS3_LOSV));
877       suni_pm7345->suni_ds3_frm_intr_enbl = 0x17;
878       suni_pm7345->suni_ds3_frm_cfg = 1;
879       suni_pm7345->suni_ds3_tran_cfg = 1;
880       suni_pm7345->suni_config = 0;
881       suni_pm7345->suni_splr_cfg = 0;
882       suni_pm7345->suni_splt_cfg = 0;
883    }
884    else
885    {
886       iadev->carrier_detect =
887           Boolean(!(suni_pm7345->suni_e3_frm_fram_intr_ind_stat & SUNI_E3_LOS));
888       suni_pm7345->suni_e3_frm_fram_options = 0x4;
889       suni_pm7345->suni_e3_frm_maint_options = 0x20;
890       suni_pm7345->suni_e3_frm_fram_intr_enbl = 0x1d;
891       suni_pm7345->suni_e3_frm_maint_intr_enbl = 0x30;
892       suni_pm7345->suni_e3_tran_stat_diag_options = 0x0;
893       suni_pm7345->suni_e3_tran_fram_options = 0x1;
894       suni_pm7345->suni_config = SUNI_PM7345_E3ENBL;
895       suni_pm7345->suni_splr_cfg = 0x41;
896       suni_pm7345->suni_splt_cfg = 0x41;
897    }
898    /*
899     * Enable RSOP loss of signal interrupt.
900     */
901    suni_pm7345->suni_intr_enbl = 0x28;
902 
903    /*
904     * Clear error counters
905     */
906    suni_pm7345->suni_id_reset = 0;
907 
908    /*
909     * Clear "PMCTST" in master test register.
910     */
911    suni_pm7345->suni_master_test = 0;
912 
913    suni_pm7345->suni_rxcp_ctrl = 0x2c;
914    suni_pm7345->suni_rxcp_fctrl = 0x81;
915 
916    suni_pm7345->suni_rxcp_idle_pat_h1 =
917    	suni_pm7345->suni_rxcp_idle_pat_h2 =
918    	suni_pm7345->suni_rxcp_idle_pat_h3 = 0;
919    suni_pm7345->suni_rxcp_idle_pat_h4 = 1;
920 
921    suni_pm7345->suni_rxcp_idle_mask_h1 = 0xff;
922    suni_pm7345->suni_rxcp_idle_mask_h2 = 0xff;
923    suni_pm7345->suni_rxcp_idle_mask_h3 = 0xff;
924    suni_pm7345->suni_rxcp_idle_mask_h4 = 0xfe;
925 
926    suni_pm7345->suni_rxcp_cell_pat_h1 =
927    	suni_pm7345->suni_rxcp_cell_pat_h2 =
928    	suni_pm7345->suni_rxcp_cell_pat_h3 = 0;
929    suni_pm7345->suni_rxcp_cell_pat_h4 = 1;
930 
931    suni_pm7345->suni_rxcp_cell_mask_h1 =
932    	suni_pm7345->suni_rxcp_cell_mask_h2 =
933    	suni_pm7345->suni_rxcp_cell_mask_h3 =
934    	suni_pm7345->suni_rxcp_cell_mask_h4 = 0xff;
935 
936    suni_pm7345->suni_txcp_ctrl = 0xa4;
937    suni_pm7345->suni_txcp_intr_en_sts = 0x10;
938    suni_pm7345->suni_txcp_idle_pat_h5 = 0x55;
939 
940    suni_pm7345->suni_config &= ~(SUNI_PM7345_LLB |
941                                  SUNI_PM7345_CLB |
942                                  SUNI_PM7345_DLB |
943                                   SUNI_PM7345_PLB);
944 #ifdef __SNMP__
945    suni_pm7345->suni_rxcp_intr_en_sts |= SUNI_OOCDE;
946 #endif /* __SNMP__ */
947    return;
948 }
949 
950 
951 /***************************** IA_LIB END *****************************/
952 
953 #ifdef CONFIG_ATM_IA_DEBUG
954 static int tcnter = 0;
955 static void xdump( u_char*  cp, int  length, char*  prefix )
956 {
957     int col, count;
958     u_char prntBuf[120];
959     u_char*  pBuf = prntBuf;
960     count = 0;
961     while(count < length){
962         pBuf += sprintf( pBuf, "%s", prefix );
963         for(col = 0;count + col < length && col < 16; col++){
964             if (col != 0 && (col % 4) == 0)
965                 pBuf += sprintf( pBuf, " " );
966             pBuf += sprintf( pBuf, "%02X ", cp[count + col] );
967         }
968         while(col++ < 16){      /* pad end of buffer with blanks */
969             if ((col % 4) == 0)
970                 sprintf( pBuf, " " );
971             pBuf += sprintf( pBuf, "   " );
972         }
973         pBuf += sprintf( pBuf, "  " );
974         for(col = 0;count + col < length && col < 16; col++){
975             if (isprint((int)cp[count + col]))
976                 pBuf += sprintf( pBuf, "%c", cp[count + col] );
977             else
978                 pBuf += sprintf( pBuf, "." );
979                 }
980         printk("%s\n", prntBuf);
981         count += col;
982         pBuf = prntBuf;
983     }
984 
985 }  /* close xdump(... */
986 #endif /* CONFIG_ATM_IA_DEBUG */
987 
988 
989 static struct atm_dev *ia_boards = NULL;
990 
991 #define ACTUAL_RAM_BASE \
992 	RAM_BASE*((iadev->mem)/(128 * 1024))
993 #define ACTUAL_SEG_RAM_BASE \
994 	IPHASE5575_FRAG_CONTROL_RAM_BASE*((iadev->mem)/(128 * 1024))
995 #define ACTUAL_REASS_RAM_BASE \
996 	IPHASE5575_REASS_CONTROL_RAM_BASE*((iadev->mem)/(128 * 1024))
997 
998 
999 /*-- some utilities and memory allocation stuff will come here -------------*/
1000 
1001 static void desc_dbg(IADEV *iadev) {
1002 
1003   u_short tcq_wr_ptr, tcq_st_ptr, tcq_ed_ptr;
1004   u32 i;
1005   void __iomem *tmp;
1006   // regval = readl((u32)ia_cmds->maddr);
1007   tcq_wr_ptr =  readw(iadev->seg_reg+TCQ_WR_PTR);
1008   printk("B_tcq_wr = 0x%x desc = %d last desc = %d\n",
1009                      tcq_wr_ptr, readw(iadev->seg_ram+tcq_wr_ptr),
1010                      readw(iadev->seg_ram+tcq_wr_ptr-2));
1011   printk(" host_tcq_wr = 0x%x  host_tcq_rd = 0x%x \n",  iadev->host_tcq_wr,
1012                    iadev->ffL.tcq_rd);
1013   tcq_st_ptr =  readw(iadev->seg_reg+TCQ_ST_ADR);
1014   tcq_ed_ptr =  readw(iadev->seg_reg+TCQ_ED_ADR);
1015   printk("tcq_st_ptr = 0x%x    tcq_ed_ptr = 0x%x \n", tcq_st_ptr, tcq_ed_ptr);
1016   i = 0;
1017   while (tcq_st_ptr != tcq_ed_ptr) {
1018       tmp = iadev->seg_ram+tcq_st_ptr;
1019       printk("TCQ slot %d desc = %d  Addr = %p\n", i++, readw(tmp), tmp);
1020       tcq_st_ptr += 2;
1021   }
1022   for(i=0; i <iadev->num_tx_desc; i++)
1023       printk("Desc_tbl[%d] = %d \n", i, iadev->desc_tbl[i].timestamp);
1024 }
1025 
1026 
1027 /*----------------------------- Recieving side stuff --------------------------*/
1028 
1029 static void rx_excp_rcvd(struct atm_dev *dev)
1030 {
1031 #if 0 /* closing the receiving size will cause too many excp int */
1032   IADEV *iadev;
1033   u_short state;
1034   u_short excpq_rd_ptr;
1035   //u_short *ptr;
1036   int vci, error = 1;
1037   iadev = INPH_IA_DEV(dev);
1038   state = readl(iadev->reass_reg + STATE_REG) & 0xffff;
1039   while((state & EXCPQ_EMPTY) != EXCPQ_EMPTY)
1040   { printk("state = %x \n", state);
1041         excpq_rd_ptr = readw(iadev->reass_reg + EXCP_Q_RD_PTR) & 0xffff;
1042  printk("state = %x excpq_rd_ptr = %x \n", state, excpq_rd_ptr);
1043         if (excpq_rd_ptr == *(u16*)(iadev->reass_reg + EXCP_Q_WR_PTR))
1044             IF_ERR(printk("excpq_rd_ptr is wrong!!!\n");)
1045         // TODO: update exception stat
1046 	vci = readw(iadev->reass_ram+excpq_rd_ptr);
1047 	error = readw(iadev->reass_ram+excpq_rd_ptr+2) & 0x0007;
1048         // pwang_test
1049 	excpq_rd_ptr += 4;
1050 	if (excpq_rd_ptr > (readw(iadev->reass_reg + EXCP_Q_ED_ADR)& 0xffff))
1051  	    excpq_rd_ptr = readw(iadev->reass_reg + EXCP_Q_ST_ADR)& 0xffff;
1052 	writew( excpq_rd_ptr, iadev->reass_reg + EXCP_Q_RD_PTR);
1053         state = readl(iadev->reass_reg + STATE_REG) & 0xffff;
1054   }
1055 #endif
1056 }
1057 
1058 static void free_desc(struct atm_dev *dev, int desc)
1059 {
1060 	IADEV *iadev;
1061 	iadev = INPH_IA_DEV(dev);
1062         writew(desc, iadev->reass_ram+iadev->rfL.fdq_wr);
1063 	iadev->rfL.fdq_wr +=2;
1064 	if (iadev->rfL.fdq_wr > iadev->rfL.fdq_ed)
1065 		iadev->rfL.fdq_wr =  iadev->rfL.fdq_st;
1066 	writew(iadev->rfL.fdq_wr, iadev->reass_reg+FREEQ_WR_PTR);
1067 }
1068 
1069 
1070 static int rx_pkt(struct atm_dev *dev)
1071 {
1072 	IADEV *iadev;
1073 	struct atm_vcc *vcc;
1074 	unsigned short status;
1075 	struct rx_buf_desc __iomem *buf_desc_ptr;
1076 	int desc;
1077 	struct dle* wr_ptr;
1078 	int len;
1079 	struct sk_buff *skb;
1080 	u_int buf_addr, dma_addr;
1081 
1082 	iadev = INPH_IA_DEV(dev);
1083 	if (iadev->rfL.pcq_rd == (readw(iadev->reass_reg+PCQ_WR_PTR)&0xffff))
1084 	{
1085    	    printk(KERN_ERR DEV_LABEL "(itf %d) Receive queue empty\n", dev->number);
1086 	    return -EINVAL;
1087 	}
1088 	/* mask 1st 3 bits to get the actual descno. */
1089 	desc = readw(iadev->reass_ram+iadev->rfL.pcq_rd) & 0x1fff;
1090         IF_RX(printk("reass_ram = %p iadev->rfL.pcq_rd = 0x%x desc = %d\n",
1091                                     iadev->reass_ram, iadev->rfL.pcq_rd, desc);
1092               printk(" pcq_wr_ptr = 0x%x\n",
1093                                readw(iadev->reass_reg+PCQ_WR_PTR)&0xffff);)
1094 	/* update the read pointer  - maybe we shud do this in the end*/
1095 	if ( iadev->rfL.pcq_rd== iadev->rfL.pcq_ed)
1096 		iadev->rfL.pcq_rd = iadev->rfL.pcq_st;
1097 	else
1098 		iadev->rfL.pcq_rd += 2;
1099 	writew(iadev->rfL.pcq_rd, iadev->reass_reg+PCQ_RD_PTR);
1100 
1101 	/* get the buffer desc entry.
1102 		update stuff. - doesn't seem to be any update necessary
1103 	*/
1104 	buf_desc_ptr = iadev->RX_DESC_BASE_ADDR;
1105 	/* make the ptr point to the corresponding buffer desc entry */
1106 	buf_desc_ptr += desc;
1107         if (!desc || (desc > iadev->num_rx_desc) ||
1108                       ((buf_desc_ptr->vc_index & 0xffff) > iadev->num_vc)) {
1109             free_desc(dev, desc);
1110             IF_ERR(printk("IA: bad descriptor desc = %d \n", desc);)
1111             return -1;
1112         }
1113 	vcc = iadev->rx_open[buf_desc_ptr->vc_index & 0xffff];
1114 	if (!vcc)
1115 	{
1116                 free_desc(dev, desc);
1117 		printk("IA: null vcc, drop PDU\n");
1118 		return -1;
1119 	}
1120 
1121 
1122 	/* might want to check the status bits for errors */
1123 	status = (u_short) (buf_desc_ptr->desc_mode);
1124 	if (status & (RX_CER | RX_PTE | RX_OFL))
1125 	{
1126                 atomic_inc(&vcc->stats->rx_err);
1127 		IF_ERR(printk("IA: bad packet, dropping it");)
1128                 if (status & RX_CER) {
1129                     IF_ERR(printk(" cause: packet CRC error\n");)
1130                 }
1131                 else if (status & RX_PTE) {
1132                     IF_ERR(printk(" cause: packet time out\n");)
1133                 }
1134                 else {
1135                     IF_ERR(printk(" cause: buffer overflow\n");)
1136                 }
1137 		goto out_free_desc;
1138 	}
1139 
1140 	/*
1141 		build DLE.
1142 	*/
1143 
1144 	buf_addr = (buf_desc_ptr->buf_start_hi << 16) | buf_desc_ptr->buf_start_lo;
1145 	dma_addr = (buf_desc_ptr->dma_start_hi << 16) | buf_desc_ptr->dma_start_lo;
1146 	len = dma_addr - buf_addr;
1147         if (len > iadev->rx_buf_sz) {
1148            printk("Over %d bytes sdu received, dropped!!!\n", iadev->rx_buf_sz);
1149            atomic_inc(&vcc->stats->rx_err);
1150 	   goto out_free_desc;
1151         }
1152 
1153         if (!(skb = atm_alloc_charge(vcc, len, GFP_ATOMIC))) {
1154            if (vcc->vci < 32)
1155               printk("Drop control packets\n");
1156 	      goto out_free_desc;
1157         }
1158 	skb_put(skb,len);
1159         // pwang_test
1160         ATM_SKB(skb)->vcc = vcc;
1161         ATM_DESC(skb) = desc;
1162 	skb_queue_tail(&iadev->rx_dma_q, skb);
1163 
1164 	/* Build the DLE structure */
1165 	wr_ptr = iadev->rx_dle_q.write;
1166 	wr_ptr->sys_pkt_addr = pci_map_single(iadev->pci, skb->data,
1167 		len, PCI_DMA_FROMDEVICE);
1168 	wr_ptr->local_pkt_addr = buf_addr;
1169 	wr_ptr->bytes = len;	/* We don't know this do we ?? */
1170 	wr_ptr->mode = DMA_INT_ENABLE;
1171 
1172 	/* shud take care of wrap around here too. */
1173         if(++wr_ptr == iadev->rx_dle_q.end)
1174              wr_ptr = iadev->rx_dle_q.start;
1175 	iadev->rx_dle_q.write = wr_ptr;
1176 	udelay(1);
1177 	/* Increment transaction counter */
1178 	writel(1, iadev->dma+IPHASE5575_RX_COUNTER);
1179 out:	return 0;
1180 out_free_desc:
1181         free_desc(dev, desc);
1182         goto out;
1183 }
1184 
1185 static void rx_intr(struct atm_dev *dev)
1186 {
1187   IADEV *iadev;
1188   u_short status;
1189   u_short state, i;
1190 
1191   iadev = INPH_IA_DEV(dev);
1192   status = readl(iadev->reass_reg+REASS_INTR_STATUS_REG) & 0xffff;
1193   IF_EVENT(printk("rx_intr: status = 0x%x\n", status);)
1194   if (status & RX_PKT_RCVD)
1195   {
1196 	/* do something */
1197 	/* Basically recvd an interrupt for receving a packet.
1198 	A descriptor would have been written to the packet complete
1199 	queue. Get all the descriptors and set up dma to move the
1200 	packets till the packet complete queue is empty..
1201 	*/
1202 	state = readl(iadev->reass_reg + STATE_REG) & 0xffff;
1203         IF_EVENT(printk("Rx intr status: RX_PKT_RCVD %08x\n", status);)
1204 	while(!(state & PCQ_EMPTY))
1205 	{
1206              rx_pkt(dev);
1207 	     state = readl(iadev->reass_reg + STATE_REG) & 0xffff;
1208 	}
1209         iadev->rxing = 1;
1210   }
1211   if (status & RX_FREEQ_EMPT)
1212   {
1213      if (iadev->rxing) {
1214         iadev->rx_tmp_cnt = iadev->rx_pkt_cnt;
1215         iadev->rx_tmp_jif = jiffies;
1216         iadev->rxing = 0;
1217      }
1218      else if ((time_after(jiffies, iadev->rx_tmp_jif + 50)) &&
1219                ((iadev->rx_pkt_cnt - iadev->rx_tmp_cnt) == 0)) {
1220         for (i = 1; i <= iadev->num_rx_desc; i++)
1221                free_desc(dev, i);
1222 printk("Test logic RUN!!!!\n");
1223         writew( ~(RX_FREEQ_EMPT|RX_EXCP_RCVD),iadev->reass_reg+REASS_MASK_REG);
1224         iadev->rxing = 1;
1225      }
1226      IF_EVENT(printk("Rx intr status: RX_FREEQ_EMPT %08x\n", status);)
1227   }
1228 
1229   if (status & RX_EXCP_RCVD)
1230   {
1231 	/* probably need to handle the exception queue also. */
1232 	IF_EVENT(printk("Rx intr status: RX_EXCP_RCVD %08x\n", status);)
1233 	rx_excp_rcvd(dev);
1234   }
1235 
1236 
1237   if (status & RX_RAW_RCVD)
1238   {
1239 	/* need to handle the raw incoming cells. This deepnds on
1240 	whether we have programmed to receive the raw cells or not.
1241 	Else ignore. */
1242 	IF_EVENT(printk("Rx intr status:  RX_RAW_RCVD %08x\n", status);)
1243   }
1244 }
1245 
1246 
1247 static void rx_dle_intr(struct atm_dev *dev)
1248 {
1249   IADEV *iadev;
1250   struct atm_vcc *vcc;
1251   struct sk_buff *skb;
1252   int desc;
1253   u_short state;
1254   struct dle *dle, *cur_dle;
1255   u_int dle_lp;
1256   int len;
1257   iadev = INPH_IA_DEV(dev);
1258 
1259   /* free all the dles done, that is just update our own dle read pointer
1260 	- do we really need to do this. Think not. */
1261   /* DMA is done, just get all the recevie buffers from the rx dma queue
1262 	and push them up to the higher layer protocol. Also free the desc
1263 	associated with the buffer. */
1264   dle = iadev->rx_dle_q.read;
1265   dle_lp = readl(iadev->dma+IPHASE5575_RX_LIST_ADDR) & (sizeof(struct dle)*DLE_ENTRIES - 1);
1266   cur_dle = (struct dle*)(iadev->rx_dle_q.start + (dle_lp >> 4));
1267   while(dle != cur_dle)
1268   {
1269       /* free the DMAed skb */
1270       skb = skb_dequeue(&iadev->rx_dma_q);
1271       if (!skb)
1272          goto INCR_DLE;
1273       desc = ATM_DESC(skb);
1274       free_desc(dev, desc);
1275 
1276       if (!(len = skb->len))
1277       {
1278           printk("rx_dle_intr: skb len 0\n");
1279 	  dev_kfree_skb_any(skb);
1280       }
1281       else
1282       {
1283           struct cpcs_trailer *trailer;
1284           u_short length;
1285           struct ia_vcc *ia_vcc;
1286 
1287 	  pci_unmap_single(iadev->pci, iadev->rx_dle_q.write->sys_pkt_addr,
1288 	  	len, PCI_DMA_FROMDEVICE);
1289           /* no VCC related housekeeping done as yet. lets see */
1290           vcc = ATM_SKB(skb)->vcc;
1291 	  if (!vcc) {
1292 	      printk("IA: null vcc\n");
1293               dev_kfree_skb_any(skb);
1294               goto INCR_DLE;
1295           }
1296           ia_vcc = INPH_IA_VCC(vcc);
1297           if (ia_vcc == NULL)
1298           {
1299              atomic_inc(&vcc->stats->rx_err);
1300              dev_kfree_skb_any(skb);
1301              atm_return(vcc, atm_guess_pdu2truesize(len));
1302              goto INCR_DLE;
1303            }
1304           // get real pkt length  pwang_test
1305           trailer = (struct cpcs_trailer*)((u_char *)skb->data +
1306                                  skb->len - sizeof(*trailer));
1307 	  length = swap_byte_order(trailer->length);
1308           if ((length > iadev->rx_buf_sz) || (length >
1309                               (skb->len - sizeof(struct cpcs_trailer))))
1310           {
1311              atomic_inc(&vcc->stats->rx_err);
1312              IF_ERR(printk("rx_dle_intr: Bad  AAL5 trailer %d (skb len %d)",
1313                                                             length, skb->len);)
1314              dev_kfree_skb_any(skb);
1315              atm_return(vcc, atm_guess_pdu2truesize(len));
1316              goto INCR_DLE;
1317           }
1318           skb_trim(skb, length);
1319 
1320 	  /* Display the packet */
1321 	  IF_RXPKT(printk("\nDmad Recvd data: len = %d \n", skb->len);
1322           xdump(skb->data, skb->len, "RX: ");
1323           printk("\n");)
1324 
1325 	  IF_RX(printk("rx_dle_intr: skb push");)
1326 	  vcc->push(vcc,skb);
1327 	  atomic_inc(&vcc->stats->rx);
1328           iadev->rx_pkt_cnt++;
1329       }
1330 INCR_DLE:
1331       if (++dle == iadev->rx_dle_q.end)
1332     	  dle = iadev->rx_dle_q.start;
1333   }
1334   iadev->rx_dle_q.read = dle;
1335 
1336   /* if the interrupts are masked because there were no free desc available,
1337 		unmask them now. */
1338   if (!iadev->rxing) {
1339      state = readl(iadev->reass_reg + STATE_REG) & 0xffff;
1340      if (!(state & FREEQ_EMPTY)) {
1341         state = readl(iadev->reass_reg + REASS_MASK_REG) & 0xffff;
1342         writel(state & ~(RX_FREEQ_EMPT |/* RX_EXCP_RCVD |*/ RX_PKT_RCVD),
1343                                       iadev->reass_reg+REASS_MASK_REG);
1344         iadev->rxing++;
1345      }
1346   }
1347 }
1348 
1349 
1350 static int open_rx(struct atm_vcc *vcc)
1351 {
1352 	IADEV *iadev;
1353 	u_short __iomem *vc_table;
1354 	u_short __iomem *reass_ptr;
1355 	IF_EVENT(printk("iadev: open_rx %d.%d\n", vcc->vpi, vcc->vci);)
1356 
1357 	if (vcc->qos.rxtp.traffic_class == ATM_NONE) return 0;
1358 	iadev = INPH_IA_DEV(vcc->dev);
1359         if (vcc->qos.rxtp.traffic_class == ATM_ABR) {
1360            if (iadev->phy_type & FE_25MBIT_PHY) {
1361                printk("IA:  ABR not support\n");
1362                return -EINVAL;
1363            }
1364         }
1365 	/* Make only this VCI in the vc table valid and let all
1366 		others be invalid entries */
1367 	vc_table = iadev->reass_ram+RX_VC_TABLE*iadev->memSize;
1368 	vc_table += vcc->vci;
1369 	/* mask the last 6 bits and OR it with 3 for 1K VCs */
1370 
1371         *vc_table = vcc->vci << 6;
1372 	/* Also keep a list of open rx vcs so that we can attach them with
1373 		incoming PDUs later. */
1374 	if ((vcc->qos.rxtp.traffic_class == ATM_ABR) ||
1375                                 (vcc->qos.txtp.traffic_class == ATM_ABR))
1376 	{
1377                 srv_cls_param_t srv_p;
1378                 init_abr_vc(iadev, &srv_p);
1379                 ia_open_abr_vc(iadev, &srv_p, vcc, 0);
1380 	}
1381        	else {  /* for UBR  later may need to add CBR logic */
1382         	reass_ptr = iadev->reass_ram+REASS_TABLE*iadev->memSize;
1383            	reass_ptr += vcc->vci;
1384            	*reass_ptr = NO_AAL5_PKT;
1385        	}
1386 
1387 	if (iadev->rx_open[vcc->vci])
1388 		printk(KERN_CRIT DEV_LABEL "(itf %d): VCI %d already open\n",
1389 			vcc->dev->number, vcc->vci);
1390 	iadev->rx_open[vcc->vci] = vcc;
1391 	return 0;
1392 }
1393 
1394 static int rx_init(struct atm_dev *dev)
1395 {
1396 	IADEV *iadev;
1397 	struct rx_buf_desc __iomem *buf_desc_ptr;
1398 	unsigned long rx_pkt_start = 0;
1399 	void *dle_addr;
1400 	struct abr_vc_table  *abr_vc_table;
1401 	u16 *vc_table;
1402 	u16 *reass_table;
1403 	int i,j, vcsize_sel;
1404 	u_short freeq_st_adr;
1405 	u_short *freeq_start;
1406 
1407 	iadev = INPH_IA_DEV(dev);
1408   //    spin_lock_init(&iadev->rx_lock);
1409 
1410 	/* Allocate 4k bytes - more aligned than needed (4k boundary) */
1411 	dle_addr = pci_alloc_consistent(iadev->pci, DLE_TOTAL_SIZE,
1412 					&iadev->rx_dle_dma);
1413 	if (!dle_addr)  {
1414 		printk(KERN_ERR DEV_LABEL "can't allocate DLEs\n");
1415 		goto err_out;
1416 	}
1417 	iadev->rx_dle_q.start = (struct dle *)dle_addr;
1418 	iadev->rx_dle_q.read = iadev->rx_dle_q.start;
1419 	iadev->rx_dle_q.write = iadev->rx_dle_q.start;
1420 	iadev->rx_dle_q.end = (struct dle*)((unsigned long)dle_addr+sizeof(struct dle)*DLE_ENTRIES);
1421 	/* the end of the dle q points to the entry after the last
1422 	DLE that can be used. */
1423 
1424 	/* write the upper 20 bits of the start address to rx list address register */
1425 	/* We know this is 32bit bus addressed so the following is safe */
1426 	writel(iadev->rx_dle_dma & 0xfffff000,
1427 	       iadev->dma + IPHASE5575_RX_LIST_ADDR);
1428 	IF_INIT(printk("Tx Dle list addr: 0x%p value: 0x%0x\n",
1429                       iadev->dma+IPHASE5575_TX_LIST_ADDR,
1430                       *(u32*)(iadev->dma+IPHASE5575_TX_LIST_ADDR));
1431 	printk("Rx Dle list addr: 0x%p value: 0x%0x\n",
1432                       iadev->dma+IPHASE5575_RX_LIST_ADDR,
1433                       *(u32*)(iadev->dma+IPHASE5575_RX_LIST_ADDR));)
1434 
1435 	writew(0xffff, iadev->reass_reg+REASS_MASK_REG);
1436 	writew(0, iadev->reass_reg+MODE_REG);
1437 	writew(RESET_REASS, iadev->reass_reg+REASS_COMMAND_REG);
1438 
1439 	/* Receive side control memory map
1440 	   -------------------------------
1441 
1442 		Buffer descr	0x0000 (736 - 23K)
1443 		VP Table	0x5c00 (256 - 512)
1444 		Except q	0x5e00 (128 - 512)
1445 		Free buffer q	0x6000 (1K - 2K)
1446 		Packet comp q	0x6800 (1K - 2K)
1447 		Reass Table	0x7000 (1K - 2K)
1448 		VC Table	0x7800 (1K - 2K)
1449 		ABR VC Table	0x8000 (1K - 32K)
1450 	*/
1451 
1452 	/* Base address for Buffer Descriptor Table */
1453 	writew(RX_DESC_BASE >> 16, iadev->reass_reg+REASS_DESC_BASE);
1454 	/* Set the buffer size register */
1455 	writew(iadev->rx_buf_sz, iadev->reass_reg+BUF_SIZE);
1456 
1457 	/* Initialize each entry in the Buffer Descriptor Table */
1458         iadev->RX_DESC_BASE_ADDR = iadev->reass_ram+RX_DESC_BASE*iadev->memSize;
1459 	buf_desc_ptr = iadev->RX_DESC_BASE_ADDR;
1460 	memset_io(buf_desc_ptr, 0, sizeof(*buf_desc_ptr));
1461 	buf_desc_ptr++;
1462 	rx_pkt_start = iadev->rx_pkt_ram;
1463 	for(i=1; i<=iadev->num_rx_desc; i++)
1464 	{
1465 		memset_io(buf_desc_ptr, 0, sizeof(*buf_desc_ptr));
1466 		buf_desc_ptr->buf_start_hi = rx_pkt_start >> 16;
1467 		buf_desc_ptr->buf_start_lo = rx_pkt_start & 0x0000ffff;
1468 		buf_desc_ptr++;
1469 		rx_pkt_start += iadev->rx_buf_sz;
1470 	}
1471 	IF_INIT(printk("Rx Buffer desc ptr: 0x%p\n", buf_desc_ptr);)
1472         i = FREE_BUF_DESC_Q*iadev->memSize;
1473 	writew(i >> 16,  iadev->reass_reg+REASS_QUEUE_BASE);
1474         writew(i, iadev->reass_reg+FREEQ_ST_ADR);
1475         writew(i+iadev->num_rx_desc*sizeof(u_short),
1476                                          iadev->reass_reg+FREEQ_ED_ADR);
1477         writew(i, iadev->reass_reg+FREEQ_RD_PTR);
1478         writew(i+iadev->num_rx_desc*sizeof(u_short),
1479                                         iadev->reass_reg+FREEQ_WR_PTR);
1480 	/* Fill the FREEQ with all the free descriptors. */
1481 	freeq_st_adr = readw(iadev->reass_reg+FREEQ_ST_ADR);
1482 	freeq_start = (u_short *)(iadev->reass_ram+freeq_st_adr);
1483 	for(i=1; i<=iadev->num_rx_desc; i++)
1484 	{
1485 		*freeq_start = (u_short)i;
1486 		freeq_start++;
1487 	}
1488 	IF_INIT(printk("freeq_start: 0x%p\n", freeq_start);)
1489         /* Packet Complete Queue */
1490         i = (PKT_COMP_Q * iadev->memSize) & 0xffff;
1491         writew(i, iadev->reass_reg+PCQ_ST_ADR);
1492         writew(i+iadev->num_vc*sizeof(u_short), iadev->reass_reg+PCQ_ED_ADR);
1493         writew(i, iadev->reass_reg+PCQ_RD_PTR);
1494         writew(i, iadev->reass_reg+PCQ_WR_PTR);
1495 
1496         /* Exception Queue */
1497         i = (EXCEPTION_Q * iadev->memSize) & 0xffff;
1498         writew(i, iadev->reass_reg+EXCP_Q_ST_ADR);
1499         writew(i + NUM_RX_EXCP * sizeof(RX_ERROR_Q),
1500                                              iadev->reass_reg+EXCP_Q_ED_ADR);
1501         writew(i, iadev->reass_reg+EXCP_Q_RD_PTR);
1502         writew(i, iadev->reass_reg+EXCP_Q_WR_PTR);
1503 
1504     	/* Load local copy of FREEQ and PCQ ptrs */
1505         iadev->rfL.fdq_st = readw(iadev->reass_reg+FREEQ_ST_ADR) & 0xffff;
1506        	iadev->rfL.fdq_ed = readw(iadev->reass_reg+FREEQ_ED_ADR) & 0xffff ;
1507 	iadev->rfL.fdq_rd = readw(iadev->reass_reg+FREEQ_RD_PTR) & 0xffff;
1508 	iadev->rfL.fdq_wr = readw(iadev->reass_reg+FREEQ_WR_PTR) & 0xffff;
1509         iadev->rfL.pcq_st = readw(iadev->reass_reg+PCQ_ST_ADR) & 0xffff;
1510 	iadev->rfL.pcq_ed = readw(iadev->reass_reg+PCQ_ED_ADR) & 0xffff;
1511 	iadev->rfL.pcq_rd = readw(iadev->reass_reg+PCQ_RD_PTR) & 0xffff;
1512 	iadev->rfL.pcq_wr = readw(iadev->reass_reg+PCQ_WR_PTR) & 0xffff;
1513 
1514         IF_INIT(printk("INIT:pcq_st:0x%x pcq_ed:0x%x pcq_rd:0x%x pcq_wr:0x%x",
1515               iadev->rfL.pcq_st, iadev->rfL.pcq_ed, iadev->rfL.pcq_rd,
1516               iadev->rfL.pcq_wr);)
1517 	/* just for check - no VP TBL */
1518 	/* VP Table */
1519 	/* writew(0x0b80, iadev->reass_reg+VP_LKUP_BASE); */
1520 	/* initialize VP Table for invalid VPIs
1521 		- I guess we can write all 1s or 0x000f in the entire memory
1522 		  space or something similar.
1523 	*/
1524 
1525 	/* This seems to work and looks right to me too !!! */
1526         i =  REASS_TABLE * iadev->memSize;
1527 	writew((i >> 3), iadev->reass_reg+REASS_TABLE_BASE);
1528  	/* initialize Reassembly table to I don't know what ???? */
1529 	reass_table = (u16 *)(iadev->reass_ram+i);
1530         j = REASS_TABLE_SZ * iadev->memSize;
1531 	for(i=0; i < j; i++)
1532 		*reass_table++ = NO_AAL5_PKT;
1533        i = 8*1024;
1534        vcsize_sel =  0;
1535        while (i != iadev->num_vc) {
1536           i /= 2;
1537           vcsize_sel++;
1538        }
1539        i = RX_VC_TABLE * iadev->memSize;
1540        writew(((i>>3) & 0xfff8) | vcsize_sel, iadev->reass_reg+VC_LKUP_BASE);
1541        vc_table = (u16 *)(iadev->reass_ram+RX_VC_TABLE*iadev->memSize);
1542         j = RX_VC_TABLE_SZ * iadev->memSize;
1543 	for(i = 0; i < j; i++)
1544 	{
1545 		/* shift the reassembly pointer by 3 + lower 3 bits of
1546 		vc_lkup_base register (=3 for 1K VCs) and the last byte
1547 		is those low 3 bits.
1548 		Shall program this later.
1549 		*/
1550 		*vc_table = (i << 6) | 15;	/* for invalid VCI */
1551 		vc_table++;
1552 	}
1553         /* ABR VC table */
1554         i =  ABR_VC_TABLE * iadev->memSize;
1555         writew(i >> 3, iadev->reass_reg+ABR_LKUP_BASE);
1556 
1557         i = ABR_VC_TABLE * iadev->memSize;
1558 	abr_vc_table = (struct abr_vc_table *)(iadev->reass_ram+i);
1559         j = REASS_TABLE_SZ * iadev->memSize;
1560         memset ((char*)abr_vc_table, 0, j * sizeof(*abr_vc_table));
1561     	for(i = 0; i < j; i++) {
1562 		abr_vc_table->rdf = 0x0003;
1563              	abr_vc_table->air = 0x5eb1;
1564 	       	abr_vc_table++;
1565         }
1566 
1567 	/* Initialize other registers */
1568 
1569 	/* VP Filter Register set for VC Reassembly only */
1570 	writew(0xff00, iadev->reass_reg+VP_FILTER);
1571         writew(0, iadev->reass_reg+XTRA_RM_OFFSET);
1572 	writew(0x1,  iadev->reass_reg+PROTOCOL_ID);
1573 
1574 	/* Packet Timeout Count  related Registers :
1575 	   Set packet timeout to occur in about 3 seconds
1576 	   Set Packet Aging Interval count register to overflow in about 4 us
1577  	*/
1578         writew(0xF6F8, iadev->reass_reg+PKT_TM_CNT );
1579 
1580         i = (j >> 6) & 0xFF;
1581         j += 2 * (j - 1);
1582         i |= ((j << 2) & 0xFF00);
1583         writew(i, iadev->reass_reg+TMOUT_RANGE);
1584 
1585         /* initiate the desc_tble */
1586         for(i=0; i<iadev->num_tx_desc;i++)
1587             iadev->desc_tbl[i].timestamp = 0;
1588 
1589 	/* to clear the interrupt status register - read it */
1590 	readw(iadev->reass_reg+REASS_INTR_STATUS_REG);
1591 
1592 	/* Mask Register - clear it */
1593 	writew(~(RX_FREEQ_EMPT|RX_PKT_RCVD), iadev->reass_reg+REASS_MASK_REG);
1594 
1595 	skb_queue_head_init(&iadev->rx_dma_q);
1596 	iadev->rx_free_desc_qhead = NULL;
1597 
1598 	iadev->rx_open = kzalloc(4 * iadev->num_vc, GFP_KERNEL);
1599 	if (!iadev->rx_open) {
1600 		printk(KERN_ERR DEV_LABEL "itf %d couldn't get free page\n",
1601 		dev->number);
1602 		goto err_free_dle;
1603 	}
1604 
1605         iadev->rxing = 1;
1606         iadev->rx_pkt_cnt = 0;
1607 	/* Mode Register */
1608 	writew(R_ONLINE, iadev->reass_reg+MODE_REG);
1609 	return 0;
1610 
1611 err_free_dle:
1612 	pci_free_consistent(iadev->pci, DLE_TOTAL_SIZE, iadev->rx_dle_q.start,
1613 			    iadev->rx_dle_dma);
1614 err_out:
1615 	return -ENOMEM;
1616 }
1617 
1618 
1619 /*
1620 	The memory map suggested in appendix A and the coding for it.
1621 	Keeping it around just in case we change our mind later.
1622 
1623 		Buffer descr	0x0000 (128 - 4K)
1624 		UBR sched	0x1000 (1K - 4K)
1625 		UBR Wait q	0x2000 (1K - 4K)
1626 		Commn queues	0x3000 Packet Ready, Trasmit comp(0x3100)
1627 					(128 - 256) each
1628 		extended VC	0x4000 (1K - 8K)
1629 		ABR sched	0x6000	and ABR wait queue (1K - 2K) each
1630 		CBR sched	0x7000 (as needed)
1631 		VC table	0x8000 (1K - 32K)
1632 */
1633 
1634 static void tx_intr(struct atm_dev *dev)
1635 {
1636 	IADEV *iadev;
1637 	unsigned short status;
1638         unsigned long flags;
1639 
1640 	iadev = INPH_IA_DEV(dev);
1641 
1642 	status = readl(iadev->seg_reg+SEG_INTR_STATUS_REG);
1643         if (status & TRANSMIT_DONE){
1644 
1645            IF_EVENT(printk("Tansmit Done Intr logic run\n");)
1646            spin_lock_irqsave(&iadev->tx_lock, flags);
1647            ia_tx_poll(iadev);
1648            spin_unlock_irqrestore(&iadev->tx_lock, flags);
1649            writew(TRANSMIT_DONE, iadev->seg_reg+SEG_INTR_STATUS_REG);
1650            if (iadev->close_pending)
1651                wake_up(&iadev->close_wait);
1652         }
1653 	if (status & TCQ_NOT_EMPTY)
1654 	{
1655 	    IF_EVENT(printk("TCQ_NOT_EMPTY int received\n");)
1656 	}
1657 }
1658 
1659 static void tx_dle_intr(struct atm_dev *dev)
1660 {
1661         IADEV *iadev;
1662         struct dle *dle, *cur_dle;
1663         struct sk_buff *skb;
1664         struct atm_vcc *vcc;
1665         struct ia_vcc  *iavcc;
1666         u_int dle_lp;
1667         unsigned long flags;
1668 
1669         iadev = INPH_IA_DEV(dev);
1670         spin_lock_irqsave(&iadev->tx_lock, flags);
1671         dle = iadev->tx_dle_q.read;
1672         dle_lp = readl(iadev->dma+IPHASE5575_TX_LIST_ADDR) &
1673                                         (sizeof(struct dle)*DLE_ENTRIES - 1);
1674         cur_dle = (struct dle*)(iadev->tx_dle_q.start + (dle_lp >> 4));
1675         while (dle != cur_dle)
1676         {
1677             /* free the DMAed skb */
1678             skb = skb_dequeue(&iadev->tx_dma_q);
1679             if (!skb) break;
1680 
1681 	    /* Revenge of the 2 dle (skb + trailer) used in ia_pkt_tx() */
1682 	    if (!((dle - iadev->tx_dle_q.start)%(2*sizeof(struct dle)))) {
1683 		pci_unmap_single(iadev->pci, dle->sys_pkt_addr, skb->len,
1684 				 PCI_DMA_TODEVICE);
1685 	    }
1686             vcc = ATM_SKB(skb)->vcc;
1687             if (!vcc) {
1688                   printk("tx_dle_intr: vcc is null\n");
1689 		  spin_unlock_irqrestore(&iadev->tx_lock, flags);
1690                   dev_kfree_skb_any(skb);
1691 
1692                   return;
1693             }
1694             iavcc = INPH_IA_VCC(vcc);
1695             if (!iavcc) {
1696                   printk("tx_dle_intr: iavcc is null\n");
1697 		  spin_unlock_irqrestore(&iadev->tx_lock, flags);
1698                   dev_kfree_skb_any(skb);
1699                   return;
1700             }
1701             if (vcc->qos.txtp.pcr >= iadev->rate_limit) {
1702                if ((vcc->pop) && (skb->len != 0))
1703                {
1704                  vcc->pop(vcc, skb);
1705                }
1706                else {
1707                  dev_kfree_skb_any(skb);
1708                }
1709             }
1710             else { /* Hold the rate-limited skb for flow control */
1711                IA_SKB_STATE(skb) |= IA_DLED;
1712                skb_queue_tail(&iavcc->txing_skb, skb);
1713             }
1714             IF_EVENT(printk("tx_dle_intr: enque skb = 0x%p \n", skb);)
1715             if (++dle == iadev->tx_dle_q.end)
1716                  dle = iadev->tx_dle_q.start;
1717         }
1718         iadev->tx_dle_q.read = dle;
1719         spin_unlock_irqrestore(&iadev->tx_lock, flags);
1720 }
1721 
1722 static int open_tx(struct atm_vcc *vcc)
1723 {
1724 	struct ia_vcc *ia_vcc;
1725 	IADEV *iadev;
1726 	struct main_vc *vc;
1727 	struct ext_vc *evc;
1728         int ret;
1729 	IF_EVENT(printk("iadev: open_tx entered vcc->vci = %d\n", vcc->vci);)
1730 	if (vcc->qos.txtp.traffic_class == ATM_NONE) return 0;
1731 	iadev = INPH_IA_DEV(vcc->dev);
1732 
1733         if (iadev->phy_type & FE_25MBIT_PHY) {
1734            if (vcc->qos.txtp.traffic_class == ATM_ABR) {
1735                printk("IA:  ABR not support\n");
1736                return -EINVAL;
1737            }
1738 	  if (vcc->qos.txtp.traffic_class == ATM_CBR) {
1739                printk("IA:  CBR not support\n");
1740                return -EINVAL;
1741           }
1742         }
1743         ia_vcc =  INPH_IA_VCC(vcc);
1744         memset((caddr_t)ia_vcc, 0, sizeof(*ia_vcc));
1745         if (vcc->qos.txtp.max_sdu >
1746                          (iadev->tx_buf_sz - sizeof(struct cpcs_trailer))){
1747            printk("IA:  SDU size over (%d) the configured SDU size %d\n",
1748 		  vcc->qos.txtp.max_sdu,iadev->tx_buf_sz);
1749 	   vcc->dev_data = NULL;
1750            kfree(ia_vcc);
1751            return -EINVAL;
1752         }
1753 	ia_vcc->vc_desc_cnt = 0;
1754         ia_vcc->txing = 1;
1755 
1756         /* find pcr */
1757         if (vcc->qos.txtp.max_pcr == ATM_MAX_PCR)
1758            vcc->qos.txtp.pcr = iadev->LineRate;
1759         else if ((vcc->qos.txtp.max_pcr == 0)&&( vcc->qos.txtp.pcr <= 0))
1760            vcc->qos.txtp.pcr = iadev->LineRate;
1761         else if ((vcc->qos.txtp.max_pcr > vcc->qos.txtp.pcr) && (vcc->qos.txtp.max_pcr> 0))
1762            vcc->qos.txtp.pcr = vcc->qos.txtp.max_pcr;
1763         if (vcc->qos.txtp.pcr > iadev->LineRate)
1764              vcc->qos.txtp.pcr = iadev->LineRate;
1765         ia_vcc->pcr = vcc->qos.txtp.pcr;
1766 
1767         if (ia_vcc->pcr > (iadev->LineRate / 6) ) ia_vcc->ltimeout = HZ / 10;
1768         else if (ia_vcc->pcr > (iadev->LineRate / 130)) ia_vcc->ltimeout = HZ;
1769         else if (ia_vcc->pcr <= 170) ia_vcc->ltimeout = 16 * HZ;
1770         else ia_vcc->ltimeout = 2700 * HZ  / ia_vcc->pcr;
1771         if (ia_vcc->pcr < iadev->rate_limit)
1772            skb_queue_head_init (&ia_vcc->txing_skb);
1773         if (ia_vcc->pcr < iadev->rate_limit) {
1774 	   struct sock *sk = sk_atm(vcc);
1775 
1776 	   if (vcc->qos.txtp.max_sdu != 0) {
1777                if (ia_vcc->pcr > 60000)
1778                   sk->sk_sndbuf = vcc->qos.txtp.max_sdu * 5;
1779                else if (ia_vcc->pcr > 2000)
1780                   sk->sk_sndbuf = vcc->qos.txtp.max_sdu * 4;
1781                else
1782                  sk->sk_sndbuf = vcc->qos.txtp.max_sdu * 3;
1783            }
1784            else
1785              sk->sk_sndbuf = 24576;
1786         }
1787 
1788 	vc = (struct main_vc *)iadev->MAIN_VC_TABLE_ADDR;
1789 	evc = (struct ext_vc *)iadev->EXT_VC_TABLE_ADDR;
1790 	vc += vcc->vci;
1791 	evc += vcc->vci;
1792 	memset((caddr_t)vc, 0, sizeof(*vc));
1793 	memset((caddr_t)evc, 0, sizeof(*evc));
1794 
1795 	/* store the most significant 4 bits of vci as the last 4 bits
1796 		of first part of atm header.
1797 	   store the last 12 bits of vci as first 12 bits of the second
1798 		part of the atm header.
1799 	*/
1800 	evc->atm_hdr1 = (vcc->vci >> 12) & 0x000f;
1801 	evc->atm_hdr2 = (vcc->vci & 0x0fff) << 4;
1802 
1803 	/* check the following for different traffic classes */
1804 	if (vcc->qos.txtp.traffic_class == ATM_UBR)
1805 	{
1806 		vc->type = UBR;
1807                 vc->status = CRC_APPEND;
1808 		vc->acr = cellrate_to_float(iadev->LineRate);
1809                 if (vcc->qos.txtp.pcr > 0)
1810                    vc->acr = cellrate_to_float(vcc->qos.txtp.pcr);
1811                 IF_UBR(printk("UBR: txtp.pcr = 0x%x f_rate = 0x%x\n",
1812                                              vcc->qos.txtp.max_pcr,vc->acr);)
1813 	}
1814 	else if (vcc->qos.txtp.traffic_class == ATM_ABR)
1815 	{       srv_cls_param_t srv_p;
1816 		IF_ABR(printk("Tx ABR VCC\n");)
1817                 init_abr_vc(iadev, &srv_p);
1818                 if (vcc->qos.txtp.pcr > 0)
1819                    srv_p.pcr = vcc->qos.txtp.pcr;
1820                 if (vcc->qos.txtp.min_pcr > 0) {
1821                    int tmpsum = iadev->sum_mcr+iadev->sum_cbr+vcc->qos.txtp.min_pcr;
1822                    if (tmpsum > iadev->LineRate)
1823                        return -EBUSY;
1824                    srv_p.mcr = vcc->qos.txtp.min_pcr;
1825                    iadev->sum_mcr += vcc->qos.txtp.min_pcr;
1826                 }
1827                 else srv_p.mcr = 0;
1828                 if (vcc->qos.txtp.icr)
1829                    srv_p.icr = vcc->qos.txtp.icr;
1830                 if (vcc->qos.txtp.tbe)
1831                    srv_p.tbe = vcc->qos.txtp.tbe;
1832                 if (vcc->qos.txtp.frtt)
1833                    srv_p.frtt = vcc->qos.txtp.frtt;
1834                 if (vcc->qos.txtp.rif)
1835                    srv_p.rif = vcc->qos.txtp.rif;
1836                 if (vcc->qos.txtp.rdf)
1837                    srv_p.rdf = vcc->qos.txtp.rdf;
1838                 if (vcc->qos.txtp.nrm_pres)
1839                    srv_p.nrm = vcc->qos.txtp.nrm;
1840                 if (vcc->qos.txtp.trm_pres)
1841                    srv_p.trm = vcc->qos.txtp.trm;
1842                 if (vcc->qos.txtp.adtf_pres)
1843                    srv_p.adtf = vcc->qos.txtp.adtf;
1844                 if (vcc->qos.txtp.cdf_pres)
1845                    srv_p.cdf = vcc->qos.txtp.cdf;
1846                 if (srv_p.icr > srv_p.pcr)
1847                    srv_p.icr = srv_p.pcr;
1848                 IF_ABR(printk("ABR:vcc->qos.txtp.max_pcr = %d  mcr = %d\n",
1849                                                       srv_p.pcr, srv_p.mcr);)
1850 		ia_open_abr_vc(iadev, &srv_p, vcc, 1);
1851 	} else if (vcc->qos.txtp.traffic_class == ATM_CBR) {
1852                 if (iadev->phy_type & FE_25MBIT_PHY) {
1853                     printk("IA:  CBR not support\n");
1854                     return -EINVAL;
1855                 }
1856                 if (vcc->qos.txtp.max_pcr > iadev->LineRate) {
1857                    IF_CBR(printk("PCR is not availble\n");)
1858                    return -1;
1859                 }
1860                 vc->type = CBR;
1861                 vc->status = CRC_APPEND;
1862                 if ((ret = ia_cbr_setup (iadev, vcc)) < 0) {
1863                     return ret;
1864                 }
1865        }
1866 	else
1867            printk("iadev:  Non UBR, ABR and CBR traffic not supportedn");
1868 
1869         iadev->testTable[vcc->vci]->vc_status |= VC_ACTIVE;
1870 	IF_EVENT(printk("ia open_tx returning \n");)
1871 	return 0;
1872 }
1873 
1874 
1875 static int tx_init(struct atm_dev *dev)
1876 {
1877 	IADEV *iadev;
1878 	struct tx_buf_desc *buf_desc_ptr;
1879 	unsigned int tx_pkt_start;
1880 	void *dle_addr;
1881 	int i;
1882 	u_short tcq_st_adr;
1883 	u_short *tcq_start;
1884 	u_short prq_st_adr;
1885 	u_short *prq_start;
1886 	struct main_vc *vc;
1887 	struct ext_vc *evc;
1888         u_short tmp16;
1889         u32 vcsize_sel;
1890 
1891 	iadev = INPH_IA_DEV(dev);
1892         spin_lock_init(&iadev->tx_lock);
1893 
1894 	IF_INIT(printk("Tx MASK REG: 0x%0x\n",
1895                                 readw(iadev->seg_reg+SEG_MASK_REG));)
1896 
1897 	/* Allocate 4k (boundary aligned) bytes */
1898 	dle_addr = pci_alloc_consistent(iadev->pci, DLE_TOTAL_SIZE,
1899 					&iadev->tx_dle_dma);
1900 	if (!dle_addr)  {
1901 		printk(KERN_ERR DEV_LABEL "can't allocate DLEs\n");
1902 		goto err_out;
1903 	}
1904 	iadev->tx_dle_q.start = (struct dle*)dle_addr;
1905 	iadev->tx_dle_q.read = iadev->tx_dle_q.start;
1906 	iadev->tx_dle_q.write = iadev->tx_dle_q.start;
1907 	iadev->tx_dle_q.end = (struct dle*)((unsigned long)dle_addr+sizeof(struct dle)*DLE_ENTRIES);
1908 
1909 	/* write the upper 20 bits of the start address to tx list address register */
1910 	writel(iadev->tx_dle_dma & 0xfffff000,
1911 	       iadev->dma + IPHASE5575_TX_LIST_ADDR);
1912 	writew(0xffff, iadev->seg_reg+SEG_MASK_REG);
1913 	writew(0, iadev->seg_reg+MODE_REG_0);
1914 	writew(RESET_SEG, iadev->seg_reg+SEG_COMMAND_REG);
1915         iadev->MAIN_VC_TABLE_ADDR = iadev->seg_ram+MAIN_VC_TABLE*iadev->memSize;
1916         iadev->EXT_VC_TABLE_ADDR = iadev->seg_ram+EXT_VC_TABLE*iadev->memSize;
1917         iadev->ABR_SCHED_TABLE_ADDR=iadev->seg_ram+ABR_SCHED_TABLE*iadev->memSize;
1918 
1919 	/*
1920 	   Transmit side control memory map
1921 	   --------------------------------
1922 	 Buffer descr 	0x0000 (128 - 4K)
1923 	 Commn queues	0x1000	Transmit comp, Packet ready(0x1400)
1924 					(512 - 1K) each
1925 					TCQ - 4K, PRQ - 5K
1926 	 CBR Table 	0x1800 (as needed) - 6K
1927 	 UBR Table	0x3000 (1K - 4K) - 12K
1928 	 UBR Wait queue	0x4000 (1K - 4K) - 16K
1929 	 ABR sched	0x5000	and ABR wait queue (1K - 2K) each
1930 				ABR Tbl - 20K, ABR Wq - 22K
1931 	 extended VC	0x6000 (1K - 8K) - 24K
1932 	 VC Table	0x8000 (1K - 32K) - 32K
1933 
1934 	Between 0x2000 (8K) and 0x3000 (12K) there is 4K space left for VBR Tbl
1935 	and Wait q, which can be allotted later.
1936 	*/
1937 
1938 	/* Buffer Descriptor Table Base address */
1939 	writew(TX_DESC_BASE, iadev->seg_reg+SEG_DESC_BASE);
1940 
1941 	/* initialize each entry in the buffer descriptor table */
1942 	buf_desc_ptr =(struct tx_buf_desc *)(iadev->seg_ram+TX_DESC_BASE);
1943 	memset((caddr_t)buf_desc_ptr, 0, sizeof(*buf_desc_ptr));
1944 	buf_desc_ptr++;
1945 	tx_pkt_start = TX_PACKET_RAM;
1946 	for(i=1; i<=iadev->num_tx_desc; i++)
1947 	{
1948 		memset((caddr_t)buf_desc_ptr, 0, sizeof(*buf_desc_ptr));
1949 		buf_desc_ptr->desc_mode = AAL5;
1950 		buf_desc_ptr->buf_start_hi = tx_pkt_start >> 16;
1951 		buf_desc_ptr->buf_start_lo = tx_pkt_start & 0x0000ffff;
1952 		buf_desc_ptr++;
1953 		tx_pkt_start += iadev->tx_buf_sz;
1954 	}
1955         iadev->tx_buf = kmalloc(iadev->num_tx_desc*sizeof(struct cpcs_trailer_desc), GFP_KERNEL);
1956         if (!iadev->tx_buf) {
1957             printk(KERN_ERR DEV_LABEL " couldn't get mem\n");
1958 	    goto err_free_dle;
1959         }
1960        	for (i= 0; i< iadev->num_tx_desc; i++)
1961        	{
1962 	    struct cpcs_trailer *cpcs;
1963 
1964        	    cpcs = kmalloc(sizeof(*cpcs), GFP_KERNEL|GFP_DMA);
1965             if(!cpcs) {
1966 		printk(KERN_ERR DEV_LABEL " couldn't get freepage\n");
1967 		goto err_free_tx_bufs;
1968             }
1969 	    iadev->tx_buf[i].cpcs = cpcs;
1970 	    iadev->tx_buf[i].dma_addr = pci_map_single(iadev->pci,
1971 		cpcs, sizeof(*cpcs), PCI_DMA_TODEVICE);
1972         }
1973         iadev->desc_tbl = kmalloc(iadev->num_tx_desc *
1974                                    sizeof(struct desc_tbl_t), GFP_KERNEL);
1975 	if (!iadev->desc_tbl) {
1976 		printk(KERN_ERR DEV_LABEL " couldn't get mem\n");
1977 		goto err_free_all_tx_bufs;
1978 	}
1979 
1980 	/* Communication Queues base address */
1981         i = TX_COMP_Q * iadev->memSize;
1982 	writew(i >> 16, iadev->seg_reg+SEG_QUEUE_BASE);
1983 
1984 	/* Transmit Complete Queue */
1985 	writew(i, iadev->seg_reg+TCQ_ST_ADR);
1986 	writew(i, iadev->seg_reg+TCQ_RD_PTR);
1987 	writew(i+iadev->num_tx_desc*sizeof(u_short),iadev->seg_reg+TCQ_WR_PTR);
1988 	iadev->host_tcq_wr = i + iadev->num_tx_desc*sizeof(u_short);
1989         writew(i+2 * iadev->num_tx_desc * sizeof(u_short),
1990                                               iadev->seg_reg+TCQ_ED_ADR);
1991 	/* Fill the TCQ with all the free descriptors. */
1992 	tcq_st_adr = readw(iadev->seg_reg+TCQ_ST_ADR);
1993 	tcq_start = (u_short *)(iadev->seg_ram+tcq_st_adr);
1994 	for(i=1; i<=iadev->num_tx_desc; i++)
1995 	{
1996 		*tcq_start = (u_short)i;
1997 		tcq_start++;
1998 	}
1999 
2000 	/* Packet Ready Queue */
2001         i = PKT_RDY_Q * iadev->memSize;
2002 	writew(i, iadev->seg_reg+PRQ_ST_ADR);
2003 	writew(i+2 * iadev->num_tx_desc * sizeof(u_short),
2004                                               iadev->seg_reg+PRQ_ED_ADR);
2005 	writew(i, iadev->seg_reg+PRQ_RD_PTR);
2006 	writew(i, iadev->seg_reg+PRQ_WR_PTR);
2007 
2008         /* Load local copy of PRQ and TCQ ptrs */
2009         iadev->ffL.prq_st = readw(iadev->seg_reg+PRQ_ST_ADR) & 0xffff;
2010 	iadev->ffL.prq_ed = readw(iadev->seg_reg+PRQ_ED_ADR) & 0xffff;
2011  	iadev->ffL.prq_wr = readw(iadev->seg_reg+PRQ_WR_PTR) & 0xffff;
2012 
2013 	iadev->ffL.tcq_st = readw(iadev->seg_reg+TCQ_ST_ADR) & 0xffff;
2014 	iadev->ffL.tcq_ed = readw(iadev->seg_reg+TCQ_ED_ADR) & 0xffff;
2015 	iadev->ffL.tcq_rd = readw(iadev->seg_reg+TCQ_RD_PTR) & 0xffff;
2016 
2017 	/* Just for safety initializing the queue to have desc 1 always */
2018 	/* Fill the PRQ with all the free descriptors. */
2019 	prq_st_adr = readw(iadev->seg_reg+PRQ_ST_ADR);
2020 	prq_start = (u_short *)(iadev->seg_ram+prq_st_adr);
2021 	for(i=1; i<=iadev->num_tx_desc; i++)
2022 	{
2023 		*prq_start = (u_short)0;	/* desc 1 in all entries */
2024 		prq_start++;
2025 	}
2026 	/* CBR Table */
2027         IF_INIT(printk("Start CBR Init\n");)
2028 #if 1  /* for 1K VC board, CBR_PTR_BASE is 0 */
2029         writew(0,iadev->seg_reg+CBR_PTR_BASE);
2030 #else /* Charlie's logic is wrong ? */
2031         tmp16 = (iadev->seg_ram+CBR_SCHED_TABLE*iadev->memSize)>>17;
2032         IF_INIT(printk("cbr_ptr_base = 0x%x ", tmp16);)
2033         writew(tmp16,iadev->seg_reg+CBR_PTR_BASE);
2034 #endif
2035 
2036         IF_INIT(printk("value in register = 0x%x\n",
2037                                    readw(iadev->seg_reg+CBR_PTR_BASE));)
2038         tmp16 = (CBR_SCHED_TABLE*iadev->memSize) >> 1;
2039         writew(tmp16, iadev->seg_reg+CBR_TAB_BEG);
2040         IF_INIT(printk("cbr_tab_beg = 0x%x in reg = 0x%x \n", tmp16,
2041                                         readw(iadev->seg_reg+CBR_TAB_BEG));)
2042         writew(tmp16, iadev->seg_reg+CBR_TAB_END+1); // CBR_PTR;
2043         tmp16 = (CBR_SCHED_TABLE*iadev->memSize + iadev->num_vc*6 - 2) >> 1;
2044         writew(tmp16, iadev->seg_reg+CBR_TAB_END);
2045         IF_INIT(printk("iadev->seg_reg = 0x%p CBR_PTR_BASE = 0x%x\n",
2046                iadev->seg_reg, readw(iadev->seg_reg+CBR_PTR_BASE));)
2047         IF_INIT(printk("CBR_TAB_BEG = 0x%x, CBR_TAB_END = 0x%x, CBR_PTR = 0x%x\n",
2048           readw(iadev->seg_reg+CBR_TAB_BEG), readw(iadev->seg_reg+CBR_TAB_END),
2049           readw(iadev->seg_reg+CBR_TAB_END+1));)
2050 
2051         /* Initialize the CBR Schedualing Table */
2052         memset_io(iadev->seg_ram+CBR_SCHED_TABLE*iadev->memSize,
2053                                                           0, iadev->num_vc*6);
2054         iadev->CbrRemEntries = iadev->CbrTotEntries = iadev->num_vc*3;
2055         iadev->CbrEntryPt = 0;
2056         iadev->Granularity = MAX_ATM_155 / iadev->CbrTotEntries;
2057         iadev->NumEnabledCBR = 0;
2058 
2059 	/* UBR scheduling Table and wait queue */
2060 	/* initialize all bytes of UBR scheduler table and wait queue to 0
2061 		- SCHEDSZ is 1K (# of entries).
2062 		- UBR Table size is 4K
2063 		- UBR wait queue is 4K
2064 	   since the table and wait queues are contiguous, all the bytes
2065 	   can be initialized by one memeset.
2066 	*/
2067 
2068         vcsize_sel = 0;
2069         i = 8*1024;
2070         while (i != iadev->num_vc) {
2071           i /= 2;
2072           vcsize_sel++;
2073         }
2074 
2075         i = MAIN_VC_TABLE * iadev->memSize;
2076         writew(vcsize_sel | ((i >> 8) & 0xfff8),iadev->seg_reg+VCT_BASE);
2077         i =  EXT_VC_TABLE * iadev->memSize;
2078         writew((i >> 8) & 0xfffe, iadev->seg_reg+VCTE_BASE);
2079         i = UBR_SCHED_TABLE * iadev->memSize;
2080         writew((i & 0xffff) >> 11,  iadev->seg_reg+UBR_SBPTR_BASE);
2081         i = UBR_WAIT_Q * iadev->memSize;
2082         writew((i >> 7) & 0xffff,  iadev->seg_reg+UBRWQ_BASE);
2083  	memset((caddr_t)(iadev->seg_ram+UBR_SCHED_TABLE*iadev->memSize),
2084                                                        0, iadev->num_vc*8);
2085 	/* ABR scheduling Table(0x5000-0x57ff) and wait queue(0x5800-0x5fff)*/
2086 	/* initialize all bytes of ABR scheduler table and wait queue to 0
2087 		- SCHEDSZ is 1K (# of entries).
2088 		- ABR Table size is 2K
2089 		- ABR wait queue is 2K
2090 	   since the table and wait queues are contiguous, all the bytes
2091 	   can be intialized by one memeset.
2092 	*/
2093         i = ABR_SCHED_TABLE * iadev->memSize;
2094         writew((i >> 11) & 0xffff, iadev->seg_reg+ABR_SBPTR_BASE);
2095         i = ABR_WAIT_Q * iadev->memSize;
2096         writew((i >> 7) & 0xffff, iadev->seg_reg+ABRWQ_BASE);
2097 
2098         i = ABR_SCHED_TABLE*iadev->memSize;
2099 	memset((caddr_t)(iadev->seg_ram+i),  0, iadev->num_vc*4);
2100 	vc = (struct main_vc *)iadev->MAIN_VC_TABLE_ADDR;
2101 	evc = (struct ext_vc *)iadev->EXT_VC_TABLE_ADDR;
2102         iadev->testTable = kmalloc(sizeof(long)*iadev->num_vc, GFP_KERNEL);
2103         if (!iadev->testTable) {
2104            printk("Get freepage  failed\n");
2105 	   goto err_free_desc_tbl;
2106         }
2107 	for(i=0; i<iadev->num_vc; i++)
2108 	{
2109 		memset((caddr_t)vc, 0, sizeof(*vc));
2110 		memset((caddr_t)evc, 0, sizeof(*evc));
2111                 iadev->testTable[i] = kmalloc(sizeof(struct testTable_t),
2112 						GFP_KERNEL);
2113 		if (!iadev->testTable[i])
2114 			goto err_free_test_tables;
2115               	iadev->testTable[i]->lastTime = 0;
2116  		iadev->testTable[i]->fract = 0;
2117                 iadev->testTable[i]->vc_status = VC_UBR;
2118 		vc++;
2119 		evc++;
2120 	}
2121 
2122 	/* Other Initialization */
2123 
2124 	/* Max Rate Register */
2125         if (iadev->phy_type & FE_25MBIT_PHY) {
2126 	   writew(RATE25, iadev->seg_reg+MAXRATE);
2127 	   writew((UBR_EN | (0x23 << 2)), iadev->seg_reg+STPARMS);
2128         }
2129         else {
2130 	   writew(cellrate_to_float(iadev->LineRate),iadev->seg_reg+MAXRATE);
2131 	   writew((UBR_EN | ABR_EN | (0x23 << 2)), iadev->seg_reg+STPARMS);
2132         }
2133 	/* Set Idle Header Reigisters to be sure */
2134 	writew(0, iadev->seg_reg+IDLEHEADHI);
2135 	writew(0, iadev->seg_reg+IDLEHEADLO);
2136 
2137 	/* Program ABR UBR Priority Register  as  PRI_ABR_UBR_EQUAL */
2138         writew(0xaa00, iadev->seg_reg+ABRUBR_ARB);
2139 
2140         iadev->close_pending = 0;
2141         init_waitqueue_head(&iadev->close_wait);
2142         init_waitqueue_head(&iadev->timeout_wait);
2143 	skb_queue_head_init(&iadev->tx_dma_q);
2144 	ia_init_rtn_q(&iadev->tx_return_q);
2145 
2146 	/* RM Cell Protocol ID and Message Type */
2147 	writew(RM_TYPE_4_0, iadev->seg_reg+RM_TYPE);
2148         skb_queue_head_init (&iadev->tx_backlog);
2149 
2150 	/* Mode Register 1 */
2151 	writew(MODE_REG_1_VAL, iadev->seg_reg+MODE_REG_1);
2152 
2153 	/* Mode Register 0 */
2154 	writew(T_ONLINE, iadev->seg_reg+MODE_REG_0);
2155 
2156 	/* Interrupt Status Register - read to clear */
2157 	readw(iadev->seg_reg+SEG_INTR_STATUS_REG);
2158 
2159 	/* Interrupt Mask Reg- don't mask TCQ_NOT_EMPTY interrupt generation */
2160         writew(~(TRANSMIT_DONE | TCQ_NOT_EMPTY), iadev->seg_reg+SEG_MASK_REG);
2161         writew(TRANSMIT_DONE, iadev->seg_reg+SEG_INTR_STATUS_REG);
2162         iadev->tx_pkt_cnt = 0;
2163         iadev->rate_limit = iadev->LineRate / 3;
2164 
2165 	return 0;
2166 
2167 err_free_test_tables:
2168 	while (--i >= 0)
2169 		kfree(iadev->testTable[i]);
2170 	kfree(iadev->testTable);
2171 err_free_desc_tbl:
2172 	kfree(iadev->desc_tbl);
2173 err_free_all_tx_bufs:
2174 	i = iadev->num_tx_desc;
2175 err_free_tx_bufs:
2176 	while (--i >= 0) {
2177 		struct cpcs_trailer_desc *desc = iadev->tx_buf + i;
2178 
2179 		pci_unmap_single(iadev->pci, desc->dma_addr,
2180 			sizeof(*desc->cpcs), PCI_DMA_TODEVICE);
2181 		kfree(desc->cpcs);
2182 	}
2183 	kfree(iadev->tx_buf);
2184 err_free_dle:
2185 	pci_free_consistent(iadev->pci, DLE_TOTAL_SIZE, iadev->tx_dle_q.start,
2186 			    iadev->tx_dle_dma);
2187 err_out:
2188 	return -ENOMEM;
2189 }
2190 
2191 static irqreturn_t ia_int(int irq, void *dev_id)
2192 {
2193    struct atm_dev *dev;
2194    IADEV *iadev;
2195    unsigned int status;
2196    int handled = 0;
2197 
2198    dev = dev_id;
2199    iadev = INPH_IA_DEV(dev);
2200    while( (status = readl(iadev->reg+IPHASE5575_BUS_STATUS_REG) & 0x7f))
2201    {
2202 	handled = 1;
2203         IF_EVENT(printk("ia_int: status = 0x%x\n", status);)
2204 	if (status & STAT_REASSINT)
2205 	{
2206 	   /* do something */
2207 	   IF_EVENT(printk("REASSINT Bus status reg: %08x\n", status);)
2208 	   rx_intr(dev);
2209 	}
2210 	if (status & STAT_DLERINT)
2211 	{
2212 	   /* Clear this bit by writing a 1 to it. */
2213 	   *(u_int *)(iadev->reg+IPHASE5575_BUS_STATUS_REG) = STAT_DLERINT;
2214 	   rx_dle_intr(dev);
2215 	}
2216 	if (status & STAT_SEGINT)
2217 	{
2218 	   /* do something */
2219            IF_EVENT(printk("IA: tx_intr \n");)
2220 	   tx_intr(dev);
2221 	}
2222 	if (status & STAT_DLETINT)
2223 	{
2224 	   *(u_int *)(iadev->reg+IPHASE5575_BUS_STATUS_REG) = STAT_DLETINT;
2225 	   tx_dle_intr(dev);
2226 	}
2227 	if (status & (STAT_FEINT | STAT_ERRINT | STAT_MARKINT))
2228 	{
2229            if (status & STAT_FEINT)
2230                IaFrontEndIntr(iadev);
2231 	}
2232    }
2233    return IRQ_RETVAL(handled);
2234 }
2235 
2236 
2237 
2238 /*----------------------------- entries --------------------------------*/
2239 static int get_esi(struct atm_dev *dev)
2240 {
2241 	IADEV *iadev;
2242 	int i;
2243 	u32 mac1;
2244 	u16 mac2;
2245 
2246 	iadev = INPH_IA_DEV(dev);
2247 	mac1 = cpu_to_be32(le32_to_cpu(readl(
2248 				iadev->reg+IPHASE5575_MAC1)));
2249 	mac2 = cpu_to_be16(le16_to_cpu(readl(iadev->reg+IPHASE5575_MAC2)));
2250 	IF_INIT(printk("ESI: 0x%08x%04x\n", mac1, mac2);)
2251 	for (i=0; i<MAC1_LEN; i++)
2252 		dev->esi[i] = mac1 >>(8*(MAC1_LEN-1-i));
2253 
2254 	for (i=0; i<MAC2_LEN; i++)
2255 		dev->esi[i+MAC1_LEN] = mac2 >>(8*(MAC2_LEN - 1 -i));
2256 	return 0;
2257 }
2258 
2259 static int reset_sar(struct atm_dev *dev)
2260 {
2261 	IADEV *iadev;
2262 	int i, error = 1;
2263 	unsigned int pci[64];
2264 
2265 	iadev = INPH_IA_DEV(dev);
2266 	for(i=0; i<64; i++)
2267 	  if ((error = pci_read_config_dword(iadev->pci,
2268 				i*4, &pci[i])) != PCIBIOS_SUCCESSFUL)
2269   	      return error;
2270 	writel(0, iadev->reg+IPHASE5575_EXT_RESET);
2271 	for(i=0; i<64; i++)
2272 	  if ((error = pci_write_config_dword(iadev->pci,
2273 					i*4, pci[i])) != PCIBIOS_SUCCESSFUL)
2274 	    return error;
2275 	udelay(5);
2276 	return 0;
2277 }
2278 
2279 
2280 static int __devinit ia_init(struct atm_dev *dev)
2281 {
2282 	IADEV *iadev;
2283 	unsigned long real_base;
2284 	void __iomem *base;
2285 	unsigned short command;
2286 	int error, i;
2287 
2288 	/* The device has been identified and registered. Now we read
2289 	   necessary configuration info like memory base address,
2290 	   interrupt number etc */
2291 
2292 	IF_INIT(printk(">ia_init\n");)
2293 	dev->ci_range.vpi_bits = 0;
2294 	dev->ci_range.vci_bits = NR_VCI_LD;
2295 
2296 	iadev = INPH_IA_DEV(dev);
2297 	real_base = pci_resource_start (iadev->pci, 0);
2298 	iadev->irq = iadev->pci->irq;
2299 
2300 	error = pci_read_config_word(iadev->pci, PCI_COMMAND, &command);
2301 	if (error) {
2302 		printk(KERN_ERR DEV_LABEL "(itf %d): init error 0x%x\n",
2303 				dev->number,error);
2304 		return -EINVAL;
2305 	}
2306 	IF_INIT(printk(DEV_LABEL "(itf %d): rev.%d,realbase=0x%lx,irq=%d\n",
2307 			dev->number, iadev->pci->revision, real_base, iadev->irq);)
2308 
2309 	/* find mapping size of board */
2310 
2311 	iadev->pci_map_size = pci_resource_len(iadev->pci, 0);
2312 
2313         if (iadev->pci_map_size == 0x100000){
2314           iadev->num_vc = 4096;
2315 	  dev->ci_range.vci_bits = NR_VCI_4K_LD;
2316           iadev->memSize = 4;
2317         }
2318         else if (iadev->pci_map_size == 0x40000) {
2319           iadev->num_vc = 1024;
2320           iadev->memSize = 1;
2321         }
2322         else {
2323            printk("Unknown pci_map_size = 0x%x\n", iadev->pci_map_size);
2324            return -EINVAL;
2325         }
2326 	IF_INIT(printk (DEV_LABEL "map size: %i\n", iadev->pci_map_size);)
2327 
2328 	/* enable bus mastering */
2329 	pci_set_master(iadev->pci);
2330 
2331 	/*
2332 	 * Delay at least 1us before doing any mem accesses (how 'bout 10?)
2333 	 */
2334 	udelay(10);
2335 
2336 	/* mapping the physical address to a virtual address in address space */
2337 	base = ioremap(real_base,iadev->pci_map_size);  /* ioremap is not resolved ??? */
2338 
2339 	if (!base)
2340 	{
2341 		printk(DEV_LABEL " (itf %d): can't set up page mapping\n",
2342 			    dev->number);
2343 		return error;
2344 	}
2345 	IF_INIT(printk(DEV_LABEL " (itf %d): rev.%d,base=%p,irq=%d\n",
2346 			dev->number, iadev->pci->revision, base, iadev->irq);)
2347 
2348 	/* filling the iphase dev structure */
2349 	iadev->mem = iadev->pci_map_size /2;
2350 	iadev->real_base = real_base;
2351 	iadev->base = base;
2352 
2353 	/* Bus Interface Control Registers */
2354 	iadev->reg = base + REG_BASE;
2355 	/* Segmentation Control Registers */
2356 	iadev->seg_reg = base + SEG_BASE;
2357 	/* Reassembly Control Registers */
2358 	iadev->reass_reg = base + REASS_BASE;
2359 	/* Front end/ DMA control registers */
2360 	iadev->phy = base + PHY_BASE;
2361 	iadev->dma = base + PHY_BASE;
2362 	/* RAM - Segmentation RAm and Reassembly RAM */
2363 	iadev->ram = base + ACTUAL_RAM_BASE;
2364 	iadev->seg_ram = base + ACTUAL_SEG_RAM_BASE;
2365 	iadev->reass_ram = base + ACTUAL_REASS_RAM_BASE;
2366 
2367 	/* lets print out the above */
2368 	IF_INIT(printk("Base addrs: %p %p %p \n %p %p %p %p\n",
2369           iadev->reg,iadev->seg_reg,iadev->reass_reg,
2370           iadev->phy, iadev->ram, iadev->seg_ram,
2371           iadev->reass_ram);)
2372 
2373 	/* lets try reading the MAC address */
2374 	error = get_esi(dev);
2375 	if (error) {
2376 	  iounmap(iadev->base);
2377 	  return error;
2378 	}
2379         printk("IA: ");
2380 	for (i=0; i < ESI_LEN; i++)
2381                 printk("%s%02X",i ? "-" : "",dev->esi[i]);
2382         printk("\n");
2383 
2384         /* reset SAR */
2385         if (reset_sar(dev)) {
2386 	   iounmap(iadev->base);
2387            printk("IA: reset SAR fail, please try again\n");
2388            return 1;
2389         }
2390 	return 0;
2391 }
2392 
2393 static void ia_update_stats(IADEV *iadev) {
2394     if (!iadev->carrier_detect)
2395         return;
2396     iadev->rx_cell_cnt += readw(iadev->reass_reg+CELL_CTR0)&0xffff;
2397     iadev->rx_cell_cnt += (readw(iadev->reass_reg+CELL_CTR1) & 0xffff) << 16;
2398     iadev->drop_rxpkt +=  readw(iadev->reass_reg + DRP_PKT_CNTR ) & 0xffff;
2399     iadev->drop_rxcell += readw(iadev->reass_reg + ERR_CNTR) & 0xffff;
2400     iadev->tx_cell_cnt += readw(iadev->seg_reg + CELL_CTR_LO_AUTO)&0xffff;
2401     iadev->tx_cell_cnt += (readw(iadev->seg_reg+CELL_CTR_HIGH_AUTO)&0xffff)<<16;
2402     return;
2403 }
2404 
2405 static void ia_led_timer(unsigned long arg) {
2406  	unsigned long flags;
2407   	static u_char blinking[8] = {0, 0, 0, 0, 0, 0, 0, 0};
2408         u_char i;
2409         static u32 ctrl_reg;
2410         for (i = 0; i < iadev_count; i++) {
2411            if (ia_dev[i]) {
2412 	      ctrl_reg = readl(ia_dev[i]->reg+IPHASE5575_BUS_CONTROL_REG);
2413 	      if (blinking[i] == 0) {
2414 		 blinking[i]++;
2415                  ctrl_reg &= (~CTRL_LED);
2416                  writel(ctrl_reg, ia_dev[i]->reg+IPHASE5575_BUS_CONTROL_REG);
2417                  ia_update_stats(ia_dev[i]);
2418               }
2419               else {
2420 		 blinking[i] = 0;
2421 		 ctrl_reg |= CTRL_LED;
2422                  writel(ctrl_reg, ia_dev[i]->reg+IPHASE5575_BUS_CONTROL_REG);
2423                  spin_lock_irqsave(&ia_dev[i]->tx_lock, flags);
2424                  if (ia_dev[i]->close_pending)
2425                     wake_up(&ia_dev[i]->close_wait);
2426                  ia_tx_poll(ia_dev[i]);
2427                  spin_unlock_irqrestore(&ia_dev[i]->tx_lock, flags);
2428               }
2429            }
2430         }
2431 	mod_timer(&ia_timer, jiffies + HZ / 4);
2432  	return;
2433 }
2434 
2435 static void ia_phy_put(struct atm_dev *dev, unsigned char value,
2436 	unsigned long addr)
2437 {
2438 	writel(value, INPH_IA_DEV(dev)->phy+addr);
2439 }
2440 
2441 static unsigned char ia_phy_get(struct atm_dev *dev, unsigned long addr)
2442 {
2443 	return readl(INPH_IA_DEV(dev)->phy+addr);
2444 }
2445 
2446 static void ia_free_tx(IADEV *iadev)
2447 {
2448 	int i;
2449 
2450 	kfree(iadev->desc_tbl);
2451 	for (i = 0; i < iadev->num_vc; i++)
2452 		kfree(iadev->testTable[i]);
2453 	kfree(iadev->testTable);
2454 	for (i = 0; i < iadev->num_tx_desc; i++) {
2455 		struct cpcs_trailer_desc *desc = iadev->tx_buf + i;
2456 
2457 		pci_unmap_single(iadev->pci, desc->dma_addr,
2458 			sizeof(*desc->cpcs), PCI_DMA_TODEVICE);
2459 		kfree(desc->cpcs);
2460 	}
2461 	kfree(iadev->tx_buf);
2462 	pci_free_consistent(iadev->pci, DLE_TOTAL_SIZE, iadev->tx_dle_q.start,
2463 			    iadev->tx_dle_dma);
2464 }
2465 
2466 static void ia_free_rx(IADEV *iadev)
2467 {
2468 	kfree(iadev->rx_open);
2469 	pci_free_consistent(iadev->pci, DLE_TOTAL_SIZE, iadev->rx_dle_q.start,
2470 			  iadev->rx_dle_dma);
2471 }
2472 
2473 static int __devinit ia_start(struct atm_dev *dev)
2474 {
2475 	IADEV *iadev;
2476 	int error;
2477 	unsigned char phy;
2478 	u32 ctrl_reg;
2479 	IF_EVENT(printk(">ia_start\n");)
2480 	iadev = INPH_IA_DEV(dev);
2481         if (request_irq(iadev->irq, &ia_int, IRQF_SHARED, DEV_LABEL, dev)) {
2482                 printk(KERN_ERR DEV_LABEL "(itf %d): IRQ%d is already in use\n",
2483                     dev->number, iadev->irq);
2484 		error = -EAGAIN;
2485 		goto err_out;
2486         }
2487         /* @@@ should release IRQ on error */
2488 	/* enabling memory + master */
2489         if ((error = pci_write_config_word(iadev->pci,
2490 				PCI_COMMAND,
2491 				PCI_COMMAND_MEMORY | PCI_COMMAND_MASTER )))
2492 	{
2493                 printk(KERN_ERR DEV_LABEL "(itf %d): can't enable memory+"
2494                     "master (0x%x)\n",dev->number, error);
2495 		error = -EIO;
2496 		goto err_free_irq;
2497         }
2498 	udelay(10);
2499 
2500 	/* Maybe we should reset the front end, initialize Bus Interface Control
2501 		Registers and see. */
2502 
2503 	IF_INIT(printk("Bus ctrl reg: %08x\n",
2504                             readl(iadev->reg+IPHASE5575_BUS_CONTROL_REG));)
2505 	ctrl_reg = readl(iadev->reg+IPHASE5575_BUS_CONTROL_REG);
2506 	ctrl_reg = (ctrl_reg & (CTRL_LED | CTRL_FE_RST))
2507 			| CTRL_B8
2508 			| CTRL_B16
2509 			| CTRL_B32
2510 			| CTRL_B48
2511 			| CTRL_B64
2512 			| CTRL_B128
2513 			| CTRL_ERRMASK
2514 			| CTRL_DLETMASK		/* shud be removed l8r */
2515 			| CTRL_DLERMASK
2516 			| CTRL_SEGMASK
2517 			| CTRL_REASSMASK
2518 			| CTRL_FEMASK
2519 			| CTRL_CSPREEMPT;
2520 
2521        writel(ctrl_reg, iadev->reg+IPHASE5575_BUS_CONTROL_REG);
2522 
2523 	IF_INIT(printk("Bus ctrl reg after initializing: %08x\n",
2524                            readl(iadev->reg+IPHASE5575_BUS_CONTROL_REG));
2525 	   printk("Bus status reg after init: %08x\n",
2526                             readl(iadev->reg+IPHASE5575_BUS_STATUS_REG));)
2527 
2528         ia_hw_type(iadev);
2529 	error = tx_init(dev);
2530 	if (error)
2531 		goto err_free_irq;
2532 	error = rx_init(dev);
2533 	if (error)
2534 		goto err_free_tx;
2535 
2536 	ctrl_reg = readl(iadev->reg+IPHASE5575_BUS_CONTROL_REG);
2537        	writel(ctrl_reg | CTRL_FE_RST, iadev->reg+IPHASE5575_BUS_CONTROL_REG);
2538 	IF_INIT(printk("Bus ctrl reg after initializing: %08x\n",
2539                                readl(iadev->reg+IPHASE5575_BUS_CONTROL_REG));)
2540         phy = 0; /* resolve compiler complaint */
2541         IF_INIT (
2542 	if ((phy=ia_phy_get(dev,0)) == 0x30)
2543 		printk("IA: pm5346,rev.%d\n",phy&0x0f);
2544 	else
2545 		printk("IA: utopia,rev.%0x\n",phy);)
2546 
2547 	if (iadev->phy_type &  FE_25MBIT_PHY)
2548            ia_mb25_init(iadev);
2549 	else if (iadev->phy_type & (FE_DS3_PHY | FE_E3_PHY))
2550            ia_suni_pm7345_init(iadev);
2551 	else {
2552 		error = suni_init(dev);
2553 		if (error)
2554 			goto err_free_rx;
2555 		if (dev->phy->start) {
2556 			error = dev->phy->start(dev);
2557 			if (error)
2558 				goto err_free_rx;
2559 		}
2560 		/* Get iadev->carrier_detect status */
2561 		IaFrontEndIntr(iadev);
2562 	}
2563 	return 0;
2564 
2565 err_free_rx:
2566 	ia_free_rx(iadev);
2567 err_free_tx:
2568 	ia_free_tx(iadev);
2569 err_free_irq:
2570 	free_irq(iadev->irq, dev);
2571 err_out:
2572 	return error;
2573 }
2574 
2575 static void ia_close(struct atm_vcc *vcc)
2576 {
2577 	DEFINE_WAIT(wait);
2578         u16 *vc_table;
2579         IADEV *iadev;
2580         struct ia_vcc *ia_vcc;
2581         struct sk_buff *skb = NULL;
2582         struct sk_buff_head tmp_tx_backlog, tmp_vcc_backlog;
2583         unsigned long closetime, flags;
2584 
2585         iadev = INPH_IA_DEV(vcc->dev);
2586         ia_vcc = INPH_IA_VCC(vcc);
2587 	if (!ia_vcc) return;
2588 
2589         IF_EVENT(printk("ia_close: ia_vcc->vc_desc_cnt = %d  vci = %d\n",
2590                                               ia_vcc->vc_desc_cnt,vcc->vci);)
2591 	clear_bit(ATM_VF_READY,&vcc->flags);
2592         skb_queue_head_init (&tmp_tx_backlog);
2593         skb_queue_head_init (&tmp_vcc_backlog);
2594         if (vcc->qos.txtp.traffic_class != ATM_NONE) {
2595            iadev->close_pending++;
2596 	   prepare_to_wait(&iadev->timeout_wait, &wait, TASK_UNINTERRUPTIBLE);
2597 	   schedule_timeout(50);
2598 	   finish_wait(&iadev->timeout_wait, &wait);
2599            spin_lock_irqsave(&iadev->tx_lock, flags);
2600            while((skb = skb_dequeue(&iadev->tx_backlog))) {
2601               if (ATM_SKB(skb)->vcc == vcc){
2602                  if (vcc->pop) vcc->pop(vcc, skb);
2603                  else dev_kfree_skb_any(skb);
2604               }
2605               else
2606                  skb_queue_tail(&tmp_tx_backlog, skb);
2607            }
2608            while((skb = skb_dequeue(&tmp_tx_backlog)))
2609              skb_queue_tail(&iadev->tx_backlog, skb);
2610            IF_EVENT(printk("IA TX Done decs_cnt = %d\n", ia_vcc->vc_desc_cnt);)
2611            closetime = 300000 / ia_vcc->pcr;
2612            if (closetime == 0)
2613               closetime = 1;
2614            spin_unlock_irqrestore(&iadev->tx_lock, flags);
2615            wait_event_timeout(iadev->close_wait, (ia_vcc->vc_desc_cnt <= 0), closetime);
2616            spin_lock_irqsave(&iadev->tx_lock, flags);
2617            iadev->close_pending--;
2618            iadev->testTable[vcc->vci]->lastTime = 0;
2619            iadev->testTable[vcc->vci]->fract = 0;
2620            iadev->testTable[vcc->vci]->vc_status = VC_UBR;
2621            if (vcc->qos.txtp.traffic_class == ATM_ABR) {
2622               if (vcc->qos.txtp.min_pcr > 0)
2623                  iadev->sum_mcr -= vcc->qos.txtp.min_pcr;
2624            }
2625            if (vcc->qos.txtp.traffic_class == ATM_CBR) {
2626               ia_vcc = INPH_IA_VCC(vcc);
2627               iadev->sum_mcr -= ia_vcc->NumCbrEntry*iadev->Granularity;
2628               ia_cbrVc_close (vcc);
2629            }
2630            spin_unlock_irqrestore(&iadev->tx_lock, flags);
2631         }
2632 
2633         if (vcc->qos.rxtp.traffic_class != ATM_NONE) {
2634            // reset reass table
2635            vc_table = (u16 *)(iadev->reass_ram+REASS_TABLE*iadev->memSize);
2636            vc_table += vcc->vci;
2637            *vc_table = NO_AAL5_PKT;
2638            // reset vc table
2639            vc_table = (u16 *)(iadev->reass_ram+RX_VC_TABLE*iadev->memSize);
2640            vc_table += vcc->vci;
2641            *vc_table = (vcc->vci << 6) | 15;
2642            if (vcc->qos.rxtp.traffic_class == ATM_ABR) {
2643               struct abr_vc_table __iomem *abr_vc_table =
2644                                 (iadev->reass_ram+ABR_VC_TABLE*iadev->memSize);
2645               abr_vc_table +=  vcc->vci;
2646               abr_vc_table->rdf = 0x0003;
2647               abr_vc_table->air = 0x5eb1;
2648            }
2649            // Drain the packets
2650            rx_dle_intr(vcc->dev);
2651            iadev->rx_open[vcc->vci] = NULL;
2652         }
2653 	kfree(INPH_IA_VCC(vcc));
2654         ia_vcc = NULL;
2655         vcc->dev_data = NULL;
2656         clear_bit(ATM_VF_ADDR,&vcc->flags);
2657         return;
2658 }
2659 
2660 static int ia_open(struct atm_vcc *vcc)
2661 {
2662 	IADEV *iadev;
2663 	struct ia_vcc *ia_vcc;
2664 	int error;
2665 	if (!test_bit(ATM_VF_PARTIAL,&vcc->flags))
2666 	{
2667 		IF_EVENT(printk("ia: not partially allocated resources\n");)
2668 		vcc->dev_data = NULL;
2669 	}
2670 	iadev = INPH_IA_DEV(vcc->dev);
2671 	if (vcc->vci != ATM_VPI_UNSPEC && vcc->vpi != ATM_VCI_UNSPEC)
2672 	{
2673 		IF_EVENT(printk("iphase open: unspec part\n");)
2674 		set_bit(ATM_VF_ADDR,&vcc->flags);
2675 	}
2676 	if (vcc->qos.aal != ATM_AAL5)
2677 		return -EINVAL;
2678 	IF_EVENT(printk(DEV_LABEL "(itf %d): open %d.%d\n",
2679                                  vcc->dev->number, vcc->vpi, vcc->vci);)
2680 
2681 	/* Device dependent initialization */
2682 	ia_vcc = kmalloc(sizeof(*ia_vcc), GFP_KERNEL);
2683 	if (!ia_vcc) return -ENOMEM;
2684 	vcc->dev_data = ia_vcc;
2685 
2686 	if ((error = open_rx(vcc)))
2687 	{
2688 		IF_EVENT(printk("iadev: error in open_rx, closing\n");)
2689 		ia_close(vcc);
2690 		return error;
2691 	}
2692 
2693 	if ((error = open_tx(vcc)))
2694 	{
2695 		IF_EVENT(printk("iadev: error in open_tx, closing\n");)
2696 		ia_close(vcc);
2697 		return error;
2698 	}
2699 
2700 	set_bit(ATM_VF_READY,&vcc->flags);
2701 
2702 #if 0
2703         {
2704            static u8 first = 1;
2705            if (first) {
2706               ia_timer.expires = jiffies + 3*HZ;
2707               add_timer(&ia_timer);
2708               first = 0;
2709            }
2710         }
2711 #endif
2712 	IF_EVENT(printk("ia open returning\n");)
2713 	return 0;
2714 }
2715 
2716 static int ia_change_qos(struct atm_vcc *vcc, struct atm_qos *qos, int flags)
2717 {
2718 	IF_EVENT(printk(">ia_change_qos\n");)
2719 	return 0;
2720 }
2721 
2722 static int ia_ioctl(struct atm_dev *dev, unsigned int cmd, void __user *arg)
2723 {
2724    IA_CMDBUF ia_cmds;
2725    IADEV *iadev;
2726    int i, board;
2727    u16 __user *tmps;
2728    IF_EVENT(printk(">ia_ioctl\n");)
2729    if (cmd != IA_CMD) {
2730       if (!dev->phy->ioctl) return -EINVAL;
2731       return dev->phy->ioctl(dev,cmd,arg);
2732    }
2733    if (copy_from_user(&ia_cmds, arg, sizeof ia_cmds)) return -EFAULT;
2734    board = ia_cmds.status;
2735    if ((board < 0) || (board > iadev_count))
2736          board = 0;
2737    iadev = ia_dev[board];
2738    switch (ia_cmds.cmd) {
2739    case MEMDUMP:
2740    {
2741 	switch (ia_cmds.sub_cmd) {
2742        	  case MEMDUMP_DEV:
2743 	     if (!capable(CAP_NET_ADMIN)) return -EPERM;
2744 	     if (copy_to_user(ia_cmds.buf, iadev, sizeof(IADEV)))
2745                 return -EFAULT;
2746              ia_cmds.status = 0;
2747              break;
2748           case MEMDUMP_SEGREG:
2749 	     if (!capable(CAP_NET_ADMIN)) return -EPERM;
2750              tmps = (u16 __user *)ia_cmds.buf;
2751              for(i=0; i<0x80; i+=2, tmps++)
2752                 if(put_user((u16)(readl(iadev->seg_reg+i) & 0xffff), tmps)) return -EFAULT;
2753              ia_cmds.status = 0;
2754              ia_cmds.len = 0x80;
2755              break;
2756           case MEMDUMP_REASSREG:
2757 	     if (!capable(CAP_NET_ADMIN)) return -EPERM;
2758              tmps = (u16 __user *)ia_cmds.buf;
2759              for(i=0; i<0x80; i+=2, tmps++)
2760                 if(put_user((u16)(readl(iadev->reass_reg+i) & 0xffff), tmps)) return -EFAULT;
2761              ia_cmds.status = 0;
2762              ia_cmds.len = 0x80;
2763              break;
2764           case MEMDUMP_FFL:
2765           {
2766              ia_regs_t       *regs_local;
2767              ffredn_t        *ffL;
2768              rfredn_t        *rfL;
2769 
2770 	     if (!capable(CAP_NET_ADMIN)) return -EPERM;
2771 	     regs_local = kmalloc(sizeof(*regs_local), GFP_KERNEL);
2772 	     if (!regs_local) return -ENOMEM;
2773 	     ffL = &regs_local->ffredn;
2774 	     rfL = &regs_local->rfredn;
2775              /* Copy real rfred registers into the local copy */
2776  	     for (i=0; i<(sizeof (rfredn_t))/4; i++)
2777                 ((u_int *)rfL)[i] = readl(iadev->reass_reg + i) & 0xffff;
2778              	/* Copy real ffred registers into the local copy */
2779 	     for (i=0; i<(sizeof (ffredn_t))/4; i++)
2780                 ((u_int *)ffL)[i] = readl(iadev->seg_reg + i) & 0xffff;
2781 
2782              if (copy_to_user(ia_cmds.buf, regs_local,sizeof(ia_regs_t))) {
2783                 kfree(regs_local);
2784                 return -EFAULT;
2785              }
2786              kfree(regs_local);
2787              printk("Board %d registers dumped\n", board);
2788              ia_cmds.status = 0;
2789 	 }
2790     	     break;
2791          case READ_REG:
2792          {
2793 	     if (!capable(CAP_NET_ADMIN)) return -EPERM;
2794              desc_dbg(iadev);
2795              ia_cmds.status = 0;
2796          }
2797              break;
2798          case 0x6:
2799          {
2800              ia_cmds.status = 0;
2801              printk("skb = 0x%lx\n", (long)skb_peek(&iadev->tx_backlog));
2802              printk("rtn_q: 0x%lx\n",(long)ia_deque_rtn_q(&iadev->tx_return_q));
2803          }
2804              break;
2805          case 0x8:
2806          {
2807              struct k_sonet_stats *stats;
2808              stats = &PRIV(_ia_dev[board])->sonet_stats;
2809              printk("section_bip: %d\n", atomic_read(&stats->section_bip));
2810              printk("line_bip   : %d\n", atomic_read(&stats->line_bip));
2811              printk("path_bip   : %d\n", atomic_read(&stats->path_bip));
2812              printk("line_febe  : %d\n", atomic_read(&stats->line_febe));
2813              printk("path_febe  : %d\n", atomic_read(&stats->path_febe));
2814              printk("corr_hcs   : %d\n", atomic_read(&stats->corr_hcs));
2815              printk("uncorr_hcs : %d\n", atomic_read(&stats->uncorr_hcs));
2816              printk("tx_cells   : %d\n", atomic_read(&stats->tx_cells));
2817              printk("rx_cells   : %d\n", atomic_read(&stats->rx_cells));
2818          }
2819             ia_cmds.status = 0;
2820             break;
2821          case 0x9:
2822 	    if (!capable(CAP_NET_ADMIN)) return -EPERM;
2823             for (i = 1; i <= iadev->num_rx_desc; i++)
2824                free_desc(_ia_dev[board], i);
2825             writew( ~(RX_FREEQ_EMPT | RX_EXCP_RCVD),
2826                                             iadev->reass_reg+REASS_MASK_REG);
2827             iadev->rxing = 1;
2828 
2829             ia_cmds.status = 0;
2830             break;
2831 
2832          case 0xb:
2833 	    if (!capable(CAP_NET_ADMIN)) return -EPERM;
2834             IaFrontEndIntr(iadev);
2835             break;
2836          case 0xa:
2837 	    if (!capable(CAP_NET_ADMIN)) return -EPERM;
2838          {
2839              ia_cmds.status = 0;
2840              IADebugFlag = ia_cmds.maddr;
2841              printk("New debug option loaded\n");
2842          }
2843              break;
2844          default:
2845              ia_cmds.status = 0;
2846              break;
2847       }
2848    }
2849       break;
2850    default:
2851       break;
2852 
2853    }
2854    return 0;
2855 }
2856 
2857 static int ia_getsockopt(struct atm_vcc *vcc, int level, int optname,
2858 	void __user *optval, int optlen)
2859 {
2860 	IF_EVENT(printk(">ia_getsockopt\n");)
2861 	return -EINVAL;
2862 }
2863 
2864 static int ia_setsockopt(struct atm_vcc *vcc, int level, int optname,
2865 	void __user *optval, unsigned int optlen)
2866 {
2867 	IF_EVENT(printk(">ia_setsockopt\n");)
2868 	return -EINVAL;
2869 }
2870 
2871 static int ia_pkt_tx (struct atm_vcc *vcc, struct sk_buff *skb) {
2872         IADEV *iadev;
2873         struct dle *wr_ptr;
2874         struct tx_buf_desc __iomem *buf_desc_ptr;
2875         int desc;
2876         int comp_code;
2877         int total_len;
2878         struct cpcs_trailer *trailer;
2879         struct ia_vcc *iavcc;
2880 
2881         iadev = INPH_IA_DEV(vcc->dev);
2882         iavcc = INPH_IA_VCC(vcc);
2883         if (!iavcc->txing) {
2884            printk("discard packet on closed VC\n");
2885            if (vcc->pop)
2886 		vcc->pop(vcc, skb);
2887            else
2888 		dev_kfree_skb_any(skb);
2889 	   return 0;
2890         }
2891 
2892         if (skb->len > iadev->tx_buf_sz - 8) {
2893            printk("Transmit size over tx buffer size\n");
2894            if (vcc->pop)
2895                  vcc->pop(vcc, skb);
2896            else
2897                  dev_kfree_skb_any(skb);
2898           return 0;
2899         }
2900         if ((unsigned long)skb->data & 3) {
2901            printk("Misaligned SKB\n");
2902            if (vcc->pop)
2903                  vcc->pop(vcc, skb);
2904            else
2905                  dev_kfree_skb_any(skb);
2906            return 0;
2907         }
2908 	/* Get a descriptor number from our free descriptor queue
2909 	   We get the descr number from the TCQ now, since I am using
2910 	   the TCQ as a free buffer queue. Initially TCQ will be
2911 	   initialized with all the descriptors and is hence, full.
2912 	*/
2913 	desc = get_desc (iadev, iavcc);
2914 	if (desc == 0xffff)
2915 	    return 1;
2916 	comp_code = desc >> 13;
2917 	desc &= 0x1fff;
2918 
2919 	if ((desc == 0) || (desc > iadev->num_tx_desc))
2920 	{
2921 		IF_ERR(printk(DEV_LABEL "invalid desc for send: %d\n", desc);)
2922                 atomic_inc(&vcc->stats->tx);
2923 		if (vcc->pop)
2924 		    vcc->pop(vcc, skb);
2925 		else
2926 		    dev_kfree_skb_any(skb);
2927 		return 0;   /* return SUCCESS */
2928 	}
2929 
2930 	if (comp_code)
2931 	{
2932 	    IF_ERR(printk(DEV_LABEL "send desc:%d completion code %d error\n",
2933                                                             desc, comp_code);)
2934 	}
2935 
2936         /* remember the desc and vcc mapping */
2937         iavcc->vc_desc_cnt++;
2938         iadev->desc_tbl[desc-1].iavcc = iavcc;
2939         iadev->desc_tbl[desc-1].txskb = skb;
2940         IA_SKB_STATE(skb) = 0;
2941 
2942         iadev->ffL.tcq_rd += 2;
2943         if (iadev->ffL.tcq_rd > iadev->ffL.tcq_ed)
2944 	  	iadev->ffL.tcq_rd  = iadev->ffL.tcq_st;
2945 	writew(iadev->ffL.tcq_rd, iadev->seg_reg+TCQ_RD_PTR);
2946 
2947 	/* Put the descriptor number in the packet ready queue
2948 		and put the updated write pointer in the DLE field
2949 	*/
2950 	*(u16*)(iadev->seg_ram+iadev->ffL.prq_wr) = desc;
2951 
2952  	iadev->ffL.prq_wr += 2;
2953         if (iadev->ffL.prq_wr > iadev->ffL.prq_ed)
2954                 iadev->ffL.prq_wr = iadev->ffL.prq_st;
2955 
2956 	/* Figure out the exact length of the packet and padding required to
2957            make it  aligned on a 48 byte boundary.  */
2958 	total_len = skb->len + sizeof(struct cpcs_trailer);
2959 	total_len = ((total_len + 47) / 48) * 48;
2960 	IF_TX(printk("ia packet len:%d padding:%d\n", total_len, total_len - skb->len);)
2961 
2962 	/* Put the packet in a tx buffer */
2963 	trailer = iadev->tx_buf[desc-1].cpcs;
2964         IF_TX(printk("Sent: skb = 0x%p skb->data: 0x%p len: %d, desc: %d\n",
2965                   skb, skb->data, skb->len, desc);)
2966 	trailer->control = 0;
2967         /*big endian*/
2968 	trailer->length = ((skb->len & 0xff) << 8) | ((skb->len & 0xff00) >> 8);
2969 	trailer->crc32 = 0;	/* not needed - dummy bytes */
2970 
2971 	/* Display the packet */
2972 	IF_TXPKT(printk("Sent data: len = %d MsgNum = %d\n",
2973                                                         skb->len, tcnter++);
2974         xdump(skb->data, skb->len, "TX: ");
2975         printk("\n");)
2976 
2977 	/* Build the buffer descriptor */
2978 	buf_desc_ptr = iadev->seg_ram+TX_DESC_BASE;
2979 	buf_desc_ptr += desc;	/* points to the corresponding entry */
2980 	buf_desc_ptr->desc_mode = AAL5 | EOM_EN | APP_CRC32 | CMPL_INT;
2981 	/* Huh ? p.115 of users guide describes this as a read-only register */
2982         writew(TRANSMIT_DONE, iadev->seg_reg+SEG_INTR_STATUS_REG);
2983 	buf_desc_ptr->vc_index = vcc->vci;
2984 	buf_desc_ptr->bytes = total_len;
2985 
2986         if (vcc->qos.txtp.traffic_class == ATM_ABR)
2987 	   clear_lockup (vcc, iadev);
2988 
2989 	/* Build the DLE structure */
2990 	wr_ptr = iadev->tx_dle_q.write;
2991 	memset((caddr_t)wr_ptr, 0, sizeof(*wr_ptr));
2992 	wr_ptr->sys_pkt_addr = pci_map_single(iadev->pci, skb->data,
2993 		skb->len, PCI_DMA_TODEVICE);
2994 	wr_ptr->local_pkt_addr = (buf_desc_ptr->buf_start_hi << 16) |
2995                                                   buf_desc_ptr->buf_start_lo;
2996 	/* wr_ptr->bytes = swap_byte_order(total_len); didn't seem to affect?? */
2997 	wr_ptr->bytes = skb->len;
2998 
2999         /* hw bug - DLEs of 0x2d, 0x2e, 0x2f cause DMA lockup */
3000         if ((wr_ptr->bytes >> 2) == 0xb)
3001            wr_ptr->bytes = 0x30;
3002 
3003 	wr_ptr->mode = TX_DLE_PSI;
3004 	wr_ptr->prq_wr_ptr_data = 0;
3005 
3006 	/* end is not to be used for the DLE q */
3007 	if (++wr_ptr == iadev->tx_dle_q.end)
3008 		wr_ptr = iadev->tx_dle_q.start;
3009 
3010         /* Build trailer dle */
3011         wr_ptr->sys_pkt_addr = iadev->tx_buf[desc-1].dma_addr;
3012         wr_ptr->local_pkt_addr = ((buf_desc_ptr->buf_start_hi << 16) |
3013           buf_desc_ptr->buf_start_lo) + total_len - sizeof(struct cpcs_trailer);
3014 
3015         wr_ptr->bytes = sizeof(struct cpcs_trailer);
3016         wr_ptr->mode = DMA_INT_ENABLE;
3017         wr_ptr->prq_wr_ptr_data = iadev->ffL.prq_wr;
3018 
3019         /* end is not to be used for the DLE q */
3020         if (++wr_ptr == iadev->tx_dle_q.end)
3021                 wr_ptr = iadev->tx_dle_q.start;
3022 
3023 	iadev->tx_dle_q.write = wr_ptr;
3024         ATM_DESC(skb) = vcc->vci;
3025         skb_queue_tail(&iadev->tx_dma_q, skb);
3026 
3027         atomic_inc(&vcc->stats->tx);
3028         iadev->tx_pkt_cnt++;
3029 	/* Increment transaction counter */
3030 	writel(2, iadev->dma+IPHASE5575_TX_COUNTER);
3031 
3032 #if 0
3033         /* add flow control logic */
3034         if (atomic_read(&vcc->stats->tx) % 20 == 0) {
3035           if (iavcc->vc_desc_cnt > 10) {
3036              vcc->tx_quota =  vcc->tx_quota * 3 / 4;
3037             printk("Tx1:  vcc->tx_quota = %d \n", (u32)vcc->tx_quota );
3038               iavcc->flow_inc = -1;
3039               iavcc->saved_tx_quota = vcc->tx_quota;
3040            } else if ((iavcc->flow_inc < 0) && (iavcc->vc_desc_cnt < 3)) {
3041              // vcc->tx_quota = 3 * iavcc->saved_tx_quota / 4;
3042              printk("Tx2:  vcc->tx_quota = %d \n", (u32)vcc->tx_quota );
3043               iavcc->flow_inc = 0;
3044            }
3045         }
3046 #endif
3047 	IF_TX(printk("ia send done\n");)
3048 	return 0;
3049 }
3050 
3051 static int ia_send(struct atm_vcc *vcc, struct sk_buff *skb)
3052 {
3053         IADEV *iadev;
3054         struct ia_vcc *iavcc;
3055         unsigned long flags;
3056 
3057         iadev = INPH_IA_DEV(vcc->dev);
3058         iavcc = INPH_IA_VCC(vcc);
3059         if ((!skb)||(skb->len>(iadev->tx_buf_sz-sizeof(struct cpcs_trailer))))
3060         {
3061             if (!skb)
3062                 printk(KERN_CRIT "null skb in ia_send\n");
3063             else dev_kfree_skb_any(skb);
3064             return -EINVAL;
3065         }
3066         spin_lock_irqsave(&iadev->tx_lock, flags);
3067         if (!test_bit(ATM_VF_READY,&vcc->flags)){
3068             dev_kfree_skb_any(skb);
3069             spin_unlock_irqrestore(&iadev->tx_lock, flags);
3070             return -EINVAL;
3071         }
3072         ATM_SKB(skb)->vcc = vcc;
3073 
3074         if (skb_peek(&iadev->tx_backlog)) {
3075            skb_queue_tail(&iadev->tx_backlog, skb);
3076         }
3077         else {
3078            if (ia_pkt_tx (vcc, skb)) {
3079               skb_queue_tail(&iadev->tx_backlog, skb);
3080            }
3081         }
3082         spin_unlock_irqrestore(&iadev->tx_lock, flags);
3083         return 0;
3084 
3085 }
3086 
3087 static int ia_proc_read(struct atm_dev *dev,loff_t *pos,char *page)
3088 {
3089   int   left = *pos, n;
3090   char  *tmpPtr;
3091   IADEV *iadev = INPH_IA_DEV(dev);
3092   if(!left--) {
3093      if (iadev->phy_type == FE_25MBIT_PHY) {
3094        n = sprintf(page, "  Board Type         :  Iphase5525-1KVC-128K\n");
3095        return n;
3096      }
3097      if (iadev->phy_type == FE_DS3_PHY)
3098         n = sprintf(page, "  Board Type         :  Iphase-ATM-DS3");
3099      else if (iadev->phy_type == FE_E3_PHY)
3100         n = sprintf(page, "  Board Type         :  Iphase-ATM-E3");
3101      else if (iadev->phy_type == FE_UTP_OPTION)
3102          n = sprintf(page, "  Board Type         :  Iphase-ATM-UTP155");
3103      else
3104         n = sprintf(page, "  Board Type         :  Iphase-ATM-OC3");
3105      tmpPtr = page + n;
3106      if (iadev->pci_map_size == 0x40000)
3107         n += sprintf(tmpPtr, "-1KVC-");
3108      else
3109         n += sprintf(tmpPtr, "-4KVC-");
3110      tmpPtr = page + n;
3111      if ((iadev->memType & MEM_SIZE_MASK) == MEM_SIZE_1M)
3112         n += sprintf(tmpPtr, "1M  \n");
3113      else if ((iadev->memType & MEM_SIZE_MASK) == MEM_SIZE_512K)
3114         n += sprintf(tmpPtr, "512K\n");
3115      else
3116        n += sprintf(tmpPtr, "128K\n");
3117      return n;
3118   }
3119   if (!left) {
3120      return  sprintf(page, "  Number of Tx Buffer:  %u\n"
3121                            "  Size of Tx Buffer  :  %u\n"
3122                            "  Number of Rx Buffer:  %u\n"
3123                            "  Size of Rx Buffer  :  %u\n"
3124                            "  Packets Receiverd  :  %u\n"
3125                            "  Packets Transmitted:  %u\n"
3126                            "  Cells Received     :  %u\n"
3127                            "  Cells Transmitted  :  %u\n"
3128                            "  Board Dropped Cells:  %u\n"
3129                            "  Board Dropped Pkts :  %u\n",
3130                            iadev->num_tx_desc,  iadev->tx_buf_sz,
3131                            iadev->num_rx_desc,  iadev->rx_buf_sz,
3132                            iadev->rx_pkt_cnt,   iadev->tx_pkt_cnt,
3133                            iadev->rx_cell_cnt, iadev->tx_cell_cnt,
3134                            iadev->drop_rxcell, iadev->drop_rxpkt);
3135   }
3136   return 0;
3137 }
3138 
3139 static const struct atmdev_ops ops = {
3140 	.open		= ia_open,
3141 	.close		= ia_close,
3142 	.ioctl		= ia_ioctl,
3143 	.getsockopt	= ia_getsockopt,
3144 	.setsockopt	= ia_setsockopt,
3145 	.send		= ia_send,
3146 	.phy_put	= ia_phy_put,
3147 	.phy_get	= ia_phy_get,
3148 	.change_qos	= ia_change_qos,
3149 	.proc_read	= ia_proc_read,
3150 	.owner		= THIS_MODULE,
3151 };
3152 
3153 static int __devinit ia_init_one(struct pci_dev *pdev,
3154 				 const struct pci_device_id *ent)
3155 {
3156 	struct atm_dev *dev;
3157 	IADEV *iadev;
3158         unsigned long flags;
3159 	int ret;
3160 
3161 	iadev = kzalloc(sizeof(*iadev), GFP_KERNEL);
3162 	if (!iadev) {
3163 		ret = -ENOMEM;
3164 		goto err_out;
3165 	}
3166 
3167 	iadev->pci = pdev;
3168 
3169 	IF_INIT(printk("ia detected at bus:%d dev: %d function:%d\n",
3170 		pdev->bus->number, PCI_SLOT(pdev->devfn), PCI_FUNC(pdev->devfn));)
3171 	if (pci_enable_device(pdev)) {
3172 		ret = -ENODEV;
3173 		goto err_out_free_iadev;
3174 	}
3175 	dev = atm_dev_register(DEV_LABEL, &ops, -1, NULL);
3176 	if (!dev) {
3177 		ret = -ENOMEM;
3178 		goto err_out_disable_dev;
3179 	}
3180 	dev->dev_data = iadev;
3181 	IF_INIT(printk(DEV_LABEL "registered at (itf :%d)\n", dev->number);)
3182 	IF_INIT(printk("dev_id = 0x%p iadev->LineRate = %d \n", dev,
3183 		iadev->LineRate);)
3184 
3185 	pci_set_drvdata(pdev, dev);
3186 
3187 	ia_dev[iadev_count] = iadev;
3188 	_ia_dev[iadev_count] = dev;
3189 	iadev_count++;
3190 	spin_lock_init(&iadev->misc_lock);
3191 	/* First fixes first. I don't want to think about this now. */
3192 	spin_lock_irqsave(&iadev->misc_lock, flags);
3193 	if (ia_init(dev) || ia_start(dev)) {
3194 		IF_INIT(printk("IA register failed!\n");)
3195 		iadev_count--;
3196 		ia_dev[iadev_count] = NULL;
3197 		_ia_dev[iadev_count] = NULL;
3198 		spin_unlock_irqrestore(&iadev->misc_lock, flags);
3199 		ret = -EINVAL;
3200 		goto err_out_deregister_dev;
3201 	}
3202 	spin_unlock_irqrestore(&iadev->misc_lock, flags);
3203 	IF_EVENT(printk("iadev_count = %d\n", iadev_count);)
3204 
3205 	iadev->next_board = ia_boards;
3206 	ia_boards = dev;
3207 
3208 	return 0;
3209 
3210 err_out_deregister_dev:
3211 	atm_dev_deregister(dev);
3212 err_out_disable_dev:
3213 	pci_disable_device(pdev);
3214 err_out_free_iadev:
3215 	kfree(iadev);
3216 err_out:
3217 	return ret;
3218 }
3219 
3220 static void __devexit ia_remove_one(struct pci_dev *pdev)
3221 {
3222 	struct atm_dev *dev = pci_get_drvdata(pdev);
3223 	IADEV *iadev = INPH_IA_DEV(dev);
3224 
3225 	/* Disable phy interrupts */
3226 	ia_phy_put(dev, ia_phy_get(dev, SUNI_RSOP_CIE) & ~(SUNI_RSOP_CIE_LOSE),
3227 				   SUNI_RSOP_CIE);
3228 	udelay(1);
3229 
3230 	if (dev->phy && dev->phy->stop)
3231 		dev->phy->stop(dev);
3232 
3233 	/* De-register device */
3234       	free_irq(iadev->irq, dev);
3235 	iadev_count--;
3236 	ia_dev[iadev_count] = NULL;
3237 	_ia_dev[iadev_count] = NULL;
3238 	IF_EVENT(printk("deregistering iav at (itf:%d)\n", dev->number);)
3239 	atm_dev_deregister(dev);
3240 
3241       	iounmap(iadev->base);
3242 	pci_disable_device(pdev);
3243 
3244 	ia_free_rx(iadev);
3245 	ia_free_tx(iadev);
3246 
3247       	kfree(iadev);
3248 }
3249 
3250 static struct pci_device_id ia_pci_tbl[] = {
3251 	{ PCI_VENDOR_ID_IPHASE, 0x0008, PCI_ANY_ID, PCI_ANY_ID, },
3252 	{ PCI_VENDOR_ID_IPHASE, 0x0009, PCI_ANY_ID, PCI_ANY_ID, },
3253 	{ 0,}
3254 };
3255 MODULE_DEVICE_TABLE(pci, ia_pci_tbl);
3256 
3257 static struct pci_driver ia_driver = {
3258 	.name =         DEV_LABEL,
3259 	.id_table =     ia_pci_tbl,
3260 	.probe =        ia_init_one,
3261 	.remove =       __devexit_p(ia_remove_one),
3262 };
3263 
3264 static int __init ia_module_init(void)
3265 {
3266 	int ret;
3267 
3268 	ret = pci_register_driver(&ia_driver);
3269 	if (ret >= 0) {
3270 		ia_timer.expires = jiffies + 3*HZ;
3271 		add_timer(&ia_timer);
3272 	} else
3273 		printk(KERN_ERR DEV_LABEL ": no adapter found\n");
3274 	return ret;
3275 }
3276 
3277 static void __exit ia_module_exit(void)
3278 {
3279 	pci_unregister_driver(&ia_driver);
3280 
3281         del_timer(&ia_timer);
3282 }
3283 
3284 module_init(ia_module_init);
3285 module_exit(ia_module_exit);
3286