1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18#include <linux/init.h>
19#include "hisax.h"
20#include "hfc_pci.h"
21#include "isdnl1.h"
22#include <linux/pci.h>
23#include <linux/sched.h>
24#include <linux/interrupt.h>
25
26static const char *hfcpci_revision = "$Revision: 1.48.2.4 $";
27
28
29typedef struct {
30 int vendor_id;
31 int device_id;
32 char *vendor_name;
33 char *card_name;
34} PCI_ENTRY;
35
36#define NT_T1_COUNT 20
37#define CLKDEL_TE 0x0e
38#define CLKDEL_NT 0x6c
39
40static const PCI_ENTRY id_list[] =
41{
42 {PCI_VENDOR_ID_CCD, PCI_DEVICE_ID_CCD_2BD0, "CCD/Billion/Asuscom", "2BD0"},
43 {PCI_VENDOR_ID_CCD, PCI_DEVICE_ID_CCD_B000, "Billion", "B000"},
44 {PCI_VENDOR_ID_CCD, PCI_DEVICE_ID_CCD_B006, "Billion", "B006"},
45 {PCI_VENDOR_ID_CCD, PCI_DEVICE_ID_CCD_B007, "Billion", "B007"},
46 {PCI_VENDOR_ID_CCD, PCI_DEVICE_ID_CCD_B008, "Billion", "B008"},
47 {PCI_VENDOR_ID_CCD, PCI_DEVICE_ID_CCD_B009, "Billion", "B009"},
48 {PCI_VENDOR_ID_CCD, PCI_DEVICE_ID_CCD_B00A, "Billion", "B00A"},
49 {PCI_VENDOR_ID_CCD, PCI_DEVICE_ID_CCD_B00B, "Billion", "B00B"},
50 {PCI_VENDOR_ID_CCD, PCI_DEVICE_ID_CCD_B00C, "Billion", "B00C"},
51 {PCI_VENDOR_ID_CCD, PCI_DEVICE_ID_CCD_B100, "Seyeon", "B100"},
52 {PCI_VENDOR_ID_CCD, PCI_DEVICE_ID_CCD_B700, "Primux II S0", "B700"},
53 {PCI_VENDOR_ID_CCD, PCI_DEVICE_ID_CCD_B701, "Primux II S0 NT", "B701"},
54 {PCI_VENDOR_ID_ABOCOM, PCI_DEVICE_ID_ABOCOM_2BD1, "Abocom/Magitek", "2BD1"},
55 {PCI_VENDOR_ID_ASUSTEK, PCI_DEVICE_ID_ASUSTEK_0675, "Asuscom/Askey", "675"},
56 {PCI_VENDOR_ID_BERKOM, PCI_DEVICE_ID_BERKOM_T_CONCEPT, "German telekom", "T-Concept"},
57 {PCI_VENDOR_ID_BERKOM, PCI_DEVICE_ID_BERKOM_A1T, "German telekom", "A1T"},
58 {PCI_VENDOR_ID_ANIGMA, PCI_DEVICE_ID_ANIGMA_MC145575, "Motorola MC145575", "MC145575"},
59 {PCI_VENDOR_ID_ZOLTRIX, PCI_DEVICE_ID_ZOLTRIX_2BD0, "Zoltrix", "2BD0"},
60 {PCI_VENDOR_ID_DIGI, PCI_DEVICE_ID_DIGI_DF_M_IOM2_E, "Digi International", "Digi DataFire Micro V IOM2 (Europe)"},
61 {PCI_VENDOR_ID_DIGI, PCI_DEVICE_ID_DIGI_DF_M_E, "Digi International", "Digi DataFire Micro V (Europe)"},
62 {PCI_VENDOR_ID_DIGI, PCI_DEVICE_ID_DIGI_DF_M_IOM2_A, "Digi International", "Digi DataFire Micro V IOM2 (North America)"},
63 {PCI_VENDOR_ID_DIGI, PCI_DEVICE_ID_DIGI_DF_M_A, "Digi International", "Digi DataFire Micro V (North America)"},
64 {PCI_VENDOR_ID_SITECOM, PCI_DEVICE_ID_SITECOM_DC105V2, "Sitecom Europe", "DC-105 ISDN PCI"},
65 {0, 0, NULL, NULL},
66};
67
68
69
70
71
72static void
73release_io_hfcpci(struct IsdnCardState *cs)
74{
75 printk(KERN_INFO "HiSax: release hfcpci at %p\n",
76 cs->hw.hfcpci.pci_io);
77 cs->hw.hfcpci.int_m2 = 0;
78 Write_hfc(cs, HFCPCI_INT_M2, cs->hw.hfcpci.int_m2);
79 Write_hfc(cs, HFCPCI_CIRM, HFCPCI_RESET);
80 mdelay(10);
81 Write_hfc(cs, HFCPCI_CIRM, 0);
82 mdelay(10);
83 Write_hfc(cs, HFCPCI_INT_M2, cs->hw.hfcpci.int_m2);
84 pci_write_config_word(cs->hw.hfcpci.dev, PCI_COMMAND, 0);
85 del_timer(&cs->hw.hfcpci.timer);
86 pci_free_consistent(cs->hw.hfcpci.dev, 0x8000,
87 cs->hw.hfcpci.fifos, cs->hw.hfcpci.dma);
88 cs->hw.hfcpci.fifos = NULL;
89 iounmap((void *)cs->hw.hfcpci.pci_io);
90}
91
92
93
94
95
96static void
97reset_hfcpci(struct IsdnCardState *cs)
98{
99 pci_write_config_word(cs->hw.hfcpci.dev, PCI_COMMAND, PCI_ENA_MEMIO);
100 cs->hw.hfcpci.int_m2 = 0;
101 Write_hfc(cs, HFCPCI_INT_M2, cs->hw.hfcpci.int_m2);
102
103 printk(KERN_INFO "HFC_PCI: resetting card\n");
104 pci_write_config_word(cs->hw.hfcpci.dev, PCI_COMMAND, PCI_ENA_MEMIO + PCI_ENA_MASTER);
105 Write_hfc(cs, HFCPCI_CIRM, HFCPCI_RESET);
106 mdelay(10);
107 Write_hfc(cs, HFCPCI_CIRM, 0);
108 mdelay(10);
109 if (Read_hfc(cs, HFCPCI_STATUS) & 2)
110 printk(KERN_WARNING "HFC-PCI init bit busy\n");
111
112 cs->hw.hfcpci.fifo_en = 0x30;
113 Write_hfc(cs, HFCPCI_FIFO_EN, cs->hw.hfcpci.fifo_en);
114
115 cs->hw.hfcpci.trm = 0 + HFCPCI_BTRANS_THRESMASK;
116 Write_hfc(cs, HFCPCI_TRM, cs->hw.hfcpci.trm);
117
118 Write_hfc(cs, HFCPCI_CLKDEL, CLKDEL_TE);
119 cs->hw.hfcpci.sctrl_e = HFCPCI_AUTO_AWAKE;
120 Write_hfc(cs, HFCPCI_SCTRL_E, cs->hw.hfcpci.sctrl_e);
121 cs->hw.hfcpci.bswapped = 0;
122 cs->hw.hfcpci.nt_mode = 0;
123 cs->hw.hfcpci.ctmt = HFCPCI_TIM3_125 | HFCPCI_AUTO_TIMER;
124 Write_hfc(cs, HFCPCI_CTMT, cs->hw.hfcpci.ctmt);
125
126 cs->hw.hfcpci.int_m1 = HFCPCI_INTS_DTRANS | HFCPCI_INTS_DREC |
127 HFCPCI_INTS_L1STATE | HFCPCI_INTS_TIMER;
128 Write_hfc(cs, HFCPCI_INT_M1, cs->hw.hfcpci.int_m1);
129
130
131 if (Read_hfc(cs, HFCPCI_INT_S1));
132
133 Write_hfc(cs, HFCPCI_STATES, HFCPCI_LOAD_STATE | 2);
134 udelay(10);
135 Write_hfc(cs, HFCPCI_STATES, 2);
136 cs->hw.hfcpci.mst_m = HFCPCI_MASTER;
137
138 Write_hfc(cs, HFCPCI_MST_MODE, cs->hw.hfcpci.mst_m);
139 cs->hw.hfcpci.sctrl = 0x40;
140 Write_hfc(cs, HFCPCI_SCTRL, cs->hw.hfcpci.sctrl);
141 cs->hw.hfcpci.sctrl_r = 0;
142 Write_hfc(cs, HFCPCI_SCTRL_R, cs->hw.hfcpci.sctrl_r);
143
144
145
146
147
148
149
150
151 cs->hw.hfcpci.conn = 0x36;
152 Write_hfc(cs, HFCPCI_CONNECT, cs->hw.hfcpci.conn);
153 Write_hfc(cs, HFCPCI_B1_SSL, 0x80);
154 Write_hfc(cs, HFCPCI_B2_SSL, 0x81);
155 Write_hfc(cs, HFCPCI_B1_RSL, 0x80);
156 Write_hfc(cs, HFCPCI_B2_RSL, 0x81);
157
158
159 cs->hw.hfcpci.int_m2 = HFCPCI_IRQ_ENABLE;
160 Write_hfc(cs, HFCPCI_INT_M2, cs->hw.hfcpci.int_m2);
161 if (Read_hfc(cs, HFCPCI_INT_S1));
162}
163
164
165
166
167static void
168hfcpci_Timer(struct IsdnCardState *cs)
169{
170 cs->hw.hfcpci.timer.expires = jiffies + 75;
171
172
173
174
175}
176
177
178
179
180
181static void
182sched_event_D_pci(struct IsdnCardState *cs, int event)
183{
184 test_and_set_bit(event, &cs->event);
185 schedule_work(&cs->tqueue);
186}
187
188
189
190
191static void
192hfcpci_sched_event(struct BCState *bcs, int event)
193{
194 test_and_set_bit(event, &bcs->event);
195 schedule_work(&bcs->tqueue);
196}
197
198
199
200
201static
202struct BCState *
203Sel_BCS(struct IsdnCardState *cs, int channel)
204{
205 if (cs->bcs[0].mode && (cs->bcs[0].channel == channel))
206 return (&cs->bcs[0]);
207 else if (cs->bcs[1].mode && (cs->bcs[1].channel == channel))
208 return (&cs->bcs[1]);
209 else
210 return (NULL);
211}
212
213
214
215
216static void hfcpci_clear_fifo_rx(struct IsdnCardState *cs, int fifo)
217{ u_char fifo_state;
218 bzfifo_type *bzr;
219
220 if (fifo) {
221 bzr = &((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.rxbz_b2;
222 fifo_state = cs->hw.hfcpci.fifo_en & HFCPCI_FIFOEN_B2RX;
223 } else {
224 bzr = &((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.rxbz_b1;
225 fifo_state = cs->hw.hfcpci.fifo_en & HFCPCI_FIFOEN_B1RX;
226 }
227 if (fifo_state)
228 cs->hw.hfcpci.fifo_en ^= fifo_state;
229 Write_hfc(cs, HFCPCI_FIFO_EN, cs->hw.hfcpci.fifo_en);
230 cs->hw.hfcpci.last_bfifo_cnt[fifo] = 0;
231 bzr->za[MAX_B_FRAMES].z1 = B_FIFO_SIZE + B_SUB_VAL - 1;
232 bzr->za[MAX_B_FRAMES].z2 = bzr->za[MAX_B_FRAMES].z1;
233 bzr->f1 = MAX_B_FRAMES;
234 bzr->f2 = bzr->f1;
235 if (fifo_state)
236 cs->hw.hfcpci.fifo_en |= fifo_state;
237 Write_hfc(cs, HFCPCI_FIFO_EN, cs->hw.hfcpci.fifo_en);
238}
239
240
241
242
243static void hfcpci_clear_fifo_tx(struct IsdnCardState *cs, int fifo)
244{ u_char fifo_state;
245 bzfifo_type *bzt;
246
247 if (fifo) {
248 bzt = &((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.txbz_b2;
249 fifo_state = cs->hw.hfcpci.fifo_en & HFCPCI_FIFOEN_B2TX;
250 } else {
251 bzt = &((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.txbz_b1;
252 fifo_state = cs->hw.hfcpci.fifo_en & HFCPCI_FIFOEN_B1TX;
253 }
254 if (fifo_state)
255 cs->hw.hfcpci.fifo_en ^= fifo_state;
256 Write_hfc(cs, HFCPCI_FIFO_EN, cs->hw.hfcpci.fifo_en);
257 bzt->za[MAX_B_FRAMES].z1 = B_FIFO_SIZE + B_SUB_VAL - 1;
258 bzt->za[MAX_B_FRAMES].z2 = bzt->za[MAX_B_FRAMES].z1;
259 bzt->f1 = MAX_B_FRAMES;
260 bzt->f2 = bzt->f1;
261 if (fifo_state)
262 cs->hw.hfcpci.fifo_en |= fifo_state;
263 Write_hfc(cs, HFCPCI_FIFO_EN, cs->hw.hfcpci.fifo_en);
264}
265
266
267
268
269static struct sk_buff
270*
271hfcpci_empty_fifo(struct BCState *bcs, bzfifo_type *bz, u_char *bdata, int count)
272{
273 u_char *ptr, *ptr1, new_f2;
274 struct sk_buff *skb;
275 struct IsdnCardState *cs = bcs->cs;
276 int total, maxlen, new_z2;
277 z_type *zp;
278
279 if ((cs->debug & L1_DEB_HSCX) && !(cs->debug & L1_DEB_HSCX_FIFO))
280 debugl1(cs, "hfcpci_empty_fifo");
281 zp = &bz->za[bz->f2];
282 new_z2 = zp->z2 + count;
283 if (new_z2 >= (B_FIFO_SIZE + B_SUB_VAL))
284 new_z2 -= B_FIFO_SIZE;
285 new_f2 = (bz->f2 + 1) & MAX_B_FRAMES;
286 if ((count > HSCX_BUFMAX + 3) || (count < 4) ||
287 (*(bdata + (zp->z1 - B_SUB_VAL)))) {
288 if (cs->debug & L1_DEB_WARN)
289 debugl1(cs, "hfcpci_empty_fifo: incoming packet invalid length %d or crc", count);
290#ifdef ERROR_STATISTIC
291 bcs->err_inv++;
292#endif
293 bz->za[new_f2].z2 = new_z2;
294 bz->f2 = new_f2;
295 skb = NULL;
296 } else if (!(skb = dev_alloc_skb(count - 3)))
297 printk(KERN_WARNING "HFCPCI: receive out of memory\n");
298 else {
299 total = count;
300 count -= 3;
301 ptr = skb_put(skb, count);
302
303 if (zp->z2 + count <= B_FIFO_SIZE + B_SUB_VAL)
304 maxlen = count;
305 else
306 maxlen = B_FIFO_SIZE + B_SUB_VAL - zp->z2;
307
308 ptr1 = bdata + (zp->z2 - B_SUB_VAL);
309 memcpy(ptr, ptr1, maxlen);
310 count -= maxlen;
311
312 if (count) {
313 ptr += maxlen;
314 ptr1 = bdata;
315 memcpy(ptr, ptr1, count);
316 }
317 bz->za[new_f2].z2 = new_z2;
318 bz->f2 = new_f2;
319
320 }
321 return (skb);
322}
323
324
325
326
327static
328int
329receive_dmsg(struct IsdnCardState *cs)
330{
331 struct sk_buff *skb;
332 int maxlen;
333 int rcnt, total;
334 int count = 5;
335 u_char *ptr, *ptr1;
336 dfifo_type *df;
337 z_type *zp;
338
339 df = &((fifo_area *) (cs->hw.hfcpci.fifos))->d_chan.d_rx;
340 if (test_and_set_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
341 debugl1(cs, "rec_dmsg blocked");
342 return (1);
343 }
344 while (((df->f1 & D_FREG_MASK) != (df->f2 & D_FREG_MASK)) && count--) {
345 zp = &df->za[df->f2 & D_FREG_MASK];
346 rcnt = zp->z1 - zp->z2;
347 if (rcnt < 0)
348 rcnt += D_FIFO_SIZE;
349 rcnt++;
350 if (cs->debug & L1_DEB_ISAC)
351 debugl1(cs, "hfcpci recd f1(%d) f2(%d) z1(%x) z2(%x) cnt(%d)",
352 df->f1, df->f2, zp->z1, zp->z2, rcnt);
353
354 if ((rcnt > MAX_DFRAME_LEN + 3) || (rcnt < 4) ||
355 (df->data[zp->z1])) {
356 if (cs->debug & L1_DEB_WARN)
357 debugl1(cs, "empty_fifo hfcpci packet inv. len %d or crc %d", rcnt, df->data[zp->z1]);
358#ifdef ERROR_STATISTIC
359 cs->err_rx++;
360#endif
361 df->f2 = ((df->f2 + 1) & MAX_D_FRAMES) | (MAX_D_FRAMES + 1);
362 df->za[df->f2 & D_FREG_MASK].z2 = (zp->z2 + rcnt) & (D_FIFO_SIZE - 1);
363 } else if ((skb = dev_alloc_skb(rcnt - 3))) {
364 total = rcnt;
365 rcnt -= 3;
366 ptr = skb_put(skb, rcnt);
367
368 if (zp->z2 + rcnt <= D_FIFO_SIZE)
369 maxlen = rcnt;
370 else
371 maxlen = D_FIFO_SIZE - zp->z2;
372
373 ptr1 = df->data + zp->z2;
374 memcpy(ptr, ptr1, maxlen);
375 rcnt -= maxlen;
376
377 if (rcnt) {
378 ptr += maxlen;
379 ptr1 = df->data;
380 memcpy(ptr, ptr1, rcnt);
381 }
382 df->f2 = ((df->f2 + 1) & MAX_D_FRAMES) | (MAX_D_FRAMES + 1);
383 df->za[df->f2 & D_FREG_MASK].z2 = (zp->z2 + total) & (D_FIFO_SIZE - 1);
384
385 skb_queue_tail(&cs->rq, skb);
386 sched_event_D_pci(cs, D_RCVBUFREADY);
387 } else
388 printk(KERN_WARNING "HFC-PCI: D receive out of memory\n");
389 }
390 test_and_clear_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags);
391 return (1);
392}
393
394
395
396
397static int
398hfcpci_empty_fifo_trans(struct BCState *bcs, bzfifo_type *bz, u_char *bdata)
399{
400 unsigned short *z1r, *z2r;
401 int new_z2, fcnt, maxlen;
402 struct sk_buff *skb;
403 u_char *ptr, *ptr1;
404
405 z1r = &bz->za[MAX_B_FRAMES].z1;
406 z2r = z1r + 1;
407
408 if (!(fcnt = *z1r - *z2r))
409 return (0);
410
411 if (fcnt <= 0)
412 fcnt += B_FIFO_SIZE;
413 if (fcnt > HFCPCI_BTRANS_THRESHOLD)
414 fcnt = HFCPCI_BTRANS_THRESHOLD;
415
416 new_z2 = *z2r + fcnt;
417 if (new_z2 >= (B_FIFO_SIZE + B_SUB_VAL))
418 new_z2 -= B_FIFO_SIZE;
419
420 if (!(skb = dev_alloc_skb(fcnt)))
421 printk(KERN_WARNING "HFCPCI: receive out of memory\n");
422 else {
423 ptr = skb_put(skb, fcnt);
424 if (*z2r + fcnt <= B_FIFO_SIZE + B_SUB_VAL)
425 maxlen = fcnt;
426 else
427 maxlen = B_FIFO_SIZE + B_SUB_VAL - *z2r;
428
429 ptr1 = bdata + (*z2r - B_SUB_VAL);
430 memcpy(ptr, ptr1, maxlen);
431 fcnt -= maxlen;
432
433 if (fcnt) {
434 ptr += maxlen;
435 ptr1 = bdata;
436 memcpy(ptr, ptr1, fcnt);
437 }
438 skb_queue_tail(&bcs->rqueue, skb);
439 hfcpci_sched_event(bcs, B_RCVBUFREADY);
440 }
441
442 *z2r = new_z2;
443 return (1);
444}
445
446
447
448
449static void
450main_rec_hfcpci(struct BCState *bcs)
451{
452 struct IsdnCardState *cs = bcs->cs;
453 int rcnt, real_fifo;
454 int receive, count = 5;
455 struct sk_buff *skb;
456 bzfifo_type *bz;
457 u_char *bdata;
458 z_type *zp;
459
460
461 if ((bcs->channel) && (!cs->hw.hfcpci.bswapped)) {
462 bz = &((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.rxbz_b2;
463 bdata = ((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.rxdat_b2;
464 real_fifo = 1;
465 } else {
466 bz = &((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.rxbz_b1;
467 bdata = ((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.rxdat_b1;
468 real_fifo = 0;
469 }
470Begin:
471 count--;
472 if (test_and_set_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
473 debugl1(cs, "rec_data %d blocked", bcs->channel);
474 return;
475 }
476 if (bz->f1 != bz->f2) {
477 if (cs->debug & L1_DEB_HSCX)
478 debugl1(cs, "hfcpci rec %d f1(%d) f2(%d)",
479 bcs->channel, bz->f1, bz->f2);
480 zp = &bz->za[bz->f2];
481
482 rcnt = zp->z1 - zp->z2;
483 if (rcnt < 0)
484 rcnt += B_FIFO_SIZE;
485 rcnt++;
486 if (cs->debug & L1_DEB_HSCX)
487 debugl1(cs, "hfcpci rec %d z1(%x) z2(%x) cnt(%d)",
488 bcs->channel, zp->z1, zp->z2, rcnt);
489 if ((skb = hfcpci_empty_fifo(bcs, bz, bdata, rcnt))) {
490 skb_queue_tail(&bcs->rqueue, skb);
491 hfcpci_sched_event(bcs, B_RCVBUFREADY);
492 }
493 rcnt = bz->f1 - bz->f2;
494 if (rcnt < 0)
495 rcnt += MAX_B_FRAMES + 1;
496 if (cs->hw.hfcpci.last_bfifo_cnt[real_fifo] > rcnt + 1) {
497 rcnt = 0;
498 hfcpci_clear_fifo_rx(cs, real_fifo);
499 }
500 cs->hw.hfcpci.last_bfifo_cnt[real_fifo] = rcnt;
501 if (rcnt > 1)
502 receive = 1;
503 else
504 receive = 0;
505 } else if (bcs->mode == L1_MODE_TRANS)
506 receive = hfcpci_empty_fifo_trans(bcs, bz, bdata);
507 else
508 receive = 0;
509 test_and_clear_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags);
510 if (count && receive)
511 goto Begin;
512}
513
514
515
516
517static void
518hfcpci_fill_dfifo(struct IsdnCardState *cs)
519{
520 int fcnt;
521 int count, new_z1, maxlen;
522 dfifo_type *df;
523 u_char *src, *dst, new_f1;
524
525 if (!cs->tx_skb)
526 return;
527 if (cs->tx_skb->len <= 0)
528 return;
529
530 df = &((fifo_area *) (cs->hw.hfcpci.fifos))->d_chan.d_tx;
531
532 if (cs->debug & L1_DEB_ISAC)
533 debugl1(cs, "hfcpci_fill_Dfifo f1(%d) f2(%d) z1(f1)(%x)",
534 df->f1, df->f2,
535 df->za[df->f1 & D_FREG_MASK].z1);
536 fcnt = df->f1 - df->f2;
537 if (fcnt < 0)
538 fcnt += (MAX_D_FRAMES + 1);
539 if (fcnt > (MAX_D_FRAMES - 1)) {
540 if (cs->debug & L1_DEB_ISAC)
541 debugl1(cs, "hfcpci_fill_Dfifo more as 14 frames");
542#ifdef ERROR_STATISTIC
543 cs->err_tx++;
544#endif
545 return;
546 }
547
548 count = df->za[df->f2 & D_FREG_MASK].z2 - df->za[df->f1 & D_FREG_MASK].z1 - 1;
549 if (count <= 0)
550 count += D_FIFO_SIZE;
551
552 if (cs->debug & L1_DEB_ISAC)
553 debugl1(cs, "hfcpci_fill_Dfifo count(%u/%d)",
554 cs->tx_skb->len, count);
555 if (count < cs->tx_skb->len) {
556 if (cs->debug & L1_DEB_ISAC)
557 debugl1(cs, "hfcpci_fill_Dfifo no fifo mem");
558 return;
559 }
560 count = cs->tx_skb->len;
561 new_z1 = (df->za[df->f1 & D_FREG_MASK].z1 + count) & (D_FIFO_SIZE - 1);
562 new_f1 = ((df->f1 + 1) & D_FREG_MASK) | (D_FREG_MASK + 1);
563 src = cs->tx_skb->data;
564 dst = df->data + df->za[df->f1 & D_FREG_MASK].z1;
565 maxlen = D_FIFO_SIZE - df->za[df->f1 & D_FREG_MASK].z1;
566 if (maxlen > count)
567 maxlen = count;
568 memcpy(dst, src, maxlen);
569
570 count -= maxlen;
571 if (count) {
572 dst = df->data;
573 src += maxlen;
574 memcpy(dst, src, count);
575 }
576 df->za[new_f1 & D_FREG_MASK].z1 = new_z1;
577 df->za[df->f1 & D_FREG_MASK].z1 = new_z1;
578 df->f1 = new_f1;
579
580 dev_kfree_skb_any(cs->tx_skb);
581 cs->tx_skb = NULL;
582}
583
584
585
586
587static void
588hfcpci_fill_fifo(struct BCState *bcs)
589{
590 struct IsdnCardState *cs = bcs->cs;
591 int maxlen, fcnt;
592 int count, new_z1;
593 bzfifo_type *bz;
594 u_char *bdata;
595 u_char new_f1, *src, *dst;
596 unsigned short *z1t, *z2t;
597
598 if (!bcs->tx_skb)
599 return;
600 if (bcs->tx_skb->len <= 0)
601 return;
602
603 if ((bcs->channel) && (!cs->hw.hfcpci.bswapped)) {
604 bz = &((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.txbz_b2;
605 bdata = ((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.txdat_b2;
606 } else {
607 bz = &((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.txbz_b1;
608 bdata = ((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.txdat_b1;
609 }
610
611 if (bcs->mode == L1_MODE_TRANS) {
612 z1t = &bz->za[MAX_B_FRAMES].z1;
613 z2t = z1t + 1;
614 if (cs->debug & L1_DEB_HSCX)
615 debugl1(cs, "hfcpci_fill_fifo_trans %d z1(%x) z2(%x)",
616 bcs->channel, *z1t, *z2t);
617 fcnt = *z2t - *z1t;
618 if (fcnt <= 0)
619 fcnt += B_FIFO_SIZE;
620 fcnt = B_FIFO_SIZE - fcnt;
621
622 while ((fcnt < 2 * HFCPCI_BTRANS_THRESHOLD) && (bcs->tx_skb)) {
623 if (bcs->tx_skb->len < B_FIFO_SIZE - fcnt) {
624
625 count = bcs->tx_skb->len;
626
627 new_z1 = *z1t + count;
628 if (new_z1 >= (B_FIFO_SIZE + B_SUB_VAL))
629 new_z1 -= B_FIFO_SIZE;
630 src = bcs->tx_skb->data;
631 dst = bdata + (*z1t - B_SUB_VAL);
632 maxlen = (B_FIFO_SIZE + B_SUB_VAL) - *z1t;
633 if (maxlen > count)
634 maxlen = count;
635 memcpy(dst, src, maxlen);
636
637 count -= maxlen;
638 if (count) {
639 dst = bdata;
640 src += maxlen;
641 memcpy(dst, src, count);
642 }
643 bcs->tx_cnt -= bcs->tx_skb->len;
644 fcnt += bcs->tx_skb->len;
645 *z1t = new_z1;
646 } else if (cs->debug & L1_DEB_HSCX)
647 debugl1(cs, "hfcpci_fill_fifo_trans %d frame length %d discarded",
648 bcs->channel, bcs->tx_skb->len);
649
650 if (test_bit(FLG_LLI_L1WAKEUP, &bcs->st->lli.flag) &&
651 (PACKET_NOACK != bcs->tx_skb->pkt_type)) {
652 u_long flags;
653 spin_lock_irqsave(&bcs->aclock, flags);
654 bcs->ackcnt += bcs->tx_skb->len;
655 spin_unlock_irqrestore(&bcs->aclock, flags);
656 schedule_event(bcs, B_ACKPENDING);
657 }
658
659 dev_kfree_skb_any(bcs->tx_skb);
660 bcs->tx_skb = skb_dequeue(&bcs->squeue);
661 }
662 test_and_clear_bit(BC_FLG_BUSY, &bcs->Flag);
663 return;
664 }
665 if (cs->debug & L1_DEB_HSCX)
666 debugl1(cs, "hfcpci_fill_fifo_hdlc %d f1(%d) f2(%d) z1(f1)(%x)",
667 bcs->channel, bz->f1, bz->f2,
668 bz->za[bz->f1].z1);
669
670 fcnt = bz->f1 - bz->f2;
671 if (fcnt < 0)
672 fcnt += (MAX_B_FRAMES + 1);
673 if (fcnt > (MAX_B_FRAMES - 1)) {
674 if (cs->debug & L1_DEB_HSCX)
675 debugl1(cs, "hfcpci_fill_Bfifo more as 14 frames");
676 return;
677 }
678
679 count = bz->za[bz->f2].z2 - bz->za[bz->f1].z1 - 1;
680 if (count <= 0)
681 count += B_FIFO_SIZE;
682
683 if (cs->debug & L1_DEB_HSCX)
684 debugl1(cs, "hfcpci_fill_fifo %d count(%u/%d),%lx",
685 bcs->channel, bcs->tx_skb->len,
686 count, current->state);
687
688 if (count < bcs->tx_skb->len) {
689 if (cs->debug & L1_DEB_HSCX)
690 debugl1(cs, "hfcpci_fill_fifo no fifo mem");
691 return;
692 }
693 count = bcs->tx_skb->len;
694 new_z1 = bz->za[bz->f1].z1 + count;
695 if (new_z1 >= (B_FIFO_SIZE + B_SUB_VAL))
696 new_z1 -= B_FIFO_SIZE;
697
698 new_f1 = ((bz->f1 + 1) & MAX_B_FRAMES);
699 src = bcs->tx_skb->data;
700 dst = bdata + (bz->za[bz->f1].z1 - B_SUB_VAL);
701 maxlen = (B_FIFO_SIZE + B_SUB_VAL) - bz->za[bz->f1].z1;
702 if (maxlen > count)
703 maxlen = count;
704 memcpy(dst, src, maxlen);
705
706 count -= maxlen;
707 if (count) {
708 dst = bdata;
709 src += maxlen;
710 memcpy(dst, src, count);
711 }
712 bcs->tx_cnt -= bcs->tx_skb->len;
713 if (test_bit(FLG_LLI_L1WAKEUP, &bcs->st->lli.flag) &&
714 (PACKET_NOACK != bcs->tx_skb->pkt_type)) {
715 u_long flags;
716 spin_lock_irqsave(&bcs->aclock, flags);
717 bcs->ackcnt += bcs->tx_skb->len;
718 spin_unlock_irqrestore(&bcs->aclock, flags);
719 schedule_event(bcs, B_ACKPENDING);
720 }
721
722 bz->za[new_f1].z1 = new_z1;
723 bz->f1 = new_f1;
724
725 dev_kfree_skb_any(bcs->tx_skb);
726 bcs->tx_skb = NULL;
727 test_and_clear_bit(BC_FLG_BUSY, &bcs->Flag);
728}
729
730
731
732
733static void
734dch_nt_l2l1(struct PStack *st, int pr, void *arg)
735{
736 struct IsdnCardState *cs = (struct IsdnCardState *) st->l1.hardware;
737
738 switch (pr) {
739 case (PH_DATA | REQUEST):
740 case (PH_PULL | REQUEST):
741 case (PH_PULL | INDICATION):
742 st->l1.l1hw(st, pr, arg);
743 break;
744 case (PH_ACTIVATE | REQUEST):
745 st->l1.l1l2(st, PH_ACTIVATE | CONFIRM, NULL);
746 break;
747 case (PH_TESTLOOP | REQUEST):
748 if (1 & (long) arg)
749 debugl1(cs, "PH_TEST_LOOP B1");
750 if (2 & (long) arg)
751 debugl1(cs, "PH_TEST_LOOP B2");
752 if (!(3 & (long) arg))
753 debugl1(cs, "PH_TEST_LOOP DISABLED");
754 st->l1.l1hw(st, HW_TESTLOOP | REQUEST, arg);
755 break;
756 default:
757 if (cs->debug)
758 debugl1(cs, "dch_nt_l2l1 msg %04X unhandled", pr);
759 break;
760 }
761}
762
763
764
765
766
767
768static int
769hfcpci_auxcmd(struct IsdnCardState *cs, isdn_ctrl *ic)
770{
771 u_long flags;
772 int i = *(unsigned int *) ic->parm.num;
773
774 if ((ic->arg == 98) &&
775 (!(cs->hw.hfcpci.int_m1 & (HFCPCI_INTS_B2TRANS + HFCPCI_INTS_B2REC + HFCPCI_INTS_B1TRANS + HFCPCI_INTS_B1REC)))) {
776 spin_lock_irqsave(&cs->lock, flags);
777 Write_hfc(cs, HFCPCI_CLKDEL, CLKDEL_NT);
778 Write_hfc(cs, HFCPCI_STATES, HFCPCI_LOAD_STATE | 0);
779 udelay(10);
780 cs->hw.hfcpci.sctrl |= SCTRL_MODE_NT;
781 Write_hfc(cs, HFCPCI_SCTRL, cs->hw.hfcpci.sctrl);
782 udelay(10);
783 Write_hfc(cs, HFCPCI_STATES, HFCPCI_LOAD_STATE | 1);
784 udelay(10);
785 Write_hfc(cs, HFCPCI_STATES, 1 | HFCPCI_ACTIVATE | HFCPCI_DO_ACTION);
786 cs->dc.hfcpci.ph_state = 1;
787 cs->hw.hfcpci.nt_mode = 1;
788 cs->hw.hfcpci.nt_timer = 0;
789 cs->stlist->l2.l2l1 = dch_nt_l2l1;
790 spin_unlock_irqrestore(&cs->lock, flags);
791 debugl1(cs, "NT mode activated");
792 return (0);
793 }
794 if ((cs->chanlimit > 1) || (cs->hw.hfcpci.bswapped) ||
795 (cs->hw.hfcpci.nt_mode) || (ic->arg != 12))
796 return (-EINVAL);
797
798 spin_lock_irqsave(&cs->lock, flags);
799 if (i) {
800 cs->logecho = 1;
801 cs->hw.hfcpci.trm |= 0x20;
802 cs->hw.hfcpci.int_m1 |= HFCPCI_INTS_B2REC;
803 cs->hw.hfcpci.fifo_en |= HFCPCI_FIFOEN_B2RX;
804 } else {
805 cs->logecho = 0;
806 cs->hw.hfcpci.trm &= ~0x20;
807 cs->hw.hfcpci.int_m1 &= ~HFCPCI_INTS_B2REC;
808 cs->hw.hfcpci.fifo_en &= ~HFCPCI_FIFOEN_B2RX;
809 }
810 cs->hw.hfcpci.sctrl_r &= ~SCTRL_B2_ENA;
811 cs->hw.hfcpci.sctrl &= ~SCTRL_B2_ENA;
812 cs->hw.hfcpci.conn |= 0x10;
813 cs->hw.hfcpci.ctmt &= ~2;
814 Write_hfc(cs, HFCPCI_CTMT, cs->hw.hfcpci.ctmt);
815 Write_hfc(cs, HFCPCI_SCTRL_R, cs->hw.hfcpci.sctrl_r);
816 Write_hfc(cs, HFCPCI_SCTRL, cs->hw.hfcpci.sctrl);
817 Write_hfc(cs, HFCPCI_CONNECT, cs->hw.hfcpci.conn);
818 Write_hfc(cs, HFCPCI_TRM, cs->hw.hfcpci.trm);
819 Write_hfc(cs, HFCPCI_FIFO_EN, cs->hw.hfcpci.fifo_en);
820 Write_hfc(cs, HFCPCI_INT_M1, cs->hw.hfcpci.int_m1);
821 spin_unlock_irqrestore(&cs->lock, flags);
822 return (0);
823}
824
825
826
827
828static void
829receive_emsg(struct IsdnCardState *cs)
830{
831 int rcnt;
832 int receive, count = 5;
833 bzfifo_type *bz;
834 u_char *bdata;
835 z_type *zp;
836 u_char *ptr, *ptr1, new_f2;
837 int total, maxlen, new_z2;
838 u_char e_buffer[256];
839
840 bz = &((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.rxbz_b2;
841 bdata = ((fifo_area *) (cs->hw.hfcpci.fifos))->b_chans.rxdat_b2;
842Begin:
843 count--;
844 if (test_and_set_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
845 debugl1(cs, "echo_rec_data blocked");
846 return;
847 }
848 if (bz->f1 != bz->f2) {
849 if (cs->debug & L1_DEB_ISAC)
850 debugl1(cs, "hfcpci e_rec f1(%d) f2(%d)",
851 bz->f1, bz->f2);
852 zp = &bz->za[bz->f2];
853
854 rcnt = zp->z1 - zp->z2;
855 if (rcnt < 0)
856 rcnt += B_FIFO_SIZE;
857 rcnt++;
858 if (cs->debug & L1_DEB_ISAC)
859 debugl1(cs, "hfcpci e_rec z1(%x) z2(%x) cnt(%d)",
860 zp->z1, zp->z2, rcnt);
861 new_z2 = zp->z2 + rcnt;
862 if (new_z2 >= (B_FIFO_SIZE + B_SUB_VAL))
863 new_z2 -= B_FIFO_SIZE;
864 new_f2 = (bz->f2 + 1) & MAX_B_FRAMES;
865 if ((rcnt > 256 + 3) || (count < 4) ||
866 (*(bdata + (zp->z1 - B_SUB_VAL)))) {
867 if (cs->debug & L1_DEB_WARN)
868 debugl1(cs, "hfcpci_empty_echan: incoming packet invalid length %d or crc", rcnt);
869 bz->za[new_f2].z2 = new_z2;
870 bz->f2 = new_f2;
871 } else {
872 total = rcnt;
873 rcnt -= 3;
874 ptr = e_buffer;
875
876 if (zp->z2 <= B_FIFO_SIZE + B_SUB_VAL)
877 maxlen = rcnt;
878 else
879 maxlen = B_FIFO_SIZE + B_SUB_VAL - zp->z2;
880
881 ptr1 = bdata + (zp->z2 - B_SUB_VAL);
882 memcpy(ptr, ptr1, maxlen);
883 rcnt -= maxlen;
884
885 if (rcnt) {
886 ptr += maxlen;
887 ptr1 = bdata;
888 memcpy(ptr, ptr1, rcnt);
889 }
890 bz->za[new_f2].z2 = new_z2;
891 bz->f2 = new_f2;
892 if (cs->debug & DEB_DLOG_HEX) {
893 ptr = cs->dlog;
894 if ((total - 3) < MAX_DLOG_SPACE / 3 - 10) {
895 *ptr++ = 'E';
896 *ptr++ = 'C';
897 *ptr++ = 'H';
898 *ptr++ = 'O';
899 *ptr++ = ':';
900 ptr += QuickHex(ptr, e_buffer, total - 3);
901 ptr--;
902 *ptr++ = '\n';
903 *ptr = 0;
904 HiSax_putstatus(cs, NULL, "%s", cs->dlog);
905 } else
906 HiSax_putstatus(cs, "LogEcho: ", "warning Frame too big (%d)", total - 3);
907 }
908 }
909
910 rcnt = bz->f1 - bz->f2;
911 if (rcnt < 0)
912 rcnt += MAX_B_FRAMES + 1;
913 if (rcnt > 1)
914 receive = 1;
915 else
916 receive = 0;
917 } else
918 receive = 0;
919 test_and_clear_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags);
920 if (count && receive)
921 goto Begin;
922}
923
924
925
926
927static irqreturn_t
928hfcpci_interrupt(int intno, void *dev_id)
929{
930 u_long flags;
931 struct IsdnCardState *cs = dev_id;
932 u_char exval;
933 struct BCState *bcs;
934 int count = 15;
935 u_char val, stat;
936
937 if (!(cs->hw.hfcpci.int_m2 & 0x08)) {
938 debugl1(cs, "HFC-PCI: int_m2 %x not initialised", cs->hw.hfcpci.int_m2);
939 return IRQ_NONE;
940 }
941 spin_lock_irqsave(&cs->lock, flags);
942 if (HFCPCI_ANYINT & (stat = Read_hfc(cs, HFCPCI_STATUS))) {
943 val = Read_hfc(cs, HFCPCI_INT_S1);
944 if (cs->debug & L1_DEB_ISAC)
945 debugl1(cs, "HFC-PCI: stat(%02x) s1(%02x)", stat, val);
946 } else {
947 spin_unlock_irqrestore(&cs->lock, flags);
948 return IRQ_NONE;
949 }
950 if (cs->debug & L1_DEB_ISAC)
951 debugl1(cs, "HFC-PCI irq %x %s", val,
952 test_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags) ?
953 "locked" : "unlocked");
954 val &= cs->hw.hfcpci.int_m1;
955 if (val & 0x40) {
956 exval = Read_hfc(cs, HFCPCI_STATES) & 0xf;
957 if (cs->debug & L1_DEB_ISAC)
958 debugl1(cs, "ph_state chg %d->%d", cs->dc.hfcpci.ph_state,
959 exval);
960 cs->dc.hfcpci.ph_state = exval;
961 sched_event_D_pci(cs, D_L1STATECHANGE);
962 val &= ~0x40;
963 }
964 if (val & 0x80) {
965 if (cs->hw.hfcpci.nt_mode) {
966 if ((--cs->hw.hfcpci.nt_timer) < 0)
967 sched_event_D_pci(cs, D_L1STATECHANGE);
968 }
969 val &= ~0x80;
970 Write_hfc(cs, HFCPCI_CTMT, cs->hw.hfcpci.ctmt | HFCPCI_CLTIMER);
971 }
972 while (val) {
973 if (test_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
974 cs->hw.hfcpci.int_s1 |= val;
975 spin_unlock_irqrestore(&cs->lock, flags);
976 return IRQ_HANDLED;
977 }
978 if (cs->hw.hfcpci.int_s1 & 0x18) {
979 exval = val;
980 val = cs->hw.hfcpci.int_s1;
981 cs->hw.hfcpci.int_s1 = exval;
982 }
983 if (val & 0x08) {
984 if (!(bcs = Sel_BCS(cs, cs->hw.hfcpci.bswapped ? 1 : 0))) {
985 if (cs->debug)
986 debugl1(cs, "hfcpci spurious 0x08 IRQ");
987 } else
988 main_rec_hfcpci(bcs);
989 }
990 if (val & 0x10) {
991 if (cs->logecho)
992 receive_emsg(cs);
993 else if (!(bcs = Sel_BCS(cs, 1))) {
994 if (cs->debug)
995 debugl1(cs, "hfcpci spurious 0x10 IRQ");
996 } else
997 main_rec_hfcpci(bcs);
998 }
999 if (val & 0x01) {
1000 if (!(bcs = Sel_BCS(cs, cs->hw.hfcpci.bswapped ? 1 : 0))) {
1001 if (cs->debug)
1002 debugl1(cs, "hfcpci spurious 0x01 IRQ");
1003 } else {
1004 if (bcs->tx_skb) {
1005 if (!test_and_set_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
1006 hfcpci_fill_fifo(bcs);
1007 test_and_clear_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags);
1008 } else
1009 debugl1(cs, "fill_data %d blocked", bcs->channel);
1010 } else {
1011 if ((bcs->tx_skb = skb_dequeue(&bcs->squeue))) {
1012 if (!test_and_set_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
1013 hfcpci_fill_fifo(bcs);
1014 test_and_clear_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags);
1015 } else
1016 debugl1(cs, "fill_data %d blocked", bcs->channel);
1017 } else {
1018 hfcpci_sched_event(bcs, B_XMTBUFREADY);
1019 }
1020 }
1021 }
1022 }
1023 if (val & 0x02) {
1024 if (!(bcs = Sel_BCS(cs, 1))) {
1025 if (cs->debug)
1026 debugl1(cs, "hfcpci spurious 0x02 IRQ");
1027 } else {
1028 if (bcs->tx_skb) {
1029 if (!test_and_set_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
1030 hfcpci_fill_fifo(bcs);
1031 test_and_clear_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags);
1032 } else
1033 debugl1(cs, "fill_data %d blocked", bcs->channel);
1034 } else {
1035 if ((bcs->tx_skb = skb_dequeue(&bcs->squeue))) {
1036 if (!test_and_set_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
1037 hfcpci_fill_fifo(bcs);
1038 test_and_clear_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags);
1039 } else
1040 debugl1(cs, "fill_data %d blocked", bcs->channel);
1041 } else {
1042 hfcpci_sched_event(bcs, B_XMTBUFREADY);
1043 }
1044 }
1045 }
1046 }
1047 if (val & 0x20) {
1048 receive_dmsg(cs);
1049 }
1050 if (val & 0x04) {
1051 if (test_and_clear_bit(FLG_DBUSY_TIMER, &cs->HW_Flags))
1052 del_timer(&cs->dbusytimer);
1053 if (test_and_clear_bit(FLG_L1_DBUSY, &cs->HW_Flags))
1054 sched_event_D_pci(cs, D_CLEARBUSY);
1055 if (cs->tx_skb) {
1056 if (cs->tx_skb->len) {
1057 if (!test_and_set_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
1058 hfcpci_fill_dfifo(cs);
1059 test_and_clear_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags);
1060 } else {
1061 debugl1(cs, "hfcpci_fill_dfifo irq blocked");
1062 }
1063 goto afterXPR;
1064 } else {
1065 dev_kfree_skb_irq(cs->tx_skb);
1066 cs->tx_cnt = 0;
1067 cs->tx_skb = NULL;
1068 }
1069 }
1070 if ((cs->tx_skb = skb_dequeue(&cs->sq))) {
1071 cs->tx_cnt = 0;
1072 if (!test_and_set_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
1073 hfcpci_fill_dfifo(cs);
1074 test_and_clear_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags);
1075 } else {
1076 debugl1(cs, "hfcpci_fill_dfifo irq blocked");
1077 }
1078 } else
1079 sched_event_D_pci(cs, D_XMTBUFREADY);
1080 }
1081 afterXPR:
1082 if (cs->hw.hfcpci.int_s1 && count--) {
1083 val = cs->hw.hfcpci.int_s1;
1084 cs->hw.hfcpci.int_s1 = 0;
1085 if (cs->debug & L1_DEB_ISAC)
1086 debugl1(cs, "HFC-PCI irq %x loop %d", val, 15 - count);
1087 } else
1088 val = 0;
1089 }
1090 spin_unlock_irqrestore(&cs->lock, flags);
1091 return IRQ_HANDLED;
1092}
1093
1094
1095
1096
1097static void
1098hfcpci_dbusy_timer(struct IsdnCardState *cs)
1099{
1100}
1101
1102
1103
1104
1105static void
1106HFCPCI_l1hw(struct PStack *st, int pr, void *arg)
1107{
1108 u_long flags;
1109 struct IsdnCardState *cs = (struct IsdnCardState *) st->l1.hardware;
1110 struct sk_buff *skb = arg;
1111
1112 switch (pr) {
1113 case (PH_DATA | REQUEST):
1114 if (cs->debug & DEB_DLOG_HEX)
1115 LogFrame(cs, skb->data, skb->len);
1116 if (cs->debug & DEB_DLOG_VERBOSE)
1117 dlogframe(cs, skb, 0);
1118 spin_lock_irqsave(&cs->lock, flags);
1119 if (cs->tx_skb) {
1120 skb_queue_tail(&cs->sq, skb);
1121#ifdef L2FRAME_DEBUG
1122 if (cs->debug & L1_DEB_LAPD)
1123 Logl2Frame(cs, skb, "PH_DATA Queued", 0);
1124#endif
1125 } else {
1126 cs->tx_skb = skb;
1127 cs->tx_cnt = 0;
1128#ifdef L2FRAME_DEBUG
1129 if (cs->debug & L1_DEB_LAPD)
1130 Logl2Frame(cs, skb, "PH_DATA", 0);
1131#endif
1132 if (!test_and_set_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
1133 hfcpci_fill_dfifo(cs);
1134 test_and_clear_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags);
1135 } else
1136 debugl1(cs, "hfcpci_fill_dfifo blocked");
1137
1138 }
1139 spin_unlock_irqrestore(&cs->lock, flags);
1140 break;
1141 case (PH_PULL | INDICATION):
1142 spin_lock_irqsave(&cs->lock, flags);
1143 if (cs->tx_skb) {
1144 if (cs->debug & L1_DEB_WARN)
1145 debugl1(cs, " l2l1 tx_skb exist this shouldn't happen");
1146 skb_queue_tail(&cs->sq, skb);
1147 spin_unlock_irqrestore(&cs->lock, flags);
1148 break;
1149 }
1150 if (cs->debug & DEB_DLOG_HEX)
1151 LogFrame(cs, skb->data, skb->len);
1152 if (cs->debug & DEB_DLOG_VERBOSE)
1153 dlogframe(cs, skb, 0);
1154 cs->tx_skb = skb;
1155 cs->tx_cnt = 0;
1156#ifdef L2FRAME_DEBUG
1157 if (cs->debug & L1_DEB_LAPD)
1158 Logl2Frame(cs, skb, "PH_DATA_PULLED", 0);
1159#endif
1160 if (!test_and_set_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
1161 hfcpci_fill_dfifo(cs);
1162 test_and_clear_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags);
1163 } else
1164 debugl1(cs, "hfcpci_fill_dfifo blocked");
1165 spin_unlock_irqrestore(&cs->lock, flags);
1166 break;
1167 case (PH_PULL | REQUEST):
1168#ifdef L2FRAME_DEBUG
1169 if (cs->debug & L1_DEB_LAPD)
1170 debugl1(cs, "-> PH_REQUEST_PULL");
1171#endif
1172 if (!cs->tx_skb) {
1173 test_and_clear_bit(FLG_L1_PULL_REQ, &st->l1.Flags);
1174 st->l1.l1l2(st, PH_PULL | CONFIRM, NULL);
1175 } else
1176 test_and_set_bit(FLG_L1_PULL_REQ, &st->l1.Flags);
1177 break;
1178 case (HW_RESET | REQUEST):
1179 spin_lock_irqsave(&cs->lock, flags);
1180 Write_hfc(cs, HFCPCI_STATES, HFCPCI_LOAD_STATE | 3);
1181 udelay(6);
1182 Write_hfc(cs, HFCPCI_STATES, 3);
1183 cs->hw.hfcpci.mst_m |= HFCPCI_MASTER;
1184 Write_hfc(cs, HFCPCI_MST_MODE, cs->hw.hfcpci.mst_m);
1185 Write_hfc(cs, HFCPCI_STATES, HFCPCI_ACTIVATE | HFCPCI_DO_ACTION);
1186 spin_unlock_irqrestore(&cs->lock, flags);
1187 l1_msg(cs, HW_POWERUP | CONFIRM, NULL);
1188 break;
1189 case (HW_ENABLE | REQUEST):
1190 spin_lock_irqsave(&cs->lock, flags);
1191 Write_hfc(cs, HFCPCI_STATES, HFCPCI_DO_ACTION);
1192 spin_unlock_irqrestore(&cs->lock, flags);
1193 break;
1194 case (HW_DEACTIVATE | REQUEST):
1195 spin_lock_irqsave(&cs->lock, flags);
1196 cs->hw.hfcpci.mst_m &= ~HFCPCI_MASTER;
1197 Write_hfc(cs, HFCPCI_MST_MODE, cs->hw.hfcpci.mst_m);
1198 spin_unlock_irqrestore(&cs->lock, flags);
1199 break;
1200 case (HW_INFO3 | REQUEST):
1201 spin_lock_irqsave(&cs->lock, flags);
1202 cs->hw.hfcpci.mst_m |= HFCPCI_MASTER;
1203 Write_hfc(cs, HFCPCI_MST_MODE, cs->hw.hfcpci.mst_m);
1204 spin_unlock_irqrestore(&cs->lock, flags);
1205 break;
1206 case (HW_TESTLOOP | REQUEST):
1207 spin_lock_irqsave(&cs->lock, flags);
1208 switch ((long) arg) {
1209 case (1):
1210 Write_hfc(cs, HFCPCI_B1_SSL, 0x80);
1211 Write_hfc(cs, HFCPCI_B1_RSL, 0x80);
1212 cs->hw.hfcpci.conn = (cs->hw.hfcpci.conn & ~7) | 1;
1213 Write_hfc(cs, HFCPCI_CONNECT, cs->hw.hfcpci.conn);
1214 break;
1215
1216 case (2):
1217 Write_hfc(cs, HFCPCI_B2_SSL, 0x81);
1218 Write_hfc(cs, HFCPCI_B2_RSL, 0x81);
1219 cs->hw.hfcpci.conn = (cs->hw.hfcpci.conn & ~0x38) | 0x08;
1220 Write_hfc(cs, HFCPCI_CONNECT, cs->hw.hfcpci.conn);
1221 break;
1222
1223 default:
1224 spin_unlock_irqrestore(&cs->lock, flags);
1225 if (cs->debug & L1_DEB_WARN)
1226 debugl1(cs, "hfcpci_l1hw loop invalid %4lx", (long) arg);
1227 return;
1228 }
1229 cs->hw.hfcpci.trm |= 0x80;
1230 Write_hfc(cs, HFCPCI_TRM, cs->hw.hfcpci.trm);
1231 spin_unlock_irqrestore(&cs->lock, flags);
1232 break;
1233 default:
1234 if (cs->debug & L1_DEB_WARN)
1235 debugl1(cs, "hfcpci_l1hw unknown pr %4x", pr);
1236 break;
1237 }
1238}
1239
1240
1241
1242
1243static void
1244setstack_hfcpci(struct PStack *st, struct IsdnCardState *cs)
1245{
1246 st->l1.l1hw = HFCPCI_l1hw;
1247}
1248
1249
1250
1251
1252static void
1253hfcpci_send_data(struct BCState *bcs)
1254{
1255 struct IsdnCardState *cs = bcs->cs;
1256
1257 if (!test_and_set_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags)) {
1258 hfcpci_fill_fifo(bcs);
1259 test_and_clear_bit(FLG_LOCK_ATOMIC, &cs->HW_Flags);
1260 } else
1261 debugl1(cs, "send_data %d blocked", bcs->channel);
1262}
1263
1264
1265
1266
1267static void
1268mode_hfcpci(struct BCState *bcs, int mode, int bc)
1269{
1270 struct IsdnCardState *cs = bcs->cs;
1271 int fifo2;
1272
1273 if (cs->debug & L1_DEB_HSCX)
1274 debugl1(cs, "HFCPCI bchannel mode %d bchan %d/%d",
1275 mode, bc, bcs->channel);
1276 bcs->mode = mode;
1277 bcs->channel = bc;
1278 fifo2 = bc;
1279 if (cs->chanlimit > 1) {
1280 cs->hw.hfcpci.bswapped = 0;
1281 cs->hw.hfcpci.sctrl_e &= ~0x80;
1282 } else {
1283 if (bc) {
1284 if (mode != L1_MODE_NULL) {
1285 cs->hw.hfcpci.bswapped = 1;
1286 cs->hw.hfcpci.sctrl_e |= 0x80;
1287 } else {
1288 cs->hw.hfcpci.bswapped = 0;
1289 cs->hw.hfcpci.sctrl_e &= ~0x80;
1290 }
1291 fifo2 = 0;
1292 } else {
1293 cs->hw.hfcpci.bswapped = 0;
1294 cs->hw.hfcpci.sctrl_e &= ~0x80;
1295 }
1296 }
1297 switch (mode) {
1298 case (L1_MODE_NULL):
1299 if (bc) {
1300 cs->hw.hfcpci.sctrl &= ~SCTRL_B2_ENA;
1301 cs->hw.hfcpci.sctrl_r &= ~SCTRL_B2_ENA;
1302 } else {
1303 cs->hw.hfcpci.sctrl &= ~SCTRL_B1_ENA;
1304 cs->hw.hfcpci.sctrl_r &= ~SCTRL_B1_ENA;
1305 }
1306 if (fifo2) {
1307 cs->hw.hfcpci.fifo_en &= ~HFCPCI_FIFOEN_B2;
1308 cs->hw.hfcpci.int_m1 &= ~(HFCPCI_INTS_B2TRANS + HFCPCI_INTS_B2REC);
1309 } else {
1310 cs->hw.hfcpci.fifo_en &= ~HFCPCI_FIFOEN_B1;
1311 cs->hw.hfcpci.int_m1 &= ~(HFCPCI_INTS_B1TRANS + HFCPCI_INTS_B1REC);
1312 }
1313 break;
1314 case (L1_MODE_TRANS):
1315 hfcpci_clear_fifo_rx(cs, fifo2);
1316 hfcpci_clear_fifo_tx(cs, fifo2);
1317 if (bc) {
1318 cs->hw.hfcpci.sctrl |= SCTRL_B2_ENA;
1319 cs->hw.hfcpci.sctrl_r |= SCTRL_B2_ENA;
1320 } else {
1321 cs->hw.hfcpci.sctrl |= SCTRL_B1_ENA;
1322 cs->hw.hfcpci.sctrl_r |= SCTRL_B1_ENA;
1323 }
1324 if (fifo2) {
1325 cs->hw.hfcpci.fifo_en |= HFCPCI_FIFOEN_B2;
1326 cs->hw.hfcpci.int_m1 |= (HFCPCI_INTS_B2TRANS + HFCPCI_INTS_B2REC);
1327 cs->hw.hfcpci.ctmt |= 2;
1328 cs->hw.hfcpci.conn &= ~0x18;
1329 } else {
1330 cs->hw.hfcpci.fifo_en |= HFCPCI_FIFOEN_B1;
1331 cs->hw.hfcpci.int_m1 |= (HFCPCI_INTS_B1TRANS + HFCPCI_INTS_B1REC);
1332 cs->hw.hfcpci.ctmt |= 1;
1333 cs->hw.hfcpci.conn &= ~0x03;
1334 }
1335 break;
1336 case (L1_MODE_HDLC):
1337 hfcpci_clear_fifo_rx(cs, fifo2);
1338 hfcpci_clear_fifo_tx(cs, fifo2);
1339 if (bc) {
1340 cs->hw.hfcpci.sctrl |= SCTRL_B2_ENA;
1341 cs->hw.hfcpci.sctrl_r |= SCTRL_B2_ENA;
1342 } else {
1343 cs->hw.hfcpci.sctrl |= SCTRL_B1_ENA;
1344 cs->hw.hfcpci.sctrl_r |= SCTRL_B1_ENA;
1345 }
1346 if (fifo2) {
1347 cs->hw.hfcpci.last_bfifo_cnt[1] = 0;
1348 cs->hw.hfcpci.fifo_en |= HFCPCI_FIFOEN_B2;
1349 cs->hw.hfcpci.int_m1 |= (HFCPCI_INTS_B2TRANS + HFCPCI_INTS_B2REC);
1350 cs->hw.hfcpci.ctmt &= ~2;
1351 cs->hw.hfcpci.conn &= ~0x18;
1352 } else {
1353 cs->hw.hfcpci.last_bfifo_cnt[0] = 0;
1354 cs->hw.hfcpci.fifo_en |= HFCPCI_FIFOEN_B1;
1355 cs->hw.hfcpci.int_m1 |= (HFCPCI_INTS_B1TRANS + HFCPCI_INTS_B1REC);
1356 cs->hw.hfcpci.ctmt &= ~1;
1357 cs->hw.hfcpci.conn &= ~0x03;
1358 }
1359 break;
1360 case (L1_MODE_EXTRN):
1361 if (bc) {
1362 cs->hw.hfcpci.conn |= 0x10;
1363 cs->hw.hfcpci.sctrl |= SCTRL_B2_ENA;
1364 cs->hw.hfcpci.sctrl_r |= SCTRL_B2_ENA;
1365 cs->hw.hfcpci.fifo_en &= ~HFCPCI_FIFOEN_B2;
1366 cs->hw.hfcpci.int_m1 &= ~(HFCPCI_INTS_B2TRANS + HFCPCI_INTS_B2REC);
1367 } else {
1368 cs->hw.hfcpci.conn |= 0x02;
1369 cs->hw.hfcpci.sctrl |= SCTRL_B1_ENA;
1370 cs->hw.hfcpci.sctrl_r |= SCTRL_B1_ENA;
1371 cs->hw.hfcpci.fifo_en &= ~HFCPCI_FIFOEN_B1;
1372 cs->hw.hfcpci.int_m1 &= ~(HFCPCI_INTS_B1TRANS + HFCPCI_INTS_B1REC);
1373 }
1374 break;
1375 }
1376 Write_hfc(cs, HFCPCI_SCTRL_E, cs->hw.hfcpci.sctrl_e);
1377 Write_hfc(cs, HFCPCI_INT_M1, cs->hw.hfcpci.int_m1);
1378 Write_hfc(cs, HFCPCI_FIFO_EN, cs->hw.hfcpci.fifo_en);
1379 Write_hfc(cs, HFCPCI_SCTRL, cs->hw.hfcpci.sctrl);
1380 Write_hfc(cs, HFCPCI_SCTRL_R, cs->hw.hfcpci.sctrl_r);
1381 Write_hfc(cs, HFCPCI_CTMT, cs->hw.hfcpci.ctmt);
1382 Write_hfc(cs, HFCPCI_CONNECT, cs->hw.hfcpci.conn);
1383}
1384
1385
1386
1387
1388static void
1389hfcpci_l2l1(struct PStack *st, int pr, void *arg)
1390{
1391 struct BCState *bcs = st->l1.bcs;
1392 u_long flags;
1393 struct sk_buff *skb = arg;
1394
1395 switch (pr) {
1396 case (PH_DATA | REQUEST):
1397 spin_lock_irqsave(&bcs->cs->lock, flags);
1398 if (bcs->tx_skb) {
1399 skb_queue_tail(&bcs->squeue, skb);
1400 } else {
1401 bcs->tx_skb = skb;
1402
1403 bcs->cs->BC_Send_Data(bcs);
1404 }
1405 spin_unlock_irqrestore(&bcs->cs->lock, flags);
1406 break;
1407 case (PH_PULL | INDICATION):
1408 spin_lock_irqsave(&bcs->cs->lock, flags);
1409 if (bcs->tx_skb) {
1410 spin_unlock_irqrestore(&bcs->cs->lock, flags);
1411 printk(KERN_WARNING "hfc_l2l1: this shouldn't happen\n");
1412 break;
1413 }
1414
1415 bcs->tx_skb = skb;
1416 bcs->cs->BC_Send_Data(bcs);
1417 spin_unlock_irqrestore(&bcs->cs->lock, flags);
1418 break;
1419 case (PH_PULL | REQUEST):
1420 if (!bcs->tx_skb) {
1421 test_and_clear_bit(FLG_L1_PULL_REQ, &st->l1.Flags);
1422 st->l1.l1l2(st, PH_PULL | CONFIRM, NULL);
1423 } else
1424 test_and_set_bit(FLG_L1_PULL_REQ, &st->l1.Flags);
1425 break;
1426 case (PH_ACTIVATE | REQUEST):
1427 spin_lock_irqsave(&bcs->cs->lock, flags);
1428 test_and_set_bit(BC_FLG_ACTIV, &bcs->Flag);
1429 mode_hfcpci(bcs, st->l1.mode, st->l1.bc);
1430 spin_unlock_irqrestore(&bcs->cs->lock, flags);
1431 l1_msg_b(st, pr, arg);
1432 break;
1433 case (PH_DEACTIVATE | REQUEST):
1434 l1_msg_b(st, pr, arg);
1435 break;
1436 case (PH_DEACTIVATE | CONFIRM):
1437 spin_lock_irqsave(&bcs->cs->lock, flags);
1438 test_and_clear_bit(BC_FLG_ACTIV, &bcs->Flag);
1439 test_and_clear_bit(BC_FLG_BUSY, &bcs->Flag);
1440 mode_hfcpci(bcs, 0, st->l1.bc);
1441 spin_unlock_irqrestore(&bcs->cs->lock, flags);
1442 st->l1.l1l2(st, PH_DEACTIVATE | CONFIRM, NULL);
1443 break;
1444 }
1445}
1446
1447
1448
1449
1450static void
1451close_hfcpci(struct BCState *bcs)
1452{
1453 mode_hfcpci(bcs, 0, bcs->channel);
1454 if (test_and_clear_bit(BC_FLG_INIT, &bcs->Flag)) {
1455 skb_queue_purge(&bcs->rqueue);
1456 skb_queue_purge(&bcs->squeue);
1457 if (bcs->tx_skb) {
1458 dev_kfree_skb_any(bcs->tx_skb);
1459 bcs->tx_skb = NULL;
1460 test_and_clear_bit(BC_FLG_BUSY, &bcs->Flag);
1461 }
1462 }
1463}
1464
1465
1466
1467
1468static int
1469open_hfcpcistate(struct IsdnCardState *cs, struct BCState *bcs)
1470{
1471 if (!test_and_set_bit(BC_FLG_INIT, &bcs->Flag)) {
1472 skb_queue_head_init(&bcs->rqueue);
1473 skb_queue_head_init(&bcs->squeue);
1474 }
1475 bcs->tx_skb = NULL;
1476 test_and_clear_bit(BC_FLG_BUSY, &bcs->Flag);
1477 bcs->event = 0;
1478 bcs->tx_cnt = 0;
1479 return (0);
1480}
1481
1482
1483
1484
1485static int
1486setstack_2b(struct PStack *st, struct BCState *bcs)
1487{
1488 bcs->channel = st->l1.bc;
1489 if (open_hfcpcistate(st->l1.hardware, bcs))
1490 return (-1);
1491 st->l1.bcs = bcs;
1492 st->l2.l2l1 = hfcpci_l2l1;
1493 setstack_manager(st);
1494 bcs->st = st;
1495 setstack_l1_B(st);
1496 return (0);
1497}
1498
1499
1500
1501
1502static void
1503hfcpci_bh(struct work_struct *work)
1504{
1505 struct IsdnCardState *cs =
1506 container_of(work, struct IsdnCardState, tqueue);
1507 u_long flags;
1508
1509
1510 if (test_and_clear_bit(D_L1STATECHANGE, &cs->event)) {
1511 if (!cs->hw.hfcpci.nt_mode)
1512 switch (cs->dc.hfcpci.ph_state) {
1513 case (0):
1514 l1_msg(cs, HW_RESET | INDICATION, NULL);
1515 break;
1516 case (3):
1517 l1_msg(cs, HW_DEACTIVATE | INDICATION, NULL);
1518 break;
1519 case (8):
1520 l1_msg(cs, HW_RSYNC | INDICATION, NULL);
1521 break;
1522 case (6):
1523 l1_msg(cs, HW_INFO2 | INDICATION, NULL);
1524 break;
1525 case (7):
1526 l1_msg(cs, HW_INFO4_P8 | INDICATION, NULL);
1527 break;
1528 default:
1529 break;
1530 } else {
1531 spin_lock_irqsave(&cs->lock, flags);
1532 switch (cs->dc.hfcpci.ph_state) {
1533 case (2):
1534 if (cs->hw.hfcpci.nt_timer < 0) {
1535 cs->hw.hfcpci.nt_timer = 0;
1536 cs->hw.hfcpci.int_m1 &= ~HFCPCI_INTS_TIMER;
1537 Write_hfc(cs, HFCPCI_INT_M1, cs->hw.hfcpci.int_m1);
1538
1539 if (Read_hfc(cs, HFCPCI_INT_S1));
1540 Write_hfc(cs, HFCPCI_STATES, 4 | HFCPCI_LOAD_STATE);
1541 udelay(10);
1542 Write_hfc(cs, HFCPCI_STATES, 4);
1543 cs->dc.hfcpci.ph_state = 4;
1544 } else {
1545 cs->hw.hfcpci.int_m1 |= HFCPCI_INTS_TIMER;
1546 Write_hfc(cs, HFCPCI_INT_M1, cs->hw.hfcpci.int_m1);
1547 cs->hw.hfcpci.ctmt &= ~HFCPCI_AUTO_TIMER;
1548 cs->hw.hfcpci.ctmt |= HFCPCI_TIM3_125;
1549 Write_hfc(cs, HFCPCI_CTMT, cs->hw.hfcpci.ctmt | HFCPCI_CLTIMER);
1550 Write_hfc(cs, HFCPCI_CTMT, cs->hw.hfcpci.ctmt | HFCPCI_CLTIMER);
1551 cs->hw.hfcpci.nt_timer = NT_T1_COUNT;
1552 Write_hfc(cs, HFCPCI_STATES, 2 | HFCPCI_NT_G2_G3);
1553 }
1554 break;
1555 case (1):
1556 case (3):
1557 case (4):
1558 cs->hw.hfcpci.nt_timer = 0;
1559 cs->hw.hfcpci.int_m1 &= ~HFCPCI_INTS_TIMER;
1560 Write_hfc(cs, HFCPCI_INT_M1, cs->hw.hfcpci.int_m1);
1561 break;
1562 default:
1563 break;
1564 }
1565 spin_unlock_irqrestore(&cs->lock, flags);
1566 }
1567 }
1568 if (test_and_clear_bit(D_RCVBUFREADY, &cs->event))
1569 DChannel_proc_rcv(cs);
1570 if (test_and_clear_bit(D_XMTBUFREADY, &cs->event))
1571 DChannel_proc_xmt(cs);
1572}
1573
1574
1575
1576
1577
1578static void
1579inithfcpci(struct IsdnCardState *cs)
1580{
1581 cs->bcs[0].BC_SetStack = setstack_2b;
1582 cs->bcs[1].BC_SetStack = setstack_2b;
1583 cs->bcs[0].BC_Close = close_hfcpci;
1584 cs->bcs[1].BC_Close = close_hfcpci;
1585 cs->dbusytimer.function = (void *) hfcpci_dbusy_timer;
1586 cs->dbusytimer.data = (long) cs;
1587 init_timer(&cs->dbusytimer);
1588 mode_hfcpci(cs->bcs, 0, 0);
1589 mode_hfcpci(cs->bcs + 1, 0, 1);
1590}
1591
1592
1593
1594
1595
1596
1597static int
1598hfcpci_card_msg(struct IsdnCardState *cs, int mt, void *arg)
1599{
1600 u_long flags;
1601
1602 if (cs->debug & L1_DEB_ISAC)
1603 debugl1(cs, "HFCPCI: card_msg %x", mt);
1604 switch (mt) {
1605 case CARD_RESET:
1606 spin_lock_irqsave(&cs->lock, flags);
1607 reset_hfcpci(cs);
1608 spin_unlock_irqrestore(&cs->lock, flags);
1609 return (0);
1610 case CARD_RELEASE:
1611 release_io_hfcpci(cs);
1612 return (0);
1613 case CARD_INIT:
1614 spin_lock_irqsave(&cs->lock, flags);
1615 inithfcpci(cs);
1616 reset_hfcpci(cs);
1617 spin_unlock_irqrestore(&cs->lock, flags);
1618 msleep(80);
1619
1620 spin_lock_irqsave(&cs->lock, flags);
1621 cs->hw.hfcpci.int_m1 &= ~HFCPCI_INTS_TIMER;
1622 Write_hfc(cs, HFCPCI_INT_M1, cs->hw.hfcpci.int_m1);
1623
1624 Write_hfc(cs, HFCPCI_MST_MODE, cs->hw.hfcpci.mst_m);
1625 spin_unlock_irqrestore(&cs->lock, flags);
1626 return (0);
1627 case CARD_TEST:
1628 return (0);
1629 }
1630 return (0);
1631}
1632
1633
1634
1635static struct pci_dev *dev_hfcpci = NULL;
1636
1637int
1638setup_hfcpci(struct IsdnCard *card)
1639{
1640 u_long flags;
1641 struct IsdnCardState *cs = card->cs;
1642 char tmp[64];
1643 int i;
1644 struct pci_dev *tmp_hfcpci = NULL;
1645
1646#ifdef __BIG_ENDIAN
1647#error "not running on big endian machines now"
1648#endif
1649
1650 strcpy(tmp, hfcpci_revision);
1651 printk(KERN_INFO "HiSax: HFC-PCI driver Rev. %s\n", HiSax_getrev(tmp));
1652
1653 cs->hw.hfcpci.int_s1 = 0;
1654 cs->dc.hfcpci.ph_state = 0;
1655 cs->hw.hfcpci.fifo = 255;
1656 if (cs->typ != ISDN_CTYPE_HFC_PCI)
1657 return (0);
1658
1659 i = 0;
1660 while (id_list[i].vendor_id) {
1661 tmp_hfcpci = hisax_find_pci_device(id_list[i].vendor_id,
1662 id_list[i].device_id,
1663 dev_hfcpci);
1664 i++;
1665 if (tmp_hfcpci) {
1666 dma_addr_t dma_mask = DMA_BIT_MASK(32) & ~0x7fffUL;
1667 if (pci_enable_device(tmp_hfcpci))
1668 continue;
1669 if (pci_set_dma_mask(tmp_hfcpci, dma_mask)) {
1670 printk(KERN_WARNING
1671 "HiSax hfc_pci: No suitable DMA available.\n");
1672 continue;
1673 }
1674 if (pci_set_consistent_dma_mask(tmp_hfcpci, dma_mask)) {
1675 printk(KERN_WARNING
1676 "HiSax hfc_pci: No suitable consistent DMA available.\n");
1677 continue;
1678 }
1679 pci_set_master(tmp_hfcpci);
1680 if ((card->para[0]) && (card->para[0] != (tmp_hfcpci->resource[0].start & PCI_BASE_ADDRESS_IO_MASK)))
1681 continue;
1682 else
1683 break;
1684 }
1685 }
1686
1687 if (!tmp_hfcpci) {
1688 printk(KERN_WARNING "HFC-PCI: No PCI card found\n");
1689 return (0);
1690 }
1691
1692 i--;
1693 dev_hfcpci = tmp_hfcpci;
1694 cs->hw.hfcpci.dev = dev_hfcpci;
1695 cs->irq = dev_hfcpci->irq;
1696 if (!cs->irq) {
1697 printk(KERN_WARNING "HFC-PCI: No IRQ for PCI card found\n");
1698 return (0);
1699 }
1700 cs->hw.hfcpci.pci_io = (char *)(unsigned long)dev_hfcpci->resource[1].start;
1701 printk(KERN_INFO "HiSax: HFC-PCI card manufacturer: %s card name: %s\n", id_list[i].vendor_name, id_list[i].card_name);
1702
1703 if (!cs->hw.hfcpci.pci_io) {
1704 printk(KERN_WARNING "HFC-PCI: No IO-Mem for PCI card found\n");
1705 return (0);
1706 }
1707
1708
1709 cs->hw.hfcpci.fifos = pci_alloc_consistent(cs->hw.hfcpci.dev,
1710 0x8000, &cs->hw.hfcpci.dma);
1711 if (!cs->hw.hfcpci.fifos) {
1712 printk(KERN_WARNING "HFC-PCI: Error allocating FIFO memory!\n");
1713 return 0;
1714 }
1715 if (cs->hw.hfcpci.dma & 0x7fff) {
1716 printk(KERN_WARNING
1717 "HFC-PCI: Error DMA memory not on 32K boundary (%lx)\n",
1718 (u_long)cs->hw.hfcpci.dma);
1719 pci_free_consistent(cs->hw.hfcpci.dev, 0x8000,
1720 cs->hw.hfcpci.fifos, cs->hw.hfcpci.dma);
1721 return 0;
1722 }
1723 pci_write_config_dword(cs->hw.hfcpci.dev, 0x80, (u32)cs->hw.hfcpci.dma);
1724 cs->hw.hfcpci.pci_io = ioremap((ulong) cs->hw.hfcpci.pci_io, 256);
1725 printk(KERN_INFO
1726 "HFC-PCI: defined at mem %p fifo %p(%lx) IRQ %d HZ %d\n",
1727 cs->hw.hfcpci.pci_io,
1728 cs->hw.hfcpci.fifos,
1729 (u_long)cs->hw.hfcpci.dma,
1730 cs->irq, HZ);
1731
1732 spin_lock_irqsave(&cs->lock, flags);
1733
1734 pci_write_config_word(cs->hw.hfcpci.dev, PCI_COMMAND, PCI_ENA_MEMIO);
1735 cs->hw.hfcpci.int_m2 = 0;
1736 cs->hw.hfcpci.int_m1 = 0;
1737 Write_hfc(cs, HFCPCI_INT_M1, cs->hw.hfcpci.int_m1);
1738 Write_hfc(cs, HFCPCI_INT_M2, cs->hw.hfcpci.int_m2);
1739
1740
1741
1742 INIT_WORK(&cs->tqueue, hfcpci_bh);
1743 cs->setstack_d = setstack_hfcpci;
1744 cs->BC_Send_Data = &hfcpci_send_data;
1745 cs->readisac = NULL;
1746 cs->writeisac = NULL;
1747 cs->readisacfifo = NULL;
1748 cs->writeisacfifo = NULL;
1749 cs->BC_Read_Reg = NULL;
1750 cs->BC_Write_Reg = NULL;
1751 cs->irq_func = &hfcpci_interrupt;
1752 cs->irq_flags |= IRQF_SHARED;
1753 cs->hw.hfcpci.timer.function = (void *) hfcpci_Timer;
1754 cs->hw.hfcpci.timer.data = (long) cs;
1755 init_timer(&cs->hw.hfcpci.timer);
1756 cs->cardmsg = &hfcpci_card_msg;
1757 cs->auxcmd = &hfcpci_auxcmd;
1758
1759 spin_unlock_irqrestore(&cs->lock, flags);
1760
1761 return (1);
1762}
1763