xref: /netbsd-src/sys/dev/tc/if_le_ioasic.c (revision 4472dbe5e3bd91ef2540bada7a7ca7384627ff9b)
1 /*	$NetBSD: if_le_ioasic.c,v 1.15 2000/03/30 12:45:43 augustss Exp $	*/
2 
3 /*
4  * Copyright (c) 1996 Carnegie-Mellon University.
5  * All rights reserved.
6  *
7  * Author: Chris G. Demetriou
8  *
9  * Permission to use, copy, modify and distribute this software and
10  * its documentation is hereby granted, provided that both the copyright
11  * notice and this permission notice appear in all copies of the
12  * software, derivative works or modified versions, and any portions
13  * thereof, and that both notices appear in supporting documentation.
14  *
15  * CARNEGIE MELLON ALLOWS FREE USE OF THIS SOFTWARE IN ITS "AS IS"
16  * CONDITION.  CARNEGIE MELLON DISCLAIMS ANY LIABILITY OF ANY KIND
17  * FOR ANY DAMAGES WHATSOEVER RESULTING FROM THE USE OF THIS SOFTWARE.
18  *
19  * Carnegie Mellon requests users of this software to return to
20  *
21  *  Software Distribution Coordinator  or  Software.Distribution@CS.CMU.EDU
22  *  School of Computer Science
23  *  Carnegie Mellon University
24  *  Pittsburgh PA 15213-3890
25  *
26  * any improvements or extensions that they make and grant Carnegie the
27  * rights to redistribute these changes.
28  */
29 
30 /*
31  * LANCE on DEC IOCTL ASIC.
32  */
33 
34 #include <sys/cdefs.h>			/* RCS ID &  macro defns */
35 __KERNEL_RCSID(0, "$NetBSD: if_le_ioasic.c,v 1.15 2000/03/30 12:45:43 augustss Exp $");
36 
37 #include "opt_inet.h"
38 
39 #include <sys/param.h>
40 #include <sys/systm.h>
41 #include <sys/mbuf.h>
42 #include <sys/syslog.h>
43 #include <sys/socket.h>
44 #include <sys/device.h>
45 
46 #include <net/if.h>
47 #include <net/if_ether.h>
48 #include <net/if_media.h>
49 
50 #ifdef INET
51 #include <netinet/in.h>
52 #include <netinet/if_inarp.h>
53 #endif
54 
55 #include <dev/ic/lancereg.h>
56 #include <dev/ic/lancevar.h>
57 #include <dev/ic/am7990reg.h>
58 #include <dev/ic/am7990var.h>
59 
60 #include <dev/tc/if_levar.h>
61 #include <dev/tc/tcvar.h>
62 #include <dev/tc/ioasicreg.h>
63 #include <dev/tc/ioasicvar.h>
64 
65 #if defined(_KERNEL) && !defined(_LKM)
66 #include "opt_ddb.h"
67 #endif
68 
69 caddr_t le_iomem;
70 
71 static int  le_ioasic_match __P((struct device *, struct cfdata *, void *));
72 static void le_ioasic_attach __P((struct device *, struct device *, void *));
73 
74 struct cfattach le_ioasic_ca = {
75 	sizeof(struct le_softc), le_ioasic_match, le_ioasic_attach
76 };
77 
78 static void ioasic_lance_dma_setup __P((struct device *));
79 static char *ioasic_lance_ether_address __P((void));
80 
81 #ifdef DDB
82 #define	integrate
83 #define hide
84 #else
85 #define	integrate	static __inline
86 #define hide		static
87 #endif
88 
89 hide void le_ioasic_copytobuf_gap2 __P((struct lance_softc *, void *,
90 	    int, int));
91 hide void le_ioasic_copyfrombuf_gap2 __P((struct lance_softc *, void *,
92 	    int, int));
93 
94 hide void le_ioasic_copytobuf_gap16 __P((struct lance_softc *, void *,
95 	    int, int));
96 hide void le_ioasic_copyfrombuf_gap16 __P((struct lance_softc *, void *,
97 	    int, int));
98 hide void le_ioasic_zerobuf_gap16 __P((struct lance_softc *, int, int));
99 
100 int
101 le_ioasic_match(parent, match, aux)
102 	struct device *parent;
103 	struct cfdata *match;
104 	void *aux;
105 {
106 	struct ioasicdev_attach_args *d = aux;
107 
108 	if (!ioasic_submatch(match, aux))
109 		return (0);
110 	if (strncmp("lance", d->iada_modname, TC_ROM_LLEN))
111 		return (0);
112 
113 	return (1);
114 }
115 
116 void
117 le_ioasic_attach(parent, self, aux)
118 	struct device *parent, *self;
119 	void *aux;
120 {
121 	struct ioasicdev_attach_args *d = aux;
122 	struct le_softc *lesc = (void *)self;
123 	struct lance_softc *sc = &lesc->sc_am7990.lsc;
124 
125 	ioasic_lance_dma_setup(parent);
126 
127 	if (le_iomem == 0) {
128 		printf("%s: DMA area not set up\n", sc->sc_dev.dv_xname);
129 		return;
130 	}
131 
132 	lesc->sc_r1 = (struct lereg1 *)
133 		TC_DENSE_TO_SPARSE(TC_PHYS_TO_UNCACHED(d->iada_addr));
134 	sc->sc_mem = (void *)TC_PHYS_TO_UNCACHED(le_iomem);
135 
136 	sc->sc_copytodesc = le_ioasic_copytobuf_gap2;
137 	sc->sc_copyfromdesc = le_ioasic_copyfrombuf_gap2;
138 	sc->sc_copytobuf = le_ioasic_copytobuf_gap16;
139 	sc->sc_copyfrombuf = le_ioasic_copyfrombuf_gap16;
140 	sc->sc_zerobuf = le_ioasic_zerobuf_gap16;
141 
142 	dec_le_common_attach(&lesc->sc_am7990, ioasic_lance_ether_address());
143 
144 	ioasic_intr_establish(parent, d->iada_cookie, TC_IPL_NET,
145 	    am7990_intr, sc);
146 }
147 
148 /*
149  * Special memory access functions needed by ioasic-attached LANCE
150  * chips.
151  */
152 
153 /*
154  * gap2: two bytes of data followed by two bytes of pad.
155  *
156  * Buffers must be 4-byte aligned.  The code doesn't worry about
157  * doing an extra byte.
158  */
159 
160 void
161 le_ioasic_copytobuf_gap2(sc, fromv, boff, len)
162 	struct lance_softc *sc;
163 	void *fromv;
164 	int boff;
165 	int len;
166 {
167 	volatile caddr_t buf = sc->sc_mem;
168 	caddr_t from = fromv;
169 	volatile u_int16_t *bptr;
170 
171 	if (boff & 0x1) {
172 		/* handle unaligned first byte */
173 		bptr = ((volatile u_int16_t *)buf) + (boff - 1);
174 		*bptr = (*from++ << 8) | (*bptr & 0xff);
175 		bptr += 2;
176 		len--;
177 	} else
178 		bptr = ((volatile u_int16_t *)buf) + boff;
179 	while (len > 1) {
180 		*bptr = (from[1] << 8) | (from[0] & 0xff);
181 		bptr += 2;
182 		from += 2;
183 		len -= 2;
184 	}
185 	if (len == 1)
186 		*bptr = (u_int16_t)*from;
187 }
188 
189 void
190 le_ioasic_copyfrombuf_gap2(sc, tov, boff, len)
191 	struct lance_softc *sc;
192 	void *tov;
193 	int boff, len;
194 {
195 	volatile caddr_t buf = sc->sc_mem;
196 	caddr_t to = tov;
197 	volatile u_int16_t *bptr;
198 	u_int16_t tmp;
199 
200 	if (boff & 0x1) {
201 		/* handle unaligned first byte */
202 		bptr = ((volatile u_int16_t *)buf) + (boff - 1);
203 		*to++ = (*bptr >> 8) & 0xff;
204 		bptr += 2;
205 		len--;
206 	} else
207 		bptr = ((volatile u_int16_t *)buf) + boff;
208 	while (len > 1) {
209 		tmp = *bptr;
210 		*to++ = tmp & 0xff;
211 		*to++ = (tmp >> 8) & 0xff;
212 		bptr += 2;
213 		len -= 2;
214 	}
215 	if (len == 1)
216 		*to = *bptr & 0xff;
217 }
218 
219 /*
220  * gap16: 16 bytes of data followed by 16 bytes of pad.
221  *
222  * Buffers must be 32-byte aligned.
223  */
224 
225 void
226 le_ioasic_copytobuf_gap16(sc, fromv, boff, len)
227 	struct lance_softc *sc;
228 	void *fromv;
229 	int boff;
230 	int len;
231 {
232 	volatile caddr_t buf = sc->sc_mem;
233 	caddr_t from = fromv;
234 	caddr_t bptr;
235 
236 	bptr = buf + ((boff << 1) & ~0x1f);
237 	boff &= 0xf;
238 
239 	/*
240 	 * Dispose of boff so destination of subsequent copies is
241 	 * 16-byte aligned.
242 	 */
243 	if (boff) {
244 		int xfer;
245 		xfer = min(len, 16 - boff);
246 		bcopy(from, bptr + boff, xfer);
247 		from += xfer;
248 		bptr += 32;
249 		len -= xfer;
250 	}
251 
252 	/* Destination of  copies is now 16-byte aligned. */
253 	if (len >= 16)
254 		switch ((u_long)from & (sizeof(u_int32_t) -1)) {
255 		case 2:
256 			/*  Ethernet headers make this the dominant case. */
257 		do {
258 			u_int32_t *dst = (u_int32_t*)bptr;
259 			u_int16_t t0;
260 			u_int32_t t1,  t2, t3, t4;
261 
262 			/* read from odd-16-bit-aligned, cached src */
263 			t0 = *(u_int16_t*)from;
264 			t1 = *(u_int32_t*)(from+2);
265 			t2 = *(u_int32_t*)(from+6);
266 			t3 = *(u_int32_t*)(from+10);
267 			t4 = *(u_int16_t*)(from+14);
268 
269 			/* DMA buffer is uncached on mips */
270 			dst[0] =         t0 |  (t1 << 16);
271 			dst[1] = (t1 >> 16) |  (t2 << 16);
272 			dst[2] = (t2 >> 16) |  (t3 << 16);
273 			dst[3] = (t3 >> 16) |  (t4 << 16);
274 
275 			from += 16;
276 			bptr += 32;
277 			len -= 16;
278 		} while (len >= 16);
279 		break;
280 
281 		case 0:
282 		do {
283 			u_int32_t *src = (u_int32_t*)from;
284 			u_int32_t *dst = (u_int32_t*)bptr;
285 			u_int32_t t0, t1, t2, t3;
286 
287 			t0 = src[0]; t1 = src[1]; t2 = src[2]; t3 = src[3];
288 			dst[0] = t0; dst[1] = t1; dst[2] = t2; dst[3] = t3;
289 
290 			from += 16;
291 			bptr += 32;
292 			len -= 16;
293 		} while (len >= 16);
294 		break;
295 
296 		default:
297 		/* Does odd-aligned case ever happen? */
298 		do {
299 			bcopy(from, bptr, 16);
300 			from += 16;
301 			bptr += 32;
302 			len -= 16;
303 		} while (len >= 16);
304 		break;
305 	}
306 	if (len)
307 		bcopy(from, bptr, len);
308 }
309 
310 void
311 le_ioasic_copyfrombuf_gap16(sc, tov, boff, len)
312 	struct lance_softc *sc;
313 	void *tov;
314 	int boff, len;
315 {
316 	volatile caddr_t buf = sc->sc_mem;
317 	caddr_t to = tov;
318 	caddr_t bptr;
319 
320 	bptr = buf + ((boff << 1) & ~0x1f);
321 	boff &= 0xf;
322 
323 	/* Dispose of boff. source of copy is subsequently 16-byte aligned. */
324 	if (boff) {
325 		int xfer;
326 		xfer = min(len, 16 - boff);
327 		bcopy(bptr+boff, to, xfer);
328 		to += xfer;
329 		bptr += 32;
330 		len -= xfer;
331 	}
332 	if (len >= 16)
333 	switch ((u_long)to & (sizeof(u_int32_t) -1)) {
334 	case 2:
335 		/*
336 		 * to is aligned to an odd 16-bit boundary.  Ethernet headers
337 		 * make this the dominant case (98% or more).
338 		 */
339 		do {
340 			u_int32_t *src = (u_int32_t*)bptr;
341 			u_int32_t t0, t1, t2, t3;
342 
343 			/* read from uncached aligned DMA buf */
344 			t0 = src[0]; t1 = src[1]; t2 = src[2]; t3 = src[3];
345 
346 			/* write to odd-16-bit-word aligned dst */
347 			*(u_int16_t *) (to+0)  = (u_short)  t0;
348 			*(u_int32_t *) (to+2)  = (t0 >> 16) |  (t1 << 16);
349 			*(u_int32_t *) (to+6)  = (t1 >> 16) |  (t2 << 16);
350 			*(u_int32_t *) (to+10) = (t2 >> 16) |  (t3 << 16);
351 			*(u_int16_t *) (to+14) = (t3 >> 16);
352 			bptr += 32;
353 			to += 16;
354 			len -= 16;
355 		} while (len > 16);
356 		break;
357 	case 0:
358 		/* 32-bit aligned aligned copy. Rare. */
359 		do {
360 			u_int32_t *src = (u_int32_t*)bptr;
361 			u_int32_t *dst = (u_int32_t*)to;
362 			u_int32_t t0, t1, t2, t3;
363 
364 			t0 = src[0]; t1 = src[1]; t2 = src[2]; t3 = src[3];
365 			dst[0] = t0; dst[1] = t1; dst[2] = t2; dst[3] = t3;
366 			to += 16;
367 			bptr += 32;
368 			len -= 16;
369 		} while (len  > 16);
370 		break;
371 
372 	/* XXX Does odd-byte-aligned case ever happen? */
373 	default:
374 		do {
375 			bcopy(bptr, to, 16);
376 			to += 16;
377 			bptr += 32;
378 			len -= 16;
379 		} while (len  > 16);
380 		break;
381 	}
382 	if (len)
383 		bcopy(bptr, to, len);
384 }
385 
386 void
387 le_ioasic_zerobuf_gap16(sc, boff, len)
388 	struct lance_softc *sc;
389 	int boff, len;
390 {
391 	volatile caddr_t buf = sc->sc_mem;
392 	caddr_t bptr;
393 	int xfer;
394 
395 	bptr = buf + ((boff << 1) & ~0x1f);
396 	boff &= 0xf;
397 	xfer = min(len, 16 - boff);
398 	while (len > 0) {
399 		bzero(bptr + boff, xfer);
400 		bptr += 32;
401 		boff = 0;
402 		len -= xfer;
403 		xfer = min(len, 16);
404 	}
405 }
406 
407 #define	LE_IOASIC_MEMSIZE	(128*1024)
408 #define	LE_IOASIC_MEMALIGN	(128*1024)
409 
410 void
411 ioasic_lance_dma_setup(parent)
412 	struct device *parent;
413 {
414 	struct ioasic_softc *sc = (void *)parent;
415 	bus_dma_tag_t dmat = sc->sc_dmat;
416 	bus_dma_segment_t seg;
417 	tc_addr_t tca;
418 	u_int32_t ssr;
419 	int rseg;
420 
421 	/*
422 	 * Allocate a DMA area for the chip.
423 	 */
424 	if (bus_dmamem_alloc(dmat, LE_IOASIC_MEMSIZE, LE_IOASIC_MEMALIGN,
425 	    0, &seg, 1, &rseg, BUS_DMA_NOWAIT)) {
426 		printf("%s: can't allocate DMA area for LANCE\n",
427 		    sc->sc_dv.dv_xname);
428 		return;
429 	}
430 	if (bus_dmamem_map(dmat, &seg, rseg, LE_IOASIC_MEMSIZE,
431 	    &le_iomem, BUS_DMA_NOWAIT|BUS_DMA_COHERENT)) {
432 		printf("%s: can't map DMA area for LANCE\n",
433 		    sc->sc_dv.dv_xname);
434 		bus_dmamem_free(dmat, &seg, rseg);
435 		return;
436 	}
437 
438 	/*
439 	 * Create and load the DMA map for the DMA area.
440 	 */
441 	if (bus_dmamap_create(dmat, LE_IOASIC_MEMSIZE, 1,
442 	    LE_IOASIC_MEMSIZE, 0, BUS_DMA_NOWAIT, &sc->sc_lance_dmam)) {
443 		printf("%s: can't create DMA map\n", sc->sc_dv.dv_xname);
444 		goto bad;
445 	}
446 	if (bus_dmamap_load(dmat, sc->sc_lance_dmam,
447 	    le_iomem, LE_IOASIC_MEMSIZE, NULL, BUS_DMA_NOWAIT)) {
448 		printf("%s: can't load DMA map\n", sc->sc_dv.dv_xname);
449 		goto bad;
450 	}
451 
452 	tca = (tc_addr_t)sc->sc_lance_dmam->dm_segs[0].ds_addr;
453 	tca = ((tca << 3) & ~0x1f) | ((tca >> 29) & 0x1f);
454 	bus_space_write_4(sc->sc_bst, sc->sc_bsh, IOASIC_LANCE_DMAPTR, tca);
455 	ssr = bus_space_read_4(sc->sc_bst, sc->sc_bsh, IOASIC_CSR);
456 	ssr |= IOASIC_CSR_DMAEN_LANCE;
457 	bus_space_write_4(sc->sc_bst, sc->sc_bsh, IOASIC_CSR, ssr);
458 	return;
459 
460  bad:
461 	bus_dmamem_unmap(dmat, le_iomem, LE_IOASIC_MEMSIZE);
462 	bus_dmamem_free(dmat, &seg, rseg);
463 	le_iomem = 0;
464 }
465 
466 /* XXX */
467 char *
468 ioasic_lance_ether_address()
469 {
470 
471         return (char *)(ioasic_base + IOASIC_SLOT_2_START);
472 }
473