xref: /netbsd-src/sys/arch/mips/cavium/dev/if_cnmac.c (revision e89934bbf778a6d6d6894877c4da59d0c7835b0f)
1 /*	$NetBSD: if_cnmac.c,v 1.5 2016/12/15 09:28:03 ozaki-r Exp $	*/
2 
3 #include <sys/cdefs.h>
4 #if 0
5 __KERNEL_RCSID(0, "$NetBSD: if_cnmac.c,v 1.5 2016/12/15 09:28:03 ozaki-r Exp $");
6 #endif
7 
8 #include "opt_octeon.h"
9 
10 #ifdef	OCTEON_ETH_DEBUG
11 
12 #ifndef DIAGNOSTIC
13 #define	DIAGNOSTIC
14 #endif
15 
16 #ifndef DEBUG
17 #define	DEBUG
18 #endif
19 
20 #endif
21 
22 /*
23  * If no free send buffer is available, free all the sent buffer and bail out.
24  */
25 #define OCTEON_ETH_SEND_QUEUE_CHECK
26 
27 /* XXX XXX XXX XXX XXX XXX */
28 
29 #include <sys/param.h>
30 #include <sys/systm.h>
31 #include <sys/pool.h>
32 #include <sys/mbuf.h>
33 #include <sys/malloc.h>
34 #include <sys/kernel.h>
35 #include <sys/socket.h>
36 #include <sys/ioctl.h>
37 #include <sys/errno.h>
38 #include <sys/device.h>
39 #include <sys/queue.h>
40 #include <sys/conf.h>
41 #include <sys/sysctl.h>
42 #include <sys/syslog.h>
43 
44 #include <net/if.h>
45 #include <net/if_dl.h>
46 #include <net/if_media.h>
47 #include <net/if_ether.h>
48 #include <net/route.h>
49 
50 #include <net/bpf.h>
51 
52 #include <netinet/in.h>
53 #include <netinet/in_systm.h>
54 #include <netinet/in_var.h>
55 #include <netinet/ip.h>
56 
57 #include <sys/bus.h>
58 #include <machine/intr.h>
59 #include <machine/endian.h>
60 #include <machine/locore.h>
61 
62 #include <dev/mii/mii.h>
63 #include <dev/mii/miivar.h>
64 
65 #include <mips/cpuregs.h>
66 
67 #include <mips/cavium/dev/octeon_asxreg.h>
68 #include <mips/cavium/dev/octeon_ciureg.h>
69 #include <mips/cavium/dev/octeon_npireg.h>
70 #include <mips/cavium/dev/octeon_gmxreg.h>
71 #include <mips/cavium/dev/octeon_ipdreg.h>
72 #include <mips/cavium/dev/octeon_pipreg.h>
73 #include <mips/cavium/dev/octeon_powreg.h>
74 #include <mips/cavium/dev/octeon_faureg.h>
75 #include <mips/cavium/dev/octeon_fpareg.h>
76 #include <mips/cavium/dev/octeon_bootbusreg.h>
77 #include <mips/cavium/include/iobusvar.h>
78 #include <mips/cavium/octeonvar.h>
79 #include <mips/cavium/dev/octeon_fpavar.h>
80 #include <mips/cavium/dev/octeon_gmxvar.h>
81 #include <mips/cavium/dev/octeon_fauvar.h>
82 #include <mips/cavium/dev/octeon_powvar.h>
83 #include <mips/cavium/dev/octeon_ipdvar.h>
84 #include <mips/cavium/dev/octeon_pipvar.h>
85 #include <mips/cavium/dev/octeon_pkovar.h>
86 #include <mips/cavium/dev/octeon_asxvar.h>
87 #include <mips/cavium/dev/octeon_smivar.h>
88 #include <mips/cavium/dev/if_cnmacvar.h>
89 
90 #ifdef OCTEON_ETH_DEBUG
91 #define	OCTEON_ETH_KASSERT(x)	KASSERT(x)
92 #define	OCTEON_ETH_KDASSERT(x)	KDASSERT(x)
93 #else
94 #define	OCTEON_ETH_KASSERT(x)
95 #define	OCTEON_ETH_KDASSERT(x)
96 #endif
97 
98 /*
99  * Set the PKO to think command buffers are an odd length.  This makes it so we
100  * never have to divide a comamnd across two buffers.
101  */
102 #define OCTEON_POOL_NWORDS_CMD	\
103 	    (((uint32_t)OCTEON_POOL_SIZE_CMD / sizeof(uint64_t)) - 1)
104 #define FPA_COMMAND_BUFFER_POOL_NWORDS	OCTEON_POOL_NWORDS_CMD	/* XXX */
105 
106 static void		octeon_eth_buf_init(struct octeon_eth_softc *);
107 
108 static int	octeon_eth_match(device_t, struct cfdata *, void *);
109 static void	octeon_eth_attach(device_t, device_t, void *);
110 static void	octeon_eth_pip_init(struct octeon_eth_softc *);
111 static void	octeon_eth_ipd_init(struct octeon_eth_softc *);
112 static void	octeon_eth_pko_init(struct octeon_eth_softc *);
113 static void	octeon_eth_asx_init(struct octeon_eth_softc *);
114 static void	octeon_eth_smi_init(struct octeon_eth_softc *);
115 
116 static void	octeon_eth_board_mac_addr(uint8_t *, size_t, struct octeon_eth_softc *);
117 
118 static int	octeon_eth_mii_readreg(device_t, int, int);
119 static void	octeon_eth_mii_writereg(device_t, int, int, int);
120 static void	octeon_eth_mii_statchg(struct ifnet *);
121 
122 static int	octeon_eth_mediainit(struct octeon_eth_softc *);
123 static void	octeon_eth_mediastatus(struct ifnet *, struct ifmediareq *);
124 static int	octeon_eth_mediachange(struct ifnet *);
125 
126 static inline void	octeon_eth_send_queue_flush_prefetch(struct octeon_eth_softc *);
127 static inline void	octeon_eth_send_queue_flush_fetch(struct octeon_eth_softc *);
128 static inline void	octeon_eth_send_queue_flush(struct octeon_eth_softc *);
129 static inline void	octeon_eth_send_queue_flush_sync(struct octeon_eth_softc *);
130 static inline int	octeon_eth_send_queue_is_full(struct octeon_eth_softc *);
131 static inline void	octeon_eth_send_queue_add(struct octeon_eth_softc *,
132 			    struct mbuf *, uint64_t *);
133 static inline void	octeon_eth_send_queue_del(struct octeon_eth_softc *,
134 			    struct mbuf **, uint64_t **);
135 static inline int	octeon_eth_buf_free_work(struct octeon_eth_softc *,
136 			    uint64_t *, uint64_t);
137 static inline void	octeon_eth_buf_ext_free_m(struct mbuf *, void *, size_t, void *);
138 static inline void	octeon_eth_buf_ext_free_ext(struct mbuf *, void *, size_t, void *);
139 
140 static int	octeon_eth_ioctl(struct ifnet *, u_long, void *);
141 static void	octeon_eth_watchdog(struct ifnet *);
142 static int	octeon_eth_init(struct ifnet *);
143 static void	octeon_eth_stop(struct ifnet *, int);
144 static void	octeon_eth_start(struct ifnet *);
145 
146 static inline int	octeon_eth_send_cmd(struct octeon_eth_softc *, uint64_t,
147 			    uint64_t);
148 static inline uint64_t	octeon_eth_send_makecmd_w1(int, paddr_t);
149 static inline uint64_t 	octeon_eth_send_makecmd_w0(uint64_t, uint64_t, size_t,
150 			    int);
151 static inline int	octeon_eth_send_makecmd_gbuf(struct octeon_eth_softc *,
152 			    struct mbuf *, uint64_t *, int *);
153 static inline int	octeon_eth_send_makecmd(struct octeon_eth_softc *,
154 			    struct mbuf *, uint64_t *, uint64_t *, uint64_t *);
155 static inline int	octeon_eth_send_buf(struct octeon_eth_softc *,
156 			    struct mbuf *, uint64_t *);
157 static inline int	octeon_eth_send(struct octeon_eth_softc *,
158 			    struct mbuf *);
159 
160 static int	octeon_eth_reset(struct octeon_eth_softc *);
161 static int	octeon_eth_configure(struct octeon_eth_softc *);
162 static int	octeon_eth_configure_common(struct octeon_eth_softc *);
163 
164 static void	octeon_eth_tick_free(void *arg);
165 static void	octeon_eth_tick_misc(void *);
166 
167 static inline int	octeon_eth_recv_mbuf(struct octeon_eth_softc *,
168 			    uint64_t *, struct mbuf **);
169 static inline int	octeon_eth_recv_check_code(struct octeon_eth_softc *,
170 			    uint64_t);
171 static inline int	octeon_eth_recv_check_jumbo(struct octeon_eth_softc *,
172 			    uint64_t);
173 static inline int	octeon_eth_recv_check_link(struct octeon_eth_softc *,
174 			    uint64_t);
175 static inline int	octeon_eth_recv_check(struct octeon_eth_softc *,
176 			    uint64_t);
177 static inline int	octeon_eth_recv(struct octeon_eth_softc *, uint64_t *);
178 static void		octeon_eth_recv_redir(struct ifnet *, struct mbuf *);
179 static inline void	octeon_eth_recv_intr(void *, uint64_t *);
180 
181 /* device driver context */
182 static struct	octeon_eth_softc *octeon_eth_gsc[GMX_PORT_NUNITS];
183 static void	*octeon_eth_pow_recv_ih;
184 
185 /* sysctl'able parameters */
186 int		octeon_eth_param_pko_cmd_w0_n2 = 1;
187 int		octeon_eth_param_pip_dyn_rs = 1;
188 int		octeon_eth_param_redir = 0;
189 int		octeon_eth_param_pktbuf = 0;
190 int		octeon_eth_param_rate = 0;
191 int		octeon_eth_param_intr = 0;
192 
193 CFATTACH_DECL_NEW(cnmac, sizeof(struct octeon_eth_softc),
194     octeon_eth_match, octeon_eth_attach, NULL, NULL);
195 
196 #ifdef OCTEON_ETH_DEBUG
197 
198 static const struct octeon_evcnt_entry octeon_evcnt_entries[] = {
199 #define	_ENTRY(name, type, parent, descr) \
200 	OCTEON_EVCNT_ENTRY(struct octeon_eth_softc, name, type, parent, descr)
201 	_ENTRY(rx,			MISC, NULL, "rx"),
202 	_ENTRY(rxint,			INTR, NULL, "rx intr"),
203 	_ENTRY(rxrs,			MISC, NULL, "rx dynamic short"),
204 	_ENTRY(rxbufpkalloc,		MISC, NULL, "rx buf pkt alloc"),
205 	_ENTRY(rxbufpkput,		MISC, NULL, "rx buf pkt put"),
206 	_ENTRY(rxbufwqalloc,		MISC, NULL, "rx buf wqe alloc"),
207 	_ENTRY(rxbufwqput,		MISC, NULL, "rx buf wqe put"),
208 	_ENTRY(rxerrcode,		MISC, NULL, "rx code error"),
209 	_ENTRY(rxerrfix,		MISC, NULL, "rx fixup error"),
210 	_ENTRY(rxerrjmb,		MISC, NULL, "rx jmb error"),
211 	_ENTRY(rxerrlink,		MISC, NULL, "rx link error"),
212 	_ENTRY(rxerroff,		MISC, NULL, "rx offload error"),
213 	_ENTRY(rxonperrshort,		MISC, NULL, "rx onp fixup short error"),
214 	_ENTRY(rxonperrpreamble,	MISC, NULL, "rx onp fixup preamble error"),
215 	_ENTRY(rxonperrcrc,		MISC, NULL, "rx onp fixup crc error"),
216 	_ENTRY(rxonperraddress,		MISC, NULL, "rx onp fixup address error"),
217 	_ENTRY(rxonponp,		MISC, NULL, "rx onp fixup onp packets"),
218 	_ENTRY(rxonpok,			MISC, NULL, "rx onp fixup success packets"),
219 	_ENTRY(tx,			MISC, NULL, "tx"),
220 	_ENTRY(txadd,			MISC, NULL, "tx add"),
221 	_ENTRY(txbufcballoc,		MISC, NULL, "tx buf cb alloc"),
222 	_ENTRY(txbufcbget,		MISC, NULL, "tx buf cb get"),
223 	_ENTRY(txbufgballoc,		MISC, NULL, "tx buf gb alloc"),
224 	_ENTRY(txbufgbget,		MISC, NULL, "tx buf gb get"),
225 	_ENTRY(txbufgbput,		MISC, NULL, "tx buf gb put"),
226 	_ENTRY(txdel,			MISC, NULL, "tx del"),
227 	_ENTRY(txerr,			MISC, NULL, "tx error"),
228 	_ENTRY(txerrcmd,		MISC, NULL, "tx cmd error"),
229 	_ENTRY(txerrgbuf,		MISC, NULL, "tx gbuf error"),
230 	_ENTRY(txerrlink,		MISC, NULL, "tx link error"),
231 	_ENTRY(txerrmkcmd,		MISC, NULL, "tx makecmd error"),
232 #undef	_ENTRY
233 };
234 #endif
235 
236 /* ---- buffer management */
237 
238 static const struct octeon_eth_pool_param {
239 	int			poolno;
240 	size_t			size;
241 	size_t			nelems;
242 } octeon_eth_pool_params[] = {
243 #define	_ENTRY(x)	{ OCTEON_POOL_NO_##x, OCTEON_POOL_SIZE_##x, OCTEON_POOL_NELEMS_##x }
244 	_ENTRY(PKT),
245 	_ENTRY(WQE),
246 	_ENTRY(CMD),
247 	_ENTRY(SG)
248 #undef	_ENTRY
249 };
250 struct octeon_fpa_buf	*octeon_eth_pools[8/* XXX */];
251 #define	octeon_eth_fb_pkt	octeon_eth_pools[OCTEON_POOL_NO_PKT]
252 #define	octeon_eth_fb_wqe	octeon_eth_pools[OCTEON_POOL_NO_WQE]
253 #define	octeon_eth_fb_cmd	octeon_eth_pools[OCTEON_POOL_NO_CMD]
254 #define	octeon_eth_fb_sg	octeon_eth_pools[OCTEON_POOL_NO_SG]
255 
256 static void
257 octeon_eth_buf_init(struct octeon_eth_softc *sc)
258 {
259 	static int once;
260 	int i;
261 	const struct octeon_eth_pool_param *pp;
262 	struct octeon_fpa_buf *fb;
263 
264 	if (once == 1)
265 		return;
266 	once = 1;
267 
268 	for (i = 0; i < (int)__arraycount(octeon_eth_pool_params); i++) {
269 		pp = &octeon_eth_pool_params[i];
270 		octeon_fpa_buf_init(pp->poolno, pp->size, pp->nelems, &fb);
271 		octeon_eth_pools[i] = fb;
272 	}
273 }
274 
275 /* ---- autoconf */
276 
277 static int
278 octeon_eth_match(device_t parent, struct cfdata *match, void *aux)
279 {
280 	struct octeon_gmx_attach_args *ga = aux;
281 
282 	if (strcmp(match->cf_name, ga->ga_name) != 0) {
283 		return 0;
284 	}
285 	return 1;
286 }
287 
288 static void
289 octeon_eth_attach(device_t parent, device_t self, void *aux)
290 {
291 	struct octeon_eth_softc *sc = device_private(self);
292 	struct octeon_gmx_attach_args *ga = aux;
293 	struct ifnet *ifp = &sc->sc_ethercom.ec_if;
294 	uint8_t enaddr[ETHER_ADDR_LEN];
295 
296 	sc->sc_dev = self;
297 	sc->sc_regt = ga->ga_regt;
298 	sc->sc_port = ga->ga_portno;
299 	sc->sc_port_type = ga->ga_port_type;
300 	sc->sc_gmx = ga->ga_gmx;
301 	sc->sc_gmx_port = ga->ga_gmx_port;
302 
303 	sc->sc_init_flag = 0;
304 	/*
305 	 * XXXUEBAYASI
306 	 * Setting PIP_IP_OFFSET[OFFSET] to 8 causes panic ... why???
307 	 */
308 	sc->sc_ip_offset = 0/* XXX */;
309 
310 	if (MIPS_PRID_IMPL(mips_options.mips_cpu_id) <= MIPS_CN30XX) {
311 		SET(sc->sc_quirks, OCTEON_ETH_QUIRKS_NO_PRE_ALIGN);
312 		SET(sc->sc_quirks, OCTEON_ETH_QUIRKS_NO_RX_INBND);
313 	}
314 
315 	octeon_eth_board_mac_addr(enaddr, sizeof(enaddr), sc);
316 	printf("%s: Ethernet address %s\n", device_xname(sc->sc_dev),
317 	    ether_sprintf(enaddr));
318 
319 	octeon_eth_gsc[sc->sc_port] = sc;
320 
321 	SIMPLEQ_INIT(&sc->sc_sendq);
322 	sc->sc_soft_req_thresh = 15/* XXX */;
323 	sc->sc_ext_callback_cnt = 0;
324 
325 	octeon_gmx_stats_init(sc->sc_gmx_port);
326 
327 	callout_init(&sc->sc_tick_misc_ch, 0);
328 	callout_init(&sc->sc_tick_free_ch, 0);
329 
330 	octeon_fau_op_init(&sc->sc_fau_done,
331 	    OCTEON_CVMSEG_ETHER_OFFSET(sc->sc_port, csm_ether_fau_done),
332 	    OCT_FAU_REG_ADDR_END - (8 * (sc->sc_port + 1))/* XXX */);
333 	octeon_fau_op_set_8(&sc->sc_fau_done, 0);
334 
335 	octeon_eth_pip_init(sc);
336 	octeon_eth_ipd_init(sc);
337 	octeon_eth_pko_init(sc);
338 	octeon_eth_asx_init(sc);
339 	octeon_eth_smi_init(sc);
340 
341 	sc->sc_gmx_port->sc_ipd = sc->sc_ipd;
342 	sc->sc_gmx_port->sc_port_asx = sc->sc_asx;
343 	sc->sc_gmx_port->sc_port_mii = &sc->sc_mii;
344 	sc->sc_gmx_port->sc_port_ec = &sc->sc_ethercom;
345 	/* XXX */
346 	sc->sc_gmx_port->sc_quirks = sc->sc_quirks;
347 
348 	/* XXX */
349 	sc->sc_pow = &octeon_pow_softc;
350 
351 	octeon_eth_mediainit(sc);
352 
353 	strncpy(ifp->if_xname, device_xname(sc->sc_dev), sizeof(ifp->if_xname));
354 	ifp->if_softc = sc;
355 	ifp->if_flags = IFF_BROADCAST | IFF_SIMPLEX | IFF_MULTICAST;
356 	ifp->if_ioctl = octeon_eth_ioctl;
357 	ifp->if_start = octeon_eth_start;
358 	ifp->if_watchdog = octeon_eth_watchdog;
359 	ifp->if_init = octeon_eth_init;
360 	ifp->if_stop = octeon_eth_stop;
361 	IFQ_SET_MAXLEN(&ifp->if_snd, max(GATHER_QUEUE_SIZE, IFQ_MAXLEN));
362 	IFQ_SET_READY(&ifp->if_snd);
363 
364 	/* XXX: not yet tx checksum */
365 	ifp->if_capabilities =
366 		IFCAP_CSUM_IPv4_Rx | IFCAP_CSUM_TCPv4_Rx | IFCAP_CSUM_UDPv4_Rx |
367 		IFCAP_CSUM_TCPv6_Rx | IFCAP_CSUM_UDPv6_Rx;
368 
369 	octeon_gmx_set_mac_addr(sc->sc_gmx_port, enaddr);
370 	octeon_gmx_set_filter(sc->sc_gmx_port);
371 
372 	if_attach(ifp);
373 	ether_ifattach(ifp, enaddr);
374 
375 	/* XXX */
376 	sc->sc_rate_recv_check_link_cap.tv_sec = 1;
377 	sc->sc_rate_recv_check_jumbo_cap.tv_sec = 1;
378 	sc->sc_rate_recv_check_code_cap.tv_sec = 1;
379 	sc->sc_rate_recv_fixup_odd_nibble_short_cap.tv_sec = 1;
380 	sc->sc_rate_recv_fixup_odd_nibble_preamble_cap.tv_sec = 1;
381 	sc->sc_rate_recv_fixup_odd_nibble_crc_cap.tv_sec = 1;
382 #ifdef OCTEON_ETH_DEBUG
383 	sc->sc_rate_recv_fixup_odd_nibble_addr_cap.tv_sec = 1;
384 #endif
385 	/* XXX */
386 
387 #if 1
388 	octeon_eth_buf_init(sc);
389 #endif
390 
391 	if (octeon_eth_pow_recv_ih == NULL)
392 		octeon_eth_pow_recv_ih = octeon_pow_intr_establish(OCTEON_POW_GROUP_PIP,
393 		    IPL_NET, octeon_eth_recv_intr, NULL, NULL);
394 
395 	OCTEON_EVCNT_ATTACH_EVCNTS(sc, octeon_evcnt_entries,
396 	    device_xname(sc->sc_dev));
397 }
398 
399 /* ---- submodules */
400 
401 /* XXX */
402 static void
403 octeon_eth_pip_init(struct octeon_eth_softc *sc)
404 {
405 	struct octeon_pip_attach_args pip_aa;
406 
407 	pip_aa.aa_port = sc->sc_port;
408 	pip_aa.aa_regt = sc->sc_regt;
409 	pip_aa.aa_tag_type = POW_TAG_TYPE_ORDERED/* XXX */;
410 	pip_aa.aa_receive_group = OCTEON_POW_GROUP_PIP;
411 	pip_aa.aa_ip_offset = sc->sc_ip_offset;
412 	octeon_pip_init(&pip_aa, &sc->sc_pip);
413 }
414 
415 /* XXX */
416 static void
417 octeon_eth_ipd_init(struct octeon_eth_softc *sc)
418 {
419 	struct octeon_ipd_attach_args ipd_aa;
420 
421 	ipd_aa.aa_port = sc->sc_port;
422 	ipd_aa.aa_regt = sc->sc_regt;
423 	ipd_aa.aa_first_mbuff_skip = 184/* XXX */;
424 	ipd_aa.aa_not_first_mbuff_skip = 0/* XXX */;
425 	octeon_ipd_init(&ipd_aa, &sc->sc_ipd);
426 }
427 
428 /* XXX */
429 static void
430 octeon_eth_pko_init(struct octeon_eth_softc *sc)
431 {
432 	struct octeon_pko_attach_args pko_aa;
433 
434 	pko_aa.aa_port = sc->sc_port;
435 	pko_aa.aa_regt = sc->sc_regt;
436 	pko_aa.aa_cmdptr = &sc->sc_cmdptr;
437 	pko_aa.aa_cmd_buf_pool = OCTEON_POOL_NO_CMD;
438 	pko_aa.aa_cmd_buf_size = OCTEON_POOL_NWORDS_CMD;
439 	octeon_pko_init(&pko_aa, &sc->sc_pko);
440 }
441 
442 /* XXX */
443 static void
444 octeon_eth_asx_init(struct octeon_eth_softc *sc)
445 {
446 	struct octeon_asx_attach_args asx_aa;
447 
448 	asx_aa.aa_port = sc->sc_port;
449 	asx_aa.aa_regt = sc->sc_regt;
450 	octeon_asx_init(&asx_aa, &sc->sc_asx);
451 }
452 
453 static void
454 octeon_eth_smi_init(struct octeon_eth_softc *sc)
455 {
456 	struct octeon_smi_attach_args smi_aa;
457 
458 	smi_aa.aa_port = sc->sc_port;
459 	smi_aa.aa_regt = sc->sc_regt;
460 	octeon_smi_init(&smi_aa, &sc->sc_smi);
461 	octeon_smi_set_clock(sc->sc_smi, 0x1464ULL); /* XXX */
462 }
463 
464 /* ---- XXX */
465 
466 #define	ADDR2UINT64(u, a) \
467 	do { \
468 		u = \
469 		    (((uint64_t)a[0] << 40) | ((uint64_t)a[1] << 32) | \
470 		     ((uint64_t)a[2] << 24) | ((uint64_t)a[3] << 16) | \
471 		     ((uint64_t)a[4] <<  8) | ((uint64_t)a[5] <<  0)); \
472 	} while (0)
473 #define	UINT642ADDR(a, u) \
474 	do { \
475 		a[0] = (uint8_t)((u) >> 40); a[1] = (uint8_t)((u) >> 32); \
476 		a[2] = (uint8_t)((u) >> 24); a[3] = (uint8_t)((u) >> 16); \
477 		a[4] = (uint8_t)((u) >>  8); a[5] = (uint8_t)((u) >>  0); \
478 	} while (0)
479 
480 static void
481 octeon_eth_board_mac_addr(uint8_t *enaddr, size_t size, struct octeon_eth_softc *sc)
482 {
483 	prop_dictionary_t dict;
484 	prop_data_t ea;
485 
486 	dict = device_properties(sc->sc_dev);
487 	KASSERT(dict != NULL);
488 	ea = prop_dictionary_get(dict, "mac-address");
489 	KASSERT(ea != NULL);
490 	memcpy(enaddr, prop_data_data_nocopy(ea), size);
491 }
492 
493 /* ---- media */
494 
495 static int
496 octeon_eth_mii_readreg(device_t self, int phy_addr, int reg)
497 {
498 	struct octeon_eth_softc *sc = device_private(self);
499 
500 	return octeon_smi_read(sc->sc_smi, phy_addr, reg);
501 }
502 
503 static void
504 octeon_eth_mii_writereg(device_t self, int phy_addr, int reg, int value)
505 {
506 	struct octeon_eth_softc *sc = device_private(self);
507 
508 	octeon_smi_write(sc->sc_smi, phy_addr, reg, value);
509 }
510 
511 static void
512 octeon_eth_mii_statchg(struct ifnet *ifp)
513 {
514 	struct octeon_eth_softc *sc = ifp->if_softc;
515 
516 	octeon_pko_port_enable(sc->sc_pko, 0);
517 	octeon_gmx_port_enable(sc->sc_gmx_port, 0);
518 
519 	octeon_eth_reset(sc);
520 
521 	if (ISSET(ifp->if_flags, IFF_RUNNING))
522 		octeon_gmx_set_filter(sc->sc_gmx_port);
523 
524 	octeon_pko_port_enable(sc->sc_pko, 1);
525 	octeon_gmx_port_enable(sc->sc_gmx_port, 1);
526 }
527 
528 static int
529 octeon_eth_mediainit(struct octeon_eth_softc *sc)
530 {
531 	struct ifnet *ifp = &sc->sc_ethercom.ec_if;
532 	prop_object_t phy;
533 
534 	sc->sc_mii.mii_ifp = ifp;
535 	sc->sc_mii.mii_readreg = octeon_eth_mii_readreg;
536 	sc->sc_mii.mii_writereg = octeon_eth_mii_writereg;
537 	sc->sc_mii.mii_statchg = octeon_eth_mii_statchg;
538 	ifmedia_init(&sc->sc_mii.mii_media, 0, octeon_eth_mediachange,
539 	    octeon_eth_mediastatus);
540 
541 	phy = prop_dictionary_get(device_properties(sc->sc_dev), "phy-addr");
542 	KASSERT(phy != NULL);
543 
544 	mii_attach(sc->sc_dev, &sc->sc_mii,
545 	    0xffffffff, prop_number_integer_value(phy),
546 	    MII_OFFSET_ANY, MIIF_DOPAUSE);
547 
548 	/* XXX XXX XXX */
549 	if (LIST_FIRST(&sc->sc_mii.mii_phys) != NULL) {
550 		/* XXX XXX XXX */
551 		ifmedia_set(&sc->sc_mii.mii_media, IFM_ETHER | IFM_AUTO);
552 		/* XXX XXX XXX */
553 	} else {
554 		/* XXX XXX XXX */
555 		ifmedia_add(&sc->sc_mii.mii_media, IFM_ETHER | IFM_NONE,
556 		    MII_MEDIA_NONE, NULL);
557 		/* XXX XXX XXX */
558 		ifmedia_set(&sc->sc_mii.mii_media, IFM_ETHER | IFM_NONE);
559 		/* XXX XXX XXX */
560 	}
561 	/* XXX XXX XXX */
562 
563 	return 0;
564 }
565 
566 static void
567 octeon_eth_mediastatus(struct ifnet *ifp, struct ifmediareq *ifmr)
568 {
569 	struct octeon_eth_softc *sc = ifp->if_softc;
570 
571 	mii_pollstat(&sc->sc_mii);
572 
573 	ifmr->ifm_status = sc->sc_mii.mii_media_status;
574 	ifmr->ifm_active = sc->sc_mii.mii_media_active;
575 	ifmr->ifm_active = (sc->sc_mii.mii_media_active & ~IFM_ETH_FMASK) |
576 	    sc->sc_gmx_port->sc_port_flowflags;
577 }
578 
579 static int
580 octeon_eth_mediachange(struct ifnet *ifp)
581 {
582 	struct octeon_eth_softc *sc = ifp->if_softc;
583 
584 	mii_mediachg(&sc->sc_mii);
585 
586 	return 0;
587 }
588 
589 /* ---- send buffer garbage collection */
590 
591 static inline void
592 octeon_eth_send_queue_flush_prefetch(struct octeon_eth_softc *sc)
593 {
594 	OCTEON_ETH_KASSERT(sc->sc_prefetch == 0);
595 	octeon_fau_op_inc_fetch_8(&sc->sc_fau_done, 0);
596 	sc->sc_prefetch = 1;
597 }
598 
599 static inline void
600 octeon_eth_send_queue_flush_fetch(struct octeon_eth_softc *sc)
601 {
602 #ifndef  OCTEON_ETH_DEBUG
603 	if (!sc->sc_prefetch)
604 		return;
605 #endif
606 	OCTEON_ETH_KASSERT(sc->sc_prefetch == 1);
607 	sc->sc_hard_done_cnt = octeon_fau_op_inc_read_8(&sc->sc_fau_done);
608 	OCTEON_ETH_KASSERT(sc->sc_hard_done_cnt <= 0);
609 	sc->sc_prefetch = 0;
610 }
611 
612 static inline void
613 octeon_eth_send_queue_flush(struct octeon_eth_softc *sc)
614 {
615 	const int64_t sent_count = sc->sc_hard_done_cnt;
616 	int i;
617 
618 	OCTEON_ETH_KASSERT(sc->sc_flush == 0);
619 	OCTEON_ETH_KASSERT(sent_count <= 0);
620 
621 	for (i = 0; i < 0 - sent_count; i++) {
622 		struct mbuf *m;
623 		uint64_t *gbuf;
624 
625 		octeon_eth_send_queue_del(sc, &m, &gbuf);
626 
627 		octeon_fpa_buf_put(octeon_eth_fb_sg, gbuf);
628 		OCTEON_EVCNT_INC(sc, txbufgbput);
629 
630 		m_freem(m);
631 	}
632 
633 	octeon_fau_op_inc_fetch_8(&sc->sc_fau_done, i);
634 	sc->sc_flush = i;
635 }
636 
637 static inline void
638 octeon_eth_send_queue_flush_sync(struct octeon_eth_softc *sc)
639 {
640 	if (sc->sc_flush == 0)
641 		return;
642 
643 	OCTEON_ETH_KASSERT(sc->sc_flush > 0);
644 
645 	/* XXX XXX XXX */
646 	octeon_fau_op_inc_read_8(&sc->sc_fau_done);
647 	sc->sc_soft_req_cnt -= sc->sc_flush;
648 	OCTEON_ETH_KASSERT(sc->sc_soft_req_cnt >= 0);
649 	/* XXX XXX XXX */
650 
651 	sc->sc_flush = 0;
652 }
653 
654 static inline int
655 octeon_eth_send_queue_is_full(struct octeon_eth_softc *sc)
656 {
657 #ifdef OCTEON_ETH_SEND_QUEUE_CHECK
658 	int64_t nofree_cnt;
659 
660 	nofree_cnt = sc->sc_soft_req_cnt + sc->sc_hard_done_cnt;
661 
662 	if (__predict_false(nofree_cnt == GATHER_QUEUE_SIZE - 1)) {
663 		octeon_eth_send_queue_flush(sc);
664 		OCTEON_EVCNT_INC(sc, txerrgbuf);
665 		octeon_eth_send_queue_flush_sync(sc);
666 		return 1;
667 	}
668 
669 #endif
670 	return 0;
671 }
672 
673 /*
674  * (Ab)use m_nextpkt and m_paddr to maintain mbuf chain and pointer to gather
675  * buffer.  Other mbuf members may be used by m_freem(), so don't touch them!
676  */
677 
678 struct _send_queue_entry {
679 	union {
680 		struct mbuf _sqe_s_mbuf;
681 		struct {
682 			char _sqe_s_entry_pad[offsetof(struct mbuf, m_nextpkt)];
683 			SIMPLEQ_ENTRY(_send_queue_entry) _sqe_s_entry_entry;
684 		} _sqe_s_entry;
685 		struct {
686 			char _sqe_s_gbuf_pad[offsetof(struct mbuf, m_paddr)];
687 			uint64_t *_sqe_s_gbuf_gbuf;
688 		} _sqe_s_gbuf;
689 	} _sqe_u;
690 #define	_sqe_entry	_sqe_u._sqe_s_entry._sqe_s_entry_entry
691 #define	_sqe_gbuf	_sqe_u._sqe_s_gbuf._sqe_s_gbuf_gbuf
692 };
693 
694 static inline void
695 octeon_eth_send_queue_add(struct octeon_eth_softc *sc, struct mbuf *m,
696     uint64_t *gbuf)
697 {
698 	struct _send_queue_entry *sqe = (struct _send_queue_entry *)m;
699 
700 	sqe->_sqe_gbuf = gbuf;
701 	SIMPLEQ_INSERT_TAIL(&sc->sc_sendq, sqe, _sqe_entry);
702 
703 	if ((m->m_flags & M_EXT) && m->m_ext.ext_free != NULL)
704 		sc->sc_ext_callback_cnt++;
705 
706 	OCTEON_EVCNT_INC(sc, txadd);
707 }
708 
709 static inline void
710 octeon_eth_send_queue_del(struct octeon_eth_softc *sc, struct mbuf **rm,
711     uint64_t **rgbuf)
712 {
713 	struct _send_queue_entry *sqe;
714 
715 	sqe = SIMPLEQ_FIRST(&sc->sc_sendq);
716 	OCTEON_ETH_KASSERT(sqe != NULL);
717 	SIMPLEQ_REMOVE_HEAD(&sc->sc_sendq, _sqe_entry);
718 
719 	*rm = (void *)sqe;
720 	*rgbuf = sqe->_sqe_gbuf;
721 
722 	if (((*rm)->m_flags & M_EXT) && (*rm)->m_ext.ext_free != NULL) {
723 		sc->sc_ext_callback_cnt--;
724 		OCTEON_ETH_KASSERT(sc->sc_ext_callback_cnt >= 0);
725 	}
726 
727 	OCTEON_EVCNT_INC(sc, txdel);
728 }
729 
730 static inline int
731 octeon_eth_buf_free_work(struct octeon_eth_softc *sc, uint64_t *work,
732     uint64_t word2)
733 {
734 	/* XXX when jumbo frame */
735 	if (ISSET(word2, PIP_WQE_WORD2_IP_BUFS)) {
736 		paddr_t addr;
737 		paddr_t start_buffer;
738 
739 		addr = work[3] & PIP_WQE_WORD3_ADDR;
740 		start_buffer = addr & ~(2048 - 1);
741 
742 		octeon_fpa_buf_put_paddr(octeon_eth_fb_pkt, start_buffer);
743 		OCTEON_EVCNT_INC(sc, rxbufpkput);
744 	}
745 
746 	octeon_fpa_buf_put(octeon_eth_fb_wqe, work);
747 	OCTEON_EVCNT_INC(sc, rxbufwqput);
748 
749 	return 0;
750 }
751 
752 static inline void
753 octeon_eth_buf_ext_free_m(struct mbuf *m, void *buf, size_t size, void *arg)
754 {
755 	uint64_t *work = (void *)arg;
756 #ifdef OCTEON_ETH_DEBUG
757 	struct octeon_eth_softc *sc = (void *)(uintptr_t)work[0];
758 #endif
759 	int s = splnet();
760 
761 	OCTEON_EVCNT_INC(sc, rxrs);
762 
763 	octeon_fpa_buf_put(octeon_eth_fb_wqe, work);
764 	OCTEON_EVCNT_INC(sc, rxbufwqput);
765 
766 	OCTEON_ETH_KASSERT(m != NULL);
767 
768 	pool_cache_put(mb_cache, m);
769 
770 	splx(s);
771 }
772 
773 static inline void
774 octeon_eth_buf_ext_free_ext(struct mbuf *m, void *buf, size_t size,
775     void *arg)
776 {
777 	uint64_t *work = (void *)arg;
778 #ifdef OCTEON_ETH_DEBUG
779 	struct octeon_eth_softc *sc = (void *)(uintptr_t)work[0];
780 #endif
781 	int s = splnet();
782 
783 	octeon_fpa_buf_put(octeon_eth_fb_wqe, work);
784 	OCTEON_EVCNT_INC(sc, rxbufwqput);
785 
786 	octeon_fpa_buf_put(octeon_eth_fb_pkt, buf);
787 	OCTEON_EVCNT_INC(sc, rxbufpkput);
788 
789 	OCTEON_ETH_KASSERT(m != NULL);
790 
791 	pool_cache_put(mb_cache, m);
792 
793 	splx(s);
794 }
795 
796 /* ---- ifnet interfaces */
797 
798 static int
799 octeon_eth_ioctl(struct ifnet *ifp, u_long cmd, void *data)
800 {
801 	struct octeon_eth_softc *sc = ifp->if_softc;
802 	struct ifreq *ifr = (struct ifreq *)data;
803 	int s, error;
804 
805 	s = splnet();
806 	switch (cmd) {
807 	case SIOCSIFMEDIA:
808 		/* Flow control requires full-duplex mode. */
809 		if (IFM_SUBTYPE(ifr->ifr_media) == IFM_AUTO ||
810 		    (ifr->ifr_media & IFM_FDX) == 0) {
811 			ifr->ifr_media &= ~IFM_ETH_FMASK;
812 		}
813 		if (IFM_SUBTYPE(ifr->ifr_media) != IFM_AUTO) {
814 			if ((ifr->ifr_media & IFM_ETH_FMASK) == IFM_FLOW) {
815 				ifr->ifr_media |=
816 				    IFM_ETH_TXPAUSE | IFM_ETH_RXPAUSE;
817 			}
818 			sc->sc_gmx_port->sc_port_flowflags =
819 				ifr->ifr_media & IFM_ETH_FMASK;
820 		}
821 		/* FALLTHROUGH */
822 	case SIOCGIFMEDIA:
823 		/* XXX: Flow contorol */
824 		error = ifmedia_ioctl(ifp, ifr, &sc->sc_mii.mii_media, cmd);
825 		break;
826 	default:
827 		error = ether_ioctl(ifp, cmd, data);
828 		if (error == ENETRESET) {
829 			/*
830 			 * Multicast list has changed; set the hardware filter
831 			 * accordingly.
832 			 */
833 			if (ISSET(ifp->if_flags, IFF_RUNNING))
834 				octeon_gmx_set_filter(sc->sc_gmx_port);
835 			error = 0;
836 		}
837 		break;
838 	}
839 	octeon_eth_start(ifp);
840 	splx(s);
841 
842 	return (error);
843 }
844 
845 /* ---- send (output) */
846 
847 static inline uint64_t
848 octeon_eth_send_makecmd_w0(uint64_t fau0, uint64_t fau1, size_t len, int segs)
849 {
850 	return octeon_pko_cmd_word0(
851 		OCT_FAU_OP_SIZE_64,		/* sz1 */
852 		OCT_FAU_OP_SIZE_64,		/* sz0 */
853 		1, fau1, 1, fau0,		/* s1, reg1, s0, reg0 */
854 		0,				/* le */
855 		octeon_eth_param_pko_cmd_w0_n2,	/* n2 */
856 		1, 0,				/* q, r */
857 		(segs == 1) ? 0 : 1,		/* g */
858 		0, 0, 1,			/* ipoffp1, ii, df */
859 		segs, (int)len);		/* segs, totalbytes */
860 }
861 
862 static inline uint64_t
863 octeon_eth_send_makecmd_w1(int size, paddr_t addr)
864 {
865 	return octeon_pko_cmd_word1(
866 		0, 0,				/* i, back */
867 		FPA_GATHER_BUFFER_POOL,		/* pool */
868 		size, addr);			/* size, addr */
869 }
870 
871 static inline int
872 octeon_eth_send_makecmd_gbuf(struct octeon_eth_softc *sc, struct mbuf *m0,
873     uint64_t *gbuf, int *rsegs)
874 {
875 	struct mbuf *m;
876 	int segs = 0;
877 	uintptr_t laddr, rlen, nlen;
878 
879 	for (m = m0; m != NULL; m = m->m_next) {
880 
881 		if (__predict_false(m->m_len == 0))
882 			continue;
883 
884 #if 0
885 		OCTEON_ETH_KASSERT(((uint32_t)m->m_data & (PAGE_SIZE - 1))
886 		   == (kvtophys((vaddr_t)m->m_data) & (PAGE_SIZE - 1)));
887 #endif
888 
889 		/*
890 		 * aligned 4k
891 		 */
892 		laddr = (uintptr_t)m->m_data & (PAGE_SIZE - 1);
893 
894 		if (laddr + m->m_len > PAGE_SIZE) {
895 			/* XXX XXX XXX */
896 			rlen = PAGE_SIZE - laddr;
897 			nlen = m->m_len - rlen;
898 			*(gbuf + segs) = octeon_eth_send_makecmd_w1(rlen,
899 			    kvtophys((vaddr_t)m->m_data));
900 			segs++;
901 			if (segs > 63) {
902 				return 1;
903 			}
904 			/* XXX XXX XXX */
905 		} else {
906 			rlen = 0;
907 			nlen = m->m_len;
908 		}
909 
910 		*(gbuf + segs) = octeon_eth_send_makecmd_w1(nlen,
911 		    kvtophys((vaddr_t)(m->m_data + rlen)));
912 		segs++;
913 		if (segs > 63) {
914 			return 1;
915 		}
916 	}
917 
918 	OCTEON_ETH_KASSERT(m == NULL);
919 
920 	*rsegs = segs;
921 
922 	return 0;
923 }
924 
925 static inline int
926 octeon_eth_send_makecmd(struct octeon_eth_softc *sc, struct mbuf *m,
927     uint64_t *gbuf, uint64_t *rpko_cmd_w0, uint64_t *rpko_cmd_w1)
928 {
929 	uint64_t pko_cmd_w0, pko_cmd_w1;
930 	int segs;
931 	int result = 0;
932 
933 	if (octeon_eth_send_makecmd_gbuf(sc, m, gbuf, &segs)) {
934 		log(LOG_WARNING, "%s: there are a lot of number of segments"
935 		    " of transmission data", device_xname(sc->sc_dev));
936 		result = 1;
937 		goto done;
938 	}
939 
940 	/*
941 	 * segs == 1	-> link mode (single continuous buffer)
942 	 *		   WORD1[size] is number of bytes pointed by segment
943 	 *
944 	 * segs > 1	-> gather mode (scatter-gather buffer)
945 	 *		   WORD1[size] is number of segments
946 	 */
947 	pko_cmd_w0 = octeon_eth_send_makecmd_w0(sc->sc_fau_done.fd_regno,
948 	    0, m->m_pkthdr.len, segs);
949 	if (segs == 1) {
950 		pko_cmd_w1 = octeon_eth_send_makecmd_w1(
951 		    m->m_pkthdr.len, kvtophys((vaddr_t)m->m_data));
952 	} else {
953 #ifdef __mips_n32
954 		KASSERT(MIPS_KSEG0_P(gbuf));
955 		pko_cmd_w1 = octeon_eth_send_makecmd_w1(segs,
956 		    MIPS_KSEG0_TO_PHYS(gbuf));
957 #else
958 		pko_cmd_w1 = octeon_eth_send_makecmd_w1(segs,
959 		    MIPS_XKPHYS_TO_PHYS(gbuf));
960 #endif
961 	}
962 
963 	*rpko_cmd_w0 = pko_cmd_w0;
964 	*rpko_cmd_w1 = pko_cmd_w1;
965 
966 done:
967 	return result;
968 }
969 
970 static inline int
971 octeon_eth_send_cmd(struct octeon_eth_softc *sc, uint64_t pko_cmd_w0,
972     uint64_t pko_cmd_w1)
973 {
974 	uint64_t *cmdptr;
975 	int result = 0;
976 
977 #ifdef __mips_n32
978 	KASSERT((sc->sc_cmdptr.cmdptr & ~MIPS_PHYS_MASK) == 0);
979 	cmdptr = (uint64_t *)MIPS_PHYS_TO_KSEG0(sc->sc_cmdptr.cmdptr);
980 #else
981 	cmdptr = (uint64_t *)MIPS_PHYS_TO_XKPHYS_CACHED(sc->sc_cmdptr.cmdptr);
982 #endif
983 	cmdptr += sc->sc_cmdptr.cmdptr_idx;
984 
985 	OCTEON_ETH_KASSERT(cmdptr != NULL);
986 
987 	*cmdptr++ = pko_cmd_w0;
988 	*cmdptr++ = pko_cmd_w1;
989 
990 	OCTEON_ETH_KASSERT(sc->sc_cmdptr.cmdptr_idx + 2 <= FPA_COMMAND_BUFFER_POOL_NWORDS - 1);
991 
992 	if (sc->sc_cmdptr.cmdptr_idx + 2 == FPA_COMMAND_BUFFER_POOL_NWORDS - 1) {
993 		paddr_t buf;
994 
995 		buf = octeon_fpa_buf_get_paddr(octeon_eth_fb_cmd);
996 		if (buf == 0) {
997 			log(LOG_WARNING,
998 			    "%s: can not allocate command buffer from free pool allocator\n",
999 			    device_xname(sc->sc_dev));
1000 			result = 1;
1001 			goto done;
1002 		}
1003 		OCTEON_EVCNT_INC(sc, txbufcbget);
1004 		*cmdptr++ = buf;
1005 		sc->sc_cmdptr.cmdptr = (uint64_t)buf;
1006 		sc->sc_cmdptr.cmdptr_idx = 0;
1007 	} else {
1008 		sc->sc_cmdptr.cmdptr_idx += 2;
1009 	}
1010 
1011 	octeon_pko_op_doorbell_write(sc->sc_port, sc->sc_port, 2);
1012 
1013 done:
1014 	return result;
1015 }
1016 
1017 static inline int
1018 octeon_eth_send_buf(struct octeon_eth_softc *sc, struct mbuf *m,
1019     uint64_t *gbuf)
1020 {
1021 	int result = 0, error;
1022 	uint64_t pko_cmd_w0, pko_cmd_w1;
1023 
1024 	error = octeon_eth_send_makecmd(sc, m, gbuf, &pko_cmd_w0, &pko_cmd_w1);
1025 	if (error != 0) {
1026 		/* already logging */
1027 		OCTEON_EVCNT_INC(sc, txerrmkcmd);
1028 		result = error;
1029 		goto done;
1030 	}
1031 
1032 	error = octeon_eth_send_cmd(sc, pko_cmd_w0, pko_cmd_w1);
1033 	if (error != 0) {
1034 		/* already logging */
1035 		OCTEON_EVCNT_INC(sc, txerrcmd);
1036 		result = error;
1037 	}
1038 
1039 done:
1040 	return result;
1041 }
1042 
1043 static inline int
1044 octeon_eth_send(struct octeon_eth_softc *sc, struct mbuf *m)
1045 {
1046 	paddr_t gaddr = 0;
1047 	uint64_t *gbuf = NULL;
1048 	int result = 0, error;
1049 
1050 	OCTEON_EVCNT_INC(sc, tx);
1051 
1052 	gaddr = octeon_fpa_buf_get_paddr(octeon_eth_fb_sg);
1053 	if (gaddr == 0) {
1054 		log(LOG_WARNING,
1055 		    "%s: can not allocate gather buffer from free pool allocator\n",
1056 		    device_xname(sc->sc_dev));
1057 		OCTEON_EVCNT_INC(sc, txerrgbuf);
1058 		result = 1;
1059 		goto done;
1060 	}
1061 	OCTEON_EVCNT_INC(sc, txbufgbget);
1062 
1063 #ifdef __mips_n32
1064 	KASSERT((gaddr & ~MIPS_PHYS_MASK) == 0);
1065 	gbuf = (uint64_t *)(uintptr_t)MIPS_PHYS_TO_KSEG0(gaddr);
1066 #else
1067 	gbuf = (uint64_t *)(uintptr_t)MIPS_PHYS_TO_XKPHYS_CACHED(gaddr);
1068 #endif
1069 
1070 	OCTEON_ETH_KASSERT(gbuf != NULL);
1071 
1072 	error = octeon_eth_send_buf(sc, m, gbuf);
1073 	if (error != 0) {
1074 		/* already logging */
1075 		octeon_fpa_buf_put_paddr(octeon_eth_fb_sg, gaddr);
1076 		OCTEON_EVCNT_INC(sc, txbufgbput);
1077 		result = error;
1078 		goto done;
1079 	}
1080 
1081 	octeon_eth_send_queue_add(sc, m, gbuf);
1082 
1083 done:
1084 	return result;
1085 }
1086 
1087 static void
1088 octeon_eth_start(struct ifnet *ifp)
1089 {
1090 	struct octeon_eth_softc *sc = ifp->if_softc;
1091 	struct mbuf *m;
1092 
1093 	/*
1094 	 * performance tuning
1095 	 * presend iobdma request
1096 	 */
1097 	octeon_eth_send_queue_flush_prefetch(sc);
1098 
1099 	if ((ifp->if_flags & (IFF_RUNNING | IFF_OACTIVE)) != IFF_RUNNING)
1100 		goto last;
1101 
1102 	/* XXX assume that OCTEON doesn't buffer packets */
1103 	if (__predict_false(!octeon_gmx_link_status(sc->sc_gmx_port))) {
1104 		/* dequeue and drop them */
1105 		while (1) {
1106 			IFQ_DEQUEUE(&ifp->if_snd, m);
1107 			if (m == NULL)
1108 				break;
1109 
1110 			m_freem(m);
1111 			IF_DROP(&ifp->if_snd);
1112 			OCTEON_EVCNT_INC(sc, txerrlink);
1113 		}
1114 		goto last;
1115 	}
1116 
1117 	for (;;) {
1118 		IFQ_POLL(&ifp->if_snd, m);
1119 		if (__predict_false(m == NULL))
1120 			break;
1121 
1122 		/* XXX XXX XXX */
1123 		octeon_eth_send_queue_flush_fetch(sc);
1124 
1125 		/*
1126 		 * If no free send buffer is available, free all the sent buffer
1127 		 * and bail out.
1128 		 */
1129 		if (octeon_eth_send_queue_is_full(sc)) {
1130 			return;
1131 		}
1132 		/* XXX XXX XXX */
1133 
1134 		IFQ_DEQUEUE(&ifp->if_snd, m);
1135 
1136 		bpf_mtap(ifp, m);
1137 
1138 		/* XXX XXX XXX */
1139 		if (sc->sc_soft_req_cnt > sc->sc_soft_req_thresh)
1140 			octeon_eth_send_queue_flush(sc);
1141 		if (octeon_eth_send(sc, m)) {
1142 			IF_DROP(&ifp->if_snd);
1143 			m_freem(m);
1144 			log(LOG_WARNING,
1145 		  	  "%s: failed in the transmission of the packet\n",
1146 		    	  device_xname(sc->sc_dev));
1147 			OCTEON_EVCNT_INC(sc, txerr);
1148 		} else {
1149 			sc->sc_soft_req_cnt++;
1150 		}
1151 		if (sc->sc_flush)
1152 			octeon_eth_send_queue_flush_sync(sc);
1153 		/* XXX XXX XXX */
1154 
1155 		/*
1156 		 * send next iobdma request
1157 		 */
1158 		octeon_eth_send_queue_flush_prefetch(sc);
1159 	}
1160 
1161 /*
1162  * Don't schedule send-buffer-free callout every time - those buffers are freed
1163  * by "free tick".  This makes some packets like NFS slower.
1164  */
1165 #ifdef OCTEON_ETH_USENFS
1166 	if (__predict_false(sc->sc_ext_callback_cnt > 0)) {
1167 		int timo;
1168 
1169 		/* ??? */
1170 		timo = hz - (100 * sc->sc_ext_callback_cnt);
1171 		if (timo < 10)
1172 			timo = 10;
1173 		callout_schedule(&sc->sc_tick_free_ch, timo);
1174 	}
1175 #endif
1176 
1177 last:
1178 	octeon_eth_send_queue_flush_fetch(sc);
1179 }
1180 
1181 static void
1182 octeon_eth_watchdog(struct ifnet *ifp)
1183 {
1184 	struct octeon_eth_softc *sc = ifp->if_softc;
1185 
1186 	printf("%s: device timeout\n", device_xname(sc->sc_dev));
1187 
1188 	octeon_eth_configure(sc);
1189 
1190 	SET(ifp->if_flags, IFF_RUNNING);
1191 	CLR(ifp->if_flags, IFF_OACTIVE);
1192 	ifp->if_timer = 0;
1193 
1194 	octeon_eth_start(ifp);
1195 }
1196 
1197 static int
1198 octeon_eth_init(struct ifnet *ifp)
1199 {
1200 	struct octeon_eth_softc *sc = ifp->if_softc;
1201 
1202 	/* XXX don't disable commonly used parts!!! XXX */
1203 	if (sc->sc_init_flag == 0) {
1204 		/* Cancel any pending I/O. */
1205 		octeon_eth_stop(ifp, 0);
1206 
1207 		/* Initialize the device */
1208 		octeon_eth_configure(sc);
1209 
1210 		octeon_pko_enable(sc->sc_pko);
1211 		octeon_ipd_enable(sc->sc_ipd);
1212 
1213 		sc->sc_init_flag = 1;
1214 	} else {
1215 		octeon_gmx_port_enable(sc->sc_gmx_port, 1);
1216 	}
1217 	octeon_eth_mediachange(ifp);
1218 
1219 	octeon_gmx_set_filter(sc->sc_gmx_port);
1220 
1221 	callout_reset(&sc->sc_tick_misc_ch, hz, octeon_eth_tick_misc, sc);
1222 	callout_reset(&sc->sc_tick_free_ch, hz, octeon_eth_tick_free, sc);
1223 
1224 	SET(ifp->if_flags, IFF_RUNNING);
1225 	CLR(ifp->if_flags, IFF_OACTIVE);
1226 
1227 	return 0;
1228 }
1229 
1230 static void
1231 octeon_eth_stop(struct ifnet *ifp, int disable)
1232 {
1233 	struct octeon_eth_softc *sc = ifp->if_softc;
1234 
1235 	callout_stop(&sc->sc_tick_misc_ch);
1236 	callout_stop(&sc->sc_tick_free_ch);
1237 
1238 	mii_down(&sc->sc_mii);
1239 
1240 	octeon_gmx_port_enable(sc->sc_gmx_port, 0);
1241 
1242 	/* Mark the interface as down and cancel the watchdog timer. */
1243 	CLR(ifp->if_flags, IFF_RUNNING | IFF_OACTIVE);
1244 	ifp->if_timer = 0;
1245 }
1246 
1247 /* ---- misc */
1248 
1249 #define PKO_INDEX_MASK	((1ULL << 12/* XXX */) - 1)
1250 
1251 static int
1252 octeon_eth_reset(struct octeon_eth_softc *sc)
1253 {
1254 	octeon_gmx_reset_speed(sc->sc_gmx_port);
1255 	octeon_gmx_reset_flowctl(sc->sc_gmx_port);
1256 	octeon_gmx_reset_timing(sc->sc_gmx_port);
1257 
1258 	return 0;
1259 }
1260 
1261 static int
1262 octeon_eth_configure(struct octeon_eth_softc *sc)
1263 {
1264 	octeon_gmx_port_enable(sc->sc_gmx_port, 0);
1265 
1266 	octeon_eth_reset(sc);
1267 
1268 	octeon_eth_configure_common(sc);
1269 
1270 	octeon_pko_port_config(sc->sc_pko);
1271 	octeon_pko_port_enable(sc->sc_pko, 1);
1272 	octeon_pip_port_config(sc->sc_pip);
1273 
1274 	octeon_gmx_tx_stats_rd_clr(sc->sc_gmx_port, 1);
1275 	octeon_gmx_rx_stats_rd_clr(sc->sc_gmx_port, 1);
1276 
1277 	octeon_gmx_port_enable(sc->sc_gmx_port, 1);
1278 
1279 	return 0;
1280 }
1281 
1282 static int
1283 octeon_eth_configure_common(struct octeon_eth_softc *sc)
1284 {
1285 	static int once;
1286 
1287 	if (once == 1)
1288 		return 0;
1289 	once = 1;
1290 
1291 	octeon_ipd_config(sc->sc_ipd);
1292 #ifdef OCTEON_ETH_IPD_RED
1293 	octeon_ipd_red(sc->sc_ipd, RECV_QUEUE_SIZE >> 2, RECV_QUEUE_SIZE >> 3);
1294 #endif
1295 	octeon_pko_config(sc->sc_pko);
1296 
1297 	octeon_pow_config(sc->sc_pow, OCTEON_POW_GROUP_PIP);
1298 
1299 	return 0;
1300 }
1301 
1302 /* ---- receive (input) */
1303 
1304 static inline int
1305 octeon_eth_recv_mbuf(struct octeon_eth_softc *sc, uint64_t *work,
1306     struct mbuf **rm)
1307 {
1308 	struct mbuf *m;
1309 	void (*ext_free)(struct mbuf *, void *, size_t, void *);
1310 	void *ext_buf;
1311 	size_t ext_size;
1312 	void *data;
1313 	uint64_t word1 = work[1];
1314 	uint64_t word2 = work[2];
1315 	uint64_t word3 = work[3];
1316 
1317 	MGETHDR(m, M_NOWAIT, MT_DATA);
1318 	if (m == NULL)
1319 		return 1;
1320 	OCTEON_ETH_KASSERT(m != NULL);
1321 
1322 	if ((word2 & PIP_WQE_WORD2_IP_BUFS) == 0) {
1323 		/* Dynamic short */
1324 		ext_free = octeon_eth_buf_ext_free_m;
1325 		ext_buf = &work[4];
1326 		ext_size = 96;
1327 
1328 		data = &work[4 + sc->sc_ip_offset / sizeof(uint64_t)];
1329 	} else {
1330 		vaddr_t addr;
1331 		vaddr_t start_buffer;
1332 
1333 #ifdef __mips_n32
1334 		KASSERT((word3 & ~MIPS_PHYS_MASK) == 0);
1335 		addr = MIPS_PHYS_TO_KSEG0(word3 & PIP_WQE_WORD3_ADDR);
1336 #else
1337 		addr = MIPS_PHYS_TO_XKPHYS_CACHED(word3 & PIP_WQE_WORD3_ADDR);
1338 #endif
1339 		start_buffer = addr & ~(2048 - 1);
1340 
1341 		ext_free = octeon_eth_buf_ext_free_ext;
1342 		ext_buf = (void *)start_buffer;
1343 		ext_size = 2048;
1344 
1345 		data = (void *)addr;
1346 	}
1347 
1348 	/* embed sc pointer into work[0] for _ext_free evcnt */
1349 	work[0] = (uintptr_t)sc;
1350 
1351 	MEXTADD(m, ext_buf, ext_size, 0, ext_free, work);
1352 	OCTEON_ETH_KASSERT(ISSET(m->m_flags, M_EXT));
1353 
1354 	m->m_data = data;
1355 	m->m_len = m->m_pkthdr.len = (word1 & PIP_WQE_WORD1_LEN) >> 48;
1356 	m_set_rcvif(m, &sc->sc_ethercom.ec_if);
1357 	/*
1358 	 * not readonly buffer
1359 	 */
1360 	m->m_flags |= M_EXT_RW;
1361 
1362 	*rm = m;
1363 
1364 	OCTEON_ETH_KASSERT(*rm != NULL);
1365 
1366 	return 0;
1367 }
1368 
1369 static inline int
1370 octeon_eth_recv_check_code(struct octeon_eth_softc *sc, uint64_t word2)
1371 {
1372 	uint64_t opecode = word2 & PIP_WQE_WORD2_NOIP_OPECODE;
1373 
1374 	if (__predict_true(!ISSET(word2, PIP_WQE_WORD2_NOIP_RE)))
1375 		return 0;
1376 
1377 	/* this error is harmless */
1378 	if (opecode == PIP_OVER_ERR)
1379 		return 0;
1380 
1381 	return 1;
1382 }
1383 
1384 static inline int
1385 octeon_eth_recv_check_jumbo(struct octeon_eth_softc *sc, uint64_t word2)
1386 {
1387 	if (__predict_false((word2 & PIP_WQE_WORD2_IP_BUFS) > (1ULL << 56)))
1388 		return 1;
1389 	return 0;
1390 }
1391 
1392 static inline int
1393 octeon_eth_recv_check_link(struct octeon_eth_softc *sc, uint64_t word2)
1394 {
1395 	if (__predict_false(!octeon_gmx_link_status(sc->sc_gmx_port)))
1396 		return 1;
1397 	return 0;
1398 }
1399 
1400 static inline int
1401 octeon_eth_recv_check(struct octeon_eth_softc *sc, uint64_t word2)
1402 {
1403 	if (__predict_false(octeon_eth_recv_check_link(sc, word2)) != 0) {
1404 		if (ratecheck(&sc->sc_rate_recv_check_link_last,
1405 		    &sc->sc_rate_recv_check_link_cap))
1406 			log(LOG_DEBUG,
1407 			    "%s: link is not up, the packet was dropped\n",
1408 			    device_xname(sc->sc_dev));
1409 		OCTEON_EVCNT_INC(sc, rxerrlink);
1410 		return 1;
1411 	}
1412 
1413 #if 0 /* XXX Performance tunig (Jumbo-frame is not supported yet!) */
1414 	if (__predict_false(octeon_eth_recv_check_jumbo(sc, word2)) != 0) {
1415 		/* XXX jumbo frame */
1416 		if (ratecheck(&sc->sc_rate_recv_check_jumbo_last,
1417 		    &sc->sc_rate_recv_check_jumbo_cap))
1418 			log(LOG_DEBUG,
1419 			    "jumbo frame was received\n");
1420 		OCTEON_EVCNT_INC(sc, rxerrjmb);
1421 		return 1;
1422 	}
1423 #endif
1424 
1425 	if (__predict_false(octeon_eth_recv_check_code(sc, word2)) != 0) {
1426 
1427 		if ((word2 & PIP_WQE_WORD2_NOIP_OPECODE) ==
1428 				PIP_WQE_WORD2_RE_OPCODE_LENGTH) {
1429 			/* no logging */
1430 			/* XXX inclement special error count */
1431 		} else if ((word2 & PIP_WQE_WORD2_NOIP_OPECODE) ==
1432 				PIP_WQE_WORD2_RE_OPCODE_PARTIAL) {
1433 			/* not an erorr. it's because of overload */
1434 		} else {
1435 
1436 			if (ratecheck(&sc->sc_rate_recv_check_code_last,
1437 			    &sc->sc_rate_recv_check_code_cap))
1438 				log(LOG_WARNING,
1439 				    "%s: the reception error had occured, "
1440 				    "the packet was dropped (error code = %" PRId64 ")\n",
1441 				    device_xname(sc->sc_dev), word2 & PIP_WQE_WORD2_NOIP_OPECODE);
1442 		}
1443 		OCTEON_EVCNT_INC(sc, rxerrcode);
1444 		return 1;
1445 	}
1446 
1447 	return 0;
1448 }
1449 
1450 static inline int
1451 octeon_eth_recv(struct octeon_eth_softc *sc, uint64_t *work)
1452 {
1453 	int result = 0;
1454 	struct ifnet *ifp;
1455 	struct mbuf *m;
1456 	uint64_t word2;
1457 
1458 	/* XXX XXX XXX */
1459 	/*
1460  	 * performance tuning
1461 	 * presend iobdma request
1462 	 */
1463 	if (sc->sc_soft_req_cnt > sc->sc_soft_req_thresh) {
1464 		octeon_eth_send_queue_flush_prefetch(sc);
1465 	}
1466 	/* XXX XXX XXX */
1467 
1468 	OCTEON_ETH_KASSERT(sc != NULL);
1469 	OCTEON_ETH_KASSERT(work != NULL);
1470 
1471 	OCTEON_EVCNT_INC(sc, rx);
1472 
1473 	word2 = work[2];
1474 	ifp = &sc->sc_ethercom.ec_if;
1475 
1476 	OCTEON_ETH_KASSERT(ifp != NULL);
1477 
1478 	if (__predict_false(octeon_eth_recv_check(sc, word2) != 0)) {
1479 		ifp->if_ierrors++;
1480 		result = 1;
1481 		octeon_eth_buf_free_work(sc, work, word2);
1482 		goto drop;
1483 	}
1484 
1485 	if (__predict_false(octeon_eth_recv_mbuf(sc, work, &m) != 0)) {
1486 		ifp->if_ierrors++;
1487 		result = 1;
1488 		octeon_eth_buf_free_work(sc, work, word2);
1489 		goto drop;
1490 	}
1491 
1492 	/* work[0] .. work[3] may not be valid any more */
1493 
1494 	OCTEON_ETH_KASSERT(m != NULL);
1495 
1496 	octeon_ipd_offload(word2, m->m_data, &m->m_pkthdr.csum_flags);
1497 
1498 	/* XXX XXX XXX */
1499 	if (sc->sc_soft_req_cnt > sc->sc_soft_req_thresh) {
1500 		octeon_eth_send_queue_flush_fetch(sc);
1501 		octeon_eth_send_queue_flush(sc);
1502 	}
1503 
1504 	/* XXX XXX XXX */
1505 	if (sc->sc_flush)
1506 		octeon_eth_send_queue_flush_sync(sc);
1507 	/* XXX XXX XXX */
1508 
1509 	if_percpuq_enqueue(ifp->if_percpuq, m);
1510 
1511 	return 0;
1512 
1513 drop:
1514 	/* XXX XXX XXX */
1515 	if (sc->sc_soft_req_cnt > sc->sc_soft_req_thresh) {
1516 		octeon_eth_send_queue_flush_fetch(sc);
1517 	}
1518 	/* XXX XXX XXX */
1519 
1520 	return result;
1521 }
1522 
1523 static void
1524 octeon_eth_recv_redir(struct ifnet *ifp, struct mbuf *m)
1525 {
1526 	struct octeon_eth_softc *rsc = ifp->if_softc;
1527 	struct octeon_eth_softc *sc = NULL;
1528 	int i;
1529 
1530 	for (i = 0; i < 3 /* XXX */; i++) {
1531 		if (rsc->sc_redir & (1 << i))
1532 			sc = octeon_eth_gsc[i];
1533 	}
1534 
1535 	if (sc == NULL) {
1536 		m_freem(m);
1537 		return;
1538 	}
1539 	octeon_eth_send_queue_flush_prefetch(sc);
1540 
1541 	octeon_eth_send_queue_flush_fetch(sc);
1542 
1543 	if (octeon_eth_send_queue_is_full(sc)) {
1544 		m_freem(m);
1545 		return;
1546 	}
1547 	if (sc->sc_soft_req_cnt > sc->sc_soft_req_thresh)
1548 		octeon_eth_send_queue_flush(sc);
1549 
1550 	if (octeon_eth_send(sc, m)) {
1551 		IF_DROP(&ifp->if_snd);
1552 		m_freem(m);
1553 	} else {
1554 		sc->sc_soft_req_cnt++;
1555 	}
1556 
1557 	if (sc->sc_flush)
1558 		octeon_eth_send_queue_flush_sync(sc);
1559 }
1560 
1561 static inline void
1562 octeon_eth_recv_intr(void *data, uint64_t *work)
1563 {
1564 	struct octeon_eth_softc *sc;
1565 	int port;
1566 
1567 	OCTEON_ETH_KASSERT(work != NULL);
1568 
1569 	port = (work[1] & PIP_WQE_WORD1_IPRT) >> 42;
1570 
1571 	OCTEON_ETH_KASSERT(port < GMX_PORT_NUNITS);
1572 
1573 	sc = octeon_eth_gsc[port];
1574 
1575 	OCTEON_ETH_KASSERT(sc != NULL);
1576 	OCTEON_ETH_KASSERT(port == sc->sc_port);
1577 
1578 	/* XXX process all work queue entries anyway */
1579 
1580 	(void)octeon_eth_recv(sc, work);
1581 }
1582 
1583 /* ---- tick */
1584 
1585 /*
1586  * octeon_eth_tick_free
1587  *
1588  * => garbage collect send gather buffer / mbuf
1589  * => called at softclock
1590  */
1591 static void
1592 octeon_eth_tick_free(void *arg)
1593 {
1594 	struct octeon_eth_softc *sc = arg;
1595 	int timo;
1596 	int s;
1597 
1598 	s = splnet();
1599 	/* XXX XXX XXX */
1600 	if (sc->sc_soft_req_cnt > 0) {
1601 		octeon_eth_send_queue_flush_prefetch(sc);
1602 		octeon_eth_send_queue_flush_fetch(sc);
1603 		octeon_eth_send_queue_flush(sc);
1604 		octeon_eth_send_queue_flush_sync(sc);
1605 	}
1606 	/* XXX XXX XXX */
1607 
1608 	/* XXX XXX XXX */
1609 	/* ??? */
1610 	timo = hz - (100 * sc->sc_ext_callback_cnt);
1611 	if (timo < 10)
1612 		 timo = 10;
1613 	callout_schedule(&sc->sc_tick_free_ch, timo);
1614 	/* XXX XXX XXX */
1615 	splx(s);
1616 }
1617 
1618 /*
1619  * octeon_eth_tick_misc
1620  *
1621  * => collect statistics
1622  * => check link status
1623  * => called at softclock
1624  */
1625 static void
1626 octeon_eth_tick_misc(void *arg)
1627 {
1628 	struct octeon_eth_softc *sc = arg;
1629 	struct ifnet *ifp;
1630 	int s;
1631 
1632 	s = splnet();
1633 
1634 	ifp = &sc->sc_ethercom.ec_if;
1635 
1636 	octeon_gmx_stats(sc->sc_gmx_port);
1637 	octeon_pip_stats(sc->sc_pip, ifp, sc->sc_port);
1638 	mii_tick(&sc->sc_mii);
1639 
1640 	splx(s);
1641 
1642 	callout_schedule(&sc->sc_tick_misc_ch, hz);
1643 }
1644 
1645 /* ---- odd nibble preamble workaround (software CRC processing) */
1646 
1647 /* ---- sysctl */
1648 
1649 static int	octeon_eth_sysctl_verify(SYSCTLFN_ARGS);
1650 static int	octeon_eth_sysctl_pool(SYSCTLFN_ARGS);
1651 static int	octeon_eth_sysctl_rd(SYSCTLFN_ARGS);
1652 
1653 static int	octeon_eth_sysctl_pkocmdw0n2_num;
1654 static int	octeon_eth_sysctl_pipdynrs_num;
1655 static int	octeon_eth_sysctl_redir_num;
1656 static int	octeon_eth_sysctl_pkt_pool_num;
1657 static int	octeon_eth_sysctl_wqe_pool_num;
1658 static int	octeon_eth_sysctl_cmd_pool_num;
1659 static int	octeon_eth_sysctl_sg_pool_num;
1660 static int	octeon_eth_sysctl_pktbuf_num;
1661 
1662 /*
1663  * Set up sysctl(3) MIB, hw.cnmac.*.
1664  */
1665 SYSCTL_SETUP(sysctl_octeon_eth, "sysctl cnmac subtree setup")
1666 {
1667 	int rc;
1668 	int octeon_eth_sysctl_root_num;
1669 	const struct sysctlnode *node;
1670 
1671 	if ((rc = sysctl_createv(clog, 0, NULL, NULL,
1672 	    0, CTLTYPE_NODE, "hw", NULL,
1673 	    NULL, 0, NULL, 0, CTL_HW, CTL_EOL)) != 0) {
1674 		goto err;
1675 	}
1676 
1677 	if ((rc = sysctl_createv(clog, 0, NULL, &node,
1678 	    0, CTLTYPE_NODE, "cnmac",
1679 	    SYSCTL_DESCR("cnmac interface controls"),
1680 	    NULL, 0, NULL, 0, CTL_HW, CTL_CREATE, CTL_EOL)) != 0) {
1681 		goto err;
1682 	}
1683 
1684 	octeon_eth_sysctl_root_num = node->sysctl_num;
1685 
1686 	if ((rc = sysctl_createv(clog, 0, NULL, &node,
1687 	    CTLFLAG_PERMANENT | CTLFLAG_READWRITE,
1688 	    CTLTYPE_INT, "pko_cmd_w0_n2",
1689 	    SYSCTL_DESCR("PKO command WORD0 N2 bit"),
1690 	    octeon_eth_sysctl_verify, 0,
1691 	    &octeon_eth_param_pko_cmd_w0_n2,
1692 	    0, CTL_HW, octeon_eth_sysctl_root_num, CTL_CREATE,
1693 	    CTL_EOL)) != 0) {
1694 	    goto err;
1695 	}
1696 
1697 	octeon_eth_sysctl_pkocmdw0n2_num = node->sysctl_num;
1698 
1699 	if ((rc = sysctl_createv(clog, 0, NULL, &node,
1700 	    CTLFLAG_PERMANENT | CTLFLAG_READWRITE,
1701 	    CTLTYPE_INT, "pip_dyn_rs",
1702 	    SYSCTL_DESCR("PIP dynamic short in WQE"),
1703 	    octeon_eth_sysctl_verify, 0,
1704 	    &octeon_eth_param_pip_dyn_rs,
1705 	    0, CTL_HW, octeon_eth_sysctl_root_num, CTL_CREATE,
1706 	    CTL_EOL)) != 0) {
1707 	    goto err;
1708 	}
1709 
1710 	octeon_eth_sysctl_pipdynrs_num = node->sysctl_num;
1711 
1712 	if ((rc = sysctl_createv(clog, 0, NULL, &node,
1713 	    CTLFLAG_PERMANENT | CTLFLAG_READWRITE,
1714 	    CTLTYPE_INT, "redir",
1715 	    SYSCTL_DESCR("input port redirection"),
1716 	    octeon_eth_sysctl_verify, 0,
1717 	    &octeon_eth_param_redir,
1718 	    0, CTL_HW, octeon_eth_sysctl_root_num, CTL_CREATE,
1719 	    CTL_EOL)) != 0) {
1720 	    goto err;
1721 	}
1722 
1723 	octeon_eth_sysctl_redir_num = node->sysctl_num;
1724 
1725 	if ((rc = sysctl_createv(clog, 0, NULL, &node,
1726 	    CTLFLAG_PERMANENT,
1727 	    CTLTYPE_INT, "pkt_pool",
1728 	    SYSCTL_DESCR("packet pool available"),
1729 	    octeon_eth_sysctl_pool, 0, NULL,
1730 	    0, CTL_HW, octeon_eth_sysctl_root_num, CTL_CREATE,
1731 	    CTL_EOL)) != 0) {
1732 	    goto err;
1733 	}
1734 
1735 	octeon_eth_sysctl_pkt_pool_num = node->sysctl_num;
1736 
1737 	if ((rc = sysctl_createv(clog, 0, NULL, &node,
1738 	    CTLFLAG_PERMANENT,
1739 	    CTLTYPE_INT, "wqe_pool",
1740 	    SYSCTL_DESCR("wqe pool available"),
1741 	    octeon_eth_sysctl_pool, 0, NULL,
1742 	    0, CTL_HW, octeon_eth_sysctl_root_num, CTL_CREATE,
1743 	    CTL_EOL)) != 0) {
1744 	    goto err;
1745 	}
1746 
1747 	octeon_eth_sysctl_wqe_pool_num = node->sysctl_num;
1748 
1749 	if ((rc = sysctl_createv(clog, 0, NULL, &node,
1750 	    CTLFLAG_PERMANENT,
1751 	    CTLTYPE_INT, "cmd_pool",
1752 	    SYSCTL_DESCR("cmd pool available"),
1753 	    octeon_eth_sysctl_pool, 0, NULL,
1754 	    0, CTL_HW, octeon_eth_sysctl_root_num, CTL_CREATE,
1755 	    CTL_EOL)) != 0) {
1756 	    goto err;
1757 	}
1758 
1759 	octeon_eth_sysctl_cmd_pool_num = node->sysctl_num;
1760 
1761 	if ((rc = sysctl_createv(clog, 0, NULL, &node,
1762 	    CTLFLAG_PERMANENT,
1763 	    CTLTYPE_INT, "sg_pool",
1764 	    SYSCTL_DESCR("sg pool available"),
1765 	    octeon_eth_sysctl_pool, 0, NULL,
1766 	    0, CTL_HW, octeon_eth_sysctl_root_num, CTL_CREATE,
1767 	    CTL_EOL)) != 0) {
1768 	    goto err;
1769 	}
1770 
1771 	octeon_eth_sysctl_sg_pool_num = node->sysctl_num;
1772 
1773 	if ((rc = sysctl_createv(clog, 0, NULL, &node,
1774 	    CTLFLAG_PERMANENT | CTLFLAG_READONLY,
1775 	    CTLTYPE_INT, "pktbuf",
1776 	    SYSCTL_DESCR("input packet buffer size on POW"),
1777 	    octeon_eth_sysctl_rd, 0,
1778 	    &octeon_eth_param_pktbuf,
1779 	    0, CTL_HW, octeon_eth_sysctl_root_num, CTL_CREATE,
1780 	    CTL_EOL)) != 0) {
1781 	    goto err;
1782 	}
1783 
1784 	octeon_eth_sysctl_pktbuf_num = node->sysctl_num;
1785 
1786 	return;
1787 
1788 err:
1789 	aprint_error("%s: syctl_createv failed (rc = %d)\n", __func__, rc);
1790 }
1791 
1792 static int
1793 octeon_eth_sysctl_verify(SYSCTLFN_ARGS)
1794 {
1795 	int error, v;
1796 	struct sysctlnode node;
1797 	struct octeon_eth_softc *sc;
1798 	int i;
1799 	int s;
1800 
1801 	node = *rnode;
1802 	v = *(int *)rnode->sysctl_data;
1803 	node.sysctl_data = &v;
1804 	error = sysctl_lookup(SYSCTLFN_CALL(&node));
1805 	if (error || newp == NULL)
1806 		return error;
1807 
1808 	if (node.sysctl_num == octeon_eth_sysctl_pkocmdw0n2_num) {
1809 		if (v < 0 || v > 1)
1810 			return EINVAL;
1811 		*(int *)rnode->sysctl_data = v;
1812 		return 0;
1813 	}
1814 
1815 	if (node.sysctl_num == octeon_eth_sysctl_pipdynrs_num) {
1816 		if (v < 0 || v > 1)
1817 			return EINVAL;
1818 		*(int *)rnode->sysctl_data = v;
1819 		s = splnet();
1820 		for (i = 0; i < 3/* XXX */; i++) {
1821 			sc = octeon_eth_gsc[i];	/* XXX */
1822 			octeon_pip_prt_cfg_enable(sc->sc_pip, PIP_PRT_CFGN_DYN_RS, v);
1823 		}
1824 		splx(s);
1825 		return 0;
1826 	}
1827 
1828 	if (node.sysctl_num == octeon_eth_sysctl_redir_num) {
1829 		if (v & ~((0x7 << (4 * 0)) | (0x7 << (4 * 1)) | (0x7 << (4 * 2))))
1830 			return EINVAL;
1831 		*(int *)rnode->sysctl_data = v;
1832 		s = splnet();
1833 		for (i = 0; i < 3/* XXX */; i++) {
1834 			struct ifnet *ifp;
1835 
1836 			sc = octeon_eth_gsc[i];	/* XXX */
1837 			ifp = &sc->sc_ethercom.ec_if;
1838 
1839 			sc->sc_redir = (octeon_eth_param_redir >> (4 * i)) & 0x7;
1840 			if (sc->sc_redir == 0) {
1841 				if (ISSET(ifp->if_flags, IFF_PROMISC)) {
1842 					CLR(ifp->if_flags, IFF_PROMISC);
1843 					octeon_eth_mii_statchg(ifp);
1844 					/* octeon_gmx_set_filter(sc->sc_gmx_port); */
1845 				}
1846 				ifp->_if_input = ether_input;
1847 			}
1848 			else {
1849 				if (!ISSET(ifp->if_flags, IFF_PROMISC)) {
1850 					SET(ifp->if_flags, IFF_PROMISC);
1851 					octeon_eth_mii_statchg(ifp);
1852 					/* octeon_gmx_set_filter(sc->sc_gmx_port); */
1853 				}
1854 				ifp->_if_input = octeon_eth_recv_redir;
1855 			}
1856 		}
1857 		splx(s);
1858 		return 0;
1859 	}
1860 
1861 	return EINVAL;
1862 }
1863 
1864 static int
1865 octeon_eth_sysctl_pool(SYSCTLFN_ARGS)
1866 {
1867 	int error, newval = 0;
1868 	struct sysctlnode node;
1869 	int s;
1870 
1871 	node = *rnode;
1872 	node.sysctl_data = &newval;
1873 	s = splnet();
1874 	if (node.sysctl_num == octeon_eth_sysctl_pkt_pool_num) {
1875 		error = octeon_fpa_available_fpa_pool(&newval, OCTEON_POOL_NO_PKT);
1876 	} else if (node.sysctl_num == octeon_eth_sysctl_wqe_pool_num) {
1877 		error = octeon_fpa_available_fpa_pool(&newval, OCTEON_POOL_NO_WQE);
1878 	} else if (node.sysctl_num == octeon_eth_sysctl_cmd_pool_num) {
1879 		error = octeon_fpa_available_fpa_pool(&newval, OCTEON_POOL_NO_CMD);
1880 	} else if (node.sysctl_num == octeon_eth_sysctl_sg_pool_num) {
1881 		error = octeon_fpa_available_fpa_pool(&newval, OCTEON_POOL_NO_SG);
1882 	} else {
1883 		splx(s);
1884 		return EINVAL;
1885 	}
1886 	splx(s);
1887 	if (error)
1888 		return error;
1889 	error = sysctl_lookup(SYSCTLFN_CALL(&node));
1890 	if (error || newp == NULL)
1891 		return error;
1892 
1893 	return 0;
1894 }
1895 
1896 static int
1897 octeon_eth_sysctl_rd(SYSCTLFN_ARGS)
1898 {
1899 	int error, v;
1900 	struct sysctlnode node;
1901 	int s;
1902 
1903 	node = *rnode;
1904 	v = *(int *)rnode->sysctl_data;
1905 	node.sysctl_data = &v;
1906 	error = sysctl_lookup(SYSCTLFN_CALL(&node));
1907 	if (error || newp != NULL)
1908 		return error;
1909 
1910 	if (node.sysctl_num == octeon_eth_sysctl_pktbuf_num) {
1911 		uint64_t tmp;
1912 		int n;
1913 
1914 		s = splnet();
1915 		tmp = octeon_fpa_query(0);
1916 		n = (int)tmp;
1917 		splx(s);
1918 		*(int *)rnode->sysctl_data = n;
1919 		octeon_eth_param_pktbuf = n;
1920 		*(int *)oldp = n;
1921 		return 0;
1922 	}
1923 
1924 	return EINVAL;
1925 }
1926