xref: /dpdk/drivers/net/bnx2x/bnx2x_ethdev.c (revision 0857b942113874c69dc3db5df11a828ee3cc9b6b)
1 /*
2  * Copyright (c) 2013-2015 Brocade Communications Systems, Inc.
3  *
4  * Copyright (c) 2015 QLogic Corporation.
5  * All rights reserved.
6  * www.qlogic.com
7  *
8  * See LICENSE.bnx2x_pmd for copyright and licensing details.
9  */
10 
11 #include "bnx2x.h"
12 #include "bnx2x_rxtx.h"
13 
14 #include <rte_dev.h>
15 
16 /*
17  * The set of PCI devices this driver supports
18  */
19 #define BROADCOM_PCI_VENDOR_ID 0x14E4
20 static const struct rte_pci_id pci_id_bnx2x_map[] = {
21 	{ RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57800) },
22 	{ RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57711) },
23 	{ RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57810) },
24 	{ RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57811) },
25 	{ RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57840_OBS) },
26 	{ RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57840_4_10) },
27 	{ RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57840_2_20) },
28 #ifdef RTE_LIBRTE_BNX2X_MF_SUPPORT
29 	{ RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57810_MF) },
30 	{ RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57811_MF) },
31 	{ RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57840_MF) },
32 #endif
33 	{ .vendor_id = 0, }
34 };
35 
36 static const struct rte_pci_id pci_id_bnx2xvf_map[] = {
37 	{ RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57800_VF) },
38 	{ RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57810_VF) },
39 	{ RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57811_VF) },
40 	{ RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57840_VF) },
41 	{ .vendor_id = 0, }
42 };
43 
44 struct rte_bnx2x_xstats_name_off {
45 	char name[RTE_ETH_XSTATS_NAME_SIZE];
46 	uint32_t offset_hi;
47 	uint32_t offset_lo;
48 };
49 
50 static const struct rte_bnx2x_xstats_name_off bnx2x_xstats_strings[] = {
51 	{"rx_buffer_drops",
52 		offsetof(struct bnx2x_eth_stats, brb_drop_hi),
53 		offsetof(struct bnx2x_eth_stats, brb_drop_lo)},
54 	{"rx_buffer_truncates",
55 		offsetof(struct bnx2x_eth_stats, brb_truncate_hi),
56 		offsetof(struct bnx2x_eth_stats, brb_truncate_lo)},
57 	{"rx_buffer_truncate_discard",
58 		offsetof(struct bnx2x_eth_stats, brb_truncate_discard),
59 		offsetof(struct bnx2x_eth_stats, brb_truncate_discard)},
60 	{"mac_filter_discard",
61 		offsetof(struct bnx2x_eth_stats, mac_filter_discard),
62 		offsetof(struct bnx2x_eth_stats, mac_filter_discard)},
63 	{"no_match_vlan_tag_discard",
64 		offsetof(struct bnx2x_eth_stats, mf_tag_discard),
65 		offsetof(struct bnx2x_eth_stats, mf_tag_discard)},
66 	{"tx_pause",
67 		offsetof(struct bnx2x_eth_stats, pause_frames_sent_hi),
68 		offsetof(struct bnx2x_eth_stats, pause_frames_sent_lo)},
69 	{"rx_pause",
70 		offsetof(struct bnx2x_eth_stats, pause_frames_received_hi),
71 		offsetof(struct bnx2x_eth_stats, pause_frames_received_lo)},
72 	{"tx_priority_flow_control",
73 		offsetof(struct bnx2x_eth_stats, pfc_frames_sent_hi),
74 		offsetof(struct bnx2x_eth_stats, pfc_frames_sent_lo)},
75 	{"rx_priority_flow_control",
76 		offsetof(struct bnx2x_eth_stats, pfc_frames_received_hi),
77 		offsetof(struct bnx2x_eth_stats, pfc_frames_received_lo)}
78 };
79 
80 static void
81 bnx2x_link_update(struct rte_eth_dev *dev)
82 {
83 	struct bnx2x_softc *sc = dev->data->dev_private;
84 
85 	PMD_INIT_FUNC_TRACE();
86 	bnx2x_link_status_update(sc);
87 	mb();
88 	dev->data->dev_link.link_speed = sc->link_vars.line_speed;
89 	switch (sc->link_vars.duplex) {
90 		case DUPLEX_FULL:
91 			dev->data->dev_link.link_duplex = ETH_LINK_FULL_DUPLEX;
92 			break;
93 		case DUPLEX_HALF:
94 			dev->data->dev_link.link_duplex = ETH_LINK_HALF_DUPLEX;
95 			break;
96 	}
97 	dev->data->dev_link.link_autoneg = !(dev->data->dev_conf.link_speeds &
98 			ETH_LINK_SPEED_FIXED);
99 	dev->data->dev_link.link_status = sc->link_vars.link_up;
100 }
101 
102 static void
103 bnx2x_interrupt_action(struct rte_eth_dev *dev)
104 {
105 	struct bnx2x_softc *sc = dev->data->dev_private;
106 	uint32_t link_status;
107 
108 	PMD_DEBUG_PERIODIC_LOG(INFO, "Interrupt handled");
109 
110 	bnx2x_intr_legacy(sc, 0);
111 
112 	if (sc->periodic_flags & PERIODIC_GO)
113 		bnx2x_periodic_callout(sc);
114 	link_status = REG_RD(sc, sc->link_params.shmem_base +
115 			offsetof(struct shmem_region,
116 				port_mb[sc->link_params.port].link_status));
117 	if ((link_status & LINK_STATUS_LINK_UP) != dev->data->dev_link.link_status)
118 		bnx2x_link_update(dev);
119 }
120 
121 static __rte_unused void
122 bnx2x_interrupt_handler(struct rte_intr_handle *handle, void *param)
123 {
124 	struct rte_eth_dev *dev = (struct rte_eth_dev *)param;
125 
126 	bnx2x_interrupt_action(dev);
127 	rte_intr_enable(handle);
128 }
129 
130 /*
131  * Devops - helper functions can be called from user application
132  */
133 
134 static int
135 bnx2x_dev_configure(struct rte_eth_dev *dev)
136 {
137 	struct bnx2x_softc *sc = dev->data->dev_private;
138 	int mp_ncpus = sysconf(_SC_NPROCESSORS_CONF);
139 
140 	PMD_INIT_FUNC_TRACE();
141 
142 	if (dev->data->dev_conf.rxmode.jumbo_frame)
143 		sc->mtu = dev->data->dev_conf.rxmode.max_rx_pkt_len;
144 
145 	if (dev->data->nb_tx_queues > dev->data->nb_rx_queues) {
146 		PMD_DRV_LOG(ERR, "The number of TX queues is greater than number of RX queues");
147 		return -EINVAL;
148 	}
149 
150 	sc->num_queues = MAX(dev->data->nb_rx_queues, dev->data->nb_tx_queues);
151 	if (sc->num_queues > mp_ncpus) {
152 		PMD_DRV_LOG(ERR, "The number of queues is more than number of CPUs");
153 		return -EINVAL;
154 	}
155 
156 	PMD_DRV_LOG(DEBUG, "num_queues=%d, mtu=%d",
157 		       sc->num_queues, sc->mtu);
158 
159 	/* allocate ilt */
160 	if (bnx2x_alloc_ilt_mem(sc) != 0) {
161 		PMD_DRV_LOG(ERR, "bnx2x_alloc_ilt_mem was failed");
162 		return -ENXIO;
163 	}
164 
165 	/* allocate the host hardware/software hsi structures */
166 	if (bnx2x_alloc_hsi_mem(sc) != 0) {
167 		PMD_DRV_LOG(ERR, "bnx2x_alloc_hsi_mem was failed");
168 		bnx2x_free_ilt_mem(sc);
169 		return -ENXIO;
170 	}
171 
172 	return 0;
173 }
174 
175 static int
176 bnx2x_dev_start(struct rte_eth_dev *dev)
177 {
178 	struct bnx2x_softc *sc = dev->data->dev_private;
179 	int ret = 0;
180 
181 	PMD_INIT_FUNC_TRACE();
182 
183 	ret = bnx2x_init(sc);
184 	if (ret) {
185 		PMD_DRV_LOG(DEBUG, "bnx2x_init failed (%d)", ret);
186 		return -1;
187 	}
188 
189 	if (IS_PF(sc)) {
190 		rte_intr_callback_register(&sc->pci_dev->intr_handle,
191 				bnx2x_interrupt_handler, (void *)dev);
192 
193 		if (rte_intr_enable(&sc->pci_dev->intr_handle))
194 			PMD_DRV_LOG(ERR, "rte_intr_enable failed");
195 	}
196 
197 	ret = bnx2x_dev_rx_init(dev);
198 	if (ret != 0) {
199 		PMD_DRV_LOG(DEBUG, "bnx2x_dev_rx_init returned error code");
200 		return -3;
201 	}
202 
203 	/* Print important adapter info for the user. */
204 	bnx2x_print_adapter_info(sc);
205 
206 	return ret;
207 }
208 
209 static void
210 bnx2x_dev_stop(struct rte_eth_dev *dev)
211 {
212 	struct bnx2x_softc *sc = dev->data->dev_private;
213 	int ret = 0;
214 
215 	PMD_INIT_FUNC_TRACE();
216 
217 	if (IS_PF(sc)) {
218 		rte_intr_disable(&sc->pci_dev->intr_handle);
219 		rte_intr_callback_unregister(&sc->pci_dev->intr_handle,
220 				bnx2x_interrupt_handler, (void *)dev);
221 	}
222 
223 	ret = bnx2x_nic_unload(sc, UNLOAD_NORMAL, FALSE);
224 	if (ret) {
225 		PMD_DRV_LOG(DEBUG, "bnx2x_nic_unload failed (%d)", ret);
226 		return;
227 	}
228 
229 	return;
230 }
231 
232 static void
233 bnx2x_dev_close(struct rte_eth_dev *dev)
234 {
235 	struct bnx2x_softc *sc = dev->data->dev_private;
236 
237 	PMD_INIT_FUNC_TRACE();
238 
239 	if (IS_VF(sc))
240 		bnx2x_vf_close(sc);
241 
242 	bnx2x_dev_clear_queues(dev);
243 	memset(&(dev->data->dev_link), 0 , sizeof(struct rte_eth_link));
244 
245 	/* free the host hardware/software hsi structures */
246 	bnx2x_free_hsi_mem(sc);
247 
248 	/* free ilt */
249 	bnx2x_free_ilt_mem(sc);
250 }
251 
252 static void
253 bnx2x_promisc_enable(struct rte_eth_dev *dev)
254 {
255 	struct bnx2x_softc *sc = dev->data->dev_private;
256 
257 	PMD_INIT_FUNC_TRACE();
258 	sc->rx_mode = BNX2X_RX_MODE_PROMISC;
259 	if (rte_eth_allmulticast_get(dev->data->port_id) == 1)
260 		sc->rx_mode = BNX2X_RX_MODE_ALLMULTI_PROMISC;
261 	bnx2x_set_rx_mode(sc);
262 }
263 
264 static void
265 bnx2x_promisc_disable(struct rte_eth_dev *dev)
266 {
267 	struct bnx2x_softc *sc = dev->data->dev_private;
268 
269 	PMD_INIT_FUNC_TRACE();
270 	sc->rx_mode = BNX2X_RX_MODE_NORMAL;
271 	if (rte_eth_allmulticast_get(dev->data->port_id) == 1)
272 		sc->rx_mode = BNX2X_RX_MODE_ALLMULTI;
273 	bnx2x_set_rx_mode(sc);
274 }
275 
276 static void
277 bnx2x_dev_allmulticast_enable(struct rte_eth_dev *dev)
278 {
279 	struct bnx2x_softc *sc = dev->data->dev_private;
280 
281 	PMD_INIT_FUNC_TRACE();
282 	sc->rx_mode = BNX2X_RX_MODE_ALLMULTI;
283 	if (rte_eth_promiscuous_get(dev->data->port_id) == 1)
284 		sc->rx_mode = BNX2X_RX_MODE_ALLMULTI_PROMISC;
285 	bnx2x_set_rx_mode(sc);
286 }
287 
288 static void
289 bnx2x_dev_allmulticast_disable(struct rte_eth_dev *dev)
290 {
291 	struct bnx2x_softc *sc = dev->data->dev_private;
292 
293 	PMD_INIT_FUNC_TRACE();
294 	sc->rx_mode = BNX2X_RX_MODE_NORMAL;
295 	if (rte_eth_promiscuous_get(dev->data->port_id) == 1)
296 		sc->rx_mode = BNX2X_RX_MODE_PROMISC;
297 	bnx2x_set_rx_mode(sc);
298 }
299 
300 static int
301 bnx2x_dev_link_update(struct rte_eth_dev *dev, __rte_unused int wait_to_complete)
302 {
303 	PMD_INIT_FUNC_TRACE();
304 
305 	int old_link_status = dev->data->dev_link.link_status;
306 
307 	bnx2x_link_update(dev);
308 
309 	return old_link_status == dev->data->dev_link.link_status ? -1 : 0;
310 }
311 
312 static int
313 bnx2xvf_dev_link_update(struct rte_eth_dev *dev, __rte_unused int wait_to_complete)
314 {
315 	int old_link_status = dev->data->dev_link.link_status;
316 	struct bnx2x_softc *sc = dev->data->dev_private;
317 
318 	bnx2x_link_update(dev);
319 
320 	bnx2x_check_bull(sc);
321 	if (sc->old_bulletin.valid_bitmap & (1 << CHANNEL_DOWN)) {
322 		PMD_DRV_LOG(ERR, "PF indicated channel is down."
323 				"VF device is no longer operational");
324 		dev->data->dev_link.link_status = ETH_LINK_DOWN;
325 	}
326 
327 	return old_link_status == dev->data->dev_link.link_status ? -1 : 0;
328 }
329 
330 static void
331 bnx2x_dev_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats)
332 {
333 	struct bnx2x_softc *sc = dev->data->dev_private;
334 	uint32_t brb_truncate_discard;
335 	uint64_t brb_drops;
336 	uint64_t brb_truncates;
337 
338 	PMD_INIT_FUNC_TRACE();
339 
340 	bnx2x_stats_handle(sc, STATS_EVENT_UPDATE);
341 
342 	memset(stats, 0, sizeof (struct rte_eth_stats));
343 
344 	stats->ipackets =
345 		HILO_U64(sc->eth_stats.total_unicast_packets_received_hi,
346 				sc->eth_stats.total_unicast_packets_received_lo) +
347 		HILO_U64(sc->eth_stats.total_multicast_packets_received_hi,
348 				sc->eth_stats.total_multicast_packets_received_lo) +
349 		HILO_U64(sc->eth_stats.total_broadcast_packets_received_hi,
350 				sc->eth_stats.total_broadcast_packets_received_lo);
351 
352 	stats->opackets =
353 		HILO_U64(sc->eth_stats.total_unicast_packets_transmitted_hi,
354 				sc->eth_stats.total_unicast_packets_transmitted_lo) +
355 		HILO_U64(sc->eth_stats.total_multicast_packets_transmitted_hi,
356 				sc->eth_stats.total_multicast_packets_transmitted_lo) +
357 		HILO_U64(sc->eth_stats.total_broadcast_packets_transmitted_hi,
358 				sc->eth_stats.total_broadcast_packets_transmitted_lo);
359 
360 	stats->ibytes =
361 		HILO_U64(sc->eth_stats.total_bytes_received_hi,
362 				sc->eth_stats.total_bytes_received_lo);
363 
364 	stats->obytes =
365 		HILO_U64(sc->eth_stats.total_bytes_transmitted_hi,
366 				sc->eth_stats.total_bytes_transmitted_lo);
367 
368 	stats->ierrors =
369 		HILO_U64(sc->eth_stats.error_bytes_received_hi,
370 				sc->eth_stats.error_bytes_received_lo);
371 
372 	stats->oerrors = 0;
373 
374 	stats->rx_nombuf =
375 		HILO_U64(sc->eth_stats.no_buff_discard_hi,
376 				sc->eth_stats.no_buff_discard_lo);
377 
378 	brb_drops =
379 		HILO_U64(sc->eth_stats.brb_drop_hi,
380 			 sc->eth_stats.brb_drop_lo);
381 
382 	brb_truncates =
383 		HILO_U64(sc->eth_stats.brb_truncate_hi,
384 			 sc->eth_stats.brb_truncate_lo);
385 
386 	brb_truncate_discard = sc->eth_stats.brb_truncate_discard;
387 
388 	stats->imissed = brb_drops + brb_truncates +
389 			 brb_truncate_discard + stats->rx_nombuf;
390 }
391 
392 static int
393 bnx2x_get_xstats_names(__rte_unused struct rte_eth_dev *dev,
394 		       struct rte_eth_xstat_name *xstats_names,
395 		       __rte_unused unsigned limit)
396 {
397 	unsigned int i, stat_cnt = RTE_DIM(bnx2x_xstats_strings);
398 
399 	if (xstats_names != NULL)
400 		for (i = 0; i < stat_cnt; i++)
401 			snprintf(xstats_names[i].name,
402 				sizeof(xstats_names[i].name),
403 				"%s",
404 				bnx2x_xstats_strings[i].name);
405 
406 	return stat_cnt;
407 }
408 
409 static int
410 bnx2x_dev_xstats_get(struct rte_eth_dev *dev, struct rte_eth_xstat *xstats,
411 		     unsigned int n)
412 {
413 	struct bnx2x_softc *sc = dev->data->dev_private;
414 	unsigned int num = RTE_DIM(bnx2x_xstats_strings);
415 
416 	if (n < num)
417 		return num;
418 
419 	bnx2x_stats_handle(sc, STATS_EVENT_UPDATE);
420 
421 	for (num = 0; num < n; num++) {
422 		if (bnx2x_xstats_strings[num].offset_hi !=
423 		    bnx2x_xstats_strings[num].offset_lo)
424 			xstats[num].value = HILO_U64(
425 					  *(uint32_t *)((char *)&sc->eth_stats +
426 					  bnx2x_xstats_strings[num].offset_hi),
427 					  *(uint32_t *)((char *)&sc->eth_stats +
428 					  bnx2x_xstats_strings[num].offset_lo));
429 		else
430 			xstats[num].value =
431 					  *(uint64_t *)((char *)&sc->eth_stats +
432 					  bnx2x_xstats_strings[num].offset_lo);
433 		xstats[num].id = num;
434 	}
435 
436 	return num;
437 }
438 
439 static void
440 bnx2x_dev_infos_get(struct rte_eth_dev *dev, __rte_unused struct rte_eth_dev_info *dev_info)
441 {
442 	struct bnx2x_softc *sc = dev->data->dev_private;
443 	dev_info->pci_dev = RTE_DEV_TO_PCI(dev->device);
444 	dev_info->max_rx_queues  = sc->max_rx_queues;
445 	dev_info->max_tx_queues  = sc->max_tx_queues;
446 	dev_info->min_rx_bufsize = BNX2X_MIN_RX_BUF_SIZE;
447 	dev_info->max_rx_pktlen  = BNX2X_MAX_RX_PKT_LEN;
448 	dev_info->max_mac_addrs  = BNX2X_MAX_MAC_ADDRS;
449 	dev_info->speed_capa = ETH_LINK_SPEED_10G | ETH_LINK_SPEED_20G;
450 }
451 
452 static void
453 bnx2x_mac_addr_add(struct rte_eth_dev *dev, struct ether_addr *mac_addr,
454 		uint32_t index, uint32_t pool)
455 {
456 	struct bnx2x_softc *sc = dev->data->dev_private;
457 
458 	if (sc->mac_ops.mac_addr_add)
459 		sc->mac_ops.mac_addr_add(dev, mac_addr, index, pool);
460 }
461 
462 static void
463 bnx2x_mac_addr_remove(struct rte_eth_dev *dev, uint32_t index)
464 {
465 	struct bnx2x_softc *sc = dev->data->dev_private;
466 
467 	if (sc->mac_ops.mac_addr_remove)
468 		sc->mac_ops.mac_addr_remove(dev, index);
469 }
470 
471 static const struct eth_dev_ops bnx2x_eth_dev_ops = {
472 	.dev_configure                = bnx2x_dev_configure,
473 	.dev_start                    = bnx2x_dev_start,
474 	.dev_stop                     = bnx2x_dev_stop,
475 	.dev_close                    = bnx2x_dev_close,
476 	.promiscuous_enable           = bnx2x_promisc_enable,
477 	.promiscuous_disable          = bnx2x_promisc_disable,
478 	.allmulticast_enable          = bnx2x_dev_allmulticast_enable,
479 	.allmulticast_disable         = bnx2x_dev_allmulticast_disable,
480 	.link_update                  = bnx2x_dev_link_update,
481 	.stats_get                    = bnx2x_dev_stats_get,
482 	.xstats_get                   = bnx2x_dev_xstats_get,
483 	.xstats_get_names             = bnx2x_get_xstats_names,
484 	.dev_infos_get                = bnx2x_dev_infos_get,
485 	.rx_queue_setup               = bnx2x_dev_rx_queue_setup,
486 	.rx_queue_release             = bnx2x_dev_rx_queue_release,
487 	.tx_queue_setup               = bnx2x_dev_tx_queue_setup,
488 	.tx_queue_release             = bnx2x_dev_tx_queue_release,
489 	.mac_addr_add                 = bnx2x_mac_addr_add,
490 	.mac_addr_remove              = bnx2x_mac_addr_remove,
491 };
492 
493 /*
494  * dev_ops for virtual function
495  */
496 static const struct eth_dev_ops bnx2xvf_eth_dev_ops = {
497 	.dev_configure                = bnx2x_dev_configure,
498 	.dev_start                    = bnx2x_dev_start,
499 	.dev_stop                     = bnx2x_dev_stop,
500 	.dev_close                    = bnx2x_dev_close,
501 	.promiscuous_enable           = bnx2x_promisc_enable,
502 	.promiscuous_disable          = bnx2x_promisc_disable,
503 	.allmulticast_enable          = bnx2x_dev_allmulticast_enable,
504 	.allmulticast_disable         = bnx2x_dev_allmulticast_disable,
505 	.link_update                  = bnx2xvf_dev_link_update,
506 	.stats_get                    = bnx2x_dev_stats_get,
507 	.xstats_get                   = bnx2x_dev_xstats_get,
508 	.xstats_get_names             = bnx2x_get_xstats_names,
509 	.dev_infos_get                = bnx2x_dev_infos_get,
510 	.rx_queue_setup               = bnx2x_dev_rx_queue_setup,
511 	.rx_queue_release             = bnx2x_dev_rx_queue_release,
512 	.tx_queue_setup               = bnx2x_dev_tx_queue_setup,
513 	.tx_queue_release             = bnx2x_dev_tx_queue_release,
514 	.mac_addr_add                 = bnx2x_mac_addr_add,
515 	.mac_addr_remove              = bnx2x_mac_addr_remove,
516 };
517 
518 
519 static int
520 bnx2x_common_dev_init(struct rte_eth_dev *eth_dev, int is_vf)
521 {
522 	int ret = 0;
523 	struct rte_pci_device *pci_dev;
524 	struct bnx2x_softc *sc;
525 
526 	PMD_INIT_FUNC_TRACE();
527 
528 	eth_dev->dev_ops = is_vf ? &bnx2xvf_eth_dev_ops : &bnx2x_eth_dev_ops;
529 	pci_dev = RTE_DEV_TO_PCI(eth_dev->device);
530 
531 	rte_eth_copy_pci_info(eth_dev, pci_dev);
532 
533 	sc = eth_dev->data->dev_private;
534 	sc->pcie_bus    = pci_dev->addr.bus;
535 	sc->pcie_device = pci_dev->addr.devid;
536 
537 	if (is_vf)
538 		sc->flags = BNX2X_IS_VF_FLAG;
539 
540 	sc->devinfo.vendor_id    = pci_dev->id.vendor_id;
541 	sc->devinfo.device_id    = pci_dev->id.device_id;
542 	sc->devinfo.subvendor_id = pci_dev->id.subsystem_vendor_id;
543 	sc->devinfo.subdevice_id = pci_dev->id.subsystem_device_id;
544 
545 	sc->pcie_func = pci_dev->addr.function;
546 	sc->bar[BAR0].base_addr = (void *)pci_dev->mem_resource[0].addr;
547 	if (is_vf)
548 		sc->bar[BAR1].base_addr = (void *)
549 			((uintptr_t)pci_dev->mem_resource[0].addr + PXP_VF_ADDR_DB_START);
550 	else
551 		sc->bar[BAR1].base_addr = pci_dev->mem_resource[2].addr;
552 
553 	assert(sc->bar[BAR0].base_addr);
554 	assert(sc->bar[BAR1].base_addr);
555 
556 	bnx2x_load_firmware(sc);
557 	assert(sc->firmware);
558 
559 	if (eth_dev->data->dev_conf.rx_adv_conf.rss_conf.rss_hf & ETH_RSS_NONFRAG_IPV4_UDP)
560 		sc->udp_rss = 1;
561 
562 	sc->rx_budget = BNX2X_RX_BUDGET;
563 	sc->hc_rx_ticks = BNX2X_RX_TICKS;
564 	sc->hc_tx_ticks = BNX2X_TX_TICKS;
565 
566 	sc->interrupt_mode = INTR_MODE_SINGLE_MSIX;
567 	sc->rx_mode = BNX2X_RX_MODE_NORMAL;
568 
569 	sc->pci_dev = pci_dev;
570 	ret = bnx2x_attach(sc);
571 	if (ret) {
572 		PMD_DRV_LOG(ERR, "bnx2x_attach failed (%d)", ret);
573 		return ret;
574 	}
575 
576 	eth_dev->data->mac_addrs = (struct ether_addr *)sc->link_params.mac_addr;
577 
578 	PMD_DRV_LOG(INFO, "pcie_bus=%d, pcie_device=%d",
579 			sc->pcie_bus, sc->pcie_device);
580 	PMD_DRV_LOG(INFO, "bar0.addr=%p, bar1.addr=%p",
581 			sc->bar[BAR0].base_addr, sc->bar[BAR1].base_addr);
582 	PMD_DRV_LOG(INFO, "port=%d, path=%d, vnic=%d, func=%d",
583 			PORT_ID(sc), PATH_ID(sc), VNIC_ID(sc), FUNC_ID(sc));
584 	PMD_DRV_LOG(INFO, "portID=%d vendorID=0x%x deviceID=0x%x",
585 			eth_dev->data->port_id, pci_dev->id.vendor_id, pci_dev->id.device_id);
586 
587 	if (IS_VF(sc)) {
588 		rte_spinlock_init(&sc->vf2pf_lock);
589 
590 		if (bnx2x_dma_alloc(sc, sizeof(struct bnx2x_vf_mbx_msg),
591 				    &sc->vf2pf_mbox_mapping, "vf2pf_mbox",
592 				    RTE_CACHE_LINE_SIZE) != 0)
593 			return -ENOMEM;
594 
595 		sc->vf2pf_mbox = (struct bnx2x_vf_mbx_msg *)
596 					 sc->vf2pf_mbox_mapping.vaddr;
597 
598 		if (bnx2x_dma_alloc(sc, sizeof(struct bnx2x_vf_bulletin),
599 				    &sc->pf2vf_bulletin_mapping, "vf2pf_bull",
600 				    RTE_CACHE_LINE_SIZE) != 0)
601 			return -ENOMEM;
602 
603 		sc->pf2vf_bulletin = (struct bnx2x_vf_bulletin *)
604 					     sc->pf2vf_bulletin_mapping.vaddr;
605 
606 		ret = bnx2x_vf_get_resources(sc, sc->max_tx_queues,
607 					     sc->max_rx_queues);
608 		if (ret)
609 			return ret;
610 	}
611 
612 	return 0;
613 }
614 
615 static int
616 eth_bnx2x_dev_init(struct rte_eth_dev *eth_dev)
617 {
618 	PMD_INIT_FUNC_TRACE();
619 	return bnx2x_common_dev_init(eth_dev, 0);
620 }
621 
622 static int
623 eth_bnx2xvf_dev_init(struct rte_eth_dev *eth_dev)
624 {
625 	PMD_INIT_FUNC_TRACE();
626 	return bnx2x_common_dev_init(eth_dev, 1);
627 }
628 
629 static struct eth_driver rte_bnx2x_pmd = {
630 	.pci_drv = {
631 		.id_table = pci_id_bnx2x_map,
632 		.drv_flags = RTE_PCI_DRV_NEED_MAPPING | RTE_PCI_DRV_INTR_LSC,
633 		.probe = rte_eth_dev_pci_probe,
634 		.remove = rte_eth_dev_pci_remove,
635 	},
636 	.eth_dev_init = eth_bnx2x_dev_init,
637 	.dev_private_size = sizeof(struct bnx2x_softc),
638 };
639 
640 /*
641  * virtual function driver struct
642  */
643 static struct eth_driver rte_bnx2xvf_pmd = {
644 	.pci_drv = {
645 		.id_table = pci_id_bnx2xvf_map,
646 		.drv_flags = RTE_PCI_DRV_NEED_MAPPING,
647 		.probe = rte_eth_dev_pci_probe,
648 		.remove = rte_eth_dev_pci_remove,
649 	},
650 	.eth_dev_init = eth_bnx2xvf_dev_init,
651 	.dev_private_size = sizeof(struct bnx2x_softc),
652 };
653 
654 RTE_PMD_REGISTER_PCI(net_bnx2x, rte_bnx2x_pmd.pci_drv);
655 RTE_PMD_REGISTER_PCI_TABLE(net_bnx2x, pci_id_bnx2x_map);
656 RTE_PMD_REGISTER_KMOD_DEP(net_bnx2x, "* igb_uio | uio_pci_generic | vfio");
657 RTE_PMD_REGISTER_PCI(net_bnx2xvf, rte_bnx2xvf_pmd.pci_drv);
658 RTE_PMD_REGISTER_PCI_TABLE(net_bnx2xvf, pci_id_bnx2xvf_map);
659 RTE_PMD_REGISTER_KMOD_DEP(net_bnx2xvf, "* igb_uio | vfio");
660