1 /* SPDX-License-Identifier: BSD-3-Clause 2 * Copyright (c) 2013-2015 Brocade Communications Systems, Inc. 3 * Copyright (c) 2015-2018 Cavium Inc. 4 * All rights reserved. 5 * www.cavium.com 6 */ 7 8 #include "bnx2x.h" 9 #include "bnx2x_rxtx.h" 10 11 #include <rte_string_fns.h> 12 #include <rte_dev.h> 13 #include <rte_ethdev_pci.h> 14 #include <rte_alarm.h> 15 16 int bnx2x_logtype_init; 17 int bnx2x_logtype_driver; 18 19 /* 20 * The set of PCI devices this driver supports 21 */ 22 #define BROADCOM_PCI_VENDOR_ID 0x14E4 23 static const struct rte_pci_id pci_id_bnx2x_map[] = { 24 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57800) }, 25 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57711) }, 26 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57810) }, 27 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57811) }, 28 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57840_OBS) }, 29 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57840_4_10) }, 30 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57840_2_20) }, 31 #ifdef RTE_LIBRTE_BNX2X_MF_SUPPORT 32 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57810_MF) }, 33 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57811_MF) }, 34 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57840_MF) }, 35 #endif 36 { .vendor_id = 0, } 37 }; 38 39 static const struct rte_pci_id pci_id_bnx2xvf_map[] = { 40 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57800_VF) }, 41 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57810_VF) }, 42 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57811_VF) }, 43 { RTE_PCI_DEVICE(BROADCOM_PCI_VENDOR_ID, CHIP_NUM_57840_VF) }, 44 { .vendor_id = 0, } 45 }; 46 47 struct rte_bnx2x_xstats_name_off { 48 char name[RTE_ETH_XSTATS_NAME_SIZE]; 49 uint32_t offset_hi; 50 uint32_t offset_lo; 51 }; 52 53 static const struct rte_bnx2x_xstats_name_off bnx2x_xstats_strings[] = { 54 {"rx_buffer_drops", 55 offsetof(struct bnx2x_eth_stats, brb_drop_hi), 56 offsetof(struct bnx2x_eth_stats, brb_drop_lo)}, 57 {"rx_buffer_truncates", 58 offsetof(struct bnx2x_eth_stats, brb_truncate_hi), 59 offsetof(struct bnx2x_eth_stats, brb_truncate_lo)}, 60 {"rx_buffer_truncate_discard", 61 offsetof(struct bnx2x_eth_stats, brb_truncate_discard), 62 offsetof(struct bnx2x_eth_stats, brb_truncate_discard)}, 63 {"mac_filter_discard", 64 offsetof(struct bnx2x_eth_stats, mac_filter_discard), 65 offsetof(struct bnx2x_eth_stats, mac_filter_discard)}, 66 {"no_match_vlan_tag_discard", 67 offsetof(struct bnx2x_eth_stats, mf_tag_discard), 68 offsetof(struct bnx2x_eth_stats, mf_tag_discard)}, 69 {"tx_pause", 70 offsetof(struct bnx2x_eth_stats, pause_frames_sent_hi), 71 offsetof(struct bnx2x_eth_stats, pause_frames_sent_lo)}, 72 {"rx_pause", 73 offsetof(struct bnx2x_eth_stats, pause_frames_received_hi), 74 offsetof(struct bnx2x_eth_stats, pause_frames_received_lo)}, 75 {"tx_priority_flow_control", 76 offsetof(struct bnx2x_eth_stats, pfc_frames_sent_hi), 77 offsetof(struct bnx2x_eth_stats, pfc_frames_sent_lo)}, 78 {"rx_priority_flow_control", 79 offsetof(struct bnx2x_eth_stats, pfc_frames_received_hi), 80 offsetof(struct bnx2x_eth_stats, pfc_frames_received_lo)} 81 }; 82 83 static int 84 bnx2x_link_update(struct rte_eth_dev *dev) 85 { 86 struct bnx2x_softc *sc = dev->data->dev_private; 87 struct rte_eth_link link; 88 89 PMD_INIT_FUNC_TRACE(sc); 90 91 bnx2x_link_status_update(sc); 92 memset(&link, 0, sizeof(link)); 93 mb(); 94 link.link_speed = sc->link_vars.line_speed; 95 switch (sc->link_vars.duplex) { 96 case DUPLEX_FULL: 97 link.link_duplex = ETH_LINK_FULL_DUPLEX; 98 break; 99 case DUPLEX_HALF: 100 link.link_duplex = ETH_LINK_HALF_DUPLEX; 101 break; 102 } 103 link.link_autoneg = !(dev->data->dev_conf.link_speeds & 104 ETH_LINK_SPEED_FIXED); 105 link.link_status = sc->link_vars.link_up; 106 107 return rte_eth_linkstatus_set(dev, &link); 108 } 109 110 static void 111 bnx2x_interrupt_action(struct rte_eth_dev *dev) 112 { 113 struct bnx2x_softc *sc = dev->data->dev_private; 114 uint32_t link_status; 115 116 bnx2x_intr_legacy(sc, 0); 117 118 if (sc->periodic_flags & PERIODIC_GO) 119 bnx2x_periodic_callout(sc); 120 link_status = REG_RD(sc, sc->link_params.shmem_base + 121 offsetof(struct shmem_region, 122 port_mb[sc->link_params.port].link_status)); 123 if ((link_status & LINK_STATUS_LINK_UP) != dev->data->dev_link.link_status) 124 bnx2x_link_update(dev); 125 } 126 127 static void 128 bnx2x_interrupt_handler(void *param) 129 { 130 struct rte_eth_dev *dev = (struct rte_eth_dev *)param; 131 struct bnx2x_softc *sc = dev->data->dev_private; 132 133 PMD_DEBUG_PERIODIC_LOG(INFO, sc, "Interrupt handled"); 134 135 atomic_store_rel_long(&sc->periodic_flags, PERIODIC_STOP); 136 bnx2x_interrupt_action(dev); 137 atomic_store_rel_long(&sc->periodic_flags, PERIODIC_GO); 138 rte_intr_enable(&sc->pci_dev->intr_handle); 139 } 140 141 static void bnx2x_periodic_start(void *param) 142 { 143 struct rte_eth_dev *dev = (struct rte_eth_dev *)param; 144 struct bnx2x_softc *sc = dev->data->dev_private; 145 int ret = 0; 146 147 atomic_store_rel_long(&sc->periodic_flags, PERIODIC_GO); 148 bnx2x_interrupt_action(dev); 149 if (IS_PF(sc)) { 150 ret = rte_eal_alarm_set(BNX2X_SP_TIMER_PERIOD, 151 bnx2x_periodic_start, (void *)dev); 152 if (ret) { 153 PMD_DRV_LOG(ERR, sc, "Unable to start periodic" 154 " timer rc %d", ret); 155 assert(false && "Unable to start periodic timer"); 156 } 157 } 158 } 159 160 void bnx2x_periodic_stop(void *param) 161 { 162 struct rte_eth_dev *dev = (struct rte_eth_dev *)param; 163 struct bnx2x_softc *sc = dev->data->dev_private; 164 165 atomic_store_rel_long(&sc->periodic_flags, PERIODIC_STOP); 166 167 rte_eal_alarm_cancel(bnx2x_periodic_start, (void *)dev); 168 } 169 170 /* 171 * Devops - helper functions can be called from user application 172 */ 173 174 static int 175 bnx2x_dev_configure(struct rte_eth_dev *dev) 176 { 177 struct bnx2x_softc *sc = dev->data->dev_private; 178 struct rte_eth_rxmode *rxmode = &dev->data->dev_conf.rxmode; 179 180 int mp_ncpus = sysconf(_SC_NPROCESSORS_CONF); 181 182 PMD_INIT_FUNC_TRACE(sc); 183 184 if (rxmode->offloads & DEV_RX_OFFLOAD_JUMBO_FRAME) 185 sc->mtu = dev->data->dev_conf.rxmode.max_rx_pkt_len; 186 187 if (dev->data->nb_tx_queues > dev->data->nb_rx_queues) { 188 PMD_DRV_LOG(ERR, sc, "The number of TX queues is greater than number of RX queues"); 189 return -EINVAL; 190 } 191 192 sc->num_queues = MAX(dev->data->nb_rx_queues, dev->data->nb_tx_queues); 193 if (sc->num_queues > mp_ncpus) { 194 PMD_DRV_LOG(ERR, sc, "The number of queues is more than number of CPUs"); 195 return -EINVAL; 196 } 197 198 PMD_DRV_LOG(DEBUG, sc, "num_queues=%d, mtu=%d", 199 sc->num_queues, sc->mtu); 200 201 /* allocate ilt */ 202 if (bnx2x_alloc_ilt_mem(sc) != 0) { 203 PMD_DRV_LOG(ERR, sc, "bnx2x_alloc_ilt_mem was failed"); 204 return -ENXIO; 205 } 206 207 /* allocate the host hardware/software hsi structures */ 208 if (bnx2x_alloc_hsi_mem(sc) != 0) { 209 PMD_DRV_LOG(ERR, sc, "bnx2x_alloc_hsi_mem was failed"); 210 bnx2x_free_ilt_mem(sc); 211 return -ENXIO; 212 } 213 214 return 0; 215 } 216 217 static int 218 bnx2x_dev_start(struct rte_eth_dev *dev) 219 { 220 struct bnx2x_softc *sc = dev->data->dev_private; 221 int ret = 0; 222 223 PMD_INIT_FUNC_TRACE(sc); 224 225 /* start the periodic callout */ 226 if (sc->periodic_flags & PERIODIC_STOP) 227 bnx2x_periodic_start(dev); 228 229 ret = bnx2x_init(sc); 230 if (ret) { 231 PMD_DRV_LOG(DEBUG, sc, "bnx2x_init failed (%d)", ret); 232 return -1; 233 } 234 235 if (IS_PF(sc)) { 236 rte_intr_callback_register(&sc->pci_dev->intr_handle, 237 bnx2x_interrupt_handler, (void *)dev); 238 239 if (rte_intr_enable(&sc->pci_dev->intr_handle)) 240 PMD_DRV_LOG(ERR, sc, "rte_intr_enable failed"); 241 } 242 243 ret = bnx2x_dev_rx_init(dev); 244 if (ret != 0) { 245 PMD_DRV_LOG(DEBUG, sc, "bnx2x_dev_rx_init returned error code"); 246 return -3; 247 } 248 249 bnx2x_print_device_info(sc); 250 251 return ret; 252 } 253 254 static void 255 bnx2x_dev_stop(struct rte_eth_dev *dev) 256 { 257 struct bnx2x_softc *sc = dev->data->dev_private; 258 int ret = 0; 259 260 PMD_INIT_FUNC_TRACE(sc); 261 262 if (IS_PF(sc)) { 263 rte_intr_disable(&sc->pci_dev->intr_handle); 264 rte_intr_callback_unregister(&sc->pci_dev->intr_handle, 265 bnx2x_interrupt_handler, (void *)dev); 266 } 267 268 /* stop the periodic callout */ 269 bnx2x_periodic_stop(dev); 270 271 ret = bnx2x_nic_unload(sc, UNLOAD_NORMAL, FALSE); 272 if (ret) { 273 PMD_DRV_LOG(DEBUG, sc, "bnx2x_nic_unload failed (%d)", ret); 274 return; 275 } 276 277 return; 278 } 279 280 static void 281 bnx2x_dev_close(struct rte_eth_dev *dev) 282 { 283 struct bnx2x_softc *sc = dev->data->dev_private; 284 285 PMD_INIT_FUNC_TRACE(sc); 286 287 if (IS_VF(sc)) 288 bnx2x_vf_close(sc); 289 290 bnx2x_dev_clear_queues(dev); 291 memset(&(dev->data->dev_link), 0 , sizeof(struct rte_eth_link)); 292 293 /* free the host hardware/software hsi structures */ 294 bnx2x_free_hsi_mem(sc); 295 296 /* free ilt */ 297 bnx2x_free_ilt_mem(sc); 298 } 299 300 static void 301 bnx2x_promisc_enable(struct rte_eth_dev *dev) 302 { 303 struct bnx2x_softc *sc = dev->data->dev_private; 304 305 PMD_INIT_FUNC_TRACE(sc); 306 sc->rx_mode = BNX2X_RX_MODE_PROMISC; 307 if (rte_eth_allmulticast_get(dev->data->port_id) == 1) 308 sc->rx_mode = BNX2X_RX_MODE_ALLMULTI_PROMISC; 309 bnx2x_set_rx_mode(sc); 310 } 311 312 static void 313 bnx2x_promisc_disable(struct rte_eth_dev *dev) 314 { 315 struct bnx2x_softc *sc = dev->data->dev_private; 316 317 PMD_INIT_FUNC_TRACE(sc); 318 sc->rx_mode = BNX2X_RX_MODE_NORMAL; 319 if (rte_eth_allmulticast_get(dev->data->port_id) == 1) 320 sc->rx_mode = BNX2X_RX_MODE_ALLMULTI; 321 bnx2x_set_rx_mode(sc); 322 } 323 324 static void 325 bnx2x_dev_allmulticast_enable(struct rte_eth_dev *dev) 326 { 327 struct bnx2x_softc *sc = dev->data->dev_private; 328 329 PMD_INIT_FUNC_TRACE(sc); 330 sc->rx_mode = BNX2X_RX_MODE_ALLMULTI; 331 if (rte_eth_promiscuous_get(dev->data->port_id) == 1) 332 sc->rx_mode = BNX2X_RX_MODE_ALLMULTI_PROMISC; 333 bnx2x_set_rx_mode(sc); 334 } 335 336 static void 337 bnx2x_dev_allmulticast_disable(struct rte_eth_dev *dev) 338 { 339 struct bnx2x_softc *sc = dev->data->dev_private; 340 341 PMD_INIT_FUNC_TRACE(sc); 342 sc->rx_mode = BNX2X_RX_MODE_NORMAL; 343 if (rte_eth_promiscuous_get(dev->data->port_id) == 1) 344 sc->rx_mode = BNX2X_RX_MODE_PROMISC; 345 bnx2x_set_rx_mode(sc); 346 } 347 348 static int 349 bnx2x_dev_link_update(struct rte_eth_dev *dev, __rte_unused int wait_to_complete) 350 { 351 struct bnx2x_softc *sc = dev->data->dev_private; 352 353 PMD_INIT_FUNC_TRACE(sc); 354 355 return bnx2x_link_update(dev); 356 } 357 358 static int 359 bnx2xvf_dev_link_update(struct rte_eth_dev *dev, __rte_unused int wait_to_complete) 360 { 361 struct bnx2x_softc *sc = dev->data->dev_private; 362 int ret = 0; 363 364 ret = bnx2x_link_update(dev); 365 366 bnx2x_check_bull(sc); 367 if (sc->old_bulletin.valid_bitmap & (1 << CHANNEL_DOWN)) { 368 PMD_DRV_LOG(ERR, sc, "PF indicated channel is down." 369 "VF device is no longer operational"); 370 dev->data->dev_link.link_status = ETH_LINK_DOWN; 371 } 372 373 return ret; 374 } 375 376 static int 377 bnx2x_dev_stats_get(struct rte_eth_dev *dev, struct rte_eth_stats *stats) 378 { 379 struct bnx2x_softc *sc = dev->data->dev_private; 380 uint32_t brb_truncate_discard; 381 uint64_t brb_drops; 382 uint64_t brb_truncates; 383 384 PMD_INIT_FUNC_TRACE(sc); 385 386 bnx2x_stats_handle(sc, STATS_EVENT_UPDATE); 387 388 memset(stats, 0, sizeof (struct rte_eth_stats)); 389 390 stats->ipackets = 391 HILO_U64(sc->eth_stats.total_unicast_packets_received_hi, 392 sc->eth_stats.total_unicast_packets_received_lo) + 393 HILO_U64(sc->eth_stats.total_multicast_packets_received_hi, 394 sc->eth_stats.total_multicast_packets_received_lo) + 395 HILO_U64(sc->eth_stats.total_broadcast_packets_received_hi, 396 sc->eth_stats.total_broadcast_packets_received_lo); 397 398 stats->opackets = 399 HILO_U64(sc->eth_stats.total_unicast_packets_transmitted_hi, 400 sc->eth_stats.total_unicast_packets_transmitted_lo) + 401 HILO_U64(sc->eth_stats.total_multicast_packets_transmitted_hi, 402 sc->eth_stats.total_multicast_packets_transmitted_lo) + 403 HILO_U64(sc->eth_stats.total_broadcast_packets_transmitted_hi, 404 sc->eth_stats.total_broadcast_packets_transmitted_lo); 405 406 stats->ibytes = 407 HILO_U64(sc->eth_stats.total_bytes_received_hi, 408 sc->eth_stats.total_bytes_received_lo); 409 410 stats->obytes = 411 HILO_U64(sc->eth_stats.total_bytes_transmitted_hi, 412 sc->eth_stats.total_bytes_transmitted_lo); 413 414 stats->ierrors = 415 HILO_U64(sc->eth_stats.error_bytes_received_hi, 416 sc->eth_stats.error_bytes_received_lo); 417 418 stats->oerrors = 0; 419 420 stats->rx_nombuf = 421 HILO_U64(sc->eth_stats.no_buff_discard_hi, 422 sc->eth_stats.no_buff_discard_lo); 423 424 brb_drops = 425 HILO_U64(sc->eth_stats.brb_drop_hi, 426 sc->eth_stats.brb_drop_lo); 427 428 brb_truncates = 429 HILO_U64(sc->eth_stats.brb_truncate_hi, 430 sc->eth_stats.brb_truncate_lo); 431 432 brb_truncate_discard = sc->eth_stats.brb_truncate_discard; 433 434 stats->imissed = brb_drops + brb_truncates + 435 brb_truncate_discard + stats->rx_nombuf; 436 437 return 0; 438 } 439 440 static int 441 bnx2x_get_xstats_names(__rte_unused struct rte_eth_dev *dev, 442 struct rte_eth_xstat_name *xstats_names, 443 __rte_unused unsigned limit) 444 { 445 unsigned int i, stat_cnt = RTE_DIM(bnx2x_xstats_strings); 446 447 if (xstats_names != NULL) 448 for (i = 0; i < stat_cnt; i++) 449 strlcpy(xstats_names[i].name, 450 bnx2x_xstats_strings[i].name, 451 sizeof(xstats_names[i].name)); 452 453 return stat_cnt; 454 } 455 456 static int 457 bnx2x_dev_xstats_get(struct rte_eth_dev *dev, struct rte_eth_xstat *xstats, 458 unsigned int n) 459 { 460 struct bnx2x_softc *sc = dev->data->dev_private; 461 unsigned int num = RTE_DIM(bnx2x_xstats_strings); 462 463 if (n < num) 464 return num; 465 466 bnx2x_stats_handle(sc, STATS_EVENT_UPDATE); 467 468 for (num = 0; num < n; num++) { 469 if (bnx2x_xstats_strings[num].offset_hi != 470 bnx2x_xstats_strings[num].offset_lo) 471 xstats[num].value = HILO_U64( 472 *(uint32_t *)((char *)&sc->eth_stats + 473 bnx2x_xstats_strings[num].offset_hi), 474 *(uint32_t *)((char *)&sc->eth_stats + 475 bnx2x_xstats_strings[num].offset_lo)); 476 else 477 xstats[num].value = 478 *(uint64_t *)((char *)&sc->eth_stats + 479 bnx2x_xstats_strings[num].offset_lo); 480 xstats[num].id = num; 481 } 482 483 return num; 484 } 485 486 static void 487 bnx2x_dev_infos_get(struct rte_eth_dev *dev, struct rte_eth_dev_info *dev_info) 488 { 489 struct bnx2x_softc *sc = dev->data->dev_private; 490 dev_info->max_rx_queues = sc->max_rx_queues; 491 dev_info->max_tx_queues = sc->max_tx_queues; 492 dev_info->min_rx_bufsize = BNX2X_MIN_RX_BUF_SIZE; 493 dev_info->max_rx_pktlen = BNX2X_MAX_RX_PKT_LEN; 494 dev_info->max_mac_addrs = BNX2X_MAX_MAC_ADDRS; 495 dev_info->speed_capa = ETH_LINK_SPEED_10G | ETH_LINK_SPEED_20G; 496 dev_info->rx_offload_capa = DEV_RX_OFFLOAD_JUMBO_FRAME; 497 } 498 499 static int 500 bnx2x_mac_addr_add(struct rte_eth_dev *dev, struct ether_addr *mac_addr, 501 uint32_t index, uint32_t pool) 502 { 503 struct bnx2x_softc *sc = dev->data->dev_private; 504 505 if (sc->mac_ops.mac_addr_add) { 506 sc->mac_ops.mac_addr_add(dev, mac_addr, index, pool); 507 return 0; 508 } 509 return -ENOTSUP; 510 } 511 512 static void 513 bnx2x_mac_addr_remove(struct rte_eth_dev *dev, uint32_t index) 514 { 515 struct bnx2x_softc *sc = dev->data->dev_private; 516 517 if (sc->mac_ops.mac_addr_remove) 518 sc->mac_ops.mac_addr_remove(dev, index); 519 } 520 521 static const struct eth_dev_ops bnx2x_eth_dev_ops = { 522 .dev_configure = bnx2x_dev_configure, 523 .dev_start = bnx2x_dev_start, 524 .dev_stop = bnx2x_dev_stop, 525 .dev_close = bnx2x_dev_close, 526 .promiscuous_enable = bnx2x_promisc_enable, 527 .promiscuous_disable = bnx2x_promisc_disable, 528 .allmulticast_enable = bnx2x_dev_allmulticast_enable, 529 .allmulticast_disable = bnx2x_dev_allmulticast_disable, 530 .link_update = bnx2x_dev_link_update, 531 .stats_get = bnx2x_dev_stats_get, 532 .xstats_get = bnx2x_dev_xstats_get, 533 .xstats_get_names = bnx2x_get_xstats_names, 534 .dev_infos_get = bnx2x_dev_infos_get, 535 .rx_queue_setup = bnx2x_dev_rx_queue_setup, 536 .rx_queue_release = bnx2x_dev_rx_queue_release, 537 .tx_queue_setup = bnx2x_dev_tx_queue_setup, 538 .tx_queue_release = bnx2x_dev_tx_queue_release, 539 .mac_addr_add = bnx2x_mac_addr_add, 540 .mac_addr_remove = bnx2x_mac_addr_remove, 541 }; 542 543 /* 544 * dev_ops for virtual function 545 */ 546 static const struct eth_dev_ops bnx2xvf_eth_dev_ops = { 547 .dev_configure = bnx2x_dev_configure, 548 .dev_start = bnx2x_dev_start, 549 .dev_stop = bnx2x_dev_stop, 550 .dev_close = bnx2x_dev_close, 551 .promiscuous_enable = bnx2x_promisc_enable, 552 .promiscuous_disable = bnx2x_promisc_disable, 553 .allmulticast_enable = bnx2x_dev_allmulticast_enable, 554 .allmulticast_disable = bnx2x_dev_allmulticast_disable, 555 .link_update = bnx2xvf_dev_link_update, 556 .stats_get = bnx2x_dev_stats_get, 557 .xstats_get = bnx2x_dev_xstats_get, 558 .xstats_get_names = bnx2x_get_xstats_names, 559 .dev_infos_get = bnx2x_dev_infos_get, 560 .rx_queue_setup = bnx2x_dev_rx_queue_setup, 561 .rx_queue_release = bnx2x_dev_rx_queue_release, 562 .tx_queue_setup = bnx2x_dev_tx_queue_setup, 563 .tx_queue_release = bnx2x_dev_tx_queue_release, 564 .mac_addr_add = bnx2x_mac_addr_add, 565 .mac_addr_remove = bnx2x_mac_addr_remove, 566 }; 567 568 569 static int 570 bnx2x_common_dev_init(struct rte_eth_dev *eth_dev, int is_vf) 571 { 572 int ret = 0; 573 struct rte_pci_device *pci_dev; 574 struct rte_pci_addr pci_addr; 575 struct bnx2x_softc *sc; 576 static bool adapter_info = true; 577 578 /* Extract key data structures */ 579 sc = eth_dev->data->dev_private; 580 pci_dev = RTE_DEV_TO_PCI(eth_dev->device); 581 pci_addr = pci_dev->addr; 582 583 snprintf(sc->devinfo.name, NAME_SIZE, PCI_SHORT_PRI_FMT ":dpdk-port-%u", 584 pci_addr.bus, pci_addr.devid, pci_addr.function, 585 eth_dev->data->port_id); 586 587 PMD_INIT_FUNC_TRACE(sc); 588 589 eth_dev->dev_ops = is_vf ? &bnx2xvf_eth_dev_ops : &bnx2x_eth_dev_ops; 590 591 rte_eth_copy_pci_info(eth_dev, pci_dev); 592 593 sc->pcie_bus = pci_dev->addr.bus; 594 sc->pcie_device = pci_dev->addr.devid; 595 596 sc->devinfo.vendor_id = pci_dev->id.vendor_id; 597 sc->devinfo.device_id = pci_dev->id.device_id; 598 sc->devinfo.subvendor_id = pci_dev->id.subsystem_vendor_id; 599 sc->devinfo.subdevice_id = pci_dev->id.subsystem_device_id; 600 601 if (is_vf) 602 sc->flags = BNX2X_IS_VF_FLAG; 603 604 sc->pcie_func = pci_dev->addr.function; 605 sc->bar[BAR0].base_addr = (void *)pci_dev->mem_resource[0].addr; 606 if (is_vf) 607 sc->bar[BAR1].base_addr = (void *) 608 ((uintptr_t)pci_dev->mem_resource[0].addr + PXP_VF_ADDR_DB_START); 609 else 610 sc->bar[BAR1].base_addr = pci_dev->mem_resource[2].addr; 611 612 assert(sc->bar[BAR0].base_addr); 613 assert(sc->bar[BAR1].base_addr); 614 615 bnx2x_load_firmware(sc); 616 assert(sc->firmware); 617 618 if (eth_dev->data->dev_conf.rx_adv_conf.rss_conf.rss_hf & ETH_RSS_NONFRAG_IPV4_UDP) 619 sc->udp_rss = 1; 620 621 sc->rx_budget = BNX2X_RX_BUDGET; 622 sc->hc_rx_ticks = BNX2X_RX_TICKS; 623 sc->hc_tx_ticks = BNX2X_TX_TICKS; 624 625 sc->interrupt_mode = INTR_MODE_SINGLE_MSIX; 626 sc->rx_mode = BNX2X_RX_MODE_NORMAL; 627 628 sc->pci_dev = pci_dev; 629 ret = bnx2x_attach(sc); 630 if (ret) { 631 PMD_DRV_LOG(ERR, sc, "bnx2x_attach failed (%d)", ret); 632 return ret; 633 } 634 635 /* Print important adapter info for the user. */ 636 if (adapter_info) { 637 bnx2x_print_adapter_info(sc); 638 adapter_info = false; 639 } 640 641 /* schedule periodic poll for slowpath link events */ 642 if (IS_PF(sc)) { 643 PMD_DRV_LOG(DEBUG, sc, "Scheduling periodic poll for slowpath link events"); 644 ret = rte_eal_alarm_set(BNX2X_SP_TIMER_PERIOD, 645 bnx2x_periodic_start, (void *)eth_dev); 646 if (ret) { 647 PMD_DRV_LOG(ERR, sc, "Unable to start periodic" 648 " timer rc %d", ret); 649 return -EINVAL; 650 } 651 } 652 653 eth_dev->data->mac_addrs = (struct ether_addr *)sc->link_params.mac_addr; 654 655 if (IS_VF(sc)) { 656 rte_spinlock_init(&sc->vf2pf_lock); 657 658 ret = bnx2x_dma_alloc(sc, sizeof(struct bnx2x_vf_mbx_msg), 659 &sc->vf2pf_mbox_mapping, "vf2pf_mbox", 660 RTE_CACHE_LINE_SIZE); 661 if (ret) 662 goto out; 663 664 sc->vf2pf_mbox = (struct bnx2x_vf_mbx_msg *) 665 sc->vf2pf_mbox_mapping.vaddr; 666 667 ret = bnx2x_dma_alloc(sc, sizeof(struct bnx2x_vf_bulletin), 668 &sc->pf2vf_bulletin_mapping, "vf2pf_bull", 669 RTE_CACHE_LINE_SIZE); 670 if (ret) 671 goto out; 672 673 sc->pf2vf_bulletin = (struct bnx2x_vf_bulletin *) 674 sc->pf2vf_bulletin_mapping.vaddr; 675 676 ret = bnx2x_vf_get_resources(sc, sc->max_tx_queues, 677 sc->max_rx_queues); 678 if (ret) 679 goto out; 680 } 681 682 return 0; 683 684 out: 685 bnx2x_periodic_stop(eth_dev); 686 return ret; 687 } 688 689 static int 690 eth_bnx2x_dev_init(struct rte_eth_dev *eth_dev) 691 { 692 struct bnx2x_softc *sc = eth_dev->data->dev_private; 693 PMD_INIT_FUNC_TRACE(sc); 694 return bnx2x_common_dev_init(eth_dev, 0); 695 } 696 697 static int 698 eth_bnx2xvf_dev_init(struct rte_eth_dev *eth_dev) 699 { 700 struct bnx2x_softc *sc = eth_dev->data->dev_private; 701 PMD_INIT_FUNC_TRACE(sc); 702 return bnx2x_common_dev_init(eth_dev, 1); 703 } 704 705 static struct rte_pci_driver rte_bnx2x_pmd; 706 static struct rte_pci_driver rte_bnx2xvf_pmd; 707 708 static int eth_bnx2x_pci_probe(struct rte_pci_driver *pci_drv, 709 struct rte_pci_device *pci_dev) 710 { 711 if (pci_drv == &rte_bnx2x_pmd) 712 return rte_eth_dev_pci_generic_probe(pci_dev, 713 sizeof(struct bnx2x_softc), eth_bnx2x_dev_init); 714 else if (pci_drv == &rte_bnx2xvf_pmd) 715 return rte_eth_dev_pci_generic_probe(pci_dev, 716 sizeof(struct bnx2x_softc), eth_bnx2xvf_dev_init); 717 else 718 return -EINVAL; 719 } 720 721 static int eth_bnx2x_pci_remove(struct rte_pci_device *pci_dev) 722 { 723 return rte_eth_dev_pci_generic_remove(pci_dev, NULL); 724 } 725 726 static struct rte_pci_driver rte_bnx2x_pmd = { 727 .id_table = pci_id_bnx2x_map, 728 .drv_flags = RTE_PCI_DRV_NEED_MAPPING | RTE_PCI_DRV_INTR_LSC, 729 .probe = eth_bnx2x_pci_probe, 730 .remove = eth_bnx2x_pci_remove, 731 }; 732 733 /* 734 * virtual function driver struct 735 */ 736 static struct rte_pci_driver rte_bnx2xvf_pmd = { 737 .id_table = pci_id_bnx2xvf_map, 738 .drv_flags = RTE_PCI_DRV_NEED_MAPPING, 739 .probe = eth_bnx2x_pci_probe, 740 .remove = eth_bnx2x_pci_remove, 741 }; 742 743 RTE_PMD_REGISTER_PCI(net_bnx2x, rte_bnx2x_pmd); 744 RTE_PMD_REGISTER_PCI_TABLE(net_bnx2x, pci_id_bnx2x_map); 745 RTE_PMD_REGISTER_KMOD_DEP(net_bnx2x, "* igb_uio | uio_pci_generic | vfio-pci"); 746 RTE_PMD_REGISTER_PCI(net_bnx2xvf, rte_bnx2xvf_pmd); 747 RTE_PMD_REGISTER_PCI_TABLE(net_bnx2xvf, pci_id_bnx2xvf_map); 748 RTE_PMD_REGISTER_KMOD_DEP(net_bnx2xvf, "* igb_uio | vfio-pci"); 749 750 RTE_INIT(bnx2x_init_log) 751 { 752 bnx2x_logtype_init = rte_log_register("pmd.net.bnx2x.init"); 753 if (bnx2x_logtype_init >= 0) 754 rte_log_set_level(bnx2x_logtype_init, RTE_LOG_NOTICE); 755 bnx2x_logtype_driver = rte_log_register("pmd.net.bnx2x.driver"); 756 if (bnx2x_logtype_driver >= 0) 757 rte_log_set_level(bnx2x_logtype_driver, RTE_LOG_NOTICE); 758 } 759