1 /* SPDX-License-Identifier: BSD-3-Clause 2 * Copyright(c) 2010-2016 Intel Corporation 3 */ 4 5 #include <stdio.h> 6 #include <stdlib.h> 7 #include <string.h> 8 #include <stdint.h> 9 #include <inttypes.h> 10 #include <sys/types.h> 11 #include <sys/queue.h> 12 #include <netinet/in.h> 13 #include <setjmp.h> 14 #include <stdarg.h> 15 #include <ctype.h> 16 #include <errno.h> 17 #include <getopt.h> 18 #include <signal.h> 19 20 #include <rte_common.h> 21 #include <rte_log.h> 22 #include <rte_malloc.h> 23 #include <rte_memory.h> 24 #include <rte_memcpy.h> 25 #include <rte_eal.h> 26 #include <rte_launch.h> 27 #include <rte_atomic.h> 28 #include <rte_cycles.h> 29 #include <rte_prefetch.h> 30 #include <rte_lcore.h> 31 #include <rte_per_lcore.h> 32 #include <rte_branch_prediction.h> 33 #include <rte_interrupts.h> 34 #include <rte_random.h> 35 #include <rte_debug.h> 36 #include <rte_ether.h> 37 #include <rte_ethdev.h> 38 #include <rte_mempool.h> 39 #include <rte_mbuf.h> 40 #include <rte_timer.h> 41 #include <rte_keepalive.h> 42 43 #include "shm.h" 44 45 #define RTE_LOGTYPE_L2FWD RTE_LOGTYPE_USER1 46 47 #define NB_MBUF 8192 48 49 #define MAX_PKT_BURST 32 50 #define BURST_TX_DRAIN_US 100 /* TX drain every ~100us */ 51 52 /* 53 * Configurable number of RX/TX ring descriptors 54 */ 55 #define RTE_TEST_RX_DESC_DEFAULT 128 56 #define RTE_TEST_TX_DESC_DEFAULT 512 57 static uint16_t nb_rxd = RTE_TEST_RX_DESC_DEFAULT; 58 static uint16_t nb_txd = RTE_TEST_TX_DESC_DEFAULT; 59 60 /* ethernet addresses of ports */ 61 static struct ether_addr l2fwd_ports_eth_addr[RTE_MAX_ETHPORTS]; 62 63 /* mask of enabled ports */ 64 static uint32_t l2fwd_enabled_port_mask; 65 66 /* list of enabled ports */ 67 static uint32_t l2fwd_dst_ports[RTE_MAX_ETHPORTS]; 68 69 static unsigned int l2fwd_rx_queue_per_lcore = 1; 70 71 #define MAX_RX_QUEUE_PER_LCORE 16 72 #define MAX_TX_QUEUE_PER_PORT 16 73 struct lcore_queue_conf { 74 unsigned n_rx_port; 75 unsigned rx_port_list[MAX_RX_QUEUE_PER_LCORE]; 76 } __rte_cache_aligned; 77 struct lcore_queue_conf lcore_queue_conf[RTE_MAX_LCORE]; 78 79 struct rte_eth_dev_tx_buffer *tx_buffer[RTE_MAX_ETHPORTS]; 80 81 static const struct rte_eth_conf port_conf = { 82 .rxmode = { 83 .split_hdr_size = 0, 84 .header_split = 0, /**< Header Split disabled */ 85 .hw_ip_checksum = 0, /**< IP checksum offload disabled */ 86 .hw_vlan_filter = 0, /**< VLAN filtering disabled */ 87 .jumbo_frame = 0, /**< Jumbo Frame Support disabled */ 88 .hw_strip_crc = 1, /**< CRC stripped by hardware */ 89 }, 90 .txmode = { 91 .mq_mode = ETH_MQ_TX_NONE, 92 }, 93 }; 94 95 struct rte_mempool *l2fwd_pktmbuf_pool = NULL; 96 97 /* Per-port statistics struct */ 98 struct l2fwd_port_statistics { 99 uint64_t tx; 100 uint64_t rx; 101 uint64_t dropped; 102 } __rte_cache_aligned; 103 struct l2fwd_port_statistics port_statistics[RTE_MAX_ETHPORTS]; 104 105 /* A tsc-based timer responsible for triggering statistics printout */ 106 #define TIMER_MILLISECOND 1 107 #define MAX_TIMER_PERIOD 86400 /* 1 day max */ 108 static int64_t timer_period = 10 * TIMER_MILLISECOND * 1000; /* 10 seconds */ 109 static int64_t check_period = 5; /* default check cycle is 5ms */ 110 111 /* Keepalive structure */ 112 struct rte_keepalive *rte_global_keepalive_info; 113 114 /* Termination signalling */ 115 static int terminate_signal_received; 116 117 /* Termination signal handler */ 118 static void handle_sigterm(__rte_unused int value) 119 { 120 terminate_signal_received = 1; 121 } 122 123 /* Print out statistics on packets dropped */ 124 static void 125 print_stats(__attribute__((unused)) struct rte_timer *ptr_timer, 126 __attribute__((unused)) void *ptr_data) 127 { 128 uint64_t total_packets_dropped, total_packets_tx, total_packets_rx; 129 uint16_t portid; 130 131 total_packets_dropped = 0; 132 total_packets_tx = 0; 133 total_packets_rx = 0; 134 135 const char clr[] = { 27, '[', '2', 'J', '\0' }; 136 const char topLeft[] = { 27, '[', '1', ';', '1', 'H', '\0' }; 137 138 /* Clear screen and move to top left */ 139 printf("%s%s", clr, topLeft); 140 141 printf("\nPort statistics ===================================="); 142 143 for (portid = 0; portid < RTE_MAX_ETHPORTS; portid++) { 144 /* skip disabled ports */ 145 if ((l2fwd_enabled_port_mask & (1 << portid)) == 0) 146 continue; 147 printf("\nStatistics for port %u ------------------------------" 148 "\nPackets sent: %24"PRIu64 149 "\nPackets received: %20"PRIu64 150 "\nPackets dropped: %21"PRIu64, 151 portid, 152 port_statistics[portid].tx, 153 port_statistics[portid].rx, 154 port_statistics[portid].dropped); 155 156 total_packets_dropped += port_statistics[portid].dropped; 157 total_packets_tx += port_statistics[portid].tx; 158 total_packets_rx += port_statistics[portid].rx; 159 } 160 printf("\nAggregate statistics ===============================" 161 "\nTotal packets sent: %18"PRIu64 162 "\nTotal packets received: %14"PRIu64 163 "\nTotal packets dropped: %15"PRIu64, 164 total_packets_tx, 165 total_packets_rx, 166 total_packets_dropped); 167 printf("\n====================================================\n"); 168 } 169 170 static void 171 l2fwd_simple_forward(struct rte_mbuf *m, unsigned portid) 172 { 173 struct ether_hdr *eth; 174 void *tmp; 175 int sent; 176 unsigned dst_port; 177 struct rte_eth_dev_tx_buffer *buffer; 178 179 dst_port = l2fwd_dst_ports[portid]; 180 eth = rte_pktmbuf_mtod(m, struct ether_hdr *); 181 182 /* 02:00:00:00:00:xx */ 183 tmp = ð->d_addr.addr_bytes[0]; 184 *((uint64_t *)tmp) = 0x000000000002 + ((uint64_t)dst_port << 40); 185 186 /* src addr */ 187 ether_addr_copy(&l2fwd_ports_eth_addr[dst_port], ð->s_addr); 188 189 buffer = tx_buffer[dst_port]; 190 sent = rte_eth_tx_buffer(dst_port, 0, buffer, m); 191 if (sent) 192 port_statistics[dst_port].tx += sent; 193 } 194 195 /* main processing loop */ 196 static void 197 l2fwd_main_loop(void) 198 { 199 struct rte_mbuf *pkts_burst[MAX_PKT_BURST]; 200 struct rte_mbuf *m; 201 int sent; 202 unsigned lcore_id; 203 uint64_t prev_tsc, diff_tsc, cur_tsc; 204 unsigned i, j, portid, nb_rx; 205 struct lcore_queue_conf *qconf; 206 const uint64_t drain_tsc = (rte_get_tsc_hz() + US_PER_S - 1) 207 / US_PER_S * BURST_TX_DRAIN_US; 208 struct rte_eth_dev_tx_buffer *buffer; 209 210 prev_tsc = 0; 211 212 lcore_id = rte_lcore_id(); 213 qconf = &lcore_queue_conf[lcore_id]; 214 215 if (qconf->n_rx_port == 0) { 216 RTE_LOG(INFO, L2FWD, "lcore %u has nothing to do\n", lcore_id); 217 return; 218 } 219 220 RTE_LOG(INFO, L2FWD, "entering main loop on lcore %u\n", lcore_id); 221 222 for (i = 0; i < qconf->n_rx_port; i++) { 223 224 portid = qconf->rx_port_list[i]; 225 RTE_LOG(INFO, L2FWD, " -- lcoreid=%u portid=%u\n", lcore_id, 226 portid); 227 } 228 229 uint64_t tsc_initial = rte_rdtsc(); 230 uint64_t tsc_lifetime = (rand()&0x07) * rte_get_tsc_hz(); 231 232 while (!terminate_signal_received) { 233 /* Keepalive heartbeat */ 234 rte_keepalive_mark_alive(rte_global_keepalive_info); 235 236 cur_tsc = rte_rdtsc(); 237 238 /* 239 * Die randomly within 7 secs for demo purposes if 240 * keepalive enabled 241 */ 242 if (check_period > 0 && cur_tsc - tsc_initial > tsc_lifetime) 243 break; 244 245 /* 246 * TX burst queue drain 247 */ 248 diff_tsc = cur_tsc - prev_tsc; 249 if (unlikely(diff_tsc > drain_tsc)) { 250 251 for (i = 0; i < qconf->n_rx_port; i++) { 252 253 portid = l2fwd_dst_ports[qconf->rx_port_list[i]]; 254 buffer = tx_buffer[portid]; 255 256 sent = rte_eth_tx_buffer_flush(portid, 0, buffer); 257 if (sent) 258 port_statistics[portid].tx += sent; 259 260 } 261 262 prev_tsc = cur_tsc; 263 } 264 265 /* 266 * Read packet from RX queues 267 */ 268 for (i = 0; i < qconf->n_rx_port; i++) { 269 270 portid = qconf->rx_port_list[i]; 271 nb_rx = rte_eth_rx_burst(portid, 0, 272 pkts_burst, MAX_PKT_BURST); 273 274 port_statistics[portid].rx += nb_rx; 275 276 for (j = 0; j < nb_rx; j++) { 277 m = pkts_burst[j]; 278 rte_prefetch0(rte_pktmbuf_mtod(m, void *)); 279 l2fwd_simple_forward(m, portid); 280 } 281 } 282 } 283 } 284 285 static int 286 l2fwd_launch_one_lcore(__attribute__((unused)) void *dummy) 287 { 288 l2fwd_main_loop(); 289 return 0; 290 } 291 292 /* display usage */ 293 static void 294 l2fwd_usage(const char *prgname) 295 { 296 printf("%s [EAL options] -- -p PORTMASK [-q NQ]\n" 297 " -p PORTMASK: hexadecimal bitmask of ports to configure\n" 298 " -q NQ: number of queue (=ports) per lcore (default is 1)\n" 299 " -K PERIOD: Keepalive check period (5 default; 86400 max)\n" 300 " -T PERIOD: statistics will be refreshed each PERIOD seconds (0 to disable, 10 default, 86400 maximum)\n", 301 prgname); 302 } 303 304 static int 305 l2fwd_parse_portmask(const char *portmask) 306 { 307 char *end = NULL; 308 unsigned long pm; 309 310 /* parse hexadecimal string */ 311 pm = strtoul(portmask, &end, 16); 312 if ((portmask[0] == '\0') || (end == NULL) || (*end != '\0')) 313 return -1; 314 315 if (pm == 0) 316 return -1; 317 318 return pm; 319 } 320 321 static unsigned int 322 l2fwd_parse_nqueue(const char *q_arg) 323 { 324 char *end = NULL; 325 unsigned long n; 326 327 /* parse hexadecimal string */ 328 n = strtoul(q_arg, &end, 10); 329 if ((q_arg[0] == '\0') || (end == NULL) || (*end != '\0')) 330 return 0; 331 if (n == 0) 332 return 0; 333 if (n >= MAX_RX_QUEUE_PER_LCORE) 334 return 0; 335 336 return n; 337 } 338 339 static int 340 l2fwd_parse_timer_period(const char *q_arg) 341 { 342 char *end = NULL; 343 int n; 344 345 /* parse number string */ 346 n = strtol(q_arg, &end, 10); 347 if ((q_arg[0] == '\0') || (end == NULL) || (*end != '\0')) 348 return -1; 349 if (n >= MAX_TIMER_PERIOD) 350 return -1; 351 352 return n; 353 } 354 355 static int 356 l2fwd_parse_check_period(const char *q_arg) 357 { 358 char *end = NULL; 359 int n; 360 361 /* parse number string */ 362 n = strtol(q_arg, &end, 10); 363 if ((q_arg[0] == '\0') || (end == NULL) || (*end != '\0')) 364 return -1; 365 if (n >= MAX_TIMER_PERIOD) 366 return -1; 367 368 return n; 369 } 370 371 /* Parse the argument given in the command line of the application */ 372 static int 373 l2fwd_parse_args(int argc, char **argv) 374 { 375 int opt, ret; 376 char **argvopt; 377 int option_index; 378 char *prgname = argv[0]; 379 static struct option lgopts[] = { 380 {NULL, 0, 0, 0} 381 }; 382 383 argvopt = argv; 384 385 while ((opt = getopt_long(argc, argvopt, "p:q:T:K:", 386 lgopts, &option_index)) != EOF) { 387 388 switch (opt) { 389 /* portmask */ 390 case 'p': 391 l2fwd_enabled_port_mask = l2fwd_parse_portmask(optarg); 392 if (l2fwd_enabled_port_mask == 0) { 393 printf("invalid portmask\n"); 394 l2fwd_usage(prgname); 395 return -1; 396 } 397 break; 398 399 /* nqueue */ 400 case 'q': 401 l2fwd_rx_queue_per_lcore = l2fwd_parse_nqueue(optarg); 402 if (l2fwd_rx_queue_per_lcore == 0) { 403 printf("invalid queue number\n"); 404 l2fwd_usage(prgname); 405 return -1; 406 } 407 break; 408 409 /* timer period */ 410 case 'T': 411 timer_period = l2fwd_parse_timer_period(optarg) 412 * (int64_t)(1000 * TIMER_MILLISECOND); 413 if (timer_period < 0) { 414 printf("invalid timer period\n"); 415 l2fwd_usage(prgname); 416 return -1; 417 } 418 break; 419 420 /* Check period */ 421 case 'K': 422 check_period = l2fwd_parse_check_period(optarg); 423 if (check_period < 0) { 424 printf("invalid check period\n"); 425 l2fwd_usage(prgname); 426 return -1; 427 } 428 break; 429 430 /* long options */ 431 case 0: 432 l2fwd_usage(prgname); 433 return -1; 434 435 default: 436 l2fwd_usage(prgname); 437 return -1; 438 } 439 } 440 441 if (optind >= 0) 442 argv[optind-1] = prgname; 443 444 ret = optind-1; 445 optind = 1; /* reset getopt lib */ 446 return ret; 447 } 448 449 /* Check the link status of all ports in up to 9s, and print them finally */ 450 static void 451 check_all_ports_link_status(uint16_t port_num, uint32_t port_mask) 452 { 453 #define CHECK_INTERVAL 100 /* 100ms */ 454 #define MAX_CHECK_TIME 90 /* 9s (90 * 100ms) in total */ 455 uint16_t portid; 456 uint8_t count, all_ports_up, print_flag = 0; 457 struct rte_eth_link link; 458 459 printf("\nChecking link status"); 460 fflush(stdout); 461 for (count = 0; count <= MAX_CHECK_TIME; count++) { 462 all_ports_up = 1; 463 for (portid = 0; portid < port_num; portid++) { 464 if ((port_mask & (1 << portid)) == 0) 465 continue; 466 memset(&link, 0, sizeof(link)); 467 rte_eth_link_get_nowait(portid, &link); 468 /* print link status if flag set */ 469 if (print_flag == 1) { 470 if (link.link_status) 471 printf( 472 "Port%d Link Up. Speed %u Mbps - %s\n", 473 portid, link.link_speed, 474 (link.link_duplex == ETH_LINK_FULL_DUPLEX) ? 475 ("full-duplex") : ("half-duplex\n")); 476 else 477 printf("Port %d Link Down\n", portid); 478 continue; 479 } 480 /* clear all_ports_up flag if any link down */ 481 if (link.link_status == ETH_LINK_DOWN) { 482 all_ports_up = 0; 483 break; 484 } 485 } 486 /* after finally printing all link status, get out */ 487 if (print_flag == 1) 488 break; 489 490 if (all_ports_up == 0) { 491 printf("."); 492 fflush(stdout); 493 rte_delay_ms(CHECK_INTERVAL); 494 } 495 496 /* set the print_flag if all ports up or timeout */ 497 if (all_ports_up == 1 || count == (MAX_CHECK_TIME - 1)) { 498 print_flag = 1; 499 printf("done\n"); 500 } 501 } 502 } 503 504 static void 505 dead_core(__rte_unused void *ptr_data, const int id_core) 506 { 507 if (terminate_signal_received) 508 return; 509 printf("Dead core %i - restarting..\n", id_core); 510 if (rte_eal_get_lcore_state(id_core) == FINISHED) { 511 rte_eal_wait_lcore(id_core); 512 rte_eal_remote_launch(l2fwd_launch_one_lcore, NULL, id_core); 513 } else { 514 printf("..false positive!\n"); 515 } 516 } 517 518 static void 519 relay_core_state(void *ptr_data, const int id_core, 520 const enum rte_keepalive_state core_state, uint64_t last_alive) 521 { 522 rte_keepalive_relayed_state((struct rte_keepalive_shm *)ptr_data, 523 id_core, core_state, last_alive); 524 } 525 526 int 527 main(int argc, char **argv) 528 { 529 struct lcore_queue_conf *qconf; 530 struct rte_eth_dev_info dev_info; 531 int ret; 532 uint16_t nb_ports; 533 uint16_t nb_ports_available; 534 uint16_t portid, last_port; 535 unsigned lcore_id, rx_lcore_id; 536 unsigned nb_ports_in_mask = 0; 537 struct sigaction signal_handler; 538 struct rte_keepalive_shm *ka_shm; 539 540 memset(&signal_handler, 0, sizeof(signal_handler)); 541 terminate_signal_received = 0; 542 signal_handler.sa_handler = &handle_sigterm; 543 if (sigaction(SIGINT, &signal_handler, NULL) == -1 || 544 sigaction(SIGTERM, &signal_handler, NULL) == -1) 545 rte_exit(EXIT_FAILURE, "SIGNAL\n"); 546 547 548 /* init EAL */ 549 ret = rte_eal_init(argc, argv); 550 if (ret < 0) 551 rte_exit(EXIT_FAILURE, "Invalid EAL arguments\n"); 552 argc -= ret; 553 argv += ret; 554 555 l2fwd_enabled_port_mask = 0; 556 557 /* parse application arguments (after the EAL ones) */ 558 ret = l2fwd_parse_args(argc, argv); 559 if (ret < 0) 560 rte_exit(EXIT_FAILURE, "Invalid L2FWD arguments\n"); 561 562 /* create the mbuf pool */ 563 l2fwd_pktmbuf_pool = rte_pktmbuf_pool_create("mbuf_pool", NB_MBUF, 32, 564 0, RTE_MBUF_DEFAULT_BUF_SIZE, rte_socket_id()); 565 if (l2fwd_pktmbuf_pool == NULL) 566 rte_exit(EXIT_FAILURE, "Cannot init mbuf pool\n"); 567 568 nb_ports = rte_eth_dev_count(); 569 if (nb_ports == 0) 570 rte_exit(EXIT_FAILURE, "No Ethernet ports - bye\n"); 571 572 /* reset l2fwd_dst_ports */ 573 for (portid = 0; portid < RTE_MAX_ETHPORTS; portid++) 574 l2fwd_dst_ports[portid] = 0; 575 last_port = 0; 576 577 /* 578 * Each logical core is assigned a dedicated TX queue on each port. 579 */ 580 for (portid = 0; portid < nb_ports; portid++) { 581 /* skip ports that are not enabled */ 582 if ((l2fwd_enabled_port_mask & (1 << portid)) == 0) 583 continue; 584 585 if (nb_ports_in_mask % 2) { 586 l2fwd_dst_ports[portid] = last_port; 587 l2fwd_dst_ports[last_port] = portid; 588 } else 589 last_port = portid; 590 591 nb_ports_in_mask++; 592 593 rte_eth_dev_info_get(portid, &dev_info); 594 } 595 if (nb_ports_in_mask % 2) { 596 printf("Notice: odd number of ports in portmask.\n"); 597 l2fwd_dst_ports[last_port] = last_port; 598 } 599 600 rx_lcore_id = 1; 601 qconf = NULL; 602 603 /* Initialize the port/queue configuration of each logical core */ 604 for (portid = 0; portid < nb_ports; portid++) { 605 /* skip ports that are not enabled */ 606 if ((l2fwd_enabled_port_mask & (1 << portid)) == 0) 607 continue; 608 609 /* get the lcore_id for this port */ 610 while (rte_lcore_is_enabled(rx_lcore_id) == 0 || 611 lcore_queue_conf[rx_lcore_id].n_rx_port == 612 l2fwd_rx_queue_per_lcore) { 613 rx_lcore_id++; 614 if (rx_lcore_id >= RTE_MAX_LCORE) 615 rte_exit(EXIT_FAILURE, "Not enough cores\n"); 616 } 617 618 if (qconf != &lcore_queue_conf[rx_lcore_id]) 619 /* Assigned a new logical core in the loop above. */ 620 qconf = &lcore_queue_conf[rx_lcore_id]; 621 622 qconf->rx_port_list[qconf->n_rx_port] = portid; 623 qconf->n_rx_port++; 624 printf("Lcore %u: RX port %u\n", 625 rx_lcore_id, portid); 626 } 627 628 nb_ports_available = nb_ports; 629 630 /* Initialise each port */ 631 for (portid = 0; portid < nb_ports; portid++) { 632 /* skip ports that are not enabled */ 633 if ((l2fwd_enabled_port_mask & (1 << portid)) == 0) { 634 printf("Skipping disabled port %u\n", portid); 635 nb_ports_available--; 636 continue; 637 } 638 /* init port */ 639 printf("Initializing port %u... ", portid); 640 fflush(stdout); 641 ret = rte_eth_dev_configure(portid, 1, 1, &port_conf); 642 if (ret < 0) 643 rte_exit(EXIT_FAILURE, 644 "Cannot configure device: err=%d, port=%u\n", 645 ret, portid); 646 647 ret = rte_eth_dev_adjust_nb_rx_tx_desc(portid, &nb_rxd, 648 &nb_txd); 649 if (ret < 0) 650 rte_exit(EXIT_FAILURE, 651 "Cannot adjust number of descriptors: err=%d, port=%u\n", 652 ret, portid); 653 654 rte_eth_macaddr_get(portid, &l2fwd_ports_eth_addr[portid]); 655 656 /* init one RX queue */ 657 fflush(stdout); 658 ret = rte_eth_rx_queue_setup(portid, 0, nb_rxd, 659 rte_eth_dev_socket_id(portid), 660 NULL, 661 l2fwd_pktmbuf_pool); 662 if (ret < 0) 663 rte_exit(EXIT_FAILURE, 664 "rte_eth_rx_queue_setup:err=%d, port=%u\n", 665 ret, portid); 666 667 /* init one TX queue on each port */ 668 fflush(stdout); 669 ret = rte_eth_tx_queue_setup(portid, 0, nb_txd, 670 rte_eth_dev_socket_id(portid), 671 NULL); 672 if (ret < 0) 673 rte_exit(EXIT_FAILURE, 674 "rte_eth_tx_queue_setup:err=%d, port=%u\n", 675 ret, portid); 676 677 /* Initialize TX buffers */ 678 tx_buffer[portid] = rte_zmalloc_socket("tx_buffer", 679 RTE_ETH_TX_BUFFER_SIZE(MAX_PKT_BURST), 0, 680 rte_eth_dev_socket_id(portid)); 681 if (tx_buffer[portid] == NULL) 682 rte_exit(EXIT_FAILURE, "Cannot allocate buffer for tx on port %u\n", 683 portid); 684 685 rte_eth_tx_buffer_init(tx_buffer[portid], MAX_PKT_BURST); 686 687 ret = rte_eth_tx_buffer_set_err_callback(tx_buffer[portid], 688 rte_eth_tx_buffer_count_callback, 689 &port_statistics[portid].dropped); 690 if (ret < 0) 691 rte_exit(EXIT_FAILURE, 692 "Cannot set error callback for tx buffer on port %u\n", 693 portid); 694 695 /* Start device */ 696 ret = rte_eth_dev_start(portid); 697 if (ret < 0) 698 rte_exit(EXIT_FAILURE, 699 "rte_eth_dev_start:err=%d, port=%u\n", 700 ret, portid); 701 702 rte_eth_promiscuous_enable(portid); 703 704 printf("Port %u, MAC address: " 705 "%02X:%02X:%02X:%02X:%02X:%02X\n\n", 706 portid, 707 l2fwd_ports_eth_addr[portid].addr_bytes[0], 708 l2fwd_ports_eth_addr[portid].addr_bytes[1], 709 l2fwd_ports_eth_addr[portid].addr_bytes[2], 710 l2fwd_ports_eth_addr[portid].addr_bytes[3], 711 l2fwd_ports_eth_addr[portid].addr_bytes[4], 712 l2fwd_ports_eth_addr[portid].addr_bytes[5]); 713 714 /* initialize port stats */ 715 memset(&port_statistics, 0, sizeof(port_statistics)); 716 } 717 718 if (!nb_ports_available) { 719 rte_exit(EXIT_FAILURE, 720 "All available ports are disabled. Please set portmask.\n"); 721 } 722 723 check_all_ports_link_status(nb_ports, l2fwd_enabled_port_mask); 724 725 struct rte_timer hb_timer, stats_timer; 726 727 rte_timer_subsystem_init(); 728 rte_timer_init(&stats_timer); 729 730 ka_shm = NULL; 731 if (check_period > 0) { 732 ka_shm = rte_keepalive_shm_create(); 733 if (ka_shm == NULL) 734 rte_exit(EXIT_FAILURE, 735 "rte_keepalive_shm_create() failed"); 736 rte_global_keepalive_info = 737 rte_keepalive_create(&dead_core, ka_shm); 738 if (rte_global_keepalive_info == NULL) 739 rte_exit(EXIT_FAILURE, "init_keep_alive() failed"); 740 rte_keepalive_register_relay_callback(rte_global_keepalive_info, 741 relay_core_state, ka_shm); 742 rte_timer_init(&hb_timer); 743 if (rte_timer_reset(&hb_timer, 744 (check_period * rte_get_timer_hz()) / 1000, 745 PERIODICAL, 746 rte_lcore_id(), 747 (void(*)(struct rte_timer*, void*)) 748 &rte_keepalive_dispatch_pings, 749 rte_global_keepalive_info 750 ) != 0 ) 751 rte_exit(EXIT_FAILURE, "Keepalive setup failure.\n"); 752 } 753 if (timer_period > 0) { 754 if (rte_timer_reset(&stats_timer, 755 (timer_period * rte_get_timer_hz()) / 1000, 756 PERIODICAL, 757 rte_lcore_id(), 758 &print_stats, NULL 759 ) != 0 ) 760 rte_exit(EXIT_FAILURE, "Stats setup failure.\n"); 761 } 762 /* launch per-lcore init on every slave lcore */ 763 RTE_LCORE_FOREACH_SLAVE(lcore_id) { 764 struct lcore_queue_conf *qconf = &lcore_queue_conf[lcore_id]; 765 766 if (qconf->n_rx_port == 0) 767 RTE_LOG(INFO, L2FWD, 768 "lcore %u has nothing to do\n", 769 lcore_id 770 ); 771 else { 772 rte_eal_remote_launch( 773 l2fwd_launch_one_lcore, 774 NULL, 775 lcore_id 776 ); 777 rte_keepalive_register_core(rte_global_keepalive_info, 778 lcore_id); 779 } 780 } 781 while (!terminate_signal_received) { 782 rte_timer_manage(); 783 rte_delay_ms(5); 784 } 785 786 RTE_LCORE_FOREACH_SLAVE(lcore_id) { 787 if (rte_eal_wait_lcore(lcore_id) < 0) 788 return -1; 789 } 790 791 if (ka_shm != NULL) 792 rte_keepalive_shm_cleanup(ka_shm); 793 return 0; 794 } 795