1 /*- 2 * BSD LICENSE 3 * 4 * Copyright(c) 2010-2014 Intel Corporation. All rights reserved. 5 * All rights reserved. 6 * 7 * Redistribution and use in source and binary forms, with or without 8 * modification, are permitted provided that the following conditions 9 * are met: 10 * 11 * * Redistributions of source code must retain the above copyright 12 * notice, this list of conditions and the following disclaimer. 13 * * Redistributions in binary form must reproduce the above copyright 14 * notice, this list of conditions and the following disclaimer in 15 * the documentation and/or other materials provided with the 16 * distribution. 17 * * Neither the name of Intel Corporation nor the names of its 18 * contributors may be used to endorse or promote products derived 19 * from this software without specific prior written permission. 20 * 21 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS 22 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT 23 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR 24 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT 25 * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, 26 * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT 27 * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, 28 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 29 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 30 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE 31 * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 32 */ 33 34 #include <stdint.h> 35 #include <memory.h> 36 37 #include <rte_log.h> 38 #include <rte_mbuf.h> 39 #include <rte_debug.h> 40 #include <rte_ethdev.h> 41 #include <rte_mempool.h> 42 #include <rte_sched.h> 43 #include <rte_cycles.h> 44 #include <rte_string_fns.h> 45 #include <rte_cfgfile.h> 46 47 #include "main.h" 48 #include "cfg_file.h" 49 50 uint32_t app_numa_mask = 0; 51 static uint32_t app_inited_port_mask = 0; 52 53 int app_pipe_to_profile[MAX_SCHED_SUBPORTS][MAX_SCHED_PIPES]; 54 55 #define MAX_NAME_LEN 32 56 57 struct ring_conf ring_conf = { 58 .rx_size = APP_RX_DESC_DEFAULT, 59 .ring_size = APP_RING_SIZE, 60 .tx_size = APP_TX_DESC_DEFAULT, 61 }; 62 63 struct burst_conf burst_conf = { 64 .rx_burst = MAX_PKT_RX_BURST, 65 .ring_burst = PKT_ENQUEUE, 66 .qos_dequeue = PKT_DEQUEUE, 67 .tx_burst = MAX_PKT_TX_BURST, 68 }; 69 70 struct ring_thresh rx_thresh = { 71 .pthresh = RX_PTHRESH, 72 .hthresh = RX_HTHRESH, 73 .wthresh = RX_WTHRESH, 74 }; 75 76 struct ring_thresh tx_thresh = { 77 .pthresh = TX_PTHRESH, 78 .hthresh = TX_HTHRESH, 79 .wthresh = TX_WTHRESH, 80 }; 81 82 uint32_t nb_pfc; 83 const char *cfg_profile = NULL; 84 int mp_size = NB_MBUF; 85 struct flow_conf qos_conf[MAX_DATA_STREAMS]; 86 87 static const struct rte_eth_conf port_conf = { 88 .rxmode = { 89 .max_rx_pkt_len = ETHER_MAX_LEN, 90 .split_hdr_size = 0, 91 .header_split = 0, /**< Header Split disabled */ 92 .hw_ip_checksum = 0, /**< IP checksum offload disabled */ 93 .hw_vlan_filter = 0, /**< VLAN filtering disabled */ 94 .jumbo_frame = 0, /**< Jumbo Frame Support disabled */ 95 .hw_strip_crc = 1, /**< CRC stripped by hardware */ 96 }, 97 .txmode = { 98 .mq_mode = ETH_DCB_NONE, 99 }, 100 }; 101 102 static int 103 app_init_port(uint16_t portid, struct rte_mempool *mp) 104 { 105 int ret; 106 struct rte_eth_link link; 107 struct rte_eth_rxconf rx_conf; 108 struct rte_eth_txconf tx_conf; 109 uint16_t rx_size; 110 uint16_t tx_size; 111 112 /* check if port already initialized (multistream configuration) */ 113 if (app_inited_port_mask & (1u << portid)) 114 return 0; 115 116 rx_conf.rx_thresh.pthresh = rx_thresh.pthresh; 117 rx_conf.rx_thresh.hthresh = rx_thresh.hthresh; 118 rx_conf.rx_thresh.wthresh = rx_thresh.wthresh; 119 rx_conf.rx_free_thresh = 32; 120 rx_conf.rx_drop_en = 0; 121 rx_conf.rx_deferred_start = 0; 122 123 tx_conf.tx_thresh.pthresh = tx_thresh.pthresh; 124 tx_conf.tx_thresh.hthresh = tx_thresh.hthresh; 125 tx_conf.tx_thresh.wthresh = tx_thresh.wthresh; 126 tx_conf.tx_free_thresh = 0; 127 tx_conf.tx_rs_thresh = 0; 128 tx_conf.txq_flags = ETH_TXQ_FLAGS_NOMULTSEGS | ETH_TXQ_FLAGS_NOOFFLOADS; 129 tx_conf.tx_deferred_start = 0; 130 131 /* init port */ 132 RTE_LOG(INFO, APP, "Initializing port %"PRIu16"... ", portid); 133 fflush(stdout); 134 ret = rte_eth_dev_configure(portid, 1, 1, &port_conf); 135 if (ret < 0) 136 rte_exit(EXIT_FAILURE, 137 "Cannot configure device: err=%d, port=%u\n", 138 ret, portid); 139 140 rx_size = ring_conf.rx_size; 141 tx_size = ring_conf.tx_size; 142 ret = rte_eth_dev_adjust_nb_rx_tx_desc(portid, &rx_size, &tx_size); 143 if (ret < 0) 144 rte_exit(EXIT_FAILURE, 145 "rte_eth_dev_adjust_nb_rx_tx_desc: err=%d,port=%u\n", 146 ret, portid); 147 ring_conf.rx_size = rx_size; 148 ring_conf.tx_size = tx_size; 149 150 /* init one RX queue */ 151 fflush(stdout); 152 ret = rte_eth_rx_queue_setup(portid, 0, (uint16_t)ring_conf.rx_size, 153 rte_eth_dev_socket_id(portid), &rx_conf, mp); 154 if (ret < 0) 155 rte_exit(EXIT_FAILURE, 156 "rte_eth_tx_queue_setup: err=%d, port=%u\n", 157 ret, portid); 158 159 /* init one TX queue */ 160 fflush(stdout); 161 ret = rte_eth_tx_queue_setup(portid, 0, 162 (uint16_t)ring_conf.tx_size, rte_eth_dev_socket_id(portid), &tx_conf); 163 if (ret < 0) 164 rte_exit(EXIT_FAILURE, 165 "rte_eth_tx_queue_setup: err=%d, port=%u queue=%d\n", 166 ret, portid, 0); 167 168 /* Start device */ 169 ret = rte_eth_dev_start(portid); 170 if (ret < 0) 171 rte_exit(EXIT_FAILURE, 172 "rte_pmd_port_start: err=%d, port=%u\n", 173 ret, portid); 174 175 printf("done: "); 176 177 /* get link status */ 178 rte_eth_link_get(portid, &link); 179 if (link.link_status) { 180 printf(" Link Up - speed %u Mbps - %s\n", 181 (uint32_t) link.link_speed, 182 (link.link_duplex == ETH_LINK_FULL_DUPLEX) ? 183 ("full-duplex") : ("half-duplex\n")); 184 } else { 185 printf(" Link Down\n"); 186 } 187 rte_eth_promiscuous_enable(portid); 188 189 /* mark port as initialized */ 190 app_inited_port_mask |= 1u << portid; 191 192 return 0; 193 } 194 195 static struct rte_sched_subport_params subport_params[MAX_SCHED_SUBPORTS] = { 196 { 197 .tb_rate = 1250000000, 198 .tb_size = 1000000, 199 200 .tc_rate = {1250000000, 1250000000, 1250000000, 1250000000}, 201 .tc_period = 10, 202 }, 203 }; 204 205 static struct rte_sched_pipe_params pipe_profiles[RTE_SCHED_PIPE_PROFILES_PER_PORT] = { 206 { /* Profile #0 */ 207 .tb_rate = 305175, 208 .tb_size = 1000000, 209 210 .tc_rate = {305175, 305175, 305175, 305175}, 211 .tc_period = 40, 212 #ifdef RTE_SCHED_SUBPORT_TC_OV 213 .tc_ov_weight = 1, 214 #endif 215 216 .wrr_weights = {1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1, 1}, 217 }, 218 }; 219 220 struct rte_sched_port_params port_params = { 221 .name = "port_scheduler_0", 222 .socket = 0, /* computed */ 223 .rate = 0, /* computed */ 224 .mtu = 6 + 6 + 4 + 4 + 2 + 1500, 225 .frame_overhead = RTE_SCHED_FRAME_OVERHEAD_DEFAULT, 226 .n_subports_per_port = 1, 227 .n_pipes_per_subport = 4096, 228 .qsize = {64, 64, 64, 64}, 229 .pipe_profiles = pipe_profiles, 230 .n_pipe_profiles = sizeof(pipe_profiles) / sizeof(struct rte_sched_pipe_params), 231 232 #ifdef RTE_SCHED_RED 233 .red_params = { 234 /* Traffic Class 0 Colors Green / Yellow / Red */ 235 [0][0] = {.min_th = 48, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9}, 236 [0][1] = {.min_th = 40, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9}, 237 [0][2] = {.min_th = 32, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9}, 238 239 /* Traffic Class 1 - Colors Green / Yellow / Red */ 240 [1][0] = {.min_th = 48, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9}, 241 [1][1] = {.min_th = 40, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9}, 242 [1][2] = {.min_th = 32, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9}, 243 244 /* Traffic Class 2 - Colors Green / Yellow / Red */ 245 [2][0] = {.min_th = 48, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9}, 246 [2][1] = {.min_th = 40, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9}, 247 [2][2] = {.min_th = 32, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9}, 248 249 /* Traffic Class 3 - Colors Green / Yellow / Red */ 250 [3][0] = {.min_th = 48, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9}, 251 [3][1] = {.min_th = 40, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9}, 252 [3][2] = {.min_th = 32, .max_th = 64, .maxp_inv = 10, .wq_log2 = 9} 253 } 254 #endif /* RTE_SCHED_RED */ 255 }; 256 257 static struct rte_sched_port * 258 app_init_sched_port(uint32_t portid, uint32_t socketid) 259 { 260 static char port_name[32]; /* static as referenced from global port_params*/ 261 struct rte_eth_link link; 262 struct rte_sched_port *port = NULL; 263 uint32_t pipe, subport; 264 int err; 265 266 rte_eth_link_get(portid, &link); 267 268 port_params.socket = socketid; 269 port_params.rate = (uint64_t) link.link_speed * 1000 * 1000 / 8; 270 snprintf(port_name, sizeof(port_name), "port_%d", portid); 271 port_params.name = port_name; 272 273 port = rte_sched_port_config(&port_params); 274 if (port == NULL){ 275 rte_exit(EXIT_FAILURE, "Unable to config sched port\n"); 276 } 277 278 for (subport = 0; subport < port_params.n_subports_per_port; subport ++) { 279 err = rte_sched_subport_config(port, subport, &subport_params[subport]); 280 if (err) { 281 rte_exit(EXIT_FAILURE, "Unable to config sched subport %u, err=%d\n", 282 subport, err); 283 } 284 285 for (pipe = 0; pipe < port_params.n_pipes_per_subport; pipe ++) { 286 if (app_pipe_to_profile[subport][pipe] != -1) { 287 err = rte_sched_pipe_config(port, subport, pipe, 288 app_pipe_to_profile[subport][pipe]); 289 if (err) { 290 rte_exit(EXIT_FAILURE, "Unable to config sched pipe %u " 291 "for profile %d, err=%d\n", pipe, 292 app_pipe_to_profile[subport][pipe], err); 293 } 294 } 295 } 296 } 297 298 return port; 299 } 300 301 static int 302 app_load_cfg_profile(const char *profile) 303 { 304 if (profile == NULL) 305 return 0; 306 struct rte_cfgfile *file = rte_cfgfile_load(profile, 0); 307 if (file == NULL) 308 rte_exit(EXIT_FAILURE, "Cannot load configuration profile %s\n", profile); 309 310 cfg_load_port(file, &port_params); 311 cfg_load_subport(file, subport_params); 312 cfg_load_pipe(file, pipe_profiles); 313 314 rte_cfgfile_close(file); 315 316 return 0; 317 } 318 319 int app_init(void) 320 { 321 uint32_t i; 322 char ring_name[MAX_NAME_LEN]; 323 char pool_name[MAX_NAME_LEN]; 324 325 if (rte_eth_dev_count() == 0) 326 rte_exit(EXIT_FAILURE, "No Ethernet port - bye\n"); 327 328 /* load configuration profile */ 329 if (app_load_cfg_profile(cfg_profile) != 0) 330 rte_exit(EXIT_FAILURE, "Invalid configuration profile\n"); 331 332 /* Initialize each active flow */ 333 for(i = 0; i < nb_pfc; i++) { 334 uint32_t socket = rte_lcore_to_socket_id(qos_conf[i].rx_core); 335 struct rte_ring *ring; 336 337 snprintf(ring_name, MAX_NAME_LEN, "ring-%u-%u", i, qos_conf[i].rx_core); 338 ring = rte_ring_lookup(ring_name); 339 if (ring == NULL) 340 qos_conf[i].rx_ring = rte_ring_create(ring_name, ring_conf.ring_size, 341 socket, RING_F_SP_ENQ | RING_F_SC_DEQ); 342 else 343 qos_conf[i].rx_ring = ring; 344 345 snprintf(ring_name, MAX_NAME_LEN, "ring-%u-%u", i, qos_conf[i].tx_core); 346 ring = rte_ring_lookup(ring_name); 347 if (ring == NULL) 348 qos_conf[i].tx_ring = rte_ring_create(ring_name, ring_conf.ring_size, 349 socket, RING_F_SP_ENQ | RING_F_SC_DEQ); 350 else 351 qos_conf[i].tx_ring = ring; 352 353 354 /* create the mbuf pools for each RX Port */ 355 snprintf(pool_name, MAX_NAME_LEN, "mbuf_pool%u", i); 356 qos_conf[i].mbuf_pool = rte_pktmbuf_pool_create(pool_name, 357 mp_size, burst_conf.rx_burst * 4, 0, 358 RTE_MBUF_DEFAULT_BUF_SIZE, 359 rte_eth_dev_socket_id(qos_conf[i].rx_port)); 360 if (qos_conf[i].mbuf_pool == NULL) 361 rte_exit(EXIT_FAILURE, "Cannot init mbuf pool for socket %u\n", i); 362 363 app_init_port(qos_conf[i].rx_port, qos_conf[i].mbuf_pool); 364 app_init_port(qos_conf[i].tx_port, qos_conf[i].mbuf_pool); 365 366 qos_conf[i].sched_port = app_init_sched_port(qos_conf[i].tx_port, socket); 367 } 368 369 RTE_LOG(INFO, APP, "time stamp clock running at %" PRIu64 " Hz\n", 370 rte_get_timer_hz()); 371 372 RTE_LOG(INFO, APP, "Ring sizes: NIC RX = %u, Mempool = %d SW queue = %u," 373 "NIC TX = %u\n", ring_conf.rx_size, mp_size, ring_conf.ring_size, 374 ring_conf.tx_size); 375 376 RTE_LOG(INFO, APP, "Burst sizes: RX read = %hu, RX write = %hu,\n" 377 " Worker read/QoS enqueue = %hu,\n" 378 " QoS dequeue = %hu, Worker write = %hu\n", 379 burst_conf.rx_burst, burst_conf.ring_burst, burst_conf.ring_burst, 380 burst_conf.qos_dequeue, burst_conf.tx_burst); 381 382 RTE_LOG(INFO, APP, "NIC thresholds RX (p = %hhu, h = %hhu, w = %hhu)," 383 "TX (p = %hhu, h = %hhu, w = %hhu)\n", 384 rx_thresh.pthresh, rx_thresh.hthresh, rx_thresh.wthresh, 385 tx_thresh.pthresh, tx_thresh.hthresh, tx_thresh.wthresh); 386 387 return 0; 388 } 389