1 /* SPDX-License-Identifier: BSD-3-Clause 2 * Copyright(c) 2010-2014 Intel Corporation 3 * Copyright 2018 Mellanox Technologies, Ltd 4 */ 5 6 #include <stdio.h> 7 8 #include <rte_net.h> 9 #include <rte_mbuf.h> 10 #include <rte_ether.h> 11 #include <rte_vxlan.h> 12 #include <rte_ethdev.h> 13 #include <rte_flow.h> 14 15 #include "testpmd.h" 16 17 static inline void 18 print_ether_addr(const char *what, const struct rte_ether_addr *eth_addr) 19 { 20 char buf[RTE_ETHER_ADDR_FMT_SIZE]; 21 rte_ether_format_addr(buf, RTE_ETHER_ADDR_FMT_SIZE, eth_addr); 22 printf("%s%s", what, buf); 23 } 24 25 static inline void 26 dump_pkt_burst(uint16_t port_id, uint16_t queue, struct rte_mbuf *pkts[], 27 uint16_t nb_pkts, int is_rx) 28 { 29 struct rte_mbuf *mb; 30 const struct rte_ether_hdr *eth_hdr; 31 struct rte_ether_hdr _eth_hdr; 32 uint16_t eth_type; 33 uint64_t ol_flags; 34 uint16_t i, packet_type; 35 uint16_t is_encapsulation; 36 char buf[256]; 37 struct rte_net_hdr_lens hdr_lens; 38 uint32_t sw_packet_type; 39 uint16_t udp_port; 40 uint32_t vx_vni; 41 const char *reason; 42 int dynf_index; 43 44 if (!nb_pkts) 45 return; 46 printf("port %u/queue %u: %s %u packets\n", 47 port_id, queue, 48 is_rx ? "received" : "sent", 49 (unsigned int) nb_pkts); 50 for (i = 0; i < nb_pkts; i++) { 51 int ret; 52 struct rte_flow_error error; 53 struct rte_flow_restore_info info = { 0, }; 54 55 mb = pkts[i]; 56 eth_hdr = rte_pktmbuf_read(mb, 0, sizeof(_eth_hdr), &_eth_hdr); 57 eth_type = RTE_BE_TO_CPU_16(eth_hdr->ether_type); 58 packet_type = mb->packet_type; 59 is_encapsulation = RTE_ETH_IS_TUNNEL_PKT(packet_type); 60 ret = rte_flow_get_restore_info(port_id, mb, &info, &error); 61 if (!ret) { 62 printf("restore info:"); 63 if (info.flags & RTE_FLOW_RESTORE_INFO_TUNNEL) { 64 struct port_flow_tunnel *port_tunnel; 65 66 port_tunnel = port_flow_locate_tunnel 67 (port_id, &info.tunnel); 68 printf(" - tunnel"); 69 if (port_tunnel) 70 printf(" #%u", port_tunnel->id); 71 else 72 printf(" %s", "-none-"); 73 printf(" type %s", 74 port_flow_tunnel_type(&info.tunnel)); 75 } else { 76 printf(" - no tunnel info"); 77 } 78 if (info.flags & RTE_FLOW_RESTORE_INFO_ENCAPSULATED) 79 printf(" - outer header present"); 80 else 81 printf(" - no outer header"); 82 if (info.flags & RTE_FLOW_RESTORE_INFO_GROUP_ID) 83 printf(" - miss group %u", info.group_id); 84 else 85 printf(" - no miss group"); 86 printf("\n"); 87 } 88 print_ether_addr(" src=", ð_hdr->s_addr); 89 print_ether_addr(" - dst=", ð_hdr->d_addr); 90 printf(" - type=0x%04x - length=%u - nb_segs=%d", 91 eth_type, (unsigned int) mb->pkt_len, 92 (int)mb->nb_segs); 93 ol_flags = mb->ol_flags; 94 if (ol_flags & PKT_RX_RSS_HASH) { 95 printf(" - RSS hash=0x%x", (unsigned int) mb->hash.rss); 96 printf(" - RSS queue=0x%x", (unsigned int) queue); 97 } 98 if (ol_flags & PKT_RX_FDIR) { 99 printf(" - FDIR matched "); 100 if (ol_flags & PKT_RX_FDIR_ID) 101 printf("ID=0x%x", 102 mb->hash.fdir.hi); 103 else if (ol_flags & PKT_RX_FDIR_FLX) 104 printf("flex bytes=0x%08x %08x", 105 mb->hash.fdir.hi, mb->hash.fdir.lo); 106 else 107 printf("hash=0x%x ID=0x%x ", 108 mb->hash.fdir.hash, mb->hash.fdir.id); 109 } 110 if (ol_flags & PKT_RX_TIMESTAMP) 111 printf(" - timestamp %"PRIu64" ", mb->timestamp); 112 if (ol_flags & PKT_RX_QINQ) 113 printf(" - QinQ VLAN tci=0x%x, VLAN tci outer=0x%x", 114 mb->vlan_tci, mb->vlan_tci_outer); 115 else if (ol_flags & PKT_RX_VLAN) 116 printf(" - VLAN tci=0x%x", mb->vlan_tci); 117 if (!is_rx && (ol_flags & PKT_TX_DYNF_METADATA)) 118 printf(" - Tx metadata: 0x%x", 119 *RTE_FLOW_DYNF_METADATA(mb)); 120 if (is_rx && (ol_flags & PKT_RX_DYNF_METADATA)) 121 printf(" - Rx metadata: 0x%x", 122 *RTE_FLOW_DYNF_METADATA(mb)); 123 for (dynf_index = 0; dynf_index < 64; dynf_index++) { 124 if (dynf_names[dynf_index][0] != '\0') 125 printf(" - dynf %s: %d", 126 dynf_names[dynf_index], 127 !!(ol_flags & (1UL << dynf_index))); 128 } 129 if (mb->packet_type) { 130 rte_get_ptype_name(mb->packet_type, buf, sizeof(buf)); 131 printf(" - hw ptype: %s", buf); 132 } 133 sw_packet_type = rte_net_get_ptype(mb, &hdr_lens, 134 RTE_PTYPE_ALL_MASK); 135 rte_get_ptype_name(sw_packet_type, buf, sizeof(buf)); 136 printf(" - sw ptype: %s", buf); 137 if (sw_packet_type & RTE_PTYPE_L2_MASK) 138 printf(" - l2_len=%d", hdr_lens.l2_len); 139 if (sw_packet_type & RTE_PTYPE_L3_MASK) 140 printf(" - l3_len=%d", hdr_lens.l3_len); 141 if (sw_packet_type & RTE_PTYPE_L4_MASK) 142 printf(" - l4_len=%d", hdr_lens.l4_len); 143 if (sw_packet_type & RTE_PTYPE_TUNNEL_MASK) 144 printf(" - tunnel_len=%d", hdr_lens.tunnel_len); 145 if (sw_packet_type & RTE_PTYPE_INNER_L2_MASK) 146 printf(" - inner_l2_len=%d", hdr_lens.inner_l2_len); 147 if (sw_packet_type & RTE_PTYPE_INNER_L3_MASK) 148 printf(" - inner_l3_len=%d", hdr_lens.inner_l3_len); 149 if (sw_packet_type & RTE_PTYPE_INNER_L4_MASK) 150 printf(" - inner_l4_len=%d", hdr_lens.inner_l4_len); 151 if (is_encapsulation) { 152 struct rte_ipv4_hdr *ipv4_hdr; 153 struct rte_ipv6_hdr *ipv6_hdr; 154 struct rte_udp_hdr *udp_hdr; 155 uint8_t l2_len; 156 uint8_t l3_len; 157 uint8_t l4_len; 158 uint8_t l4_proto; 159 struct rte_vxlan_hdr *vxlan_hdr; 160 161 l2_len = sizeof(struct rte_ether_hdr); 162 163 /* Do not support ipv4 option field */ 164 if (RTE_ETH_IS_IPV4_HDR(packet_type)) { 165 l3_len = sizeof(struct rte_ipv4_hdr); 166 ipv4_hdr = rte_pktmbuf_mtod_offset(mb, 167 struct rte_ipv4_hdr *, 168 l2_len); 169 l4_proto = ipv4_hdr->next_proto_id; 170 } else { 171 l3_len = sizeof(struct rte_ipv6_hdr); 172 ipv6_hdr = rte_pktmbuf_mtod_offset(mb, 173 struct rte_ipv6_hdr *, 174 l2_len); 175 l4_proto = ipv6_hdr->proto; 176 } 177 if (l4_proto == IPPROTO_UDP) { 178 udp_hdr = rte_pktmbuf_mtod_offset(mb, 179 struct rte_udp_hdr *, 180 l2_len + l3_len); 181 l4_len = sizeof(struct rte_udp_hdr); 182 vxlan_hdr = rte_pktmbuf_mtod_offset(mb, 183 struct rte_vxlan_hdr *, 184 l2_len + l3_len + l4_len); 185 udp_port = RTE_BE_TO_CPU_16(udp_hdr->dst_port); 186 vx_vni = rte_be_to_cpu_32(vxlan_hdr->vx_vni); 187 printf(" - VXLAN packet: packet type =%d, " 188 "Destination UDP port =%d, VNI = %d", 189 packet_type, udp_port, vx_vni >> 8); 190 } 191 } 192 printf(" - %s queue=0x%x", is_rx ? "Receive" : "Send", 193 (unsigned int) queue); 194 printf("\n"); 195 rte_get_rx_ol_flag_list(mb->ol_flags, buf, sizeof(buf)); 196 printf(" ol_flags: %s\n", buf); 197 if (rte_mbuf_check(mb, 1, &reason) < 0) 198 printf("INVALID mbuf: %s\n", reason); 199 } 200 } 201 202 uint16_t 203 dump_rx_pkts(uint16_t port_id, uint16_t queue, struct rte_mbuf *pkts[], 204 uint16_t nb_pkts, __rte_unused uint16_t max_pkts, 205 __rte_unused void *user_param) 206 { 207 dump_pkt_burst(port_id, queue, pkts, nb_pkts, 1); 208 return nb_pkts; 209 } 210 211 uint16_t 212 dump_tx_pkts(uint16_t port_id, uint16_t queue, struct rte_mbuf *pkts[], 213 uint16_t nb_pkts, __rte_unused void *user_param) 214 { 215 dump_pkt_burst(port_id, queue, pkts, nb_pkts, 0); 216 return nb_pkts; 217 } 218 219 uint16_t 220 tx_pkt_set_md(uint16_t port_id, __rte_unused uint16_t queue, 221 struct rte_mbuf *pkts[], uint16_t nb_pkts, 222 __rte_unused void *user_param) 223 { 224 uint16_t i = 0; 225 226 /* 227 * Add metadata value to every Tx packet, 228 * and set ol_flags accordingly. 229 */ 230 if (rte_flow_dynf_metadata_avail()) 231 for (i = 0; i < nb_pkts; i++) { 232 *RTE_FLOW_DYNF_METADATA(pkts[i]) = 233 ports[port_id].tx_metadata; 234 pkts[i]->ol_flags |= PKT_TX_DYNF_METADATA; 235 } 236 return nb_pkts; 237 } 238 239 void 240 add_tx_md_callback(portid_t portid) 241 { 242 struct rte_eth_dev_info dev_info; 243 uint16_t queue; 244 int ret; 245 246 if (port_id_is_invalid(portid, ENABLED_WARN)) 247 return; 248 249 ret = eth_dev_info_get_print_err(portid, &dev_info); 250 if (ret != 0) 251 return; 252 253 for (queue = 0; queue < dev_info.nb_tx_queues; queue++) 254 if (!ports[portid].tx_set_md_cb[queue]) 255 ports[portid].tx_set_md_cb[queue] = 256 rte_eth_add_tx_callback(portid, queue, 257 tx_pkt_set_md, NULL); 258 } 259 260 void 261 remove_tx_md_callback(portid_t portid) 262 { 263 struct rte_eth_dev_info dev_info; 264 uint16_t queue; 265 int ret; 266 267 if (port_id_is_invalid(portid, ENABLED_WARN)) 268 return; 269 270 ret = eth_dev_info_get_print_err(portid, &dev_info); 271 if (ret != 0) 272 return; 273 274 for (queue = 0; queue < dev_info.nb_tx_queues; queue++) 275 if (ports[portid].tx_set_md_cb[queue]) { 276 rte_eth_remove_tx_callback(portid, queue, 277 ports[portid].tx_set_md_cb[queue]); 278 ports[portid].tx_set_md_cb[queue] = NULL; 279 } 280 } 281 282 uint16_t 283 tx_pkt_set_dynf(uint16_t port_id, __rte_unused uint16_t queue, 284 struct rte_mbuf *pkts[], uint16_t nb_pkts, 285 __rte_unused void *user_param) 286 { 287 uint16_t i = 0; 288 289 if (ports[port_id].mbuf_dynf) 290 for (i = 0; i < nb_pkts; i++) 291 pkts[i]->ol_flags |= ports[port_id].mbuf_dynf; 292 return nb_pkts; 293 } 294 295 void 296 add_tx_dynf_callback(portid_t portid) 297 { 298 struct rte_eth_dev_info dev_info; 299 uint16_t queue; 300 int ret; 301 302 if (port_id_is_invalid(portid, ENABLED_WARN)) 303 return; 304 305 ret = eth_dev_info_get_print_err(portid, &dev_info); 306 if (ret != 0) 307 return; 308 309 for (queue = 0; queue < dev_info.nb_tx_queues; queue++) 310 if (!ports[portid].tx_set_dynf_cb[queue]) 311 ports[portid].tx_set_dynf_cb[queue] = 312 rte_eth_add_tx_callback(portid, queue, 313 tx_pkt_set_dynf, NULL); 314 } 315 316 void 317 remove_tx_dynf_callback(portid_t portid) 318 { 319 struct rte_eth_dev_info dev_info; 320 uint16_t queue; 321 int ret; 322 323 if (port_id_is_invalid(portid, ENABLED_WARN)) 324 return; 325 326 ret = eth_dev_info_get_print_err(portid, &dev_info); 327 if (ret != 0) 328 return; 329 330 for (queue = 0; queue < dev_info.nb_tx_queues; queue++) 331 if (ports[portid].tx_set_dynf_cb[queue]) { 332 rte_eth_remove_tx_callback(portid, queue, 333 ports[portid].tx_set_dynf_cb[queue]); 334 ports[portid].tx_set_dynf_cb[queue] = NULL; 335 } 336 } 337 338 int 339 eth_dev_info_get_print_err(uint16_t port_id, 340 struct rte_eth_dev_info *dev_info) 341 { 342 int ret; 343 344 ret = rte_eth_dev_info_get(port_id, dev_info); 345 if (ret != 0) 346 printf("Error during getting device (port %u) info: %s\n", 347 port_id, strerror(-ret)); 348 349 return ret; 350 } 351 352 void 353 eth_set_promisc_mode(uint16_t port, int enable) 354 { 355 int ret; 356 357 if (enable) 358 ret = rte_eth_promiscuous_enable(port); 359 else 360 ret = rte_eth_promiscuous_disable(port); 361 362 if (ret != 0) 363 printf("Error during %s promiscuous mode for port %u: %s\n", 364 enable ? "enabling" : "disabling", 365 port, rte_strerror(-ret)); 366 } 367 368 void 369 eth_set_allmulticast_mode(uint16_t port, int enable) 370 { 371 int ret; 372 373 if (enable) 374 ret = rte_eth_allmulticast_enable(port); 375 else 376 ret = rte_eth_allmulticast_disable(port); 377 378 if (ret != 0) 379 printf("Error during %s all-multicast mode for port %u: %s\n", 380 enable ? "enabling" : "disabling", 381 port, rte_strerror(-ret)); 382 } 383 384 int 385 eth_link_get_nowait_print_err(uint16_t port_id, struct rte_eth_link *link) 386 { 387 int ret; 388 389 ret = rte_eth_link_get_nowait(port_id, link); 390 if (ret < 0) 391 printf("Device (port %u) link get (without wait) failed: %s\n", 392 port_id, rte_strerror(-ret)); 393 394 return ret; 395 } 396 397 int 398 eth_macaddr_get_print_err(uint16_t port_id, struct rte_ether_addr *mac_addr) 399 { 400 int ret; 401 402 ret = rte_eth_macaddr_get(port_id, mac_addr); 403 if (ret != 0) 404 printf("Error getting device (port %u) mac address: %s\n", 405 port_id, rte_strerror(-ret)); 406 407 return ret; 408 } 409