1 /* SPDX-License-Identifier: BSD-3-Clause 2 * Copyright(c) 2010-2014 Intel Corporation 3 */ 4 5 #include <stdarg.h> 6 #include <string.h> 7 #include <stdio.h> 8 #include <errno.h> 9 #include <stdint.h> 10 #include <unistd.h> 11 #include <inttypes.h> 12 13 #include <sys/queue.h> 14 #include <sys/stat.h> 15 16 #include <rte_common.h> 17 #include <rte_byteorder.h> 18 #include <rte_log.h> 19 #include <rte_debug.h> 20 #include <rte_cycles.h> 21 #include <rte_memory.h> 22 #include <rte_memcpy.h> 23 #include <rte_launch.h> 24 #include <rte_eal.h> 25 #include <rte_per_lcore.h> 26 #include <rte_lcore.h> 27 #include <rte_branch_prediction.h> 28 #include <rte_mempool.h> 29 #include <rte_mbuf.h> 30 #include <rte_interrupts.h> 31 #include <rte_pci.h> 32 #include <rte_ether.h> 33 #include <rte_ethdev.h> 34 #include <rte_ip.h> 35 #include <rte_string_fns.h> 36 #include <rte_flow.h> 37 38 #include "testpmd.h" 39 40 /* 41 * Forwarding of packets in MAC mode. 42 * Change the source and the destination Ethernet addressed of packets 43 * before forwarding them. 44 */ 45 static void 46 pkt_burst_mac_forward(struct fwd_stream *fs) 47 { 48 struct rte_mbuf *pkts_burst[MAX_PKT_BURST]; 49 struct rte_port *txp; 50 struct rte_mbuf *mb; 51 struct rte_ether_hdr *eth_hdr; 52 uint32_t retry; 53 uint16_t nb_rx; 54 uint16_t nb_tx; 55 uint16_t i; 56 uint64_t ol_flags = 0; 57 uint64_t tx_offloads; 58 uint64_t start_tsc = 0; 59 60 get_start_cycles(&start_tsc); 61 62 /* 63 * Receive a burst of packets and forward them. 64 */ 65 nb_rx = rte_eth_rx_burst(fs->rx_port, fs->rx_queue, pkts_burst, 66 nb_pkt_per_burst); 67 inc_rx_burst_stats(fs, nb_rx); 68 if (unlikely(nb_rx == 0)) 69 return; 70 71 fs->rx_packets += nb_rx; 72 txp = &ports[fs->tx_port]; 73 tx_offloads = txp->dev_conf.txmode.offloads; 74 if (tx_offloads & RTE_ETH_TX_OFFLOAD_VLAN_INSERT) 75 ol_flags = RTE_MBUF_F_TX_VLAN; 76 if (tx_offloads & RTE_ETH_TX_OFFLOAD_QINQ_INSERT) 77 ol_flags |= RTE_MBUF_F_TX_QINQ; 78 if (tx_offloads & RTE_ETH_TX_OFFLOAD_MACSEC_INSERT) 79 ol_flags |= RTE_MBUF_F_TX_MACSEC; 80 for (i = 0; i < nb_rx; i++) { 81 if (likely(i < nb_rx - 1)) 82 rte_prefetch0(rte_pktmbuf_mtod(pkts_burst[i + 1], 83 void *)); 84 mb = pkts_burst[i]; 85 eth_hdr = rte_pktmbuf_mtod(mb, struct rte_ether_hdr *); 86 rte_ether_addr_copy(&peer_eth_addrs[fs->peer_addr], 87 ð_hdr->dst_addr); 88 rte_ether_addr_copy(&ports[fs->tx_port].eth_addr, 89 ð_hdr->src_addr); 90 mb->ol_flags &= RTE_MBUF_F_INDIRECT | RTE_MBUF_F_EXTERNAL; 91 mb->ol_flags |= ol_flags; 92 mb->l2_len = sizeof(struct rte_ether_hdr); 93 mb->l3_len = sizeof(struct rte_ipv4_hdr); 94 mb->vlan_tci = txp->tx_vlan_id; 95 mb->vlan_tci_outer = txp->tx_vlan_id_outer; 96 } 97 nb_tx = rte_eth_tx_burst(fs->tx_port, fs->tx_queue, pkts_burst, nb_rx); 98 /* 99 * Retry if necessary 100 */ 101 if (unlikely(nb_tx < nb_rx) && fs->retry_enabled) { 102 retry = 0; 103 while (nb_tx < nb_rx && retry++ < burst_tx_retry_num) { 104 rte_delay_us(burst_tx_delay_time); 105 nb_tx += rte_eth_tx_burst(fs->tx_port, fs->tx_queue, 106 &pkts_burst[nb_tx], nb_rx - nb_tx); 107 } 108 } 109 110 fs->tx_packets += nb_tx; 111 inc_tx_burst_stats(fs, nb_tx); 112 if (unlikely(nb_tx < nb_rx)) { 113 fs->fwd_dropped += (nb_rx - nb_tx); 114 do { 115 rte_pktmbuf_free(pkts_burst[nb_tx]); 116 } while (++nb_tx < nb_rx); 117 } 118 119 get_end_cycles(fs, start_tsc); 120 } 121 122 struct fwd_engine mac_fwd_engine = { 123 .fwd_mode_name = "mac", 124 .port_fwd_begin = NULL, 125 .port_fwd_end = NULL, 126 .packet_fwd = pkt_burst_mac_forward, 127 }; 128