1 /* SPDX-License-Identifier: BSD-3-Clause 2 * Copyright(c) 2010-2014 Intel Corporation 3 */ 4 5 #include <stdarg.h> 6 #include <string.h> 7 #include <stdio.h> 8 #include <errno.h> 9 #include <stdint.h> 10 #include <unistd.h> 11 #include <inttypes.h> 12 13 #include <sys/queue.h> 14 #include <sys/stat.h> 15 16 #include <rte_common.h> 17 #include <rte_byteorder.h> 18 #include <rte_log.h> 19 #include <rte_debug.h> 20 #include <rte_cycles.h> 21 #include <rte_memory.h> 22 #include <rte_memcpy.h> 23 #include <rte_launch.h> 24 #include <rte_eal.h> 25 #include <rte_per_lcore.h> 26 #include <rte_lcore.h> 27 #include <rte_atomic.h> 28 #include <rte_branch_prediction.h> 29 #include <rte_mempool.h> 30 #include <rte_mbuf.h> 31 #include <rte_interrupts.h> 32 #include <rte_pci.h> 33 #include <rte_ether.h> 34 #include <rte_ethdev.h> 35 #include <rte_ip.h> 36 #include <rte_string_fns.h> 37 #include <rte_flow.h> 38 39 #include "testpmd.h" 40 41 /* 42 * Forwarding of packets in MAC mode. 43 * Change the source and the destination Ethernet addressed of packets 44 * before forwarding them. 45 */ 46 static void 47 pkt_burst_mac_forward(struct fwd_stream *fs) 48 { 49 struct rte_mbuf *pkts_burst[MAX_PKT_BURST]; 50 struct rte_port *txp; 51 struct rte_mbuf *mb; 52 struct rte_ether_hdr *eth_hdr; 53 uint32_t retry; 54 uint16_t nb_rx; 55 uint16_t nb_tx; 56 uint16_t i; 57 uint64_t ol_flags = 0; 58 uint64_t tx_offloads; 59 uint64_t start_tsc = 0; 60 61 get_start_cycles(&start_tsc); 62 63 /* 64 * Receive a burst of packets and forward them. 65 */ 66 nb_rx = rte_eth_rx_burst(fs->rx_port, fs->rx_queue, pkts_burst, 67 nb_pkt_per_burst); 68 inc_rx_burst_stats(fs, nb_rx); 69 if (unlikely(nb_rx == 0)) 70 return; 71 72 fs->rx_packets += nb_rx; 73 txp = &ports[fs->tx_port]; 74 tx_offloads = txp->dev_conf.txmode.offloads; 75 if (tx_offloads & DEV_TX_OFFLOAD_VLAN_INSERT) 76 ol_flags = PKT_TX_VLAN_PKT; 77 if (tx_offloads & DEV_TX_OFFLOAD_QINQ_INSERT) 78 ol_flags |= PKT_TX_QINQ_PKT; 79 if (tx_offloads & DEV_TX_OFFLOAD_MACSEC_INSERT) 80 ol_flags |= PKT_TX_MACSEC; 81 for (i = 0; i < nb_rx; i++) { 82 if (likely(i < nb_rx - 1)) 83 rte_prefetch0(rte_pktmbuf_mtod(pkts_burst[i + 1], 84 void *)); 85 mb = pkts_burst[i]; 86 eth_hdr = rte_pktmbuf_mtod(mb, struct rte_ether_hdr *); 87 rte_ether_addr_copy(&peer_eth_addrs[fs->peer_addr], 88 ð_hdr->d_addr); 89 rte_ether_addr_copy(&ports[fs->tx_port].eth_addr, 90 ð_hdr->s_addr); 91 mb->ol_flags &= IND_ATTACHED_MBUF | EXT_ATTACHED_MBUF; 92 mb->ol_flags |= ol_flags; 93 mb->l2_len = sizeof(struct rte_ether_hdr); 94 mb->l3_len = sizeof(struct rte_ipv4_hdr); 95 mb->vlan_tci = txp->tx_vlan_id; 96 mb->vlan_tci_outer = txp->tx_vlan_id_outer; 97 } 98 nb_tx = rte_eth_tx_burst(fs->tx_port, fs->tx_queue, pkts_burst, nb_rx); 99 /* 100 * Retry if necessary 101 */ 102 if (unlikely(nb_tx < nb_rx) && fs->retry_enabled) { 103 retry = 0; 104 while (nb_tx < nb_rx && retry++ < burst_tx_retry_num) { 105 rte_delay_us(burst_tx_delay_time); 106 nb_tx += rte_eth_tx_burst(fs->tx_port, fs->tx_queue, 107 &pkts_burst[nb_tx], nb_rx - nb_tx); 108 } 109 } 110 111 fs->tx_packets += nb_tx; 112 inc_tx_burst_stats(fs, nb_tx); 113 if (unlikely(nb_tx < nb_rx)) { 114 fs->fwd_dropped += (nb_rx - nb_tx); 115 do { 116 rte_pktmbuf_free(pkts_burst[nb_tx]); 117 } while (++nb_tx < nb_rx); 118 } 119 120 get_end_cycles(fs, start_tsc); 121 } 122 123 struct fwd_engine mac_fwd_engine = { 124 .fwd_mode_name = "mac", 125 .port_fwd_begin = NULL, 126 .port_fwd_end = NULL, 127 .packet_fwd = pkt_burst_mac_forward, 128 }; 129