1 /* SPDX-License-Identifier: BSD-3-Clause 2 * Copyright(c) 2016 Cavium, Inc 3 */ 4 5 #ifndef __THUNDERX_NICVF_RXTX_H__ 6 #define __THUNDERX_NICVF_RXTX_H__ 7 8 #include <rte_byteorder.h> 9 #include <ethdev_driver.h> 10 11 #define NICVF_RX_OFFLOAD_NONE 0x1 12 #define NICVF_RX_OFFLOAD_CKSUM 0x2 13 #define NICVF_RX_OFFLOAD_VLAN_STRIP 0x4 14 15 #define NICVF_TX_OFFLOAD_MASK (PKT_TX_IP_CKSUM | PKT_TX_L4_MASK) 16 17 #if RTE_BYTE_ORDER == RTE_BIG_ENDIAN 18 static inline uint16_t __attribute__((const)) 19 nicvf_frag_num(uint16_t i) 20 { 21 return (i & ~3) + 3 - (i & 3); 22 } 23 24 static inline void __rte_hot 25 fill_sq_desc_gather(union sq_entry_t *entry, struct rte_mbuf *pkt) 26 { 27 /* Local variable sqe to avoid read from sq desc memory*/ 28 union sq_entry_t sqe; 29 30 /* Fill the SQ gather entry */ 31 sqe.buff[0] = 0; sqe.buff[1] = 0; 32 sqe.gather.subdesc_type = SQ_DESC_TYPE_GATHER; 33 sqe.gather.ld_type = NIC_SEND_LD_TYPE_E_LDT; 34 sqe.gather.size = pkt->data_len; 35 sqe.gather.addr = rte_mbuf_data_iova(pkt); 36 37 entry->buff[0] = sqe.buff[0]; 38 entry->buff[1] = sqe.buff[1]; 39 } 40 41 #else 42 43 static inline uint16_t __attribute__((const)) 44 nicvf_frag_num(uint16_t i) 45 { 46 return i; 47 } 48 49 static inline void __rte_hot 50 fill_sq_desc_gather(union sq_entry_t *entry, struct rte_mbuf *pkt) 51 { 52 entry->buff[0] = (uint64_t)SQ_DESC_TYPE_GATHER << 60 | 53 (uint64_t)NIC_SEND_LD_TYPE_E_LDT << 58 | 54 pkt->data_len; 55 entry->buff[1] = rte_mbuf_data_iova(pkt); 56 } 57 #endif 58 59 static inline void 60 nicvf_mbuff_init_update(struct rte_mbuf *pkt, const uint64_t mbuf_init, 61 uint16_t apad) 62 { 63 union mbuf_initializer init = {.value = mbuf_init}; 64 #if RTE_BYTE_ORDER == RTE_BIG_ENDIAN 65 init.fields.data_off += apad; 66 #else 67 init.value += apad; 68 #endif 69 *(uint64_t *)(&pkt->rearm_data) = init.value; 70 } 71 72 static inline void 73 nicvf_mbuff_init_mseg_update(struct rte_mbuf *pkt, const uint64_t mbuf_init, 74 uint16_t apad, uint16_t nb_segs) 75 { 76 union mbuf_initializer init = {.value = mbuf_init}; 77 #if RTE_BYTE_ORDER == RTE_BIG_ENDIAN 78 init.fields.data_off += apad; 79 #else 80 init.value += apad; 81 #endif 82 init.fields.nb_segs = nb_segs; 83 *(uint64_t *)(&pkt->rearm_data) = init.value; 84 } 85 86 uint32_t nicvf_dev_rx_queue_count(struct rte_eth_dev *dev, uint16_t queue_idx); 87 uint32_t nicvf_dev_rbdr_refill(struct rte_eth_dev *dev, uint16_t queue_idx); 88 89 uint16_t nicvf_recv_pkts_no_offload(void *rxq, struct rte_mbuf **rx_pkts, 90 uint16_t pkts); 91 uint16_t nicvf_recv_pkts_cksum(void *rxq, struct rte_mbuf **rx_pkts, 92 uint16_t pkts); 93 uint16_t nicvf_recv_pkts_vlan_strip(void *rx_queue, struct rte_mbuf **rx_pkts, 94 uint16_t nb_pkts); 95 uint16_t nicvf_recv_pkts_cksum_vlan_strip(void *rx_queue, 96 struct rte_mbuf **rx_pkts, uint16_t nb_pkts); 97 98 uint16_t nicvf_recv_pkts_multiseg_no_offload(void *rx_queue, 99 struct rte_mbuf **rx_pkts, uint16_t nb_pkts); 100 uint16_t nicvf_recv_pkts_multiseg_cksum(void *rx_queue, 101 struct rte_mbuf **rx_pkts, uint16_t nb_pkts); 102 uint16_t nicvf_recv_pkts_multiseg_vlan_strip(void *rx_queue, 103 struct rte_mbuf **rx_pkts, uint16_t nb_pkts); 104 uint16_t nicvf_recv_pkts_multiseg_cksum_vlan_strip(void *rx_queue, 105 struct rte_mbuf **rx_pkts, uint16_t nb_pkts); 106 107 uint16_t nicvf_xmit_pkts(void *txq, struct rte_mbuf **tx_pkts, uint16_t pkts); 108 uint16_t nicvf_xmit_pkts_multiseg(void *txq, struct rte_mbuf **tx_pkts, 109 uint16_t pkts); 110 111 void nicvf_single_pool_free_xmited_buffers(struct nicvf_txq *sq); 112 void nicvf_multi_pool_free_xmited_buffers(struct nicvf_txq *sq); 113 114 #endif /* __THUNDERX_NICVF_RXTX_H__ */ 115