1 /* SPDX-License-Identifier: BSD-3-Clause
2 * Copyright(c) 2016 Cavium, Inc
3 */
4
5 #ifndef __THUNDERX_NICVF_RXTX_H__
6 #define __THUNDERX_NICVF_RXTX_H__
7
8 #include <rte_byteorder.h>
9 #include <ethdev_driver.h>
10
11 #define NICVF_RX_OFFLOAD_NONE 0x1
12 #define NICVF_RX_OFFLOAD_CKSUM 0x2
13 #define NICVF_RX_OFFLOAD_VLAN_STRIP 0x4
14
15 #define NICVF_TX_OFFLOAD_MASK (RTE_MBUF_F_TX_IP_CKSUM | RTE_MBUF_F_TX_L4_MASK)
16
17 #if RTE_BYTE_ORDER == RTE_BIG_ENDIAN
18 static inline uint16_t __attribute__((const))
nicvf_frag_num(uint16_t i)19 nicvf_frag_num(uint16_t i)
20 {
21 return (i & ~3) + 3 - (i & 3);
22 }
23
24 static inline void __rte_hot
fill_sq_desc_gather(union sq_entry_t * entry,struct rte_mbuf * pkt)25 fill_sq_desc_gather(union sq_entry_t *entry, struct rte_mbuf *pkt)
26 {
27 /* Local variable sqe to avoid read from sq desc memory*/
28 union sq_entry_t sqe;
29
30 /* Fill the SQ gather entry */
31 sqe.buff[0] = 0; sqe.buff[1] = 0;
32 sqe.gather.subdesc_type = SQ_DESC_TYPE_GATHER;
33 sqe.gather.ld_type = NIC_SEND_LD_TYPE_E_LDT;
34 sqe.gather.size = pkt->data_len;
35 sqe.gather.addr = rte_mbuf_data_iova(pkt);
36
37 entry->buff[0] = sqe.buff[0];
38 entry->buff[1] = sqe.buff[1];
39 }
40
41 #else
42
43 static inline uint16_t __attribute__((const))
nicvf_frag_num(uint16_t i)44 nicvf_frag_num(uint16_t i)
45 {
46 return i;
47 }
48
49 static inline void __rte_hot
fill_sq_desc_gather(union sq_entry_t * entry,struct rte_mbuf * pkt)50 fill_sq_desc_gather(union sq_entry_t *entry, struct rte_mbuf *pkt)
51 {
52 entry->buff[0] = (uint64_t)SQ_DESC_TYPE_GATHER << 60 |
53 (uint64_t)NIC_SEND_LD_TYPE_E_LDT << 58 |
54 pkt->data_len;
55 entry->buff[1] = rte_mbuf_data_iova(pkt);
56 }
57 #endif
58
59 static inline void
nicvf_mbuff_init_update(struct rte_mbuf * pkt,const uint64_t mbuf_init,uint16_t apad)60 nicvf_mbuff_init_update(struct rte_mbuf *pkt, const uint64_t mbuf_init,
61 uint16_t apad)
62 {
63 union mbuf_initializer init = {.value = mbuf_init};
64 #if RTE_BYTE_ORDER == RTE_BIG_ENDIAN
65 init.fields.data_off += apad;
66 #else
67 init.value += apad;
68 #endif
69 *(uint64_t *)(&pkt->rearm_data) = init.value;
70 }
71
72 static inline void
nicvf_mbuff_init_mseg_update(struct rte_mbuf * pkt,const uint64_t mbuf_init,uint16_t apad,uint16_t nb_segs)73 nicvf_mbuff_init_mseg_update(struct rte_mbuf *pkt, const uint64_t mbuf_init,
74 uint16_t apad, uint16_t nb_segs)
75 {
76 union mbuf_initializer init = {.value = mbuf_init};
77 #if RTE_BYTE_ORDER == RTE_BIG_ENDIAN
78 init.fields.data_off += apad;
79 #else
80 init.value += apad;
81 #endif
82 init.fields.nb_segs = nb_segs;
83 *(uint64_t *)(&pkt->rearm_data) = init.value;
84 }
85
86 uint32_t nicvf_dev_rx_queue_count(void *rx_queue);
87 uint32_t nicvf_dev_rbdr_refill(struct rte_eth_dev *dev, uint16_t queue_idx);
88
89 uint16_t nicvf_recv_pkts_no_offload(void *rxq, struct rte_mbuf **rx_pkts,
90 uint16_t pkts);
91 uint16_t nicvf_recv_pkts_cksum(void *rxq, struct rte_mbuf **rx_pkts,
92 uint16_t pkts);
93 uint16_t nicvf_recv_pkts_vlan_strip(void *rx_queue, struct rte_mbuf **rx_pkts,
94 uint16_t nb_pkts);
95 uint16_t nicvf_recv_pkts_cksum_vlan_strip(void *rx_queue,
96 struct rte_mbuf **rx_pkts, uint16_t nb_pkts);
97
98 uint16_t nicvf_recv_pkts_multiseg_no_offload(void *rx_queue,
99 struct rte_mbuf **rx_pkts, uint16_t nb_pkts);
100 uint16_t nicvf_recv_pkts_multiseg_cksum(void *rx_queue,
101 struct rte_mbuf **rx_pkts, uint16_t nb_pkts);
102 uint16_t nicvf_recv_pkts_multiseg_vlan_strip(void *rx_queue,
103 struct rte_mbuf **rx_pkts, uint16_t nb_pkts);
104 uint16_t nicvf_recv_pkts_multiseg_cksum_vlan_strip(void *rx_queue,
105 struct rte_mbuf **rx_pkts, uint16_t nb_pkts);
106
107 uint16_t nicvf_xmit_pkts(void *txq, struct rte_mbuf **tx_pkts, uint16_t pkts);
108 uint16_t nicvf_xmit_pkts_multiseg(void *txq, struct rte_mbuf **tx_pkts,
109 uint16_t pkts);
110
111 void nicvf_single_pool_free_xmited_buffers(struct nicvf_txq *sq);
112 void nicvf_multi_pool_free_xmited_buffers(struct nicvf_txq *sq);
113
114 #endif /* __THUNDERX_NICVF_RXTX_H__ */
115