1aaf4363eSJerin Jacob /* SPDX-License-Identifier: BSD-3-Clause
2aaf4363eSJerin Jacob * Copyright(c) 2016 Cavium, Inc
3e4387966SJerin Jacob */
4e4387966SJerin Jacob
5e4387966SJerin Jacob #ifndef __THUNDERX_NICVF_ETHDEV_H__
6e4387966SJerin Jacob #define __THUNDERX_NICVF_ETHDEV_H__
7e4387966SJerin Jacob
8df96fd0dSBruce Richardson #include <ethdev_driver.h>
9e4387966SJerin Jacob
1046091359SKamil Rytarowski #define THUNDERX_NICVF_PMD_VERSION "2.0"
11001a1c0fSZyta Szpak #define THUNDERX_REG_BYTES 8
12e4387966SJerin Jacob
13e4387966SJerin Jacob #define NICVF_INTR_POLL_INTERVAL_MS 50
14*1f7b83b8SHanumanth Pothula /* Poll for link state for every 2 sec */
15*1f7b83b8SHanumanth Pothula #define NICVF_INTR_LINK_POLL_INTERVAL_MS 2000
168fc70464SJerin Jacob #define NICVF_HALF_DUPLEX 0x00
178fc70464SJerin Jacob #define NICVF_FULL_DUPLEX 0x01
188fc70464SJerin Jacob #define NICVF_UNKNOWN_DUPLEX 0xff
198fc70464SJerin Jacob
20dcd7b1e1SJerin Jacob #define NICVF_RSS_OFFLOAD_PASS1 ( \
21295968d1SFerruh Yigit RTE_ETH_RSS_PORT | \
22295968d1SFerruh Yigit RTE_ETH_RSS_IPV4 | \
23295968d1SFerruh Yigit RTE_ETH_RSS_NONFRAG_IPV4_TCP | \
24295968d1SFerruh Yigit RTE_ETH_RSS_NONFRAG_IPV4_UDP | \
25295968d1SFerruh Yigit RTE_ETH_RSS_IPV6 | \
26295968d1SFerruh Yigit RTE_ETH_RSS_NONFRAG_IPV6_TCP | \
27295968d1SFerruh Yigit RTE_ETH_RSS_NONFRAG_IPV6_UDP)
28dcd7b1e1SJerin Jacob
29dcd7b1e1SJerin Jacob #define NICVF_RSS_OFFLOAD_TUNNEL ( \
30295968d1SFerruh Yigit RTE_ETH_RSS_VXLAN | \
31295968d1SFerruh Yigit RTE_ETH_RSS_GENEVE | \
32295968d1SFerruh Yigit RTE_ETH_RSS_NVGRE)
33dcd7b1e1SJerin Jacob
34c97da2cbSMaciej Czekaj #define NICVF_TX_OFFLOAD_CAPA ( \
35295968d1SFerruh Yigit RTE_ETH_TX_OFFLOAD_IPV4_CKSUM | \
36295968d1SFerruh Yigit RTE_ETH_TX_OFFLOAD_UDP_CKSUM | \
37295968d1SFerruh Yigit RTE_ETH_TX_OFFLOAD_TCP_CKSUM | \
38295968d1SFerruh Yigit RTE_ETH_TX_OFFLOAD_OUTER_IPV4_CKSUM | \
39295968d1SFerruh Yigit RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE | \
40295968d1SFerruh Yigit RTE_ETH_TX_OFFLOAD_MULTI_SEGS)
41c97da2cbSMaciej Czekaj
42c97da2cbSMaciej Czekaj #define NICVF_RX_OFFLOAD_CAPA ( \
43295968d1SFerruh Yigit RTE_ETH_RX_OFFLOAD_CHECKSUM | \
44295968d1SFerruh Yigit RTE_ETH_RX_OFFLOAD_VLAN_STRIP | \
45295968d1SFerruh Yigit RTE_ETH_RX_OFFLOAD_SCATTER | \
46295968d1SFerruh Yigit RTE_ETH_RX_OFFLOAD_RSS_HASH)
47c97da2cbSMaciej Czekaj
48dcd7b1e1SJerin Jacob #define NICVF_DEFAULT_RX_FREE_THRESH 224
49dcd7b1e1SJerin Jacob #define NICVF_DEFAULT_TX_FREE_THRESH 224
50dcd7b1e1SJerin Jacob #define NICVF_TX_FREE_MPOOL_THRESH 16
51aa0d976eSJerin Jacob #define NICVF_MAX_RX_FREE_THRESH 1024
52aa0d976eSJerin Jacob #define NICVF_MAX_TX_FREE_THRESH 1024
53e4387966SJerin Jacob
5465d9804eSJerin Jacob #define VLAN_TAG_SIZE 4 /* 802.3ac tag */
5565d9804eSJerin Jacob
56279d3319SRakesh Kudurumalla #define SKIP_DATA_BYTES "skip_data_bytes"
57e4387966SJerin Jacob static inline struct nicvf *
nicvf_pmd_priv(struct rte_eth_dev * eth_dev)58e4387966SJerin Jacob nicvf_pmd_priv(struct rte_eth_dev *eth_dev)
59e4387966SJerin Jacob {
60e4387966SJerin Jacob return eth_dev->data->dev_private;
61e4387966SJerin Jacob }
62e4387966SJerin Jacob
63e2d7fc9fSJerin Jacob static inline uint64_t
nicvf_mempool_phy_offset(struct rte_mempool * mp)64e2d7fc9fSJerin Jacob nicvf_mempool_phy_offset(struct rte_mempool *mp)
65e2d7fc9fSJerin Jacob {
66e2d7fc9fSJerin Jacob struct rte_mempool_memhdr *hdr;
67e2d7fc9fSJerin Jacob
68e2d7fc9fSJerin Jacob hdr = STAILQ_FIRST(&mp->mem_list);
69e2d7fc9fSJerin Jacob assert(hdr != NULL);
70efd785f9SThomas Monjalon return (uint64_t)((uintptr_t)hdr->addr - hdr->iova);
71e2d7fc9fSJerin Jacob }
72e2d7fc9fSJerin Jacob
73e2d7fc9fSJerin Jacob static inline uint16_t
nicvf_mbuff_meta_length(struct rte_mbuf * mbuf)74e2d7fc9fSJerin Jacob nicvf_mbuff_meta_length(struct rte_mbuf *mbuf)
75e2d7fc9fSJerin Jacob {
76e2d7fc9fSJerin Jacob return (uint16_t)((uintptr_t)mbuf->buf_addr - (uintptr_t)mbuf);
77e2d7fc9fSJerin Jacob }
78e2d7fc9fSJerin Jacob
7953a3ebfeSKamil Rytarowski static inline uint16_t
nicvf_netdev_qidx(struct nicvf * nic,uint8_t local_qidx)8053a3ebfeSKamil Rytarowski nicvf_netdev_qidx(struct nicvf *nic, uint8_t local_qidx)
8153a3ebfeSKamil Rytarowski {
8253a3ebfeSKamil Rytarowski uint16_t global_qidx = local_qidx;
8353a3ebfeSKamil Rytarowski
8453a3ebfeSKamil Rytarowski if (nic->sqs_mode)
8553a3ebfeSKamil Rytarowski global_qidx += ((nic->sqs_id + 1) * MAX_CMP_QUEUES_PER_QS);
8653a3ebfeSKamil Rytarowski
8753a3ebfeSKamil Rytarowski return global_qidx;
8853a3ebfeSKamil Rytarowski }
8953a3ebfeSKamil Rytarowski
90e2d7fc9fSJerin Jacob /*
91e2d7fc9fSJerin Jacob * Simple phy2virt functions assuming mbufs are in a single huge page
92e2d7fc9fSJerin Jacob * V = P + offset
93e2d7fc9fSJerin Jacob * P = V - offset
94e2d7fc9fSJerin Jacob */
95e2d7fc9fSJerin Jacob static inline uintptr_t
nicvf_mbuff_phy2virt(rte_iova_t phy,uint64_t mbuf_phys_off)96df6e0a06SSantosh Shukla nicvf_mbuff_phy2virt(rte_iova_t phy, uint64_t mbuf_phys_off)
97e2d7fc9fSJerin Jacob {
98e2d7fc9fSJerin Jacob return (uintptr_t)(phy + mbuf_phys_off);
99e2d7fc9fSJerin Jacob }
100e2d7fc9fSJerin Jacob
101e2d7fc9fSJerin Jacob static inline uintptr_t
nicvf_mbuff_virt2phy(uintptr_t virt,uint64_t mbuf_phys_off)102e2d7fc9fSJerin Jacob nicvf_mbuff_virt2phy(uintptr_t virt, uint64_t mbuf_phys_off)
103e2d7fc9fSJerin Jacob {
104df6e0a06SSantosh Shukla return (rte_iova_t)(virt - mbuf_phys_off);
105e2d7fc9fSJerin Jacob }
106e2d7fc9fSJerin Jacob
10753a3ebfeSKamil Rytarowski static inline void
nicvf_tx_range(struct rte_eth_dev * dev,struct nicvf * nic,uint16_t * tx_start,uint16_t * tx_end)10853a3ebfeSKamil Rytarowski nicvf_tx_range(struct rte_eth_dev *dev, struct nicvf *nic, uint16_t *tx_start,
10953a3ebfeSKamil Rytarowski uint16_t *tx_end)
11053a3ebfeSKamil Rytarowski {
11153a3ebfeSKamil Rytarowski uint16_t tmp;
11253a3ebfeSKamil Rytarowski
11353a3ebfeSKamil Rytarowski *tx_start = RTE_ALIGN_FLOOR(nicvf_netdev_qidx(nic, 0),
11453a3ebfeSKamil Rytarowski MAX_SND_QUEUES_PER_QS);
11553a3ebfeSKamil Rytarowski tmp = RTE_ALIGN_CEIL(nicvf_netdev_qidx(nic, 0) + 1,
11653a3ebfeSKamil Rytarowski MAX_SND_QUEUES_PER_QS) - 1;
11753a3ebfeSKamil Rytarowski *tx_end = dev->data->nb_tx_queues ?
11853a3ebfeSKamil Rytarowski RTE_MIN(tmp, dev->data->nb_tx_queues - 1) : 0;
11953a3ebfeSKamil Rytarowski }
12053a3ebfeSKamil Rytarowski
12153a3ebfeSKamil Rytarowski static inline void
nicvf_rx_range(struct rte_eth_dev * dev,struct nicvf * nic,uint16_t * rx_start,uint16_t * rx_end)12253a3ebfeSKamil Rytarowski nicvf_rx_range(struct rte_eth_dev *dev, struct nicvf *nic, uint16_t *rx_start,
12353a3ebfeSKamil Rytarowski uint16_t *rx_end)
12453a3ebfeSKamil Rytarowski {
12553a3ebfeSKamil Rytarowski uint16_t tmp;
12653a3ebfeSKamil Rytarowski
12753a3ebfeSKamil Rytarowski *rx_start = RTE_ALIGN_FLOOR(nicvf_netdev_qidx(nic, 0),
12853a3ebfeSKamil Rytarowski MAX_RCV_QUEUES_PER_QS);
12953a3ebfeSKamil Rytarowski tmp = RTE_ALIGN_CEIL(nicvf_netdev_qidx(nic, 0) + 1,
13053a3ebfeSKamil Rytarowski MAX_RCV_QUEUES_PER_QS) - 1;
13153a3ebfeSKamil Rytarowski *rx_end = dev->data->nb_rx_queues ?
13253a3ebfeSKamil Rytarowski RTE_MIN(tmp, dev->data->nb_rx_queues - 1) : 0;
13353a3ebfeSKamil Rytarowski }
13453a3ebfeSKamil Rytarowski
135e4387966SJerin Jacob #endif /* __THUNDERX_NICVF_ETHDEV_H__ */
136