xref: /dpdk/drivers/net/hns3/hns3_rxtx_vec.c (revision d49b64477f246e53210488825fdd92ccf53fa184)
1e31f123dSWei Hu (Xavier) /* SPDX-License-Identifier: BSD-3-Clause
253e6f86cSMin Hu (Connor)  * Copyright(c) 2020-2021 HiSilicon Limited.
3e31f123dSWei Hu (Xavier)  */
4e31f123dSWei Hu (Xavier) 
5e31f123dSWei Hu (Xavier) #include <rte_io.h>
6df96fd0dSBruce Richardson #include <ethdev_driver.h>
7e31f123dSWei Hu (Xavier) 
8e31f123dSWei Hu (Xavier) #include "hns3_ethdev.h"
9e31f123dSWei Hu (Xavier) #include "hns3_rxtx.h"
10e31f123dSWei Hu (Xavier) #include "hns3_rxtx_vec.h"
11e31f123dSWei Hu (Xavier) 
12e31f123dSWei Hu (Xavier) #if defined RTE_ARCH_ARM64
13e31f123dSWei Hu (Xavier) #include "hns3_rxtx_vec_neon.h"
14e31f123dSWei Hu (Xavier) #endif
15e31f123dSWei Hu (Xavier) 
16e31f123dSWei Hu (Xavier) int
hns3_tx_check_vec_support(struct rte_eth_dev * dev)17e31f123dSWei Hu (Xavier) hns3_tx_check_vec_support(struct rte_eth_dev *dev)
18e31f123dSWei Hu (Xavier) {
19e31f123dSWei Hu (Xavier) 	struct rte_eth_txmode *txmode = &dev->data->dev_conf.txmode;
20*4ac14c1dSHuisong Li 	struct rte_eth_rxmode *rxmode = &dev->data->dev_conf.rxmode;
2138b539d9SMin Hu (Connor) 
22295968d1SFerruh Yigit 	/* Only support RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE */
23295968d1SFerruh Yigit 	if (txmode->offloads != RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE)
24e31f123dSWei Hu (Xavier) 		return -ENOTSUP;
25e31f123dSWei Hu (Xavier) 
26*4ac14c1dSHuisong Li 	/*
27*4ac14c1dSHuisong Li 	 * PTP function requires the cooperation of Rx and Tx.
28*4ac14c1dSHuisong Li 	 * Tx vector isn't supported if RTE_ETH_RX_OFFLOAD_TIMESTAMP is set
29*4ac14c1dSHuisong Li 	 * in Rx offloads.
30*4ac14c1dSHuisong Li 	 */
31*4ac14c1dSHuisong Li 	if (rxmode->offloads & RTE_ETH_RX_OFFLOAD_TIMESTAMP)
323ca3dcd6SMin Hu (Connor) 		return -ENOTSUP;
333ca3dcd6SMin Hu (Connor) 
34e31f123dSWei Hu (Xavier) 	return 0;
35e31f123dSWei Hu (Xavier) }
36e31f123dSWei Hu (Xavier) 
37e31f123dSWei Hu (Xavier) uint16_t
hns3_xmit_pkts_vec(void * tx_queue,struct rte_mbuf ** tx_pkts,uint16_t nb_pkts)38e31f123dSWei Hu (Xavier) hns3_xmit_pkts_vec(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts)
39e31f123dSWei Hu (Xavier) {
40e31f123dSWei Hu (Xavier) 	struct hns3_tx_queue *txq = (struct hns3_tx_queue *)tx_queue;
41e31f123dSWei Hu (Xavier) 	uint16_t nb_tx = 0;
42e31f123dSWei Hu (Xavier) 
43e31f123dSWei Hu (Xavier) 	while (nb_pkts) {
44e31f123dSWei Hu (Xavier) 		uint16_t ret, new_burst;
45e31f123dSWei Hu (Xavier) 
46e31f123dSWei Hu (Xavier) 		new_burst = RTE_MIN(nb_pkts, txq->tx_rs_thresh);
47e31f123dSWei Hu (Xavier) 		ret = hns3_xmit_fixed_burst_vec(tx_queue, &tx_pkts[nb_tx],
48e31f123dSWei Hu (Xavier) 						new_burst);
49e31f123dSWei Hu (Xavier) 		nb_tx += ret;
50e31f123dSWei Hu (Xavier) 		nb_pkts -= ret;
51e31f123dSWei Hu (Xavier) 		if (ret < new_burst)
52e31f123dSWei Hu (Xavier) 			break;
53e31f123dSWei Hu (Xavier) 	}
54e31f123dSWei Hu (Xavier) 
55e31f123dSWei Hu (Xavier) 	return nb_tx;
56e31f123dSWei Hu (Xavier) }
57a3d4f4d2SWei Hu (Xavier) 
58a3d4f4d2SWei Hu (Xavier) uint16_t
hns3_recv_pkts_vec(void * __restrict rx_queue,struct rte_mbuf ** __restrict rx_pkts,uint16_t nb_pkts)59a3d4f4d2SWei Hu (Xavier) hns3_recv_pkts_vec(void *__restrict rx_queue,
60a3d4f4d2SWei Hu (Xavier) 		   struct rte_mbuf **__restrict rx_pkts,
61a3d4f4d2SWei Hu (Xavier) 		   uint16_t nb_pkts)
62a3d4f4d2SWei Hu (Xavier) {
63a3d4f4d2SWei Hu (Xavier) 	struct hns3_rx_queue *rxq = rx_queue;
64a3d4f4d2SWei Hu (Xavier) 	struct hns3_desc *rxdp = &rxq->rx_ring[rxq->next_to_use];
652d408d06SChengwen Feng 	uint64_t pkt_err_mask;  /* bit mask indicate whick pkts is error */
66a3d4f4d2SWei Hu (Xavier) 	uint16_t nb_rx;
67a3d4f4d2SWei Hu (Xavier) 
68a3d4f4d2SWei Hu (Xavier) 	rte_prefetch_non_temporal(rxdp);
69a3d4f4d2SWei Hu (Xavier) 
702d408d06SChengwen Feng 	nb_pkts = RTE_ALIGN_FLOOR(nb_pkts, HNS3_DEFAULT_DESCS_PER_LOOP);
712d408d06SChengwen Feng 
72a3d4f4d2SWei Hu (Xavier) 	if (rxq->rx_rearm_nb > HNS3_DEFAULT_RXQ_REARM_THRESH)
73a3d4f4d2SWei Hu (Xavier) 		hns3_rxq_rearm_mbuf(rxq);
74a3d4f4d2SWei Hu (Xavier) 
75a3d4f4d2SWei Hu (Xavier) 	if (unlikely(!(rxdp->rx.bd_base_info &
76a3d4f4d2SWei Hu (Xavier) 			rte_cpu_to_le_32(1u << HNS3_RXD_VLD_B))))
77a3d4f4d2SWei Hu (Xavier) 		return 0;
78a3d4f4d2SWei Hu (Xavier) 
79a3d4f4d2SWei Hu (Xavier) 	rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 0].mbuf);
80a3d4f4d2SWei Hu (Xavier) 	rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 1].mbuf);
81a3d4f4d2SWei Hu (Xavier) 	rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 2].mbuf);
82a3d4f4d2SWei Hu (Xavier) 	rte_prefetch0(rxq->sw_ring[rxq->next_to_use + 3].mbuf);
83a3d4f4d2SWei Hu (Xavier) 
842d408d06SChengwen Feng 	if (likely(nb_pkts <= HNS3_DEFAULT_RX_BURST)) {
852d408d06SChengwen Feng 		pkt_err_mask = 0;
862d408d06SChengwen Feng 		nb_rx = hns3_recv_burst_vec(rxq, rx_pkts, nb_pkts,
872d408d06SChengwen Feng 					    &pkt_err_mask);
882d408d06SChengwen Feng 		nb_rx = hns3_rx_reassemble_pkts(rx_pkts, nb_rx, pkt_err_mask);
892d408d06SChengwen Feng 		return nb_rx;
902d408d06SChengwen Feng 	}
912d408d06SChengwen Feng 
922d408d06SChengwen Feng 	nb_rx = 0;
932d408d06SChengwen Feng 	while (nb_pkts > 0) {
942d408d06SChengwen Feng 		uint16_t ret, n;
952d408d06SChengwen Feng 
962d408d06SChengwen Feng 		n = RTE_MIN(nb_pkts, HNS3_DEFAULT_RX_BURST);
972d408d06SChengwen Feng 		pkt_err_mask = 0;
982d408d06SChengwen Feng 		ret = hns3_recv_burst_vec(rxq, &rx_pkts[nb_rx], n,
992d408d06SChengwen Feng 					  &pkt_err_mask);
1002d408d06SChengwen Feng 		nb_pkts -= ret;
1012d408d06SChengwen Feng 		nb_rx += hns3_rx_reassemble_pkts(&rx_pkts[nb_rx], ret,
1022d408d06SChengwen Feng 						 pkt_err_mask);
1032d408d06SChengwen Feng 		if (ret < n)
1042d408d06SChengwen Feng 			break;
1052d408d06SChengwen Feng 
1062d408d06SChengwen Feng 		if (rxq->rx_rearm_nb > HNS3_DEFAULT_RXQ_REARM_THRESH)
1072d408d06SChengwen Feng 			hns3_rxq_rearm_mbuf(rxq);
1082d408d06SChengwen Feng 	}
109a3d4f4d2SWei Hu (Xavier) 
110a3d4f4d2SWei Hu (Xavier) 	return nb_rx;
111a3d4f4d2SWei Hu (Xavier) }
112a3d4f4d2SWei Hu (Xavier) 
113a3d4f4d2SWei Hu (Xavier) static void
hns3_rxq_vec_setup_rearm_data(struct hns3_rx_queue * rxq)114a3d4f4d2SWei Hu (Xavier) hns3_rxq_vec_setup_rearm_data(struct hns3_rx_queue *rxq)
115a3d4f4d2SWei Hu (Xavier) {
116a3d4f4d2SWei Hu (Xavier) 	uintptr_t p;
117a3d4f4d2SWei Hu (Xavier) 	struct rte_mbuf mb_def = { .buf_addr = 0 }; /* zeroed mbuf */
118a3d4f4d2SWei Hu (Xavier) 
119a3d4f4d2SWei Hu (Xavier) 	mb_def.nb_segs = 1;
120a3d4f4d2SWei Hu (Xavier) 	mb_def.data_off = RTE_PKTMBUF_HEADROOM;
121a3d4f4d2SWei Hu (Xavier) 	mb_def.port = rxq->port_id;
122a3d4f4d2SWei Hu (Xavier) 	rte_mbuf_refcnt_set(&mb_def, 1);
123a3d4f4d2SWei Hu (Xavier) 
124cb12e988SChengwen Feng 	/* compile-time verifies the rearm_data first 8bytes */
125cb12e988SChengwen Feng 	RTE_BUILD_BUG_ON(offsetof(struct rte_mbuf, data_off) <
126cb12e988SChengwen Feng 			 offsetof(struct rte_mbuf, rearm_data));
127cb12e988SChengwen Feng 	RTE_BUILD_BUG_ON(offsetof(struct rte_mbuf, refcnt) <
128cb12e988SChengwen Feng 			 offsetof(struct rte_mbuf, rearm_data));
129cb12e988SChengwen Feng 	RTE_BUILD_BUG_ON(offsetof(struct rte_mbuf, nb_segs) <
130cb12e988SChengwen Feng 			 offsetof(struct rte_mbuf, rearm_data));
131cb12e988SChengwen Feng 	RTE_BUILD_BUG_ON(offsetof(struct rte_mbuf, port) <
132cb12e988SChengwen Feng 			 offsetof(struct rte_mbuf, rearm_data));
133cb12e988SChengwen Feng 	RTE_BUILD_BUG_ON(offsetof(struct rte_mbuf, data_off) -
134cb12e988SChengwen Feng 			 offsetof(struct rte_mbuf, rearm_data) > 6);
135cb12e988SChengwen Feng 	RTE_BUILD_BUG_ON(offsetof(struct rte_mbuf, refcnt) -
136cb12e988SChengwen Feng 			 offsetof(struct rte_mbuf, rearm_data) > 6);
137cb12e988SChengwen Feng 	RTE_BUILD_BUG_ON(offsetof(struct rte_mbuf, nb_segs) -
138cb12e988SChengwen Feng 			 offsetof(struct rte_mbuf, rearm_data) > 6);
139cb12e988SChengwen Feng 	RTE_BUILD_BUG_ON(offsetof(struct rte_mbuf, port) -
140cb12e988SChengwen Feng 			 offsetof(struct rte_mbuf, rearm_data) > 6);
141cb12e988SChengwen Feng 
142a3d4f4d2SWei Hu (Xavier) 	/* prevent compiler reordering: rearm_data covers previous fields */
143a3d4f4d2SWei Hu (Xavier) 	rte_compiler_barrier();
144a3d4f4d2SWei Hu (Xavier) 	p = (uintptr_t)&mb_def.rearm_data;
145a3d4f4d2SWei Hu (Xavier) 	rxq->mbuf_initializer = *(uint64_t *)p;
146a3d4f4d2SWei Hu (Xavier) }
147a3d4f4d2SWei Hu (Xavier) 
148a3d4f4d2SWei Hu (Xavier) void
hns3_rxq_vec_setup(struct hns3_rx_queue * rxq)149a3d4f4d2SWei Hu (Xavier) hns3_rxq_vec_setup(struct hns3_rx_queue *rxq)
150a3d4f4d2SWei Hu (Xavier) {
151a3d4f4d2SWei Hu (Xavier) 	struct hns3_entry *sw_ring = &rxq->sw_ring[rxq->nb_rx_desc];
152a3d4f4d2SWei Hu (Xavier) 	unsigned int i;
153a3d4f4d2SWei Hu (Xavier) 
154a3d4f4d2SWei Hu (Xavier) 	memset(&rxq->rx_ring[rxq->nb_rx_desc], 0,
155a3d4f4d2SWei Hu (Xavier) 		sizeof(struct hns3_desc) * HNS3_DEFAULT_RX_BURST);
156a3d4f4d2SWei Hu (Xavier) 
157a3d4f4d2SWei Hu (Xavier) 	memset(&rxq->fake_mbuf, 0, sizeof(rxq->fake_mbuf));
158a3d4f4d2SWei Hu (Xavier) 	for (i = 0; i < HNS3_DEFAULT_RX_BURST; i++)
159a3d4f4d2SWei Hu (Xavier) 		sw_ring[i].mbuf = &rxq->fake_mbuf;
160a3d4f4d2SWei Hu (Xavier) 
161a3d4f4d2SWei Hu (Xavier) 	hns3_rxq_vec_setup_rearm_data(rxq);
162a3d4f4d2SWei Hu (Xavier) 
163a3d4f4d2SWei Hu (Xavier) 	memset(rxq->offset_table, 0, sizeof(rxq->offset_table));
164a3d4f4d2SWei Hu (Xavier) }
165a3d4f4d2SWei Hu (Xavier) 
166a3d4f4d2SWei Hu (Xavier) static int
hns3_rxq_vec_check(struct hns3_rx_queue * rxq,void * arg)167a3d4f4d2SWei Hu (Xavier) hns3_rxq_vec_check(struct hns3_rx_queue *rxq, void *arg)
168a3d4f4d2SWei Hu (Xavier) {
169a3d4f4d2SWei Hu (Xavier) 	uint32_t min_vec_bds = HNS3_DEFAULT_RXQ_REARM_THRESH +
170a3d4f4d2SWei Hu (Xavier) 				HNS3_DEFAULT_RX_BURST;
171a3d4f4d2SWei Hu (Xavier) 
172a3d4f4d2SWei Hu (Xavier) 	if (rxq->nb_rx_desc < min_vec_bds)
173a3d4f4d2SWei Hu (Xavier) 		return -ENOTSUP;
174a3d4f4d2SWei Hu (Xavier) 
175a3d4f4d2SWei Hu (Xavier) 	if (rxq->nb_rx_desc % HNS3_DEFAULT_RXQ_REARM_THRESH)
176a3d4f4d2SWei Hu (Xavier) 		return -ENOTSUP;
177a3d4f4d2SWei Hu (Xavier) 
178a3d4f4d2SWei Hu (Xavier) 	RTE_SET_USED(arg);
179a3d4f4d2SWei Hu (Xavier) 	return 0;
180a3d4f4d2SWei Hu (Xavier) }
181a3d4f4d2SWei Hu (Xavier) 
182a3d4f4d2SWei Hu (Xavier) int
hns3_rx_check_vec_support(struct rte_eth_dev * dev)183a3d4f4d2SWei Hu (Xavier) hns3_rx_check_vec_support(struct rte_eth_dev *dev)
184a3d4f4d2SWei Hu (Xavier) {
185a3d4f4d2SWei Hu (Xavier) 	struct rte_eth_rxmode *rxmode = &dev->data->dev_conf.rxmode;
186295968d1SFerruh Yigit 	uint64_t offloads_mask = RTE_ETH_RX_OFFLOAD_TCP_LRO |
187*4ac14c1dSHuisong Li 				 RTE_ETH_RX_OFFLOAD_VLAN |
188*4ac14c1dSHuisong Li 				 RTE_ETH_RX_OFFLOAD_TIMESTAMP;
18938b539d9SMin Hu (Connor) 
190a3d4f4d2SWei Hu (Xavier) 	if (dev->data->scattered_rx)
191a3d4f4d2SWei Hu (Xavier) 		return -ENOTSUP;
192a3d4f4d2SWei Hu (Xavier) 
193a3d4f4d2SWei Hu (Xavier) 	if (rxmode->offloads & offloads_mask)
194a3d4f4d2SWei Hu (Xavier) 		return -ENOTSUP;
195a3d4f4d2SWei Hu (Xavier) 
196a3d4f4d2SWei Hu (Xavier) 	if (hns3_rxq_iterate(dev, hns3_rxq_vec_check, NULL) != 0)
197a3d4f4d2SWei Hu (Xavier) 		return -ENOTSUP;
198a3d4f4d2SWei Hu (Xavier) 
199a3d4f4d2SWei Hu (Xavier) 	return 0;
200a3d4f4d2SWei Hu (Xavier) }
201