xref: /dpdk/drivers/net/mlx5/mlx5_rxtx.c (revision 6703d836346fa32a59cac8264efae5ea6acfbb7a)
12e22920bSAdrien Mazarguil /*-
22e22920bSAdrien Mazarguil  *   BSD LICENSE
32e22920bSAdrien Mazarguil  *
42e22920bSAdrien Mazarguil  *   Copyright 2015 6WIND S.A.
52e22920bSAdrien Mazarguil  *   Copyright 2015 Mellanox.
62e22920bSAdrien Mazarguil  *
72e22920bSAdrien Mazarguil  *   Redistribution and use in source and binary forms, with or without
82e22920bSAdrien Mazarguil  *   modification, are permitted provided that the following conditions
92e22920bSAdrien Mazarguil  *   are met:
102e22920bSAdrien Mazarguil  *
112e22920bSAdrien Mazarguil  *     * Redistributions of source code must retain the above copyright
122e22920bSAdrien Mazarguil  *       notice, this list of conditions and the following disclaimer.
132e22920bSAdrien Mazarguil  *     * Redistributions in binary form must reproduce the above copyright
142e22920bSAdrien Mazarguil  *       notice, this list of conditions and the following disclaimer in
152e22920bSAdrien Mazarguil  *       the documentation and/or other materials provided with the
162e22920bSAdrien Mazarguil  *       distribution.
172e22920bSAdrien Mazarguil  *     * Neither the name of 6WIND S.A. nor the names of its
182e22920bSAdrien Mazarguil  *       contributors may be used to endorse or promote products derived
192e22920bSAdrien Mazarguil  *       from this software without specific prior written permission.
202e22920bSAdrien Mazarguil  *
212e22920bSAdrien Mazarguil  *   THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
222e22920bSAdrien Mazarguil  *   "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
232e22920bSAdrien Mazarguil  *   LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
242e22920bSAdrien Mazarguil  *   A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
252e22920bSAdrien Mazarguil  *   OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
262e22920bSAdrien Mazarguil  *   SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
272e22920bSAdrien Mazarguil  *   LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
282e22920bSAdrien Mazarguil  *   DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
292e22920bSAdrien Mazarguil  *   THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
302e22920bSAdrien Mazarguil  *   (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
312e22920bSAdrien Mazarguil  *   OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
322e22920bSAdrien Mazarguil  */
332e22920bSAdrien Mazarguil 
342e22920bSAdrien Mazarguil #include <assert.h>
352e22920bSAdrien Mazarguil #include <stdint.h>
362e22920bSAdrien Mazarguil #include <string.h>
372e22920bSAdrien Mazarguil #include <stdlib.h>
382e22920bSAdrien Mazarguil 
392e22920bSAdrien Mazarguil /* Verbs header. */
402e22920bSAdrien Mazarguil /* ISO C doesn't support unnamed structs/unions, disabling -pedantic. */
412e22920bSAdrien Mazarguil #ifdef PEDANTIC
42fc5b160fSBruce Richardson #pragma GCC diagnostic ignored "-Wpedantic"
432e22920bSAdrien Mazarguil #endif
442e22920bSAdrien Mazarguil #include <infiniband/verbs.h>
456218063bSNélio Laranjeiro #include <infiniband/mlx5_hw.h>
466218063bSNélio Laranjeiro #include <infiniband/arch.h>
472e22920bSAdrien Mazarguil #ifdef PEDANTIC
48fc5b160fSBruce Richardson #pragma GCC diagnostic error "-Wpedantic"
492e22920bSAdrien Mazarguil #endif
502e22920bSAdrien Mazarguil 
512e22920bSAdrien Mazarguil /* DPDK headers don't like -pedantic. */
522e22920bSAdrien Mazarguil #ifdef PEDANTIC
53fc5b160fSBruce Richardson #pragma GCC diagnostic ignored "-Wpedantic"
542e22920bSAdrien Mazarguil #endif
552e22920bSAdrien Mazarguil #include <rte_mbuf.h>
562e22920bSAdrien Mazarguil #include <rte_mempool.h>
572e22920bSAdrien Mazarguil #include <rte_prefetch.h>
582e22920bSAdrien Mazarguil #include <rte_common.h>
592e22920bSAdrien Mazarguil #include <rte_branch_prediction.h>
606218063bSNélio Laranjeiro #include <rte_ether.h>
612e22920bSAdrien Mazarguil #ifdef PEDANTIC
62fc5b160fSBruce Richardson #pragma GCC diagnostic error "-Wpedantic"
632e22920bSAdrien Mazarguil #endif
642e22920bSAdrien Mazarguil 
652e22920bSAdrien Mazarguil #include "mlx5.h"
662e22920bSAdrien Mazarguil #include "mlx5_utils.h"
672e22920bSAdrien Mazarguil #include "mlx5_rxtx.h"
68f3db9489SYaacov Hazan #include "mlx5_autoconf.h"
692e22920bSAdrien Mazarguil #include "mlx5_defs.h"
706218063bSNélio Laranjeiro #include "mlx5_prm.h"
716218063bSNélio Laranjeiro 
72ff1807a3SNélio Laranjeiro static inline int
73ff1807a3SNélio Laranjeiro check_cqe(volatile struct mlx5_cqe *cqe,
74ff1807a3SNélio Laranjeiro 	  unsigned int cqes_n, const uint16_t ci)
75ff1807a3SNélio Laranjeiro 	  __attribute__((always_inline));
76ff1807a3SNélio Laranjeiro 
7761b09ae4SNélio Laranjeiro static inline void
7861b09ae4SNélio Laranjeiro txq_complete(struct txq *txq) __attribute__((always_inline));
7961b09ae4SNélio Laranjeiro 
80ff1807a3SNélio Laranjeiro static inline uint32_t
81ff1807a3SNélio Laranjeiro txq_mp2mr(struct txq *txq, struct rte_mempool *mp)
82ff1807a3SNélio Laranjeiro 	__attribute__((always_inline));
83ff1807a3SNélio Laranjeiro 
84ff1807a3SNélio Laranjeiro static inline void
8530807f62SNélio Laranjeiro mlx5_tx_dbrec(struct txq *txq, volatile struct mlx5_wqe *wqe)
8630807f62SNélio Laranjeiro 	__attribute__((always_inline));
87ff1807a3SNélio Laranjeiro 
88ff1807a3SNélio Laranjeiro static inline uint32_t
89ff1807a3SNélio Laranjeiro rxq_cq_to_pkt_type(volatile struct mlx5_cqe *cqe)
90ff1807a3SNélio Laranjeiro 	__attribute__((always_inline));
91ff1807a3SNélio Laranjeiro 
92ff1807a3SNélio Laranjeiro static inline int
93ff1807a3SNélio Laranjeiro mlx5_rx_poll_len(struct rxq *rxq, volatile struct mlx5_cqe *cqe,
94ff1807a3SNélio Laranjeiro 		 uint16_t cqe_cnt, uint32_t *rss_hash)
95ff1807a3SNélio Laranjeiro 		 __attribute__((always_inline));
96ff1807a3SNélio Laranjeiro 
97ff1807a3SNélio Laranjeiro static inline uint32_t
98ff1807a3SNélio Laranjeiro rxq_cq_to_ol_flags(struct rxq *rxq, volatile struct mlx5_cqe *cqe)
99ff1807a3SNélio Laranjeiro 		   __attribute__((always_inline));
100ff1807a3SNélio Laranjeiro 
10199c12dccSNélio Laranjeiro #ifndef NDEBUG
10299c12dccSNélio Laranjeiro 
10399c12dccSNélio Laranjeiro /**
10499c12dccSNélio Laranjeiro  * Verify or set magic value in CQE.
10599c12dccSNélio Laranjeiro  *
10699c12dccSNélio Laranjeiro  * @param cqe
10799c12dccSNélio Laranjeiro  *   Pointer to CQE.
10899c12dccSNélio Laranjeiro  *
10999c12dccSNélio Laranjeiro  * @return
11099c12dccSNélio Laranjeiro  *   0 the first time.
11199c12dccSNélio Laranjeiro  */
11299c12dccSNélio Laranjeiro static inline int
11397267b8eSNelio Laranjeiro check_cqe_seen(volatile struct mlx5_cqe *cqe)
11499c12dccSNélio Laranjeiro {
11599c12dccSNélio Laranjeiro 	static const uint8_t magic[] = "seen";
116ea3bc3b1SNélio Laranjeiro 	volatile uint8_t (*buf)[sizeof(cqe->rsvd0)] = &cqe->rsvd0;
11799c12dccSNélio Laranjeiro 	int ret = 1;
11899c12dccSNélio Laranjeiro 	unsigned int i;
11999c12dccSNélio Laranjeiro 
12099c12dccSNélio Laranjeiro 	for (i = 0; i < sizeof(magic) && i < sizeof(*buf); ++i)
12199c12dccSNélio Laranjeiro 		if (!ret || (*buf)[i] != magic[i]) {
12299c12dccSNélio Laranjeiro 			ret = 0;
12399c12dccSNélio Laranjeiro 			(*buf)[i] = magic[i];
12499c12dccSNélio Laranjeiro 		}
12599c12dccSNélio Laranjeiro 	return ret;
12699c12dccSNélio Laranjeiro }
12799c12dccSNélio Laranjeiro 
12899c12dccSNélio Laranjeiro #endif /* NDEBUG */
1296218063bSNélio Laranjeiro 
13099c12dccSNélio Laranjeiro /**
13199c12dccSNélio Laranjeiro  * Check whether CQE is valid.
13299c12dccSNélio Laranjeiro  *
13399c12dccSNélio Laranjeiro  * @param cqe
13499c12dccSNélio Laranjeiro  *   Pointer to CQE.
13599c12dccSNélio Laranjeiro  * @param cqes_n
13699c12dccSNélio Laranjeiro  *   Size of completion queue.
13799c12dccSNélio Laranjeiro  * @param ci
13899c12dccSNélio Laranjeiro  *   Consumer index.
13999c12dccSNélio Laranjeiro  *
14099c12dccSNélio Laranjeiro  * @return
14199c12dccSNélio Laranjeiro  *   0 on success, 1 on failure.
14299c12dccSNélio Laranjeiro  */
14399c12dccSNélio Laranjeiro static inline int
14497267b8eSNelio Laranjeiro check_cqe(volatile struct mlx5_cqe *cqe,
14599c12dccSNélio Laranjeiro 	  unsigned int cqes_n, const uint16_t ci)
1466218063bSNélio Laranjeiro {
14799c12dccSNélio Laranjeiro 	uint16_t idx = ci & cqes_n;
14899c12dccSNélio Laranjeiro 	uint8_t op_own = cqe->op_own;
14999c12dccSNélio Laranjeiro 	uint8_t op_owner = MLX5_CQE_OWNER(op_own);
15099c12dccSNélio Laranjeiro 	uint8_t op_code = MLX5_CQE_OPCODE(op_own);
1516218063bSNélio Laranjeiro 
15299c12dccSNélio Laranjeiro 	if (unlikely((op_owner != (!!(idx))) || (op_code == MLX5_CQE_INVALID)))
15399c12dccSNélio Laranjeiro 		return 1; /* No CQE. */
15499c12dccSNélio Laranjeiro #ifndef NDEBUG
15599c12dccSNélio Laranjeiro 	if ((op_code == MLX5_CQE_RESP_ERR) ||
15699c12dccSNélio Laranjeiro 	    (op_code == MLX5_CQE_REQ_ERR)) {
15799c12dccSNélio Laranjeiro 		volatile struct mlx5_err_cqe *err_cqe = (volatile void *)cqe;
15899c12dccSNélio Laranjeiro 		uint8_t syndrome = err_cqe->syndrome;
15999c12dccSNélio Laranjeiro 
16099c12dccSNélio Laranjeiro 		if ((syndrome == MLX5_CQE_SYNDROME_LOCAL_LENGTH_ERR) ||
16199c12dccSNélio Laranjeiro 		    (syndrome == MLX5_CQE_SYNDROME_REMOTE_ABORTED_ERR))
16299c12dccSNélio Laranjeiro 			return 0;
16397267b8eSNelio Laranjeiro 		if (!check_cqe_seen(cqe))
16499c12dccSNélio Laranjeiro 			ERROR("unexpected CQE error %u (0x%02x)"
16599c12dccSNélio Laranjeiro 			      " syndrome 0x%02x",
16699c12dccSNélio Laranjeiro 			      op_code, op_code, syndrome);
16799c12dccSNélio Laranjeiro 		return 1;
16899c12dccSNélio Laranjeiro 	} else if ((op_code != MLX5_CQE_RESP_SEND) &&
16999c12dccSNélio Laranjeiro 		   (op_code != MLX5_CQE_REQ)) {
17097267b8eSNelio Laranjeiro 		if (!check_cqe_seen(cqe))
17199c12dccSNélio Laranjeiro 			ERROR("unexpected CQE opcode %u (0x%02x)",
17299c12dccSNélio Laranjeiro 			      op_code, op_code);
17399c12dccSNélio Laranjeiro 		return 1;
1746218063bSNélio Laranjeiro 	}
17599c12dccSNélio Laranjeiro #endif /* NDEBUG */
17699c12dccSNélio Laranjeiro 	return 0;
1776218063bSNélio Laranjeiro }
1782e22920bSAdrien Mazarguil 
179fdcb0f53SNélio Laranjeiro /**
180fdcb0f53SNélio Laranjeiro  * Return the address of the WQE.
181fdcb0f53SNélio Laranjeiro  *
182fdcb0f53SNélio Laranjeiro  * @param txq
183fdcb0f53SNélio Laranjeiro  *   Pointer to TX queue structure.
184fdcb0f53SNélio Laranjeiro  * @param  wqe_ci
185fdcb0f53SNélio Laranjeiro  *   WQE consumer index.
186fdcb0f53SNélio Laranjeiro  *
187fdcb0f53SNélio Laranjeiro  * @return
188fdcb0f53SNélio Laranjeiro  *   WQE address.
189fdcb0f53SNélio Laranjeiro  */
190fdcb0f53SNélio Laranjeiro static inline uintptr_t *
191fdcb0f53SNélio Laranjeiro tx_mlx5_wqe(struct txq *txq, uint16_t ci)
192fdcb0f53SNélio Laranjeiro {
193fdcb0f53SNélio Laranjeiro 	ci &= ((1 << txq->wqe_n) - 1);
194fdcb0f53SNélio Laranjeiro 	return (uintptr_t *)((uintptr_t)txq->wqes + ci * MLX5_WQE_SIZE);
195fdcb0f53SNélio Laranjeiro }
196fdcb0f53SNélio Laranjeiro 
1972e22920bSAdrien Mazarguil /**
1986ce84bd8SYongseok Koh  * Return the size of tailroom of WQ.
1996ce84bd8SYongseok Koh  *
2006ce84bd8SYongseok Koh  * @param txq
2016ce84bd8SYongseok Koh  *   Pointer to TX queue structure.
2026ce84bd8SYongseok Koh  * @param addr
2036ce84bd8SYongseok Koh  *   Pointer to tail of WQ.
2046ce84bd8SYongseok Koh  *
2056ce84bd8SYongseok Koh  * @return
2066ce84bd8SYongseok Koh  *   Size of tailroom.
2076ce84bd8SYongseok Koh  */
2086ce84bd8SYongseok Koh static inline size_t
2096ce84bd8SYongseok Koh tx_mlx5_wq_tailroom(struct txq *txq, void *addr)
2106ce84bd8SYongseok Koh {
2116ce84bd8SYongseok Koh 	size_t tailroom;
2126ce84bd8SYongseok Koh 	tailroom = (uintptr_t)(txq->wqes) +
2136ce84bd8SYongseok Koh 		   (1 << txq->wqe_n) * MLX5_WQE_SIZE -
2146ce84bd8SYongseok Koh 		   (uintptr_t)addr;
2156ce84bd8SYongseok Koh 	return tailroom;
2166ce84bd8SYongseok Koh }
2176ce84bd8SYongseok Koh 
2186ce84bd8SYongseok Koh /**
2196ce84bd8SYongseok Koh  * Copy data to tailroom of circular queue.
2206ce84bd8SYongseok Koh  *
2216ce84bd8SYongseok Koh  * @param dst
2226ce84bd8SYongseok Koh  *   Pointer to destination.
2236ce84bd8SYongseok Koh  * @param src
2246ce84bd8SYongseok Koh  *   Pointer to source.
2256ce84bd8SYongseok Koh  * @param n
2266ce84bd8SYongseok Koh  *   Number of bytes to copy.
2276ce84bd8SYongseok Koh  * @param base
2286ce84bd8SYongseok Koh  *   Pointer to head of queue.
2296ce84bd8SYongseok Koh  * @param tailroom
2306ce84bd8SYongseok Koh  *   Size of tailroom from dst.
2316ce84bd8SYongseok Koh  *
2326ce84bd8SYongseok Koh  * @return
2336ce84bd8SYongseok Koh  *   Pointer after copied data.
2346ce84bd8SYongseok Koh  */
2356ce84bd8SYongseok Koh static inline void *
2366ce84bd8SYongseok Koh mlx5_copy_to_wq(void *dst, const void *src, size_t n,
2376ce84bd8SYongseok Koh 		void *base, size_t tailroom)
2386ce84bd8SYongseok Koh {
2396ce84bd8SYongseok Koh 	void *ret;
2406ce84bd8SYongseok Koh 
2416ce84bd8SYongseok Koh 	if (n > tailroom) {
2426ce84bd8SYongseok Koh 		rte_memcpy(dst, src, tailroom);
2436ce84bd8SYongseok Koh 		rte_memcpy(base, (void *)((uintptr_t)src + tailroom),
2446ce84bd8SYongseok Koh 			   n - tailroom);
2456ce84bd8SYongseok Koh 		ret = (uint8_t *)base + n - tailroom;
2466ce84bd8SYongseok Koh 	} else {
2476ce84bd8SYongseok Koh 		rte_memcpy(dst, src, n);
2486ce84bd8SYongseok Koh 		ret = (n == tailroom) ? base : (uint8_t *)dst + n;
2496ce84bd8SYongseok Koh 	}
2506ce84bd8SYongseok Koh 	return ret;
2516ce84bd8SYongseok Koh }
2526ce84bd8SYongseok Koh 
2536ce84bd8SYongseok Koh /**
2542e22920bSAdrien Mazarguil  * Manage TX completions.
2552e22920bSAdrien Mazarguil  *
2562e22920bSAdrien Mazarguil  * When sending a burst, mlx5_tx_burst() posts several WRs.
2572e22920bSAdrien Mazarguil  *
2582e22920bSAdrien Mazarguil  * @param txq
2592e22920bSAdrien Mazarguil  *   Pointer to TX queue structure.
2602e22920bSAdrien Mazarguil  */
261a6ca35aaSNélio Laranjeiro static inline void
2622e22920bSAdrien Mazarguil txq_complete(struct txq *txq)
2632e22920bSAdrien Mazarguil {
264b4b12e55SNélio Laranjeiro 	const unsigned int elts_n = 1 << txq->elts_n;
265e2f116eeSNélio Laranjeiro 	const unsigned int cqe_n = 1 << txq->cqe_n;
26699c12dccSNélio Laranjeiro 	const unsigned int cqe_cnt = cqe_n - 1;
2671d88ba17SNélio Laranjeiro 	uint16_t elts_free = txq->elts_tail;
2681d88ba17SNélio Laranjeiro 	uint16_t elts_tail;
2691d88ba17SNélio Laranjeiro 	uint16_t cq_ci = txq->cq_ci;
27097267b8eSNelio Laranjeiro 	volatile struct mlx5_cqe *cqe = NULL;
271fdcb0f53SNélio Laranjeiro 	volatile struct mlx5_wqe_ctrl *ctrl;
2722e22920bSAdrien Mazarguil 
27399c12dccSNélio Laranjeiro 	do {
27497267b8eSNelio Laranjeiro 		volatile struct mlx5_cqe *tmp;
2751d88ba17SNélio Laranjeiro 
27697267b8eSNelio Laranjeiro 		tmp = &(*txq->cqes)[cq_ci & cqe_cnt];
27797267b8eSNelio Laranjeiro 		if (check_cqe(tmp, cqe_n, cq_ci))
2781d88ba17SNélio Laranjeiro 			break;
279c305090bSAdrien Mazarguil 		cqe = tmp;
28099c12dccSNélio Laranjeiro #ifndef NDEBUG
28199c12dccSNélio Laranjeiro 		if (MLX5_CQE_FORMAT(cqe->op_own) == MLX5_COMPRESSED) {
28297267b8eSNelio Laranjeiro 			if (!check_cqe_seen(cqe))
28399c12dccSNélio Laranjeiro 				ERROR("unexpected compressed CQE, TX stopped");
28499c12dccSNélio Laranjeiro 			return;
2852e22920bSAdrien Mazarguil 		}
28699c12dccSNélio Laranjeiro 		if ((MLX5_CQE_OPCODE(cqe->op_own) == MLX5_CQE_RESP_ERR) ||
28799c12dccSNélio Laranjeiro 		    (MLX5_CQE_OPCODE(cqe->op_own) == MLX5_CQE_REQ_ERR)) {
28897267b8eSNelio Laranjeiro 			if (!check_cqe_seen(cqe))
28999c12dccSNélio Laranjeiro 				ERROR("unexpected error CQE, TX stopped");
29099c12dccSNélio Laranjeiro 			return;
29199c12dccSNélio Laranjeiro 		}
29299c12dccSNélio Laranjeiro #endif /* NDEBUG */
29399c12dccSNélio Laranjeiro 		++cq_ci;
29499c12dccSNélio Laranjeiro 	} while (1);
295c305090bSAdrien Mazarguil 	if (unlikely(cqe == NULL))
2961d88ba17SNélio Laranjeiro 		return;
297f04f1d51SNélio Laranjeiro 	txq->wqe_pi = ntohs(cqe->wqe_counter);
298fdcb0f53SNélio Laranjeiro 	ctrl = (volatile struct mlx5_wqe_ctrl *)
299f04f1d51SNélio Laranjeiro 		tx_mlx5_wqe(txq, txq->wqe_pi);
300fdcb0f53SNélio Laranjeiro 	elts_tail = ctrl->ctrl3;
301a821d09dSNélio Laranjeiro 	assert(elts_tail < (1 << txq->wqe_n));
3021d88ba17SNélio Laranjeiro 	/* Free buffers. */
303c305090bSAdrien Mazarguil 	while (elts_free != elts_tail) {
3041d88ba17SNélio Laranjeiro 		struct rte_mbuf *elt = (*txq->elts)[elts_free];
305a859e8a9SNelio Laranjeiro 		unsigned int elts_free_next =
3061d88ba17SNélio Laranjeiro 			(elts_free + 1) & (elts_n - 1);
3071d88ba17SNélio Laranjeiro 		struct rte_mbuf *elt_next = (*txq->elts)[elts_free_next];
308a859e8a9SNelio Laranjeiro 
309b185e63fSAdrien Mazarguil #ifndef NDEBUG
310b185e63fSAdrien Mazarguil 		/* Poisoning. */
3111d88ba17SNélio Laranjeiro 		memset(&(*txq->elts)[elts_free],
3121d88ba17SNélio Laranjeiro 		       0x66,
3131d88ba17SNélio Laranjeiro 		       sizeof((*txq->elts)[elts_free]));
314b185e63fSAdrien Mazarguil #endif
3151d88ba17SNélio Laranjeiro 		RTE_MBUF_PREFETCH_TO_FREE(elt_next);
3161d88ba17SNélio Laranjeiro 		/* Only one segment needs to be freed. */
3171d88ba17SNélio Laranjeiro 		rte_pktmbuf_free_seg(elt);
318a859e8a9SNelio Laranjeiro 		elts_free = elts_free_next;
319c305090bSAdrien Mazarguil 	}
3201d88ba17SNélio Laranjeiro 	txq->cq_ci = cq_ci;
3212e22920bSAdrien Mazarguil 	txq->elts_tail = elts_tail;
3221d88ba17SNélio Laranjeiro 	/* Update the consumer index. */
3231d88ba17SNélio Laranjeiro 	rte_wmb();
3241d88ba17SNélio Laranjeiro 	*txq->cq_db = htonl(cq_ci);
3252e22920bSAdrien Mazarguil }
3262e22920bSAdrien Mazarguil 
3272e22920bSAdrien Mazarguil /**
3288340392eSAdrien Mazarguil  * Get Memory Pool (MP) from mbuf. If mbuf is indirect, the pool from which
3298340392eSAdrien Mazarguil  * the cloned mbuf is allocated is returned instead.
3308340392eSAdrien Mazarguil  *
3318340392eSAdrien Mazarguil  * @param buf
3328340392eSAdrien Mazarguil  *   Pointer to mbuf.
3338340392eSAdrien Mazarguil  *
3348340392eSAdrien Mazarguil  * @return
3358340392eSAdrien Mazarguil  *   Memory pool where data is located for given mbuf.
3368340392eSAdrien Mazarguil  */
3378340392eSAdrien Mazarguil static struct rte_mempool *
3388340392eSAdrien Mazarguil txq_mb2mp(struct rte_mbuf *buf)
3398340392eSAdrien Mazarguil {
3408340392eSAdrien Mazarguil 	if (unlikely(RTE_MBUF_INDIRECT(buf)))
3418340392eSAdrien Mazarguil 		return rte_mbuf_from_indirect(buf)->pool;
3428340392eSAdrien Mazarguil 	return buf->pool;
3438340392eSAdrien Mazarguil }
3448340392eSAdrien Mazarguil 
3458340392eSAdrien Mazarguil /**
3462e22920bSAdrien Mazarguil  * Get Memory Region (MR) <-> Memory Pool (MP) association from txq->mp2mr[].
3472e22920bSAdrien Mazarguil  * Add MP to txq->mp2mr[] if it's not registered yet. If mp2mr[] is full,
3482e22920bSAdrien Mazarguil  * remove an entry first.
3492e22920bSAdrien Mazarguil  *
3502e22920bSAdrien Mazarguil  * @param txq
3512e22920bSAdrien Mazarguil  *   Pointer to TX queue structure.
3522e22920bSAdrien Mazarguil  * @param[in] mp
3532e22920bSAdrien Mazarguil  *   Memory Pool for which a Memory Region lkey must be returned.
3542e22920bSAdrien Mazarguil  *
3552e22920bSAdrien Mazarguil  * @return
3562e22920bSAdrien Mazarguil  *   mr->lkey on success, (uint32_t)-1 on failure.
3572e22920bSAdrien Mazarguil  */
358491770faSNélio Laranjeiro static inline uint32_t
359d1d914ebSOlivier Matz txq_mp2mr(struct txq *txq, struct rte_mempool *mp)
3602e22920bSAdrien Mazarguil {
3612e22920bSAdrien Mazarguil 	unsigned int i;
362491770faSNélio Laranjeiro 	uint32_t lkey = (uint32_t)-1;
3632e22920bSAdrien Mazarguil 
3642e22920bSAdrien Mazarguil 	for (i = 0; (i != RTE_DIM(txq->mp2mr)); ++i) {
3652e22920bSAdrien Mazarguil 		if (unlikely(txq->mp2mr[i].mp == NULL)) {
3662e22920bSAdrien Mazarguil 			/* Unknown MP, add a new MR for it. */
3672e22920bSAdrien Mazarguil 			break;
3682e22920bSAdrien Mazarguil 		}
3692e22920bSAdrien Mazarguil 		if (txq->mp2mr[i].mp == mp) {
3702e22920bSAdrien Mazarguil 			assert(txq->mp2mr[i].lkey != (uint32_t)-1);
3711d88ba17SNélio Laranjeiro 			assert(htonl(txq->mp2mr[i].mr->lkey) ==
3721d88ba17SNélio Laranjeiro 			       txq->mp2mr[i].lkey);
373491770faSNélio Laranjeiro 			lkey = txq->mp2mr[i].lkey;
374491770faSNélio Laranjeiro 			break;
3752e22920bSAdrien Mazarguil 		}
3762e22920bSAdrien Mazarguil 	}
377491770faSNélio Laranjeiro 	if (unlikely(lkey == (uint32_t)-1))
378491770faSNélio Laranjeiro 		lkey = txq_mp2mr_reg(txq, mp, i);
379491770faSNélio Laranjeiro 	return lkey;
3800a3b350dSOlga Shern }
3810a3b350dSOlga Shern 
382e192ef80SYaacov Hazan /**
3831d88ba17SNélio Laranjeiro  * Ring TX queue doorbell.
3841d88ba17SNélio Laranjeiro  *
3851d88ba17SNélio Laranjeiro  * @param txq
3861d88ba17SNélio Laranjeiro  *   Pointer to TX queue structure.
38730807f62SNélio Laranjeiro  * @param wqe
38830807f62SNélio Laranjeiro  *   Pointer to the last WQE posted in the NIC.
3891d88ba17SNélio Laranjeiro  */
3901d88ba17SNélio Laranjeiro static inline void
39130807f62SNélio Laranjeiro mlx5_tx_dbrec(struct txq *txq, volatile struct mlx5_wqe *wqe)
3921d88ba17SNélio Laranjeiro {
39330807f62SNélio Laranjeiro 	uint64_t *dst = (uint64_t *)((uintptr_t)txq->bf_reg);
39430807f62SNélio Laranjeiro 	volatile uint64_t *src = ((volatile uint64_t *)wqe);
39530807f62SNélio Laranjeiro 
3961d88ba17SNélio Laranjeiro 	rte_wmb();
3971d88ba17SNélio Laranjeiro 	*txq->qp_db = htonl(txq->wqe_ci);
3981d88ba17SNélio Laranjeiro 	/* Ensure ordering between DB record and BF copy. */
3991d88ba17SNélio Laranjeiro 	rte_wmb();
40030807f62SNélio Laranjeiro 	*dst = *src;
4011d88ba17SNélio Laranjeiro }
402e192ef80SYaacov Hazan 
4031d88ba17SNélio Laranjeiro /**
4048788fec1SOlivier Matz  * DPDK callback to check the status of a tx descriptor.
4058788fec1SOlivier Matz  *
4068788fec1SOlivier Matz  * @param tx_queue
4078788fec1SOlivier Matz  *   The tx queue.
4088788fec1SOlivier Matz  * @param[in] offset
4098788fec1SOlivier Matz  *   The index of the descriptor in the ring.
4108788fec1SOlivier Matz  *
4118788fec1SOlivier Matz  * @return
4128788fec1SOlivier Matz  *   The status of the tx descriptor.
4138788fec1SOlivier Matz  */
4148788fec1SOlivier Matz int
4158788fec1SOlivier Matz mlx5_tx_descriptor_status(void *tx_queue, uint16_t offset)
4168788fec1SOlivier Matz {
4178788fec1SOlivier Matz 	struct txq *txq = tx_queue;
4188788fec1SOlivier Matz 	const unsigned int elts_n = 1 << txq->elts_n;
4198788fec1SOlivier Matz 	const unsigned int elts_cnt = elts_n - 1;
4208788fec1SOlivier Matz 	unsigned int used;
4218788fec1SOlivier Matz 
4228788fec1SOlivier Matz 	txq_complete(txq);
4238788fec1SOlivier Matz 	used = (txq->elts_head - txq->elts_tail) & elts_cnt;
4248788fec1SOlivier Matz 	if (offset < used)
4258788fec1SOlivier Matz 		return RTE_ETH_TX_DESC_FULL;
4268788fec1SOlivier Matz 	return RTE_ETH_TX_DESC_DONE;
4278788fec1SOlivier Matz }
4288788fec1SOlivier Matz 
4298788fec1SOlivier Matz /**
4308788fec1SOlivier Matz  * DPDK callback to check the status of a rx descriptor.
4318788fec1SOlivier Matz  *
4328788fec1SOlivier Matz  * @param rx_queue
4338788fec1SOlivier Matz  *   The rx queue.
4348788fec1SOlivier Matz  * @param[in] offset
4358788fec1SOlivier Matz  *   The index of the descriptor in the ring.
4368788fec1SOlivier Matz  *
4378788fec1SOlivier Matz  * @return
4388788fec1SOlivier Matz  *   The status of the tx descriptor.
4398788fec1SOlivier Matz  */
4408788fec1SOlivier Matz int
4418788fec1SOlivier Matz mlx5_rx_descriptor_status(void *rx_queue, uint16_t offset)
4428788fec1SOlivier Matz {
4438788fec1SOlivier Matz 	struct rxq *rxq = rx_queue;
4448788fec1SOlivier Matz 	struct rxq_zip *zip = &rxq->zip;
4458788fec1SOlivier Matz 	volatile struct mlx5_cqe *cqe;
4468788fec1SOlivier Matz 	const unsigned int cqe_n = (1 << rxq->cqe_n);
4478788fec1SOlivier Matz 	const unsigned int cqe_cnt = cqe_n - 1;
4488788fec1SOlivier Matz 	unsigned int cq_ci;
4498788fec1SOlivier Matz 	unsigned int used;
4508788fec1SOlivier Matz 
4518788fec1SOlivier Matz 	/* if we are processing a compressed cqe */
4528788fec1SOlivier Matz 	if (zip->ai) {
4538788fec1SOlivier Matz 		used = zip->cqe_cnt - zip->ca;
4548788fec1SOlivier Matz 		cq_ci = zip->cq_ci;
4558788fec1SOlivier Matz 	} else {
4568788fec1SOlivier Matz 		used = 0;
4578788fec1SOlivier Matz 		cq_ci = rxq->cq_ci;
4588788fec1SOlivier Matz 	}
4598788fec1SOlivier Matz 	cqe = &(*rxq->cqes)[cq_ci & cqe_cnt];
4608788fec1SOlivier Matz 	while (check_cqe(cqe, cqe_n, cq_ci) == 0) {
4618788fec1SOlivier Matz 		int8_t op_own;
4628788fec1SOlivier Matz 		unsigned int n;
4638788fec1SOlivier Matz 
4648788fec1SOlivier Matz 		op_own = cqe->op_own;
4658788fec1SOlivier Matz 		if (MLX5_CQE_FORMAT(op_own) == MLX5_COMPRESSED)
4668788fec1SOlivier Matz 			n = ntohl(cqe->byte_cnt);
4678788fec1SOlivier Matz 		else
4688788fec1SOlivier Matz 			n = 1;
4698788fec1SOlivier Matz 		cq_ci += n;
4708788fec1SOlivier Matz 		used += n;
4718788fec1SOlivier Matz 		cqe = &(*rxq->cqes)[cq_ci & cqe_cnt];
4728788fec1SOlivier Matz 	}
4738788fec1SOlivier Matz 	used = RTE_MIN(used, (1U << rxq->elts_n) - 1);
4748788fec1SOlivier Matz 	if (offset < used)
4758788fec1SOlivier Matz 		return RTE_ETH_RX_DESC_DONE;
4768788fec1SOlivier Matz 	return RTE_ETH_RX_DESC_AVAIL;
4778788fec1SOlivier Matz }
4788788fec1SOlivier Matz 
4798788fec1SOlivier Matz /**
4802e22920bSAdrien Mazarguil  * DPDK callback for TX.
4812e22920bSAdrien Mazarguil  *
4822e22920bSAdrien Mazarguil  * @param dpdk_txq
4832e22920bSAdrien Mazarguil  *   Generic pointer to TX queue structure.
4842e22920bSAdrien Mazarguil  * @param[in] pkts
4852e22920bSAdrien Mazarguil  *   Packets to transmit.
4862e22920bSAdrien Mazarguil  * @param pkts_n
4872e22920bSAdrien Mazarguil  *   Number of packets in array.
4882e22920bSAdrien Mazarguil  *
4892e22920bSAdrien Mazarguil  * @return
4902e22920bSAdrien Mazarguil  *   Number of packets successfully transmitted (<= pkts_n).
4912e22920bSAdrien Mazarguil  */
4922e22920bSAdrien Mazarguil uint16_t
4932e22920bSAdrien Mazarguil mlx5_tx_burst(void *dpdk_txq, struct rte_mbuf **pkts, uint16_t pkts_n)
4942e22920bSAdrien Mazarguil {
4952e22920bSAdrien Mazarguil 	struct txq *txq = (struct txq *)dpdk_txq;
4961d88ba17SNélio Laranjeiro 	uint16_t elts_head = txq->elts_head;
497b4b12e55SNélio Laranjeiro 	const unsigned int elts_n = 1 << txq->elts_n;
498c3d62cc9SAdrien Mazarguil 	unsigned int i = 0;
499a5bf6af9SAdrien Mazarguil 	unsigned int j = 0;
5003f13f8c2SShahaf Shuler 	unsigned int k = 0;
5012e22920bSAdrien Mazarguil 	unsigned int max;
502f04f1d51SNélio Laranjeiro 	uint16_t max_wqe;
503c305090bSAdrien Mazarguil 	unsigned int comp;
5049a7fa9f7SNélio Laranjeiro 	volatile struct mlx5_wqe_v *wqe = NULL;
5056579c27cSNélio Laranjeiro 	unsigned int segs_n = 0;
5066579c27cSNélio Laranjeiro 	struct rte_mbuf *buf = NULL;
5076579c27cSNélio Laranjeiro 	uint8_t *raw;
5082e22920bSAdrien Mazarguil 
5091d88ba17SNélio Laranjeiro 	if (unlikely(!pkts_n))
5101d88ba17SNélio Laranjeiro 		return 0;
5115e1d11ecSNelio Laranjeiro 	/* Prefetch first packet cacheline. */
512c3d62cc9SAdrien Mazarguil 	rte_prefetch0(*pkts);
5131d88ba17SNélio Laranjeiro 	/* Start processing. */
5142e22920bSAdrien Mazarguil 	txq_complete(txq);
5154f52bbfbSNelio Laranjeiro 	max = (elts_n - (elts_head - txq->elts_tail));
5162e22920bSAdrien Mazarguil 	if (max > elts_n)
5172e22920bSAdrien Mazarguil 		max -= elts_n;
518f04f1d51SNélio Laranjeiro 	max_wqe = (1u << txq->wqe_n) - (txq->wqe_ci - txq->wqe_pi);
519f04f1d51SNélio Laranjeiro 	if (unlikely(!max_wqe))
520f04f1d51SNélio Laranjeiro 		return 0;
521c3d62cc9SAdrien Mazarguil 	do {
5229a7fa9f7SNélio Laranjeiro 		volatile rte_v128u32_t *dseg = NULL;
523573f54afSNélio Laranjeiro 		uint32_t length;
5248688b2f8SNélio Laranjeiro 		unsigned int ds = 0;
5256579c27cSNélio Laranjeiro 		uintptr_t addr;
5269a7fa9f7SNélio Laranjeiro 		uint64_t naddr;
5270d637a34SNélio Laranjeiro 		uint16_t pkt_inline_sz = MLX5_WQE_DWORD_SIZE + 2;
5283f13f8c2SShahaf Shuler 		uint16_t tso_header_sz = 0;
529eef822ddSNélio Laranjeiro 		uint16_t ehdr;
5309a7fa9f7SNélio Laranjeiro 		uint8_t cs_flags = 0;
5313f13f8c2SShahaf Shuler 		uint64_t tso = 0;
5326579c27cSNélio Laranjeiro #ifdef MLX5_PMD_SOFT_COUNTERS
5336579c27cSNélio Laranjeiro 		uint32_t total_length = 0;
5346579c27cSNélio Laranjeiro #endif
5352e22920bSAdrien Mazarguil 
5366579c27cSNélio Laranjeiro 		/* first_seg */
5376579c27cSNélio Laranjeiro 		buf = *(pkts++);
5386579c27cSNélio Laranjeiro 		segs_n = buf->nb_segs;
539c3d62cc9SAdrien Mazarguil 		/*
540c3d62cc9SAdrien Mazarguil 		 * Make sure there is enough room to store this packet and
541c3d62cc9SAdrien Mazarguil 		 * that one ring entry remains unused.
542c3d62cc9SAdrien Mazarguil 		 */
543a5bf6af9SAdrien Mazarguil 		assert(segs_n);
544a5bf6af9SAdrien Mazarguil 		if (max < segs_n + 1)
545c3d62cc9SAdrien Mazarguil 			break;
546a5bf6af9SAdrien Mazarguil 		max -= segs_n;
5476579c27cSNélio Laranjeiro 		--segs_n;
5486579c27cSNélio Laranjeiro 		if (!segs_n)
549c3d62cc9SAdrien Mazarguil 			--pkts_n;
550f04f1d51SNélio Laranjeiro 		if (unlikely(--max_wqe == 0))
551f04f1d51SNélio Laranjeiro 			break;
5529a7fa9f7SNélio Laranjeiro 		wqe = (volatile struct mlx5_wqe_v *)
553fdcb0f53SNélio Laranjeiro 			tx_mlx5_wqe(txq, txq->wqe_ci);
554fdcb0f53SNélio Laranjeiro 		rte_prefetch0(tx_mlx5_wqe(txq, txq->wqe_ci + 1));
5556579c27cSNélio Laranjeiro 		if (pkts_n > 1)
556c3d62cc9SAdrien Mazarguil 			rte_prefetch0(*pkts);
5576579c27cSNélio Laranjeiro 		addr = rte_pktmbuf_mtod(buf, uintptr_t);
5582e22920bSAdrien Mazarguil 		length = DATA_LEN(buf);
559eef822ddSNélio Laranjeiro 		ehdr = (((uint8_t *)addr)[1] << 8) |
560eef822ddSNélio Laranjeiro 		       ((uint8_t *)addr)[0];
5616579c27cSNélio Laranjeiro #ifdef MLX5_PMD_SOFT_COUNTERS
5626579c27cSNélio Laranjeiro 		total_length = length;
5636579c27cSNélio Laranjeiro #endif
564959be52eSNélio Laranjeiro 		if (length < (MLX5_WQE_DWORD_SIZE + 2))
565959be52eSNélio Laranjeiro 			break;
5662e22920bSAdrien Mazarguil 		/* Update element. */
5671d88ba17SNélio Laranjeiro 		(*txq->elts)[elts_head] = buf;
5686579c27cSNélio Laranjeiro 		elts_head = (elts_head + 1) & (elts_n - 1);
5695e1d11ecSNelio Laranjeiro 		/* Prefetch next buffer data. */
5706579c27cSNélio Laranjeiro 		if (pkts_n > 1) {
5716579c27cSNélio Laranjeiro 			volatile void *pkt_addr;
5726579c27cSNélio Laranjeiro 
5736579c27cSNélio Laranjeiro 			pkt_addr = rte_pktmbuf_mtod(*pkts, volatile void *);
5746579c27cSNélio Laranjeiro 			rte_prefetch0(pkt_addr);
5756579c27cSNélio Laranjeiro 		}
5761d88ba17SNélio Laranjeiro 		/* Should we enable HW CKSUM offload */
5771d88ba17SNélio Laranjeiro 		if (buf->ol_flags &
5781d88ba17SNélio Laranjeiro 		    (PKT_TX_IP_CKSUM | PKT_TX_TCP_CKSUM | PKT_TX_UDP_CKSUM)) {
579f5fde520SShahaf Shuler 			const uint64_t is_tunneled = buf->ol_flags &
580f5fde520SShahaf Shuler 						     (PKT_TX_TUNNEL_GRE |
581f5fde520SShahaf Shuler 						      PKT_TX_TUNNEL_VXLAN);
582f5fde520SShahaf Shuler 
583f5fde520SShahaf Shuler 			if (is_tunneled && txq->tunnel_en) {
584f5fde520SShahaf Shuler 				cs_flags = MLX5_ETH_WQE_L3_INNER_CSUM |
585f5fde520SShahaf Shuler 					   MLX5_ETH_WQE_L4_INNER_CSUM;
586f5fde520SShahaf Shuler 				if (buf->ol_flags & PKT_TX_OUTER_IP_CKSUM)
587f5fde520SShahaf Shuler 					cs_flags |= MLX5_ETH_WQE_L3_CSUM;
588f5fde520SShahaf Shuler 			} else {
589f5fde520SShahaf Shuler 				cs_flags = MLX5_ETH_WQE_L3_CSUM |
590f5fde520SShahaf Shuler 					   MLX5_ETH_WQE_L4_CSUM;
591f5fde520SShahaf Shuler 			}
5921d88ba17SNélio Laranjeiro 		}
593b8fe952eSNélio Laranjeiro 		raw = ((uint8_t *)(uintptr_t)wqe) + 2 * MLX5_WQE_DWORD_SIZE;
5946579c27cSNélio Laranjeiro 		/* Replace the Ethernet type by the VLAN if necessary. */
5956579c27cSNélio Laranjeiro 		if (buf->ol_flags & PKT_TX_VLAN_PKT) {
5966579c27cSNélio Laranjeiro 			uint32_t vlan = htonl(0x81000000 | buf->vlan_tci);
5970d637a34SNélio Laranjeiro 			unsigned int len = 2 * ETHER_ADDR_LEN - 2;
5986579c27cSNélio Laranjeiro 
5990d637a34SNélio Laranjeiro 			addr += 2;
6000d637a34SNélio Laranjeiro 			length -= 2;
6010d637a34SNélio Laranjeiro 			/* Copy Destination and source mac address. */
6020d637a34SNélio Laranjeiro 			memcpy((uint8_t *)raw, ((uint8_t *)addr), len);
6030d637a34SNélio Laranjeiro 			/* Copy VLAN. */
6040d637a34SNélio Laranjeiro 			memcpy((uint8_t *)raw + len, &vlan, sizeof(vlan));
6050d637a34SNélio Laranjeiro 			/* Copy missing two bytes to end the DSeg. */
6060d637a34SNélio Laranjeiro 			memcpy((uint8_t *)raw + len + sizeof(vlan),
6070d637a34SNélio Laranjeiro 			       ((uint8_t *)addr) + len, 2);
6080d637a34SNélio Laranjeiro 			addr += len + 2;
6090d637a34SNélio Laranjeiro 			length -= (len + 2);
6100d637a34SNélio Laranjeiro 		} else {
6110d637a34SNélio Laranjeiro 			memcpy((uint8_t *)raw, ((uint8_t *)addr) + 2,
6120d637a34SNélio Laranjeiro 			       MLX5_WQE_DWORD_SIZE);
6130d637a34SNélio Laranjeiro 			length -= pkt_inline_sz;
6140d637a34SNélio Laranjeiro 			addr += pkt_inline_sz;
6156579c27cSNélio Laranjeiro 		}
6163f13f8c2SShahaf Shuler 		if (txq->tso_en) {
6173f13f8c2SShahaf Shuler 			tso = buf->ol_flags & PKT_TX_TCP_SEG;
6183f13f8c2SShahaf Shuler 			if (tso) {
6193f13f8c2SShahaf Shuler 				uintptr_t end = (uintptr_t)
6203f13f8c2SShahaf Shuler 						(((uintptr_t)txq->wqes) +
6213f13f8c2SShahaf Shuler 						(1 << txq->wqe_n) *
6223f13f8c2SShahaf Shuler 						MLX5_WQE_SIZE);
6233f13f8c2SShahaf Shuler 				unsigned int copy_b;
6243f13f8c2SShahaf Shuler 				uint8_t vlan_sz = (buf->ol_flags &
6253f13f8c2SShahaf Shuler 						  PKT_TX_VLAN_PKT) ? 4 : 0;
626b247f346SShahaf Shuler 				const uint64_t is_tunneled =
627b247f346SShahaf Shuler 							buf->ol_flags &
628b247f346SShahaf Shuler 							(PKT_TX_TUNNEL_GRE |
629b247f346SShahaf Shuler 							 PKT_TX_TUNNEL_VXLAN);
6303f13f8c2SShahaf Shuler 
6313f13f8c2SShahaf Shuler 				tso_header_sz = buf->l2_len + vlan_sz +
6323f13f8c2SShahaf Shuler 						buf->l3_len + buf->l4_len;
6333f13f8c2SShahaf Shuler 
634b247f346SShahaf Shuler 				if (is_tunneled	&& txq->tunnel_en) {
635b247f346SShahaf Shuler 					tso_header_sz += buf->outer_l2_len +
636b247f346SShahaf Shuler 							 buf->outer_l3_len;
6372a6c96beSShahaf Shuler 					cs_flags |= MLX5_ETH_WQE_L4_INNER_CSUM;
6382a6c96beSShahaf Shuler 				} else {
6392a6c96beSShahaf Shuler 					cs_flags |= MLX5_ETH_WQE_L4_CSUM;
640b247f346SShahaf Shuler 				}
6413f13f8c2SShahaf Shuler 				if (unlikely(tso_header_sz >
6423f13f8c2SShahaf Shuler 					     MLX5_MAX_TSO_HEADER))
6433f13f8c2SShahaf Shuler 					break;
6443f13f8c2SShahaf Shuler 				copy_b = tso_header_sz - pkt_inline_sz;
6453f13f8c2SShahaf Shuler 				/* First seg must contain all headers. */
6463f13f8c2SShahaf Shuler 				assert(copy_b <= length);
6473f13f8c2SShahaf Shuler 				raw += MLX5_WQE_DWORD_SIZE;
6483f13f8c2SShahaf Shuler 				if (copy_b &&
6493f13f8c2SShahaf Shuler 				   ((end - (uintptr_t)raw) > copy_b)) {
6503f13f8c2SShahaf Shuler 					uint16_t n = (MLX5_WQE_DS(copy_b) -
6513f13f8c2SShahaf Shuler 						      1 + 3) / 4;
6523f13f8c2SShahaf Shuler 
6533f13f8c2SShahaf Shuler 					if (unlikely(max_wqe < n))
6543f13f8c2SShahaf Shuler 						break;
6553f13f8c2SShahaf Shuler 					max_wqe -= n;
6563f13f8c2SShahaf Shuler 					rte_memcpy((void *)raw,
6573f13f8c2SShahaf Shuler 						   (void *)addr, copy_b);
6583f13f8c2SShahaf Shuler 					addr += copy_b;
6593f13f8c2SShahaf Shuler 					length -= copy_b;
6603f13f8c2SShahaf Shuler 					pkt_inline_sz += copy_b;
6613f13f8c2SShahaf Shuler 					/*
6623f13f8c2SShahaf Shuler 					 * Another DWORD will be added
6633f13f8c2SShahaf Shuler 					 * in the inline part.
6643f13f8c2SShahaf Shuler 					 */
6653f13f8c2SShahaf Shuler 					raw += MLX5_WQE_DS(copy_b) *
6663f13f8c2SShahaf Shuler 					       MLX5_WQE_DWORD_SIZE -
6673f13f8c2SShahaf Shuler 					       MLX5_WQE_DWORD_SIZE;
6683f13f8c2SShahaf Shuler 				} else {
6693f13f8c2SShahaf Shuler 					/* NOP WQE. */
6703f13f8c2SShahaf Shuler 					wqe->ctrl = (rte_v128u32_t){
6713f13f8c2SShahaf Shuler 						     htonl(txq->wqe_ci << 8),
6723f13f8c2SShahaf Shuler 						     htonl(txq->qp_num_8s | 1),
6733f13f8c2SShahaf Shuler 						     0,
6743f13f8c2SShahaf Shuler 						     0,
6753f13f8c2SShahaf Shuler 					};
6763f13f8c2SShahaf Shuler 					ds = 1;
6773f13f8c2SShahaf Shuler 					total_length = 0;
6783f13f8c2SShahaf Shuler 					pkts--;
6793f13f8c2SShahaf Shuler 					pkts_n++;
6803f13f8c2SShahaf Shuler 					elts_head = (elts_head - 1) &
6813f13f8c2SShahaf Shuler 						    (elts_n - 1);
6823f13f8c2SShahaf Shuler 					k++;
6833f13f8c2SShahaf Shuler 					goto next_wqe;
6843f13f8c2SShahaf Shuler 				}
6853f13f8c2SShahaf Shuler 			}
6863f13f8c2SShahaf Shuler 		}
6876579c27cSNélio Laranjeiro 		/* Inline if enough room. */
6883f13f8c2SShahaf Shuler 		if (txq->inline_en || tso) {
689fdcb0f53SNélio Laranjeiro 			uintptr_t end = (uintptr_t)
690fdcb0f53SNélio Laranjeiro 				(((uintptr_t)txq->wqes) +
691fdcb0f53SNélio Laranjeiro 				 (1 << txq->wqe_n) * MLX5_WQE_SIZE);
6928fcd6c2cSNélio Laranjeiro 			unsigned int max_inline = txq->max_inline *
6938fcd6c2cSNélio Laranjeiro 						  RTE_CACHE_LINE_SIZE -
6943f13f8c2SShahaf Shuler 						  (pkt_inline_sz - 2);
6956579c27cSNélio Laranjeiro 			uintptr_t addr_end = (addr + max_inline) &
6966579c27cSNélio Laranjeiro 					     ~(RTE_CACHE_LINE_SIZE - 1);
6978fcd6c2cSNélio Laranjeiro 			unsigned int copy_b = (addr_end > addr) ?
6988fcd6c2cSNélio Laranjeiro 				RTE_MIN((addr_end - addr), length) :
6998fcd6c2cSNélio Laranjeiro 				0;
7006579c27cSNélio Laranjeiro 
7018fcd6c2cSNélio Laranjeiro 			raw += MLX5_WQE_DWORD_SIZE;
7028fcd6c2cSNélio Laranjeiro 			if (copy_b && ((end - (uintptr_t)raw) > copy_b)) {
703f04f1d51SNélio Laranjeiro 				/*
704f04f1d51SNélio Laranjeiro 				 * One Dseg remains in the current WQE.  To
705f04f1d51SNélio Laranjeiro 				 * keep the computation positive, it is
706f04f1d51SNélio Laranjeiro 				 * removed after the bytes to Dseg conversion.
707f04f1d51SNélio Laranjeiro 				 */
7088fcd6c2cSNélio Laranjeiro 				uint16_t n = (MLX5_WQE_DS(copy_b) - 1 + 3) / 4;
7098fcd6c2cSNélio Laranjeiro 
710f04f1d51SNélio Laranjeiro 				if (unlikely(max_wqe < n))
711f04f1d51SNélio Laranjeiro 					break;
712f04f1d51SNélio Laranjeiro 				max_wqe -= n;
7133f13f8c2SShahaf Shuler 				if (tso) {
7143f13f8c2SShahaf Shuler 					uint32_t inl =
7153f13f8c2SShahaf Shuler 						htonl(copy_b | MLX5_INLINE_SEG);
7163f13f8c2SShahaf Shuler 
7173f13f8c2SShahaf Shuler 					pkt_inline_sz =
7183f13f8c2SShahaf Shuler 						MLX5_WQE_DS(tso_header_sz) *
7193f13f8c2SShahaf Shuler 						MLX5_WQE_DWORD_SIZE;
7203f13f8c2SShahaf Shuler 					rte_memcpy((void *)raw,
7213f13f8c2SShahaf Shuler 						   (void *)&inl, sizeof(inl));
7223f13f8c2SShahaf Shuler 					raw += sizeof(inl);
7233f13f8c2SShahaf Shuler 					pkt_inline_sz += sizeof(inl);
7243f13f8c2SShahaf Shuler 				}
7256579c27cSNélio Laranjeiro 				rte_memcpy((void *)raw, (void *)addr, copy_b);
7266579c27cSNélio Laranjeiro 				addr += copy_b;
7276579c27cSNélio Laranjeiro 				length -= copy_b;
7286579c27cSNélio Laranjeiro 				pkt_inline_sz += copy_b;
7296579c27cSNélio Laranjeiro 			}
7306579c27cSNélio Laranjeiro 			/*
731786b5c2dSShahaf Shuler 			 * 2 DWORDs consumed by the WQE header + ETH segment +
7326579c27cSNélio Laranjeiro 			 * the size of the inline part of the packet.
7336579c27cSNélio Laranjeiro 			 */
7346579c27cSNélio Laranjeiro 			ds = 2 + MLX5_WQE_DS(pkt_inline_sz - 2);
7356579c27cSNélio Laranjeiro 			if (length > 0) {
736f04f1d51SNélio Laranjeiro 				if (ds % (MLX5_WQE_SIZE /
737f04f1d51SNélio Laranjeiro 					  MLX5_WQE_DWORD_SIZE) == 0) {
738f04f1d51SNélio Laranjeiro 					if (unlikely(--max_wqe == 0))
739f04f1d51SNélio Laranjeiro 						break;
740f04f1d51SNélio Laranjeiro 					dseg = (volatile rte_v128u32_t *)
741f04f1d51SNélio Laranjeiro 					       tx_mlx5_wqe(txq, txq->wqe_ci +
742f04f1d51SNélio Laranjeiro 							   ds / 4);
743f04f1d51SNélio Laranjeiro 				} else {
7449a7fa9f7SNélio Laranjeiro 					dseg = (volatile rte_v128u32_t *)
7456579c27cSNélio Laranjeiro 						((uintptr_t)wqe +
7466579c27cSNélio Laranjeiro 						 (ds * MLX5_WQE_DWORD_SIZE));
747f04f1d51SNélio Laranjeiro 				}
7486579c27cSNélio Laranjeiro 				goto use_dseg;
7496579c27cSNélio Laranjeiro 			} else if (!segs_n) {
7506579c27cSNélio Laranjeiro 				goto next_pkt;
7516579c27cSNélio Laranjeiro 			} else {
752786b5c2dSShahaf Shuler 				/* dseg will be advance as part of next_seg */
753786b5c2dSShahaf Shuler 				dseg = (volatile rte_v128u32_t *)
754786b5c2dSShahaf Shuler 					((uintptr_t)wqe +
755786b5c2dSShahaf Shuler 					 ((ds - 1) * MLX5_WQE_DWORD_SIZE));
7566579c27cSNélio Laranjeiro 				goto next_seg;
7576579c27cSNélio Laranjeiro 			}
7586579c27cSNélio Laranjeiro 		} else {
7596579c27cSNélio Laranjeiro 			/*
7606579c27cSNélio Laranjeiro 			 * No inline has been done in the packet, only the
7616579c27cSNélio Laranjeiro 			 * Ethernet Header as been stored.
7626579c27cSNélio Laranjeiro 			 */
7639a7fa9f7SNélio Laranjeiro 			dseg = (volatile rte_v128u32_t *)
7646579c27cSNélio Laranjeiro 				((uintptr_t)wqe + (3 * MLX5_WQE_DWORD_SIZE));
7656579c27cSNélio Laranjeiro 			ds = 3;
7666579c27cSNélio Laranjeiro use_dseg:
7676579c27cSNélio Laranjeiro 			/* Add the remaining packet as a simple ds. */
7689a7fa9f7SNélio Laranjeiro 			naddr = htonll(addr);
7699a7fa9f7SNélio Laranjeiro 			*dseg = (rte_v128u32_t){
7709a7fa9f7SNélio Laranjeiro 				htonl(length),
7719a7fa9f7SNélio Laranjeiro 				txq_mp2mr(txq, txq_mb2mp(buf)),
7729a7fa9f7SNélio Laranjeiro 				naddr,
7739a7fa9f7SNélio Laranjeiro 				naddr >> 32,
7746579c27cSNélio Laranjeiro 			};
7756579c27cSNélio Laranjeiro 			++ds;
7766579c27cSNélio Laranjeiro 			if (!segs_n)
7776579c27cSNélio Laranjeiro 				goto next_pkt;
7786579c27cSNélio Laranjeiro 		}
7796579c27cSNélio Laranjeiro next_seg:
7806579c27cSNélio Laranjeiro 		assert(buf);
7816579c27cSNélio Laranjeiro 		assert(ds);
7826579c27cSNélio Laranjeiro 		assert(wqe);
783a5bf6af9SAdrien Mazarguil 		/*
784a5bf6af9SAdrien Mazarguil 		 * Spill on next WQE when the current one does not have
785a5bf6af9SAdrien Mazarguil 		 * enough room left. Size of WQE must a be a multiple
786a5bf6af9SAdrien Mazarguil 		 * of data segment size.
787a5bf6af9SAdrien Mazarguil 		 */
7888688b2f8SNélio Laranjeiro 		assert(!(MLX5_WQE_SIZE % MLX5_WQE_DWORD_SIZE));
7896579c27cSNélio Laranjeiro 		if (!(ds % (MLX5_WQE_SIZE / MLX5_WQE_DWORD_SIZE))) {
790f04f1d51SNélio Laranjeiro 			if (unlikely(--max_wqe == 0))
791f04f1d51SNélio Laranjeiro 				break;
7929a7fa9f7SNélio Laranjeiro 			dseg = (volatile rte_v128u32_t *)
793f04f1d51SNélio Laranjeiro 			       tx_mlx5_wqe(txq, txq->wqe_ci + ds / 4);
794f04f1d51SNélio Laranjeiro 			rte_prefetch0(tx_mlx5_wqe(txq,
795f04f1d51SNélio Laranjeiro 						  txq->wqe_ci + ds / 4 + 1));
7966579c27cSNélio Laranjeiro 		} else {
797a5bf6af9SAdrien Mazarguil 			++dseg;
7986579c27cSNélio Laranjeiro 		}
799a5bf6af9SAdrien Mazarguil 		++ds;
800a5bf6af9SAdrien Mazarguil 		buf = buf->next;
801a5bf6af9SAdrien Mazarguil 		assert(buf);
8026579c27cSNélio Laranjeiro 		length = DATA_LEN(buf);
803a5bf6af9SAdrien Mazarguil #ifdef MLX5_PMD_SOFT_COUNTERS
8046579c27cSNélio Laranjeiro 		total_length += length;
805a5bf6af9SAdrien Mazarguil #endif
8066579c27cSNélio Laranjeiro 		/* Store segment information. */
8079a7fa9f7SNélio Laranjeiro 		naddr = htonll(rte_pktmbuf_mtod(buf, uintptr_t));
8089a7fa9f7SNélio Laranjeiro 		*dseg = (rte_v128u32_t){
8099a7fa9f7SNélio Laranjeiro 			htonl(length),
8109a7fa9f7SNélio Laranjeiro 			txq_mp2mr(txq, txq_mb2mp(buf)),
8119a7fa9f7SNélio Laranjeiro 			naddr,
8129a7fa9f7SNélio Laranjeiro 			naddr >> 32,
8136579c27cSNélio Laranjeiro 		};
8146579c27cSNélio Laranjeiro 		(*txq->elts)[elts_head] = buf;
8156579c27cSNélio Laranjeiro 		elts_head = (elts_head + 1) & (elts_n - 1);
816a5bf6af9SAdrien Mazarguil 		++j;
8176579c27cSNélio Laranjeiro 		--segs_n;
8186579c27cSNélio Laranjeiro 		if (segs_n)
8196579c27cSNélio Laranjeiro 			goto next_seg;
8206579c27cSNélio Laranjeiro 		else
8216579c27cSNélio Laranjeiro 			--pkts_n;
8226579c27cSNélio Laranjeiro next_pkt:
8236579c27cSNélio Laranjeiro 		++i;
824b8fe952eSNélio Laranjeiro 		/* Initialize known and common part of the WQE structure. */
8253f13f8c2SShahaf Shuler 		if (tso) {
8263f13f8c2SShahaf Shuler 			wqe->ctrl = (rte_v128u32_t){
8273f13f8c2SShahaf Shuler 				htonl((txq->wqe_ci << 8) | MLX5_OPCODE_TSO),
8283f13f8c2SShahaf Shuler 				htonl(txq->qp_num_8s | ds),
8293f13f8c2SShahaf Shuler 				0,
8303f13f8c2SShahaf Shuler 				0,
8313f13f8c2SShahaf Shuler 			};
8323f13f8c2SShahaf Shuler 			wqe->eseg = (rte_v128u32_t){
8333f13f8c2SShahaf Shuler 				0,
8343f13f8c2SShahaf Shuler 				cs_flags | (htons(buf->tso_segsz) << 16),
8353f13f8c2SShahaf Shuler 				0,
8363f13f8c2SShahaf Shuler 				(ehdr << 16) | htons(tso_header_sz),
8373f13f8c2SShahaf Shuler 			};
8383f13f8c2SShahaf Shuler 		} else {
8399a7fa9f7SNélio Laranjeiro 			wqe->ctrl = (rte_v128u32_t){
8409a7fa9f7SNélio Laranjeiro 				htonl((txq->wqe_ci << 8) | MLX5_OPCODE_SEND),
8419a7fa9f7SNélio Laranjeiro 				htonl(txq->qp_num_8s | ds),
8429a7fa9f7SNélio Laranjeiro 				0,
8439a7fa9f7SNélio Laranjeiro 				0,
8449a7fa9f7SNélio Laranjeiro 			};
8459a7fa9f7SNélio Laranjeiro 			wqe->eseg = (rte_v128u32_t){
8469a7fa9f7SNélio Laranjeiro 				0,
8479a7fa9f7SNélio Laranjeiro 				cs_flags,
8489a7fa9f7SNélio Laranjeiro 				0,
849eef822ddSNélio Laranjeiro 				(ehdr << 16) | htons(pkt_inline_sz),
8509a7fa9f7SNélio Laranjeiro 			};
8513f13f8c2SShahaf Shuler 		}
8523f13f8c2SShahaf Shuler next_wqe:
8536579c27cSNélio Laranjeiro 		txq->wqe_ci += (ds + 3) / 4;
85487011737SAdrien Mazarguil #ifdef MLX5_PMD_SOFT_COUNTERS
855573f54afSNélio Laranjeiro 		/* Increment sent bytes counter. */
8566579c27cSNélio Laranjeiro 		txq->stats.obytes += total_length;
85787011737SAdrien Mazarguil #endif
858c3d62cc9SAdrien Mazarguil 	} while (pkts_n);
8592e22920bSAdrien Mazarguil 	/* Take a shortcut if nothing must be sent. */
8603f13f8c2SShahaf Shuler 	if (unlikely((i + k) == 0))
8612e22920bSAdrien Mazarguil 		return 0;
862c305090bSAdrien Mazarguil 	/* Check whether completion threshold has been reached. */
8633f13f8c2SShahaf Shuler 	comp = txq->elts_comp + i + j + k;
864c305090bSAdrien Mazarguil 	if (comp >= MLX5_TX_COMP_THRESH) {
8659a7fa9f7SNélio Laranjeiro 		volatile struct mlx5_wqe_ctrl *w =
8669a7fa9f7SNélio Laranjeiro 			(volatile struct mlx5_wqe_ctrl *)wqe;
8679a7fa9f7SNélio Laranjeiro 
868c305090bSAdrien Mazarguil 		/* Request completion on last WQE. */
8699a7fa9f7SNélio Laranjeiro 		w->ctrl2 = htonl(8);
870c305090bSAdrien Mazarguil 		/* Save elts_head in unused "immediate" field of WQE. */
8719a7fa9f7SNélio Laranjeiro 		w->ctrl3 = elts_head;
872c305090bSAdrien Mazarguil 		txq->elts_comp = 0;
873c305090bSAdrien Mazarguil 	} else {
874c305090bSAdrien Mazarguil 		txq->elts_comp = comp;
875c305090bSAdrien Mazarguil 	}
87687011737SAdrien Mazarguil #ifdef MLX5_PMD_SOFT_COUNTERS
87787011737SAdrien Mazarguil 	/* Increment sent packets counter. */
87887011737SAdrien Mazarguil 	txq->stats.opackets += i;
87987011737SAdrien Mazarguil #endif
8802e22920bSAdrien Mazarguil 	/* Ring QP doorbell. */
88130807f62SNélio Laranjeiro 	mlx5_tx_dbrec(txq, (volatile struct mlx5_wqe *)wqe);
8822e22920bSAdrien Mazarguil 	txq->elts_head = elts_head;
8832e22920bSAdrien Mazarguil 	return i;
8842e22920bSAdrien Mazarguil }
8852e22920bSAdrien Mazarguil 
8862e22920bSAdrien Mazarguil /**
887230189d9SNélio Laranjeiro  * Open a MPW session.
888230189d9SNélio Laranjeiro  *
889230189d9SNélio Laranjeiro  * @param txq
890230189d9SNélio Laranjeiro  *   Pointer to TX queue structure.
891230189d9SNélio Laranjeiro  * @param mpw
892230189d9SNélio Laranjeiro  *   Pointer to MPW session structure.
893230189d9SNélio Laranjeiro  * @param length
894230189d9SNélio Laranjeiro  *   Packet length.
895230189d9SNélio Laranjeiro  */
896230189d9SNélio Laranjeiro static inline void
897230189d9SNélio Laranjeiro mlx5_mpw_new(struct txq *txq, struct mlx5_mpw *mpw, uint32_t length)
898230189d9SNélio Laranjeiro {
899a821d09dSNélio Laranjeiro 	uint16_t idx = txq->wqe_ci & ((1 << txq->wqe_n) - 1);
900230189d9SNélio Laranjeiro 	volatile struct mlx5_wqe_data_seg (*dseg)[MLX5_MPW_DSEG_MAX] =
901230189d9SNélio Laranjeiro 		(volatile struct mlx5_wqe_data_seg (*)[])
902fdcb0f53SNélio Laranjeiro 		tx_mlx5_wqe(txq, idx + 1);
903230189d9SNélio Laranjeiro 
904230189d9SNélio Laranjeiro 	mpw->state = MLX5_MPW_STATE_OPENED;
905230189d9SNélio Laranjeiro 	mpw->pkts_n = 0;
906230189d9SNélio Laranjeiro 	mpw->len = length;
907230189d9SNélio Laranjeiro 	mpw->total_len = 0;
908fdcb0f53SNélio Laranjeiro 	mpw->wqe = (volatile struct mlx5_wqe *)tx_mlx5_wqe(txq, idx);
9098688b2f8SNélio Laranjeiro 	mpw->wqe->eseg.mss = htons(length);
9108688b2f8SNélio Laranjeiro 	mpw->wqe->eseg.inline_hdr_sz = 0;
9118688b2f8SNélio Laranjeiro 	mpw->wqe->eseg.rsvd0 = 0;
9128688b2f8SNélio Laranjeiro 	mpw->wqe->eseg.rsvd1 = 0;
9138688b2f8SNélio Laranjeiro 	mpw->wqe->eseg.rsvd2 = 0;
9148688b2f8SNélio Laranjeiro 	mpw->wqe->ctrl[0] = htonl((MLX5_OPC_MOD_MPW << 24) |
915c904ae25SNélio Laranjeiro 				  (txq->wqe_ci << 8) | MLX5_OPCODE_TSO);
9168688b2f8SNélio Laranjeiro 	mpw->wqe->ctrl[2] = 0;
9178688b2f8SNélio Laranjeiro 	mpw->wqe->ctrl[3] = 0;
9188688b2f8SNélio Laranjeiro 	mpw->data.dseg[0] = (volatile struct mlx5_wqe_data_seg *)
9198688b2f8SNélio Laranjeiro 		(((uintptr_t)mpw->wqe) + (2 * MLX5_WQE_DWORD_SIZE));
9208688b2f8SNélio Laranjeiro 	mpw->data.dseg[1] = (volatile struct mlx5_wqe_data_seg *)
9218688b2f8SNélio Laranjeiro 		(((uintptr_t)mpw->wqe) + (3 * MLX5_WQE_DWORD_SIZE));
922230189d9SNélio Laranjeiro 	mpw->data.dseg[2] = &(*dseg)[0];
923230189d9SNélio Laranjeiro 	mpw->data.dseg[3] = &(*dseg)[1];
924230189d9SNélio Laranjeiro 	mpw->data.dseg[4] = &(*dseg)[2];
925230189d9SNélio Laranjeiro }
926230189d9SNélio Laranjeiro 
927230189d9SNélio Laranjeiro /**
928230189d9SNélio Laranjeiro  * Close a MPW session.
929230189d9SNélio Laranjeiro  *
930230189d9SNélio Laranjeiro  * @param txq
931230189d9SNélio Laranjeiro  *   Pointer to TX queue structure.
932230189d9SNélio Laranjeiro  * @param mpw
933230189d9SNélio Laranjeiro  *   Pointer to MPW session structure.
934230189d9SNélio Laranjeiro  */
935230189d9SNélio Laranjeiro static inline void
936230189d9SNélio Laranjeiro mlx5_mpw_close(struct txq *txq, struct mlx5_mpw *mpw)
937230189d9SNélio Laranjeiro {
938230189d9SNélio Laranjeiro 	unsigned int num = mpw->pkts_n;
939230189d9SNélio Laranjeiro 
940230189d9SNélio Laranjeiro 	/*
941230189d9SNélio Laranjeiro 	 * Store size in multiple of 16 bytes. Control and Ethernet segments
942230189d9SNélio Laranjeiro 	 * count as 2.
943230189d9SNélio Laranjeiro 	 */
9448688b2f8SNélio Laranjeiro 	mpw->wqe->ctrl[1] = htonl(txq->qp_num_8s | (2 + num));
945230189d9SNélio Laranjeiro 	mpw->state = MLX5_MPW_STATE_CLOSED;
946230189d9SNélio Laranjeiro 	if (num < 3)
947230189d9SNélio Laranjeiro 		++txq->wqe_ci;
948230189d9SNélio Laranjeiro 	else
949230189d9SNélio Laranjeiro 		txq->wqe_ci += 2;
950fdcb0f53SNélio Laranjeiro 	rte_prefetch0(tx_mlx5_wqe(txq, txq->wqe_ci));
951fdcb0f53SNélio Laranjeiro 	rte_prefetch0(tx_mlx5_wqe(txq, txq->wqe_ci + 1));
952230189d9SNélio Laranjeiro }
953230189d9SNélio Laranjeiro 
954230189d9SNélio Laranjeiro /**
955230189d9SNélio Laranjeiro  * DPDK callback for TX with MPW support.
956230189d9SNélio Laranjeiro  *
957230189d9SNélio Laranjeiro  * @param dpdk_txq
958230189d9SNélio Laranjeiro  *   Generic pointer to TX queue structure.
959230189d9SNélio Laranjeiro  * @param[in] pkts
960230189d9SNélio Laranjeiro  *   Packets to transmit.
961230189d9SNélio Laranjeiro  * @param pkts_n
962230189d9SNélio Laranjeiro  *   Number of packets in array.
963230189d9SNélio Laranjeiro  *
964230189d9SNélio Laranjeiro  * @return
965230189d9SNélio Laranjeiro  *   Number of packets successfully transmitted (<= pkts_n).
966230189d9SNélio Laranjeiro  */
967230189d9SNélio Laranjeiro uint16_t
968230189d9SNélio Laranjeiro mlx5_tx_burst_mpw(void *dpdk_txq, struct rte_mbuf **pkts, uint16_t pkts_n)
969230189d9SNélio Laranjeiro {
970230189d9SNélio Laranjeiro 	struct txq *txq = (struct txq *)dpdk_txq;
971230189d9SNélio Laranjeiro 	uint16_t elts_head = txq->elts_head;
972b4b12e55SNélio Laranjeiro 	const unsigned int elts_n = 1 << txq->elts_n;
973c3d62cc9SAdrien Mazarguil 	unsigned int i = 0;
974a5bf6af9SAdrien Mazarguil 	unsigned int j = 0;
975230189d9SNélio Laranjeiro 	unsigned int max;
976f04f1d51SNélio Laranjeiro 	uint16_t max_wqe;
977230189d9SNélio Laranjeiro 	unsigned int comp;
978230189d9SNélio Laranjeiro 	struct mlx5_mpw mpw = {
979230189d9SNélio Laranjeiro 		.state = MLX5_MPW_STATE_CLOSED,
980230189d9SNélio Laranjeiro 	};
981230189d9SNélio Laranjeiro 
982c3d62cc9SAdrien Mazarguil 	if (unlikely(!pkts_n))
983c3d62cc9SAdrien Mazarguil 		return 0;
984230189d9SNélio Laranjeiro 	/* Prefetch first packet cacheline. */
985fdcb0f53SNélio Laranjeiro 	rte_prefetch0(tx_mlx5_wqe(txq, txq->wqe_ci));
986fdcb0f53SNélio Laranjeiro 	rte_prefetch0(tx_mlx5_wqe(txq, txq->wqe_ci + 1));
987230189d9SNélio Laranjeiro 	/* Start processing. */
988230189d9SNélio Laranjeiro 	txq_complete(txq);
989230189d9SNélio Laranjeiro 	max = (elts_n - (elts_head - txq->elts_tail));
990230189d9SNélio Laranjeiro 	if (max > elts_n)
991230189d9SNélio Laranjeiro 		max -= elts_n;
992f04f1d51SNélio Laranjeiro 	max_wqe = (1u << txq->wqe_n) - (txq->wqe_ci - txq->wqe_pi);
993f04f1d51SNélio Laranjeiro 	if (unlikely(!max_wqe))
994f04f1d51SNélio Laranjeiro 		return 0;
995c3d62cc9SAdrien Mazarguil 	do {
996a5bf6af9SAdrien Mazarguil 		struct rte_mbuf *buf = *(pkts++);
997c3d62cc9SAdrien Mazarguil 		unsigned int elts_head_next;
998230189d9SNélio Laranjeiro 		uint32_t length;
999a5bf6af9SAdrien Mazarguil 		unsigned int segs_n = buf->nb_segs;
1000230189d9SNélio Laranjeiro 		uint32_t cs_flags = 0;
1001230189d9SNélio Laranjeiro 
1002c3d62cc9SAdrien Mazarguil 		/*
1003c3d62cc9SAdrien Mazarguil 		 * Make sure there is enough room to store this packet and
1004c3d62cc9SAdrien Mazarguil 		 * that one ring entry remains unused.
1005c3d62cc9SAdrien Mazarguil 		 */
1006a5bf6af9SAdrien Mazarguil 		assert(segs_n);
1007a5bf6af9SAdrien Mazarguil 		if (max < segs_n + 1)
1008c3d62cc9SAdrien Mazarguil 			break;
1009a5bf6af9SAdrien Mazarguil 		/* Do not bother with large packets MPW cannot handle. */
1010a5bf6af9SAdrien Mazarguil 		if (segs_n > MLX5_MPW_DSEG_MAX)
1011a5bf6af9SAdrien Mazarguil 			break;
1012a5bf6af9SAdrien Mazarguil 		max -= segs_n;
1013c3d62cc9SAdrien Mazarguil 		--pkts_n;
1014230189d9SNélio Laranjeiro 		/* Should we enable HW CKSUM offload */
1015230189d9SNélio Laranjeiro 		if (buf->ol_flags &
1016230189d9SNélio Laranjeiro 		    (PKT_TX_IP_CKSUM | PKT_TX_TCP_CKSUM | PKT_TX_UDP_CKSUM))
1017230189d9SNélio Laranjeiro 			cs_flags = MLX5_ETH_WQE_L3_CSUM | MLX5_ETH_WQE_L4_CSUM;
1018a5bf6af9SAdrien Mazarguil 		/* Retrieve packet information. */
1019a5bf6af9SAdrien Mazarguil 		length = PKT_LEN(buf);
1020a5bf6af9SAdrien Mazarguil 		assert(length);
1021230189d9SNélio Laranjeiro 		/* Start new session if packet differs. */
1022230189d9SNélio Laranjeiro 		if ((mpw.state == MLX5_MPW_STATE_OPENED) &&
1023230189d9SNélio Laranjeiro 		    ((mpw.len != length) ||
1024a5bf6af9SAdrien Mazarguil 		     (segs_n != 1) ||
10258688b2f8SNélio Laranjeiro 		     (mpw.wqe->eseg.cs_flags != cs_flags)))
1026230189d9SNélio Laranjeiro 			mlx5_mpw_close(txq, &mpw);
1027230189d9SNélio Laranjeiro 		if (mpw.state == MLX5_MPW_STATE_CLOSED) {
1028f04f1d51SNélio Laranjeiro 			/*
1029f04f1d51SNélio Laranjeiro 			 * Multi-Packet WQE consumes at most two WQE.
1030f04f1d51SNélio Laranjeiro 			 * mlx5_mpw_new() expects to be able to use such
1031f04f1d51SNélio Laranjeiro 			 * resources.
1032f04f1d51SNélio Laranjeiro 			 */
1033f04f1d51SNélio Laranjeiro 			if (unlikely(max_wqe < 2))
1034f04f1d51SNélio Laranjeiro 				break;
1035f04f1d51SNélio Laranjeiro 			max_wqe -= 2;
1036230189d9SNélio Laranjeiro 			mlx5_mpw_new(txq, &mpw, length);
10378688b2f8SNélio Laranjeiro 			mpw.wqe->eseg.cs_flags = cs_flags;
1038230189d9SNélio Laranjeiro 		}
1039a5bf6af9SAdrien Mazarguil 		/* Multi-segment packets must be alone in their MPW. */
1040a5bf6af9SAdrien Mazarguil 		assert((segs_n == 1) || (mpw.pkts_n == 0));
1041a5bf6af9SAdrien Mazarguil #if defined(MLX5_PMD_SOFT_COUNTERS) || !defined(NDEBUG)
1042a5bf6af9SAdrien Mazarguil 		length = 0;
1043a5bf6af9SAdrien Mazarguil #endif
1044a5bf6af9SAdrien Mazarguil 		do {
1045a5bf6af9SAdrien Mazarguil 			volatile struct mlx5_wqe_data_seg *dseg;
1046a5bf6af9SAdrien Mazarguil 			uintptr_t addr;
1047a5bf6af9SAdrien Mazarguil 
1048a5bf6af9SAdrien Mazarguil 			elts_head_next = (elts_head + 1) & (elts_n - 1);
1049a5bf6af9SAdrien Mazarguil 			assert(buf);
1050a5bf6af9SAdrien Mazarguil 			(*txq->elts)[elts_head] = buf;
1051230189d9SNélio Laranjeiro 			dseg = mpw.data.dseg[mpw.pkts_n];
1052a5bf6af9SAdrien Mazarguil 			addr = rte_pktmbuf_mtod(buf, uintptr_t);
1053230189d9SNélio Laranjeiro 			*dseg = (struct mlx5_wqe_data_seg){
1054a5bf6af9SAdrien Mazarguil 				.byte_count = htonl(DATA_LEN(buf)),
1055230189d9SNélio Laranjeiro 				.lkey = txq_mp2mr(txq, txq_mb2mp(buf)),
1056230189d9SNélio Laranjeiro 				.addr = htonll(addr),
1057230189d9SNélio Laranjeiro 			};
1058a5bf6af9SAdrien Mazarguil 			elts_head = elts_head_next;
1059a5bf6af9SAdrien Mazarguil #if defined(MLX5_PMD_SOFT_COUNTERS) || !defined(NDEBUG)
1060a5bf6af9SAdrien Mazarguil 			length += DATA_LEN(buf);
1061a5bf6af9SAdrien Mazarguil #endif
1062a5bf6af9SAdrien Mazarguil 			buf = buf->next;
1063230189d9SNélio Laranjeiro 			++mpw.pkts_n;
1064a5bf6af9SAdrien Mazarguil 			++j;
1065a5bf6af9SAdrien Mazarguil 		} while (--segs_n);
1066a5bf6af9SAdrien Mazarguil 		assert(length == mpw.len);
1067230189d9SNélio Laranjeiro 		if (mpw.pkts_n == MLX5_MPW_DSEG_MAX)
1068230189d9SNélio Laranjeiro 			mlx5_mpw_close(txq, &mpw);
1069230189d9SNélio Laranjeiro 		elts_head = elts_head_next;
1070230189d9SNélio Laranjeiro #ifdef MLX5_PMD_SOFT_COUNTERS
1071230189d9SNélio Laranjeiro 		/* Increment sent bytes counter. */
1072230189d9SNélio Laranjeiro 		txq->stats.obytes += length;
1073230189d9SNélio Laranjeiro #endif
1074c3d62cc9SAdrien Mazarguil 		++i;
1075c3d62cc9SAdrien Mazarguil 	} while (pkts_n);
1076230189d9SNélio Laranjeiro 	/* Take a shortcut if nothing must be sent. */
1077230189d9SNélio Laranjeiro 	if (unlikely(i == 0))
1078230189d9SNélio Laranjeiro 		return 0;
1079230189d9SNélio Laranjeiro 	/* Check whether completion threshold has been reached. */
1080a5bf6af9SAdrien Mazarguil 	/* "j" includes both packets and segments. */
1081a5bf6af9SAdrien Mazarguil 	comp = txq->elts_comp + j;
1082230189d9SNélio Laranjeiro 	if (comp >= MLX5_TX_COMP_THRESH) {
10838688b2f8SNélio Laranjeiro 		volatile struct mlx5_wqe *wqe = mpw.wqe;
1084230189d9SNélio Laranjeiro 
1085230189d9SNélio Laranjeiro 		/* Request completion on last WQE. */
10868688b2f8SNélio Laranjeiro 		wqe->ctrl[2] = htonl(8);
1087230189d9SNélio Laranjeiro 		/* Save elts_head in unused "immediate" field of WQE. */
10888688b2f8SNélio Laranjeiro 		wqe->ctrl[3] = elts_head;
1089230189d9SNélio Laranjeiro 		txq->elts_comp = 0;
1090230189d9SNélio Laranjeiro 	} else {
1091230189d9SNélio Laranjeiro 		txq->elts_comp = comp;
1092230189d9SNélio Laranjeiro 	}
1093230189d9SNélio Laranjeiro #ifdef MLX5_PMD_SOFT_COUNTERS
1094230189d9SNélio Laranjeiro 	/* Increment sent packets counter. */
1095230189d9SNélio Laranjeiro 	txq->stats.opackets += i;
1096230189d9SNélio Laranjeiro #endif
1097230189d9SNélio Laranjeiro 	/* Ring QP doorbell. */
1098230189d9SNélio Laranjeiro 	if (mpw.state == MLX5_MPW_STATE_OPENED)
1099230189d9SNélio Laranjeiro 		mlx5_mpw_close(txq, &mpw);
110030807f62SNélio Laranjeiro 	mlx5_tx_dbrec(txq, mpw.wqe);
1101230189d9SNélio Laranjeiro 	txq->elts_head = elts_head;
1102230189d9SNélio Laranjeiro 	return i;
1103230189d9SNélio Laranjeiro }
1104230189d9SNélio Laranjeiro 
1105230189d9SNélio Laranjeiro /**
1106230189d9SNélio Laranjeiro  * Open a MPW inline session.
1107230189d9SNélio Laranjeiro  *
1108230189d9SNélio Laranjeiro  * @param txq
1109230189d9SNélio Laranjeiro  *   Pointer to TX queue structure.
1110230189d9SNélio Laranjeiro  * @param mpw
1111230189d9SNélio Laranjeiro  *   Pointer to MPW session structure.
1112230189d9SNélio Laranjeiro  * @param length
1113230189d9SNélio Laranjeiro  *   Packet length.
1114230189d9SNélio Laranjeiro  */
1115230189d9SNélio Laranjeiro static inline void
1116230189d9SNélio Laranjeiro mlx5_mpw_inline_new(struct txq *txq, struct mlx5_mpw *mpw, uint32_t length)
1117230189d9SNélio Laranjeiro {
1118a821d09dSNélio Laranjeiro 	uint16_t idx = txq->wqe_ci & ((1 << txq->wqe_n) - 1);
11198688b2f8SNélio Laranjeiro 	struct mlx5_wqe_inl_small *inl;
1120230189d9SNélio Laranjeiro 
1121230189d9SNélio Laranjeiro 	mpw->state = MLX5_MPW_INL_STATE_OPENED;
1122230189d9SNélio Laranjeiro 	mpw->pkts_n = 0;
1123230189d9SNélio Laranjeiro 	mpw->len = length;
1124230189d9SNélio Laranjeiro 	mpw->total_len = 0;
1125fdcb0f53SNélio Laranjeiro 	mpw->wqe = (volatile struct mlx5_wqe *)tx_mlx5_wqe(txq, idx);
11268688b2f8SNélio Laranjeiro 	mpw->wqe->ctrl[0] = htonl((MLX5_OPC_MOD_MPW << 24) |
1127230189d9SNélio Laranjeiro 				  (txq->wqe_ci << 8) |
1128c904ae25SNélio Laranjeiro 				  MLX5_OPCODE_TSO);
11298688b2f8SNélio Laranjeiro 	mpw->wqe->ctrl[2] = 0;
11308688b2f8SNélio Laranjeiro 	mpw->wqe->ctrl[3] = 0;
11318688b2f8SNélio Laranjeiro 	mpw->wqe->eseg.mss = htons(length);
11328688b2f8SNélio Laranjeiro 	mpw->wqe->eseg.inline_hdr_sz = 0;
11338688b2f8SNélio Laranjeiro 	mpw->wqe->eseg.cs_flags = 0;
11348688b2f8SNélio Laranjeiro 	mpw->wqe->eseg.rsvd0 = 0;
11358688b2f8SNélio Laranjeiro 	mpw->wqe->eseg.rsvd1 = 0;
11368688b2f8SNélio Laranjeiro 	mpw->wqe->eseg.rsvd2 = 0;
11378688b2f8SNélio Laranjeiro 	inl = (struct mlx5_wqe_inl_small *)
11388688b2f8SNélio Laranjeiro 		(((uintptr_t)mpw->wqe) + 2 * MLX5_WQE_DWORD_SIZE);
11398688b2f8SNélio Laranjeiro 	mpw->data.raw = (uint8_t *)&inl->raw;
1140230189d9SNélio Laranjeiro }
1141230189d9SNélio Laranjeiro 
1142230189d9SNélio Laranjeiro /**
1143230189d9SNélio Laranjeiro  * Close a MPW inline session.
1144230189d9SNélio Laranjeiro  *
1145230189d9SNélio Laranjeiro  * @param txq
1146230189d9SNélio Laranjeiro  *   Pointer to TX queue structure.
1147230189d9SNélio Laranjeiro  * @param mpw
1148230189d9SNélio Laranjeiro  *   Pointer to MPW session structure.
1149230189d9SNélio Laranjeiro  */
1150230189d9SNélio Laranjeiro static inline void
1151230189d9SNélio Laranjeiro mlx5_mpw_inline_close(struct txq *txq, struct mlx5_mpw *mpw)
1152230189d9SNélio Laranjeiro {
1153230189d9SNélio Laranjeiro 	unsigned int size;
11548688b2f8SNélio Laranjeiro 	struct mlx5_wqe_inl_small *inl = (struct mlx5_wqe_inl_small *)
11558688b2f8SNélio Laranjeiro 		(((uintptr_t)mpw->wqe) + (2 * MLX5_WQE_DWORD_SIZE));
1156230189d9SNélio Laranjeiro 
11578688b2f8SNélio Laranjeiro 	size = MLX5_WQE_SIZE - MLX5_MWQE64_INL_DATA + mpw->total_len;
1158230189d9SNélio Laranjeiro 	/*
1159230189d9SNélio Laranjeiro 	 * Store size in multiple of 16 bytes. Control and Ethernet segments
1160230189d9SNélio Laranjeiro 	 * count as 2.
1161230189d9SNélio Laranjeiro 	 */
11628688b2f8SNélio Laranjeiro 	mpw->wqe->ctrl[1] = htonl(txq->qp_num_8s | MLX5_WQE_DS(size));
1163230189d9SNélio Laranjeiro 	mpw->state = MLX5_MPW_STATE_CLOSED;
11648688b2f8SNélio Laranjeiro 	inl->byte_cnt = htonl(mpw->total_len | MLX5_INLINE_SEG);
11658688b2f8SNélio Laranjeiro 	txq->wqe_ci += (size + (MLX5_WQE_SIZE - 1)) / MLX5_WQE_SIZE;
1166230189d9SNélio Laranjeiro }
1167230189d9SNélio Laranjeiro 
1168230189d9SNélio Laranjeiro /**
1169230189d9SNélio Laranjeiro  * DPDK callback for TX with MPW inline support.
1170230189d9SNélio Laranjeiro  *
1171230189d9SNélio Laranjeiro  * @param dpdk_txq
1172230189d9SNélio Laranjeiro  *   Generic pointer to TX queue structure.
1173230189d9SNélio Laranjeiro  * @param[in] pkts
1174230189d9SNélio Laranjeiro  *   Packets to transmit.
1175230189d9SNélio Laranjeiro  * @param pkts_n
1176230189d9SNélio Laranjeiro  *   Number of packets in array.
1177230189d9SNélio Laranjeiro  *
1178230189d9SNélio Laranjeiro  * @return
1179230189d9SNélio Laranjeiro  *   Number of packets successfully transmitted (<= pkts_n).
1180230189d9SNélio Laranjeiro  */
1181230189d9SNélio Laranjeiro uint16_t
1182230189d9SNélio Laranjeiro mlx5_tx_burst_mpw_inline(void *dpdk_txq, struct rte_mbuf **pkts,
1183230189d9SNélio Laranjeiro 			 uint16_t pkts_n)
1184230189d9SNélio Laranjeiro {
1185230189d9SNélio Laranjeiro 	struct txq *txq = (struct txq *)dpdk_txq;
1186230189d9SNélio Laranjeiro 	uint16_t elts_head = txq->elts_head;
1187b4b12e55SNélio Laranjeiro 	const unsigned int elts_n = 1 << txq->elts_n;
1188c3d62cc9SAdrien Mazarguil 	unsigned int i = 0;
1189a5bf6af9SAdrien Mazarguil 	unsigned int j = 0;
1190230189d9SNélio Laranjeiro 	unsigned int max;
1191f04f1d51SNélio Laranjeiro 	uint16_t max_wqe;
1192230189d9SNélio Laranjeiro 	unsigned int comp;
11930e8679fcSNélio Laranjeiro 	unsigned int inline_room = txq->max_inline * RTE_CACHE_LINE_SIZE;
1194230189d9SNélio Laranjeiro 	struct mlx5_mpw mpw = {
1195230189d9SNélio Laranjeiro 		.state = MLX5_MPW_STATE_CLOSED,
1196230189d9SNélio Laranjeiro 	};
1197f04f1d51SNélio Laranjeiro 	/*
1198f04f1d51SNélio Laranjeiro 	 * Compute the maximum number of WQE which can be consumed by inline
1199f04f1d51SNélio Laranjeiro 	 * code.
1200f04f1d51SNélio Laranjeiro 	 * - 2 DSEG for:
1201f04f1d51SNélio Laranjeiro 	 *   - 1 control segment,
1202f04f1d51SNélio Laranjeiro 	 *   - 1 Ethernet segment,
1203f04f1d51SNélio Laranjeiro 	 * - N Dseg from the inline request.
1204f04f1d51SNélio Laranjeiro 	 */
1205f04f1d51SNélio Laranjeiro 	const unsigned int wqe_inl_n =
1206f04f1d51SNélio Laranjeiro 		((2 * MLX5_WQE_DWORD_SIZE +
1207f04f1d51SNélio Laranjeiro 		  txq->max_inline * RTE_CACHE_LINE_SIZE) +
1208f04f1d51SNélio Laranjeiro 		 RTE_CACHE_LINE_SIZE - 1) / RTE_CACHE_LINE_SIZE;
1209230189d9SNélio Laranjeiro 
1210c3d62cc9SAdrien Mazarguil 	if (unlikely(!pkts_n))
1211c3d62cc9SAdrien Mazarguil 		return 0;
1212230189d9SNélio Laranjeiro 	/* Prefetch first packet cacheline. */
1213fdcb0f53SNélio Laranjeiro 	rte_prefetch0(tx_mlx5_wqe(txq, txq->wqe_ci));
1214fdcb0f53SNélio Laranjeiro 	rte_prefetch0(tx_mlx5_wqe(txq, txq->wqe_ci + 1));
1215230189d9SNélio Laranjeiro 	/* Start processing. */
1216230189d9SNélio Laranjeiro 	txq_complete(txq);
1217230189d9SNélio Laranjeiro 	max = (elts_n - (elts_head - txq->elts_tail));
1218230189d9SNélio Laranjeiro 	if (max > elts_n)
1219230189d9SNélio Laranjeiro 		max -= elts_n;
1220c3d62cc9SAdrien Mazarguil 	do {
1221a5bf6af9SAdrien Mazarguil 		struct rte_mbuf *buf = *(pkts++);
1222c3d62cc9SAdrien Mazarguil 		unsigned int elts_head_next;
1223230189d9SNélio Laranjeiro 		uintptr_t addr;
1224230189d9SNélio Laranjeiro 		uint32_t length;
1225a5bf6af9SAdrien Mazarguil 		unsigned int segs_n = buf->nb_segs;
1226230189d9SNélio Laranjeiro 		uint32_t cs_flags = 0;
1227230189d9SNélio Laranjeiro 
1228c3d62cc9SAdrien Mazarguil 		/*
1229c3d62cc9SAdrien Mazarguil 		 * Make sure there is enough room to store this packet and
1230c3d62cc9SAdrien Mazarguil 		 * that one ring entry remains unused.
1231c3d62cc9SAdrien Mazarguil 		 */
1232a5bf6af9SAdrien Mazarguil 		assert(segs_n);
1233a5bf6af9SAdrien Mazarguil 		if (max < segs_n + 1)
1234c3d62cc9SAdrien Mazarguil 			break;
1235a5bf6af9SAdrien Mazarguil 		/* Do not bother with large packets MPW cannot handle. */
1236a5bf6af9SAdrien Mazarguil 		if (segs_n > MLX5_MPW_DSEG_MAX)
1237a5bf6af9SAdrien Mazarguil 			break;
1238a5bf6af9SAdrien Mazarguil 		max -= segs_n;
1239c3d62cc9SAdrien Mazarguil 		--pkts_n;
1240f04f1d51SNélio Laranjeiro 		/*
1241f04f1d51SNélio Laranjeiro 		 * Compute max_wqe in case less WQE were consumed in previous
1242f04f1d51SNélio Laranjeiro 		 * iteration.
1243f04f1d51SNélio Laranjeiro 		 */
1244f04f1d51SNélio Laranjeiro 		max_wqe = (1u << txq->wqe_n) - (txq->wqe_ci - txq->wqe_pi);
1245230189d9SNélio Laranjeiro 		/* Should we enable HW CKSUM offload */
1246230189d9SNélio Laranjeiro 		if (buf->ol_flags &
1247230189d9SNélio Laranjeiro 		    (PKT_TX_IP_CKSUM | PKT_TX_TCP_CKSUM | PKT_TX_UDP_CKSUM))
1248230189d9SNélio Laranjeiro 			cs_flags = MLX5_ETH_WQE_L3_CSUM | MLX5_ETH_WQE_L4_CSUM;
1249a5bf6af9SAdrien Mazarguil 		/* Retrieve packet information. */
1250a5bf6af9SAdrien Mazarguil 		length = PKT_LEN(buf);
1251230189d9SNélio Laranjeiro 		/* Start new session if packet differs. */
1252230189d9SNélio Laranjeiro 		if (mpw.state == MLX5_MPW_STATE_OPENED) {
1253230189d9SNélio Laranjeiro 			if ((mpw.len != length) ||
1254a5bf6af9SAdrien Mazarguil 			    (segs_n != 1) ||
12558688b2f8SNélio Laranjeiro 			    (mpw.wqe->eseg.cs_flags != cs_flags))
1256230189d9SNélio Laranjeiro 				mlx5_mpw_close(txq, &mpw);
1257230189d9SNélio Laranjeiro 		} else if (mpw.state == MLX5_MPW_INL_STATE_OPENED) {
1258230189d9SNélio Laranjeiro 			if ((mpw.len != length) ||
1259a5bf6af9SAdrien Mazarguil 			    (segs_n != 1) ||
1260230189d9SNélio Laranjeiro 			    (length > inline_room) ||
12618688b2f8SNélio Laranjeiro 			    (mpw.wqe->eseg.cs_flags != cs_flags)) {
1262230189d9SNélio Laranjeiro 				mlx5_mpw_inline_close(txq, &mpw);
12630e8679fcSNélio Laranjeiro 				inline_room =
12640e8679fcSNélio Laranjeiro 					txq->max_inline * RTE_CACHE_LINE_SIZE;
1265230189d9SNélio Laranjeiro 			}
1266230189d9SNélio Laranjeiro 		}
1267230189d9SNélio Laranjeiro 		if (mpw.state == MLX5_MPW_STATE_CLOSED) {
1268a5bf6af9SAdrien Mazarguil 			if ((segs_n != 1) ||
1269a5bf6af9SAdrien Mazarguil 			    (length > inline_room)) {
1270f04f1d51SNélio Laranjeiro 				/*
1271f04f1d51SNélio Laranjeiro 				 * Multi-Packet WQE consumes at most two WQE.
1272f04f1d51SNélio Laranjeiro 				 * mlx5_mpw_new() expects to be able to use
1273f04f1d51SNélio Laranjeiro 				 * such resources.
1274f04f1d51SNélio Laranjeiro 				 */
1275f04f1d51SNélio Laranjeiro 				if (unlikely(max_wqe < 2))
1276f04f1d51SNélio Laranjeiro 					break;
1277f04f1d51SNélio Laranjeiro 				max_wqe -= 2;
1278230189d9SNélio Laranjeiro 				mlx5_mpw_new(txq, &mpw, length);
12798688b2f8SNélio Laranjeiro 				mpw.wqe->eseg.cs_flags = cs_flags;
1280230189d9SNélio Laranjeiro 			} else {
1281f04f1d51SNélio Laranjeiro 				if (unlikely(max_wqe < wqe_inl_n))
1282f04f1d51SNélio Laranjeiro 					break;
1283f04f1d51SNélio Laranjeiro 				max_wqe -= wqe_inl_n;
1284230189d9SNélio Laranjeiro 				mlx5_mpw_inline_new(txq, &mpw, length);
12858688b2f8SNélio Laranjeiro 				mpw.wqe->eseg.cs_flags = cs_flags;
1286230189d9SNélio Laranjeiro 			}
1287230189d9SNélio Laranjeiro 		}
1288a5bf6af9SAdrien Mazarguil 		/* Multi-segment packets must be alone in their MPW. */
1289a5bf6af9SAdrien Mazarguil 		assert((segs_n == 1) || (mpw.pkts_n == 0));
1290230189d9SNélio Laranjeiro 		if (mpw.state == MLX5_MPW_STATE_OPENED) {
12910e8679fcSNélio Laranjeiro 			assert(inline_room ==
12920e8679fcSNélio Laranjeiro 			       txq->max_inline * RTE_CACHE_LINE_SIZE);
1293a5bf6af9SAdrien Mazarguil #if defined(MLX5_PMD_SOFT_COUNTERS) || !defined(NDEBUG)
1294a5bf6af9SAdrien Mazarguil 			length = 0;
1295a5bf6af9SAdrien Mazarguil #endif
1296a5bf6af9SAdrien Mazarguil 			do {
1297230189d9SNélio Laranjeiro 				volatile struct mlx5_wqe_data_seg *dseg;
1298230189d9SNélio Laranjeiro 
1299a5bf6af9SAdrien Mazarguil 				elts_head_next =
1300a5bf6af9SAdrien Mazarguil 					(elts_head + 1) & (elts_n - 1);
1301a5bf6af9SAdrien Mazarguil 				assert(buf);
1302a5bf6af9SAdrien Mazarguil 				(*txq->elts)[elts_head] = buf;
1303230189d9SNélio Laranjeiro 				dseg = mpw.data.dseg[mpw.pkts_n];
1304a5bf6af9SAdrien Mazarguil 				addr = rte_pktmbuf_mtod(buf, uintptr_t);
1305230189d9SNélio Laranjeiro 				*dseg = (struct mlx5_wqe_data_seg){
1306a5bf6af9SAdrien Mazarguil 					.byte_count = htonl(DATA_LEN(buf)),
1307230189d9SNélio Laranjeiro 					.lkey = txq_mp2mr(txq, txq_mb2mp(buf)),
1308230189d9SNélio Laranjeiro 					.addr = htonll(addr),
1309230189d9SNélio Laranjeiro 				};
1310a5bf6af9SAdrien Mazarguil 				elts_head = elts_head_next;
1311a5bf6af9SAdrien Mazarguil #if defined(MLX5_PMD_SOFT_COUNTERS) || !defined(NDEBUG)
1312a5bf6af9SAdrien Mazarguil 				length += DATA_LEN(buf);
1313a5bf6af9SAdrien Mazarguil #endif
1314a5bf6af9SAdrien Mazarguil 				buf = buf->next;
1315230189d9SNélio Laranjeiro 				++mpw.pkts_n;
1316a5bf6af9SAdrien Mazarguil 				++j;
1317a5bf6af9SAdrien Mazarguil 			} while (--segs_n);
1318a5bf6af9SAdrien Mazarguil 			assert(length == mpw.len);
1319230189d9SNélio Laranjeiro 			if (mpw.pkts_n == MLX5_MPW_DSEG_MAX)
1320230189d9SNélio Laranjeiro 				mlx5_mpw_close(txq, &mpw);
1321230189d9SNélio Laranjeiro 		} else {
1322230189d9SNélio Laranjeiro 			unsigned int max;
1323230189d9SNélio Laranjeiro 
1324230189d9SNélio Laranjeiro 			assert(mpw.state == MLX5_MPW_INL_STATE_OPENED);
1325230189d9SNélio Laranjeiro 			assert(length <= inline_room);
1326a5bf6af9SAdrien Mazarguil 			assert(length == DATA_LEN(buf));
1327a5bf6af9SAdrien Mazarguil 			elts_head_next = (elts_head + 1) & (elts_n - 1);
1328a5bf6af9SAdrien Mazarguil 			addr = rte_pktmbuf_mtod(buf, uintptr_t);
1329a5bf6af9SAdrien Mazarguil 			(*txq->elts)[elts_head] = buf;
1330230189d9SNélio Laranjeiro 			/* Maximum number of bytes before wrapping. */
1331fdcb0f53SNélio Laranjeiro 			max = ((((uintptr_t)(txq->wqes)) +
1332fdcb0f53SNélio Laranjeiro 				(1 << txq->wqe_n) *
1333fdcb0f53SNélio Laranjeiro 				MLX5_WQE_SIZE) -
1334230189d9SNélio Laranjeiro 			       (uintptr_t)mpw.data.raw);
1335230189d9SNélio Laranjeiro 			if (length > max) {
1336230189d9SNélio Laranjeiro 				rte_memcpy((void *)(uintptr_t)mpw.data.raw,
1337230189d9SNélio Laranjeiro 					   (void *)addr,
1338230189d9SNélio Laranjeiro 					   max);
1339fdcb0f53SNélio Laranjeiro 				mpw.data.raw = (volatile void *)txq->wqes;
1340230189d9SNélio Laranjeiro 				rte_memcpy((void *)(uintptr_t)mpw.data.raw,
1341230189d9SNélio Laranjeiro 					   (void *)(addr + max),
1342230189d9SNélio Laranjeiro 					   length - max);
1343230189d9SNélio Laranjeiro 				mpw.data.raw += length - max;
1344230189d9SNélio Laranjeiro 			} else {
1345230189d9SNélio Laranjeiro 				rte_memcpy((void *)(uintptr_t)mpw.data.raw,
1346230189d9SNélio Laranjeiro 					   (void *)addr,
1347230189d9SNélio Laranjeiro 					   length);
134816c64768SYongseok Koh 
134916c64768SYongseok Koh 				if (length == max)
135016c64768SYongseok Koh 					mpw.data.raw =
135116c64768SYongseok Koh 						(volatile void *)txq->wqes;
135216c64768SYongseok Koh 				else
1353230189d9SNélio Laranjeiro 					mpw.data.raw += length;
1354230189d9SNélio Laranjeiro 			}
1355230189d9SNélio Laranjeiro 			++mpw.pkts_n;
135676bf1574SYongseok Koh 			mpw.total_len += length;
1357a5bf6af9SAdrien Mazarguil 			++j;
1358230189d9SNélio Laranjeiro 			if (mpw.pkts_n == MLX5_MPW_DSEG_MAX) {
1359230189d9SNélio Laranjeiro 				mlx5_mpw_inline_close(txq, &mpw);
13600e8679fcSNélio Laranjeiro 				inline_room =
13610e8679fcSNélio Laranjeiro 					txq->max_inline * RTE_CACHE_LINE_SIZE;
1362230189d9SNélio Laranjeiro 			} else {
1363230189d9SNélio Laranjeiro 				inline_room -= length;
1364230189d9SNélio Laranjeiro 			}
1365230189d9SNélio Laranjeiro 		}
1366230189d9SNélio Laranjeiro 		elts_head = elts_head_next;
1367230189d9SNélio Laranjeiro #ifdef MLX5_PMD_SOFT_COUNTERS
1368230189d9SNélio Laranjeiro 		/* Increment sent bytes counter. */
1369230189d9SNélio Laranjeiro 		txq->stats.obytes += length;
1370230189d9SNélio Laranjeiro #endif
1371c3d62cc9SAdrien Mazarguil 		++i;
1372c3d62cc9SAdrien Mazarguil 	} while (pkts_n);
1373230189d9SNélio Laranjeiro 	/* Take a shortcut if nothing must be sent. */
1374230189d9SNélio Laranjeiro 	if (unlikely(i == 0))
1375230189d9SNélio Laranjeiro 		return 0;
1376230189d9SNélio Laranjeiro 	/* Check whether completion threshold has been reached. */
1377a5bf6af9SAdrien Mazarguil 	/* "j" includes both packets and segments. */
1378a5bf6af9SAdrien Mazarguil 	comp = txq->elts_comp + j;
1379230189d9SNélio Laranjeiro 	if (comp >= MLX5_TX_COMP_THRESH) {
13808688b2f8SNélio Laranjeiro 		volatile struct mlx5_wqe *wqe = mpw.wqe;
1381230189d9SNélio Laranjeiro 
1382230189d9SNélio Laranjeiro 		/* Request completion on last WQE. */
13838688b2f8SNélio Laranjeiro 		wqe->ctrl[2] = htonl(8);
1384230189d9SNélio Laranjeiro 		/* Save elts_head in unused "immediate" field of WQE. */
13858688b2f8SNélio Laranjeiro 		wqe->ctrl[3] = elts_head;
1386230189d9SNélio Laranjeiro 		txq->elts_comp = 0;
1387230189d9SNélio Laranjeiro 	} else {
1388230189d9SNélio Laranjeiro 		txq->elts_comp = comp;
1389230189d9SNélio Laranjeiro 	}
1390230189d9SNélio Laranjeiro #ifdef MLX5_PMD_SOFT_COUNTERS
1391230189d9SNélio Laranjeiro 	/* Increment sent packets counter. */
1392230189d9SNélio Laranjeiro 	txq->stats.opackets += i;
1393230189d9SNélio Laranjeiro #endif
1394230189d9SNélio Laranjeiro 	/* Ring QP doorbell. */
1395230189d9SNélio Laranjeiro 	if (mpw.state == MLX5_MPW_INL_STATE_OPENED)
1396230189d9SNélio Laranjeiro 		mlx5_mpw_inline_close(txq, &mpw);
1397230189d9SNélio Laranjeiro 	else if (mpw.state == MLX5_MPW_STATE_OPENED)
1398230189d9SNélio Laranjeiro 		mlx5_mpw_close(txq, &mpw);
139930807f62SNélio Laranjeiro 	mlx5_tx_dbrec(txq, mpw.wqe);
1400230189d9SNélio Laranjeiro 	txq->elts_head = elts_head;
1401230189d9SNélio Laranjeiro 	return i;
1402230189d9SNélio Laranjeiro }
1403230189d9SNélio Laranjeiro 
1404230189d9SNélio Laranjeiro /**
14056ce84bd8SYongseok Koh  * Open an Enhanced MPW session.
14066ce84bd8SYongseok Koh  *
14076ce84bd8SYongseok Koh  * @param txq
14086ce84bd8SYongseok Koh  *   Pointer to TX queue structure.
14096ce84bd8SYongseok Koh  * @param mpw
14106ce84bd8SYongseok Koh  *   Pointer to MPW session structure.
14116ce84bd8SYongseok Koh  * @param length
14126ce84bd8SYongseok Koh  *   Packet length.
14136ce84bd8SYongseok Koh  */
14146ce84bd8SYongseok Koh static inline void
14156ce84bd8SYongseok Koh mlx5_empw_new(struct txq *txq, struct mlx5_mpw *mpw, int padding)
14166ce84bd8SYongseok Koh {
14176ce84bd8SYongseok Koh 	uint16_t idx = txq->wqe_ci & ((1 << txq->wqe_n) - 1);
14186ce84bd8SYongseok Koh 
14196ce84bd8SYongseok Koh 	mpw->state = MLX5_MPW_ENHANCED_STATE_OPENED;
14206ce84bd8SYongseok Koh 	mpw->pkts_n = 0;
14216ce84bd8SYongseok Koh 	mpw->total_len = sizeof(struct mlx5_wqe);
14226ce84bd8SYongseok Koh 	mpw->wqe = (volatile struct mlx5_wqe *)tx_mlx5_wqe(txq, idx);
14236ce84bd8SYongseok Koh 	mpw->wqe->ctrl[0] = htonl((MLX5_OPC_MOD_ENHANCED_MPSW << 24) |
14246ce84bd8SYongseok Koh 				  (txq->wqe_ci << 8) |
14256ce84bd8SYongseok Koh 				  MLX5_OPCODE_ENHANCED_MPSW);
14266ce84bd8SYongseok Koh 	mpw->wqe->ctrl[2] = 0;
14276ce84bd8SYongseok Koh 	mpw->wqe->ctrl[3] = 0;
14286ce84bd8SYongseok Koh 	memset((void *)(uintptr_t)&mpw->wqe->eseg, 0, MLX5_WQE_DWORD_SIZE);
14296ce84bd8SYongseok Koh 	if (unlikely(padding)) {
14306ce84bd8SYongseok Koh 		uintptr_t addr = (uintptr_t)(mpw->wqe + 1);
14316ce84bd8SYongseok Koh 
14326ce84bd8SYongseok Koh 		/* Pad the first 2 DWORDs with zero-length inline header. */
14336ce84bd8SYongseok Koh 		*(volatile uint32_t *)addr = htonl(MLX5_INLINE_SEG);
14346ce84bd8SYongseok Koh 		*(volatile uint32_t *)(addr + MLX5_WQE_DWORD_SIZE) =
14356ce84bd8SYongseok Koh 			htonl(MLX5_INLINE_SEG);
14366ce84bd8SYongseok Koh 		mpw->total_len += 2 * MLX5_WQE_DWORD_SIZE;
14376ce84bd8SYongseok Koh 		/* Start from the next WQEBB. */
14386ce84bd8SYongseok Koh 		mpw->data.raw = (volatile void *)(tx_mlx5_wqe(txq, idx + 1));
14396ce84bd8SYongseok Koh 	} else {
14406ce84bd8SYongseok Koh 		mpw->data.raw = (volatile void *)(mpw->wqe + 1);
14416ce84bd8SYongseok Koh 	}
14426ce84bd8SYongseok Koh }
14436ce84bd8SYongseok Koh 
14446ce84bd8SYongseok Koh /**
14456ce84bd8SYongseok Koh  * Close an Enhanced MPW session.
14466ce84bd8SYongseok Koh  *
14476ce84bd8SYongseok Koh  * @param txq
14486ce84bd8SYongseok Koh  *   Pointer to TX queue structure.
14496ce84bd8SYongseok Koh  * @param mpw
14506ce84bd8SYongseok Koh  *   Pointer to MPW session structure.
14516ce84bd8SYongseok Koh  *
14526ce84bd8SYongseok Koh  * @return
14536ce84bd8SYongseok Koh  *   Number of consumed WQEs.
14546ce84bd8SYongseok Koh  */
14556ce84bd8SYongseok Koh static inline uint16_t
14566ce84bd8SYongseok Koh mlx5_empw_close(struct txq *txq, struct mlx5_mpw *mpw)
14576ce84bd8SYongseok Koh {
14586ce84bd8SYongseok Koh 	uint16_t ret;
14596ce84bd8SYongseok Koh 
14606ce84bd8SYongseok Koh 	/* Store size in multiple of 16 bytes. Control and Ethernet segments
14616ce84bd8SYongseok Koh 	 * count as 2.
14626ce84bd8SYongseok Koh 	 */
14636ce84bd8SYongseok Koh 	mpw->wqe->ctrl[1] = htonl(txq->qp_num_8s | MLX5_WQE_DS(mpw->total_len));
14646ce84bd8SYongseok Koh 	mpw->state = MLX5_MPW_STATE_CLOSED;
14656ce84bd8SYongseok Koh 	ret = (mpw->total_len + (MLX5_WQE_SIZE - 1)) / MLX5_WQE_SIZE;
14666ce84bd8SYongseok Koh 	txq->wqe_ci += ret;
14676ce84bd8SYongseok Koh 	return ret;
14686ce84bd8SYongseok Koh }
14696ce84bd8SYongseok Koh 
14706ce84bd8SYongseok Koh /**
14716ce84bd8SYongseok Koh  * DPDK callback for TX with Enhanced MPW support.
14726ce84bd8SYongseok Koh  *
14736ce84bd8SYongseok Koh  * @param dpdk_txq
14746ce84bd8SYongseok Koh  *   Generic pointer to TX queue structure.
14756ce84bd8SYongseok Koh  * @param[in] pkts
14766ce84bd8SYongseok Koh  *   Packets to transmit.
14776ce84bd8SYongseok Koh  * @param pkts_n
14786ce84bd8SYongseok Koh  *   Number of packets in array.
14796ce84bd8SYongseok Koh  *
14806ce84bd8SYongseok Koh  * @return
14816ce84bd8SYongseok Koh  *   Number of packets successfully transmitted (<= pkts_n).
14826ce84bd8SYongseok Koh  */
14836ce84bd8SYongseok Koh uint16_t
14846ce84bd8SYongseok Koh mlx5_tx_burst_empw(void *dpdk_txq, struct rte_mbuf **pkts, uint16_t pkts_n)
14856ce84bd8SYongseok Koh {
14866ce84bd8SYongseok Koh 	struct txq *txq = (struct txq *)dpdk_txq;
14876ce84bd8SYongseok Koh 	uint16_t elts_head = txq->elts_head;
14886ce84bd8SYongseok Koh 	const unsigned int elts_n = 1 << txq->elts_n;
14896ce84bd8SYongseok Koh 	unsigned int i = 0;
14906ce84bd8SYongseok Koh 	unsigned int j = 0;
14916ce84bd8SYongseok Koh 	unsigned int max_elts;
14926ce84bd8SYongseok Koh 	uint16_t max_wqe;
14936ce84bd8SYongseok Koh 	unsigned int max_inline = txq->max_inline * RTE_CACHE_LINE_SIZE;
14946ce84bd8SYongseok Koh 	unsigned int mpw_room = 0;
14956ce84bd8SYongseok Koh 	unsigned int inl_pad = 0;
14966ce84bd8SYongseok Koh 	uint32_t inl_hdr;
14976ce84bd8SYongseok Koh 	struct mlx5_mpw mpw = {
14986ce84bd8SYongseok Koh 		.state = MLX5_MPW_STATE_CLOSED,
14996ce84bd8SYongseok Koh 	};
15006ce84bd8SYongseok Koh 
15016ce84bd8SYongseok Koh 	if (unlikely(!pkts_n))
15026ce84bd8SYongseok Koh 		return 0;
15036ce84bd8SYongseok Koh 	/* Start processing. */
15046ce84bd8SYongseok Koh 	txq_complete(txq);
15056ce84bd8SYongseok Koh 	max_elts = (elts_n - (elts_head - txq->elts_tail));
15066ce84bd8SYongseok Koh 	if (max_elts > elts_n)
15076ce84bd8SYongseok Koh 		max_elts -= elts_n;
15086ce84bd8SYongseok Koh 	/* A CQE slot must always be available. */
15096ce84bd8SYongseok Koh 	assert((1u << txq->cqe_n) - (txq->cq_pi - txq->cq_ci));
15106ce84bd8SYongseok Koh 	max_wqe = (1u << txq->wqe_n) - (txq->wqe_ci - txq->wqe_pi);
15116ce84bd8SYongseok Koh 	if (unlikely(!max_wqe))
15126ce84bd8SYongseok Koh 		return 0;
15136ce84bd8SYongseok Koh 	do {
15146ce84bd8SYongseok Koh 		struct rte_mbuf *buf = *(pkts++);
15156ce84bd8SYongseok Koh 		unsigned int elts_head_next;
15166ce84bd8SYongseok Koh 		uintptr_t addr;
15176ce84bd8SYongseok Koh 		uint64_t naddr;
15186ce84bd8SYongseok Koh 		unsigned int n;
15196ce84bd8SYongseok Koh 		unsigned int do_inline = 0; /* Whether inline is possible. */
15206ce84bd8SYongseok Koh 		uint32_t length;
15216ce84bd8SYongseok Koh 		unsigned int segs_n = buf->nb_segs;
15226ce84bd8SYongseok Koh 		uint32_t cs_flags = 0;
15236ce84bd8SYongseok Koh 
15246ce84bd8SYongseok Koh 		/*
15256ce84bd8SYongseok Koh 		 * Make sure there is enough room to store this packet and
15266ce84bd8SYongseok Koh 		 * that one ring entry remains unused.
15276ce84bd8SYongseok Koh 		 */
15286ce84bd8SYongseok Koh 		assert(segs_n);
15296ce84bd8SYongseok Koh 		if (max_elts - j < segs_n + 1)
15306ce84bd8SYongseok Koh 			break;
15316ce84bd8SYongseok Koh 		/* Do not bother with large packets MPW cannot handle. */
15326ce84bd8SYongseok Koh 		if (segs_n > MLX5_MPW_DSEG_MAX)
15336ce84bd8SYongseok Koh 			break;
15346ce84bd8SYongseok Koh 		/* Should we enable HW CKSUM offload. */
15356ce84bd8SYongseok Koh 		if (buf->ol_flags &
15366ce84bd8SYongseok Koh 		    (PKT_TX_IP_CKSUM | PKT_TX_TCP_CKSUM | PKT_TX_UDP_CKSUM))
15376ce84bd8SYongseok Koh 			cs_flags = MLX5_ETH_WQE_L3_CSUM | MLX5_ETH_WQE_L4_CSUM;
15386ce84bd8SYongseok Koh 		/* Retrieve packet information. */
15396ce84bd8SYongseok Koh 		length = PKT_LEN(buf);
15406ce84bd8SYongseok Koh 		/* Start new session if:
15416ce84bd8SYongseok Koh 		 * - multi-segment packet
15426ce84bd8SYongseok Koh 		 * - no space left even for a dseg
15436ce84bd8SYongseok Koh 		 * - next packet can be inlined with a new WQE
15446ce84bd8SYongseok Koh 		 * - cs_flag differs
15456ce84bd8SYongseok Koh 		 * It can't be MLX5_MPW_STATE_OPENED as always have a single
15466ce84bd8SYongseok Koh 		 * segmented packet.
15476ce84bd8SYongseok Koh 		 */
15486ce84bd8SYongseok Koh 		if (mpw.state == MLX5_MPW_ENHANCED_STATE_OPENED) {
15496ce84bd8SYongseok Koh 			if ((segs_n != 1) ||
15506ce84bd8SYongseok Koh 			    (inl_pad + sizeof(struct mlx5_wqe_data_seg) >
15516ce84bd8SYongseok Koh 			      mpw_room) ||
15526ce84bd8SYongseok Koh 			    (length <= txq->inline_max_packet_sz &&
15536ce84bd8SYongseok Koh 			     inl_pad + sizeof(inl_hdr) + length >
15546ce84bd8SYongseok Koh 			      mpw_room) ||
15556ce84bd8SYongseok Koh 			    (mpw.wqe->eseg.cs_flags != cs_flags))
15566ce84bd8SYongseok Koh 				max_wqe -= mlx5_empw_close(txq, &mpw);
15576ce84bd8SYongseok Koh 		}
15586ce84bd8SYongseok Koh 		if (unlikely(mpw.state == MLX5_MPW_STATE_CLOSED)) {
15596ce84bd8SYongseok Koh 			if (unlikely(segs_n != 1)) {
15606ce84bd8SYongseok Koh 				/* Fall back to legacy MPW.
15616ce84bd8SYongseok Koh 				 * A MPW session consumes 2 WQEs at most to
15626ce84bd8SYongseok Koh 				 * include MLX5_MPW_DSEG_MAX pointers.
15636ce84bd8SYongseok Koh 				 */
15646ce84bd8SYongseok Koh 				if (unlikely(max_wqe < 2))
15656ce84bd8SYongseok Koh 					break;
15666ce84bd8SYongseok Koh 				mlx5_mpw_new(txq, &mpw, length);
15676ce84bd8SYongseok Koh 			} else {
15686ce84bd8SYongseok Koh 				/* In Enhanced MPW, inline as much as the budget
15696ce84bd8SYongseok Koh 				 * is allowed. The remaining space is to be
15706ce84bd8SYongseok Koh 				 * filled with dsegs. If the title WQEBB isn't
15716ce84bd8SYongseok Koh 				 * padded, it will have 2 dsegs there.
15726ce84bd8SYongseok Koh 				 */
15736ce84bd8SYongseok Koh 				mpw_room = RTE_MIN(MLX5_WQE_SIZE_MAX,
15746ce84bd8SYongseok Koh 					    (max_inline ? max_inline :
15756ce84bd8SYongseok Koh 					     pkts_n * MLX5_WQE_DWORD_SIZE) +
15766ce84bd8SYongseok Koh 					    MLX5_WQE_SIZE);
15776ce84bd8SYongseok Koh 				if (unlikely(max_wqe * MLX5_WQE_SIZE <
15786ce84bd8SYongseok Koh 					      mpw_room))
15796ce84bd8SYongseok Koh 					break;
15806ce84bd8SYongseok Koh 				/* Don't pad the title WQEBB to not waste WQ. */
15816ce84bd8SYongseok Koh 				mlx5_empw_new(txq, &mpw, 0);
15826ce84bd8SYongseok Koh 				mpw_room -= mpw.total_len;
15836ce84bd8SYongseok Koh 				inl_pad = 0;
15846ce84bd8SYongseok Koh 				do_inline =
15856ce84bd8SYongseok Koh 					length <= txq->inline_max_packet_sz &&
15866ce84bd8SYongseok Koh 					sizeof(inl_hdr) + length <= mpw_room &&
15876ce84bd8SYongseok Koh 					!txq->mpw_hdr_dseg;
15886ce84bd8SYongseok Koh 			}
15896ce84bd8SYongseok Koh 			mpw.wqe->eseg.cs_flags = cs_flags;
15906ce84bd8SYongseok Koh 		} else {
15916ce84bd8SYongseok Koh 			/* Evaluate whether the next packet can be inlined.
15926ce84bd8SYongseok Koh 			 * Inlininig is possible when:
15936ce84bd8SYongseok Koh 			 * - length is less than configured value
15946ce84bd8SYongseok Koh 			 * - length fits for remaining space
15956ce84bd8SYongseok Koh 			 * - not required to fill the title WQEBB with dsegs
15966ce84bd8SYongseok Koh 			 */
15976ce84bd8SYongseok Koh 			do_inline =
15986ce84bd8SYongseok Koh 				length <= txq->inline_max_packet_sz &&
15996ce84bd8SYongseok Koh 				inl_pad + sizeof(inl_hdr) + length <=
16006ce84bd8SYongseok Koh 				 mpw_room &&
16016ce84bd8SYongseok Koh 				(!txq->mpw_hdr_dseg ||
16026ce84bd8SYongseok Koh 				 mpw.total_len >= MLX5_WQE_SIZE);
16036ce84bd8SYongseok Koh 		}
16046ce84bd8SYongseok Koh 		/* Multi-segment packets must be alone in their MPW. */
16056ce84bd8SYongseok Koh 		assert((segs_n == 1) || (mpw.pkts_n == 0));
16066ce84bd8SYongseok Koh 		if (unlikely(mpw.state == MLX5_MPW_STATE_OPENED)) {
16076ce84bd8SYongseok Koh #if defined(MLX5_PMD_SOFT_COUNTERS) || !defined(NDEBUG)
16086ce84bd8SYongseok Koh 			length = 0;
16096ce84bd8SYongseok Koh #endif
16106ce84bd8SYongseok Koh 			do {
16116ce84bd8SYongseok Koh 				volatile struct mlx5_wqe_data_seg *dseg;
16126ce84bd8SYongseok Koh 
16136ce84bd8SYongseok Koh 				elts_head_next =
16146ce84bd8SYongseok Koh 					(elts_head + 1) & (elts_n - 1);
16156ce84bd8SYongseok Koh 				assert(buf);
16166ce84bd8SYongseok Koh 				(*txq->elts)[elts_head] = buf;
16176ce84bd8SYongseok Koh 				dseg = mpw.data.dseg[mpw.pkts_n];
16186ce84bd8SYongseok Koh 				addr = rte_pktmbuf_mtod(buf, uintptr_t);
16196ce84bd8SYongseok Koh 				*dseg = (struct mlx5_wqe_data_seg){
16206ce84bd8SYongseok Koh 					.byte_count = htonl(DATA_LEN(buf)),
16216ce84bd8SYongseok Koh 					.lkey = txq_mp2mr(txq, txq_mb2mp(buf)),
16226ce84bd8SYongseok Koh 					.addr = htonll(addr),
16236ce84bd8SYongseok Koh 				};
16246ce84bd8SYongseok Koh 				elts_head = elts_head_next;
16256ce84bd8SYongseok Koh #if defined(MLX5_PMD_SOFT_COUNTERS) || !defined(NDEBUG)
16266ce84bd8SYongseok Koh 				length += DATA_LEN(buf);
16276ce84bd8SYongseok Koh #endif
16286ce84bd8SYongseok Koh 				buf = buf->next;
16296ce84bd8SYongseok Koh 				++j;
16306ce84bd8SYongseok Koh 				++mpw.pkts_n;
16316ce84bd8SYongseok Koh 			} while (--segs_n);
16326ce84bd8SYongseok Koh 			/* A multi-segmented packet takes one MPW session.
16336ce84bd8SYongseok Koh 			 * TODO: Pack more multi-segmented packets if possible.
16346ce84bd8SYongseok Koh 			 */
16356ce84bd8SYongseok Koh 			mlx5_mpw_close(txq, &mpw);
16366ce84bd8SYongseok Koh 			if (mpw.pkts_n < 3)
16376ce84bd8SYongseok Koh 				max_wqe--;
16386ce84bd8SYongseok Koh 			else
16396ce84bd8SYongseok Koh 				max_wqe -= 2;
16406ce84bd8SYongseok Koh 		} else if (do_inline) {
16416ce84bd8SYongseok Koh 			/* Inline packet into WQE. */
16426ce84bd8SYongseok Koh 			unsigned int max;
16436ce84bd8SYongseok Koh 
16446ce84bd8SYongseok Koh 			assert(mpw.state == MLX5_MPW_ENHANCED_STATE_OPENED);
16456ce84bd8SYongseok Koh 			assert(length == DATA_LEN(buf));
16466ce84bd8SYongseok Koh 			inl_hdr = htonl(length | MLX5_INLINE_SEG);
16476ce84bd8SYongseok Koh 			addr = rte_pktmbuf_mtod(buf, uintptr_t);
16486ce84bd8SYongseok Koh 			mpw.data.raw = (volatile void *)
16496ce84bd8SYongseok Koh 				((uintptr_t)mpw.data.raw + inl_pad);
16506ce84bd8SYongseok Koh 			max = tx_mlx5_wq_tailroom(txq,
16516ce84bd8SYongseok Koh 					(void *)(uintptr_t)mpw.data.raw);
16526ce84bd8SYongseok Koh 			/* Copy inline header. */
16536ce84bd8SYongseok Koh 			mpw.data.raw = (volatile void *)
16546ce84bd8SYongseok Koh 				mlx5_copy_to_wq(
16556ce84bd8SYongseok Koh 					  (void *)(uintptr_t)mpw.data.raw,
16566ce84bd8SYongseok Koh 					  &inl_hdr,
16576ce84bd8SYongseok Koh 					  sizeof(inl_hdr),
16586ce84bd8SYongseok Koh 					  (void *)(uintptr_t)txq->wqes,
16596ce84bd8SYongseok Koh 					  max);
16606ce84bd8SYongseok Koh 			max = tx_mlx5_wq_tailroom(txq,
16616ce84bd8SYongseok Koh 					(void *)(uintptr_t)mpw.data.raw);
16626ce84bd8SYongseok Koh 			/* Copy packet data. */
16636ce84bd8SYongseok Koh 			mpw.data.raw = (volatile void *)
16646ce84bd8SYongseok Koh 				mlx5_copy_to_wq(
16656ce84bd8SYongseok Koh 					  (void *)(uintptr_t)mpw.data.raw,
16666ce84bd8SYongseok Koh 					  (void *)addr,
16676ce84bd8SYongseok Koh 					  length,
16686ce84bd8SYongseok Koh 					  (void *)(uintptr_t)txq->wqes,
16696ce84bd8SYongseok Koh 					  max);
16706ce84bd8SYongseok Koh 			++mpw.pkts_n;
16716ce84bd8SYongseok Koh 			mpw.total_len += (inl_pad + sizeof(inl_hdr) + length);
16726ce84bd8SYongseok Koh 			/* No need to get completion as the entire packet is
16736ce84bd8SYongseok Koh 			 * copied to WQ. Free the buf right away.
16746ce84bd8SYongseok Koh 			 */
16756ce84bd8SYongseok Koh 			elts_head_next = elts_head;
16766ce84bd8SYongseok Koh 			rte_pktmbuf_free_seg(buf);
16776ce84bd8SYongseok Koh 			mpw_room -= (inl_pad + sizeof(inl_hdr) + length);
16786ce84bd8SYongseok Koh 			/* Add pad in the next packet if any. */
16796ce84bd8SYongseok Koh 			inl_pad = (((uintptr_t)mpw.data.raw +
16806ce84bd8SYongseok Koh 					(MLX5_WQE_DWORD_SIZE - 1)) &
16816ce84bd8SYongseok Koh 					~(MLX5_WQE_DWORD_SIZE - 1)) -
16826ce84bd8SYongseok Koh 				  (uintptr_t)mpw.data.raw;
16836ce84bd8SYongseok Koh 		} else {
16846ce84bd8SYongseok Koh 			/* No inline. Load a dseg of packet pointer. */
16856ce84bd8SYongseok Koh 			volatile rte_v128u32_t *dseg;
16866ce84bd8SYongseok Koh 
16876ce84bd8SYongseok Koh 			assert(mpw.state == MLX5_MPW_ENHANCED_STATE_OPENED);
16886ce84bd8SYongseok Koh 			assert((inl_pad + sizeof(*dseg)) <= mpw_room);
16896ce84bd8SYongseok Koh 			assert(length == DATA_LEN(buf));
16906ce84bd8SYongseok Koh 			if (!tx_mlx5_wq_tailroom(txq,
16916ce84bd8SYongseok Koh 					(void *)((uintptr_t)mpw.data.raw
16926ce84bd8SYongseok Koh 						+ inl_pad)))
16936ce84bd8SYongseok Koh 				dseg = (volatile void *)txq->wqes;
16946ce84bd8SYongseok Koh 			else
16956ce84bd8SYongseok Koh 				dseg = (volatile void *)
16966ce84bd8SYongseok Koh 					((uintptr_t)mpw.data.raw +
16976ce84bd8SYongseok Koh 					 inl_pad);
16986ce84bd8SYongseok Koh 			elts_head_next = (elts_head + 1) & (elts_n - 1);
16996ce84bd8SYongseok Koh 			(*txq->elts)[elts_head] = buf;
17006ce84bd8SYongseok Koh 			addr = rte_pktmbuf_mtod(buf, uintptr_t);
17016ce84bd8SYongseok Koh 			for (n = 0; n * RTE_CACHE_LINE_SIZE < length; n++)
17026ce84bd8SYongseok Koh 				rte_prefetch2((void *)(addr +
17036ce84bd8SYongseok Koh 						n * RTE_CACHE_LINE_SIZE));
17046ce84bd8SYongseok Koh 			naddr = htonll(addr);
17056ce84bd8SYongseok Koh 			*dseg = (rte_v128u32_t) {
17066ce84bd8SYongseok Koh 				htonl(length),
17076ce84bd8SYongseok Koh 				txq_mp2mr(txq, txq_mb2mp(buf)),
17086ce84bd8SYongseok Koh 				naddr,
17096ce84bd8SYongseok Koh 				naddr >> 32,
17106ce84bd8SYongseok Koh 			};
17116ce84bd8SYongseok Koh 			mpw.data.raw = (volatile void *)(dseg + 1);
17126ce84bd8SYongseok Koh 			mpw.total_len += (inl_pad + sizeof(*dseg));
17136ce84bd8SYongseok Koh 			++j;
17146ce84bd8SYongseok Koh 			++mpw.pkts_n;
17156ce84bd8SYongseok Koh 			mpw_room -= (inl_pad + sizeof(*dseg));
17166ce84bd8SYongseok Koh 			inl_pad = 0;
17176ce84bd8SYongseok Koh 		}
17186ce84bd8SYongseok Koh 		elts_head = elts_head_next;
17196ce84bd8SYongseok Koh #ifdef MLX5_PMD_SOFT_COUNTERS
17206ce84bd8SYongseok Koh 		/* Increment sent bytes counter. */
17216ce84bd8SYongseok Koh 		txq->stats.obytes += length;
17226ce84bd8SYongseok Koh #endif
17236ce84bd8SYongseok Koh 		++i;
17246ce84bd8SYongseok Koh 	} while (i < pkts_n);
17256ce84bd8SYongseok Koh 	/* Take a shortcut if nothing must be sent. */
17266ce84bd8SYongseok Koh 	if (unlikely(i == 0))
17276ce84bd8SYongseok Koh 		return 0;
17286ce84bd8SYongseok Koh 	/* Check whether completion threshold has been reached. */
17296ce84bd8SYongseok Koh 	if (txq->elts_comp + j >= MLX5_TX_COMP_THRESH ||
17306ce84bd8SYongseok Koh 			(uint16_t)(txq->wqe_ci - txq->mpw_comp) >=
17316ce84bd8SYongseok Koh 			 (1 << txq->wqe_n) / MLX5_TX_COMP_THRESH_INLINE_DIV) {
17326ce84bd8SYongseok Koh 		volatile struct mlx5_wqe *wqe = mpw.wqe;
17336ce84bd8SYongseok Koh 
17346ce84bd8SYongseok Koh 		/* Request completion on last WQE. */
17356ce84bd8SYongseok Koh 		wqe->ctrl[2] = htonl(8);
17366ce84bd8SYongseok Koh 		/* Save elts_head in unused "immediate" field of WQE. */
17376ce84bd8SYongseok Koh 		wqe->ctrl[3] = elts_head;
17386ce84bd8SYongseok Koh 		txq->elts_comp = 0;
17396ce84bd8SYongseok Koh 		txq->mpw_comp = txq->wqe_ci;
17406ce84bd8SYongseok Koh 		txq->cq_pi++;
17416ce84bd8SYongseok Koh 	} else {
17426ce84bd8SYongseok Koh 		txq->elts_comp += j;
17436ce84bd8SYongseok Koh 	}
17446ce84bd8SYongseok Koh #ifdef MLX5_PMD_SOFT_COUNTERS
17456ce84bd8SYongseok Koh 	/* Increment sent packets counter. */
17466ce84bd8SYongseok Koh 	txq->stats.opackets += i;
17476ce84bd8SYongseok Koh #endif
17486ce84bd8SYongseok Koh 	if (mpw.state == MLX5_MPW_ENHANCED_STATE_OPENED)
17496ce84bd8SYongseok Koh 		mlx5_empw_close(txq, &mpw);
17506ce84bd8SYongseok Koh 	else if (mpw.state == MLX5_MPW_STATE_OPENED)
17516ce84bd8SYongseok Koh 		mlx5_mpw_close(txq, &mpw);
17526ce84bd8SYongseok Koh 	/* Ring QP doorbell. */
17536ce84bd8SYongseok Koh 	mlx5_tx_dbrec(txq, mpw.wqe);
17546ce84bd8SYongseok Koh 	txq->elts_head = elts_head;
17556ce84bd8SYongseok Koh 	return i;
17566ce84bd8SYongseok Koh }
17576ce84bd8SYongseok Koh 
17586ce84bd8SYongseok Koh /**
175967fa62bcSAdrien Mazarguil  * Translate RX completion flags to packet type.
176067fa62bcSAdrien Mazarguil  *
17616218063bSNélio Laranjeiro  * @param[in] cqe
17626218063bSNélio Laranjeiro  *   Pointer to CQE.
176367fa62bcSAdrien Mazarguil  *
176478a38edfSJianfeng Tan  * @note: fix mlx5_dev_supported_ptypes_get() if any change here.
176578a38edfSJianfeng Tan  *
176667fa62bcSAdrien Mazarguil  * @return
176767fa62bcSAdrien Mazarguil  *   Packet type for struct rte_mbuf.
176867fa62bcSAdrien Mazarguil  */
176967fa62bcSAdrien Mazarguil static inline uint32_t
177097267b8eSNelio Laranjeiro rxq_cq_to_pkt_type(volatile struct mlx5_cqe *cqe)
177167fa62bcSAdrien Mazarguil {
177267fa62bcSAdrien Mazarguil 	uint32_t pkt_type;
17730603df73SNélio Laranjeiro 	uint16_t flags = ntohs(cqe->hdr_type_etc);
177467fa62bcSAdrien Mazarguil 
17750603df73SNélio Laranjeiro 	if (cqe->pkt_info & MLX5_CQE_RX_TUNNEL_PACKET) {
177667fa62bcSAdrien Mazarguil 		pkt_type =
177767fa62bcSAdrien Mazarguil 			TRANSPOSE(flags,
1778350f4c48SNelio Laranjeiro 				  MLX5_CQE_RX_IPV4_PACKET,
1779501505c5SMatthieu Ternisien d'Ouville 				  RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN) |
178067fa62bcSAdrien Mazarguil 			TRANSPOSE(flags,
1781350f4c48SNelio Laranjeiro 				  MLX5_CQE_RX_IPV6_PACKET,
1782501505c5SMatthieu Ternisien d'Ouville 				  RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN);
17830603df73SNélio Laranjeiro 		pkt_type |= ((cqe->pkt_info & MLX5_CQE_RX_OUTER_PACKET) ?
17840603df73SNélio Laranjeiro 			     RTE_PTYPE_L3_IPV6_EXT_UNKNOWN :
17850603df73SNélio Laranjeiro 			     RTE_PTYPE_L3_IPV4_EXT_UNKNOWN);
17860603df73SNélio Laranjeiro 	} else {
178767fa62bcSAdrien Mazarguil 		pkt_type =
178867fa62bcSAdrien Mazarguil 			TRANSPOSE(flags,
17896218063bSNélio Laranjeiro 				  MLX5_CQE_L3_HDR_TYPE_IPV6,
1790501505c5SMatthieu Ternisien d'Ouville 				  RTE_PTYPE_L3_IPV6_EXT_UNKNOWN) |
179167fa62bcSAdrien Mazarguil 			TRANSPOSE(flags,
17926218063bSNélio Laranjeiro 				  MLX5_CQE_L3_HDR_TYPE_IPV4,
1793501505c5SMatthieu Ternisien d'Ouville 				  RTE_PTYPE_L3_IPV4_EXT_UNKNOWN);
17940603df73SNélio Laranjeiro 	}
179567fa62bcSAdrien Mazarguil 	return pkt_type;
179667fa62bcSAdrien Mazarguil }
179767fa62bcSAdrien Mazarguil 
179867fa62bcSAdrien Mazarguil /**
179999c12dccSNélio Laranjeiro  * Get size of the next packet for a given CQE. For compressed CQEs, the
180099c12dccSNélio Laranjeiro  * consumer index is updated only once all packets of the current one have
180199c12dccSNélio Laranjeiro  * been processed.
180299c12dccSNélio Laranjeiro  *
180399c12dccSNélio Laranjeiro  * @param rxq
180499c12dccSNélio Laranjeiro  *   Pointer to RX queue.
180599c12dccSNélio Laranjeiro  * @param cqe
180699c12dccSNélio Laranjeiro  *   CQE to process.
1807ecf60761SNélio Laranjeiro  * @param[out] rss_hash
1808ecf60761SNélio Laranjeiro  *   Packet RSS Hash result.
180999c12dccSNélio Laranjeiro  *
181099c12dccSNélio Laranjeiro  * @return
181199c12dccSNélio Laranjeiro  *   Packet size in bytes (0 if there is none), -1 in case of completion
181299c12dccSNélio Laranjeiro  *   with error.
181399c12dccSNélio Laranjeiro  */
181499c12dccSNélio Laranjeiro static inline int
181597267b8eSNelio Laranjeiro mlx5_rx_poll_len(struct rxq *rxq, volatile struct mlx5_cqe *cqe,
1816ecf60761SNélio Laranjeiro 		 uint16_t cqe_cnt, uint32_t *rss_hash)
181799c12dccSNélio Laranjeiro {
181899c12dccSNélio Laranjeiro 	struct rxq_zip *zip = &rxq->zip;
181999c12dccSNélio Laranjeiro 	uint16_t cqe_n = cqe_cnt + 1;
182099c12dccSNélio Laranjeiro 	int len = 0;
1821d2e842d0SYongseok Koh 	uint16_t idx, end;
182299c12dccSNélio Laranjeiro 
182399c12dccSNélio Laranjeiro 	/* Process compressed data in the CQE and mini arrays. */
182499c12dccSNélio Laranjeiro 	if (zip->ai) {
182599c12dccSNélio Laranjeiro 		volatile struct mlx5_mini_cqe8 (*mc)[8] =
182699c12dccSNélio Laranjeiro 			(volatile struct mlx5_mini_cqe8 (*)[8])
182797267b8eSNelio Laranjeiro 			(uintptr_t)(&(*rxq->cqes)[zip->ca & cqe_cnt]);
182899c12dccSNélio Laranjeiro 
182999c12dccSNélio Laranjeiro 		len = ntohl((*mc)[zip->ai & 7].byte_cnt);
1830ecf60761SNélio Laranjeiro 		*rss_hash = ntohl((*mc)[zip->ai & 7].rx_hash_result);
183199c12dccSNélio Laranjeiro 		if ((++zip->ai & 7) == 0) {
1832d2e842d0SYongseok Koh 			/* Invalidate consumed CQEs */
1833d2e842d0SYongseok Koh 			idx = zip->ca;
1834d2e842d0SYongseok Koh 			end = zip->na;
1835d2e842d0SYongseok Koh 			while (idx != end) {
1836d2e842d0SYongseok Koh 				(*rxq->cqes)[idx & cqe_cnt].op_own =
1837d2e842d0SYongseok Koh 					MLX5_CQE_INVALIDATE;
1838d2e842d0SYongseok Koh 				++idx;
1839d2e842d0SYongseok Koh 			}
184099c12dccSNélio Laranjeiro 			/*
184199c12dccSNélio Laranjeiro 			 * Increment consumer index to skip the number of
184299c12dccSNélio Laranjeiro 			 * CQEs consumed. Hardware leaves holes in the CQ
184399c12dccSNélio Laranjeiro 			 * ring for software use.
184499c12dccSNélio Laranjeiro 			 */
184599c12dccSNélio Laranjeiro 			zip->ca = zip->na;
184699c12dccSNélio Laranjeiro 			zip->na += 8;
184799c12dccSNélio Laranjeiro 		}
184899c12dccSNélio Laranjeiro 		if (unlikely(rxq->zip.ai == rxq->zip.cqe_cnt)) {
1849d2e842d0SYongseok Koh 			/* Invalidate the rest */
1850d2e842d0SYongseok Koh 			idx = zip->ca;
1851d2e842d0SYongseok Koh 			end = zip->cq_ci;
185299c12dccSNélio Laranjeiro 
185399c12dccSNélio Laranjeiro 			while (idx != end) {
185497267b8eSNelio Laranjeiro 				(*rxq->cqes)[idx & cqe_cnt].op_own =
185599c12dccSNélio Laranjeiro 					MLX5_CQE_INVALIDATE;
185699c12dccSNélio Laranjeiro 				++idx;
185799c12dccSNélio Laranjeiro 			}
185899c12dccSNélio Laranjeiro 			rxq->cq_ci = zip->cq_ci;
185999c12dccSNélio Laranjeiro 			zip->ai = 0;
186099c12dccSNélio Laranjeiro 		}
186199c12dccSNélio Laranjeiro 	/* No compressed data, get next CQE and verify if it is compressed. */
186299c12dccSNélio Laranjeiro 	} else {
186399c12dccSNélio Laranjeiro 		int ret;
186499c12dccSNélio Laranjeiro 		int8_t op_own;
186599c12dccSNélio Laranjeiro 
186697267b8eSNelio Laranjeiro 		ret = check_cqe(cqe, cqe_n, rxq->cq_ci);
186799c12dccSNélio Laranjeiro 		if (unlikely(ret == 1))
186899c12dccSNélio Laranjeiro 			return 0;
186999c12dccSNélio Laranjeiro 		++rxq->cq_ci;
187099c12dccSNélio Laranjeiro 		op_own = cqe->op_own;
187199c12dccSNélio Laranjeiro 		if (MLX5_CQE_FORMAT(op_own) == MLX5_COMPRESSED) {
187299c12dccSNélio Laranjeiro 			volatile struct mlx5_mini_cqe8 (*mc)[8] =
187399c12dccSNélio Laranjeiro 				(volatile struct mlx5_mini_cqe8 (*)[8])
187499c12dccSNélio Laranjeiro 				(uintptr_t)(&(*rxq->cqes)[rxq->cq_ci &
187597267b8eSNelio Laranjeiro 							  cqe_cnt]);
187699c12dccSNélio Laranjeiro 
187799c12dccSNélio Laranjeiro 			/* Fix endianness. */
187899c12dccSNélio Laranjeiro 			zip->cqe_cnt = ntohl(cqe->byte_cnt);
187999c12dccSNélio Laranjeiro 			/*
188099c12dccSNélio Laranjeiro 			 * Current mini array position is the one returned by
188199c12dccSNélio Laranjeiro 			 * check_cqe64().
188299c12dccSNélio Laranjeiro 			 *
188399c12dccSNélio Laranjeiro 			 * If completion comprises several mini arrays, as a
188499c12dccSNélio Laranjeiro 			 * special case the second one is located 7 CQEs after
188599c12dccSNélio Laranjeiro 			 * the initial CQE instead of 8 for subsequent ones.
188699c12dccSNélio Laranjeiro 			 */
1887d2e842d0SYongseok Koh 			zip->ca = rxq->cq_ci;
188899c12dccSNélio Laranjeiro 			zip->na = zip->ca + 7;
188999c12dccSNélio Laranjeiro 			/* Compute the next non compressed CQE. */
189099c12dccSNélio Laranjeiro 			--rxq->cq_ci;
189199c12dccSNélio Laranjeiro 			zip->cq_ci = rxq->cq_ci + zip->cqe_cnt;
189299c12dccSNélio Laranjeiro 			/* Get packet size to return. */
189399c12dccSNélio Laranjeiro 			len = ntohl((*mc)[0].byte_cnt);
1894ecf60761SNélio Laranjeiro 			*rss_hash = ntohl((*mc)[0].rx_hash_result);
189599c12dccSNélio Laranjeiro 			zip->ai = 1;
1896d2e842d0SYongseok Koh 			/* Prefetch all the entries to be invalidated */
1897d2e842d0SYongseok Koh 			idx = zip->ca;
1898d2e842d0SYongseok Koh 			end = zip->cq_ci;
1899d2e842d0SYongseok Koh 			while (idx != end) {
1900d2e842d0SYongseok Koh 				rte_prefetch0(&(*rxq->cqes)[(idx) & cqe_cnt]);
1901d2e842d0SYongseok Koh 				++idx;
1902d2e842d0SYongseok Koh 			}
190399c12dccSNélio Laranjeiro 		} else {
190499c12dccSNélio Laranjeiro 			len = ntohl(cqe->byte_cnt);
1905ecf60761SNélio Laranjeiro 			*rss_hash = ntohl(cqe->rx_hash_res);
190699c12dccSNélio Laranjeiro 		}
190799c12dccSNélio Laranjeiro 		/* Error while receiving packet. */
190899c12dccSNélio Laranjeiro 		if (unlikely(MLX5_CQE_OPCODE(op_own) == MLX5_CQE_RESP_ERR))
190999c12dccSNélio Laranjeiro 			return -1;
191099c12dccSNélio Laranjeiro 	}
191199c12dccSNélio Laranjeiro 	return len;
191299c12dccSNélio Laranjeiro }
191399c12dccSNélio Laranjeiro 
191499c12dccSNélio Laranjeiro /**
191567fa62bcSAdrien Mazarguil  * Translate RX completion flags to offload flags.
191667fa62bcSAdrien Mazarguil  *
191767fa62bcSAdrien Mazarguil  * @param[in] rxq
191867fa62bcSAdrien Mazarguil  *   Pointer to RX queue structure.
19196218063bSNélio Laranjeiro  * @param[in] cqe
19206218063bSNélio Laranjeiro  *   Pointer to CQE.
192167fa62bcSAdrien Mazarguil  *
192267fa62bcSAdrien Mazarguil  * @return
192367fa62bcSAdrien Mazarguil  *   Offload flags (ol_flags) for struct rte_mbuf.
192467fa62bcSAdrien Mazarguil  */
192567fa62bcSAdrien Mazarguil static inline uint32_t
192697267b8eSNelio Laranjeiro rxq_cq_to_ol_flags(struct rxq *rxq, volatile struct mlx5_cqe *cqe)
192767fa62bcSAdrien Mazarguil {
192867fa62bcSAdrien Mazarguil 	uint32_t ol_flags = 0;
19290603df73SNélio Laranjeiro 	uint16_t flags = ntohs(cqe->hdr_type_etc);
193067fa62bcSAdrien Mazarguil 
19310603df73SNélio Laranjeiro 	ol_flags =
19320603df73SNélio Laranjeiro 		TRANSPOSE(flags,
19330603df73SNélio Laranjeiro 			  MLX5_CQE_RX_L3_HDR_VALID,
19340603df73SNélio Laranjeiro 			  PKT_RX_IP_CKSUM_GOOD) |
19350603df73SNélio Laranjeiro 		TRANSPOSE(flags,
19360603df73SNélio Laranjeiro 			  MLX5_CQE_RX_L4_HDR_VALID,
193783e9d9a3SNelio Laranjeiro 			  PKT_RX_L4_CKSUM_GOOD);
193897267b8eSNelio Laranjeiro 	if ((cqe->pkt_info & MLX5_CQE_RX_TUNNEL_PACKET) && (rxq->csum_l2tun))
193967fa62bcSAdrien Mazarguil 		ol_flags |=
19400603df73SNélio Laranjeiro 			TRANSPOSE(flags,
19410603df73SNélio Laranjeiro 				  MLX5_CQE_RX_L3_HDR_VALID,
194283e9d9a3SNelio Laranjeiro 				  PKT_RX_IP_CKSUM_GOOD) |
19430603df73SNélio Laranjeiro 			TRANSPOSE(flags,
19440603df73SNélio Laranjeiro 				  MLX5_CQE_RX_L4_HDR_VALID,
194583e9d9a3SNelio Laranjeiro 				  PKT_RX_L4_CKSUM_GOOD);
194667fa62bcSAdrien Mazarguil 	return ol_flags;
194767fa62bcSAdrien Mazarguil }
194867fa62bcSAdrien Mazarguil 
194967fa62bcSAdrien Mazarguil /**
19502e22920bSAdrien Mazarguil  * DPDK callback for RX.
19512e22920bSAdrien Mazarguil  *
19522e22920bSAdrien Mazarguil  * @param dpdk_rxq
19532e22920bSAdrien Mazarguil  *   Generic pointer to RX queue structure.
19542e22920bSAdrien Mazarguil  * @param[out] pkts
19552e22920bSAdrien Mazarguil  *   Array to store received packets.
19562e22920bSAdrien Mazarguil  * @param pkts_n
19572e22920bSAdrien Mazarguil  *   Maximum number of packets in array.
19582e22920bSAdrien Mazarguil  *
19592e22920bSAdrien Mazarguil  * @return
19602e22920bSAdrien Mazarguil  *   Number of packets successfully received (<= pkts_n).
19612e22920bSAdrien Mazarguil  */
19622e22920bSAdrien Mazarguil uint16_t
19632e22920bSAdrien Mazarguil mlx5_rx_burst(void *dpdk_rxq, struct rte_mbuf **pkts, uint16_t pkts_n)
19642e22920bSAdrien Mazarguil {
19656218063bSNélio Laranjeiro 	struct rxq *rxq = dpdk_rxq;
1966b4b12e55SNélio Laranjeiro 	const unsigned int wqe_cnt = (1 << rxq->elts_n) - 1;
1967e2f116eeSNélio Laranjeiro 	const unsigned int cqe_cnt = (1 << rxq->cqe_n) - 1;
19689964b965SNélio Laranjeiro 	const unsigned int sges_n = rxq->sges_n;
19699964b965SNélio Laranjeiro 	struct rte_mbuf *pkt = NULL;
19709964b965SNélio Laranjeiro 	struct rte_mbuf *seg = NULL;
197197267b8eSNelio Laranjeiro 	volatile struct mlx5_cqe *cqe =
197297267b8eSNelio Laranjeiro 		&(*rxq->cqes)[rxq->cq_ci & cqe_cnt];
19739964b965SNélio Laranjeiro 	unsigned int i = 0;
19749964b965SNélio Laranjeiro 	unsigned int rq_ci = rxq->rq_ci << sges_n;
19754e66a6feSNelio Laranjeiro 	int len = 0; /* keep its value across iterations. */
19762e22920bSAdrien Mazarguil 
19779964b965SNélio Laranjeiro 	while (pkts_n) {
19789964b965SNélio Laranjeiro 		unsigned int idx = rq_ci & wqe_cnt;
19799964b965SNélio Laranjeiro 		volatile struct mlx5_wqe_data_seg *wqe = &(*rxq->wqes)[idx];
19809964b965SNélio Laranjeiro 		struct rte_mbuf *rep = (*rxq->elts)[idx];
1981ecf60761SNélio Laranjeiro 		uint32_t rss_hash_res = 0;
19829964b965SNélio Laranjeiro 
19839964b965SNélio Laranjeiro 		if (pkt)
19849964b965SNélio Laranjeiro 			NEXT(seg) = rep;
19859964b965SNélio Laranjeiro 		seg = rep;
19869964b965SNélio Laranjeiro 		rte_prefetch0(seg);
19876218063bSNélio Laranjeiro 		rte_prefetch0(cqe);
19889964b965SNélio Laranjeiro 		rte_prefetch0(wqe);
1989fbfd9955SOlivier Matz 		rep = rte_mbuf_raw_alloc(rxq->mp);
19902e22920bSAdrien Mazarguil 		if (unlikely(rep == NULL)) {
199115a756b6SSagi Grimberg 			++rxq->stats.rx_nombuf;
199215a756b6SSagi Grimberg 			if (!pkt) {
199315a756b6SSagi Grimberg 				/*
199415a756b6SSagi Grimberg 				 * no buffers before we even started,
199515a756b6SSagi Grimberg 				 * bail out silently.
199615a756b6SSagi Grimberg 				 */
199715a756b6SSagi Grimberg 				break;
199815a756b6SSagi Grimberg 			}
1999a1bdb71aSNélio Laranjeiro 			while (pkt != seg) {
2000a1bdb71aSNélio Laranjeiro 				assert(pkt != (*rxq->elts)[idx]);
2001fe5fe382SNélio Laranjeiro 				rep = NEXT(pkt);
20028f094a9aSOlivier Matz 				NEXT(pkt) = NULL;
20038f094a9aSOlivier Matz 				NB_SEGS(pkt) = 1;
20041f88c0a2SOlivier Matz 				rte_mbuf_raw_free(pkt);
2005fe5fe382SNélio Laranjeiro 				pkt = rep;
20069964b965SNélio Laranjeiro 			}
20076218063bSNélio Laranjeiro 			break;
20082e22920bSAdrien Mazarguil 		}
20099964b965SNélio Laranjeiro 		if (!pkt) {
201097267b8eSNelio Laranjeiro 			cqe = &(*rxq->cqes)[rxq->cq_ci & cqe_cnt];
2011ecf60761SNélio Laranjeiro 			len = mlx5_rx_poll_len(rxq, cqe, cqe_cnt,
2012ecf60761SNélio Laranjeiro 					       &rss_hash_res);
2013ecf60761SNélio Laranjeiro 			if (!len) {
20141f88c0a2SOlivier Matz 				rte_mbuf_raw_free(rep);
20156218063bSNélio Laranjeiro 				break;
20166218063bSNélio Laranjeiro 			}
201799c12dccSNélio Laranjeiro 			if (unlikely(len == -1)) {
201899c12dccSNélio Laranjeiro 				/* RX error, packet is likely too large. */
20191f88c0a2SOlivier Matz 				rte_mbuf_raw_free(rep);
202099c12dccSNélio Laranjeiro 				++rxq->stats.idropped;
202199c12dccSNélio Laranjeiro 				goto skip;
202299c12dccSNélio Laranjeiro 			}
20239964b965SNélio Laranjeiro 			pkt = seg;
20249964b965SNélio Laranjeiro 			assert(len >= (rxq->crc_present << 2));
20259964b965SNélio Laranjeiro 			/* Update packet information. */
20260ac64846SMaxime Leroy 			pkt->packet_type = 0;
20270ac64846SMaxime Leroy 			pkt->ol_flags = 0;
202836ba0c00SNélio Laranjeiro 			if (rss_hash_res && rxq->rss_hash) {
2029ecf60761SNélio Laranjeiro 				pkt->hash.rss = rss_hash_res;
2030ecf60761SNélio Laranjeiro 				pkt->ol_flags = PKT_RX_RSS_HASH;
2031ecf60761SNélio Laranjeiro 			}
2032c604f619SNélio Laranjeiro 			if (rxq->mark &&
2033c604f619SNélio Laranjeiro 			    MLX5_FLOW_MARK_IS_VALID(cqe->sop_drop_qpn)) {
2034b268a3eeSNélio Laranjeiro 				pkt->ol_flags |= PKT_RX_FDIR;
2035b268a3eeSNélio Laranjeiro 				if (cqe->sop_drop_qpn !=
2036b268a3eeSNélio Laranjeiro 				    htonl(MLX5_FLOW_MARK_DEFAULT)) {
2037b268a3eeSNélio Laranjeiro 					uint32_t mark = cqe->sop_drop_qpn;
2038b268a3eeSNélio Laranjeiro 
2039b268a3eeSNélio Laranjeiro 					pkt->ol_flags |= PKT_RX_FDIR_ID;
2040ea3bc3b1SNélio Laranjeiro 					pkt->hash.fdir.hi =
2041b268a3eeSNélio Laranjeiro 						mlx5_flow_mark_get(mark);
2042b268a3eeSNélio Laranjeiro 				}
2043ea3bc3b1SNélio Laranjeiro 			}
2044*6703d836SNélio Laranjeiro 			if (rxq->csum | rxq->csum_l2tun) {
2045*6703d836SNélio Laranjeiro 				pkt->packet_type = rxq_cq_to_pkt_type(cqe);
2046*6703d836SNélio Laranjeiro 				pkt->ol_flags |= rxq_cq_to_ol_flags(rxq, cqe);
20476218063bSNélio Laranjeiro 			}
2048*6703d836SNélio Laranjeiro 			if (rxq->vlan_strip &&
2049*6703d836SNélio Laranjeiro 			    (cqe->hdr_type_etc &
2050*6703d836SNélio Laranjeiro 			     htons(MLX5_CQE_VLAN_STRIPPED))) {
20516218063bSNélio Laranjeiro 				pkt->ol_flags |= PKT_RX_VLAN_PKT |
2052b37b528dSOlivier Matz 					PKT_RX_VLAN_STRIPPED;
20536218063bSNélio Laranjeiro 				pkt->vlan_tci = ntohs(cqe->vlan_info);
2054f3db9489SYaacov Hazan 			}
20556218063bSNélio Laranjeiro 			if (rxq->crc_present)
20566218063bSNélio Laranjeiro 				len -= ETHER_CRC_LEN;
20576218063bSNélio Laranjeiro 			PKT_LEN(pkt) = len;
20589964b965SNélio Laranjeiro 		}
20599964b965SNélio Laranjeiro 		DATA_LEN(rep) = DATA_LEN(seg);
20609964b965SNélio Laranjeiro 		PKT_LEN(rep) = PKT_LEN(seg);
20619964b965SNélio Laranjeiro 		SET_DATA_OFF(rep, DATA_OFF(seg));
20629964b965SNélio Laranjeiro 		NB_SEGS(rep) = NB_SEGS(seg);
20639964b965SNélio Laranjeiro 		PORT(rep) = PORT(seg);
20649964b965SNélio Laranjeiro 		NEXT(rep) = NULL;
20659964b965SNélio Laranjeiro 		(*rxq->elts)[idx] = rep;
20669964b965SNélio Laranjeiro 		/*
20679964b965SNélio Laranjeiro 		 * Fill NIC descriptor with the new buffer.  The lkey and size
20689964b965SNélio Laranjeiro 		 * of the buffers are already known, only the buffer address
20699964b965SNélio Laranjeiro 		 * changes.
20709964b965SNélio Laranjeiro 		 */
20719964b965SNélio Laranjeiro 		wqe->addr = htonll(rte_pktmbuf_mtod(rep, uintptr_t));
20729964b965SNélio Laranjeiro 		if (len > DATA_LEN(seg)) {
20739964b965SNélio Laranjeiro 			len -= DATA_LEN(seg);
20749964b965SNélio Laranjeiro 			++NB_SEGS(pkt);
20759964b965SNélio Laranjeiro 			++rq_ci;
20769964b965SNélio Laranjeiro 			continue;
20779964b965SNélio Laranjeiro 		}
20789964b965SNélio Laranjeiro 		DATA_LEN(seg) = len;
207987011737SAdrien Mazarguil #ifdef MLX5_PMD_SOFT_COUNTERS
208087011737SAdrien Mazarguil 		/* Increment bytes counter. */
20819964b965SNélio Laranjeiro 		rxq->stats.ibytes += PKT_LEN(pkt);
208287011737SAdrien Mazarguil #endif
20836218063bSNélio Laranjeiro 		/* Return packet. */
20846218063bSNélio Laranjeiro 		*(pkts++) = pkt;
20859964b965SNélio Laranjeiro 		pkt = NULL;
20869964b965SNélio Laranjeiro 		--pkts_n;
20879964b965SNélio Laranjeiro 		++i;
208899c12dccSNélio Laranjeiro skip:
20899964b965SNélio Laranjeiro 		/* Align consumer index to the next stride. */
20909964b965SNélio Laranjeiro 		rq_ci >>= sges_n;
20916218063bSNélio Laranjeiro 		++rq_ci;
20929964b965SNélio Laranjeiro 		rq_ci <<= sges_n;
20932e22920bSAdrien Mazarguil 	}
20949964b965SNélio Laranjeiro 	if (unlikely((i == 0) && ((rq_ci >> sges_n) == rxq->rq_ci)))
20952e22920bSAdrien Mazarguil 		return 0;
20966218063bSNélio Laranjeiro 	/* Update the consumer index. */
20979964b965SNélio Laranjeiro 	rxq->rq_ci = rq_ci >> sges_n;
20986218063bSNélio Laranjeiro 	rte_wmb();
20996218063bSNélio Laranjeiro 	*rxq->cq_db = htonl(rxq->cq_ci);
21006218063bSNélio Laranjeiro 	rte_wmb();
21016218063bSNélio Laranjeiro 	*rxq->rq_db = htonl(rxq->rq_ci);
210287011737SAdrien Mazarguil #ifdef MLX5_PMD_SOFT_COUNTERS
210387011737SAdrien Mazarguil 	/* Increment packets counter. */
21049964b965SNélio Laranjeiro 	rxq->stats.ipackets += i;
210587011737SAdrien Mazarguil #endif
21069964b965SNélio Laranjeiro 	return i;
21072e22920bSAdrien Mazarguil }
21082e22920bSAdrien Mazarguil 
21092e22920bSAdrien Mazarguil /**
21102e22920bSAdrien Mazarguil  * Dummy DPDK callback for TX.
21112e22920bSAdrien Mazarguil  *
21122e22920bSAdrien Mazarguil  * This function is used to temporarily replace the real callback during
21132e22920bSAdrien Mazarguil  * unsafe control operations on the queue, or in case of error.
21142e22920bSAdrien Mazarguil  *
21152e22920bSAdrien Mazarguil  * @param dpdk_txq
21162e22920bSAdrien Mazarguil  *   Generic pointer to TX queue structure.
21172e22920bSAdrien Mazarguil  * @param[in] pkts
21182e22920bSAdrien Mazarguil  *   Packets to transmit.
21192e22920bSAdrien Mazarguil  * @param pkts_n
21202e22920bSAdrien Mazarguil  *   Number of packets in array.
21212e22920bSAdrien Mazarguil  *
21222e22920bSAdrien Mazarguil  * @return
21232e22920bSAdrien Mazarguil  *   Number of packets successfully transmitted (<= pkts_n).
21242e22920bSAdrien Mazarguil  */
21252e22920bSAdrien Mazarguil uint16_t
21262e22920bSAdrien Mazarguil removed_tx_burst(void *dpdk_txq, struct rte_mbuf **pkts, uint16_t pkts_n)
21272e22920bSAdrien Mazarguil {
21282e22920bSAdrien Mazarguil 	(void)dpdk_txq;
21292e22920bSAdrien Mazarguil 	(void)pkts;
21302e22920bSAdrien Mazarguil 	(void)pkts_n;
21312e22920bSAdrien Mazarguil 	return 0;
21322e22920bSAdrien Mazarguil }
21332e22920bSAdrien Mazarguil 
21342e22920bSAdrien Mazarguil /**
21352e22920bSAdrien Mazarguil  * Dummy DPDK callback for RX.
21362e22920bSAdrien Mazarguil  *
21372e22920bSAdrien Mazarguil  * This function is used to temporarily replace the real callback during
21382e22920bSAdrien Mazarguil  * unsafe control operations on the queue, or in case of error.
21392e22920bSAdrien Mazarguil  *
21402e22920bSAdrien Mazarguil  * @param dpdk_rxq
21412e22920bSAdrien Mazarguil  *   Generic pointer to RX queue structure.
21422e22920bSAdrien Mazarguil  * @param[out] pkts
21432e22920bSAdrien Mazarguil  *   Array to store received packets.
21442e22920bSAdrien Mazarguil  * @param pkts_n
21452e22920bSAdrien Mazarguil  *   Maximum number of packets in array.
21462e22920bSAdrien Mazarguil  *
21472e22920bSAdrien Mazarguil  * @return
21482e22920bSAdrien Mazarguil  *   Number of packets successfully received (<= pkts_n).
21492e22920bSAdrien Mazarguil  */
21502e22920bSAdrien Mazarguil uint16_t
21512e22920bSAdrien Mazarguil removed_rx_burst(void *dpdk_rxq, struct rte_mbuf **pkts, uint16_t pkts_n)
21522e22920bSAdrien Mazarguil {
21532e22920bSAdrien Mazarguil 	(void)dpdk_rxq;
21542e22920bSAdrien Mazarguil 	(void)pkts;
21552e22920bSAdrien Mazarguil 	(void)pkts_n;
21562e22920bSAdrien Mazarguil 	return 0;
21572e22920bSAdrien Mazarguil }
21583c7d44afSShahaf Shuler 
21593c7d44afSShahaf Shuler /**
21603c7d44afSShahaf Shuler  * DPDK callback for rx queue interrupt enable.
21613c7d44afSShahaf Shuler  *
21623c7d44afSShahaf Shuler  * @param dev
21633c7d44afSShahaf Shuler  *   Pointer to Ethernet device structure.
21643c7d44afSShahaf Shuler  * @param rx_queue_id
21653c7d44afSShahaf Shuler  *   RX queue number
21663c7d44afSShahaf Shuler  *
21673c7d44afSShahaf Shuler  * @return
21683c7d44afSShahaf Shuler  *   0 on success, negative on failure.
21693c7d44afSShahaf Shuler  */
21703c7d44afSShahaf Shuler int
21713c7d44afSShahaf Shuler mlx5_rx_intr_enable(struct rte_eth_dev *dev, uint16_t rx_queue_id)
21723c7d44afSShahaf Shuler {
21733c7d44afSShahaf Shuler #ifdef HAVE_UPDATE_CQ_CI
21743c7d44afSShahaf Shuler 	struct priv *priv = mlx5_get_priv(dev);
21753c7d44afSShahaf Shuler 	struct rxq *rxq = (*priv->rxqs)[rx_queue_id];
21763c7d44afSShahaf Shuler 	struct rxq_ctrl *rxq_ctrl = container_of(rxq, struct rxq_ctrl, rxq);
21773c7d44afSShahaf Shuler 	struct ibv_cq *cq = rxq_ctrl->cq;
21783c7d44afSShahaf Shuler 	uint16_t ci = rxq->cq_ci;
21793c7d44afSShahaf Shuler 	int ret = 0;
21803c7d44afSShahaf Shuler 
21813c7d44afSShahaf Shuler 	ibv_mlx5_exp_update_cq_ci(cq, ci);
21823c7d44afSShahaf Shuler 	ret = ibv_req_notify_cq(cq, 0);
21833c7d44afSShahaf Shuler #else
21843c7d44afSShahaf Shuler 	int ret = -1;
21853c7d44afSShahaf Shuler 	(void)dev;
21863c7d44afSShahaf Shuler 	(void)rx_queue_id;
21873c7d44afSShahaf Shuler #endif
21883c7d44afSShahaf Shuler 	if (ret)
21893c7d44afSShahaf Shuler 		WARN("unable to arm interrupt on rx queue %d", rx_queue_id);
21903c7d44afSShahaf Shuler 	return ret;
21913c7d44afSShahaf Shuler }
21923c7d44afSShahaf Shuler 
21933c7d44afSShahaf Shuler /**
21943c7d44afSShahaf Shuler  * DPDK callback for rx queue interrupt disable.
21953c7d44afSShahaf Shuler  *
21963c7d44afSShahaf Shuler  * @param dev
21973c7d44afSShahaf Shuler  *   Pointer to Ethernet device structure.
21983c7d44afSShahaf Shuler  * @param rx_queue_id
21993c7d44afSShahaf Shuler  *   RX queue number
22003c7d44afSShahaf Shuler  *
22013c7d44afSShahaf Shuler  * @return
22023c7d44afSShahaf Shuler  *   0 on success, negative on failure.
22033c7d44afSShahaf Shuler  */
22043c7d44afSShahaf Shuler int
22053c7d44afSShahaf Shuler mlx5_rx_intr_disable(struct rte_eth_dev *dev, uint16_t rx_queue_id)
22063c7d44afSShahaf Shuler {
22073c7d44afSShahaf Shuler #ifdef HAVE_UPDATE_CQ_CI
22083c7d44afSShahaf Shuler 	struct priv *priv = mlx5_get_priv(dev);
22093c7d44afSShahaf Shuler 	struct rxq *rxq = (*priv->rxqs)[rx_queue_id];
22103c7d44afSShahaf Shuler 	struct rxq_ctrl *rxq_ctrl = container_of(rxq, struct rxq_ctrl, rxq);
22113c7d44afSShahaf Shuler 	struct ibv_cq *cq = rxq_ctrl->cq;
22123c7d44afSShahaf Shuler 	struct ibv_cq *ev_cq;
22133c7d44afSShahaf Shuler 	void *ev_ctx;
22143c7d44afSShahaf Shuler 	int ret = 0;
22153c7d44afSShahaf Shuler 
22163c7d44afSShahaf Shuler 	ret = ibv_get_cq_event(cq->channel, &ev_cq, &ev_ctx);
22173c7d44afSShahaf Shuler 	if (ret || ev_cq != cq)
22183c7d44afSShahaf Shuler 		ret = -1;
22193c7d44afSShahaf Shuler 	else
22203c7d44afSShahaf Shuler 		ibv_ack_cq_events(cq, 1);
22213c7d44afSShahaf Shuler #else
22223c7d44afSShahaf Shuler 	int ret = -1;
22233c7d44afSShahaf Shuler 	(void)dev;
22243c7d44afSShahaf Shuler 	(void)rx_queue_id;
22253c7d44afSShahaf Shuler #endif
22263c7d44afSShahaf Shuler 	if (ret)
22273c7d44afSShahaf Shuler 		WARN("unable to disable interrupt on rx queue %d",
22283c7d44afSShahaf Shuler 		     rx_queue_id);
22293c7d44afSShahaf Shuler 	return ret;
22303c7d44afSShahaf Shuler }
2231