xref: /dpdk/drivers/net/mlx5/mlx5_rxtx.c (revision 7d6bf6b866b8c25ec06539b3eeed1db4f785577c)
18fd92a66SOlivier Matz /* SPDX-License-Identifier: BSD-3-Clause
22e22920bSAdrien Mazarguil  * Copyright 2015 6WIND S.A.
35feecc57SShahaf Shuler  * Copyright 2015 Mellanox Technologies, Ltd
42e22920bSAdrien Mazarguil  */
52e22920bSAdrien Mazarguil 
62e22920bSAdrien Mazarguil #include <assert.h>
72e22920bSAdrien Mazarguil #include <stdint.h>
82e22920bSAdrien Mazarguil #include <string.h>
92e22920bSAdrien Mazarguil #include <stdlib.h>
102e22920bSAdrien Mazarguil 
112e22920bSAdrien Mazarguil /* Verbs header. */
122e22920bSAdrien Mazarguil /* ISO C doesn't support unnamed structs/unions, disabling -pedantic. */
132e22920bSAdrien Mazarguil #ifdef PEDANTIC
14fc5b160fSBruce Richardson #pragma GCC diagnostic ignored "-Wpedantic"
152e22920bSAdrien Mazarguil #endif
162e22920bSAdrien Mazarguil #include <infiniband/verbs.h>
1743e9d979SShachar Beiser #include <infiniband/mlx5dv.h>
182e22920bSAdrien Mazarguil #ifdef PEDANTIC
19fc5b160fSBruce Richardson #pragma GCC diagnostic error "-Wpedantic"
202e22920bSAdrien Mazarguil #endif
212e22920bSAdrien Mazarguil 
222e22920bSAdrien Mazarguil #include <rte_mbuf.h>
232e22920bSAdrien Mazarguil #include <rte_mempool.h>
242e22920bSAdrien Mazarguil #include <rte_prefetch.h>
252e22920bSAdrien Mazarguil #include <rte_common.h>
262e22920bSAdrien Mazarguil #include <rte_branch_prediction.h>
276218063bSNélio Laranjeiro #include <rte_ether.h>
282e22920bSAdrien Mazarguil 
292e22920bSAdrien Mazarguil #include "mlx5.h"
302e22920bSAdrien Mazarguil #include "mlx5_utils.h"
312e22920bSAdrien Mazarguil #include "mlx5_rxtx.h"
32f3db9489SYaacov Hazan #include "mlx5_autoconf.h"
332e22920bSAdrien Mazarguil #include "mlx5_defs.h"
346218063bSNélio Laranjeiro #include "mlx5_prm.h"
356218063bSNélio Laranjeiro 
36c0583d98SJerin Jacob static __rte_always_inline uint32_t
373cc08bc6SXueming Li rxq_cq_to_pkt_type(struct mlx5_rxq_data *rxq, volatile struct mlx5_cqe *cqe);
38ff1807a3SNélio Laranjeiro 
39c0583d98SJerin Jacob static __rte_always_inline int
4078142aacSNélio Laranjeiro mlx5_rx_poll_len(struct mlx5_rxq_data *rxq, volatile struct mlx5_cqe *cqe,
41c0583d98SJerin Jacob 		 uint16_t cqe_cnt, uint32_t *rss_hash);
42ff1807a3SNélio Laranjeiro 
43c0583d98SJerin Jacob static __rte_always_inline uint32_t
446ba07449SXueming Li rxq_cq_to_ol_flags(volatile struct mlx5_cqe *cqe);
45ff1807a3SNélio Laranjeiro 
463e1f82a1SYongseok Koh static __rte_always_inline void
473e1f82a1SYongseok Koh rxq_cq_to_mbuf(struct mlx5_rxq_data *rxq, struct rte_mbuf *pkt,
483e1f82a1SYongseok Koh 	       volatile struct mlx5_cqe *cqe, uint32_t rss_hash_res);
493e1f82a1SYongseok Koh 
50*7d6bf6b8SYongseok Koh static __rte_always_inline void
51*7d6bf6b8SYongseok Koh mprq_buf_replace(struct mlx5_rxq_data *rxq, uint16_t rq_idx);
52*7d6bf6b8SYongseok Koh 
53ea16068cSYongseok Koh uint32_t mlx5_ptype_table[] __rte_cache_aligned = {
54ea16068cSYongseok Koh 	[0xff] = RTE_PTYPE_ALL_MASK, /* Last entry for errored packet. */
55ea16068cSYongseok Koh };
56ea16068cSYongseok Koh 
575f8ba81cSXueming Li uint8_t mlx5_cksum_table[1 << 10] __rte_cache_aligned;
585f8ba81cSXueming Li uint8_t mlx5_swp_types_table[1 << 10] __rte_cache_aligned;
595f8ba81cSXueming Li 
60ea16068cSYongseok Koh /**
61ea16068cSYongseok Koh  * Build a table to translate Rx completion flags to packet type.
62ea16068cSYongseok Koh  *
63ea16068cSYongseok Koh  * @note: fix mlx5_dev_supported_ptypes_get() if any change here.
64ea16068cSYongseok Koh  */
65ea16068cSYongseok Koh void
66ea16068cSYongseok Koh mlx5_set_ptype_table(void)
67ea16068cSYongseok Koh {
68ea16068cSYongseok Koh 	unsigned int i;
69ea16068cSYongseok Koh 	uint32_t (*p)[RTE_DIM(mlx5_ptype_table)] = &mlx5_ptype_table;
70ea16068cSYongseok Koh 
719807f113SYongseok Koh 	/* Last entry must not be overwritten, reserved for errored packet. */
729807f113SYongseok Koh 	for (i = 0; i < RTE_DIM(mlx5_ptype_table) - 1; ++i)
73ea16068cSYongseok Koh 		(*p)[i] = RTE_PTYPE_UNKNOWN;
746cb559d6SYongseok Koh 	/*
756cb559d6SYongseok Koh 	 * The index to the array should have:
76ea16068cSYongseok Koh 	 * bit[1:0] = l3_hdr_type
77ea16068cSYongseok Koh 	 * bit[4:2] = l4_hdr_type
78ea16068cSYongseok Koh 	 * bit[5] = ip_frag
79ea16068cSYongseok Koh 	 * bit[6] = tunneled
80ea16068cSYongseok Koh 	 * bit[7] = outer_l3_type
8199c12dccSNélio Laranjeiro 	 */
823ca63b88SShahaf Shuler 	/* L2 */
833ca63b88SShahaf Shuler 	(*p)[0x00] = RTE_PTYPE_L2_ETHER;
84ea16068cSYongseok Koh 	/* L3 */
85ea16068cSYongseok Koh 	(*p)[0x01] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN |
86ea16068cSYongseok Koh 		     RTE_PTYPE_L4_NONFRAG;
87ea16068cSYongseok Koh 	(*p)[0x02] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN |
88ea16068cSYongseok Koh 		     RTE_PTYPE_L4_NONFRAG;
89ea16068cSYongseok Koh 	/* Fragmented */
90ea16068cSYongseok Koh 	(*p)[0x21] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN |
91ea16068cSYongseok Koh 		     RTE_PTYPE_L4_FRAG;
92ea16068cSYongseok Koh 	(*p)[0x22] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN |
93ea16068cSYongseok Koh 		     RTE_PTYPE_L4_FRAG;
94ea16068cSYongseok Koh 	/* TCP */
95ea16068cSYongseok Koh 	(*p)[0x05] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN |
96ea16068cSYongseok Koh 		     RTE_PTYPE_L4_TCP;
97ea16068cSYongseok Koh 	(*p)[0x06] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN |
98ea16068cSYongseok Koh 		     RTE_PTYPE_L4_TCP;
990915e287SBin Huang 	(*p)[0x0d] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN |
1000915e287SBin Huang 		     RTE_PTYPE_L4_TCP;
1010915e287SBin Huang 	(*p)[0x0e] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN |
1020915e287SBin Huang 		     RTE_PTYPE_L4_TCP;
1030915e287SBin Huang 	(*p)[0x11] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN |
1040915e287SBin Huang 		     RTE_PTYPE_L4_TCP;
1050915e287SBin Huang 	(*p)[0x12] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN |
1060915e287SBin Huang 		     RTE_PTYPE_L4_TCP;
107ea16068cSYongseok Koh 	/* UDP */
108ea16068cSYongseok Koh 	(*p)[0x09] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN |
109ea16068cSYongseok Koh 		     RTE_PTYPE_L4_UDP;
110ea16068cSYongseok Koh 	(*p)[0x0a] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN |
111ea16068cSYongseok Koh 		     RTE_PTYPE_L4_UDP;
112ea16068cSYongseok Koh 	/* Repeat with outer_l3_type being set. Just in case. */
113ea16068cSYongseok Koh 	(*p)[0x81] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN |
114ea16068cSYongseok Koh 		     RTE_PTYPE_L4_NONFRAG;
115ea16068cSYongseok Koh 	(*p)[0x82] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN |
116ea16068cSYongseok Koh 		     RTE_PTYPE_L4_NONFRAG;
117ea16068cSYongseok Koh 	(*p)[0xa1] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN |
118ea16068cSYongseok Koh 		     RTE_PTYPE_L4_FRAG;
119ea16068cSYongseok Koh 	(*p)[0xa2] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN |
120ea16068cSYongseok Koh 		     RTE_PTYPE_L4_FRAG;
121ea16068cSYongseok Koh 	(*p)[0x85] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN |
122ea16068cSYongseok Koh 		     RTE_PTYPE_L4_TCP;
123ea16068cSYongseok Koh 	(*p)[0x86] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN |
124ea16068cSYongseok Koh 		     RTE_PTYPE_L4_TCP;
1250915e287SBin Huang 	(*p)[0x8d] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN |
1260915e287SBin Huang 		     RTE_PTYPE_L4_TCP;
1270915e287SBin Huang 	(*p)[0x8e] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN |
1280915e287SBin Huang 		     RTE_PTYPE_L4_TCP;
1290915e287SBin Huang 	(*p)[0x91] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN |
1300915e287SBin Huang 		     RTE_PTYPE_L4_TCP;
1310915e287SBin Huang 	(*p)[0x92] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN |
1320915e287SBin Huang 		     RTE_PTYPE_L4_TCP;
133ea16068cSYongseok Koh 	(*p)[0x89] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN |
134ea16068cSYongseok Koh 		     RTE_PTYPE_L4_UDP;
135ea16068cSYongseok Koh 	(*p)[0x8a] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN |
136ea16068cSYongseok Koh 		     RTE_PTYPE_L4_UDP;
137ea16068cSYongseok Koh 	/* Tunneled - L3 */
1383cc08bc6SXueming Li 	(*p)[0x40] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN;
139ea16068cSYongseok Koh 	(*p)[0x41] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN |
140ea16068cSYongseok Koh 		     RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN |
141ea16068cSYongseok Koh 		     RTE_PTYPE_INNER_L4_NONFRAG;
142ea16068cSYongseok Koh 	(*p)[0x42] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN |
143ea16068cSYongseok Koh 		     RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN |
144ea16068cSYongseok Koh 		     RTE_PTYPE_INNER_L4_NONFRAG;
1453cc08bc6SXueming Li 	(*p)[0xc0] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN;
146ea16068cSYongseok Koh 	(*p)[0xc1] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN |
147ea16068cSYongseok Koh 		     RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN |
148ea16068cSYongseok Koh 		     RTE_PTYPE_INNER_L4_NONFRAG;
149ea16068cSYongseok Koh 	(*p)[0xc2] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN |
150ea16068cSYongseok Koh 		     RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN |
151ea16068cSYongseok Koh 		     RTE_PTYPE_INNER_L4_NONFRAG;
152ea16068cSYongseok Koh 	/* Tunneled - Fragmented */
153ea16068cSYongseok Koh 	(*p)[0x61] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN |
154ea16068cSYongseok Koh 		     RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN |
155ea16068cSYongseok Koh 		     RTE_PTYPE_INNER_L4_FRAG;
156ea16068cSYongseok Koh 	(*p)[0x62] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN |
157ea16068cSYongseok Koh 		     RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN |
158ea16068cSYongseok Koh 		     RTE_PTYPE_INNER_L4_FRAG;
159ea16068cSYongseok Koh 	(*p)[0xe1] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN |
160ea16068cSYongseok Koh 		     RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN |
161ea16068cSYongseok Koh 		     RTE_PTYPE_INNER_L4_FRAG;
162ea16068cSYongseok Koh 	(*p)[0xe2] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN |
163ea16068cSYongseok Koh 		     RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN |
164ea16068cSYongseok Koh 		     RTE_PTYPE_INNER_L4_FRAG;
165ea16068cSYongseok Koh 	/* Tunneled - TCP */
166ea16068cSYongseok Koh 	(*p)[0x45] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN |
167ea16068cSYongseok Koh 		     RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN |
1686c897093SYongseok Koh 		     RTE_PTYPE_INNER_L4_TCP;
169ea16068cSYongseok Koh 	(*p)[0x46] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN |
170ea16068cSYongseok Koh 		     RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN |
1716c897093SYongseok Koh 		     RTE_PTYPE_INNER_L4_TCP;
1720915e287SBin Huang 	(*p)[0x4d] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN |
1730915e287SBin Huang 		     RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN |
1740915e287SBin Huang 		     RTE_PTYPE_INNER_L4_TCP;
1750915e287SBin Huang 	(*p)[0x4e] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN |
1760915e287SBin Huang 		     RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN |
1770915e287SBin Huang 		     RTE_PTYPE_INNER_L4_TCP;
1780915e287SBin Huang 	(*p)[0x51] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN |
1790915e287SBin Huang 		     RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN |
1800915e287SBin Huang 		     RTE_PTYPE_INNER_L4_TCP;
1810915e287SBin Huang 	(*p)[0x52] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN |
1820915e287SBin Huang 		     RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN |
1830915e287SBin Huang 		     RTE_PTYPE_INNER_L4_TCP;
184ea16068cSYongseok Koh 	(*p)[0xc5] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN |
185ea16068cSYongseok Koh 		     RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN |
1866c897093SYongseok Koh 		     RTE_PTYPE_INNER_L4_TCP;
187ea16068cSYongseok Koh 	(*p)[0xc6] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN |
188ea16068cSYongseok Koh 		     RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN |
1896c897093SYongseok Koh 		     RTE_PTYPE_INNER_L4_TCP;
1900915e287SBin Huang 	(*p)[0xcd] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN |
1910915e287SBin Huang 		     RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN |
1920915e287SBin Huang 		     RTE_PTYPE_INNER_L4_TCP;
1930915e287SBin Huang 	(*p)[0xce] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN |
1940915e287SBin Huang 		     RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN |
1950915e287SBin Huang 		     RTE_PTYPE_INNER_L4_TCP;
1960915e287SBin Huang 	(*p)[0xd1] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN |
1970915e287SBin Huang 		     RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN |
1980915e287SBin Huang 		     RTE_PTYPE_INNER_L4_TCP;
1990915e287SBin Huang 	(*p)[0xd2] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN |
2000915e287SBin Huang 		     RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN |
2010915e287SBin Huang 		     RTE_PTYPE_INNER_L4_TCP;
202ea16068cSYongseok Koh 	/* Tunneled - UDP */
203ea16068cSYongseok Koh 	(*p)[0x49] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN |
204ea16068cSYongseok Koh 		     RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN |
2056c897093SYongseok Koh 		     RTE_PTYPE_INNER_L4_UDP;
206ea16068cSYongseok Koh 	(*p)[0x4a] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN |
207ea16068cSYongseok Koh 		     RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN |
2086c897093SYongseok Koh 		     RTE_PTYPE_INNER_L4_UDP;
209ea16068cSYongseok Koh 	(*p)[0xc9] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN |
210ea16068cSYongseok Koh 		     RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN |
2116c897093SYongseok Koh 		     RTE_PTYPE_INNER_L4_UDP;
212ea16068cSYongseok Koh 	(*p)[0xca] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN |
213ea16068cSYongseok Koh 		     RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN |
2146c897093SYongseok Koh 		     RTE_PTYPE_INNER_L4_UDP;
215ea16068cSYongseok Koh }
216fdcb0f53SNélio Laranjeiro 
2172e22920bSAdrien Mazarguil /**
2185f8ba81cSXueming Li  * Build a table to translate packet to checksum type of Verbs.
2195f8ba81cSXueming Li  */
2205f8ba81cSXueming Li void
2215f8ba81cSXueming Li mlx5_set_cksum_table(void)
2225f8ba81cSXueming Li {
2235f8ba81cSXueming Li 	unsigned int i;
2245f8ba81cSXueming Li 	uint8_t v;
2255f8ba81cSXueming Li 
2265f8ba81cSXueming Li 	/*
2275f8ba81cSXueming Li 	 * The index should have:
2285f8ba81cSXueming Li 	 * bit[0] = PKT_TX_TCP_SEG
2295f8ba81cSXueming Li 	 * bit[2:3] = PKT_TX_UDP_CKSUM, PKT_TX_TCP_CKSUM
2305f8ba81cSXueming Li 	 * bit[4] = PKT_TX_IP_CKSUM
2315f8ba81cSXueming Li 	 * bit[8] = PKT_TX_OUTER_IP_CKSUM
2325f8ba81cSXueming Li 	 * bit[9] = tunnel
2335f8ba81cSXueming Li 	 */
2345f8ba81cSXueming Li 	for (i = 0; i < RTE_DIM(mlx5_cksum_table); ++i) {
2355f8ba81cSXueming Li 		v = 0;
2365f8ba81cSXueming Li 		if (i & (1 << 9)) {
2375f8ba81cSXueming Li 			/* Tunneled packet. */
2385f8ba81cSXueming Li 			if (i & (1 << 8)) /* Outer IP. */
2395f8ba81cSXueming Li 				v |= MLX5_ETH_WQE_L3_CSUM;
2405f8ba81cSXueming Li 			if (i & (1 << 4)) /* Inner IP. */
2415f8ba81cSXueming Li 				v |= MLX5_ETH_WQE_L3_INNER_CSUM;
2425f8ba81cSXueming Li 			if (i & (3 << 2 | 1 << 0)) /* L4 or TSO. */
2435f8ba81cSXueming Li 				v |= MLX5_ETH_WQE_L4_INNER_CSUM;
2445f8ba81cSXueming Li 		} else {
2455f8ba81cSXueming Li 			/* No tunnel. */
2465f8ba81cSXueming Li 			if (i & (1 << 4)) /* IP. */
2475f8ba81cSXueming Li 				v |= MLX5_ETH_WQE_L3_CSUM;
2485f8ba81cSXueming Li 			if (i & (3 << 2 | 1 << 0)) /* L4 or TSO. */
2495f8ba81cSXueming Li 				v |= MLX5_ETH_WQE_L4_CSUM;
2505f8ba81cSXueming Li 		}
2515f8ba81cSXueming Li 		mlx5_cksum_table[i] = v;
2525f8ba81cSXueming Li 	}
2535f8ba81cSXueming Li }
2545f8ba81cSXueming Li 
2555f8ba81cSXueming Li /**
2565f8ba81cSXueming Li  * Build a table to translate packet type of mbuf to SWP type of Verbs.
2575f8ba81cSXueming Li  */
2585f8ba81cSXueming Li void
2595f8ba81cSXueming Li mlx5_set_swp_types_table(void)
2605f8ba81cSXueming Li {
2615f8ba81cSXueming Li 	unsigned int i;
2625f8ba81cSXueming Li 	uint8_t v;
2635f8ba81cSXueming Li 
2645f8ba81cSXueming Li 	/*
2655f8ba81cSXueming Li 	 * The index should have:
2665f8ba81cSXueming Li 	 * bit[0:1] = PKT_TX_L4_MASK
2675f8ba81cSXueming Li 	 * bit[4] = PKT_TX_IPV6
2685f8ba81cSXueming Li 	 * bit[8] = PKT_TX_OUTER_IPV6
2695f8ba81cSXueming Li 	 * bit[9] = PKT_TX_OUTER_UDP
2705f8ba81cSXueming Li 	 */
2715f8ba81cSXueming Li 	for (i = 0; i < RTE_DIM(mlx5_swp_types_table); ++i) {
2725f8ba81cSXueming Li 		v = 0;
2735f8ba81cSXueming Li 		if (i & (1 << 8))
2745f8ba81cSXueming Li 			v |= MLX5_ETH_WQE_L3_OUTER_IPV6;
2755f8ba81cSXueming Li 		if (i & (1 << 9))
2765f8ba81cSXueming Li 			v |= MLX5_ETH_WQE_L4_OUTER_UDP;
2775f8ba81cSXueming Li 		if (i & (1 << 4))
2785f8ba81cSXueming Li 			v |= MLX5_ETH_WQE_L3_INNER_IPV6;
2795f8ba81cSXueming Li 		if ((i & 3) == (PKT_TX_UDP_CKSUM >> 52))
2805f8ba81cSXueming Li 			v |= MLX5_ETH_WQE_L4_INNER_UDP;
2815f8ba81cSXueming Li 		mlx5_swp_types_table[i] = v;
2825f8ba81cSXueming Li 	}
2835f8ba81cSXueming Li }
2845f8ba81cSXueming Li 
2855f8ba81cSXueming Li /**
2866ce84bd8SYongseok Koh  * Return the size of tailroom of WQ.
2876ce84bd8SYongseok Koh  *
2886ce84bd8SYongseok Koh  * @param txq
2896ce84bd8SYongseok Koh  *   Pointer to TX queue structure.
2906ce84bd8SYongseok Koh  * @param addr
2916ce84bd8SYongseok Koh  *   Pointer to tail of WQ.
2926ce84bd8SYongseok Koh  *
2936ce84bd8SYongseok Koh  * @return
2946ce84bd8SYongseok Koh  *   Size of tailroom.
2956ce84bd8SYongseok Koh  */
2966ce84bd8SYongseok Koh static inline size_t
297991b04f6SNélio Laranjeiro tx_mlx5_wq_tailroom(struct mlx5_txq_data *txq, void *addr)
2986ce84bd8SYongseok Koh {
2996ce84bd8SYongseok Koh 	size_t tailroom;
3006ce84bd8SYongseok Koh 	tailroom = (uintptr_t)(txq->wqes) +
3016ce84bd8SYongseok Koh 		   (1 << txq->wqe_n) * MLX5_WQE_SIZE -
3026ce84bd8SYongseok Koh 		   (uintptr_t)addr;
3036ce84bd8SYongseok Koh 	return tailroom;
3046ce84bd8SYongseok Koh }
3056ce84bd8SYongseok Koh 
3066ce84bd8SYongseok Koh /**
3076ce84bd8SYongseok Koh  * Copy data to tailroom of circular queue.
3086ce84bd8SYongseok Koh  *
3096ce84bd8SYongseok Koh  * @param dst
3106ce84bd8SYongseok Koh  *   Pointer to destination.
3116ce84bd8SYongseok Koh  * @param src
3126ce84bd8SYongseok Koh  *   Pointer to source.
3136ce84bd8SYongseok Koh  * @param n
3146ce84bd8SYongseok Koh  *   Number of bytes to copy.
3156ce84bd8SYongseok Koh  * @param base
3166ce84bd8SYongseok Koh  *   Pointer to head of queue.
3176ce84bd8SYongseok Koh  * @param tailroom
3186ce84bd8SYongseok Koh  *   Size of tailroom from dst.
3196ce84bd8SYongseok Koh  *
3206ce84bd8SYongseok Koh  * @return
3216ce84bd8SYongseok Koh  *   Pointer after copied data.
3226ce84bd8SYongseok Koh  */
3236ce84bd8SYongseok Koh static inline void *
3246ce84bd8SYongseok Koh mlx5_copy_to_wq(void *dst, const void *src, size_t n,
3256ce84bd8SYongseok Koh 		void *base, size_t tailroom)
3266ce84bd8SYongseok Koh {
3276ce84bd8SYongseok Koh 	void *ret;
3286ce84bd8SYongseok Koh 
3296ce84bd8SYongseok Koh 	if (n > tailroom) {
3306ce84bd8SYongseok Koh 		rte_memcpy(dst, src, tailroom);
3316ce84bd8SYongseok Koh 		rte_memcpy(base, (void *)((uintptr_t)src + tailroom),
3326ce84bd8SYongseok Koh 			   n - tailroom);
3336ce84bd8SYongseok Koh 		ret = (uint8_t *)base + n - tailroom;
3346ce84bd8SYongseok Koh 	} else {
3356ce84bd8SYongseok Koh 		rte_memcpy(dst, src, n);
3366ce84bd8SYongseok Koh 		ret = (n == tailroom) ? base : (uint8_t *)dst + n;
3376ce84bd8SYongseok Koh 	}
3386ce84bd8SYongseok Koh 	return ret;
3396ce84bd8SYongseok Koh }
3406ce84bd8SYongseok Koh 
3416ce84bd8SYongseok Koh /**
342593f472cSXueming Li  * Inline TSO headers into WQE.
343593f472cSXueming Li  *
344593f472cSXueming Li  * @return
345593f472cSXueming Li  *   0 on success, negative errno value on failure.
346593f472cSXueming Li  */
347593f472cSXueming Li static int
348593f472cSXueming Li inline_tso(struct mlx5_txq_data *txq, struct rte_mbuf *buf,
349593f472cSXueming Li 	   uint32_t *length,
350593f472cSXueming Li 	   uintptr_t *addr,
351593f472cSXueming Li 	   uint16_t *pkt_inline_sz,
352593f472cSXueming Li 	   uint8_t **raw,
353593f472cSXueming Li 	   uint16_t *max_wqe,
354593f472cSXueming Li 	   uint16_t *tso_segsz,
355593f472cSXueming Li 	   uint16_t *tso_header_sz)
356593f472cSXueming Li {
357593f472cSXueming Li 	uintptr_t end = (uintptr_t)(((uintptr_t)txq->wqes) +
358593f472cSXueming Li 				    (1 << txq->wqe_n) * MLX5_WQE_SIZE);
359593f472cSXueming Li 	unsigned int copy_b;
360593f472cSXueming Li 	uint8_t vlan_sz = (buf->ol_flags & PKT_TX_VLAN_PKT) ? 4 : 0;
3615f8ba81cSXueming Li 	const uint8_t tunneled = txq->tunnel_en && (buf->ol_flags &
3625f8ba81cSXueming Li 				 PKT_TX_TUNNEL_MASK);
363593f472cSXueming Li 	uint16_t n_wqe;
364593f472cSXueming Li 
365593f472cSXueming Li 	*tso_segsz = buf->tso_segsz;
366593f472cSXueming Li 	*tso_header_sz = buf->l2_len + vlan_sz + buf->l3_len + buf->l4_len;
367593f472cSXueming Li 	if (unlikely(*tso_segsz == 0 || *tso_header_sz == 0)) {
368593f472cSXueming Li 		txq->stats.oerrors++;
369593f472cSXueming Li 		return -EINVAL;
370593f472cSXueming Li 	}
3715f8ba81cSXueming Li 	if (tunneled)
372593f472cSXueming Li 		*tso_header_sz += buf->outer_l2_len + buf->outer_l3_len;
3735f8ba81cSXueming Li 	/* First seg must contain all TSO headers. */
3745f8ba81cSXueming Li 	if (unlikely(*tso_header_sz > MLX5_MAX_TSO_HEADER) ||
3755f8ba81cSXueming Li 		     *tso_header_sz > DATA_LEN(buf)) {
376593f472cSXueming Li 		txq->stats.oerrors++;
377593f472cSXueming Li 		return -EINVAL;
378593f472cSXueming Li 	}
379593f472cSXueming Li 	copy_b = *tso_header_sz - *pkt_inline_sz;
380593f472cSXueming Li 	if (!copy_b || ((end - (uintptr_t)*raw) < copy_b))
381593f472cSXueming Li 		return -EAGAIN;
382593f472cSXueming Li 	n_wqe = (MLX5_WQE_DS(copy_b) - 1 + 3) / 4;
383593f472cSXueming Li 	if (unlikely(*max_wqe < n_wqe))
384593f472cSXueming Li 		return -EINVAL;
385593f472cSXueming Li 	*max_wqe -= n_wqe;
386593f472cSXueming Li 	rte_memcpy((void *)*raw, (void *)*addr, copy_b);
387593f472cSXueming Li 	*length -= copy_b;
388593f472cSXueming Li 	*addr += copy_b;
389593f472cSXueming Li 	copy_b = MLX5_WQE_DS(copy_b) * MLX5_WQE_DWORD_SIZE;
390593f472cSXueming Li 	*pkt_inline_sz += copy_b;
391593f472cSXueming Li 	*raw += copy_b;
392593f472cSXueming Li 	return 0;
393593f472cSXueming Li }
394593f472cSXueming Li 
395593f472cSXueming Li /**
3968788fec1SOlivier Matz  * DPDK callback to check the status of a tx descriptor.
3978788fec1SOlivier Matz  *
3988788fec1SOlivier Matz  * @param tx_queue
3998788fec1SOlivier Matz  *   The tx queue.
4008788fec1SOlivier Matz  * @param[in] offset
4018788fec1SOlivier Matz  *   The index of the descriptor in the ring.
4028788fec1SOlivier Matz  *
4038788fec1SOlivier Matz  * @return
4048788fec1SOlivier Matz  *   The status of the tx descriptor.
4058788fec1SOlivier Matz  */
4068788fec1SOlivier Matz int
4078788fec1SOlivier Matz mlx5_tx_descriptor_status(void *tx_queue, uint16_t offset)
4088788fec1SOlivier Matz {
409991b04f6SNélio Laranjeiro 	struct mlx5_txq_data *txq = tx_queue;
4108c819a69SYongseok Koh 	uint16_t used;
4118788fec1SOlivier Matz 
4126cb559d6SYongseok Koh 	mlx5_tx_complete(txq);
4138c819a69SYongseok Koh 	used = txq->elts_head - txq->elts_tail;
4148788fec1SOlivier Matz 	if (offset < used)
4158788fec1SOlivier Matz 		return RTE_ETH_TX_DESC_FULL;
4168788fec1SOlivier Matz 	return RTE_ETH_TX_DESC_DONE;
4178788fec1SOlivier Matz }
4188788fec1SOlivier Matz 
4198788fec1SOlivier Matz /**
4208788fec1SOlivier Matz  * DPDK callback to check the status of a rx descriptor.
4218788fec1SOlivier Matz  *
4228788fec1SOlivier Matz  * @param rx_queue
4238788fec1SOlivier Matz  *   The rx queue.
4248788fec1SOlivier Matz  * @param[in] offset
4258788fec1SOlivier Matz  *   The index of the descriptor in the ring.
4268788fec1SOlivier Matz  *
4278788fec1SOlivier Matz  * @return
4288788fec1SOlivier Matz  *   The status of the tx descriptor.
4298788fec1SOlivier Matz  */
4308788fec1SOlivier Matz int
4318788fec1SOlivier Matz mlx5_rx_descriptor_status(void *rx_queue, uint16_t offset)
4328788fec1SOlivier Matz {
43378142aacSNélio Laranjeiro 	struct mlx5_rxq_data *rxq = rx_queue;
4348788fec1SOlivier Matz 	struct rxq_zip *zip = &rxq->zip;
4358788fec1SOlivier Matz 	volatile struct mlx5_cqe *cqe;
4368788fec1SOlivier Matz 	const unsigned int cqe_n = (1 << rxq->cqe_n);
4378788fec1SOlivier Matz 	const unsigned int cqe_cnt = cqe_n - 1;
4388788fec1SOlivier Matz 	unsigned int cq_ci;
4398788fec1SOlivier Matz 	unsigned int used;
4408788fec1SOlivier Matz 
4418788fec1SOlivier Matz 	/* if we are processing a compressed cqe */
4428788fec1SOlivier Matz 	if (zip->ai) {
4438788fec1SOlivier Matz 		used = zip->cqe_cnt - zip->ca;
4448788fec1SOlivier Matz 		cq_ci = zip->cq_ci;
4458788fec1SOlivier Matz 	} else {
4468788fec1SOlivier Matz 		used = 0;
4478788fec1SOlivier Matz 		cq_ci = rxq->cq_ci;
4488788fec1SOlivier Matz 	}
4498788fec1SOlivier Matz 	cqe = &(*rxq->cqes)[cq_ci & cqe_cnt];
4508788fec1SOlivier Matz 	while (check_cqe(cqe, cqe_n, cq_ci) == 0) {
4518788fec1SOlivier Matz 		int8_t op_own;
4528788fec1SOlivier Matz 		unsigned int n;
4538788fec1SOlivier Matz 
4548788fec1SOlivier Matz 		op_own = cqe->op_own;
4558788fec1SOlivier Matz 		if (MLX5_CQE_FORMAT(op_own) == MLX5_COMPRESSED)
4566b30a6a8SShachar Beiser 			n = rte_be_to_cpu_32(cqe->byte_cnt);
4578788fec1SOlivier Matz 		else
4588788fec1SOlivier Matz 			n = 1;
4598788fec1SOlivier Matz 		cq_ci += n;
4608788fec1SOlivier Matz 		used += n;
4618788fec1SOlivier Matz 		cqe = &(*rxq->cqes)[cq_ci & cqe_cnt];
4628788fec1SOlivier Matz 	}
4638788fec1SOlivier Matz 	used = RTE_MIN(used, (1U << rxq->elts_n) - 1);
4648788fec1SOlivier Matz 	if (offset < used)
4658788fec1SOlivier Matz 		return RTE_ETH_RX_DESC_DONE;
4668788fec1SOlivier Matz 	return RTE_ETH_RX_DESC_AVAIL;
4678788fec1SOlivier Matz }
4688788fec1SOlivier Matz 
4698788fec1SOlivier Matz /**
4702e22920bSAdrien Mazarguil  * DPDK callback for TX.
4712e22920bSAdrien Mazarguil  *
4722e22920bSAdrien Mazarguil  * @param dpdk_txq
4732e22920bSAdrien Mazarguil  *   Generic pointer to TX queue structure.
4742e22920bSAdrien Mazarguil  * @param[in] pkts
4752e22920bSAdrien Mazarguil  *   Packets to transmit.
4762e22920bSAdrien Mazarguil  * @param pkts_n
4772e22920bSAdrien Mazarguil  *   Number of packets in array.
4782e22920bSAdrien Mazarguil  *
4792e22920bSAdrien Mazarguil  * @return
4802e22920bSAdrien Mazarguil  *   Number of packets successfully transmitted (<= pkts_n).
4812e22920bSAdrien Mazarguil  */
4822e22920bSAdrien Mazarguil uint16_t
4832e22920bSAdrien Mazarguil mlx5_tx_burst(void *dpdk_txq, struct rte_mbuf **pkts, uint16_t pkts_n)
4842e22920bSAdrien Mazarguil {
485991b04f6SNélio Laranjeiro 	struct mlx5_txq_data *txq = (struct mlx5_txq_data *)dpdk_txq;
4861d88ba17SNélio Laranjeiro 	uint16_t elts_head = txq->elts_head;
4878c819a69SYongseok Koh 	const uint16_t elts_n = 1 << txq->elts_n;
4888c819a69SYongseok Koh 	const uint16_t elts_m = elts_n - 1;
489c3d62cc9SAdrien Mazarguil 	unsigned int i = 0;
490a5bf6af9SAdrien Mazarguil 	unsigned int j = 0;
4913f13f8c2SShahaf Shuler 	unsigned int k = 0;
4928c819a69SYongseok Koh 	uint16_t max_elts;
493f04f1d51SNélio Laranjeiro 	uint16_t max_wqe;
494c305090bSAdrien Mazarguil 	unsigned int comp;
495ac180a21SYongseok Koh 	volatile struct mlx5_wqe_ctrl *last_wqe = NULL;
4966579c27cSNélio Laranjeiro 	unsigned int segs_n = 0;
49727a6b2d6SNélio Laranjeiro 	const unsigned int max_inline = txq->max_inline;
4982e22920bSAdrien Mazarguil 
4991d88ba17SNélio Laranjeiro 	if (unlikely(!pkts_n))
5001d88ba17SNélio Laranjeiro 		return 0;
5015e1d11ecSNelio Laranjeiro 	/* Prefetch first packet cacheline. */
502c3d62cc9SAdrien Mazarguil 	rte_prefetch0(*pkts);
5031d88ba17SNélio Laranjeiro 	/* Start processing. */
5046cb559d6SYongseok Koh 	mlx5_tx_complete(txq);
5058c819a69SYongseok Koh 	max_elts = (elts_n - (elts_head - txq->elts_tail));
5062eefbec5SYongseok Koh 	/* A CQE slot must always be available. */
5072eefbec5SYongseok Koh 	assert((1u << txq->cqe_n) - (txq->cq_pi - txq->cq_ci));
508f04f1d51SNélio Laranjeiro 	max_wqe = (1u << txq->wqe_n) - (txq->wqe_ci - txq->wqe_pi);
509f04f1d51SNélio Laranjeiro 	if (unlikely(!max_wqe))
510f04f1d51SNélio Laranjeiro 		return 0;
511c3d62cc9SAdrien Mazarguil 	do {
5125f8ba81cSXueming Li 		struct rte_mbuf *buf = *pkts; /* First_seg. */
5133bbae1ebSNélio Laranjeiro 		uint8_t *raw;
5143bbae1ebSNélio Laranjeiro 		volatile struct mlx5_wqe_v *wqe = NULL;
5159a7fa9f7SNélio Laranjeiro 		volatile rte_v128u32_t *dseg = NULL;
516573f54afSNélio Laranjeiro 		uint32_t length;
5178688b2f8SNélio Laranjeiro 		unsigned int ds = 0;
518ac180a21SYongseok Koh 		unsigned int sg = 0; /* counter of additional segs attached. */
5196579c27cSNélio Laranjeiro 		uintptr_t addr;
5200d637a34SNélio Laranjeiro 		uint16_t pkt_inline_sz = MLX5_WQE_DWORD_SIZE + 2;
5213f13f8c2SShahaf Shuler 		uint16_t tso_header_sz = 0;
522eef822ddSNélio Laranjeiro 		uint16_t ehdr;
5234aa15eb1SNélio Laranjeiro 		uint8_t cs_flags;
5245f8ba81cSXueming Li 		uint8_t tso = txq->tso_en && (buf->ol_flags & PKT_TX_TCP_SEG);
5255f8ba81cSXueming Li 		uint8_t is_vlan = !!(buf->ol_flags & PKT_TX_VLAN_PKT);
5265f8ba81cSXueming Li 		uint32_t swp_offsets = 0;
5275f8ba81cSXueming Li 		uint8_t swp_types = 0;
52883daf156SShahaf Shuler 		uint16_t tso_segsz = 0;
5296579c27cSNélio Laranjeiro #ifdef MLX5_PMD_SOFT_COUNTERS
5306579c27cSNélio Laranjeiro 		uint32_t total_length = 0;
5316579c27cSNélio Laranjeiro #endif
532593f472cSXueming Li 		int ret;
5332e22920bSAdrien Mazarguil 
5346579c27cSNélio Laranjeiro 		segs_n = buf->nb_segs;
535c3d62cc9SAdrien Mazarguil 		/*
536c3d62cc9SAdrien Mazarguil 		 * Make sure there is enough room to store this packet and
537c3d62cc9SAdrien Mazarguil 		 * that one ring entry remains unused.
538c3d62cc9SAdrien Mazarguil 		 */
539a5bf6af9SAdrien Mazarguil 		assert(segs_n);
5408c819a69SYongseok Koh 		if (max_elts < segs_n)
541c3d62cc9SAdrien Mazarguil 			break;
5428c819a69SYongseok Koh 		max_elts -= segs_n;
543f895536bSYongseok Koh 		sg = --segs_n;
544f04f1d51SNélio Laranjeiro 		if (unlikely(--max_wqe == 0))
545f04f1d51SNélio Laranjeiro 			break;
5469a7fa9f7SNélio Laranjeiro 		wqe = (volatile struct mlx5_wqe_v *)
547fdcb0f53SNélio Laranjeiro 			tx_mlx5_wqe(txq, txq->wqe_ci);
548fdcb0f53SNélio Laranjeiro 		rte_prefetch0(tx_mlx5_wqe(txq, txq->wqe_ci + 1));
5493730e6c6SYongseok Koh 		if (pkts_n - i > 1)
5503730e6c6SYongseok Koh 			rte_prefetch0(*(pkts + 1));
5516579c27cSNélio Laranjeiro 		addr = rte_pktmbuf_mtod(buf, uintptr_t);
5522e22920bSAdrien Mazarguil 		length = DATA_LEN(buf);
553eef822ddSNélio Laranjeiro 		ehdr = (((uint8_t *)addr)[1] << 8) |
554eef822ddSNélio Laranjeiro 		       ((uint8_t *)addr)[0];
5556579c27cSNélio Laranjeiro #ifdef MLX5_PMD_SOFT_COUNTERS
5566579c27cSNélio Laranjeiro 		total_length = length;
5576579c27cSNélio Laranjeiro #endif
55824c14430SShahaf Shuler 		if (length < (MLX5_WQE_DWORD_SIZE + 2)) {
55924c14430SShahaf Shuler 			txq->stats.oerrors++;
560959be52eSNélio Laranjeiro 			break;
56124c14430SShahaf Shuler 		}
5622e22920bSAdrien Mazarguil 		/* Update element. */
5638c819a69SYongseok Koh 		(*txq->elts)[elts_head & elts_m] = buf;
5645e1d11ecSNelio Laranjeiro 		/* Prefetch next buffer data. */
5653730e6c6SYongseok Koh 		if (pkts_n - i > 1)
5663730e6c6SYongseok Koh 			rte_prefetch0(
5673730e6c6SYongseok Koh 			    rte_pktmbuf_mtod(*(pkts + 1), volatile void *));
5685f8ba81cSXueming Li 		cs_flags = txq_ol_cksum_to_cs(buf);
5695f8ba81cSXueming Li 		txq_mbuf_to_swp(txq, buf, tso, is_vlan,
5705f8ba81cSXueming Li 				(uint8_t *)&swp_offsets, &swp_types);
571b8fe952eSNélio Laranjeiro 		raw = ((uint8_t *)(uintptr_t)wqe) + 2 * MLX5_WQE_DWORD_SIZE;
5726579c27cSNélio Laranjeiro 		/* Replace the Ethernet type by the VLAN if necessary. */
5735f8ba81cSXueming Li 		if (is_vlan) {
5746b30a6a8SShachar Beiser 			uint32_t vlan = rte_cpu_to_be_32(0x81000000 |
5756b30a6a8SShachar Beiser 							 buf->vlan_tci);
5760d637a34SNélio Laranjeiro 			unsigned int len = 2 * ETHER_ADDR_LEN - 2;
5776579c27cSNélio Laranjeiro 
5780d637a34SNélio Laranjeiro 			addr += 2;
5790d637a34SNélio Laranjeiro 			length -= 2;
5800d637a34SNélio Laranjeiro 			/* Copy Destination and source mac address. */
5810d637a34SNélio Laranjeiro 			memcpy((uint8_t *)raw, ((uint8_t *)addr), len);
5820d637a34SNélio Laranjeiro 			/* Copy VLAN. */
5830d637a34SNélio Laranjeiro 			memcpy((uint8_t *)raw + len, &vlan, sizeof(vlan));
5840d637a34SNélio Laranjeiro 			/* Copy missing two bytes to end the DSeg. */
5850d637a34SNélio Laranjeiro 			memcpy((uint8_t *)raw + len + sizeof(vlan),
5860d637a34SNélio Laranjeiro 			       ((uint8_t *)addr) + len, 2);
5870d637a34SNélio Laranjeiro 			addr += len + 2;
5880d637a34SNélio Laranjeiro 			length -= (len + 2);
5890d637a34SNélio Laranjeiro 		} else {
5900d637a34SNélio Laranjeiro 			memcpy((uint8_t *)raw, ((uint8_t *)addr) + 2,
5910d637a34SNélio Laranjeiro 			       MLX5_WQE_DWORD_SIZE);
5920d637a34SNélio Laranjeiro 			length -= pkt_inline_sz;
5930d637a34SNélio Laranjeiro 			addr += pkt_inline_sz;
5946579c27cSNélio Laranjeiro 		}
595d8292497SYongseok Koh 		raw += MLX5_WQE_DWORD_SIZE;
5963f13f8c2SShahaf Shuler 		if (tso) {
5975f8ba81cSXueming Li 			ret = inline_tso(txq, buf, &length,
598593f472cSXueming Li 					 &addr, &pkt_inline_sz,
599593f472cSXueming Li 					 &raw, &max_wqe,
600593f472cSXueming Li 					 &tso_segsz, &tso_header_sz);
601593f472cSXueming Li 			if (ret == -EINVAL) {
60296fc8d65SShahaf Shuler 				break;
603593f472cSXueming Li 			} else if (ret == -EAGAIN) {
6043f13f8c2SShahaf Shuler 				/* NOP WQE. */
6053f13f8c2SShahaf Shuler 				wqe->ctrl = (rte_v128u32_t){
60636aa55eaSYongseok Koh 					rte_cpu_to_be_32(txq->wqe_ci << 8),
60736aa55eaSYongseok Koh 					rte_cpu_to_be_32(txq->qp_num_8s | 1),
6083f13f8c2SShahaf Shuler 					0,
6093f13f8c2SShahaf Shuler 					0,
6103f13f8c2SShahaf Shuler 				};
6113f13f8c2SShahaf Shuler 				ds = 1;
612cb98affeSThierry Herbelot #ifdef MLX5_PMD_SOFT_COUNTERS
6133f13f8c2SShahaf Shuler 				total_length = 0;
614cb98affeSThierry Herbelot #endif
6153f13f8c2SShahaf Shuler 				k++;
6163f13f8c2SShahaf Shuler 				goto next_wqe;
6173f13f8c2SShahaf Shuler 			}
6183f13f8c2SShahaf Shuler 		}
6196579c27cSNélio Laranjeiro 		/* Inline if enough room. */
62027a6b2d6SNélio Laranjeiro 		if (max_inline || tso) {
621f895536bSYongseok Koh 			uint32_t inl = 0;
622fdcb0f53SNélio Laranjeiro 			uintptr_t end = (uintptr_t)
623fdcb0f53SNélio Laranjeiro 				(((uintptr_t)txq->wqes) +
624fdcb0f53SNélio Laranjeiro 				 (1 << txq->wqe_n) * MLX5_WQE_SIZE);
625ab76eab3SYongseok Koh 			unsigned int inline_room = max_inline *
6268fcd6c2cSNélio Laranjeiro 						   RTE_CACHE_LINE_SIZE -
627d8292497SYongseok Koh 						   (pkt_inline_sz - 2) -
628d8292497SYongseok Koh 						   !!tso * sizeof(inl);
629f895536bSYongseok Koh 			uintptr_t addr_end;
630f895536bSYongseok Koh 			unsigned int copy_b;
6316579c27cSNélio Laranjeiro 
632f895536bSYongseok Koh pkt_inline:
633f895536bSYongseok Koh 			addr_end = RTE_ALIGN_FLOOR(addr + inline_room,
634f895536bSYongseok Koh 						   RTE_CACHE_LINE_SIZE);
635f895536bSYongseok Koh 			copy_b = (addr_end > addr) ?
636f895536bSYongseok Koh 				 RTE_MIN((addr_end - addr), length) : 0;
6378fcd6c2cSNélio Laranjeiro 			if (copy_b && ((end - (uintptr_t)raw) > copy_b)) {
638f04f1d51SNélio Laranjeiro 				/*
639f04f1d51SNélio Laranjeiro 				 * One Dseg remains in the current WQE.  To
640f04f1d51SNélio Laranjeiro 				 * keep the computation positive, it is
641f04f1d51SNélio Laranjeiro 				 * removed after the bytes to Dseg conversion.
642f04f1d51SNélio Laranjeiro 				 */
6438fcd6c2cSNélio Laranjeiro 				uint16_t n = (MLX5_WQE_DS(copy_b) - 1 + 3) / 4;
6448fcd6c2cSNélio Laranjeiro 
645f04f1d51SNélio Laranjeiro 				if (unlikely(max_wqe < n))
646f04f1d51SNélio Laranjeiro 					break;
647f04f1d51SNélio Laranjeiro 				max_wqe -= n;
648f895536bSYongseok Koh 				if (tso && !inl) {
6496963ae8bSYongseok Koh 					inl = rte_cpu_to_be_32(copy_b |
6506b30a6a8SShachar Beiser 							       MLX5_INLINE_SEG);
6513f13f8c2SShahaf Shuler 					rte_memcpy((void *)raw,
6523f13f8c2SShahaf Shuler 						   (void *)&inl, sizeof(inl));
6533f13f8c2SShahaf Shuler 					raw += sizeof(inl);
6543f13f8c2SShahaf Shuler 					pkt_inline_sz += sizeof(inl);
6553f13f8c2SShahaf Shuler 				}
6566579c27cSNélio Laranjeiro 				rte_memcpy((void *)raw, (void *)addr, copy_b);
6576579c27cSNélio Laranjeiro 				addr += copy_b;
6586579c27cSNélio Laranjeiro 				length -= copy_b;
6596579c27cSNélio Laranjeiro 				pkt_inline_sz += copy_b;
6606579c27cSNélio Laranjeiro 			}
6616579c27cSNélio Laranjeiro 			/*
662786b5c2dSShahaf Shuler 			 * 2 DWORDs consumed by the WQE header + ETH segment +
6636579c27cSNélio Laranjeiro 			 * the size of the inline part of the packet.
6646579c27cSNélio Laranjeiro 			 */
6656579c27cSNélio Laranjeiro 			ds = 2 + MLX5_WQE_DS(pkt_inline_sz - 2);
6666579c27cSNélio Laranjeiro 			if (length > 0) {
667f04f1d51SNélio Laranjeiro 				if (ds % (MLX5_WQE_SIZE /
668f04f1d51SNélio Laranjeiro 					  MLX5_WQE_DWORD_SIZE) == 0) {
669f04f1d51SNélio Laranjeiro 					if (unlikely(--max_wqe == 0))
670f04f1d51SNélio Laranjeiro 						break;
671f04f1d51SNélio Laranjeiro 					dseg = (volatile rte_v128u32_t *)
672f04f1d51SNélio Laranjeiro 					       tx_mlx5_wqe(txq, txq->wqe_ci +
673f04f1d51SNélio Laranjeiro 							   ds / 4);
674f04f1d51SNélio Laranjeiro 				} else {
6759a7fa9f7SNélio Laranjeiro 					dseg = (volatile rte_v128u32_t *)
6766579c27cSNélio Laranjeiro 						((uintptr_t)wqe +
6776579c27cSNélio Laranjeiro 						 (ds * MLX5_WQE_DWORD_SIZE));
678f04f1d51SNélio Laranjeiro 				}
6796579c27cSNélio Laranjeiro 				goto use_dseg;
6806579c27cSNélio Laranjeiro 			} else if (!segs_n) {
6816579c27cSNélio Laranjeiro 				goto next_pkt;
6826579c27cSNélio Laranjeiro 			} else {
683f895536bSYongseok Koh 				raw += copy_b;
684f895536bSYongseok Koh 				inline_room -= copy_b;
685f895536bSYongseok Koh 				--segs_n;
686f895536bSYongseok Koh 				buf = buf->next;
687f895536bSYongseok Koh 				assert(buf);
688f895536bSYongseok Koh 				addr = rte_pktmbuf_mtod(buf, uintptr_t);
689f895536bSYongseok Koh 				length = DATA_LEN(buf);
690f895536bSYongseok Koh #ifdef MLX5_PMD_SOFT_COUNTERS
691f895536bSYongseok Koh 				total_length += length;
692f895536bSYongseok Koh #endif
693f895536bSYongseok Koh 				(*txq->elts)[++elts_head & elts_m] = buf;
694f895536bSYongseok Koh 				goto pkt_inline;
6956579c27cSNélio Laranjeiro 			}
6966579c27cSNélio Laranjeiro 		} else {
6976579c27cSNélio Laranjeiro 			/*
6986579c27cSNélio Laranjeiro 			 * No inline has been done in the packet, only the
6996579c27cSNélio Laranjeiro 			 * Ethernet Header as been stored.
7006579c27cSNélio Laranjeiro 			 */
7019a7fa9f7SNélio Laranjeiro 			dseg = (volatile rte_v128u32_t *)
7026579c27cSNélio Laranjeiro 				((uintptr_t)wqe + (3 * MLX5_WQE_DWORD_SIZE));
7036579c27cSNélio Laranjeiro 			ds = 3;
7046579c27cSNélio Laranjeiro use_dseg:
7056579c27cSNélio Laranjeiro 			/* Add the remaining packet as a simple ds. */
706ebbb81ebSNélio Laranjeiro 			addr = rte_cpu_to_be_64(addr);
7079a7fa9f7SNélio Laranjeiro 			*dseg = (rte_v128u32_t){
7086b30a6a8SShachar Beiser 				rte_cpu_to_be_32(length),
7096cb559d6SYongseok Koh 				mlx5_tx_mb2mr(txq, buf),
710ebbb81ebSNélio Laranjeiro 				addr,
711ebbb81ebSNélio Laranjeiro 				addr >> 32,
7126579c27cSNélio Laranjeiro 			};
7136579c27cSNélio Laranjeiro 			++ds;
7146579c27cSNélio Laranjeiro 			if (!segs_n)
7156579c27cSNélio Laranjeiro 				goto next_pkt;
7166579c27cSNélio Laranjeiro 		}
7176579c27cSNélio Laranjeiro next_seg:
7186579c27cSNélio Laranjeiro 		assert(buf);
7196579c27cSNélio Laranjeiro 		assert(ds);
7206579c27cSNélio Laranjeiro 		assert(wqe);
721a5bf6af9SAdrien Mazarguil 		/*
722a5bf6af9SAdrien Mazarguil 		 * Spill on next WQE when the current one does not have
723a5bf6af9SAdrien Mazarguil 		 * enough room left. Size of WQE must a be a multiple
724a5bf6af9SAdrien Mazarguil 		 * of data segment size.
725a5bf6af9SAdrien Mazarguil 		 */
7268688b2f8SNélio Laranjeiro 		assert(!(MLX5_WQE_SIZE % MLX5_WQE_DWORD_SIZE));
7276579c27cSNélio Laranjeiro 		if (!(ds % (MLX5_WQE_SIZE / MLX5_WQE_DWORD_SIZE))) {
728f04f1d51SNélio Laranjeiro 			if (unlikely(--max_wqe == 0))
729f04f1d51SNélio Laranjeiro 				break;
7309a7fa9f7SNélio Laranjeiro 			dseg = (volatile rte_v128u32_t *)
731f04f1d51SNélio Laranjeiro 			       tx_mlx5_wqe(txq, txq->wqe_ci + ds / 4);
732f04f1d51SNélio Laranjeiro 			rte_prefetch0(tx_mlx5_wqe(txq,
733f04f1d51SNélio Laranjeiro 						  txq->wqe_ci + ds / 4 + 1));
7346579c27cSNélio Laranjeiro 		} else {
735a5bf6af9SAdrien Mazarguil 			++dseg;
7366579c27cSNélio Laranjeiro 		}
737a5bf6af9SAdrien Mazarguil 		++ds;
738a5bf6af9SAdrien Mazarguil 		buf = buf->next;
739a5bf6af9SAdrien Mazarguil 		assert(buf);
7406579c27cSNélio Laranjeiro 		length = DATA_LEN(buf);
741a5bf6af9SAdrien Mazarguil #ifdef MLX5_PMD_SOFT_COUNTERS
7426579c27cSNélio Laranjeiro 		total_length += length;
743a5bf6af9SAdrien Mazarguil #endif
7446579c27cSNélio Laranjeiro 		/* Store segment information. */
745ebbb81ebSNélio Laranjeiro 		addr = rte_cpu_to_be_64(rte_pktmbuf_mtod(buf, uintptr_t));
7469a7fa9f7SNélio Laranjeiro 		*dseg = (rte_v128u32_t){
7476b30a6a8SShachar Beiser 			rte_cpu_to_be_32(length),
7486cb559d6SYongseok Koh 			mlx5_tx_mb2mr(txq, buf),
749ebbb81ebSNélio Laranjeiro 			addr,
750ebbb81ebSNélio Laranjeiro 			addr >> 32,
7516579c27cSNélio Laranjeiro 		};
7528c819a69SYongseok Koh 		(*txq->elts)[++elts_head & elts_m] = buf;
753f895536bSYongseok Koh 		if (--segs_n)
7546579c27cSNélio Laranjeiro 			goto next_seg;
7556579c27cSNélio Laranjeiro next_pkt:
756883ce172SShahaf Shuler 		if (ds > MLX5_DSEG_MAX) {
757883ce172SShahaf Shuler 			txq->stats.oerrors++;
758883ce172SShahaf Shuler 			break;
759883ce172SShahaf Shuler 		}
7608c819a69SYongseok Koh 		++elts_head;
7613730e6c6SYongseok Koh 		++pkts;
7626579c27cSNélio Laranjeiro 		++i;
763f895536bSYongseok Koh 		j += sg;
764b8fe952eSNélio Laranjeiro 		/* Initialize known and common part of the WQE structure. */
7653f13f8c2SShahaf Shuler 		if (tso) {
7663f13f8c2SShahaf Shuler 			wqe->ctrl = (rte_v128u32_t){
7676b30a6a8SShachar Beiser 				rte_cpu_to_be_32((txq->wqe_ci << 8) |
7686b30a6a8SShachar Beiser 						 MLX5_OPCODE_TSO),
7696b30a6a8SShachar Beiser 				rte_cpu_to_be_32(txq->qp_num_8s | ds),
7703f13f8c2SShahaf Shuler 				0,
7713f13f8c2SShahaf Shuler 				0,
7723f13f8c2SShahaf Shuler 			};
7733f13f8c2SShahaf Shuler 			wqe->eseg = (rte_v128u32_t){
7745f8ba81cSXueming Li 				swp_offsets,
7755f8ba81cSXueming Li 				cs_flags | (swp_types << 8) |
7765f8ba81cSXueming Li 				(rte_cpu_to_be_16(tso_segsz) << 16),
7773f13f8c2SShahaf Shuler 				0,
7786b30a6a8SShachar Beiser 				(ehdr << 16) | rte_cpu_to_be_16(tso_header_sz),
7793f13f8c2SShahaf Shuler 			};
7803f13f8c2SShahaf Shuler 		} else {
7819a7fa9f7SNélio Laranjeiro 			wqe->ctrl = (rte_v128u32_t){
7826b30a6a8SShachar Beiser 				rte_cpu_to_be_32((txq->wqe_ci << 8) |
7836b30a6a8SShachar Beiser 						 MLX5_OPCODE_SEND),
7846b30a6a8SShachar Beiser 				rte_cpu_to_be_32(txq->qp_num_8s | ds),
7859a7fa9f7SNélio Laranjeiro 				0,
7869a7fa9f7SNélio Laranjeiro 				0,
7879a7fa9f7SNélio Laranjeiro 			};
7889a7fa9f7SNélio Laranjeiro 			wqe->eseg = (rte_v128u32_t){
7895f8ba81cSXueming Li 				swp_offsets,
7905f8ba81cSXueming Li 				cs_flags | (swp_types << 8),
7919a7fa9f7SNélio Laranjeiro 				0,
7926b30a6a8SShachar Beiser 				(ehdr << 16) | rte_cpu_to_be_16(pkt_inline_sz),
7939a7fa9f7SNélio Laranjeiro 			};
7943f13f8c2SShahaf Shuler 		}
7953f13f8c2SShahaf Shuler next_wqe:
7966579c27cSNélio Laranjeiro 		txq->wqe_ci += (ds + 3) / 4;
797ac180a21SYongseok Koh 		/* Save the last successful WQE for completion request */
798ac180a21SYongseok Koh 		last_wqe = (volatile struct mlx5_wqe_ctrl *)wqe;
79987011737SAdrien Mazarguil #ifdef MLX5_PMD_SOFT_COUNTERS
800573f54afSNélio Laranjeiro 		/* Increment sent bytes counter. */
8016579c27cSNélio Laranjeiro 		txq->stats.obytes += total_length;
80287011737SAdrien Mazarguil #endif
8033730e6c6SYongseok Koh 	} while (i < pkts_n);
8042e22920bSAdrien Mazarguil 	/* Take a shortcut if nothing must be sent. */
8053f13f8c2SShahaf Shuler 	if (unlikely((i + k) == 0))
8062e22920bSAdrien Mazarguil 		return 0;
8078c819a69SYongseok Koh 	txq->elts_head += (i + j);
808c305090bSAdrien Mazarguil 	/* Check whether completion threshold has been reached. */
8093f13f8c2SShahaf Shuler 	comp = txq->elts_comp + i + j + k;
810c305090bSAdrien Mazarguil 	if (comp >= MLX5_TX_COMP_THRESH) {
811c305090bSAdrien Mazarguil 		/* Request completion on last WQE. */
8126b30a6a8SShachar Beiser 		last_wqe->ctrl2 = rte_cpu_to_be_32(8);
813c305090bSAdrien Mazarguil 		/* Save elts_head in unused "immediate" field of WQE. */
814ac180a21SYongseok Koh 		last_wqe->ctrl3 = txq->elts_head;
815c305090bSAdrien Mazarguil 		txq->elts_comp = 0;
8162eefbec5SYongseok Koh #ifndef NDEBUG
8172eefbec5SYongseok Koh 		++txq->cq_pi;
8182eefbec5SYongseok Koh #endif
819c305090bSAdrien Mazarguil 	} else {
820c305090bSAdrien Mazarguil 		txq->elts_comp = comp;
821c305090bSAdrien Mazarguil 	}
82287011737SAdrien Mazarguil #ifdef MLX5_PMD_SOFT_COUNTERS
82387011737SAdrien Mazarguil 	/* Increment sent packets counter. */
82487011737SAdrien Mazarguil 	txq->stats.opackets += i;
82587011737SAdrien Mazarguil #endif
8262e22920bSAdrien Mazarguil 	/* Ring QP doorbell. */
827ac180a21SYongseok Koh 	mlx5_tx_dbrec(txq, (volatile struct mlx5_wqe *)last_wqe);
8282e22920bSAdrien Mazarguil 	return i;
8292e22920bSAdrien Mazarguil }
8302e22920bSAdrien Mazarguil 
8312e22920bSAdrien Mazarguil /**
832230189d9SNélio Laranjeiro  * Open a MPW session.
833230189d9SNélio Laranjeiro  *
834230189d9SNélio Laranjeiro  * @param txq
835230189d9SNélio Laranjeiro  *   Pointer to TX queue structure.
836230189d9SNélio Laranjeiro  * @param mpw
837230189d9SNélio Laranjeiro  *   Pointer to MPW session structure.
838230189d9SNélio Laranjeiro  * @param length
839230189d9SNélio Laranjeiro  *   Packet length.
840230189d9SNélio Laranjeiro  */
841230189d9SNélio Laranjeiro static inline void
842991b04f6SNélio Laranjeiro mlx5_mpw_new(struct mlx5_txq_data *txq, struct mlx5_mpw *mpw, uint32_t length)
843230189d9SNélio Laranjeiro {
844a821d09dSNélio Laranjeiro 	uint16_t idx = txq->wqe_ci & ((1 << txq->wqe_n) - 1);
845230189d9SNélio Laranjeiro 	volatile struct mlx5_wqe_data_seg (*dseg)[MLX5_MPW_DSEG_MAX] =
846230189d9SNélio Laranjeiro 		(volatile struct mlx5_wqe_data_seg (*)[])
847fdcb0f53SNélio Laranjeiro 		tx_mlx5_wqe(txq, idx + 1);
848230189d9SNélio Laranjeiro 
849230189d9SNélio Laranjeiro 	mpw->state = MLX5_MPW_STATE_OPENED;
850230189d9SNélio Laranjeiro 	mpw->pkts_n = 0;
851230189d9SNélio Laranjeiro 	mpw->len = length;
852230189d9SNélio Laranjeiro 	mpw->total_len = 0;
853fdcb0f53SNélio Laranjeiro 	mpw->wqe = (volatile struct mlx5_wqe *)tx_mlx5_wqe(txq, idx);
8546b30a6a8SShachar Beiser 	mpw->wqe->eseg.mss = rte_cpu_to_be_16(length);
8558688b2f8SNélio Laranjeiro 	mpw->wqe->eseg.inline_hdr_sz = 0;
8568688b2f8SNélio Laranjeiro 	mpw->wqe->eseg.rsvd0 = 0;
8578688b2f8SNélio Laranjeiro 	mpw->wqe->eseg.rsvd1 = 0;
8588688b2f8SNélio Laranjeiro 	mpw->wqe->eseg.rsvd2 = 0;
8596b30a6a8SShachar Beiser 	mpw->wqe->ctrl[0] = rte_cpu_to_be_32((MLX5_OPC_MOD_MPW << 24) |
8606b30a6a8SShachar Beiser 					     (txq->wqe_ci << 8) |
8616b30a6a8SShachar Beiser 					     MLX5_OPCODE_TSO);
8628688b2f8SNélio Laranjeiro 	mpw->wqe->ctrl[2] = 0;
8638688b2f8SNélio Laranjeiro 	mpw->wqe->ctrl[3] = 0;
8648688b2f8SNélio Laranjeiro 	mpw->data.dseg[0] = (volatile struct mlx5_wqe_data_seg *)
8658688b2f8SNélio Laranjeiro 		(((uintptr_t)mpw->wqe) + (2 * MLX5_WQE_DWORD_SIZE));
8668688b2f8SNélio Laranjeiro 	mpw->data.dseg[1] = (volatile struct mlx5_wqe_data_seg *)
8678688b2f8SNélio Laranjeiro 		(((uintptr_t)mpw->wqe) + (3 * MLX5_WQE_DWORD_SIZE));
868230189d9SNélio Laranjeiro 	mpw->data.dseg[2] = &(*dseg)[0];
869230189d9SNélio Laranjeiro 	mpw->data.dseg[3] = &(*dseg)[1];
870230189d9SNélio Laranjeiro 	mpw->data.dseg[4] = &(*dseg)[2];
871230189d9SNélio Laranjeiro }
872230189d9SNélio Laranjeiro 
873230189d9SNélio Laranjeiro /**
874230189d9SNélio Laranjeiro  * Close a MPW session.
875230189d9SNélio Laranjeiro  *
876230189d9SNélio Laranjeiro  * @param txq
877230189d9SNélio Laranjeiro  *   Pointer to TX queue structure.
878230189d9SNélio Laranjeiro  * @param mpw
879230189d9SNélio Laranjeiro  *   Pointer to MPW session structure.
880230189d9SNélio Laranjeiro  */
881230189d9SNélio Laranjeiro static inline void
882991b04f6SNélio Laranjeiro mlx5_mpw_close(struct mlx5_txq_data *txq, struct mlx5_mpw *mpw)
883230189d9SNélio Laranjeiro {
884230189d9SNélio Laranjeiro 	unsigned int num = mpw->pkts_n;
885230189d9SNélio Laranjeiro 
886230189d9SNélio Laranjeiro 	/*
887230189d9SNélio Laranjeiro 	 * Store size in multiple of 16 bytes. Control and Ethernet segments
888230189d9SNélio Laranjeiro 	 * count as 2.
889230189d9SNélio Laranjeiro 	 */
8906b30a6a8SShachar Beiser 	mpw->wqe->ctrl[1] = rte_cpu_to_be_32(txq->qp_num_8s | (2 + num));
891230189d9SNélio Laranjeiro 	mpw->state = MLX5_MPW_STATE_CLOSED;
892230189d9SNélio Laranjeiro 	if (num < 3)
893230189d9SNélio Laranjeiro 		++txq->wqe_ci;
894230189d9SNélio Laranjeiro 	else
895230189d9SNélio Laranjeiro 		txq->wqe_ci += 2;
896fdcb0f53SNélio Laranjeiro 	rte_prefetch0(tx_mlx5_wqe(txq, txq->wqe_ci));
897fdcb0f53SNélio Laranjeiro 	rte_prefetch0(tx_mlx5_wqe(txq, txq->wqe_ci + 1));
898230189d9SNélio Laranjeiro }
899230189d9SNélio Laranjeiro 
900230189d9SNélio Laranjeiro /**
901230189d9SNélio Laranjeiro  * DPDK callback for TX with MPW support.
902230189d9SNélio Laranjeiro  *
903230189d9SNélio Laranjeiro  * @param dpdk_txq
904230189d9SNélio Laranjeiro  *   Generic pointer to TX queue structure.
905230189d9SNélio Laranjeiro  * @param[in] pkts
906230189d9SNélio Laranjeiro  *   Packets to transmit.
907230189d9SNélio Laranjeiro  * @param pkts_n
908230189d9SNélio Laranjeiro  *   Number of packets in array.
909230189d9SNélio Laranjeiro  *
910230189d9SNélio Laranjeiro  * @return
911230189d9SNélio Laranjeiro  *   Number of packets successfully transmitted (<= pkts_n).
912230189d9SNélio Laranjeiro  */
913230189d9SNélio Laranjeiro uint16_t
914230189d9SNélio Laranjeiro mlx5_tx_burst_mpw(void *dpdk_txq, struct rte_mbuf **pkts, uint16_t pkts_n)
915230189d9SNélio Laranjeiro {
916991b04f6SNélio Laranjeiro 	struct mlx5_txq_data *txq = (struct mlx5_txq_data *)dpdk_txq;
917230189d9SNélio Laranjeiro 	uint16_t elts_head = txq->elts_head;
9188c819a69SYongseok Koh 	const uint16_t elts_n = 1 << txq->elts_n;
9198c819a69SYongseok Koh 	const uint16_t elts_m = elts_n - 1;
920c3d62cc9SAdrien Mazarguil 	unsigned int i = 0;
921a5bf6af9SAdrien Mazarguil 	unsigned int j = 0;
9228c819a69SYongseok Koh 	uint16_t max_elts;
923f04f1d51SNélio Laranjeiro 	uint16_t max_wqe;
924230189d9SNélio Laranjeiro 	unsigned int comp;
925230189d9SNélio Laranjeiro 	struct mlx5_mpw mpw = {
926230189d9SNélio Laranjeiro 		.state = MLX5_MPW_STATE_CLOSED,
927230189d9SNélio Laranjeiro 	};
928230189d9SNélio Laranjeiro 
929c3d62cc9SAdrien Mazarguil 	if (unlikely(!pkts_n))
930c3d62cc9SAdrien Mazarguil 		return 0;
931230189d9SNélio Laranjeiro 	/* Prefetch first packet cacheline. */
932fdcb0f53SNélio Laranjeiro 	rte_prefetch0(tx_mlx5_wqe(txq, txq->wqe_ci));
933fdcb0f53SNélio Laranjeiro 	rte_prefetch0(tx_mlx5_wqe(txq, txq->wqe_ci + 1));
934230189d9SNélio Laranjeiro 	/* Start processing. */
9356cb559d6SYongseok Koh 	mlx5_tx_complete(txq);
9368c819a69SYongseok Koh 	max_elts = (elts_n - (elts_head - txq->elts_tail));
9372eefbec5SYongseok Koh 	/* A CQE slot must always be available. */
9382eefbec5SYongseok Koh 	assert((1u << txq->cqe_n) - (txq->cq_pi - txq->cq_ci));
939f04f1d51SNélio Laranjeiro 	max_wqe = (1u << txq->wqe_n) - (txq->wqe_ci - txq->wqe_pi);
940f04f1d51SNélio Laranjeiro 	if (unlikely(!max_wqe))
941f04f1d51SNélio Laranjeiro 		return 0;
942c3d62cc9SAdrien Mazarguil 	do {
943a5bf6af9SAdrien Mazarguil 		struct rte_mbuf *buf = *(pkts++);
944230189d9SNélio Laranjeiro 		uint32_t length;
945a5bf6af9SAdrien Mazarguil 		unsigned int segs_n = buf->nb_segs;
9464aa15eb1SNélio Laranjeiro 		uint32_t cs_flags;
947230189d9SNélio Laranjeiro 
948c3d62cc9SAdrien Mazarguil 		/*
949c3d62cc9SAdrien Mazarguil 		 * Make sure there is enough room to store this packet and
950c3d62cc9SAdrien Mazarguil 		 * that one ring entry remains unused.
951c3d62cc9SAdrien Mazarguil 		 */
952a5bf6af9SAdrien Mazarguil 		assert(segs_n);
9538c819a69SYongseok Koh 		if (max_elts < segs_n)
954c3d62cc9SAdrien Mazarguil 			break;
955a5bf6af9SAdrien Mazarguil 		/* Do not bother with large packets MPW cannot handle. */
95624c14430SShahaf Shuler 		if (segs_n > MLX5_MPW_DSEG_MAX) {
95724c14430SShahaf Shuler 			txq->stats.oerrors++;
958a5bf6af9SAdrien Mazarguil 			break;
95924c14430SShahaf Shuler 		}
9608c819a69SYongseok Koh 		max_elts -= segs_n;
961c3d62cc9SAdrien Mazarguil 		--pkts_n;
9625f8ba81cSXueming Li 		cs_flags = txq_ol_cksum_to_cs(buf);
963a5bf6af9SAdrien Mazarguil 		/* Retrieve packet information. */
964a5bf6af9SAdrien Mazarguil 		length = PKT_LEN(buf);
965a5bf6af9SAdrien Mazarguil 		assert(length);
966230189d9SNélio Laranjeiro 		/* Start new session if packet differs. */
967230189d9SNélio Laranjeiro 		if ((mpw.state == MLX5_MPW_STATE_OPENED) &&
968230189d9SNélio Laranjeiro 		    ((mpw.len != length) ||
969a5bf6af9SAdrien Mazarguil 		     (segs_n != 1) ||
9708688b2f8SNélio Laranjeiro 		     (mpw.wqe->eseg.cs_flags != cs_flags)))
971230189d9SNélio Laranjeiro 			mlx5_mpw_close(txq, &mpw);
972230189d9SNélio Laranjeiro 		if (mpw.state == MLX5_MPW_STATE_CLOSED) {
973f04f1d51SNélio Laranjeiro 			/*
974f04f1d51SNélio Laranjeiro 			 * Multi-Packet WQE consumes at most two WQE.
975f04f1d51SNélio Laranjeiro 			 * mlx5_mpw_new() expects to be able to use such
976f04f1d51SNélio Laranjeiro 			 * resources.
977f04f1d51SNélio Laranjeiro 			 */
978f04f1d51SNélio Laranjeiro 			if (unlikely(max_wqe < 2))
979f04f1d51SNélio Laranjeiro 				break;
980f04f1d51SNélio Laranjeiro 			max_wqe -= 2;
981230189d9SNélio Laranjeiro 			mlx5_mpw_new(txq, &mpw, length);
9828688b2f8SNélio Laranjeiro 			mpw.wqe->eseg.cs_flags = cs_flags;
983230189d9SNélio Laranjeiro 		}
984a5bf6af9SAdrien Mazarguil 		/* Multi-segment packets must be alone in their MPW. */
985a5bf6af9SAdrien Mazarguil 		assert((segs_n == 1) || (mpw.pkts_n == 0));
986a5bf6af9SAdrien Mazarguil #if defined(MLX5_PMD_SOFT_COUNTERS) || !defined(NDEBUG)
987a5bf6af9SAdrien Mazarguil 		length = 0;
988a5bf6af9SAdrien Mazarguil #endif
989a5bf6af9SAdrien Mazarguil 		do {
990a5bf6af9SAdrien Mazarguil 			volatile struct mlx5_wqe_data_seg *dseg;
991a5bf6af9SAdrien Mazarguil 			uintptr_t addr;
992a5bf6af9SAdrien Mazarguil 
993a5bf6af9SAdrien Mazarguil 			assert(buf);
9948c819a69SYongseok Koh 			(*txq->elts)[elts_head++ & elts_m] = buf;
995230189d9SNélio Laranjeiro 			dseg = mpw.data.dseg[mpw.pkts_n];
996a5bf6af9SAdrien Mazarguil 			addr = rte_pktmbuf_mtod(buf, uintptr_t);
997230189d9SNélio Laranjeiro 			*dseg = (struct mlx5_wqe_data_seg){
9986b30a6a8SShachar Beiser 				.byte_count = rte_cpu_to_be_32(DATA_LEN(buf)),
9996cb559d6SYongseok Koh 				.lkey = mlx5_tx_mb2mr(txq, buf),
10006b30a6a8SShachar Beiser 				.addr = rte_cpu_to_be_64(addr),
1001230189d9SNélio Laranjeiro 			};
1002a5bf6af9SAdrien Mazarguil #if defined(MLX5_PMD_SOFT_COUNTERS) || !defined(NDEBUG)
1003a5bf6af9SAdrien Mazarguil 			length += DATA_LEN(buf);
1004a5bf6af9SAdrien Mazarguil #endif
1005a5bf6af9SAdrien Mazarguil 			buf = buf->next;
1006230189d9SNélio Laranjeiro 			++mpw.pkts_n;
1007a5bf6af9SAdrien Mazarguil 			++j;
1008a5bf6af9SAdrien Mazarguil 		} while (--segs_n);
1009a5bf6af9SAdrien Mazarguil 		assert(length == mpw.len);
1010230189d9SNélio Laranjeiro 		if (mpw.pkts_n == MLX5_MPW_DSEG_MAX)
1011230189d9SNélio Laranjeiro 			mlx5_mpw_close(txq, &mpw);
1012230189d9SNélio Laranjeiro #ifdef MLX5_PMD_SOFT_COUNTERS
1013230189d9SNélio Laranjeiro 		/* Increment sent bytes counter. */
1014230189d9SNélio Laranjeiro 		txq->stats.obytes += length;
1015230189d9SNélio Laranjeiro #endif
1016c3d62cc9SAdrien Mazarguil 		++i;
1017c3d62cc9SAdrien Mazarguil 	} while (pkts_n);
1018230189d9SNélio Laranjeiro 	/* Take a shortcut if nothing must be sent. */
1019230189d9SNélio Laranjeiro 	if (unlikely(i == 0))
1020230189d9SNélio Laranjeiro 		return 0;
1021230189d9SNélio Laranjeiro 	/* Check whether completion threshold has been reached. */
1022a5bf6af9SAdrien Mazarguil 	/* "j" includes both packets and segments. */
1023a5bf6af9SAdrien Mazarguil 	comp = txq->elts_comp + j;
1024230189d9SNélio Laranjeiro 	if (comp >= MLX5_TX_COMP_THRESH) {
10258688b2f8SNélio Laranjeiro 		volatile struct mlx5_wqe *wqe = mpw.wqe;
1026230189d9SNélio Laranjeiro 
1027230189d9SNélio Laranjeiro 		/* Request completion on last WQE. */
10286b30a6a8SShachar Beiser 		wqe->ctrl[2] = rte_cpu_to_be_32(8);
1029230189d9SNélio Laranjeiro 		/* Save elts_head in unused "immediate" field of WQE. */
10308688b2f8SNélio Laranjeiro 		wqe->ctrl[3] = elts_head;
1031230189d9SNélio Laranjeiro 		txq->elts_comp = 0;
10322eefbec5SYongseok Koh #ifndef NDEBUG
10332eefbec5SYongseok Koh 		++txq->cq_pi;
10342eefbec5SYongseok Koh #endif
1035230189d9SNélio Laranjeiro 	} else {
1036230189d9SNélio Laranjeiro 		txq->elts_comp = comp;
1037230189d9SNélio Laranjeiro 	}
1038230189d9SNélio Laranjeiro #ifdef MLX5_PMD_SOFT_COUNTERS
1039230189d9SNélio Laranjeiro 	/* Increment sent packets counter. */
1040230189d9SNélio Laranjeiro 	txq->stats.opackets += i;
1041230189d9SNélio Laranjeiro #endif
1042230189d9SNélio Laranjeiro 	/* Ring QP doorbell. */
1043230189d9SNélio Laranjeiro 	if (mpw.state == MLX5_MPW_STATE_OPENED)
1044230189d9SNélio Laranjeiro 		mlx5_mpw_close(txq, &mpw);
104530807f62SNélio Laranjeiro 	mlx5_tx_dbrec(txq, mpw.wqe);
1046230189d9SNélio Laranjeiro 	txq->elts_head = elts_head;
1047230189d9SNélio Laranjeiro 	return i;
1048230189d9SNélio Laranjeiro }
1049230189d9SNélio Laranjeiro 
1050230189d9SNélio Laranjeiro /**
1051230189d9SNélio Laranjeiro  * Open a MPW inline session.
1052230189d9SNélio Laranjeiro  *
1053230189d9SNélio Laranjeiro  * @param txq
1054230189d9SNélio Laranjeiro  *   Pointer to TX queue structure.
1055230189d9SNélio Laranjeiro  * @param mpw
1056230189d9SNélio Laranjeiro  *   Pointer to MPW session structure.
1057230189d9SNélio Laranjeiro  * @param length
1058230189d9SNélio Laranjeiro  *   Packet length.
1059230189d9SNélio Laranjeiro  */
1060230189d9SNélio Laranjeiro static inline void
1061991b04f6SNélio Laranjeiro mlx5_mpw_inline_new(struct mlx5_txq_data *txq, struct mlx5_mpw *mpw,
1062991b04f6SNélio Laranjeiro 		    uint32_t length)
1063230189d9SNélio Laranjeiro {
1064a821d09dSNélio Laranjeiro 	uint16_t idx = txq->wqe_ci & ((1 << txq->wqe_n) - 1);
10658688b2f8SNélio Laranjeiro 	struct mlx5_wqe_inl_small *inl;
1066230189d9SNélio Laranjeiro 
1067230189d9SNélio Laranjeiro 	mpw->state = MLX5_MPW_INL_STATE_OPENED;
1068230189d9SNélio Laranjeiro 	mpw->pkts_n = 0;
1069230189d9SNélio Laranjeiro 	mpw->len = length;
1070230189d9SNélio Laranjeiro 	mpw->total_len = 0;
1071fdcb0f53SNélio Laranjeiro 	mpw->wqe = (volatile struct mlx5_wqe *)tx_mlx5_wqe(txq, idx);
10726b30a6a8SShachar Beiser 	mpw->wqe->ctrl[0] = rte_cpu_to_be_32((MLX5_OPC_MOD_MPW << 24) |
1073230189d9SNélio Laranjeiro 					     (txq->wqe_ci << 8) |
1074c904ae25SNélio Laranjeiro 					     MLX5_OPCODE_TSO);
10758688b2f8SNélio Laranjeiro 	mpw->wqe->ctrl[2] = 0;
10768688b2f8SNélio Laranjeiro 	mpw->wqe->ctrl[3] = 0;
10776b30a6a8SShachar Beiser 	mpw->wqe->eseg.mss = rte_cpu_to_be_16(length);
10788688b2f8SNélio Laranjeiro 	mpw->wqe->eseg.inline_hdr_sz = 0;
10798688b2f8SNélio Laranjeiro 	mpw->wqe->eseg.cs_flags = 0;
10808688b2f8SNélio Laranjeiro 	mpw->wqe->eseg.rsvd0 = 0;
10818688b2f8SNélio Laranjeiro 	mpw->wqe->eseg.rsvd1 = 0;
10828688b2f8SNélio Laranjeiro 	mpw->wqe->eseg.rsvd2 = 0;
10838688b2f8SNélio Laranjeiro 	inl = (struct mlx5_wqe_inl_small *)
10848688b2f8SNélio Laranjeiro 		(((uintptr_t)mpw->wqe) + 2 * MLX5_WQE_DWORD_SIZE);
10858688b2f8SNélio Laranjeiro 	mpw->data.raw = (uint8_t *)&inl->raw;
1086230189d9SNélio Laranjeiro }
1087230189d9SNélio Laranjeiro 
1088230189d9SNélio Laranjeiro /**
1089230189d9SNélio Laranjeiro  * Close a MPW inline session.
1090230189d9SNélio Laranjeiro  *
1091230189d9SNélio Laranjeiro  * @param txq
1092230189d9SNélio Laranjeiro  *   Pointer to TX queue structure.
1093230189d9SNélio Laranjeiro  * @param mpw
1094230189d9SNélio Laranjeiro  *   Pointer to MPW session structure.
1095230189d9SNélio Laranjeiro  */
1096230189d9SNélio Laranjeiro static inline void
1097991b04f6SNélio Laranjeiro mlx5_mpw_inline_close(struct mlx5_txq_data *txq, struct mlx5_mpw *mpw)
1098230189d9SNélio Laranjeiro {
1099230189d9SNélio Laranjeiro 	unsigned int size;
11008688b2f8SNélio Laranjeiro 	struct mlx5_wqe_inl_small *inl = (struct mlx5_wqe_inl_small *)
11018688b2f8SNélio Laranjeiro 		(((uintptr_t)mpw->wqe) + (2 * MLX5_WQE_DWORD_SIZE));
1102230189d9SNélio Laranjeiro 
11038688b2f8SNélio Laranjeiro 	size = MLX5_WQE_SIZE - MLX5_MWQE64_INL_DATA + mpw->total_len;
1104230189d9SNélio Laranjeiro 	/*
1105230189d9SNélio Laranjeiro 	 * Store size in multiple of 16 bytes. Control and Ethernet segments
1106230189d9SNélio Laranjeiro 	 * count as 2.
1107230189d9SNélio Laranjeiro 	 */
11086b30a6a8SShachar Beiser 	mpw->wqe->ctrl[1] = rte_cpu_to_be_32(txq->qp_num_8s |
11096b30a6a8SShachar Beiser 					     MLX5_WQE_DS(size));
1110230189d9SNélio Laranjeiro 	mpw->state = MLX5_MPW_STATE_CLOSED;
11116b30a6a8SShachar Beiser 	inl->byte_cnt = rte_cpu_to_be_32(mpw->total_len | MLX5_INLINE_SEG);
11128688b2f8SNélio Laranjeiro 	txq->wqe_ci += (size + (MLX5_WQE_SIZE - 1)) / MLX5_WQE_SIZE;
1113230189d9SNélio Laranjeiro }
1114230189d9SNélio Laranjeiro 
1115230189d9SNélio Laranjeiro /**
1116230189d9SNélio Laranjeiro  * DPDK callback for TX with MPW inline support.
1117230189d9SNélio Laranjeiro  *
1118230189d9SNélio Laranjeiro  * @param dpdk_txq
1119230189d9SNélio Laranjeiro  *   Generic pointer to TX queue structure.
1120230189d9SNélio Laranjeiro  * @param[in] pkts
1121230189d9SNélio Laranjeiro  *   Packets to transmit.
1122230189d9SNélio Laranjeiro  * @param pkts_n
1123230189d9SNélio Laranjeiro  *   Number of packets in array.
1124230189d9SNélio Laranjeiro  *
1125230189d9SNélio Laranjeiro  * @return
1126230189d9SNélio Laranjeiro  *   Number of packets successfully transmitted (<= pkts_n).
1127230189d9SNélio Laranjeiro  */
1128230189d9SNélio Laranjeiro uint16_t
1129230189d9SNélio Laranjeiro mlx5_tx_burst_mpw_inline(void *dpdk_txq, struct rte_mbuf **pkts,
1130230189d9SNélio Laranjeiro 			 uint16_t pkts_n)
1131230189d9SNélio Laranjeiro {
1132991b04f6SNélio Laranjeiro 	struct mlx5_txq_data *txq = (struct mlx5_txq_data *)dpdk_txq;
1133230189d9SNélio Laranjeiro 	uint16_t elts_head = txq->elts_head;
11348c819a69SYongseok Koh 	const uint16_t elts_n = 1 << txq->elts_n;
11358c819a69SYongseok Koh 	const uint16_t elts_m = elts_n - 1;
1136c3d62cc9SAdrien Mazarguil 	unsigned int i = 0;
1137a5bf6af9SAdrien Mazarguil 	unsigned int j = 0;
11388c819a69SYongseok Koh 	uint16_t max_elts;
1139f04f1d51SNélio Laranjeiro 	uint16_t max_wqe;
1140230189d9SNélio Laranjeiro 	unsigned int comp;
11410e8679fcSNélio Laranjeiro 	unsigned int inline_room = txq->max_inline * RTE_CACHE_LINE_SIZE;
1142230189d9SNélio Laranjeiro 	struct mlx5_mpw mpw = {
1143230189d9SNélio Laranjeiro 		.state = MLX5_MPW_STATE_CLOSED,
1144230189d9SNélio Laranjeiro 	};
1145f04f1d51SNélio Laranjeiro 	/*
1146f04f1d51SNélio Laranjeiro 	 * Compute the maximum number of WQE which can be consumed by inline
1147f04f1d51SNélio Laranjeiro 	 * code.
1148f04f1d51SNélio Laranjeiro 	 * - 2 DSEG for:
1149f04f1d51SNélio Laranjeiro 	 *   - 1 control segment,
1150f04f1d51SNélio Laranjeiro 	 *   - 1 Ethernet segment,
1151f04f1d51SNélio Laranjeiro 	 * - N Dseg from the inline request.
1152f04f1d51SNélio Laranjeiro 	 */
1153f04f1d51SNélio Laranjeiro 	const unsigned int wqe_inl_n =
1154f04f1d51SNélio Laranjeiro 		((2 * MLX5_WQE_DWORD_SIZE +
1155f04f1d51SNélio Laranjeiro 		  txq->max_inline * RTE_CACHE_LINE_SIZE) +
1156f04f1d51SNélio Laranjeiro 		 RTE_CACHE_LINE_SIZE - 1) / RTE_CACHE_LINE_SIZE;
1157230189d9SNélio Laranjeiro 
1158c3d62cc9SAdrien Mazarguil 	if (unlikely(!pkts_n))
1159c3d62cc9SAdrien Mazarguil 		return 0;
1160230189d9SNélio Laranjeiro 	/* Prefetch first packet cacheline. */
1161fdcb0f53SNélio Laranjeiro 	rte_prefetch0(tx_mlx5_wqe(txq, txq->wqe_ci));
1162fdcb0f53SNélio Laranjeiro 	rte_prefetch0(tx_mlx5_wqe(txq, txq->wqe_ci + 1));
1163230189d9SNélio Laranjeiro 	/* Start processing. */
11646cb559d6SYongseok Koh 	mlx5_tx_complete(txq);
11658c819a69SYongseok Koh 	max_elts = (elts_n - (elts_head - txq->elts_tail));
11662eefbec5SYongseok Koh 	/* A CQE slot must always be available. */
11672eefbec5SYongseok Koh 	assert((1u << txq->cqe_n) - (txq->cq_pi - txq->cq_ci));
1168c3d62cc9SAdrien Mazarguil 	do {
1169a5bf6af9SAdrien Mazarguil 		struct rte_mbuf *buf = *(pkts++);
1170230189d9SNélio Laranjeiro 		uintptr_t addr;
1171230189d9SNélio Laranjeiro 		uint32_t length;
1172a5bf6af9SAdrien Mazarguil 		unsigned int segs_n = buf->nb_segs;
11734aa15eb1SNélio Laranjeiro 		uint8_t cs_flags;
1174230189d9SNélio Laranjeiro 
1175c3d62cc9SAdrien Mazarguil 		/*
1176c3d62cc9SAdrien Mazarguil 		 * Make sure there is enough room to store this packet and
1177c3d62cc9SAdrien Mazarguil 		 * that one ring entry remains unused.
1178c3d62cc9SAdrien Mazarguil 		 */
1179a5bf6af9SAdrien Mazarguil 		assert(segs_n);
11808c819a69SYongseok Koh 		if (max_elts < segs_n)
1181c3d62cc9SAdrien Mazarguil 			break;
1182a5bf6af9SAdrien Mazarguil 		/* Do not bother with large packets MPW cannot handle. */
118324c14430SShahaf Shuler 		if (segs_n > MLX5_MPW_DSEG_MAX) {
118424c14430SShahaf Shuler 			txq->stats.oerrors++;
1185a5bf6af9SAdrien Mazarguil 			break;
118624c14430SShahaf Shuler 		}
11878c819a69SYongseok Koh 		max_elts -= segs_n;
1188c3d62cc9SAdrien Mazarguil 		--pkts_n;
1189f04f1d51SNélio Laranjeiro 		/*
1190f04f1d51SNélio Laranjeiro 		 * Compute max_wqe in case less WQE were consumed in previous
1191f04f1d51SNélio Laranjeiro 		 * iteration.
1192f04f1d51SNélio Laranjeiro 		 */
1193f04f1d51SNélio Laranjeiro 		max_wqe = (1u << txq->wqe_n) - (txq->wqe_ci - txq->wqe_pi);
11945f8ba81cSXueming Li 		cs_flags = txq_ol_cksum_to_cs(buf);
1195a5bf6af9SAdrien Mazarguil 		/* Retrieve packet information. */
1196a5bf6af9SAdrien Mazarguil 		length = PKT_LEN(buf);
1197230189d9SNélio Laranjeiro 		/* Start new session if packet differs. */
1198230189d9SNélio Laranjeiro 		if (mpw.state == MLX5_MPW_STATE_OPENED) {
1199230189d9SNélio Laranjeiro 			if ((mpw.len != length) ||
1200a5bf6af9SAdrien Mazarguil 			    (segs_n != 1) ||
12018688b2f8SNélio Laranjeiro 			    (mpw.wqe->eseg.cs_flags != cs_flags))
1202230189d9SNélio Laranjeiro 				mlx5_mpw_close(txq, &mpw);
1203230189d9SNélio Laranjeiro 		} else if (mpw.state == MLX5_MPW_INL_STATE_OPENED) {
1204230189d9SNélio Laranjeiro 			if ((mpw.len != length) ||
1205a5bf6af9SAdrien Mazarguil 			    (segs_n != 1) ||
1206230189d9SNélio Laranjeiro 			    (length > inline_room) ||
12078688b2f8SNélio Laranjeiro 			    (mpw.wqe->eseg.cs_flags != cs_flags)) {
1208230189d9SNélio Laranjeiro 				mlx5_mpw_inline_close(txq, &mpw);
12090e8679fcSNélio Laranjeiro 				inline_room =
12100e8679fcSNélio Laranjeiro 					txq->max_inline * RTE_CACHE_LINE_SIZE;
1211230189d9SNélio Laranjeiro 			}
1212230189d9SNélio Laranjeiro 		}
1213230189d9SNélio Laranjeiro 		if (mpw.state == MLX5_MPW_STATE_CLOSED) {
1214a5bf6af9SAdrien Mazarguil 			if ((segs_n != 1) ||
1215a5bf6af9SAdrien Mazarguil 			    (length > inline_room)) {
1216f04f1d51SNélio Laranjeiro 				/*
1217f04f1d51SNélio Laranjeiro 				 * Multi-Packet WQE consumes at most two WQE.
1218f04f1d51SNélio Laranjeiro 				 * mlx5_mpw_new() expects to be able to use
1219f04f1d51SNélio Laranjeiro 				 * such resources.
1220f04f1d51SNélio Laranjeiro 				 */
1221f04f1d51SNélio Laranjeiro 				if (unlikely(max_wqe < 2))
1222f04f1d51SNélio Laranjeiro 					break;
1223f04f1d51SNélio Laranjeiro 				max_wqe -= 2;
1224230189d9SNélio Laranjeiro 				mlx5_mpw_new(txq, &mpw, length);
12258688b2f8SNélio Laranjeiro 				mpw.wqe->eseg.cs_flags = cs_flags;
1226230189d9SNélio Laranjeiro 			} else {
1227f04f1d51SNélio Laranjeiro 				if (unlikely(max_wqe < wqe_inl_n))
1228f04f1d51SNélio Laranjeiro 					break;
1229f04f1d51SNélio Laranjeiro 				max_wqe -= wqe_inl_n;
1230230189d9SNélio Laranjeiro 				mlx5_mpw_inline_new(txq, &mpw, length);
12318688b2f8SNélio Laranjeiro 				mpw.wqe->eseg.cs_flags = cs_flags;
1232230189d9SNélio Laranjeiro 			}
1233230189d9SNélio Laranjeiro 		}
1234a5bf6af9SAdrien Mazarguil 		/* Multi-segment packets must be alone in their MPW. */
1235a5bf6af9SAdrien Mazarguil 		assert((segs_n == 1) || (mpw.pkts_n == 0));
1236230189d9SNélio Laranjeiro 		if (mpw.state == MLX5_MPW_STATE_OPENED) {
12370e8679fcSNélio Laranjeiro 			assert(inline_room ==
12380e8679fcSNélio Laranjeiro 			       txq->max_inline * RTE_CACHE_LINE_SIZE);
1239a5bf6af9SAdrien Mazarguil #if defined(MLX5_PMD_SOFT_COUNTERS) || !defined(NDEBUG)
1240a5bf6af9SAdrien Mazarguil 			length = 0;
1241a5bf6af9SAdrien Mazarguil #endif
1242a5bf6af9SAdrien Mazarguil 			do {
1243230189d9SNélio Laranjeiro 				volatile struct mlx5_wqe_data_seg *dseg;
1244230189d9SNélio Laranjeiro 
1245a5bf6af9SAdrien Mazarguil 				assert(buf);
12468c819a69SYongseok Koh 				(*txq->elts)[elts_head++ & elts_m] = buf;
1247230189d9SNélio Laranjeiro 				dseg = mpw.data.dseg[mpw.pkts_n];
1248a5bf6af9SAdrien Mazarguil 				addr = rte_pktmbuf_mtod(buf, uintptr_t);
1249230189d9SNélio Laranjeiro 				*dseg = (struct mlx5_wqe_data_seg){
12506b30a6a8SShachar Beiser 					.byte_count =
12516b30a6a8SShachar Beiser 					       rte_cpu_to_be_32(DATA_LEN(buf)),
12526cb559d6SYongseok Koh 					.lkey = mlx5_tx_mb2mr(txq, buf),
12536b30a6a8SShachar Beiser 					.addr = rte_cpu_to_be_64(addr),
1254230189d9SNélio Laranjeiro 				};
1255a5bf6af9SAdrien Mazarguil #if defined(MLX5_PMD_SOFT_COUNTERS) || !defined(NDEBUG)
1256a5bf6af9SAdrien Mazarguil 				length += DATA_LEN(buf);
1257a5bf6af9SAdrien Mazarguil #endif
1258a5bf6af9SAdrien Mazarguil 				buf = buf->next;
1259230189d9SNélio Laranjeiro 				++mpw.pkts_n;
1260a5bf6af9SAdrien Mazarguil 				++j;
1261a5bf6af9SAdrien Mazarguil 			} while (--segs_n);
1262a5bf6af9SAdrien Mazarguil 			assert(length == mpw.len);
1263230189d9SNélio Laranjeiro 			if (mpw.pkts_n == MLX5_MPW_DSEG_MAX)
1264230189d9SNélio Laranjeiro 				mlx5_mpw_close(txq, &mpw);
1265230189d9SNélio Laranjeiro 		} else {
1266230189d9SNélio Laranjeiro 			unsigned int max;
1267230189d9SNélio Laranjeiro 
1268230189d9SNélio Laranjeiro 			assert(mpw.state == MLX5_MPW_INL_STATE_OPENED);
1269230189d9SNélio Laranjeiro 			assert(length <= inline_room);
1270a5bf6af9SAdrien Mazarguil 			assert(length == DATA_LEN(buf));
1271a5bf6af9SAdrien Mazarguil 			addr = rte_pktmbuf_mtod(buf, uintptr_t);
12728c819a69SYongseok Koh 			(*txq->elts)[elts_head++ & elts_m] = buf;
1273230189d9SNélio Laranjeiro 			/* Maximum number of bytes before wrapping. */
1274fdcb0f53SNélio Laranjeiro 			max = ((((uintptr_t)(txq->wqes)) +
1275fdcb0f53SNélio Laranjeiro 				(1 << txq->wqe_n) *
1276fdcb0f53SNélio Laranjeiro 				MLX5_WQE_SIZE) -
1277230189d9SNélio Laranjeiro 			       (uintptr_t)mpw.data.raw);
1278230189d9SNélio Laranjeiro 			if (length > max) {
1279230189d9SNélio Laranjeiro 				rte_memcpy((void *)(uintptr_t)mpw.data.raw,
1280230189d9SNélio Laranjeiro 					   (void *)addr,
1281230189d9SNélio Laranjeiro 					   max);
1282fdcb0f53SNélio Laranjeiro 				mpw.data.raw = (volatile void *)txq->wqes;
1283230189d9SNélio Laranjeiro 				rte_memcpy((void *)(uintptr_t)mpw.data.raw,
1284230189d9SNélio Laranjeiro 					   (void *)(addr + max),
1285230189d9SNélio Laranjeiro 					   length - max);
1286230189d9SNélio Laranjeiro 				mpw.data.raw += length - max;
1287230189d9SNélio Laranjeiro 			} else {
1288230189d9SNélio Laranjeiro 				rte_memcpy((void *)(uintptr_t)mpw.data.raw,
1289230189d9SNélio Laranjeiro 					   (void *)addr,
1290230189d9SNélio Laranjeiro 					   length);
129116c64768SYongseok Koh 
129216c64768SYongseok Koh 				if (length == max)
129316c64768SYongseok Koh 					mpw.data.raw =
129416c64768SYongseok Koh 						(volatile void *)txq->wqes;
129516c64768SYongseok Koh 				else
1296230189d9SNélio Laranjeiro 					mpw.data.raw += length;
1297230189d9SNélio Laranjeiro 			}
1298230189d9SNélio Laranjeiro 			++mpw.pkts_n;
129976bf1574SYongseok Koh 			mpw.total_len += length;
1300a5bf6af9SAdrien Mazarguil 			++j;
1301230189d9SNélio Laranjeiro 			if (mpw.pkts_n == MLX5_MPW_DSEG_MAX) {
1302230189d9SNélio Laranjeiro 				mlx5_mpw_inline_close(txq, &mpw);
13030e8679fcSNélio Laranjeiro 				inline_room =
13040e8679fcSNélio Laranjeiro 					txq->max_inline * RTE_CACHE_LINE_SIZE;
1305230189d9SNélio Laranjeiro 			} else {
1306230189d9SNélio Laranjeiro 				inline_room -= length;
1307230189d9SNélio Laranjeiro 			}
1308230189d9SNélio Laranjeiro 		}
1309230189d9SNélio Laranjeiro #ifdef MLX5_PMD_SOFT_COUNTERS
1310230189d9SNélio Laranjeiro 		/* Increment sent bytes counter. */
1311230189d9SNélio Laranjeiro 		txq->stats.obytes += length;
1312230189d9SNélio Laranjeiro #endif
1313c3d62cc9SAdrien Mazarguil 		++i;
1314c3d62cc9SAdrien Mazarguil 	} while (pkts_n);
1315230189d9SNélio Laranjeiro 	/* Take a shortcut if nothing must be sent. */
1316230189d9SNélio Laranjeiro 	if (unlikely(i == 0))
1317230189d9SNélio Laranjeiro 		return 0;
1318230189d9SNélio Laranjeiro 	/* Check whether completion threshold has been reached. */
1319a5bf6af9SAdrien Mazarguil 	/* "j" includes both packets and segments. */
1320a5bf6af9SAdrien Mazarguil 	comp = txq->elts_comp + j;
1321230189d9SNélio Laranjeiro 	if (comp >= MLX5_TX_COMP_THRESH) {
13228688b2f8SNélio Laranjeiro 		volatile struct mlx5_wqe *wqe = mpw.wqe;
1323230189d9SNélio Laranjeiro 
1324230189d9SNélio Laranjeiro 		/* Request completion on last WQE. */
13256b30a6a8SShachar Beiser 		wqe->ctrl[2] = rte_cpu_to_be_32(8);
1326230189d9SNélio Laranjeiro 		/* Save elts_head in unused "immediate" field of WQE. */
13278688b2f8SNélio Laranjeiro 		wqe->ctrl[3] = elts_head;
1328230189d9SNélio Laranjeiro 		txq->elts_comp = 0;
13292eefbec5SYongseok Koh #ifndef NDEBUG
13302eefbec5SYongseok Koh 		++txq->cq_pi;
13312eefbec5SYongseok Koh #endif
1332230189d9SNélio Laranjeiro 	} else {
1333230189d9SNélio Laranjeiro 		txq->elts_comp = comp;
1334230189d9SNélio Laranjeiro 	}
1335230189d9SNélio Laranjeiro #ifdef MLX5_PMD_SOFT_COUNTERS
1336230189d9SNélio Laranjeiro 	/* Increment sent packets counter. */
1337230189d9SNélio Laranjeiro 	txq->stats.opackets += i;
1338230189d9SNélio Laranjeiro #endif
1339230189d9SNélio Laranjeiro 	/* Ring QP doorbell. */
1340230189d9SNélio Laranjeiro 	if (mpw.state == MLX5_MPW_INL_STATE_OPENED)
1341230189d9SNélio Laranjeiro 		mlx5_mpw_inline_close(txq, &mpw);
1342230189d9SNélio Laranjeiro 	else if (mpw.state == MLX5_MPW_STATE_OPENED)
1343230189d9SNélio Laranjeiro 		mlx5_mpw_close(txq, &mpw);
134430807f62SNélio Laranjeiro 	mlx5_tx_dbrec(txq, mpw.wqe);
1345230189d9SNélio Laranjeiro 	txq->elts_head = elts_head;
1346230189d9SNélio Laranjeiro 	return i;
1347230189d9SNélio Laranjeiro }
1348230189d9SNélio Laranjeiro 
1349230189d9SNélio Laranjeiro /**
13506ce84bd8SYongseok Koh  * Open an Enhanced MPW session.
13516ce84bd8SYongseok Koh  *
13526ce84bd8SYongseok Koh  * @param txq
13536ce84bd8SYongseok Koh  *   Pointer to TX queue structure.
13546ce84bd8SYongseok Koh  * @param mpw
13556ce84bd8SYongseok Koh  *   Pointer to MPW session structure.
13566ce84bd8SYongseok Koh  * @param length
13576ce84bd8SYongseok Koh  *   Packet length.
13586ce84bd8SYongseok Koh  */
13596ce84bd8SYongseok Koh static inline void
1360991b04f6SNélio Laranjeiro mlx5_empw_new(struct mlx5_txq_data *txq, struct mlx5_mpw *mpw, int padding)
13616ce84bd8SYongseok Koh {
13626ce84bd8SYongseok Koh 	uint16_t idx = txq->wqe_ci & ((1 << txq->wqe_n) - 1);
13636ce84bd8SYongseok Koh 
13646ce84bd8SYongseok Koh 	mpw->state = MLX5_MPW_ENHANCED_STATE_OPENED;
13656ce84bd8SYongseok Koh 	mpw->pkts_n = 0;
13666ce84bd8SYongseok Koh 	mpw->total_len = sizeof(struct mlx5_wqe);
13676ce84bd8SYongseok Koh 	mpw->wqe = (volatile struct mlx5_wqe *)tx_mlx5_wqe(txq, idx);
13686b30a6a8SShachar Beiser 	mpw->wqe->ctrl[0] =
13696b30a6a8SShachar Beiser 		rte_cpu_to_be_32((MLX5_OPC_MOD_ENHANCED_MPSW << 24) |
13706ce84bd8SYongseok Koh 				 (txq->wqe_ci << 8) |
13716ce84bd8SYongseok Koh 				 MLX5_OPCODE_ENHANCED_MPSW);
13726ce84bd8SYongseok Koh 	mpw->wqe->ctrl[2] = 0;
13736ce84bd8SYongseok Koh 	mpw->wqe->ctrl[3] = 0;
13746ce84bd8SYongseok Koh 	memset((void *)(uintptr_t)&mpw->wqe->eseg, 0, MLX5_WQE_DWORD_SIZE);
13756ce84bd8SYongseok Koh 	if (unlikely(padding)) {
13766ce84bd8SYongseok Koh 		uintptr_t addr = (uintptr_t)(mpw->wqe + 1);
13776ce84bd8SYongseok Koh 
13786ce84bd8SYongseok Koh 		/* Pad the first 2 DWORDs with zero-length inline header. */
13796b30a6a8SShachar Beiser 		*(volatile uint32_t *)addr = rte_cpu_to_be_32(MLX5_INLINE_SEG);
13806ce84bd8SYongseok Koh 		*(volatile uint32_t *)(addr + MLX5_WQE_DWORD_SIZE) =
13816b30a6a8SShachar Beiser 			rte_cpu_to_be_32(MLX5_INLINE_SEG);
13826ce84bd8SYongseok Koh 		mpw->total_len += 2 * MLX5_WQE_DWORD_SIZE;
13836ce84bd8SYongseok Koh 		/* Start from the next WQEBB. */
13846ce84bd8SYongseok Koh 		mpw->data.raw = (volatile void *)(tx_mlx5_wqe(txq, idx + 1));
13856ce84bd8SYongseok Koh 	} else {
13866ce84bd8SYongseok Koh 		mpw->data.raw = (volatile void *)(mpw->wqe + 1);
13876ce84bd8SYongseok Koh 	}
13886ce84bd8SYongseok Koh }
13896ce84bd8SYongseok Koh 
13906ce84bd8SYongseok Koh /**
13916ce84bd8SYongseok Koh  * Close an Enhanced MPW session.
13926ce84bd8SYongseok Koh  *
13936ce84bd8SYongseok Koh  * @param txq
13946ce84bd8SYongseok Koh  *   Pointer to TX queue structure.
13956ce84bd8SYongseok Koh  * @param mpw
13966ce84bd8SYongseok Koh  *   Pointer to MPW session structure.
13976ce84bd8SYongseok Koh  *
13986ce84bd8SYongseok Koh  * @return
13996ce84bd8SYongseok Koh  *   Number of consumed WQEs.
14006ce84bd8SYongseok Koh  */
14016ce84bd8SYongseok Koh static inline uint16_t
1402991b04f6SNélio Laranjeiro mlx5_empw_close(struct mlx5_txq_data *txq, struct mlx5_mpw *mpw)
14036ce84bd8SYongseok Koh {
14046ce84bd8SYongseok Koh 	uint16_t ret;
14056ce84bd8SYongseok Koh 
14066ce84bd8SYongseok Koh 	/* Store size in multiple of 16 bytes. Control and Ethernet segments
14076ce84bd8SYongseok Koh 	 * count as 2.
14086ce84bd8SYongseok Koh 	 */
14096b30a6a8SShachar Beiser 	mpw->wqe->ctrl[1] = rte_cpu_to_be_32(txq->qp_num_8s |
14106b30a6a8SShachar Beiser 					     MLX5_WQE_DS(mpw->total_len));
14116ce84bd8SYongseok Koh 	mpw->state = MLX5_MPW_STATE_CLOSED;
14126ce84bd8SYongseok Koh 	ret = (mpw->total_len + (MLX5_WQE_SIZE - 1)) / MLX5_WQE_SIZE;
14136ce84bd8SYongseok Koh 	txq->wqe_ci += ret;
14146ce84bd8SYongseok Koh 	return ret;
14156ce84bd8SYongseok Koh }
14166ce84bd8SYongseok Koh 
14176ce84bd8SYongseok Koh /**
14184b0d7b7fSYongseok Koh  * TX with Enhanced MPW support.
14196ce84bd8SYongseok Koh  *
14204b0d7b7fSYongseok Koh  * @param txq
14214b0d7b7fSYongseok Koh  *   Pointer to TX queue structure.
14226ce84bd8SYongseok Koh  * @param[in] pkts
14236ce84bd8SYongseok Koh  *   Packets to transmit.
14246ce84bd8SYongseok Koh  * @param pkts_n
14256ce84bd8SYongseok Koh  *   Number of packets in array.
14266ce84bd8SYongseok Koh  *
14276ce84bd8SYongseok Koh  * @return
14286ce84bd8SYongseok Koh  *   Number of packets successfully transmitted (<= pkts_n).
14296ce84bd8SYongseok Koh  */
14304b0d7b7fSYongseok Koh static inline uint16_t
14314b0d7b7fSYongseok Koh txq_burst_empw(struct mlx5_txq_data *txq, struct rte_mbuf **pkts,
14324b0d7b7fSYongseok Koh 	       uint16_t pkts_n)
14336ce84bd8SYongseok Koh {
14346ce84bd8SYongseok Koh 	uint16_t elts_head = txq->elts_head;
14358c819a69SYongseok Koh 	const uint16_t elts_n = 1 << txq->elts_n;
14368c819a69SYongseok Koh 	const uint16_t elts_m = elts_n - 1;
14376ce84bd8SYongseok Koh 	unsigned int i = 0;
14386ce84bd8SYongseok Koh 	unsigned int j = 0;
14398c819a69SYongseok Koh 	uint16_t max_elts;
14406ce84bd8SYongseok Koh 	uint16_t max_wqe;
14416ce84bd8SYongseok Koh 	unsigned int max_inline = txq->max_inline * RTE_CACHE_LINE_SIZE;
14426ce84bd8SYongseok Koh 	unsigned int mpw_room = 0;
14436ce84bd8SYongseok Koh 	unsigned int inl_pad = 0;
14446ce84bd8SYongseok Koh 	uint32_t inl_hdr;
14456ce84bd8SYongseok Koh 	struct mlx5_mpw mpw = {
14466ce84bd8SYongseok Koh 		.state = MLX5_MPW_STATE_CLOSED,
14476ce84bd8SYongseok Koh 	};
14486ce84bd8SYongseok Koh 
14496ce84bd8SYongseok Koh 	if (unlikely(!pkts_n))
14506ce84bd8SYongseok Koh 		return 0;
14516ce84bd8SYongseok Koh 	/* Start processing. */
14526cb559d6SYongseok Koh 	mlx5_tx_complete(txq);
14536ce84bd8SYongseok Koh 	max_elts = (elts_n - (elts_head - txq->elts_tail));
14546ce84bd8SYongseok Koh 	/* A CQE slot must always be available. */
14556ce84bd8SYongseok Koh 	assert((1u << txq->cqe_n) - (txq->cq_pi - txq->cq_ci));
14566ce84bd8SYongseok Koh 	max_wqe = (1u << txq->wqe_n) - (txq->wqe_ci - txq->wqe_pi);
14576ce84bd8SYongseok Koh 	if (unlikely(!max_wqe))
14586ce84bd8SYongseok Koh 		return 0;
14596ce84bd8SYongseok Koh 	do {
14606ce84bd8SYongseok Koh 		struct rte_mbuf *buf = *(pkts++);
14616ce84bd8SYongseok Koh 		uintptr_t addr;
14626ce84bd8SYongseok Koh 		unsigned int do_inline = 0; /* Whether inline is possible. */
14636ce84bd8SYongseok Koh 		uint32_t length;
14644aa15eb1SNélio Laranjeiro 		uint8_t cs_flags;
14656ce84bd8SYongseok Koh 
146648642ec5SYongseok Koh 		/* Multi-segmented packet is handled in slow-path outside. */
146748642ec5SYongseok Koh 		assert(NB_SEGS(buf) == 1);
146848642ec5SYongseok Koh 		/* Make sure there is enough room to store this packet. */
146948642ec5SYongseok Koh 		if (max_elts - j == 0)
14706ce84bd8SYongseok Koh 			break;
14715f8ba81cSXueming Li 		cs_flags = txq_ol_cksum_to_cs(buf);
14726ce84bd8SYongseok Koh 		/* Retrieve packet information. */
14736ce84bd8SYongseok Koh 		length = PKT_LEN(buf);
14746ce84bd8SYongseok Koh 		/* Start new session if:
14756ce84bd8SYongseok Koh 		 * - multi-segment packet
14766ce84bd8SYongseok Koh 		 * - no space left even for a dseg
14776ce84bd8SYongseok Koh 		 * - next packet can be inlined with a new WQE
14786ce84bd8SYongseok Koh 		 * - cs_flag differs
14796ce84bd8SYongseok Koh 		 */
14806ce84bd8SYongseok Koh 		if (mpw.state == MLX5_MPW_ENHANCED_STATE_OPENED) {
148148642ec5SYongseok Koh 			if ((inl_pad + sizeof(struct mlx5_wqe_data_seg) >
14826ce84bd8SYongseok Koh 			     mpw_room) ||
14836ce84bd8SYongseok Koh 			    (length <= txq->inline_max_packet_sz &&
14846ce84bd8SYongseok Koh 			     inl_pad + sizeof(inl_hdr) + length >
14856ce84bd8SYongseok Koh 			     mpw_room) ||
14866ce84bd8SYongseok Koh 			    (mpw.wqe->eseg.cs_flags != cs_flags))
14876ce84bd8SYongseok Koh 				max_wqe -= mlx5_empw_close(txq, &mpw);
14886ce84bd8SYongseok Koh 		}
14896ce84bd8SYongseok Koh 		if (unlikely(mpw.state == MLX5_MPW_STATE_CLOSED)) {
149048642ec5SYongseok Koh 			/* In Enhanced MPW, inline as much as the budget is
149148642ec5SYongseok Koh 			 * allowed. The remaining space is to be filled with
149248642ec5SYongseok Koh 			 * dsegs. If the title WQEBB isn't padded, it will have
149348642ec5SYongseok Koh 			 * 2 dsegs there.
14946ce84bd8SYongseok Koh 			 */
14956ce84bd8SYongseok Koh 			mpw_room = RTE_MIN(MLX5_WQE_SIZE_MAX,
14966ce84bd8SYongseok Koh 					   (max_inline ? max_inline :
14976ce84bd8SYongseok Koh 					    pkts_n * MLX5_WQE_DWORD_SIZE) +
14986ce84bd8SYongseok Koh 					   MLX5_WQE_SIZE);
149948642ec5SYongseok Koh 			if (unlikely(max_wqe * MLX5_WQE_SIZE < mpw_room))
15006ce84bd8SYongseok Koh 				break;
15016ce84bd8SYongseok Koh 			/* Don't pad the title WQEBB to not waste WQ. */
15026ce84bd8SYongseok Koh 			mlx5_empw_new(txq, &mpw, 0);
15036ce84bd8SYongseok Koh 			mpw_room -= mpw.total_len;
15046ce84bd8SYongseok Koh 			inl_pad = 0;
150548642ec5SYongseok Koh 			do_inline = length <= txq->inline_max_packet_sz &&
15066ce84bd8SYongseok Koh 				    sizeof(inl_hdr) + length <= mpw_room &&
15076ce84bd8SYongseok Koh 				    !txq->mpw_hdr_dseg;
15086ce84bd8SYongseok Koh 			mpw.wqe->eseg.cs_flags = cs_flags;
15096ce84bd8SYongseok Koh 		} else {
15106ce84bd8SYongseok Koh 			/* Evaluate whether the next packet can be inlined.
15116ce84bd8SYongseok Koh 			 * Inlininig is possible when:
15126ce84bd8SYongseok Koh 			 * - length is less than configured value
15136ce84bd8SYongseok Koh 			 * - length fits for remaining space
15146ce84bd8SYongseok Koh 			 * - not required to fill the title WQEBB with dsegs
15156ce84bd8SYongseok Koh 			 */
15166ce84bd8SYongseok Koh 			do_inline =
15176ce84bd8SYongseok Koh 				length <= txq->inline_max_packet_sz &&
15186ce84bd8SYongseok Koh 				inl_pad + sizeof(inl_hdr) + length <=
15196ce84bd8SYongseok Koh 				 mpw_room &&
15206ce84bd8SYongseok Koh 				(!txq->mpw_hdr_dseg ||
15216ce84bd8SYongseok Koh 				 mpw.total_len >= MLX5_WQE_SIZE);
15226ce84bd8SYongseok Koh 		}
152324a8f524SYongseok Koh 		if (max_inline && do_inline) {
15246ce84bd8SYongseok Koh 			/* Inline packet into WQE. */
15256ce84bd8SYongseok Koh 			unsigned int max;
15266ce84bd8SYongseok Koh 
15276ce84bd8SYongseok Koh 			assert(mpw.state == MLX5_MPW_ENHANCED_STATE_OPENED);
15286ce84bd8SYongseok Koh 			assert(length == DATA_LEN(buf));
15296b30a6a8SShachar Beiser 			inl_hdr = rte_cpu_to_be_32(length | MLX5_INLINE_SEG);
15306ce84bd8SYongseok Koh 			addr = rte_pktmbuf_mtod(buf, uintptr_t);
15316ce84bd8SYongseok Koh 			mpw.data.raw = (volatile void *)
15326ce84bd8SYongseok Koh 				((uintptr_t)mpw.data.raw + inl_pad);
15336ce84bd8SYongseok Koh 			max = tx_mlx5_wq_tailroom(txq,
15346ce84bd8SYongseok Koh 					(void *)(uintptr_t)mpw.data.raw);
15356ce84bd8SYongseok Koh 			/* Copy inline header. */
15366ce84bd8SYongseok Koh 			mpw.data.raw = (volatile void *)
15376ce84bd8SYongseok Koh 				mlx5_copy_to_wq(
15386ce84bd8SYongseok Koh 					  (void *)(uintptr_t)mpw.data.raw,
15396ce84bd8SYongseok Koh 					  &inl_hdr,
15406ce84bd8SYongseok Koh 					  sizeof(inl_hdr),
15416ce84bd8SYongseok Koh 					  (void *)(uintptr_t)txq->wqes,
15426ce84bd8SYongseok Koh 					  max);
15436ce84bd8SYongseok Koh 			max = tx_mlx5_wq_tailroom(txq,
15446ce84bd8SYongseok Koh 					(void *)(uintptr_t)mpw.data.raw);
15456ce84bd8SYongseok Koh 			/* Copy packet data. */
15466ce84bd8SYongseok Koh 			mpw.data.raw = (volatile void *)
15476ce84bd8SYongseok Koh 				mlx5_copy_to_wq(
15486ce84bd8SYongseok Koh 					  (void *)(uintptr_t)mpw.data.raw,
15496ce84bd8SYongseok Koh 					  (void *)addr,
15506ce84bd8SYongseok Koh 					  length,
15516ce84bd8SYongseok Koh 					  (void *)(uintptr_t)txq->wqes,
15526ce84bd8SYongseok Koh 					  max);
15536ce84bd8SYongseok Koh 			++mpw.pkts_n;
15546ce84bd8SYongseok Koh 			mpw.total_len += (inl_pad + sizeof(inl_hdr) + length);
15556ce84bd8SYongseok Koh 			/* No need to get completion as the entire packet is
15566ce84bd8SYongseok Koh 			 * copied to WQ. Free the buf right away.
15576ce84bd8SYongseok Koh 			 */
15586ce84bd8SYongseok Koh 			rte_pktmbuf_free_seg(buf);
15596ce84bd8SYongseok Koh 			mpw_room -= (inl_pad + sizeof(inl_hdr) + length);
15606ce84bd8SYongseok Koh 			/* Add pad in the next packet if any. */
15616ce84bd8SYongseok Koh 			inl_pad = (((uintptr_t)mpw.data.raw +
15626ce84bd8SYongseok Koh 					(MLX5_WQE_DWORD_SIZE - 1)) &
15636ce84bd8SYongseok Koh 					~(MLX5_WQE_DWORD_SIZE - 1)) -
15646ce84bd8SYongseok Koh 				  (uintptr_t)mpw.data.raw;
15656ce84bd8SYongseok Koh 		} else {
15666ce84bd8SYongseok Koh 			/* No inline. Load a dseg of packet pointer. */
15676ce84bd8SYongseok Koh 			volatile rte_v128u32_t *dseg;
15686ce84bd8SYongseok Koh 
15696ce84bd8SYongseok Koh 			assert(mpw.state == MLX5_MPW_ENHANCED_STATE_OPENED);
15706ce84bd8SYongseok Koh 			assert((inl_pad + sizeof(*dseg)) <= mpw_room);
15716ce84bd8SYongseok Koh 			assert(length == DATA_LEN(buf));
15726ce84bd8SYongseok Koh 			if (!tx_mlx5_wq_tailroom(txq,
15736ce84bd8SYongseok Koh 					(void *)((uintptr_t)mpw.data.raw
15746ce84bd8SYongseok Koh 						+ inl_pad)))
15756ce84bd8SYongseok Koh 				dseg = (volatile void *)txq->wqes;
15766ce84bd8SYongseok Koh 			else
15776ce84bd8SYongseok Koh 				dseg = (volatile void *)
15786ce84bd8SYongseok Koh 					((uintptr_t)mpw.data.raw +
15796ce84bd8SYongseok Koh 					 inl_pad);
15808c819a69SYongseok Koh 			(*txq->elts)[elts_head++ & elts_m] = buf;
1581f84411beSYongseok Koh 			addr = rte_cpu_to_be_64(rte_pktmbuf_mtod(buf,
1582f84411beSYongseok Koh 								 uintptr_t));
15836ce84bd8SYongseok Koh 			*dseg = (rte_v128u32_t) {
15846b30a6a8SShachar Beiser 				rte_cpu_to_be_32(length),
15856cb559d6SYongseok Koh 				mlx5_tx_mb2mr(txq, buf),
1586ebbb81ebSNélio Laranjeiro 				addr,
1587ebbb81ebSNélio Laranjeiro 				addr >> 32,
15886ce84bd8SYongseok Koh 			};
15896ce84bd8SYongseok Koh 			mpw.data.raw = (volatile void *)(dseg + 1);
15906ce84bd8SYongseok Koh 			mpw.total_len += (inl_pad + sizeof(*dseg));
15916ce84bd8SYongseok Koh 			++j;
15926ce84bd8SYongseok Koh 			++mpw.pkts_n;
15936ce84bd8SYongseok Koh 			mpw_room -= (inl_pad + sizeof(*dseg));
15946ce84bd8SYongseok Koh 			inl_pad = 0;
15956ce84bd8SYongseok Koh 		}
15966ce84bd8SYongseok Koh #ifdef MLX5_PMD_SOFT_COUNTERS
15976ce84bd8SYongseok Koh 		/* Increment sent bytes counter. */
15986ce84bd8SYongseok Koh 		txq->stats.obytes += length;
15996ce84bd8SYongseok Koh #endif
16006ce84bd8SYongseok Koh 		++i;
16016ce84bd8SYongseok Koh 	} while (i < pkts_n);
16026ce84bd8SYongseok Koh 	/* Take a shortcut if nothing must be sent. */
16036ce84bd8SYongseok Koh 	if (unlikely(i == 0))
16046ce84bd8SYongseok Koh 		return 0;
16056ce84bd8SYongseok Koh 	/* Check whether completion threshold has been reached. */
16066ce84bd8SYongseok Koh 	if (txq->elts_comp + j >= MLX5_TX_COMP_THRESH ||
16076ce84bd8SYongseok Koh 			(uint16_t)(txq->wqe_ci - txq->mpw_comp) >=
16086ce84bd8SYongseok Koh 			 (1 << txq->wqe_n) / MLX5_TX_COMP_THRESH_INLINE_DIV) {
16096ce84bd8SYongseok Koh 		volatile struct mlx5_wqe *wqe = mpw.wqe;
16106ce84bd8SYongseok Koh 
16116ce84bd8SYongseok Koh 		/* Request completion on last WQE. */
16126b30a6a8SShachar Beiser 		wqe->ctrl[2] = rte_cpu_to_be_32(8);
16136ce84bd8SYongseok Koh 		/* Save elts_head in unused "immediate" field of WQE. */
16146ce84bd8SYongseok Koh 		wqe->ctrl[3] = elts_head;
16156ce84bd8SYongseok Koh 		txq->elts_comp = 0;
16166ce84bd8SYongseok Koh 		txq->mpw_comp = txq->wqe_ci;
16172eefbec5SYongseok Koh #ifndef NDEBUG
16182eefbec5SYongseok Koh 		++txq->cq_pi;
16192eefbec5SYongseok Koh #endif
16206ce84bd8SYongseok Koh 	} else {
16216ce84bd8SYongseok Koh 		txq->elts_comp += j;
16226ce84bd8SYongseok Koh 	}
16236ce84bd8SYongseok Koh #ifdef MLX5_PMD_SOFT_COUNTERS
16246ce84bd8SYongseok Koh 	/* Increment sent packets counter. */
16256ce84bd8SYongseok Koh 	txq->stats.opackets += i;
16266ce84bd8SYongseok Koh #endif
16276ce84bd8SYongseok Koh 	if (mpw.state == MLX5_MPW_ENHANCED_STATE_OPENED)
16286ce84bd8SYongseok Koh 		mlx5_empw_close(txq, &mpw);
16296ce84bd8SYongseok Koh 	/* Ring QP doorbell. */
16306ce84bd8SYongseok Koh 	mlx5_tx_dbrec(txq, mpw.wqe);
16316ce84bd8SYongseok Koh 	txq->elts_head = elts_head;
16326ce84bd8SYongseok Koh 	return i;
16336ce84bd8SYongseok Koh }
16346ce84bd8SYongseok Koh 
16356ce84bd8SYongseok Koh /**
16364b0d7b7fSYongseok Koh  * DPDK callback for TX with Enhanced MPW support.
16374b0d7b7fSYongseok Koh  *
16384b0d7b7fSYongseok Koh  * @param dpdk_txq
16394b0d7b7fSYongseok Koh  *   Generic pointer to TX queue structure.
16404b0d7b7fSYongseok Koh  * @param[in] pkts
16414b0d7b7fSYongseok Koh  *   Packets to transmit.
16424b0d7b7fSYongseok Koh  * @param pkts_n
16434b0d7b7fSYongseok Koh  *   Number of packets in array.
16444b0d7b7fSYongseok Koh  *
16454b0d7b7fSYongseok Koh  * @return
16464b0d7b7fSYongseok Koh  *   Number of packets successfully transmitted (<= pkts_n).
16474b0d7b7fSYongseok Koh  */
16484b0d7b7fSYongseok Koh uint16_t
16494b0d7b7fSYongseok Koh mlx5_tx_burst_empw(void *dpdk_txq, struct rte_mbuf **pkts, uint16_t pkts_n)
16504b0d7b7fSYongseok Koh {
16514b0d7b7fSYongseok Koh 	struct mlx5_txq_data *txq = (struct mlx5_txq_data *)dpdk_txq;
16524b0d7b7fSYongseok Koh 	uint16_t nb_tx = 0;
16534b0d7b7fSYongseok Koh 
16544b0d7b7fSYongseok Koh 	while (pkts_n > nb_tx) {
16554b0d7b7fSYongseok Koh 		uint16_t n;
16564b0d7b7fSYongseok Koh 		uint16_t ret;
16574b0d7b7fSYongseok Koh 
16584b0d7b7fSYongseok Koh 		n = txq_count_contig_multi_seg(&pkts[nb_tx], pkts_n - nb_tx);
16594b0d7b7fSYongseok Koh 		if (n) {
16604b0d7b7fSYongseok Koh 			ret = mlx5_tx_burst(dpdk_txq, &pkts[nb_tx], n);
16614b0d7b7fSYongseok Koh 			if (!ret)
16624b0d7b7fSYongseok Koh 				break;
16634b0d7b7fSYongseok Koh 			nb_tx += ret;
16644b0d7b7fSYongseok Koh 		}
16654b0d7b7fSYongseok Koh 		n = txq_count_contig_single_seg(&pkts[nb_tx], pkts_n - nb_tx);
16664b0d7b7fSYongseok Koh 		if (n) {
16674b0d7b7fSYongseok Koh 			ret = txq_burst_empw(txq, &pkts[nb_tx], n);
16684b0d7b7fSYongseok Koh 			if (!ret)
16694b0d7b7fSYongseok Koh 				break;
16704b0d7b7fSYongseok Koh 			nb_tx += ret;
16714b0d7b7fSYongseok Koh 		}
16724b0d7b7fSYongseok Koh 	}
16734b0d7b7fSYongseok Koh 	return nb_tx;
16744b0d7b7fSYongseok Koh }
16754b0d7b7fSYongseok Koh 
16764b0d7b7fSYongseok Koh /**
167767fa62bcSAdrien Mazarguil  * Translate RX completion flags to packet type.
167867fa62bcSAdrien Mazarguil  *
16793cc08bc6SXueming Li  * @param[in] rxq
16803cc08bc6SXueming Li  *   Pointer to RX queue structure.
16816218063bSNélio Laranjeiro  * @param[in] cqe
16826218063bSNélio Laranjeiro  *   Pointer to CQE.
168367fa62bcSAdrien Mazarguil  *
168478a38edfSJianfeng Tan  * @note: fix mlx5_dev_supported_ptypes_get() if any change here.
168578a38edfSJianfeng Tan  *
168667fa62bcSAdrien Mazarguil  * @return
168767fa62bcSAdrien Mazarguil  *   Packet type for struct rte_mbuf.
168867fa62bcSAdrien Mazarguil  */
168967fa62bcSAdrien Mazarguil static inline uint32_t
16903cc08bc6SXueming Li rxq_cq_to_pkt_type(struct mlx5_rxq_data *rxq, volatile struct mlx5_cqe *cqe)
169167fa62bcSAdrien Mazarguil {
1692ea16068cSYongseok Koh 	uint8_t idx;
1693ea16068cSYongseok Koh 	uint8_t pinfo = cqe->pkt_info;
1694ea16068cSYongseok Koh 	uint16_t ptype = cqe->hdr_type_etc;
169567fa62bcSAdrien Mazarguil 
1696ea16068cSYongseok Koh 	/*
1697ea16068cSYongseok Koh 	 * The index to the array should have:
1698ea16068cSYongseok Koh 	 * bit[1:0] = l3_hdr_type
1699ea16068cSYongseok Koh 	 * bit[4:2] = l4_hdr_type
1700ea16068cSYongseok Koh 	 * bit[5] = ip_frag
1701ea16068cSYongseok Koh 	 * bit[6] = tunneled
1702ea16068cSYongseok Koh 	 * bit[7] = outer_l3_type
1703ea16068cSYongseok Koh 	 */
1704ea16068cSYongseok Koh 	idx = ((pinfo & 0x3) << 6) | ((ptype & 0xfc00) >> 10);
17053cc08bc6SXueming Li 	return mlx5_ptype_table[idx] | rxq->tunnel * !!(idx & (1 << 6));
170667fa62bcSAdrien Mazarguil }
170767fa62bcSAdrien Mazarguil 
170867fa62bcSAdrien Mazarguil /**
170999c12dccSNélio Laranjeiro  * Get size of the next packet for a given CQE. For compressed CQEs, the
171099c12dccSNélio Laranjeiro  * consumer index is updated only once all packets of the current one have
171199c12dccSNélio Laranjeiro  * been processed.
171299c12dccSNélio Laranjeiro  *
171399c12dccSNélio Laranjeiro  * @param rxq
171499c12dccSNélio Laranjeiro  *   Pointer to RX queue.
171599c12dccSNélio Laranjeiro  * @param cqe
171699c12dccSNélio Laranjeiro  *   CQE to process.
1717ecf60761SNélio Laranjeiro  * @param[out] rss_hash
1718ecf60761SNélio Laranjeiro  *   Packet RSS Hash result.
171999c12dccSNélio Laranjeiro  *
172099c12dccSNélio Laranjeiro  * @return
172199c12dccSNélio Laranjeiro  *   Packet size in bytes (0 if there is none), -1 in case of completion
172299c12dccSNélio Laranjeiro  *   with error.
172399c12dccSNélio Laranjeiro  */
172499c12dccSNélio Laranjeiro static inline int
172578142aacSNélio Laranjeiro mlx5_rx_poll_len(struct mlx5_rxq_data *rxq, volatile struct mlx5_cqe *cqe,
1726ecf60761SNélio Laranjeiro 		 uint16_t cqe_cnt, uint32_t *rss_hash)
172799c12dccSNélio Laranjeiro {
172899c12dccSNélio Laranjeiro 	struct rxq_zip *zip = &rxq->zip;
172999c12dccSNélio Laranjeiro 	uint16_t cqe_n = cqe_cnt + 1;
173099c12dccSNélio Laranjeiro 	int len = 0;
1731d2e842d0SYongseok Koh 	uint16_t idx, end;
173299c12dccSNélio Laranjeiro 
173399c12dccSNélio Laranjeiro 	/* Process compressed data in the CQE and mini arrays. */
173499c12dccSNélio Laranjeiro 	if (zip->ai) {
173599c12dccSNélio Laranjeiro 		volatile struct mlx5_mini_cqe8 (*mc)[8] =
173699c12dccSNélio Laranjeiro 			(volatile struct mlx5_mini_cqe8 (*)[8])
17374aff4bcbSYongseok Koh 			(uintptr_t)(&(*rxq->cqes)[zip->ca & cqe_cnt].pkt_info);
173899c12dccSNélio Laranjeiro 
17396b30a6a8SShachar Beiser 		len = rte_be_to_cpu_32((*mc)[zip->ai & 7].byte_cnt);
17406b30a6a8SShachar Beiser 		*rss_hash = rte_be_to_cpu_32((*mc)[zip->ai & 7].rx_hash_result);
174199c12dccSNélio Laranjeiro 		if ((++zip->ai & 7) == 0) {
1742d2e842d0SYongseok Koh 			/* Invalidate consumed CQEs */
1743d2e842d0SYongseok Koh 			idx = zip->ca;
1744d2e842d0SYongseok Koh 			end = zip->na;
1745d2e842d0SYongseok Koh 			while (idx != end) {
1746d2e842d0SYongseok Koh 				(*rxq->cqes)[idx & cqe_cnt].op_own =
1747d2e842d0SYongseok Koh 					MLX5_CQE_INVALIDATE;
1748d2e842d0SYongseok Koh 				++idx;
1749d2e842d0SYongseok Koh 			}
175099c12dccSNélio Laranjeiro 			/*
175199c12dccSNélio Laranjeiro 			 * Increment consumer index to skip the number of
175299c12dccSNélio Laranjeiro 			 * CQEs consumed. Hardware leaves holes in the CQ
175399c12dccSNélio Laranjeiro 			 * ring for software use.
175499c12dccSNélio Laranjeiro 			 */
175599c12dccSNélio Laranjeiro 			zip->ca = zip->na;
175699c12dccSNélio Laranjeiro 			zip->na += 8;
175799c12dccSNélio Laranjeiro 		}
175899c12dccSNélio Laranjeiro 		if (unlikely(rxq->zip.ai == rxq->zip.cqe_cnt)) {
1759d2e842d0SYongseok Koh 			/* Invalidate the rest */
1760d2e842d0SYongseok Koh 			idx = zip->ca;
1761d2e842d0SYongseok Koh 			end = zip->cq_ci;
176299c12dccSNélio Laranjeiro 
176399c12dccSNélio Laranjeiro 			while (idx != end) {
176497267b8eSNelio Laranjeiro 				(*rxq->cqes)[idx & cqe_cnt].op_own =
176599c12dccSNélio Laranjeiro 					MLX5_CQE_INVALIDATE;
176699c12dccSNélio Laranjeiro 				++idx;
176799c12dccSNélio Laranjeiro 			}
176899c12dccSNélio Laranjeiro 			rxq->cq_ci = zip->cq_ci;
176999c12dccSNélio Laranjeiro 			zip->ai = 0;
177099c12dccSNélio Laranjeiro 		}
177199c12dccSNélio Laranjeiro 	/* No compressed data, get next CQE and verify if it is compressed. */
177299c12dccSNélio Laranjeiro 	} else {
177399c12dccSNélio Laranjeiro 		int ret;
177499c12dccSNélio Laranjeiro 		int8_t op_own;
177599c12dccSNélio Laranjeiro 
177697267b8eSNelio Laranjeiro 		ret = check_cqe(cqe, cqe_n, rxq->cq_ci);
177799c12dccSNélio Laranjeiro 		if (unlikely(ret == 1))
177899c12dccSNélio Laranjeiro 			return 0;
177999c12dccSNélio Laranjeiro 		++rxq->cq_ci;
178099c12dccSNélio Laranjeiro 		op_own = cqe->op_own;
17811742c2d9SYongseok Koh 		rte_cio_rmb();
178299c12dccSNélio Laranjeiro 		if (MLX5_CQE_FORMAT(op_own) == MLX5_COMPRESSED) {
178399c12dccSNélio Laranjeiro 			volatile struct mlx5_mini_cqe8 (*mc)[8] =
178499c12dccSNélio Laranjeiro 				(volatile struct mlx5_mini_cqe8 (*)[8])
178599c12dccSNélio Laranjeiro 				(uintptr_t)(&(*rxq->cqes)[rxq->cq_ci &
17864aff4bcbSYongseok Koh 							  cqe_cnt].pkt_info);
178799c12dccSNélio Laranjeiro 
178899c12dccSNélio Laranjeiro 			/* Fix endianness. */
17896b30a6a8SShachar Beiser 			zip->cqe_cnt = rte_be_to_cpu_32(cqe->byte_cnt);
179099c12dccSNélio Laranjeiro 			/*
179199c12dccSNélio Laranjeiro 			 * Current mini array position is the one returned by
179299c12dccSNélio Laranjeiro 			 * check_cqe64().
179399c12dccSNélio Laranjeiro 			 *
179499c12dccSNélio Laranjeiro 			 * If completion comprises several mini arrays, as a
179599c12dccSNélio Laranjeiro 			 * special case the second one is located 7 CQEs after
179699c12dccSNélio Laranjeiro 			 * the initial CQE instead of 8 for subsequent ones.
179799c12dccSNélio Laranjeiro 			 */
1798d2e842d0SYongseok Koh 			zip->ca = rxq->cq_ci;
179999c12dccSNélio Laranjeiro 			zip->na = zip->ca + 7;
180099c12dccSNélio Laranjeiro 			/* Compute the next non compressed CQE. */
180199c12dccSNélio Laranjeiro 			--rxq->cq_ci;
180299c12dccSNélio Laranjeiro 			zip->cq_ci = rxq->cq_ci + zip->cqe_cnt;
180399c12dccSNélio Laranjeiro 			/* Get packet size to return. */
18046b30a6a8SShachar Beiser 			len = rte_be_to_cpu_32((*mc)[0].byte_cnt);
18056b30a6a8SShachar Beiser 			*rss_hash = rte_be_to_cpu_32((*mc)[0].rx_hash_result);
180699c12dccSNélio Laranjeiro 			zip->ai = 1;
1807d2e842d0SYongseok Koh 			/* Prefetch all the entries to be invalidated */
1808d2e842d0SYongseok Koh 			idx = zip->ca;
1809d2e842d0SYongseok Koh 			end = zip->cq_ci;
1810d2e842d0SYongseok Koh 			while (idx != end) {
1811d2e842d0SYongseok Koh 				rte_prefetch0(&(*rxq->cqes)[(idx) & cqe_cnt]);
1812d2e842d0SYongseok Koh 				++idx;
1813d2e842d0SYongseok Koh 			}
181499c12dccSNélio Laranjeiro 		} else {
18156b30a6a8SShachar Beiser 			len = rte_be_to_cpu_32(cqe->byte_cnt);
18166b30a6a8SShachar Beiser 			*rss_hash = rte_be_to_cpu_32(cqe->rx_hash_res);
181799c12dccSNélio Laranjeiro 		}
181899c12dccSNélio Laranjeiro 		/* Error while receiving packet. */
181999c12dccSNélio Laranjeiro 		if (unlikely(MLX5_CQE_OPCODE(op_own) == MLX5_CQE_RESP_ERR))
182099c12dccSNélio Laranjeiro 			return -1;
182199c12dccSNélio Laranjeiro 	}
182299c12dccSNélio Laranjeiro 	return len;
182399c12dccSNélio Laranjeiro }
182499c12dccSNélio Laranjeiro 
182599c12dccSNélio Laranjeiro /**
182667fa62bcSAdrien Mazarguil  * Translate RX completion flags to offload flags.
182767fa62bcSAdrien Mazarguil  *
18286218063bSNélio Laranjeiro  * @param[in] cqe
18296218063bSNélio Laranjeiro  *   Pointer to CQE.
183067fa62bcSAdrien Mazarguil  *
183167fa62bcSAdrien Mazarguil  * @return
183267fa62bcSAdrien Mazarguil  *   Offload flags (ol_flags) for struct rte_mbuf.
183367fa62bcSAdrien Mazarguil  */
183467fa62bcSAdrien Mazarguil static inline uint32_t
18356ba07449SXueming Li rxq_cq_to_ol_flags(volatile struct mlx5_cqe *cqe)
183667fa62bcSAdrien Mazarguil {
183767fa62bcSAdrien Mazarguil 	uint32_t ol_flags = 0;
18386b30a6a8SShachar Beiser 	uint16_t flags = rte_be_to_cpu_16(cqe->hdr_type_etc);
183967fa62bcSAdrien Mazarguil 
18400603df73SNélio Laranjeiro 	ol_flags =
18410603df73SNélio Laranjeiro 		TRANSPOSE(flags,
18420603df73SNélio Laranjeiro 			  MLX5_CQE_RX_L3_HDR_VALID,
18430603df73SNélio Laranjeiro 			  PKT_RX_IP_CKSUM_GOOD) |
18440603df73SNélio Laranjeiro 		TRANSPOSE(flags,
18450603df73SNélio Laranjeiro 			  MLX5_CQE_RX_L4_HDR_VALID,
184683e9d9a3SNelio Laranjeiro 			  PKT_RX_L4_CKSUM_GOOD);
184767fa62bcSAdrien Mazarguil 	return ol_flags;
184867fa62bcSAdrien Mazarguil }
184967fa62bcSAdrien Mazarguil 
185067fa62bcSAdrien Mazarguil /**
18513e1f82a1SYongseok Koh  * Fill in mbuf fields from RX completion flags.
18523e1f82a1SYongseok Koh  * Note that pkt->ol_flags should be initialized outside of this function.
18533e1f82a1SYongseok Koh  *
18543e1f82a1SYongseok Koh  * @param rxq
18553e1f82a1SYongseok Koh  *   Pointer to RX queue.
18563e1f82a1SYongseok Koh  * @param pkt
18573e1f82a1SYongseok Koh  *   mbuf to fill.
18583e1f82a1SYongseok Koh  * @param cqe
18593e1f82a1SYongseok Koh  *   CQE to process.
18603e1f82a1SYongseok Koh  * @param rss_hash_res
18613e1f82a1SYongseok Koh  *   Packet RSS Hash result.
18623e1f82a1SYongseok Koh  */
18633e1f82a1SYongseok Koh static inline void
18643e1f82a1SYongseok Koh rxq_cq_to_mbuf(struct mlx5_rxq_data *rxq, struct rte_mbuf *pkt,
18653e1f82a1SYongseok Koh 	       volatile struct mlx5_cqe *cqe, uint32_t rss_hash_res)
18663e1f82a1SYongseok Koh {
18673e1f82a1SYongseok Koh 	/* Update packet information. */
18683e1f82a1SYongseok Koh 	pkt->packet_type = rxq_cq_to_pkt_type(rxq, cqe);
18693e1f82a1SYongseok Koh 	if (rss_hash_res && rxq->rss_hash) {
18703e1f82a1SYongseok Koh 		pkt->hash.rss = rss_hash_res;
18713e1f82a1SYongseok Koh 		pkt->ol_flags |= PKT_RX_RSS_HASH;
18723e1f82a1SYongseok Koh 	}
18733e1f82a1SYongseok Koh 	if (rxq->mark && MLX5_FLOW_MARK_IS_VALID(cqe->sop_drop_qpn)) {
18743e1f82a1SYongseok Koh 		pkt->ol_flags |= PKT_RX_FDIR;
18753e1f82a1SYongseok Koh 		if (cqe->sop_drop_qpn !=
18763e1f82a1SYongseok Koh 		    rte_cpu_to_be_32(MLX5_FLOW_MARK_DEFAULT)) {
18773e1f82a1SYongseok Koh 			uint32_t mark = cqe->sop_drop_qpn;
18783e1f82a1SYongseok Koh 
18793e1f82a1SYongseok Koh 			pkt->ol_flags |= PKT_RX_FDIR_ID;
18803e1f82a1SYongseok Koh 			pkt->hash.fdir.hi = mlx5_flow_mark_get(mark);
18813e1f82a1SYongseok Koh 		}
18823e1f82a1SYongseok Koh 	}
18833e1f82a1SYongseok Koh 	if (rxq->csum)
18843e1f82a1SYongseok Koh 		pkt->ol_flags |= rxq_cq_to_ol_flags(cqe);
18853e1f82a1SYongseok Koh 	if (rxq->vlan_strip &&
18863e1f82a1SYongseok Koh 	    (cqe->hdr_type_etc & rte_cpu_to_be_16(MLX5_CQE_VLAN_STRIPPED))) {
18873e1f82a1SYongseok Koh 		pkt->ol_flags |= PKT_RX_VLAN | PKT_RX_VLAN_STRIPPED;
18883e1f82a1SYongseok Koh 		pkt->vlan_tci = rte_be_to_cpu_16(cqe->vlan_info);
18893e1f82a1SYongseok Koh 	}
18903e1f82a1SYongseok Koh 	if (rxq->hw_timestamp) {
18913e1f82a1SYongseok Koh 		pkt->timestamp = rte_be_to_cpu_64(cqe->timestamp);
18923e1f82a1SYongseok Koh 		pkt->ol_flags |= PKT_RX_TIMESTAMP;
18933e1f82a1SYongseok Koh 	}
18943e1f82a1SYongseok Koh }
18953e1f82a1SYongseok Koh 
18963e1f82a1SYongseok Koh /**
18972e22920bSAdrien Mazarguil  * DPDK callback for RX.
18982e22920bSAdrien Mazarguil  *
18992e22920bSAdrien Mazarguil  * @param dpdk_rxq
19002e22920bSAdrien Mazarguil  *   Generic pointer to RX queue structure.
19012e22920bSAdrien Mazarguil  * @param[out] pkts
19022e22920bSAdrien Mazarguil  *   Array to store received packets.
19032e22920bSAdrien Mazarguil  * @param pkts_n
19042e22920bSAdrien Mazarguil  *   Maximum number of packets in array.
19052e22920bSAdrien Mazarguil  *
19062e22920bSAdrien Mazarguil  * @return
19072e22920bSAdrien Mazarguil  *   Number of packets successfully received (<= pkts_n).
19082e22920bSAdrien Mazarguil  */
19092e22920bSAdrien Mazarguil uint16_t
19102e22920bSAdrien Mazarguil mlx5_rx_burst(void *dpdk_rxq, struct rte_mbuf **pkts, uint16_t pkts_n)
19112e22920bSAdrien Mazarguil {
191278142aacSNélio Laranjeiro 	struct mlx5_rxq_data *rxq = dpdk_rxq;
1913b4b12e55SNélio Laranjeiro 	const unsigned int wqe_cnt = (1 << rxq->elts_n) - 1;
1914e2f116eeSNélio Laranjeiro 	const unsigned int cqe_cnt = (1 << rxq->cqe_n) - 1;
19159964b965SNélio Laranjeiro 	const unsigned int sges_n = rxq->sges_n;
19169964b965SNélio Laranjeiro 	struct rte_mbuf *pkt = NULL;
19179964b965SNélio Laranjeiro 	struct rte_mbuf *seg = NULL;
191897267b8eSNelio Laranjeiro 	volatile struct mlx5_cqe *cqe =
191997267b8eSNelio Laranjeiro 		&(*rxq->cqes)[rxq->cq_ci & cqe_cnt];
19209964b965SNélio Laranjeiro 	unsigned int i = 0;
19219964b965SNélio Laranjeiro 	unsigned int rq_ci = rxq->rq_ci << sges_n;
19224e66a6feSNelio Laranjeiro 	int len = 0; /* keep its value across iterations. */
19232e22920bSAdrien Mazarguil 
19249964b965SNélio Laranjeiro 	while (pkts_n) {
19259964b965SNélio Laranjeiro 		unsigned int idx = rq_ci & wqe_cnt;
1926*7d6bf6b8SYongseok Koh 		volatile struct mlx5_wqe_data_seg *wqe =
1927*7d6bf6b8SYongseok Koh 			&((volatile struct mlx5_wqe_data_seg *)rxq->wqes)[idx];
19289964b965SNélio Laranjeiro 		struct rte_mbuf *rep = (*rxq->elts)[idx];
1929ecf60761SNélio Laranjeiro 		uint32_t rss_hash_res = 0;
19309964b965SNélio Laranjeiro 
19319964b965SNélio Laranjeiro 		if (pkt)
19329964b965SNélio Laranjeiro 			NEXT(seg) = rep;
19339964b965SNélio Laranjeiro 		seg = rep;
19349964b965SNélio Laranjeiro 		rte_prefetch0(seg);
19356218063bSNélio Laranjeiro 		rte_prefetch0(cqe);
19369964b965SNélio Laranjeiro 		rte_prefetch0(wqe);
1937fbfd9955SOlivier Matz 		rep = rte_mbuf_raw_alloc(rxq->mp);
19382e22920bSAdrien Mazarguil 		if (unlikely(rep == NULL)) {
193915a756b6SSagi Grimberg 			++rxq->stats.rx_nombuf;
194015a756b6SSagi Grimberg 			if (!pkt) {
194115a756b6SSagi Grimberg 				/*
194215a756b6SSagi Grimberg 				 * no buffers before we even started,
194315a756b6SSagi Grimberg 				 * bail out silently.
194415a756b6SSagi Grimberg 				 */
194515a756b6SSagi Grimberg 				break;
194615a756b6SSagi Grimberg 			}
1947a1bdb71aSNélio Laranjeiro 			while (pkt != seg) {
1948a1bdb71aSNélio Laranjeiro 				assert(pkt != (*rxq->elts)[idx]);
1949fe5fe382SNélio Laranjeiro 				rep = NEXT(pkt);
19508f094a9aSOlivier Matz 				NEXT(pkt) = NULL;
19518f094a9aSOlivier Matz 				NB_SEGS(pkt) = 1;
19521f88c0a2SOlivier Matz 				rte_mbuf_raw_free(pkt);
1953fe5fe382SNélio Laranjeiro 				pkt = rep;
19549964b965SNélio Laranjeiro 			}
19556218063bSNélio Laranjeiro 			break;
19562e22920bSAdrien Mazarguil 		}
19579964b965SNélio Laranjeiro 		if (!pkt) {
195897267b8eSNelio Laranjeiro 			cqe = &(*rxq->cqes)[rxq->cq_ci & cqe_cnt];
1959ecf60761SNélio Laranjeiro 			len = mlx5_rx_poll_len(rxq, cqe, cqe_cnt,
1960ecf60761SNélio Laranjeiro 					       &rss_hash_res);
1961ecf60761SNélio Laranjeiro 			if (!len) {
19621f88c0a2SOlivier Matz 				rte_mbuf_raw_free(rep);
19636218063bSNélio Laranjeiro 				break;
19646218063bSNélio Laranjeiro 			}
196599c12dccSNélio Laranjeiro 			if (unlikely(len == -1)) {
196699c12dccSNélio Laranjeiro 				/* RX error, packet is likely too large. */
19671f88c0a2SOlivier Matz 				rte_mbuf_raw_free(rep);
196899c12dccSNélio Laranjeiro 				++rxq->stats.idropped;
196999c12dccSNélio Laranjeiro 				goto skip;
197099c12dccSNélio Laranjeiro 			}
19719964b965SNélio Laranjeiro 			pkt = seg;
19729964b965SNélio Laranjeiro 			assert(len >= (rxq->crc_present << 2));
19730ac64846SMaxime Leroy 			pkt->ol_flags = 0;
19743e1f82a1SYongseok Koh 			rxq_cq_to_mbuf(rxq, pkt, cqe, rss_hash_res);
19756218063bSNélio Laranjeiro 			if (rxq->crc_present)
19766218063bSNélio Laranjeiro 				len -= ETHER_CRC_LEN;
19776218063bSNélio Laranjeiro 			PKT_LEN(pkt) = len;
19789964b965SNélio Laranjeiro 		}
19799964b965SNélio Laranjeiro 		DATA_LEN(rep) = DATA_LEN(seg);
19809964b965SNélio Laranjeiro 		PKT_LEN(rep) = PKT_LEN(seg);
19819964b965SNélio Laranjeiro 		SET_DATA_OFF(rep, DATA_OFF(seg));
19829964b965SNélio Laranjeiro 		PORT(rep) = PORT(seg);
19839964b965SNélio Laranjeiro 		(*rxq->elts)[idx] = rep;
19849964b965SNélio Laranjeiro 		/*
19859964b965SNélio Laranjeiro 		 * Fill NIC descriptor with the new buffer.  The lkey and size
19869964b965SNélio Laranjeiro 		 * of the buffers are already known, only the buffer address
19879964b965SNélio Laranjeiro 		 * changes.
19889964b965SNélio Laranjeiro 		 */
19896b30a6a8SShachar Beiser 		wqe->addr = rte_cpu_to_be_64(rte_pktmbuf_mtod(rep, uintptr_t));
1990974f1e7eSYongseok Koh 		/* If there's only one MR, no need to replace LKey in WQE. */
1991974f1e7eSYongseok Koh 		if (unlikely(mlx5_mr_btree_len(&rxq->mr_ctrl.cache_bh) > 1))
1992974f1e7eSYongseok Koh 			wqe->lkey = mlx5_rx_mb2mr(rxq, rep);
19939964b965SNélio Laranjeiro 		if (len > DATA_LEN(seg)) {
19949964b965SNélio Laranjeiro 			len -= DATA_LEN(seg);
19959964b965SNélio Laranjeiro 			++NB_SEGS(pkt);
19969964b965SNélio Laranjeiro 			++rq_ci;
19979964b965SNélio Laranjeiro 			continue;
19989964b965SNélio Laranjeiro 		}
19999964b965SNélio Laranjeiro 		DATA_LEN(seg) = len;
200087011737SAdrien Mazarguil #ifdef MLX5_PMD_SOFT_COUNTERS
200187011737SAdrien Mazarguil 		/* Increment bytes counter. */
20029964b965SNélio Laranjeiro 		rxq->stats.ibytes += PKT_LEN(pkt);
200387011737SAdrien Mazarguil #endif
20046218063bSNélio Laranjeiro 		/* Return packet. */
20056218063bSNélio Laranjeiro 		*(pkts++) = pkt;
20069964b965SNélio Laranjeiro 		pkt = NULL;
20079964b965SNélio Laranjeiro 		--pkts_n;
20089964b965SNélio Laranjeiro 		++i;
200999c12dccSNélio Laranjeiro skip:
20109964b965SNélio Laranjeiro 		/* Align consumer index to the next stride. */
20119964b965SNélio Laranjeiro 		rq_ci >>= sges_n;
20126218063bSNélio Laranjeiro 		++rq_ci;
20139964b965SNélio Laranjeiro 		rq_ci <<= sges_n;
20142e22920bSAdrien Mazarguil 	}
20159964b965SNélio Laranjeiro 	if (unlikely((i == 0) && ((rq_ci >> sges_n) == rxq->rq_ci)))
20162e22920bSAdrien Mazarguil 		return 0;
20176218063bSNélio Laranjeiro 	/* Update the consumer index. */
20189964b965SNélio Laranjeiro 	rxq->rq_ci = rq_ci >> sges_n;
20194fe7f662SYongseok Koh 	rte_cio_wmb();
20206b30a6a8SShachar Beiser 	*rxq->cq_db = rte_cpu_to_be_32(rxq->cq_ci);
20214fe7f662SYongseok Koh 	rte_cio_wmb();
20226b30a6a8SShachar Beiser 	*rxq->rq_db = rte_cpu_to_be_32(rxq->rq_ci);
202387011737SAdrien Mazarguil #ifdef MLX5_PMD_SOFT_COUNTERS
202487011737SAdrien Mazarguil 	/* Increment packets counter. */
20259964b965SNélio Laranjeiro 	rxq->stats.ipackets += i;
202687011737SAdrien Mazarguil #endif
20279964b965SNélio Laranjeiro 	return i;
20282e22920bSAdrien Mazarguil }
20292e22920bSAdrien Mazarguil 
2030*7d6bf6b8SYongseok Koh void
2031*7d6bf6b8SYongseok Koh mlx5_mprq_buf_free_cb(void *addr __rte_unused, void *opaque)
2032*7d6bf6b8SYongseok Koh {
2033*7d6bf6b8SYongseok Koh 	struct mlx5_mprq_buf *buf = opaque;
2034*7d6bf6b8SYongseok Koh 
2035*7d6bf6b8SYongseok Koh 	if (rte_atomic16_read(&buf->refcnt) == 1) {
2036*7d6bf6b8SYongseok Koh 		rte_mempool_put(buf->mp, buf);
2037*7d6bf6b8SYongseok Koh 	} else if (rte_atomic16_add_return(&buf->refcnt, -1) == 0) {
2038*7d6bf6b8SYongseok Koh 		rte_atomic16_set(&buf->refcnt, 1);
2039*7d6bf6b8SYongseok Koh 		rte_mempool_put(buf->mp, buf);
2040*7d6bf6b8SYongseok Koh 	}
2041*7d6bf6b8SYongseok Koh }
2042*7d6bf6b8SYongseok Koh 
2043*7d6bf6b8SYongseok Koh void
2044*7d6bf6b8SYongseok Koh mlx5_mprq_buf_free(struct mlx5_mprq_buf *buf)
2045*7d6bf6b8SYongseok Koh {
2046*7d6bf6b8SYongseok Koh 	mlx5_mprq_buf_free_cb(NULL, buf);
2047*7d6bf6b8SYongseok Koh }
2048*7d6bf6b8SYongseok Koh 
2049*7d6bf6b8SYongseok Koh static inline void
2050*7d6bf6b8SYongseok Koh mprq_buf_replace(struct mlx5_rxq_data *rxq, uint16_t rq_idx)
2051*7d6bf6b8SYongseok Koh {
2052*7d6bf6b8SYongseok Koh 	struct mlx5_mprq_buf *rep = rxq->mprq_repl;
2053*7d6bf6b8SYongseok Koh 	volatile struct mlx5_wqe_data_seg *wqe =
2054*7d6bf6b8SYongseok Koh 		&((volatile struct mlx5_wqe_mprq *)rxq->wqes)[rq_idx].dseg;
2055*7d6bf6b8SYongseok Koh 	void *addr;
2056*7d6bf6b8SYongseok Koh 
2057*7d6bf6b8SYongseok Koh 	assert(rep != NULL);
2058*7d6bf6b8SYongseok Koh 	/* Replace MPRQ buf. */
2059*7d6bf6b8SYongseok Koh 	(*rxq->mprq_bufs)[rq_idx] = rep;
2060*7d6bf6b8SYongseok Koh 	/* Replace WQE. */
2061*7d6bf6b8SYongseok Koh 	addr = mlx5_mprq_buf_addr(rep);
2062*7d6bf6b8SYongseok Koh 	wqe->addr = rte_cpu_to_be_64((uintptr_t)addr);
2063*7d6bf6b8SYongseok Koh 	/* If there's only one MR, no need to replace LKey in WQE. */
2064*7d6bf6b8SYongseok Koh 	if (unlikely(mlx5_mr_btree_len(&rxq->mr_ctrl.cache_bh) > 1))
2065*7d6bf6b8SYongseok Koh 		wqe->lkey = mlx5_rx_addr2mr(rxq, (uintptr_t)addr);
2066*7d6bf6b8SYongseok Koh 	/* Stash a mbuf for next replacement. */
2067*7d6bf6b8SYongseok Koh 	if (likely(!rte_mempool_get(rxq->mprq_mp, (void **)&rep)))
2068*7d6bf6b8SYongseok Koh 		rxq->mprq_repl = rep;
2069*7d6bf6b8SYongseok Koh 	else
2070*7d6bf6b8SYongseok Koh 		rxq->mprq_repl = NULL;
2071*7d6bf6b8SYongseok Koh }
2072*7d6bf6b8SYongseok Koh 
2073*7d6bf6b8SYongseok Koh /**
2074*7d6bf6b8SYongseok Koh  * DPDK callback for RX with Multi-Packet RQ support.
2075*7d6bf6b8SYongseok Koh  *
2076*7d6bf6b8SYongseok Koh  * @param dpdk_rxq
2077*7d6bf6b8SYongseok Koh  *   Generic pointer to RX queue structure.
2078*7d6bf6b8SYongseok Koh  * @param[out] pkts
2079*7d6bf6b8SYongseok Koh  *   Array to store received packets.
2080*7d6bf6b8SYongseok Koh  * @param pkts_n
2081*7d6bf6b8SYongseok Koh  *   Maximum number of packets in array.
2082*7d6bf6b8SYongseok Koh  *
2083*7d6bf6b8SYongseok Koh  * @return
2084*7d6bf6b8SYongseok Koh  *   Number of packets successfully received (<= pkts_n).
2085*7d6bf6b8SYongseok Koh  */
2086*7d6bf6b8SYongseok Koh uint16_t
2087*7d6bf6b8SYongseok Koh mlx5_rx_burst_mprq(void *dpdk_rxq, struct rte_mbuf **pkts, uint16_t pkts_n)
2088*7d6bf6b8SYongseok Koh {
2089*7d6bf6b8SYongseok Koh 	struct mlx5_rxq_data *rxq = dpdk_rxq;
2090*7d6bf6b8SYongseok Koh 	const unsigned int strd_n = 1 << rxq->strd_num_n;
2091*7d6bf6b8SYongseok Koh 	const unsigned int strd_sz = 1 << rxq->strd_sz_n;
2092*7d6bf6b8SYongseok Koh 	const unsigned int strd_shift =
2093*7d6bf6b8SYongseok Koh 		MLX5_MPRQ_STRIDE_SHIFT_BYTE * rxq->strd_shift_en;
2094*7d6bf6b8SYongseok Koh 	const unsigned int cq_mask = (1 << rxq->cqe_n) - 1;
2095*7d6bf6b8SYongseok Koh 	const unsigned int wq_mask = (1 << rxq->elts_n) - 1;
2096*7d6bf6b8SYongseok Koh 	volatile struct mlx5_cqe *cqe = &(*rxq->cqes)[rxq->cq_ci & cq_mask];
2097*7d6bf6b8SYongseok Koh 	unsigned int i = 0;
2098*7d6bf6b8SYongseok Koh 	uint16_t rq_ci = rxq->rq_ci;
2099*7d6bf6b8SYongseok Koh 	uint16_t strd_idx = rxq->strd_ci;
2100*7d6bf6b8SYongseok Koh 	struct mlx5_mprq_buf *buf = (*rxq->mprq_bufs)[rq_ci & wq_mask];
2101*7d6bf6b8SYongseok Koh 
2102*7d6bf6b8SYongseok Koh 	while (i < pkts_n) {
2103*7d6bf6b8SYongseok Koh 		struct rte_mbuf *pkt;
2104*7d6bf6b8SYongseok Koh 		void *addr;
2105*7d6bf6b8SYongseok Koh 		int ret;
2106*7d6bf6b8SYongseok Koh 		unsigned int len;
2107*7d6bf6b8SYongseok Koh 		uint16_t consumed_strd;
2108*7d6bf6b8SYongseok Koh 		uint32_t offset;
2109*7d6bf6b8SYongseok Koh 		uint32_t byte_cnt;
2110*7d6bf6b8SYongseok Koh 		uint32_t rss_hash_res = 0;
2111*7d6bf6b8SYongseok Koh 
2112*7d6bf6b8SYongseok Koh 		if (strd_idx == strd_n) {
2113*7d6bf6b8SYongseok Koh 			/* Replace WQE only if the buffer is still in use. */
2114*7d6bf6b8SYongseok Koh 			if (rte_atomic16_read(&buf->refcnt) > 1) {
2115*7d6bf6b8SYongseok Koh 				mprq_buf_replace(rxq, rq_ci & wq_mask);
2116*7d6bf6b8SYongseok Koh 				/* Release the old buffer. */
2117*7d6bf6b8SYongseok Koh 				mlx5_mprq_buf_free(buf);
2118*7d6bf6b8SYongseok Koh 			} else if (unlikely(rxq->mprq_repl == NULL)) {
2119*7d6bf6b8SYongseok Koh 				struct mlx5_mprq_buf *rep;
2120*7d6bf6b8SYongseok Koh 
2121*7d6bf6b8SYongseok Koh 				/*
2122*7d6bf6b8SYongseok Koh 				 * Currently, the MPRQ mempool is out of buffer
2123*7d6bf6b8SYongseok Koh 				 * and doing memcpy regardless of the size of Rx
2124*7d6bf6b8SYongseok Koh 				 * packet. Retry allocation to get back to
2125*7d6bf6b8SYongseok Koh 				 * normal.
2126*7d6bf6b8SYongseok Koh 				 */
2127*7d6bf6b8SYongseok Koh 				if (!rte_mempool_get(rxq->mprq_mp,
2128*7d6bf6b8SYongseok Koh 						     (void **)&rep))
2129*7d6bf6b8SYongseok Koh 					rxq->mprq_repl = rep;
2130*7d6bf6b8SYongseok Koh 			}
2131*7d6bf6b8SYongseok Koh 			/* Advance to the next WQE. */
2132*7d6bf6b8SYongseok Koh 			strd_idx = 0;
2133*7d6bf6b8SYongseok Koh 			++rq_ci;
2134*7d6bf6b8SYongseok Koh 			buf = (*rxq->mprq_bufs)[rq_ci & wq_mask];
2135*7d6bf6b8SYongseok Koh 		}
2136*7d6bf6b8SYongseok Koh 		cqe = &(*rxq->cqes)[rxq->cq_ci & cq_mask];
2137*7d6bf6b8SYongseok Koh 		ret = mlx5_rx_poll_len(rxq, cqe, cq_mask, &rss_hash_res);
2138*7d6bf6b8SYongseok Koh 		if (!ret)
2139*7d6bf6b8SYongseok Koh 			break;
2140*7d6bf6b8SYongseok Koh 		if (unlikely(ret == -1)) {
2141*7d6bf6b8SYongseok Koh 			/* RX error, packet is likely too large. */
2142*7d6bf6b8SYongseok Koh 			++rxq->stats.idropped;
2143*7d6bf6b8SYongseok Koh 			continue;
2144*7d6bf6b8SYongseok Koh 		}
2145*7d6bf6b8SYongseok Koh 		byte_cnt = ret;
2146*7d6bf6b8SYongseok Koh 		consumed_strd = (byte_cnt & MLX5_MPRQ_STRIDE_NUM_MASK) >>
2147*7d6bf6b8SYongseok Koh 				MLX5_MPRQ_STRIDE_NUM_SHIFT;
2148*7d6bf6b8SYongseok Koh 		assert(consumed_strd);
2149*7d6bf6b8SYongseok Koh 		/* Calculate offset before adding up stride index. */
2150*7d6bf6b8SYongseok Koh 		offset = strd_idx * strd_sz + strd_shift;
2151*7d6bf6b8SYongseok Koh 		strd_idx += consumed_strd;
2152*7d6bf6b8SYongseok Koh 		if (byte_cnt & MLX5_MPRQ_FILLER_MASK)
2153*7d6bf6b8SYongseok Koh 			continue;
2154*7d6bf6b8SYongseok Koh 		/*
2155*7d6bf6b8SYongseok Koh 		 * Currently configured to receive a packet per a stride. But if
2156*7d6bf6b8SYongseok Koh 		 * MTU is adjusted through kernel interface, device could
2157*7d6bf6b8SYongseok Koh 		 * consume multiple strides without raising an error. In this
2158*7d6bf6b8SYongseok Koh 		 * case, the packet should be dropped because it is bigger than
2159*7d6bf6b8SYongseok Koh 		 * the max_rx_pkt_len.
2160*7d6bf6b8SYongseok Koh 		 */
2161*7d6bf6b8SYongseok Koh 		if (unlikely(consumed_strd > 1)) {
2162*7d6bf6b8SYongseok Koh 			++rxq->stats.idropped;
2163*7d6bf6b8SYongseok Koh 			continue;
2164*7d6bf6b8SYongseok Koh 		}
2165*7d6bf6b8SYongseok Koh 		pkt = rte_pktmbuf_alloc(rxq->mp);
2166*7d6bf6b8SYongseok Koh 		if (unlikely(pkt == NULL)) {
2167*7d6bf6b8SYongseok Koh 			++rxq->stats.rx_nombuf;
2168*7d6bf6b8SYongseok Koh 			break;
2169*7d6bf6b8SYongseok Koh 		}
2170*7d6bf6b8SYongseok Koh 		len = (byte_cnt & MLX5_MPRQ_LEN_MASK) >> MLX5_MPRQ_LEN_SHIFT;
2171*7d6bf6b8SYongseok Koh 		assert((int)len >= (rxq->crc_present << 2));
2172*7d6bf6b8SYongseok Koh 		if (rxq->crc_present)
2173*7d6bf6b8SYongseok Koh 			len -= ETHER_CRC_LEN;
2174*7d6bf6b8SYongseok Koh 		addr = RTE_PTR_ADD(mlx5_mprq_buf_addr(buf), offset);
2175*7d6bf6b8SYongseok Koh 		/* Initialize the offload flag. */
2176*7d6bf6b8SYongseok Koh 		pkt->ol_flags = 0;
2177*7d6bf6b8SYongseok Koh 		/*
2178*7d6bf6b8SYongseok Koh 		 * Memcpy packets to the target mbuf if:
2179*7d6bf6b8SYongseok Koh 		 * - The size of packet is smaller than mprq_max_memcpy_len.
2180*7d6bf6b8SYongseok Koh 		 * - Out of buffer in the Mempool for Multi-Packet RQ.
2181*7d6bf6b8SYongseok Koh 		 */
2182*7d6bf6b8SYongseok Koh 		if (len <= rxq->mprq_max_memcpy_len || rxq->mprq_repl == NULL) {
2183*7d6bf6b8SYongseok Koh 			/*
2184*7d6bf6b8SYongseok Koh 			 * When memcpy'ing packet due to out-of-buffer, the
2185*7d6bf6b8SYongseok Koh 			 * packet must be smaller than the target mbuf.
2186*7d6bf6b8SYongseok Koh 			 */
2187*7d6bf6b8SYongseok Koh 			if (unlikely(rte_pktmbuf_tailroom(pkt) < len)) {
2188*7d6bf6b8SYongseok Koh 				rte_pktmbuf_free_seg(pkt);
2189*7d6bf6b8SYongseok Koh 				++rxq->stats.idropped;
2190*7d6bf6b8SYongseok Koh 				continue;
2191*7d6bf6b8SYongseok Koh 			}
2192*7d6bf6b8SYongseok Koh 			rte_memcpy(rte_pktmbuf_mtod(pkt, void *), addr, len);
2193*7d6bf6b8SYongseok Koh 		} else {
2194*7d6bf6b8SYongseok Koh 			rte_iova_t buf_iova;
2195*7d6bf6b8SYongseok Koh 			struct rte_mbuf_ext_shared_info *shinfo;
2196*7d6bf6b8SYongseok Koh 			uint16_t buf_len = consumed_strd * strd_sz;
2197*7d6bf6b8SYongseok Koh 
2198*7d6bf6b8SYongseok Koh 			/* Increment the refcnt of the whole chunk. */
2199*7d6bf6b8SYongseok Koh 			rte_atomic16_add_return(&buf->refcnt, 1);
2200*7d6bf6b8SYongseok Koh 			assert((uint16_t)rte_atomic16_read(&buf->refcnt) <=
2201*7d6bf6b8SYongseok Koh 			       strd_n + 1);
2202*7d6bf6b8SYongseok Koh 			addr = RTE_PTR_SUB(addr, RTE_PKTMBUF_HEADROOM);
2203*7d6bf6b8SYongseok Koh 			/*
2204*7d6bf6b8SYongseok Koh 			 * MLX5 device doesn't use iova but it is necessary in a
2205*7d6bf6b8SYongseok Koh 			 * case where the Rx packet is transmitted via a
2206*7d6bf6b8SYongseok Koh 			 * different PMD.
2207*7d6bf6b8SYongseok Koh 			 */
2208*7d6bf6b8SYongseok Koh 			buf_iova = rte_mempool_virt2iova(buf) +
2209*7d6bf6b8SYongseok Koh 				   RTE_PTR_DIFF(addr, buf);
2210*7d6bf6b8SYongseok Koh 			shinfo = rte_pktmbuf_ext_shinfo_init_helper(addr,
2211*7d6bf6b8SYongseok Koh 					&buf_len, mlx5_mprq_buf_free_cb, buf);
2212*7d6bf6b8SYongseok Koh 			/*
2213*7d6bf6b8SYongseok Koh 			 * EXT_ATTACHED_MBUF will be set to pkt->ol_flags when
2214*7d6bf6b8SYongseok Koh 			 * attaching the stride to mbuf and more offload flags
2215*7d6bf6b8SYongseok Koh 			 * will be added below by calling rxq_cq_to_mbuf().
2216*7d6bf6b8SYongseok Koh 			 * Other fields will be overwritten.
2217*7d6bf6b8SYongseok Koh 			 */
2218*7d6bf6b8SYongseok Koh 			rte_pktmbuf_attach_extbuf(pkt, addr, buf_iova, buf_len,
2219*7d6bf6b8SYongseok Koh 						  shinfo);
2220*7d6bf6b8SYongseok Koh 			rte_pktmbuf_reset_headroom(pkt);
2221*7d6bf6b8SYongseok Koh 			assert(pkt->ol_flags == EXT_ATTACHED_MBUF);
2222*7d6bf6b8SYongseok Koh 			/*
2223*7d6bf6b8SYongseok Koh 			 * Prevent potential overflow due to MTU change through
2224*7d6bf6b8SYongseok Koh 			 * kernel interface.
2225*7d6bf6b8SYongseok Koh 			 */
2226*7d6bf6b8SYongseok Koh 			if (unlikely(rte_pktmbuf_tailroom(pkt) < len)) {
2227*7d6bf6b8SYongseok Koh 				rte_pktmbuf_free_seg(pkt);
2228*7d6bf6b8SYongseok Koh 				++rxq->stats.idropped;
2229*7d6bf6b8SYongseok Koh 				continue;
2230*7d6bf6b8SYongseok Koh 			}
2231*7d6bf6b8SYongseok Koh 		}
2232*7d6bf6b8SYongseok Koh 		rxq_cq_to_mbuf(rxq, pkt, cqe, rss_hash_res);
2233*7d6bf6b8SYongseok Koh 		PKT_LEN(pkt) = len;
2234*7d6bf6b8SYongseok Koh 		DATA_LEN(pkt) = len;
2235*7d6bf6b8SYongseok Koh 		PORT(pkt) = rxq->port_id;
2236*7d6bf6b8SYongseok Koh #ifdef MLX5_PMD_SOFT_COUNTERS
2237*7d6bf6b8SYongseok Koh 		/* Increment bytes counter. */
2238*7d6bf6b8SYongseok Koh 		rxq->stats.ibytes += PKT_LEN(pkt);
2239*7d6bf6b8SYongseok Koh #endif
2240*7d6bf6b8SYongseok Koh 		/* Return packet. */
2241*7d6bf6b8SYongseok Koh 		*(pkts++) = pkt;
2242*7d6bf6b8SYongseok Koh 		++i;
2243*7d6bf6b8SYongseok Koh 	}
2244*7d6bf6b8SYongseok Koh 	/* Update the consumer indexes. */
2245*7d6bf6b8SYongseok Koh 	rxq->strd_ci = strd_idx;
2246*7d6bf6b8SYongseok Koh 	rte_io_wmb();
2247*7d6bf6b8SYongseok Koh 	*rxq->cq_db = rte_cpu_to_be_32(rxq->cq_ci);
2248*7d6bf6b8SYongseok Koh 	if (rq_ci != rxq->rq_ci) {
2249*7d6bf6b8SYongseok Koh 		rxq->rq_ci = rq_ci;
2250*7d6bf6b8SYongseok Koh 		rte_io_wmb();
2251*7d6bf6b8SYongseok Koh 		*rxq->rq_db = rte_cpu_to_be_32(rxq->rq_ci);
2252*7d6bf6b8SYongseok Koh 	}
2253*7d6bf6b8SYongseok Koh #ifdef MLX5_PMD_SOFT_COUNTERS
2254*7d6bf6b8SYongseok Koh 	/* Increment packets counter. */
2255*7d6bf6b8SYongseok Koh 	rxq->stats.ipackets += i;
2256*7d6bf6b8SYongseok Koh #endif
2257*7d6bf6b8SYongseok Koh 	return i;
2258*7d6bf6b8SYongseok Koh }
2259*7d6bf6b8SYongseok Koh 
22602e22920bSAdrien Mazarguil /**
22612e22920bSAdrien Mazarguil  * Dummy DPDK callback for TX.
22622e22920bSAdrien Mazarguil  *
22632e22920bSAdrien Mazarguil  * This function is used to temporarily replace the real callback during
22642e22920bSAdrien Mazarguil  * unsafe control operations on the queue, or in case of error.
22652e22920bSAdrien Mazarguil  *
22662e22920bSAdrien Mazarguil  * @param dpdk_txq
22672e22920bSAdrien Mazarguil  *   Generic pointer to TX queue structure.
22682e22920bSAdrien Mazarguil  * @param[in] pkts
22692e22920bSAdrien Mazarguil  *   Packets to transmit.
22702e22920bSAdrien Mazarguil  * @param pkts_n
22712e22920bSAdrien Mazarguil  *   Number of packets in array.
22722e22920bSAdrien Mazarguil  *
22732e22920bSAdrien Mazarguil  * @return
22742e22920bSAdrien Mazarguil  *   Number of packets successfully transmitted (<= pkts_n).
22752e22920bSAdrien Mazarguil  */
22762e22920bSAdrien Mazarguil uint16_t
227756f08e16SNélio Laranjeiro removed_tx_burst(void *dpdk_txq __rte_unused,
227856f08e16SNélio Laranjeiro 		 struct rte_mbuf **pkts __rte_unused,
227956f08e16SNélio Laranjeiro 		 uint16_t pkts_n __rte_unused)
22802e22920bSAdrien Mazarguil {
22812e22920bSAdrien Mazarguil 	return 0;
22822e22920bSAdrien Mazarguil }
22832e22920bSAdrien Mazarguil 
22842e22920bSAdrien Mazarguil /**
22852e22920bSAdrien Mazarguil  * Dummy DPDK callback for RX.
22862e22920bSAdrien Mazarguil  *
22872e22920bSAdrien Mazarguil  * This function is used to temporarily replace the real callback during
22882e22920bSAdrien Mazarguil  * unsafe control operations on the queue, or in case of error.
22892e22920bSAdrien Mazarguil  *
22902e22920bSAdrien Mazarguil  * @param dpdk_rxq
22912e22920bSAdrien Mazarguil  *   Generic pointer to RX queue structure.
22922e22920bSAdrien Mazarguil  * @param[out] pkts
22932e22920bSAdrien Mazarguil  *   Array to store received packets.
22942e22920bSAdrien Mazarguil  * @param pkts_n
22952e22920bSAdrien Mazarguil  *   Maximum number of packets in array.
22962e22920bSAdrien Mazarguil  *
22972e22920bSAdrien Mazarguil  * @return
22982e22920bSAdrien Mazarguil  *   Number of packets successfully received (<= pkts_n).
22992e22920bSAdrien Mazarguil  */
23002e22920bSAdrien Mazarguil uint16_t
230156f08e16SNélio Laranjeiro removed_rx_burst(void *dpdk_txq __rte_unused,
230256f08e16SNélio Laranjeiro 		 struct rte_mbuf **pkts __rte_unused,
230356f08e16SNélio Laranjeiro 		 uint16_t pkts_n __rte_unused)
23042e22920bSAdrien Mazarguil {
23052e22920bSAdrien Mazarguil 	return 0;
23062e22920bSAdrien Mazarguil }
23076cb559d6SYongseok Koh 
23086cb559d6SYongseok Koh /*
23096cb559d6SYongseok Koh  * Vectorized Rx/Tx routines are not compiled in when required vector
23106cb559d6SYongseok Koh  * instructions are not supported on a target architecture. The following null
23116cb559d6SYongseok Koh  * stubs are needed for linkage when those are not included outside of this file
23126cb559d6SYongseok Koh  * (e.g.  mlx5_rxtx_vec_sse.c for x86).
23136cb559d6SYongseok Koh  */
23146cb559d6SYongseok Koh 
23156cb559d6SYongseok Koh uint16_t __attribute__((weak))
231656f08e16SNélio Laranjeiro mlx5_tx_burst_raw_vec(void *dpdk_txq __rte_unused,
231756f08e16SNélio Laranjeiro 		      struct rte_mbuf **pkts __rte_unused,
231856f08e16SNélio Laranjeiro 		      uint16_t pkts_n __rte_unused)
23196cb559d6SYongseok Koh {
23206cb559d6SYongseok Koh 	return 0;
23216cb559d6SYongseok Koh }
23226cb559d6SYongseok Koh 
23236cb559d6SYongseok Koh uint16_t __attribute__((weak))
232456f08e16SNélio Laranjeiro mlx5_tx_burst_vec(void *dpdk_txq __rte_unused,
232556f08e16SNélio Laranjeiro 		  struct rte_mbuf **pkts __rte_unused,
232656f08e16SNélio Laranjeiro 		  uint16_t pkts_n __rte_unused)
23276cb559d6SYongseok Koh {
23286cb559d6SYongseok Koh 	return 0;
23296cb559d6SYongseok Koh }
23306cb559d6SYongseok Koh 
23316cb559d6SYongseok Koh uint16_t __attribute__((weak))
233256f08e16SNélio Laranjeiro mlx5_rx_burst_vec(void *dpdk_txq __rte_unused,
233356f08e16SNélio Laranjeiro 		  struct rte_mbuf **pkts __rte_unused,
233456f08e16SNélio Laranjeiro 		  uint16_t pkts_n __rte_unused)
23356cb559d6SYongseok Koh {
23366cb559d6SYongseok Koh 	return 0;
23376cb559d6SYongseok Koh }
23386cb559d6SYongseok Koh 
23396cb559d6SYongseok Koh int __attribute__((weak))
2340af4f09f2SNélio Laranjeiro mlx5_check_raw_vec_tx_support(struct rte_eth_dev *dev __rte_unused)
23416cb559d6SYongseok Koh {
23426cb559d6SYongseok Koh 	return -ENOTSUP;
23436cb559d6SYongseok Koh }
23446cb559d6SYongseok Koh 
23456cb559d6SYongseok Koh int __attribute__((weak))
2346af4f09f2SNélio Laranjeiro mlx5_check_vec_tx_support(struct rte_eth_dev *dev __rte_unused)
23476cb559d6SYongseok Koh {
23486cb559d6SYongseok Koh 	return -ENOTSUP;
23496cb559d6SYongseok Koh }
23506cb559d6SYongseok Koh 
23516cb559d6SYongseok Koh int __attribute__((weak))
2352af4f09f2SNélio Laranjeiro mlx5_rxq_check_vec_support(struct mlx5_rxq_data *rxq __rte_unused)
23536cb559d6SYongseok Koh {
23546cb559d6SYongseok Koh 	return -ENOTSUP;
23556cb559d6SYongseok Koh }
23566cb559d6SYongseok Koh 
23576cb559d6SYongseok Koh int __attribute__((weak))
2358af4f09f2SNélio Laranjeiro mlx5_check_vec_rx_support(struct rte_eth_dev *dev __rte_unused)
23596cb559d6SYongseok Koh {
23606cb559d6SYongseok Koh 	return -ENOTSUP;
23616cb559d6SYongseok Koh }
2362