18fd92a66SOlivier Matz /* SPDX-License-Identifier: BSD-3-Clause 22e22920bSAdrien Mazarguil * Copyright 2015 6WIND S.A. 35feecc57SShahaf Shuler * Copyright 2015 Mellanox Technologies, Ltd 42e22920bSAdrien Mazarguil */ 52e22920bSAdrien Mazarguil 62e22920bSAdrien Mazarguil #include <assert.h> 72e22920bSAdrien Mazarguil #include <stdint.h> 82e22920bSAdrien Mazarguil #include <string.h> 92e22920bSAdrien Mazarguil #include <stdlib.h> 102e22920bSAdrien Mazarguil 112e22920bSAdrien Mazarguil /* Verbs header. */ 122e22920bSAdrien Mazarguil /* ISO C doesn't support unnamed structs/unions, disabling -pedantic. */ 132e22920bSAdrien Mazarguil #ifdef PEDANTIC 14fc5b160fSBruce Richardson #pragma GCC diagnostic ignored "-Wpedantic" 152e22920bSAdrien Mazarguil #endif 162e22920bSAdrien Mazarguil #include <infiniband/verbs.h> 1743e9d979SShachar Beiser #include <infiniband/mlx5dv.h> 182e22920bSAdrien Mazarguil #ifdef PEDANTIC 19fc5b160fSBruce Richardson #pragma GCC diagnostic error "-Wpedantic" 202e22920bSAdrien Mazarguil #endif 212e22920bSAdrien Mazarguil 222e22920bSAdrien Mazarguil #include <rte_mbuf.h> 232e22920bSAdrien Mazarguil #include <rte_mempool.h> 242e22920bSAdrien Mazarguil #include <rte_prefetch.h> 252e22920bSAdrien Mazarguil #include <rte_common.h> 262e22920bSAdrien Mazarguil #include <rte_branch_prediction.h> 276218063bSNélio Laranjeiro #include <rte_ether.h> 282e22920bSAdrien Mazarguil 292e22920bSAdrien Mazarguil #include "mlx5.h" 302e22920bSAdrien Mazarguil #include "mlx5_utils.h" 312e22920bSAdrien Mazarguil #include "mlx5_rxtx.h" 32f3db9489SYaacov Hazan #include "mlx5_autoconf.h" 332e22920bSAdrien Mazarguil #include "mlx5_defs.h" 346218063bSNélio Laranjeiro #include "mlx5_prm.h" 356218063bSNélio Laranjeiro 36c0583d98SJerin Jacob static __rte_always_inline uint32_t 373cc08bc6SXueming Li rxq_cq_to_pkt_type(struct mlx5_rxq_data *rxq, volatile struct mlx5_cqe *cqe); 38ff1807a3SNélio Laranjeiro 39c0583d98SJerin Jacob static __rte_always_inline int 4078142aacSNélio Laranjeiro mlx5_rx_poll_len(struct mlx5_rxq_data *rxq, volatile struct mlx5_cqe *cqe, 41c0583d98SJerin Jacob uint16_t cqe_cnt, uint32_t *rss_hash); 42ff1807a3SNélio Laranjeiro 43c0583d98SJerin Jacob static __rte_always_inline uint32_t 446ba07449SXueming Li rxq_cq_to_ol_flags(volatile struct mlx5_cqe *cqe); 45ff1807a3SNélio Laranjeiro 463e1f82a1SYongseok Koh static __rte_always_inline void 473e1f82a1SYongseok Koh rxq_cq_to_mbuf(struct mlx5_rxq_data *rxq, struct rte_mbuf *pkt, 483e1f82a1SYongseok Koh volatile struct mlx5_cqe *cqe, uint32_t rss_hash_res); 493e1f82a1SYongseok Koh 50*7d6bf6b8SYongseok Koh static __rte_always_inline void 51*7d6bf6b8SYongseok Koh mprq_buf_replace(struct mlx5_rxq_data *rxq, uint16_t rq_idx); 52*7d6bf6b8SYongseok Koh 53ea16068cSYongseok Koh uint32_t mlx5_ptype_table[] __rte_cache_aligned = { 54ea16068cSYongseok Koh [0xff] = RTE_PTYPE_ALL_MASK, /* Last entry for errored packet. */ 55ea16068cSYongseok Koh }; 56ea16068cSYongseok Koh 575f8ba81cSXueming Li uint8_t mlx5_cksum_table[1 << 10] __rte_cache_aligned; 585f8ba81cSXueming Li uint8_t mlx5_swp_types_table[1 << 10] __rte_cache_aligned; 595f8ba81cSXueming Li 60ea16068cSYongseok Koh /** 61ea16068cSYongseok Koh * Build a table to translate Rx completion flags to packet type. 62ea16068cSYongseok Koh * 63ea16068cSYongseok Koh * @note: fix mlx5_dev_supported_ptypes_get() if any change here. 64ea16068cSYongseok Koh */ 65ea16068cSYongseok Koh void 66ea16068cSYongseok Koh mlx5_set_ptype_table(void) 67ea16068cSYongseok Koh { 68ea16068cSYongseok Koh unsigned int i; 69ea16068cSYongseok Koh uint32_t (*p)[RTE_DIM(mlx5_ptype_table)] = &mlx5_ptype_table; 70ea16068cSYongseok Koh 719807f113SYongseok Koh /* Last entry must not be overwritten, reserved for errored packet. */ 729807f113SYongseok Koh for (i = 0; i < RTE_DIM(mlx5_ptype_table) - 1; ++i) 73ea16068cSYongseok Koh (*p)[i] = RTE_PTYPE_UNKNOWN; 746cb559d6SYongseok Koh /* 756cb559d6SYongseok Koh * The index to the array should have: 76ea16068cSYongseok Koh * bit[1:0] = l3_hdr_type 77ea16068cSYongseok Koh * bit[4:2] = l4_hdr_type 78ea16068cSYongseok Koh * bit[5] = ip_frag 79ea16068cSYongseok Koh * bit[6] = tunneled 80ea16068cSYongseok Koh * bit[7] = outer_l3_type 8199c12dccSNélio Laranjeiro */ 823ca63b88SShahaf Shuler /* L2 */ 833ca63b88SShahaf Shuler (*p)[0x00] = RTE_PTYPE_L2_ETHER; 84ea16068cSYongseok Koh /* L3 */ 85ea16068cSYongseok Koh (*p)[0x01] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN | 86ea16068cSYongseok Koh RTE_PTYPE_L4_NONFRAG; 87ea16068cSYongseok Koh (*p)[0x02] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN | 88ea16068cSYongseok Koh RTE_PTYPE_L4_NONFRAG; 89ea16068cSYongseok Koh /* Fragmented */ 90ea16068cSYongseok Koh (*p)[0x21] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN | 91ea16068cSYongseok Koh RTE_PTYPE_L4_FRAG; 92ea16068cSYongseok Koh (*p)[0x22] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN | 93ea16068cSYongseok Koh RTE_PTYPE_L4_FRAG; 94ea16068cSYongseok Koh /* TCP */ 95ea16068cSYongseok Koh (*p)[0x05] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN | 96ea16068cSYongseok Koh RTE_PTYPE_L4_TCP; 97ea16068cSYongseok Koh (*p)[0x06] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN | 98ea16068cSYongseok Koh RTE_PTYPE_L4_TCP; 990915e287SBin Huang (*p)[0x0d] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN | 1000915e287SBin Huang RTE_PTYPE_L4_TCP; 1010915e287SBin Huang (*p)[0x0e] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN | 1020915e287SBin Huang RTE_PTYPE_L4_TCP; 1030915e287SBin Huang (*p)[0x11] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN | 1040915e287SBin Huang RTE_PTYPE_L4_TCP; 1050915e287SBin Huang (*p)[0x12] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN | 1060915e287SBin Huang RTE_PTYPE_L4_TCP; 107ea16068cSYongseok Koh /* UDP */ 108ea16068cSYongseok Koh (*p)[0x09] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN | 109ea16068cSYongseok Koh RTE_PTYPE_L4_UDP; 110ea16068cSYongseok Koh (*p)[0x0a] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN | 111ea16068cSYongseok Koh RTE_PTYPE_L4_UDP; 112ea16068cSYongseok Koh /* Repeat with outer_l3_type being set. Just in case. */ 113ea16068cSYongseok Koh (*p)[0x81] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN | 114ea16068cSYongseok Koh RTE_PTYPE_L4_NONFRAG; 115ea16068cSYongseok Koh (*p)[0x82] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN | 116ea16068cSYongseok Koh RTE_PTYPE_L4_NONFRAG; 117ea16068cSYongseok Koh (*p)[0xa1] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN | 118ea16068cSYongseok Koh RTE_PTYPE_L4_FRAG; 119ea16068cSYongseok Koh (*p)[0xa2] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN | 120ea16068cSYongseok Koh RTE_PTYPE_L4_FRAG; 121ea16068cSYongseok Koh (*p)[0x85] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN | 122ea16068cSYongseok Koh RTE_PTYPE_L4_TCP; 123ea16068cSYongseok Koh (*p)[0x86] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN | 124ea16068cSYongseok Koh RTE_PTYPE_L4_TCP; 1250915e287SBin Huang (*p)[0x8d] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN | 1260915e287SBin Huang RTE_PTYPE_L4_TCP; 1270915e287SBin Huang (*p)[0x8e] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN | 1280915e287SBin Huang RTE_PTYPE_L4_TCP; 1290915e287SBin Huang (*p)[0x91] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN | 1300915e287SBin Huang RTE_PTYPE_L4_TCP; 1310915e287SBin Huang (*p)[0x92] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN | 1320915e287SBin Huang RTE_PTYPE_L4_TCP; 133ea16068cSYongseok Koh (*p)[0x89] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN | 134ea16068cSYongseok Koh RTE_PTYPE_L4_UDP; 135ea16068cSYongseok Koh (*p)[0x8a] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN | 136ea16068cSYongseok Koh RTE_PTYPE_L4_UDP; 137ea16068cSYongseok Koh /* Tunneled - L3 */ 1383cc08bc6SXueming Li (*p)[0x40] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN; 139ea16068cSYongseok Koh (*p)[0x41] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN | 140ea16068cSYongseok Koh RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN | 141ea16068cSYongseok Koh RTE_PTYPE_INNER_L4_NONFRAG; 142ea16068cSYongseok Koh (*p)[0x42] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN | 143ea16068cSYongseok Koh RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN | 144ea16068cSYongseok Koh RTE_PTYPE_INNER_L4_NONFRAG; 1453cc08bc6SXueming Li (*p)[0xc0] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN; 146ea16068cSYongseok Koh (*p)[0xc1] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN | 147ea16068cSYongseok Koh RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN | 148ea16068cSYongseok Koh RTE_PTYPE_INNER_L4_NONFRAG; 149ea16068cSYongseok Koh (*p)[0xc2] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN | 150ea16068cSYongseok Koh RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN | 151ea16068cSYongseok Koh RTE_PTYPE_INNER_L4_NONFRAG; 152ea16068cSYongseok Koh /* Tunneled - Fragmented */ 153ea16068cSYongseok Koh (*p)[0x61] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN | 154ea16068cSYongseok Koh RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN | 155ea16068cSYongseok Koh RTE_PTYPE_INNER_L4_FRAG; 156ea16068cSYongseok Koh (*p)[0x62] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN | 157ea16068cSYongseok Koh RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN | 158ea16068cSYongseok Koh RTE_PTYPE_INNER_L4_FRAG; 159ea16068cSYongseok Koh (*p)[0xe1] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN | 160ea16068cSYongseok Koh RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN | 161ea16068cSYongseok Koh RTE_PTYPE_INNER_L4_FRAG; 162ea16068cSYongseok Koh (*p)[0xe2] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN | 163ea16068cSYongseok Koh RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN | 164ea16068cSYongseok Koh RTE_PTYPE_INNER_L4_FRAG; 165ea16068cSYongseok Koh /* Tunneled - TCP */ 166ea16068cSYongseok Koh (*p)[0x45] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN | 167ea16068cSYongseok Koh RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN | 1686c897093SYongseok Koh RTE_PTYPE_INNER_L4_TCP; 169ea16068cSYongseok Koh (*p)[0x46] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN | 170ea16068cSYongseok Koh RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN | 1716c897093SYongseok Koh RTE_PTYPE_INNER_L4_TCP; 1720915e287SBin Huang (*p)[0x4d] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN | 1730915e287SBin Huang RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN | 1740915e287SBin Huang RTE_PTYPE_INNER_L4_TCP; 1750915e287SBin Huang (*p)[0x4e] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN | 1760915e287SBin Huang RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN | 1770915e287SBin Huang RTE_PTYPE_INNER_L4_TCP; 1780915e287SBin Huang (*p)[0x51] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN | 1790915e287SBin Huang RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN | 1800915e287SBin Huang RTE_PTYPE_INNER_L4_TCP; 1810915e287SBin Huang (*p)[0x52] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN | 1820915e287SBin Huang RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN | 1830915e287SBin Huang RTE_PTYPE_INNER_L4_TCP; 184ea16068cSYongseok Koh (*p)[0xc5] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN | 185ea16068cSYongseok Koh RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN | 1866c897093SYongseok Koh RTE_PTYPE_INNER_L4_TCP; 187ea16068cSYongseok Koh (*p)[0xc6] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN | 188ea16068cSYongseok Koh RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN | 1896c897093SYongseok Koh RTE_PTYPE_INNER_L4_TCP; 1900915e287SBin Huang (*p)[0xcd] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN | 1910915e287SBin Huang RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN | 1920915e287SBin Huang RTE_PTYPE_INNER_L4_TCP; 1930915e287SBin Huang (*p)[0xce] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN | 1940915e287SBin Huang RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN | 1950915e287SBin Huang RTE_PTYPE_INNER_L4_TCP; 1960915e287SBin Huang (*p)[0xd1] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN | 1970915e287SBin Huang RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN | 1980915e287SBin Huang RTE_PTYPE_INNER_L4_TCP; 1990915e287SBin Huang (*p)[0xd2] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN | 2000915e287SBin Huang RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN | 2010915e287SBin Huang RTE_PTYPE_INNER_L4_TCP; 202ea16068cSYongseok Koh /* Tunneled - UDP */ 203ea16068cSYongseok Koh (*p)[0x49] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN | 204ea16068cSYongseok Koh RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN | 2056c897093SYongseok Koh RTE_PTYPE_INNER_L4_UDP; 206ea16068cSYongseok Koh (*p)[0x4a] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4_EXT_UNKNOWN | 207ea16068cSYongseok Koh RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN | 2086c897093SYongseok Koh RTE_PTYPE_INNER_L4_UDP; 209ea16068cSYongseok Koh (*p)[0xc9] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN | 210ea16068cSYongseok Koh RTE_PTYPE_INNER_L3_IPV6_EXT_UNKNOWN | 2116c897093SYongseok Koh RTE_PTYPE_INNER_L4_UDP; 212ea16068cSYongseok Koh (*p)[0xca] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6_EXT_UNKNOWN | 213ea16068cSYongseok Koh RTE_PTYPE_INNER_L3_IPV4_EXT_UNKNOWN | 2146c897093SYongseok Koh RTE_PTYPE_INNER_L4_UDP; 215ea16068cSYongseok Koh } 216fdcb0f53SNélio Laranjeiro 2172e22920bSAdrien Mazarguil /** 2185f8ba81cSXueming Li * Build a table to translate packet to checksum type of Verbs. 2195f8ba81cSXueming Li */ 2205f8ba81cSXueming Li void 2215f8ba81cSXueming Li mlx5_set_cksum_table(void) 2225f8ba81cSXueming Li { 2235f8ba81cSXueming Li unsigned int i; 2245f8ba81cSXueming Li uint8_t v; 2255f8ba81cSXueming Li 2265f8ba81cSXueming Li /* 2275f8ba81cSXueming Li * The index should have: 2285f8ba81cSXueming Li * bit[0] = PKT_TX_TCP_SEG 2295f8ba81cSXueming Li * bit[2:3] = PKT_TX_UDP_CKSUM, PKT_TX_TCP_CKSUM 2305f8ba81cSXueming Li * bit[4] = PKT_TX_IP_CKSUM 2315f8ba81cSXueming Li * bit[8] = PKT_TX_OUTER_IP_CKSUM 2325f8ba81cSXueming Li * bit[9] = tunnel 2335f8ba81cSXueming Li */ 2345f8ba81cSXueming Li for (i = 0; i < RTE_DIM(mlx5_cksum_table); ++i) { 2355f8ba81cSXueming Li v = 0; 2365f8ba81cSXueming Li if (i & (1 << 9)) { 2375f8ba81cSXueming Li /* Tunneled packet. */ 2385f8ba81cSXueming Li if (i & (1 << 8)) /* Outer IP. */ 2395f8ba81cSXueming Li v |= MLX5_ETH_WQE_L3_CSUM; 2405f8ba81cSXueming Li if (i & (1 << 4)) /* Inner IP. */ 2415f8ba81cSXueming Li v |= MLX5_ETH_WQE_L3_INNER_CSUM; 2425f8ba81cSXueming Li if (i & (3 << 2 | 1 << 0)) /* L4 or TSO. */ 2435f8ba81cSXueming Li v |= MLX5_ETH_WQE_L4_INNER_CSUM; 2445f8ba81cSXueming Li } else { 2455f8ba81cSXueming Li /* No tunnel. */ 2465f8ba81cSXueming Li if (i & (1 << 4)) /* IP. */ 2475f8ba81cSXueming Li v |= MLX5_ETH_WQE_L3_CSUM; 2485f8ba81cSXueming Li if (i & (3 << 2 | 1 << 0)) /* L4 or TSO. */ 2495f8ba81cSXueming Li v |= MLX5_ETH_WQE_L4_CSUM; 2505f8ba81cSXueming Li } 2515f8ba81cSXueming Li mlx5_cksum_table[i] = v; 2525f8ba81cSXueming Li } 2535f8ba81cSXueming Li } 2545f8ba81cSXueming Li 2555f8ba81cSXueming Li /** 2565f8ba81cSXueming Li * Build a table to translate packet type of mbuf to SWP type of Verbs. 2575f8ba81cSXueming Li */ 2585f8ba81cSXueming Li void 2595f8ba81cSXueming Li mlx5_set_swp_types_table(void) 2605f8ba81cSXueming Li { 2615f8ba81cSXueming Li unsigned int i; 2625f8ba81cSXueming Li uint8_t v; 2635f8ba81cSXueming Li 2645f8ba81cSXueming Li /* 2655f8ba81cSXueming Li * The index should have: 2665f8ba81cSXueming Li * bit[0:1] = PKT_TX_L4_MASK 2675f8ba81cSXueming Li * bit[4] = PKT_TX_IPV6 2685f8ba81cSXueming Li * bit[8] = PKT_TX_OUTER_IPV6 2695f8ba81cSXueming Li * bit[9] = PKT_TX_OUTER_UDP 2705f8ba81cSXueming Li */ 2715f8ba81cSXueming Li for (i = 0; i < RTE_DIM(mlx5_swp_types_table); ++i) { 2725f8ba81cSXueming Li v = 0; 2735f8ba81cSXueming Li if (i & (1 << 8)) 2745f8ba81cSXueming Li v |= MLX5_ETH_WQE_L3_OUTER_IPV6; 2755f8ba81cSXueming Li if (i & (1 << 9)) 2765f8ba81cSXueming Li v |= MLX5_ETH_WQE_L4_OUTER_UDP; 2775f8ba81cSXueming Li if (i & (1 << 4)) 2785f8ba81cSXueming Li v |= MLX5_ETH_WQE_L3_INNER_IPV6; 2795f8ba81cSXueming Li if ((i & 3) == (PKT_TX_UDP_CKSUM >> 52)) 2805f8ba81cSXueming Li v |= MLX5_ETH_WQE_L4_INNER_UDP; 2815f8ba81cSXueming Li mlx5_swp_types_table[i] = v; 2825f8ba81cSXueming Li } 2835f8ba81cSXueming Li } 2845f8ba81cSXueming Li 2855f8ba81cSXueming Li /** 2866ce84bd8SYongseok Koh * Return the size of tailroom of WQ. 2876ce84bd8SYongseok Koh * 2886ce84bd8SYongseok Koh * @param txq 2896ce84bd8SYongseok Koh * Pointer to TX queue structure. 2906ce84bd8SYongseok Koh * @param addr 2916ce84bd8SYongseok Koh * Pointer to tail of WQ. 2926ce84bd8SYongseok Koh * 2936ce84bd8SYongseok Koh * @return 2946ce84bd8SYongseok Koh * Size of tailroom. 2956ce84bd8SYongseok Koh */ 2966ce84bd8SYongseok Koh static inline size_t 297991b04f6SNélio Laranjeiro tx_mlx5_wq_tailroom(struct mlx5_txq_data *txq, void *addr) 2986ce84bd8SYongseok Koh { 2996ce84bd8SYongseok Koh size_t tailroom; 3006ce84bd8SYongseok Koh tailroom = (uintptr_t)(txq->wqes) + 3016ce84bd8SYongseok Koh (1 << txq->wqe_n) * MLX5_WQE_SIZE - 3026ce84bd8SYongseok Koh (uintptr_t)addr; 3036ce84bd8SYongseok Koh return tailroom; 3046ce84bd8SYongseok Koh } 3056ce84bd8SYongseok Koh 3066ce84bd8SYongseok Koh /** 3076ce84bd8SYongseok Koh * Copy data to tailroom of circular queue. 3086ce84bd8SYongseok Koh * 3096ce84bd8SYongseok Koh * @param dst 3106ce84bd8SYongseok Koh * Pointer to destination. 3116ce84bd8SYongseok Koh * @param src 3126ce84bd8SYongseok Koh * Pointer to source. 3136ce84bd8SYongseok Koh * @param n 3146ce84bd8SYongseok Koh * Number of bytes to copy. 3156ce84bd8SYongseok Koh * @param base 3166ce84bd8SYongseok Koh * Pointer to head of queue. 3176ce84bd8SYongseok Koh * @param tailroom 3186ce84bd8SYongseok Koh * Size of tailroom from dst. 3196ce84bd8SYongseok Koh * 3206ce84bd8SYongseok Koh * @return 3216ce84bd8SYongseok Koh * Pointer after copied data. 3226ce84bd8SYongseok Koh */ 3236ce84bd8SYongseok Koh static inline void * 3246ce84bd8SYongseok Koh mlx5_copy_to_wq(void *dst, const void *src, size_t n, 3256ce84bd8SYongseok Koh void *base, size_t tailroom) 3266ce84bd8SYongseok Koh { 3276ce84bd8SYongseok Koh void *ret; 3286ce84bd8SYongseok Koh 3296ce84bd8SYongseok Koh if (n > tailroom) { 3306ce84bd8SYongseok Koh rte_memcpy(dst, src, tailroom); 3316ce84bd8SYongseok Koh rte_memcpy(base, (void *)((uintptr_t)src + tailroom), 3326ce84bd8SYongseok Koh n - tailroom); 3336ce84bd8SYongseok Koh ret = (uint8_t *)base + n - tailroom; 3346ce84bd8SYongseok Koh } else { 3356ce84bd8SYongseok Koh rte_memcpy(dst, src, n); 3366ce84bd8SYongseok Koh ret = (n == tailroom) ? base : (uint8_t *)dst + n; 3376ce84bd8SYongseok Koh } 3386ce84bd8SYongseok Koh return ret; 3396ce84bd8SYongseok Koh } 3406ce84bd8SYongseok Koh 3416ce84bd8SYongseok Koh /** 342593f472cSXueming Li * Inline TSO headers into WQE. 343593f472cSXueming Li * 344593f472cSXueming Li * @return 345593f472cSXueming Li * 0 on success, negative errno value on failure. 346593f472cSXueming Li */ 347593f472cSXueming Li static int 348593f472cSXueming Li inline_tso(struct mlx5_txq_data *txq, struct rte_mbuf *buf, 349593f472cSXueming Li uint32_t *length, 350593f472cSXueming Li uintptr_t *addr, 351593f472cSXueming Li uint16_t *pkt_inline_sz, 352593f472cSXueming Li uint8_t **raw, 353593f472cSXueming Li uint16_t *max_wqe, 354593f472cSXueming Li uint16_t *tso_segsz, 355593f472cSXueming Li uint16_t *tso_header_sz) 356593f472cSXueming Li { 357593f472cSXueming Li uintptr_t end = (uintptr_t)(((uintptr_t)txq->wqes) + 358593f472cSXueming Li (1 << txq->wqe_n) * MLX5_WQE_SIZE); 359593f472cSXueming Li unsigned int copy_b; 360593f472cSXueming Li uint8_t vlan_sz = (buf->ol_flags & PKT_TX_VLAN_PKT) ? 4 : 0; 3615f8ba81cSXueming Li const uint8_t tunneled = txq->tunnel_en && (buf->ol_flags & 3625f8ba81cSXueming Li PKT_TX_TUNNEL_MASK); 363593f472cSXueming Li uint16_t n_wqe; 364593f472cSXueming Li 365593f472cSXueming Li *tso_segsz = buf->tso_segsz; 366593f472cSXueming Li *tso_header_sz = buf->l2_len + vlan_sz + buf->l3_len + buf->l4_len; 367593f472cSXueming Li if (unlikely(*tso_segsz == 0 || *tso_header_sz == 0)) { 368593f472cSXueming Li txq->stats.oerrors++; 369593f472cSXueming Li return -EINVAL; 370593f472cSXueming Li } 3715f8ba81cSXueming Li if (tunneled) 372593f472cSXueming Li *tso_header_sz += buf->outer_l2_len + buf->outer_l3_len; 3735f8ba81cSXueming Li /* First seg must contain all TSO headers. */ 3745f8ba81cSXueming Li if (unlikely(*tso_header_sz > MLX5_MAX_TSO_HEADER) || 3755f8ba81cSXueming Li *tso_header_sz > DATA_LEN(buf)) { 376593f472cSXueming Li txq->stats.oerrors++; 377593f472cSXueming Li return -EINVAL; 378593f472cSXueming Li } 379593f472cSXueming Li copy_b = *tso_header_sz - *pkt_inline_sz; 380593f472cSXueming Li if (!copy_b || ((end - (uintptr_t)*raw) < copy_b)) 381593f472cSXueming Li return -EAGAIN; 382593f472cSXueming Li n_wqe = (MLX5_WQE_DS(copy_b) - 1 + 3) / 4; 383593f472cSXueming Li if (unlikely(*max_wqe < n_wqe)) 384593f472cSXueming Li return -EINVAL; 385593f472cSXueming Li *max_wqe -= n_wqe; 386593f472cSXueming Li rte_memcpy((void *)*raw, (void *)*addr, copy_b); 387593f472cSXueming Li *length -= copy_b; 388593f472cSXueming Li *addr += copy_b; 389593f472cSXueming Li copy_b = MLX5_WQE_DS(copy_b) * MLX5_WQE_DWORD_SIZE; 390593f472cSXueming Li *pkt_inline_sz += copy_b; 391593f472cSXueming Li *raw += copy_b; 392593f472cSXueming Li return 0; 393593f472cSXueming Li } 394593f472cSXueming Li 395593f472cSXueming Li /** 3968788fec1SOlivier Matz * DPDK callback to check the status of a tx descriptor. 3978788fec1SOlivier Matz * 3988788fec1SOlivier Matz * @param tx_queue 3998788fec1SOlivier Matz * The tx queue. 4008788fec1SOlivier Matz * @param[in] offset 4018788fec1SOlivier Matz * The index of the descriptor in the ring. 4028788fec1SOlivier Matz * 4038788fec1SOlivier Matz * @return 4048788fec1SOlivier Matz * The status of the tx descriptor. 4058788fec1SOlivier Matz */ 4068788fec1SOlivier Matz int 4078788fec1SOlivier Matz mlx5_tx_descriptor_status(void *tx_queue, uint16_t offset) 4088788fec1SOlivier Matz { 409991b04f6SNélio Laranjeiro struct mlx5_txq_data *txq = tx_queue; 4108c819a69SYongseok Koh uint16_t used; 4118788fec1SOlivier Matz 4126cb559d6SYongseok Koh mlx5_tx_complete(txq); 4138c819a69SYongseok Koh used = txq->elts_head - txq->elts_tail; 4148788fec1SOlivier Matz if (offset < used) 4158788fec1SOlivier Matz return RTE_ETH_TX_DESC_FULL; 4168788fec1SOlivier Matz return RTE_ETH_TX_DESC_DONE; 4178788fec1SOlivier Matz } 4188788fec1SOlivier Matz 4198788fec1SOlivier Matz /** 4208788fec1SOlivier Matz * DPDK callback to check the status of a rx descriptor. 4218788fec1SOlivier Matz * 4228788fec1SOlivier Matz * @param rx_queue 4238788fec1SOlivier Matz * The rx queue. 4248788fec1SOlivier Matz * @param[in] offset 4258788fec1SOlivier Matz * The index of the descriptor in the ring. 4268788fec1SOlivier Matz * 4278788fec1SOlivier Matz * @return 4288788fec1SOlivier Matz * The status of the tx descriptor. 4298788fec1SOlivier Matz */ 4308788fec1SOlivier Matz int 4318788fec1SOlivier Matz mlx5_rx_descriptor_status(void *rx_queue, uint16_t offset) 4328788fec1SOlivier Matz { 43378142aacSNélio Laranjeiro struct mlx5_rxq_data *rxq = rx_queue; 4348788fec1SOlivier Matz struct rxq_zip *zip = &rxq->zip; 4358788fec1SOlivier Matz volatile struct mlx5_cqe *cqe; 4368788fec1SOlivier Matz const unsigned int cqe_n = (1 << rxq->cqe_n); 4378788fec1SOlivier Matz const unsigned int cqe_cnt = cqe_n - 1; 4388788fec1SOlivier Matz unsigned int cq_ci; 4398788fec1SOlivier Matz unsigned int used; 4408788fec1SOlivier Matz 4418788fec1SOlivier Matz /* if we are processing a compressed cqe */ 4428788fec1SOlivier Matz if (zip->ai) { 4438788fec1SOlivier Matz used = zip->cqe_cnt - zip->ca; 4448788fec1SOlivier Matz cq_ci = zip->cq_ci; 4458788fec1SOlivier Matz } else { 4468788fec1SOlivier Matz used = 0; 4478788fec1SOlivier Matz cq_ci = rxq->cq_ci; 4488788fec1SOlivier Matz } 4498788fec1SOlivier Matz cqe = &(*rxq->cqes)[cq_ci & cqe_cnt]; 4508788fec1SOlivier Matz while (check_cqe(cqe, cqe_n, cq_ci) == 0) { 4518788fec1SOlivier Matz int8_t op_own; 4528788fec1SOlivier Matz unsigned int n; 4538788fec1SOlivier Matz 4548788fec1SOlivier Matz op_own = cqe->op_own; 4558788fec1SOlivier Matz if (MLX5_CQE_FORMAT(op_own) == MLX5_COMPRESSED) 4566b30a6a8SShachar Beiser n = rte_be_to_cpu_32(cqe->byte_cnt); 4578788fec1SOlivier Matz else 4588788fec1SOlivier Matz n = 1; 4598788fec1SOlivier Matz cq_ci += n; 4608788fec1SOlivier Matz used += n; 4618788fec1SOlivier Matz cqe = &(*rxq->cqes)[cq_ci & cqe_cnt]; 4628788fec1SOlivier Matz } 4638788fec1SOlivier Matz used = RTE_MIN(used, (1U << rxq->elts_n) - 1); 4648788fec1SOlivier Matz if (offset < used) 4658788fec1SOlivier Matz return RTE_ETH_RX_DESC_DONE; 4668788fec1SOlivier Matz return RTE_ETH_RX_DESC_AVAIL; 4678788fec1SOlivier Matz } 4688788fec1SOlivier Matz 4698788fec1SOlivier Matz /** 4702e22920bSAdrien Mazarguil * DPDK callback for TX. 4712e22920bSAdrien Mazarguil * 4722e22920bSAdrien Mazarguil * @param dpdk_txq 4732e22920bSAdrien Mazarguil * Generic pointer to TX queue structure. 4742e22920bSAdrien Mazarguil * @param[in] pkts 4752e22920bSAdrien Mazarguil * Packets to transmit. 4762e22920bSAdrien Mazarguil * @param pkts_n 4772e22920bSAdrien Mazarguil * Number of packets in array. 4782e22920bSAdrien Mazarguil * 4792e22920bSAdrien Mazarguil * @return 4802e22920bSAdrien Mazarguil * Number of packets successfully transmitted (<= pkts_n). 4812e22920bSAdrien Mazarguil */ 4822e22920bSAdrien Mazarguil uint16_t 4832e22920bSAdrien Mazarguil mlx5_tx_burst(void *dpdk_txq, struct rte_mbuf **pkts, uint16_t pkts_n) 4842e22920bSAdrien Mazarguil { 485991b04f6SNélio Laranjeiro struct mlx5_txq_data *txq = (struct mlx5_txq_data *)dpdk_txq; 4861d88ba17SNélio Laranjeiro uint16_t elts_head = txq->elts_head; 4878c819a69SYongseok Koh const uint16_t elts_n = 1 << txq->elts_n; 4888c819a69SYongseok Koh const uint16_t elts_m = elts_n - 1; 489c3d62cc9SAdrien Mazarguil unsigned int i = 0; 490a5bf6af9SAdrien Mazarguil unsigned int j = 0; 4913f13f8c2SShahaf Shuler unsigned int k = 0; 4928c819a69SYongseok Koh uint16_t max_elts; 493f04f1d51SNélio Laranjeiro uint16_t max_wqe; 494c305090bSAdrien Mazarguil unsigned int comp; 495ac180a21SYongseok Koh volatile struct mlx5_wqe_ctrl *last_wqe = NULL; 4966579c27cSNélio Laranjeiro unsigned int segs_n = 0; 49727a6b2d6SNélio Laranjeiro const unsigned int max_inline = txq->max_inline; 4982e22920bSAdrien Mazarguil 4991d88ba17SNélio Laranjeiro if (unlikely(!pkts_n)) 5001d88ba17SNélio Laranjeiro return 0; 5015e1d11ecSNelio Laranjeiro /* Prefetch first packet cacheline. */ 502c3d62cc9SAdrien Mazarguil rte_prefetch0(*pkts); 5031d88ba17SNélio Laranjeiro /* Start processing. */ 5046cb559d6SYongseok Koh mlx5_tx_complete(txq); 5058c819a69SYongseok Koh max_elts = (elts_n - (elts_head - txq->elts_tail)); 5062eefbec5SYongseok Koh /* A CQE slot must always be available. */ 5072eefbec5SYongseok Koh assert((1u << txq->cqe_n) - (txq->cq_pi - txq->cq_ci)); 508f04f1d51SNélio Laranjeiro max_wqe = (1u << txq->wqe_n) - (txq->wqe_ci - txq->wqe_pi); 509f04f1d51SNélio Laranjeiro if (unlikely(!max_wqe)) 510f04f1d51SNélio Laranjeiro return 0; 511c3d62cc9SAdrien Mazarguil do { 5125f8ba81cSXueming Li struct rte_mbuf *buf = *pkts; /* First_seg. */ 5133bbae1ebSNélio Laranjeiro uint8_t *raw; 5143bbae1ebSNélio Laranjeiro volatile struct mlx5_wqe_v *wqe = NULL; 5159a7fa9f7SNélio Laranjeiro volatile rte_v128u32_t *dseg = NULL; 516573f54afSNélio Laranjeiro uint32_t length; 5178688b2f8SNélio Laranjeiro unsigned int ds = 0; 518ac180a21SYongseok Koh unsigned int sg = 0; /* counter of additional segs attached. */ 5196579c27cSNélio Laranjeiro uintptr_t addr; 5200d637a34SNélio Laranjeiro uint16_t pkt_inline_sz = MLX5_WQE_DWORD_SIZE + 2; 5213f13f8c2SShahaf Shuler uint16_t tso_header_sz = 0; 522eef822ddSNélio Laranjeiro uint16_t ehdr; 5234aa15eb1SNélio Laranjeiro uint8_t cs_flags; 5245f8ba81cSXueming Li uint8_t tso = txq->tso_en && (buf->ol_flags & PKT_TX_TCP_SEG); 5255f8ba81cSXueming Li uint8_t is_vlan = !!(buf->ol_flags & PKT_TX_VLAN_PKT); 5265f8ba81cSXueming Li uint32_t swp_offsets = 0; 5275f8ba81cSXueming Li uint8_t swp_types = 0; 52883daf156SShahaf Shuler uint16_t tso_segsz = 0; 5296579c27cSNélio Laranjeiro #ifdef MLX5_PMD_SOFT_COUNTERS 5306579c27cSNélio Laranjeiro uint32_t total_length = 0; 5316579c27cSNélio Laranjeiro #endif 532593f472cSXueming Li int ret; 5332e22920bSAdrien Mazarguil 5346579c27cSNélio Laranjeiro segs_n = buf->nb_segs; 535c3d62cc9SAdrien Mazarguil /* 536c3d62cc9SAdrien Mazarguil * Make sure there is enough room to store this packet and 537c3d62cc9SAdrien Mazarguil * that one ring entry remains unused. 538c3d62cc9SAdrien Mazarguil */ 539a5bf6af9SAdrien Mazarguil assert(segs_n); 5408c819a69SYongseok Koh if (max_elts < segs_n) 541c3d62cc9SAdrien Mazarguil break; 5428c819a69SYongseok Koh max_elts -= segs_n; 543f895536bSYongseok Koh sg = --segs_n; 544f04f1d51SNélio Laranjeiro if (unlikely(--max_wqe == 0)) 545f04f1d51SNélio Laranjeiro break; 5469a7fa9f7SNélio Laranjeiro wqe = (volatile struct mlx5_wqe_v *) 547fdcb0f53SNélio Laranjeiro tx_mlx5_wqe(txq, txq->wqe_ci); 548fdcb0f53SNélio Laranjeiro rte_prefetch0(tx_mlx5_wqe(txq, txq->wqe_ci + 1)); 5493730e6c6SYongseok Koh if (pkts_n - i > 1) 5503730e6c6SYongseok Koh rte_prefetch0(*(pkts + 1)); 5516579c27cSNélio Laranjeiro addr = rte_pktmbuf_mtod(buf, uintptr_t); 5522e22920bSAdrien Mazarguil length = DATA_LEN(buf); 553eef822ddSNélio Laranjeiro ehdr = (((uint8_t *)addr)[1] << 8) | 554eef822ddSNélio Laranjeiro ((uint8_t *)addr)[0]; 5556579c27cSNélio Laranjeiro #ifdef MLX5_PMD_SOFT_COUNTERS 5566579c27cSNélio Laranjeiro total_length = length; 5576579c27cSNélio Laranjeiro #endif 55824c14430SShahaf Shuler if (length < (MLX5_WQE_DWORD_SIZE + 2)) { 55924c14430SShahaf Shuler txq->stats.oerrors++; 560959be52eSNélio Laranjeiro break; 56124c14430SShahaf Shuler } 5622e22920bSAdrien Mazarguil /* Update element. */ 5638c819a69SYongseok Koh (*txq->elts)[elts_head & elts_m] = buf; 5645e1d11ecSNelio Laranjeiro /* Prefetch next buffer data. */ 5653730e6c6SYongseok Koh if (pkts_n - i > 1) 5663730e6c6SYongseok Koh rte_prefetch0( 5673730e6c6SYongseok Koh rte_pktmbuf_mtod(*(pkts + 1), volatile void *)); 5685f8ba81cSXueming Li cs_flags = txq_ol_cksum_to_cs(buf); 5695f8ba81cSXueming Li txq_mbuf_to_swp(txq, buf, tso, is_vlan, 5705f8ba81cSXueming Li (uint8_t *)&swp_offsets, &swp_types); 571b8fe952eSNélio Laranjeiro raw = ((uint8_t *)(uintptr_t)wqe) + 2 * MLX5_WQE_DWORD_SIZE; 5726579c27cSNélio Laranjeiro /* Replace the Ethernet type by the VLAN if necessary. */ 5735f8ba81cSXueming Li if (is_vlan) { 5746b30a6a8SShachar Beiser uint32_t vlan = rte_cpu_to_be_32(0x81000000 | 5756b30a6a8SShachar Beiser buf->vlan_tci); 5760d637a34SNélio Laranjeiro unsigned int len = 2 * ETHER_ADDR_LEN - 2; 5776579c27cSNélio Laranjeiro 5780d637a34SNélio Laranjeiro addr += 2; 5790d637a34SNélio Laranjeiro length -= 2; 5800d637a34SNélio Laranjeiro /* Copy Destination and source mac address. */ 5810d637a34SNélio Laranjeiro memcpy((uint8_t *)raw, ((uint8_t *)addr), len); 5820d637a34SNélio Laranjeiro /* Copy VLAN. */ 5830d637a34SNélio Laranjeiro memcpy((uint8_t *)raw + len, &vlan, sizeof(vlan)); 5840d637a34SNélio Laranjeiro /* Copy missing two bytes to end the DSeg. */ 5850d637a34SNélio Laranjeiro memcpy((uint8_t *)raw + len + sizeof(vlan), 5860d637a34SNélio Laranjeiro ((uint8_t *)addr) + len, 2); 5870d637a34SNélio Laranjeiro addr += len + 2; 5880d637a34SNélio Laranjeiro length -= (len + 2); 5890d637a34SNélio Laranjeiro } else { 5900d637a34SNélio Laranjeiro memcpy((uint8_t *)raw, ((uint8_t *)addr) + 2, 5910d637a34SNélio Laranjeiro MLX5_WQE_DWORD_SIZE); 5920d637a34SNélio Laranjeiro length -= pkt_inline_sz; 5930d637a34SNélio Laranjeiro addr += pkt_inline_sz; 5946579c27cSNélio Laranjeiro } 595d8292497SYongseok Koh raw += MLX5_WQE_DWORD_SIZE; 5963f13f8c2SShahaf Shuler if (tso) { 5975f8ba81cSXueming Li ret = inline_tso(txq, buf, &length, 598593f472cSXueming Li &addr, &pkt_inline_sz, 599593f472cSXueming Li &raw, &max_wqe, 600593f472cSXueming Li &tso_segsz, &tso_header_sz); 601593f472cSXueming Li if (ret == -EINVAL) { 60296fc8d65SShahaf Shuler break; 603593f472cSXueming Li } else if (ret == -EAGAIN) { 6043f13f8c2SShahaf Shuler /* NOP WQE. */ 6053f13f8c2SShahaf Shuler wqe->ctrl = (rte_v128u32_t){ 60636aa55eaSYongseok Koh rte_cpu_to_be_32(txq->wqe_ci << 8), 60736aa55eaSYongseok Koh rte_cpu_to_be_32(txq->qp_num_8s | 1), 6083f13f8c2SShahaf Shuler 0, 6093f13f8c2SShahaf Shuler 0, 6103f13f8c2SShahaf Shuler }; 6113f13f8c2SShahaf Shuler ds = 1; 612cb98affeSThierry Herbelot #ifdef MLX5_PMD_SOFT_COUNTERS 6133f13f8c2SShahaf Shuler total_length = 0; 614cb98affeSThierry Herbelot #endif 6153f13f8c2SShahaf Shuler k++; 6163f13f8c2SShahaf Shuler goto next_wqe; 6173f13f8c2SShahaf Shuler } 6183f13f8c2SShahaf Shuler } 6196579c27cSNélio Laranjeiro /* Inline if enough room. */ 62027a6b2d6SNélio Laranjeiro if (max_inline || tso) { 621f895536bSYongseok Koh uint32_t inl = 0; 622fdcb0f53SNélio Laranjeiro uintptr_t end = (uintptr_t) 623fdcb0f53SNélio Laranjeiro (((uintptr_t)txq->wqes) + 624fdcb0f53SNélio Laranjeiro (1 << txq->wqe_n) * MLX5_WQE_SIZE); 625ab76eab3SYongseok Koh unsigned int inline_room = max_inline * 6268fcd6c2cSNélio Laranjeiro RTE_CACHE_LINE_SIZE - 627d8292497SYongseok Koh (pkt_inline_sz - 2) - 628d8292497SYongseok Koh !!tso * sizeof(inl); 629f895536bSYongseok Koh uintptr_t addr_end; 630f895536bSYongseok Koh unsigned int copy_b; 6316579c27cSNélio Laranjeiro 632f895536bSYongseok Koh pkt_inline: 633f895536bSYongseok Koh addr_end = RTE_ALIGN_FLOOR(addr + inline_room, 634f895536bSYongseok Koh RTE_CACHE_LINE_SIZE); 635f895536bSYongseok Koh copy_b = (addr_end > addr) ? 636f895536bSYongseok Koh RTE_MIN((addr_end - addr), length) : 0; 6378fcd6c2cSNélio Laranjeiro if (copy_b && ((end - (uintptr_t)raw) > copy_b)) { 638f04f1d51SNélio Laranjeiro /* 639f04f1d51SNélio Laranjeiro * One Dseg remains in the current WQE. To 640f04f1d51SNélio Laranjeiro * keep the computation positive, it is 641f04f1d51SNélio Laranjeiro * removed after the bytes to Dseg conversion. 642f04f1d51SNélio Laranjeiro */ 6438fcd6c2cSNélio Laranjeiro uint16_t n = (MLX5_WQE_DS(copy_b) - 1 + 3) / 4; 6448fcd6c2cSNélio Laranjeiro 645f04f1d51SNélio Laranjeiro if (unlikely(max_wqe < n)) 646f04f1d51SNélio Laranjeiro break; 647f04f1d51SNélio Laranjeiro max_wqe -= n; 648f895536bSYongseok Koh if (tso && !inl) { 6496963ae8bSYongseok Koh inl = rte_cpu_to_be_32(copy_b | 6506b30a6a8SShachar Beiser MLX5_INLINE_SEG); 6513f13f8c2SShahaf Shuler rte_memcpy((void *)raw, 6523f13f8c2SShahaf Shuler (void *)&inl, sizeof(inl)); 6533f13f8c2SShahaf Shuler raw += sizeof(inl); 6543f13f8c2SShahaf Shuler pkt_inline_sz += sizeof(inl); 6553f13f8c2SShahaf Shuler } 6566579c27cSNélio Laranjeiro rte_memcpy((void *)raw, (void *)addr, copy_b); 6576579c27cSNélio Laranjeiro addr += copy_b; 6586579c27cSNélio Laranjeiro length -= copy_b; 6596579c27cSNélio Laranjeiro pkt_inline_sz += copy_b; 6606579c27cSNélio Laranjeiro } 6616579c27cSNélio Laranjeiro /* 662786b5c2dSShahaf Shuler * 2 DWORDs consumed by the WQE header + ETH segment + 6636579c27cSNélio Laranjeiro * the size of the inline part of the packet. 6646579c27cSNélio Laranjeiro */ 6656579c27cSNélio Laranjeiro ds = 2 + MLX5_WQE_DS(pkt_inline_sz - 2); 6666579c27cSNélio Laranjeiro if (length > 0) { 667f04f1d51SNélio Laranjeiro if (ds % (MLX5_WQE_SIZE / 668f04f1d51SNélio Laranjeiro MLX5_WQE_DWORD_SIZE) == 0) { 669f04f1d51SNélio Laranjeiro if (unlikely(--max_wqe == 0)) 670f04f1d51SNélio Laranjeiro break; 671f04f1d51SNélio Laranjeiro dseg = (volatile rte_v128u32_t *) 672f04f1d51SNélio Laranjeiro tx_mlx5_wqe(txq, txq->wqe_ci + 673f04f1d51SNélio Laranjeiro ds / 4); 674f04f1d51SNélio Laranjeiro } else { 6759a7fa9f7SNélio Laranjeiro dseg = (volatile rte_v128u32_t *) 6766579c27cSNélio Laranjeiro ((uintptr_t)wqe + 6776579c27cSNélio Laranjeiro (ds * MLX5_WQE_DWORD_SIZE)); 678f04f1d51SNélio Laranjeiro } 6796579c27cSNélio Laranjeiro goto use_dseg; 6806579c27cSNélio Laranjeiro } else if (!segs_n) { 6816579c27cSNélio Laranjeiro goto next_pkt; 6826579c27cSNélio Laranjeiro } else { 683f895536bSYongseok Koh raw += copy_b; 684f895536bSYongseok Koh inline_room -= copy_b; 685f895536bSYongseok Koh --segs_n; 686f895536bSYongseok Koh buf = buf->next; 687f895536bSYongseok Koh assert(buf); 688f895536bSYongseok Koh addr = rte_pktmbuf_mtod(buf, uintptr_t); 689f895536bSYongseok Koh length = DATA_LEN(buf); 690f895536bSYongseok Koh #ifdef MLX5_PMD_SOFT_COUNTERS 691f895536bSYongseok Koh total_length += length; 692f895536bSYongseok Koh #endif 693f895536bSYongseok Koh (*txq->elts)[++elts_head & elts_m] = buf; 694f895536bSYongseok Koh goto pkt_inline; 6956579c27cSNélio Laranjeiro } 6966579c27cSNélio Laranjeiro } else { 6976579c27cSNélio Laranjeiro /* 6986579c27cSNélio Laranjeiro * No inline has been done in the packet, only the 6996579c27cSNélio Laranjeiro * Ethernet Header as been stored. 7006579c27cSNélio Laranjeiro */ 7019a7fa9f7SNélio Laranjeiro dseg = (volatile rte_v128u32_t *) 7026579c27cSNélio Laranjeiro ((uintptr_t)wqe + (3 * MLX5_WQE_DWORD_SIZE)); 7036579c27cSNélio Laranjeiro ds = 3; 7046579c27cSNélio Laranjeiro use_dseg: 7056579c27cSNélio Laranjeiro /* Add the remaining packet as a simple ds. */ 706ebbb81ebSNélio Laranjeiro addr = rte_cpu_to_be_64(addr); 7079a7fa9f7SNélio Laranjeiro *dseg = (rte_v128u32_t){ 7086b30a6a8SShachar Beiser rte_cpu_to_be_32(length), 7096cb559d6SYongseok Koh mlx5_tx_mb2mr(txq, buf), 710ebbb81ebSNélio Laranjeiro addr, 711ebbb81ebSNélio Laranjeiro addr >> 32, 7126579c27cSNélio Laranjeiro }; 7136579c27cSNélio Laranjeiro ++ds; 7146579c27cSNélio Laranjeiro if (!segs_n) 7156579c27cSNélio Laranjeiro goto next_pkt; 7166579c27cSNélio Laranjeiro } 7176579c27cSNélio Laranjeiro next_seg: 7186579c27cSNélio Laranjeiro assert(buf); 7196579c27cSNélio Laranjeiro assert(ds); 7206579c27cSNélio Laranjeiro assert(wqe); 721a5bf6af9SAdrien Mazarguil /* 722a5bf6af9SAdrien Mazarguil * Spill on next WQE when the current one does not have 723a5bf6af9SAdrien Mazarguil * enough room left. Size of WQE must a be a multiple 724a5bf6af9SAdrien Mazarguil * of data segment size. 725a5bf6af9SAdrien Mazarguil */ 7268688b2f8SNélio Laranjeiro assert(!(MLX5_WQE_SIZE % MLX5_WQE_DWORD_SIZE)); 7276579c27cSNélio Laranjeiro if (!(ds % (MLX5_WQE_SIZE / MLX5_WQE_DWORD_SIZE))) { 728f04f1d51SNélio Laranjeiro if (unlikely(--max_wqe == 0)) 729f04f1d51SNélio Laranjeiro break; 7309a7fa9f7SNélio Laranjeiro dseg = (volatile rte_v128u32_t *) 731f04f1d51SNélio Laranjeiro tx_mlx5_wqe(txq, txq->wqe_ci + ds / 4); 732f04f1d51SNélio Laranjeiro rte_prefetch0(tx_mlx5_wqe(txq, 733f04f1d51SNélio Laranjeiro txq->wqe_ci + ds / 4 + 1)); 7346579c27cSNélio Laranjeiro } else { 735a5bf6af9SAdrien Mazarguil ++dseg; 7366579c27cSNélio Laranjeiro } 737a5bf6af9SAdrien Mazarguil ++ds; 738a5bf6af9SAdrien Mazarguil buf = buf->next; 739a5bf6af9SAdrien Mazarguil assert(buf); 7406579c27cSNélio Laranjeiro length = DATA_LEN(buf); 741a5bf6af9SAdrien Mazarguil #ifdef MLX5_PMD_SOFT_COUNTERS 7426579c27cSNélio Laranjeiro total_length += length; 743a5bf6af9SAdrien Mazarguil #endif 7446579c27cSNélio Laranjeiro /* Store segment information. */ 745ebbb81ebSNélio Laranjeiro addr = rte_cpu_to_be_64(rte_pktmbuf_mtod(buf, uintptr_t)); 7469a7fa9f7SNélio Laranjeiro *dseg = (rte_v128u32_t){ 7476b30a6a8SShachar Beiser rte_cpu_to_be_32(length), 7486cb559d6SYongseok Koh mlx5_tx_mb2mr(txq, buf), 749ebbb81ebSNélio Laranjeiro addr, 750ebbb81ebSNélio Laranjeiro addr >> 32, 7516579c27cSNélio Laranjeiro }; 7528c819a69SYongseok Koh (*txq->elts)[++elts_head & elts_m] = buf; 753f895536bSYongseok Koh if (--segs_n) 7546579c27cSNélio Laranjeiro goto next_seg; 7556579c27cSNélio Laranjeiro next_pkt: 756883ce172SShahaf Shuler if (ds > MLX5_DSEG_MAX) { 757883ce172SShahaf Shuler txq->stats.oerrors++; 758883ce172SShahaf Shuler break; 759883ce172SShahaf Shuler } 7608c819a69SYongseok Koh ++elts_head; 7613730e6c6SYongseok Koh ++pkts; 7626579c27cSNélio Laranjeiro ++i; 763f895536bSYongseok Koh j += sg; 764b8fe952eSNélio Laranjeiro /* Initialize known and common part of the WQE structure. */ 7653f13f8c2SShahaf Shuler if (tso) { 7663f13f8c2SShahaf Shuler wqe->ctrl = (rte_v128u32_t){ 7676b30a6a8SShachar Beiser rte_cpu_to_be_32((txq->wqe_ci << 8) | 7686b30a6a8SShachar Beiser MLX5_OPCODE_TSO), 7696b30a6a8SShachar Beiser rte_cpu_to_be_32(txq->qp_num_8s | ds), 7703f13f8c2SShahaf Shuler 0, 7713f13f8c2SShahaf Shuler 0, 7723f13f8c2SShahaf Shuler }; 7733f13f8c2SShahaf Shuler wqe->eseg = (rte_v128u32_t){ 7745f8ba81cSXueming Li swp_offsets, 7755f8ba81cSXueming Li cs_flags | (swp_types << 8) | 7765f8ba81cSXueming Li (rte_cpu_to_be_16(tso_segsz) << 16), 7773f13f8c2SShahaf Shuler 0, 7786b30a6a8SShachar Beiser (ehdr << 16) | rte_cpu_to_be_16(tso_header_sz), 7793f13f8c2SShahaf Shuler }; 7803f13f8c2SShahaf Shuler } else { 7819a7fa9f7SNélio Laranjeiro wqe->ctrl = (rte_v128u32_t){ 7826b30a6a8SShachar Beiser rte_cpu_to_be_32((txq->wqe_ci << 8) | 7836b30a6a8SShachar Beiser MLX5_OPCODE_SEND), 7846b30a6a8SShachar Beiser rte_cpu_to_be_32(txq->qp_num_8s | ds), 7859a7fa9f7SNélio Laranjeiro 0, 7869a7fa9f7SNélio Laranjeiro 0, 7879a7fa9f7SNélio Laranjeiro }; 7889a7fa9f7SNélio Laranjeiro wqe->eseg = (rte_v128u32_t){ 7895f8ba81cSXueming Li swp_offsets, 7905f8ba81cSXueming Li cs_flags | (swp_types << 8), 7919a7fa9f7SNélio Laranjeiro 0, 7926b30a6a8SShachar Beiser (ehdr << 16) | rte_cpu_to_be_16(pkt_inline_sz), 7939a7fa9f7SNélio Laranjeiro }; 7943f13f8c2SShahaf Shuler } 7953f13f8c2SShahaf Shuler next_wqe: 7966579c27cSNélio Laranjeiro txq->wqe_ci += (ds + 3) / 4; 797ac180a21SYongseok Koh /* Save the last successful WQE for completion request */ 798ac180a21SYongseok Koh last_wqe = (volatile struct mlx5_wqe_ctrl *)wqe; 79987011737SAdrien Mazarguil #ifdef MLX5_PMD_SOFT_COUNTERS 800573f54afSNélio Laranjeiro /* Increment sent bytes counter. */ 8016579c27cSNélio Laranjeiro txq->stats.obytes += total_length; 80287011737SAdrien Mazarguil #endif 8033730e6c6SYongseok Koh } while (i < pkts_n); 8042e22920bSAdrien Mazarguil /* Take a shortcut if nothing must be sent. */ 8053f13f8c2SShahaf Shuler if (unlikely((i + k) == 0)) 8062e22920bSAdrien Mazarguil return 0; 8078c819a69SYongseok Koh txq->elts_head += (i + j); 808c305090bSAdrien Mazarguil /* Check whether completion threshold has been reached. */ 8093f13f8c2SShahaf Shuler comp = txq->elts_comp + i + j + k; 810c305090bSAdrien Mazarguil if (comp >= MLX5_TX_COMP_THRESH) { 811c305090bSAdrien Mazarguil /* Request completion on last WQE. */ 8126b30a6a8SShachar Beiser last_wqe->ctrl2 = rte_cpu_to_be_32(8); 813c305090bSAdrien Mazarguil /* Save elts_head in unused "immediate" field of WQE. */ 814ac180a21SYongseok Koh last_wqe->ctrl3 = txq->elts_head; 815c305090bSAdrien Mazarguil txq->elts_comp = 0; 8162eefbec5SYongseok Koh #ifndef NDEBUG 8172eefbec5SYongseok Koh ++txq->cq_pi; 8182eefbec5SYongseok Koh #endif 819c305090bSAdrien Mazarguil } else { 820c305090bSAdrien Mazarguil txq->elts_comp = comp; 821c305090bSAdrien Mazarguil } 82287011737SAdrien Mazarguil #ifdef MLX5_PMD_SOFT_COUNTERS 82387011737SAdrien Mazarguil /* Increment sent packets counter. */ 82487011737SAdrien Mazarguil txq->stats.opackets += i; 82587011737SAdrien Mazarguil #endif 8262e22920bSAdrien Mazarguil /* Ring QP doorbell. */ 827ac180a21SYongseok Koh mlx5_tx_dbrec(txq, (volatile struct mlx5_wqe *)last_wqe); 8282e22920bSAdrien Mazarguil return i; 8292e22920bSAdrien Mazarguil } 8302e22920bSAdrien Mazarguil 8312e22920bSAdrien Mazarguil /** 832230189d9SNélio Laranjeiro * Open a MPW session. 833230189d9SNélio Laranjeiro * 834230189d9SNélio Laranjeiro * @param txq 835230189d9SNélio Laranjeiro * Pointer to TX queue structure. 836230189d9SNélio Laranjeiro * @param mpw 837230189d9SNélio Laranjeiro * Pointer to MPW session structure. 838230189d9SNélio Laranjeiro * @param length 839230189d9SNélio Laranjeiro * Packet length. 840230189d9SNélio Laranjeiro */ 841230189d9SNélio Laranjeiro static inline void 842991b04f6SNélio Laranjeiro mlx5_mpw_new(struct mlx5_txq_data *txq, struct mlx5_mpw *mpw, uint32_t length) 843230189d9SNélio Laranjeiro { 844a821d09dSNélio Laranjeiro uint16_t idx = txq->wqe_ci & ((1 << txq->wqe_n) - 1); 845230189d9SNélio Laranjeiro volatile struct mlx5_wqe_data_seg (*dseg)[MLX5_MPW_DSEG_MAX] = 846230189d9SNélio Laranjeiro (volatile struct mlx5_wqe_data_seg (*)[]) 847fdcb0f53SNélio Laranjeiro tx_mlx5_wqe(txq, idx + 1); 848230189d9SNélio Laranjeiro 849230189d9SNélio Laranjeiro mpw->state = MLX5_MPW_STATE_OPENED; 850230189d9SNélio Laranjeiro mpw->pkts_n = 0; 851230189d9SNélio Laranjeiro mpw->len = length; 852230189d9SNélio Laranjeiro mpw->total_len = 0; 853fdcb0f53SNélio Laranjeiro mpw->wqe = (volatile struct mlx5_wqe *)tx_mlx5_wqe(txq, idx); 8546b30a6a8SShachar Beiser mpw->wqe->eseg.mss = rte_cpu_to_be_16(length); 8558688b2f8SNélio Laranjeiro mpw->wqe->eseg.inline_hdr_sz = 0; 8568688b2f8SNélio Laranjeiro mpw->wqe->eseg.rsvd0 = 0; 8578688b2f8SNélio Laranjeiro mpw->wqe->eseg.rsvd1 = 0; 8588688b2f8SNélio Laranjeiro mpw->wqe->eseg.rsvd2 = 0; 8596b30a6a8SShachar Beiser mpw->wqe->ctrl[0] = rte_cpu_to_be_32((MLX5_OPC_MOD_MPW << 24) | 8606b30a6a8SShachar Beiser (txq->wqe_ci << 8) | 8616b30a6a8SShachar Beiser MLX5_OPCODE_TSO); 8628688b2f8SNélio Laranjeiro mpw->wqe->ctrl[2] = 0; 8638688b2f8SNélio Laranjeiro mpw->wqe->ctrl[3] = 0; 8648688b2f8SNélio Laranjeiro mpw->data.dseg[0] = (volatile struct mlx5_wqe_data_seg *) 8658688b2f8SNélio Laranjeiro (((uintptr_t)mpw->wqe) + (2 * MLX5_WQE_DWORD_SIZE)); 8668688b2f8SNélio Laranjeiro mpw->data.dseg[1] = (volatile struct mlx5_wqe_data_seg *) 8678688b2f8SNélio Laranjeiro (((uintptr_t)mpw->wqe) + (3 * MLX5_WQE_DWORD_SIZE)); 868230189d9SNélio Laranjeiro mpw->data.dseg[2] = &(*dseg)[0]; 869230189d9SNélio Laranjeiro mpw->data.dseg[3] = &(*dseg)[1]; 870230189d9SNélio Laranjeiro mpw->data.dseg[4] = &(*dseg)[2]; 871230189d9SNélio Laranjeiro } 872230189d9SNélio Laranjeiro 873230189d9SNélio Laranjeiro /** 874230189d9SNélio Laranjeiro * Close a MPW session. 875230189d9SNélio Laranjeiro * 876230189d9SNélio Laranjeiro * @param txq 877230189d9SNélio Laranjeiro * Pointer to TX queue structure. 878230189d9SNélio Laranjeiro * @param mpw 879230189d9SNélio Laranjeiro * Pointer to MPW session structure. 880230189d9SNélio Laranjeiro */ 881230189d9SNélio Laranjeiro static inline void 882991b04f6SNélio Laranjeiro mlx5_mpw_close(struct mlx5_txq_data *txq, struct mlx5_mpw *mpw) 883230189d9SNélio Laranjeiro { 884230189d9SNélio Laranjeiro unsigned int num = mpw->pkts_n; 885230189d9SNélio Laranjeiro 886230189d9SNélio Laranjeiro /* 887230189d9SNélio Laranjeiro * Store size in multiple of 16 bytes. Control and Ethernet segments 888230189d9SNélio Laranjeiro * count as 2. 889230189d9SNélio Laranjeiro */ 8906b30a6a8SShachar Beiser mpw->wqe->ctrl[1] = rte_cpu_to_be_32(txq->qp_num_8s | (2 + num)); 891230189d9SNélio Laranjeiro mpw->state = MLX5_MPW_STATE_CLOSED; 892230189d9SNélio Laranjeiro if (num < 3) 893230189d9SNélio Laranjeiro ++txq->wqe_ci; 894230189d9SNélio Laranjeiro else 895230189d9SNélio Laranjeiro txq->wqe_ci += 2; 896fdcb0f53SNélio Laranjeiro rte_prefetch0(tx_mlx5_wqe(txq, txq->wqe_ci)); 897fdcb0f53SNélio Laranjeiro rte_prefetch0(tx_mlx5_wqe(txq, txq->wqe_ci + 1)); 898230189d9SNélio Laranjeiro } 899230189d9SNélio Laranjeiro 900230189d9SNélio Laranjeiro /** 901230189d9SNélio Laranjeiro * DPDK callback for TX with MPW support. 902230189d9SNélio Laranjeiro * 903230189d9SNélio Laranjeiro * @param dpdk_txq 904230189d9SNélio Laranjeiro * Generic pointer to TX queue structure. 905230189d9SNélio Laranjeiro * @param[in] pkts 906230189d9SNélio Laranjeiro * Packets to transmit. 907230189d9SNélio Laranjeiro * @param pkts_n 908230189d9SNélio Laranjeiro * Number of packets in array. 909230189d9SNélio Laranjeiro * 910230189d9SNélio Laranjeiro * @return 911230189d9SNélio Laranjeiro * Number of packets successfully transmitted (<= pkts_n). 912230189d9SNélio Laranjeiro */ 913230189d9SNélio Laranjeiro uint16_t 914230189d9SNélio Laranjeiro mlx5_tx_burst_mpw(void *dpdk_txq, struct rte_mbuf **pkts, uint16_t pkts_n) 915230189d9SNélio Laranjeiro { 916991b04f6SNélio Laranjeiro struct mlx5_txq_data *txq = (struct mlx5_txq_data *)dpdk_txq; 917230189d9SNélio Laranjeiro uint16_t elts_head = txq->elts_head; 9188c819a69SYongseok Koh const uint16_t elts_n = 1 << txq->elts_n; 9198c819a69SYongseok Koh const uint16_t elts_m = elts_n - 1; 920c3d62cc9SAdrien Mazarguil unsigned int i = 0; 921a5bf6af9SAdrien Mazarguil unsigned int j = 0; 9228c819a69SYongseok Koh uint16_t max_elts; 923f04f1d51SNélio Laranjeiro uint16_t max_wqe; 924230189d9SNélio Laranjeiro unsigned int comp; 925230189d9SNélio Laranjeiro struct mlx5_mpw mpw = { 926230189d9SNélio Laranjeiro .state = MLX5_MPW_STATE_CLOSED, 927230189d9SNélio Laranjeiro }; 928230189d9SNélio Laranjeiro 929c3d62cc9SAdrien Mazarguil if (unlikely(!pkts_n)) 930c3d62cc9SAdrien Mazarguil return 0; 931230189d9SNélio Laranjeiro /* Prefetch first packet cacheline. */ 932fdcb0f53SNélio Laranjeiro rte_prefetch0(tx_mlx5_wqe(txq, txq->wqe_ci)); 933fdcb0f53SNélio Laranjeiro rte_prefetch0(tx_mlx5_wqe(txq, txq->wqe_ci + 1)); 934230189d9SNélio Laranjeiro /* Start processing. */ 9356cb559d6SYongseok Koh mlx5_tx_complete(txq); 9368c819a69SYongseok Koh max_elts = (elts_n - (elts_head - txq->elts_tail)); 9372eefbec5SYongseok Koh /* A CQE slot must always be available. */ 9382eefbec5SYongseok Koh assert((1u << txq->cqe_n) - (txq->cq_pi - txq->cq_ci)); 939f04f1d51SNélio Laranjeiro max_wqe = (1u << txq->wqe_n) - (txq->wqe_ci - txq->wqe_pi); 940f04f1d51SNélio Laranjeiro if (unlikely(!max_wqe)) 941f04f1d51SNélio Laranjeiro return 0; 942c3d62cc9SAdrien Mazarguil do { 943a5bf6af9SAdrien Mazarguil struct rte_mbuf *buf = *(pkts++); 944230189d9SNélio Laranjeiro uint32_t length; 945a5bf6af9SAdrien Mazarguil unsigned int segs_n = buf->nb_segs; 9464aa15eb1SNélio Laranjeiro uint32_t cs_flags; 947230189d9SNélio Laranjeiro 948c3d62cc9SAdrien Mazarguil /* 949c3d62cc9SAdrien Mazarguil * Make sure there is enough room to store this packet and 950c3d62cc9SAdrien Mazarguil * that one ring entry remains unused. 951c3d62cc9SAdrien Mazarguil */ 952a5bf6af9SAdrien Mazarguil assert(segs_n); 9538c819a69SYongseok Koh if (max_elts < segs_n) 954c3d62cc9SAdrien Mazarguil break; 955a5bf6af9SAdrien Mazarguil /* Do not bother with large packets MPW cannot handle. */ 95624c14430SShahaf Shuler if (segs_n > MLX5_MPW_DSEG_MAX) { 95724c14430SShahaf Shuler txq->stats.oerrors++; 958a5bf6af9SAdrien Mazarguil break; 95924c14430SShahaf Shuler } 9608c819a69SYongseok Koh max_elts -= segs_n; 961c3d62cc9SAdrien Mazarguil --pkts_n; 9625f8ba81cSXueming Li cs_flags = txq_ol_cksum_to_cs(buf); 963a5bf6af9SAdrien Mazarguil /* Retrieve packet information. */ 964a5bf6af9SAdrien Mazarguil length = PKT_LEN(buf); 965a5bf6af9SAdrien Mazarguil assert(length); 966230189d9SNélio Laranjeiro /* Start new session if packet differs. */ 967230189d9SNélio Laranjeiro if ((mpw.state == MLX5_MPW_STATE_OPENED) && 968230189d9SNélio Laranjeiro ((mpw.len != length) || 969a5bf6af9SAdrien Mazarguil (segs_n != 1) || 9708688b2f8SNélio Laranjeiro (mpw.wqe->eseg.cs_flags != cs_flags))) 971230189d9SNélio Laranjeiro mlx5_mpw_close(txq, &mpw); 972230189d9SNélio Laranjeiro if (mpw.state == MLX5_MPW_STATE_CLOSED) { 973f04f1d51SNélio Laranjeiro /* 974f04f1d51SNélio Laranjeiro * Multi-Packet WQE consumes at most two WQE. 975f04f1d51SNélio Laranjeiro * mlx5_mpw_new() expects to be able to use such 976f04f1d51SNélio Laranjeiro * resources. 977f04f1d51SNélio Laranjeiro */ 978f04f1d51SNélio Laranjeiro if (unlikely(max_wqe < 2)) 979f04f1d51SNélio Laranjeiro break; 980f04f1d51SNélio Laranjeiro max_wqe -= 2; 981230189d9SNélio Laranjeiro mlx5_mpw_new(txq, &mpw, length); 9828688b2f8SNélio Laranjeiro mpw.wqe->eseg.cs_flags = cs_flags; 983230189d9SNélio Laranjeiro } 984a5bf6af9SAdrien Mazarguil /* Multi-segment packets must be alone in their MPW. */ 985a5bf6af9SAdrien Mazarguil assert((segs_n == 1) || (mpw.pkts_n == 0)); 986a5bf6af9SAdrien Mazarguil #if defined(MLX5_PMD_SOFT_COUNTERS) || !defined(NDEBUG) 987a5bf6af9SAdrien Mazarguil length = 0; 988a5bf6af9SAdrien Mazarguil #endif 989a5bf6af9SAdrien Mazarguil do { 990a5bf6af9SAdrien Mazarguil volatile struct mlx5_wqe_data_seg *dseg; 991a5bf6af9SAdrien Mazarguil uintptr_t addr; 992a5bf6af9SAdrien Mazarguil 993a5bf6af9SAdrien Mazarguil assert(buf); 9948c819a69SYongseok Koh (*txq->elts)[elts_head++ & elts_m] = buf; 995230189d9SNélio Laranjeiro dseg = mpw.data.dseg[mpw.pkts_n]; 996a5bf6af9SAdrien Mazarguil addr = rte_pktmbuf_mtod(buf, uintptr_t); 997230189d9SNélio Laranjeiro *dseg = (struct mlx5_wqe_data_seg){ 9986b30a6a8SShachar Beiser .byte_count = rte_cpu_to_be_32(DATA_LEN(buf)), 9996cb559d6SYongseok Koh .lkey = mlx5_tx_mb2mr(txq, buf), 10006b30a6a8SShachar Beiser .addr = rte_cpu_to_be_64(addr), 1001230189d9SNélio Laranjeiro }; 1002a5bf6af9SAdrien Mazarguil #if defined(MLX5_PMD_SOFT_COUNTERS) || !defined(NDEBUG) 1003a5bf6af9SAdrien Mazarguil length += DATA_LEN(buf); 1004a5bf6af9SAdrien Mazarguil #endif 1005a5bf6af9SAdrien Mazarguil buf = buf->next; 1006230189d9SNélio Laranjeiro ++mpw.pkts_n; 1007a5bf6af9SAdrien Mazarguil ++j; 1008a5bf6af9SAdrien Mazarguil } while (--segs_n); 1009a5bf6af9SAdrien Mazarguil assert(length == mpw.len); 1010230189d9SNélio Laranjeiro if (mpw.pkts_n == MLX5_MPW_DSEG_MAX) 1011230189d9SNélio Laranjeiro mlx5_mpw_close(txq, &mpw); 1012230189d9SNélio Laranjeiro #ifdef MLX5_PMD_SOFT_COUNTERS 1013230189d9SNélio Laranjeiro /* Increment sent bytes counter. */ 1014230189d9SNélio Laranjeiro txq->stats.obytes += length; 1015230189d9SNélio Laranjeiro #endif 1016c3d62cc9SAdrien Mazarguil ++i; 1017c3d62cc9SAdrien Mazarguil } while (pkts_n); 1018230189d9SNélio Laranjeiro /* Take a shortcut if nothing must be sent. */ 1019230189d9SNélio Laranjeiro if (unlikely(i == 0)) 1020230189d9SNélio Laranjeiro return 0; 1021230189d9SNélio Laranjeiro /* Check whether completion threshold has been reached. */ 1022a5bf6af9SAdrien Mazarguil /* "j" includes both packets and segments. */ 1023a5bf6af9SAdrien Mazarguil comp = txq->elts_comp + j; 1024230189d9SNélio Laranjeiro if (comp >= MLX5_TX_COMP_THRESH) { 10258688b2f8SNélio Laranjeiro volatile struct mlx5_wqe *wqe = mpw.wqe; 1026230189d9SNélio Laranjeiro 1027230189d9SNélio Laranjeiro /* Request completion on last WQE. */ 10286b30a6a8SShachar Beiser wqe->ctrl[2] = rte_cpu_to_be_32(8); 1029230189d9SNélio Laranjeiro /* Save elts_head in unused "immediate" field of WQE. */ 10308688b2f8SNélio Laranjeiro wqe->ctrl[3] = elts_head; 1031230189d9SNélio Laranjeiro txq->elts_comp = 0; 10322eefbec5SYongseok Koh #ifndef NDEBUG 10332eefbec5SYongseok Koh ++txq->cq_pi; 10342eefbec5SYongseok Koh #endif 1035230189d9SNélio Laranjeiro } else { 1036230189d9SNélio Laranjeiro txq->elts_comp = comp; 1037230189d9SNélio Laranjeiro } 1038230189d9SNélio Laranjeiro #ifdef MLX5_PMD_SOFT_COUNTERS 1039230189d9SNélio Laranjeiro /* Increment sent packets counter. */ 1040230189d9SNélio Laranjeiro txq->stats.opackets += i; 1041230189d9SNélio Laranjeiro #endif 1042230189d9SNélio Laranjeiro /* Ring QP doorbell. */ 1043230189d9SNélio Laranjeiro if (mpw.state == MLX5_MPW_STATE_OPENED) 1044230189d9SNélio Laranjeiro mlx5_mpw_close(txq, &mpw); 104530807f62SNélio Laranjeiro mlx5_tx_dbrec(txq, mpw.wqe); 1046230189d9SNélio Laranjeiro txq->elts_head = elts_head; 1047230189d9SNélio Laranjeiro return i; 1048230189d9SNélio Laranjeiro } 1049230189d9SNélio Laranjeiro 1050230189d9SNélio Laranjeiro /** 1051230189d9SNélio Laranjeiro * Open a MPW inline session. 1052230189d9SNélio Laranjeiro * 1053230189d9SNélio Laranjeiro * @param txq 1054230189d9SNélio Laranjeiro * Pointer to TX queue structure. 1055230189d9SNélio Laranjeiro * @param mpw 1056230189d9SNélio Laranjeiro * Pointer to MPW session structure. 1057230189d9SNélio Laranjeiro * @param length 1058230189d9SNélio Laranjeiro * Packet length. 1059230189d9SNélio Laranjeiro */ 1060230189d9SNélio Laranjeiro static inline void 1061991b04f6SNélio Laranjeiro mlx5_mpw_inline_new(struct mlx5_txq_data *txq, struct mlx5_mpw *mpw, 1062991b04f6SNélio Laranjeiro uint32_t length) 1063230189d9SNélio Laranjeiro { 1064a821d09dSNélio Laranjeiro uint16_t idx = txq->wqe_ci & ((1 << txq->wqe_n) - 1); 10658688b2f8SNélio Laranjeiro struct mlx5_wqe_inl_small *inl; 1066230189d9SNélio Laranjeiro 1067230189d9SNélio Laranjeiro mpw->state = MLX5_MPW_INL_STATE_OPENED; 1068230189d9SNélio Laranjeiro mpw->pkts_n = 0; 1069230189d9SNélio Laranjeiro mpw->len = length; 1070230189d9SNélio Laranjeiro mpw->total_len = 0; 1071fdcb0f53SNélio Laranjeiro mpw->wqe = (volatile struct mlx5_wqe *)tx_mlx5_wqe(txq, idx); 10726b30a6a8SShachar Beiser mpw->wqe->ctrl[0] = rte_cpu_to_be_32((MLX5_OPC_MOD_MPW << 24) | 1073230189d9SNélio Laranjeiro (txq->wqe_ci << 8) | 1074c904ae25SNélio Laranjeiro MLX5_OPCODE_TSO); 10758688b2f8SNélio Laranjeiro mpw->wqe->ctrl[2] = 0; 10768688b2f8SNélio Laranjeiro mpw->wqe->ctrl[3] = 0; 10776b30a6a8SShachar Beiser mpw->wqe->eseg.mss = rte_cpu_to_be_16(length); 10788688b2f8SNélio Laranjeiro mpw->wqe->eseg.inline_hdr_sz = 0; 10798688b2f8SNélio Laranjeiro mpw->wqe->eseg.cs_flags = 0; 10808688b2f8SNélio Laranjeiro mpw->wqe->eseg.rsvd0 = 0; 10818688b2f8SNélio Laranjeiro mpw->wqe->eseg.rsvd1 = 0; 10828688b2f8SNélio Laranjeiro mpw->wqe->eseg.rsvd2 = 0; 10838688b2f8SNélio Laranjeiro inl = (struct mlx5_wqe_inl_small *) 10848688b2f8SNélio Laranjeiro (((uintptr_t)mpw->wqe) + 2 * MLX5_WQE_DWORD_SIZE); 10858688b2f8SNélio Laranjeiro mpw->data.raw = (uint8_t *)&inl->raw; 1086230189d9SNélio Laranjeiro } 1087230189d9SNélio Laranjeiro 1088230189d9SNélio Laranjeiro /** 1089230189d9SNélio Laranjeiro * Close a MPW inline session. 1090230189d9SNélio Laranjeiro * 1091230189d9SNélio Laranjeiro * @param txq 1092230189d9SNélio Laranjeiro * Pointer to TX queue structure. 1093230189d9SNélio Laranjeiro * @param mpw 1094230189d9SNélio Laranjeiro * Pointer to MPW session structure. 1095230189d9SNélio Laranjeiro */ 1096230189d9SNélio Laranjeiro static inline void 1097991b04f6SNélio Laranjeiro mlx5_mpw_inline_close(struct mlx5_txq_data *txq, struct mlx5_mpw *mpw) 1098230189d9SNélio Laranjeiro { 1099230189d9SNélio Laranjeiro unsigned int size; 11008688b2f8SNélio Laranjeiro struct mlx5_wqe_inl_small *inl = (struct mlx5_wqe_inl_small *) 11018688b2f8SNélio Laranjeiro (((uintptr_t)mpw->wqe) + (2 * MLX5_WQE_DWORD_SIZE)); 1102230189d9SNélio Laranjeiro 11038688b2f8SNélio Laranjeiro size = MLX5_WQE_SIZE - MLX5_MWQE64_INL_DATA + mpw->total_len; 1104230189d9SNélio Laranjeiro /* 1105230189d9SNélio Laranjeiro * Store size in multiple of 16 bytes. Control and Ethernet segments 1106230189d9SNélio Laranjeiro * count as 2. 1107230189d9SNélio Laranjeiro */ 11086b30a6a8SShachar Beiser mpw->wqe->ctrl[1] = rte_cpu_to_be_32(txq->qp_num_8s | 11096b30a6a8SShachar Beiser MLX5_WQE_DS(size)); 1110230189d9SNélio Laranjeiro mpw->state = MLX5_MPW_STATE_CLOSED; 11116b30a6a8SShachar Beiser inl->byte_cnt = rte_cpu_to_be_32(mpw->total_len | MLX5_INLINE_SEG); 11128688b2f8SNélio Laranjeiro txq->wqe_ci += (size + (MLX5_WQE_SIZE - 1)) / MLX5_WQE_SIZE; 1113230189d9SNélio Laranjeiro } 1114230189d9SNélio Laranjeiro 1115230189d9SNélio Laranjeiro /** 1116230189d9SNélio Laranjeiro * DPDK callback for TX with MPW inline support. 1117230189d9SNélio Laranjeiro * 1118230189d9SNélio Laranjeiro * @param dpdk_txq 1119230189d9SNélio Laranjeiro * Generic pointer to TX queue structure. 1120230189d9SNélio Laranjeiro * @param[in] pkts 1121230189d9SNélio Laranjeiro * Packets to transmit. 1122230189d9SNélio Laranjeiro * @param pkts_n 1123230189d9SNélio Laranjeiro * Number of packets in array. 1124230189d9SNélio Laranjeiro * 1125230189d9SNélio Laranjeiro * @return 1126230189d9SNélio Laranjeiro * Number of packets successfully transmitted (<= pkts_n). 1127230189d9SNélio Laranjeiro */ 1128230189d9SNélio Laranjeiro uint16_t 1129230189d9SNélio Laranjeiro mlx5_tx_burst_mpw_inline(void *dpdk_txq, struct rte_mbuf **pkts, 1130230189d9SNélio Laranjeiro uint16_t pkts_n) 1131230189d9SNélio Laranjeiro { 1132991b04f6SNélio Laranjeiro struct mlx5_txq_data *txq = (struct mlx5_txq_data *)dpdk_txq; 1133230189d9SNélio Laranjeiro uint16_t elts_head = txq->elts_head; 11348c819a69SYongseok Koh const uint16_t elts_n = 1 << txq->elts_n; 11358c819a69SYongseok Koh const uint16_t elts_m = elts_n - 1; 1136c3d62cc9SAdrien Mazarguil unsigned int i = 0; 1137a5bf6af9SAdrien Mazarguil unsigned int j = 0; 11388c819a69SYongseok Koh uint16_t max_elts; 1139f04f1d51SNélio Laranjeiro uint16_t max_wqe; 1140230189d9SNélio Laranjeiro unsigned int comp; 11410e8679fcSNélio Laranjeiro unsigned int inline_room = txq->max_inline * RTE_CACHE_LINE_SIZE; 1142230189d9SNélio Laranjeiro struct mlx5_mpw mpw = { 1143230189d9SNélio Laranjeiro .state = MLX5_MPW_STATE_CLOSED, 1144230189d9SNélio Laranjeiro }; 1145f04f1d51SNélio Laranjeiro /* 1146f04f1d51SNélio Laranjeiro * Compute the maximum number of WQE which can be consumed by inline 1147f04f1d51SNélio Laranjeiro * code. 1148f04f1d51SNélio Laranjeiro * - 2 DSEG for: 1149f04f1d51SNélio Laranjeiro * - 1 control segment, 1150f04f1d51SNélio Laranjeiro * - 1 Ethernet segment, 1151f04f1d51SNélio Laranjeiro * - N Dseg from the inline request. 1152f04f1d51SNélio Laranjeiro */ 1153f04f1d51SNélio Laranjeiro const unsigned int wqe_inl_n = 1154f04f1d51SNélio Laranjeiro ((2 * MLX5_WQE_DWORD_SIZE + 1155f04f1d51SNélio Laranjeiro txq->max_inline * RTE_CACHE_LINE_SIZE) + 1156f04f1d51SNélio Laranjeiro RTE_CACHE_LINE_SIZE - 1) / RTE_CACHE_LINE_SIZE; 1157230189d9SNélio Laranjeiro 1158c3d62cc9SAdrien Mazarguil if (unlikely(!pkts_n)) 1159c3d62cc9SAdrien Mazarguil return 0; 1160230189d9SNélio Laranjeiro /* Prefetch first packet cacheline. */ 1161fdcb0f53SNélio Laranjeiro rte_prefetch0(tx_mlx5_wqe(txq, txq->wqe_ci)); 1162fdcb0f53SNélio Laranjeiro rte_prefetch0(tx_mlx5_wqe(txq, txq->wqe_ci + 1)); 1163230189d9SNélio Laranjeiro /* Start processing. */ 11646cb559d6SYongseok Koh mlx5_tx_complete(txq); 11658c819a69SYongseok Koh max_elts = (elts_n - (elts_head - txq->elts_tail)); 11662eefbec5SYongseok Koh /* A CQE slot must always be available. */ 11672eefbec5SYongseok Koh assert((1u << txq->cqe_n) - (txq->cq_pi - txq->cq_ci)); 1168c3d62cc9SAdrien Mazarguil do { 1169a5bf6af9SAdrien Mazarguil struct rte_mbuf *buf = *(pkts++); 1170230189d9SNélio Laranjeiro uintptr_t addr; 1171230189d9SNélio Laranjeiro uint32_t length; 1172a5bf6af9SAdrien Mazarguil unsigned int segs_n = buf->nb_segs; 11734aa15eb1SNélio Laranjeiro uint8_t cs_flags; 1174230189d9SNélio Laranjeiro 1175c3d62cc9SAdrien Mazarguil /* 1176c3d62cc9SAdrien Mazarguil * Make sure there is enough room to store this packet and 1177c3d62cc9SAdrien Mazarguil * that one ring entry remains unused. 1178c3d62cc9SAdrien Mazarguil */ 1179a5bf6af9SAdrien Mazarguil assert(segs_n); 11808c819a69SYongseok Koh if (max_elts < segs_n) 1181c3d62cc9SAdrien Mazarguil break; 1182a5bf6af9SAdrien Mazarguil /* Do not bother with large packets MPW cannot handle. */ 118324c14430SShahaf Shuler if (segs_n > MLX5_MPW_DSEG_MAX) { 118424c14430SShahaf Shuler txq->stats.oerrors++; 1185a5bf6af9SAdrien Mazarguil break; 118624c14430SShahaf Shuler } 11878c819a69SYongseok Koh max_elts -= segs_n; 1188c3d62cc9SAdrien Mazarguil --pkts_n; 1189f04f1d51SNélio Laranjeiro /* 1190f04f1d51SNélio Laranjeiro * Compute max_wqe in case less WQE were consumed in previous 1191f04f1d51SNélio Laranjeiro * iteration. 1192f04f1d51SNélio Laranjeiro */ 1193f04f1d51SNélio Laranjeiro max_wqe = (1u << txq->wqe_n) - (txq->wqe_ci - txq->wqe_pi); 11945f8ba81cSXueming Li cs_flags = txq_ol_cksum_to_cs(buf); 1195a5bf6af9SAdrien Mazarguil /* Retrieve packet information. */ 1196a5bf6af9SAdrien Mazarguil length = PKT_LEN(buf); 1197230189d9SNélio Laranjeiro /* Start new session if packet differs. */ 1198230189d9SNélio Laranjeiro if (mpw.state == MLX5_MPW_STATE_OPENED) { 1199230189d9SNélio Laranjeiro if ((mpw.len != length) || 1200a5bf6af9SAdrien Mazarguil (segs_n != 1) || 12018688b2f8SNélio Laranjeiro (mpw.wqe->eseg.cs_flags != cs_flags)) 1202230189d9SNélio Laranjeiro mlx5_mpw_close(txq, &mpw); 1203230189d9SNélio Laranjeiro } else if (mpw.state == MLX5_MPW_INL_STATE_OPENED) { 1204230189d9SNélio Laranjeiro if ((mpw.len != length) || 1205a5bf6af9SAdrien Mazarguil (segs_n != 1) || 1206230189d9SNélio Laranjeiro (length > inline_room) || 12078688b2f8SNélio Laranjeiro (mpw.wqe->eseg.cs_flags != cs_flags)) { 1208230189d9SNélio Laranjeiro mlx5_mpw_inline_close(txq, &mpw); 12090e8679fcSNélio Laranjeiro inline_room = 12100e8679fcSNélio Laranjeiro txq->max_inline * RTE_CACHE_LINE_SIZE; 1211230189d9SNélio Laranjeiro } 1212230189d9SNélio Laranjeiro } 1213230189d9SNélio Laranjeiro if (mpw.state == MLX5_MPW_STATE_CLOSED) { 1214a5bf6af9SAdrien Mazarguil if ((segs_n != 1) || 1215a5bf6af9SAdrien Mazarguil (length > inline_room)) { 1216f04f1d51SNélio Laranjeiro /* 1217f04f1d51SNélio Laranjeiro * Multi-Packet WQE consumes at most two WQE. 1218f04f1d51SNélio Laranjeiro * mlx5_mpw_new() expects to be able to use 1219f04f1d51SNélio Laranjeiro * such resources. 1220f04f1d51SNélio Laranjeiro */ 1221f04f1d51SNélio Laranjeiro if (unlikely(max_wqe < 2)) 1222f04f1d51SNélio Laranjeiro break; 1223f04f1d51SNélio Laranjeiro max_wqe -= 2; 1224230189d9SNélio Laranjeiro mlx5_mpw_new(txq, &mpw, length); 12258688b2f8SNélio Laranjeiro mpw.wqe->eseg.cs_flags = cs_flags; 1226230189d9SNélio Laranjeiro } else { 1227f04f1d51SNélio Laranjeiro if (unlikely(max_wqe < wqe_inl_n)) 1228f04f1d51SNélio Laranjeiro break; 1229f04f1d51SNélio Laranjeiro max_wqe -= wqe_inl_n; 1230230189d9SNélio Laranjeiro mlx5_mpw_inline_new(txq, &mpw, length); 12318688b2f8SNélio Laranjeiro mpw.wqe->eseg.cs_flags = cs_flags; 1232230189d9SNélio Laranjeiro } 1233230189d9SNélio Laranjeiro } 1234a5bf6af9SAdrien Mazarguil /* Multi-segment packets must be alone in their MPW. */ 1235a5bf6af9SAdrien Mazarguil assert((segs_n == 1) || (mpw.pkts_n == 0)); 1236230189d9SNélio Laranjeiro if (mpw.state == MLX5_MPW_STATE_OPENED) { 12370e8679fcSNélio Laranjeiro assert(inline_room == 12380e8679fcSNélio Laranjeiro txq->max_inline * RTE_CACHE_LINE_SIZE); 1239a5bf6af9SAdrien Mazarguil #if defined(MLX5_PMD_SOFT_COUNTERS) || !defined(NDEBUG) 1240a5bf6af9SAdrien Mazarguil length = 0; 1241a5bf6af9SAdrien Mazarguil #endif 1242a5bf6af9SAdrien Mazarguil do { 1243230189d9SNélio Laranjeiro volatile struct mlx5_wqe_data_seg *dseg; 1244230189d9SNélio Laranjeiro 1245a5bf6af9SAdrien Mazarguil assert(buf); 12468c819a69SYongseok Koh (*txq->elts)[elts_head++ & elts_m] = buf; 1247230189d9SNélio Laranjeiro dseg = mpw.data.dseg[mpw.pkts_n]; 1248a5bf6af9SAdrien Mazarguil addr = rte_pktmbuf_mtod(buf, uintptr_t); 1249230189d9SNélio Laranjeiro *dseg = (struct mlx5_wqe_data_seg){ 12506b30a6a8SShachar Beiser .byte_count = 12516b30a6a8SShachar Beiser rte_cpu_to_be_32(DATA_LEN(buf)), 12526cb559d6SYongseok Koh .lkey = mlx5_tx_mb2mr(txq, buf), 12536b30a6a8SShachar Beiser .addr = rte_cpu_to_be_64(addr), 1254230189d9SNélio Laranjeiro }; 1255a5bf6af9SAdrien Mazarguil #if defined(MLX5_PMD_SOFT_COUNTERS) || !defined(NDEBUG) 1256a5bf6af9SAdrien Mazarguil length += DATA_LEN(buf); 1257a5bf6af9SAdrien Mazarguil #endif 1258a5bf6af9SAdrien Mazarguil buf = buf->next; 1259230189d9SNélio Laranjeiro ++mpw.pkts_n; 1260a5bf6af9SAdrien Mazarguil ++j; 1261a5bf6af9SAdrien Mazarguil } while (--segs_n); 1262a5bf6af9SAdrien Mazarguil assert(length == mpw.len); 1263230189d9SNélio Laranjeiro if (mpw.pkts_n == MLX5_MPW_DSEG_MAX) 1264230189d9SNélio Laranjeiro mlx5_mpw_close(txq, &mpw); 1265230189d9SNélio Laranjeiro } else { 1266230189d9SNélio Laranjeiro unsigned int max; 1267230189d9SNélio Laranjeiro 1268230189d9SNélio Laranjeiro assert(mpw.state == MLX5_MPW_INL_STATE_OPENED); 1269230189d9SNélio Laranjeiro assert(length <= inline_room); 1270a5bf6af9SAdrien Mazarguil assert(length == DATA_LEN(buf)); 1271a5bf6af9SAdrien Mazarguil addr = rte_pktmbuf_mtod(buf, uintptr_t); 12728c819a69SYongseok Koh (*txq->elts)[elts_head++ & elts_m] = buf; 1273230189d9SNélio Laranjeiro /* Maximum number of bytes before wrapping. */ 1274fdcb0f53SNélio Laranjeiro max = ((((uintptr_t)(txq->wqes)) + 1275fdcb0f53SNélio Laranjeiro (1 << txq->wqe_n) * 1276fdcb0f53SNélio Laranjeiro MLX5_WQE_SIZE) - 1277230189d9SNélio Laranjeiro (uintptr_t)mpw.data.raw); 1278230189d9SNélio Laranjeiro if (length > max) { 1279230189d9SNélio Laranjeiro rte_memcpy((void *)(uintptr_t)mpw.data.raw, 1280230189d9SNélio Laranjeiro (void *)addr, 1281230189d9SNélio Laranjeiro max); 1282fdcb0f53SNélio Laranjeiro mpw.data.raw = (volatile void *)txq->wqes; 1283230189d9SNélio Laranjeiro rte_memcpy((void *)(uintptr_t)mpw.data.raw, 1284230189d9SNélio Laranjeiro (void *)(addr + max), 1285230189d9SNélio Laranjeiro length - max); 1286230189d9SNélio Laranjeiro mpw.data.raw += length - max; 1287230189d9SNélio Laranjeiro } else { 1288230189d9SNélio Laranjeiro rte_memcpy((void *)(uintptr_t)mpw.data.raw, 1289230189d9SNélio Laranjeiro (void *)addr, 1290230189d9SNélio Laranjeiro length); 129116c64768SYongseok Koh 129216c64768SYongseok Koh if (length == max) 129316c64768SYongseok Koh mpw.data.raw = 129416c64768SYongseok Koh (volatile void *)txq->wqes; 129516c64768SYongseok Koh else 1296230189d9SNélio Laranjeiro mpw.data.raw += length; 1297230189d9SNélio Laranjeiro } 1298230189d9SNélio Laranjeiro ++mpw.pkts_n; 129976bf1574SYongseok Koh mpw.total_len += length; 1300a5bf6af9SAdrien Mazarguil ++j; 1301230189d9SNélio Laranjeiro if (mpw.pkts_n == MLX5_MPW_DSEG_MAX) { 1302230189d9SNélio Laranjeiro mlx5_mpw_inline_close(txq, &mpw); 13030e8679fcSNélio Laranjeiro inline_room = 13040e8679fcSNélio Laranjeiro txq->max_inline * RTE_CACHE_LINE_SIZE; 1305230189d9SNélio Laranjeiro } else { 1306230189d9SNélio Laranjeiro inline_room -= length; 1307230189d9SNélio Laranjeiro } 1308230189d9SNélio Laranjeiro } 1309230189d9SNélio Laranjeiro #ifdef MLX5_PMD_SOFT_COUNTERS 1310230189d9SNélio Laranjeiro /* Increment sent bytes counter. */ 1311230189d9SNélio Laranjeiro txq->stats.obytes += length; 1312230189d9SNélio Laranjeiro #endif 1313c3d62cc9SAdrien Mazarguil ++i; 1314c3d62cc9SAdrien Mazarguil } while (pkts_n); 1315230189d9SNélio Laranjeiro /* Take a shortcut if nothing must be sent. */ 1316230189d9SNélio Laranjeiro if (unlikely(i == 0)) 1317230189d9SNélio Laranjeiro return 0; 1318230189d9SNélio Laranjeiro /* Check whether completion threshold has been reached. */ 1319a5bf6af9SAdrien Mazarguil /* "j" includes both packets and segments. */ 1320a5bf6af9SAdrien Mazarguil comp = txq->elts_comp + j; 1321230189d9SNélio Laranjeiro if (comp >= MLX5_TX_COMP_THRESH) { 13228688b2f8SNélio Laranjeiro volatile struct mlx5_wqe *wqe = mpw.wqe; 1323230189d9SNélio Laranjeiro 1324230189d9SNélio Laranjeiro /* Request completion on last WQE. */ 13256b30a6a8SShachar Beiser wqe->ctrl[2] = rte_cpu_to_be_32(8); 1326230189d9SNélio Laranjeiro /* Save elts_head in unused "immediate" field of WQE. */ 13278688b2f8SNélio Laranjeiro wqe->ctrl[3] = elts_head; 1328230189d9SNélio Laranjeiro txq->elts_comp = 0; 13292eefbec5SYongseok Koh #ifndef NDEBUG 13302eefbec5SYongseok Koh ++txq->cq_pi; 13312eefbec5SYongseok Koh #endif 1332230189d9SNélio Laranjeiro } else { 1333230189d9SNélio Laranjeiro txq->elts_comp = comp; 1334230189d9SNélio Laranjeiro } 1335230189d9SNélio Laranjeiro #ifdef MLX5_PMD_SOFT_COUNTERS 1336230189d9SNélio Laranjeiro /* Increment sent packets counter. */ 1337230189d9SNélio Laranjeiro txq->stats.opackets += i; 1338230189d9SNélio Laranjeiro #endif 1339230189d9SNélio Laranjeiro /* Ring QP doorbell. */ 1340230189d9SNélio Laranjeiro if (mpw.state == MLX5_MPW_INL_STATE_OPENED) 1341230189d9SNélio Laranjeiro mlx5_mpw_inline_close(txq, &mpw); 1342230189d9SNélio Laranjeiro else if (mpw.state == MLX5_MPW_STATE_OPENED) 1343230189d9SNélio Laranjeiro mlx5_mpw_close(txq, &mpw); 134430807f62SNélio Laranjeiro mlx5_tx_dbrec(txq, mpw.wqe); 1345230189d9SNélio Laranjeiro txq->elts_head = elts_head; 1346230189d9SNélio Laranjeiro return i; 1347230189d9SNélio Laranjeiro } 1348230189d9SNélio Laranjeiro 1349230189d9SNélio Laranjeiro /** 13506ce84bd8SYongseok Koh * Open an Enhanced MPW session. 13516ce84bd8SYongseok Koh * 13526ce84bd8SYongseok Koh * @param txq 13536ce84bd8SYongseok Koh * Pointer to TX queue structure. 13546ce84bd8SYongseok Koh * @param mpw 13556ce84bd8SYongseok Koh * Pointer to MPW session structure. 13566ce84bd8SYongseok Koh * @param length 13576ce84bd8SYongseok Koh * Packet length. 13586ce84bd8SYongseok Koh */ 13596ce84bd8SYongseok Koh static inline void 1360991b04f6SNélio Laranjeiro mlx5_empw_new(struct mlx5_txq_data *txq, struct mlx5_mpw *mpw, int padding) 13616ce84bd8SYongseok Koh { 13626ce84bd8SYongseok Koh uint16_t idx = txq->wqe_ci & ((1 << txq->wqe_n) - 1); 13636ce84bd8SYongseok Koh 13646ce84bd8SYongseok Koh mpw->state = MLX5_MPW_ENHANCED_STATE_OPENED; 13656ce84bd8SYongseok Koh mpw->pkts_n = 0; 13666ce84bd8SYongseok Koh mpw->total_len = sizeof(struct mlx5_wqe); 13676ce84bd8SYongseok Koh mpw->wqe = (volatile struct mlx5_wqe *)tx_mlx5_wqe(txq, idx); 13686b30a6a8SShachar Beiser mpw->wqe->ctrl[0] = 13696b30a6a8SShachar Beiser rte_cpu_to_be_32((MLX5_OPC_MOD_ENHANCED_MPSW << 24) | 13706ce84bd8SYongseok Koh (txq->wqe_ci << 8) | 13716ce84bd8SYongseok Koh MLX5_OPCODE_ENHANCED_MPSW); 13726ce84bd8SYongseok Koh mpw->wqe->ctrl[2] = 0; 13736ce84bd8SYongseok Koh mpw->wqe->ctrl[3] = 0; 13746ce84bd8SYongseok Koh memset((void *)(uintptr_t)&mpw->wqe->eseg, 0, MLX5_WQE_DWORD_SIZE); 13756ce84bd8SYongseok Koh if (unlikely(padding)) { 13766ce84bd8SYongseok Koh uintptr_t addr = (uintptr_t)(mpw->wqe + 1); 13776ce84bd8SYongseok Koh 13786ce84bd8SYongseok Koh /* Pad the first 2 DWORDs with zero-length inline header. */ 13796b30a6a8SShachar Beiser *(volatile uint32_t *)addr = rte_cpu_to_be_32(MLX5_INLINE_SEG); 13806ce84bd8SYongseok Koh *(volatile uint32_t *)(addr + MLX5_WQE_DWORD_SIZE) = 13816b30a6a8SShachar Beiser rte_cpu_to_be_32(MLX5_INLINE_SEG); 13826ce84bd8SYongseok Koh mpw->total_len += 2 * MLX5_WQE_DWORD_SIZE; 13836ce84bd8SYongseok Koh /* Start from the next WQEBB. */ 13846ce84bd8SYongseok Koh mpw->data.raw = (volatile void *)(tx_mlx5_wqe(txq, idx + 1)); 13856ce84bd8SYongseok Koh } else { 13866ce84bd8SYongseok Koh mpw->data.raw = (volatile void *)(mpw->wqe + 1); 13876ce84bd8SYongseok Koh } 13886ce84bd8SYongseok Koh } 13896ce84bd8SYongseok Koh 13906ce84bd8SYongseok Koh /** 13916ce84bd8SYongseok Koh * Close an Enhanced MPW session. 13926ce84bd8SYongseok Koh * 13936ce84bd8SYongseok Koh * @param txq 13946ce84bd8SYongseok Koh * Pointer to TX queue structure. 13956ce84bd8SYongseok Koh * @param mpw 13966ce84bd8SYongseok Koh * Pointer to MPW session structure. 13976ce84bd8SYongseok Koh * 13986ce84bd8SYongseok Koh * @return 13996ce84bd8SYongseok Koh * Number of consumed WQEs. 14006ce84bd8SYongseok Koh */ 14016ce84bd8SYongseok Koh static inline uint16_t 1402991b04f6SNélio Laranjeiro mlx5_empw_close(struct mlx5_txq_data *txq, struct mlx5_mpw *mpw) 14036ce84bd8SYongseok Koh { 14046ce84bd8SYongseok Koh uint16_t ret; 14056ce84bd8SYongseok Koh 14066ce84bd8SYongseok Koh /* Store size in multiple of 16 bytes. Control and Ethernet segments 14076ce84bd8SYongseok Koh * count as 2. 14086ce84bd8SYongseok Koh */ 14096b30a6a8SShachar Beiser mpw->wqe->ctrl[1] = rte_cpu_to_be_32(txq->qp_num_8s | 14106b30a6a8SShachar Beiser MLX5_WQE_DS(mpw->total_len)); 14116ce84bd8SYongseok Koh mpw->state = MLX5_MPW_STATE_CLOSED; 14126ce84bd8SYongseok Koh ret = (mpw->total_len + (MLX5_WQE_SIZE - 1)) / MLX5_WQE_SIZE; 14136ce84bd8SYongseok Koh txq->wqe_ci += ret; 14146ce84bd8SYongseok Koh return ret; 14156ce84bd8SYongseok Koh } 14166ce84bd8SYongseok Koh 14176ce84bd8SYongseok Koh /** 14184b0d7b7fSYongseok Koh * TX with Enhanced MPW support. 14196ce84bd8SYongseok Koh * 14204b0d7b7fSYongseok Koh * @param txq 14214b0d7b7fSYongseok Koh * Pointer to TX queue structure. 14226ce84bd8SYongseok Koh * @param[in] pkts 14236ce84bd8SYongseok Koh * Packets to transmit. 14246ce84bd8SYongseok Koh * @param pkts_n 14256ce84bd8SYongseok Koh * Number of packets in array. 14266ce84bd8SYongseok Koh * 14276ce84bd8SYongseok Koh * @return 14286ce84bd8SYongseok Koh * Number of packets successfully transmitted (<= pkts_n). 14296ce84bd8SYongseok Koh */ 14304b0d7b7fSYongseok Koh static inline uint16_t 14314b0d7b7fSYongseok Koh txq_burst_empw(struct mlx5_txq_data *txq, struct rte_mbuf **pkts, 14324b0d7b7fSYongseok Koh uint16_t pkts_n) 14336ce84bd8SYongseok Koh { 14346ce84bd8SYongseok Koh uint16_t elts_head = txq->elts_head; 14358c819a69SYongseok Koh const uint16_t elts_n = 1 << txq->elts_n; 14368c819a69SYongseok Koh const uint16_t elts_m = elts_n - 1; 14376ce84bd8SYongseok Koh unsigned int i = 0; 14386ce84bd8SYongseok Koh unsigned int j = 0; 14398c819a69SYongseok Koh uint16_t max_elts; 14406ce84bd8SYongseok Koh uint16_t max_wqe; 14416ce84bd8SYongseok Koh unsigned int max_inline = txq->max_inline * RTE_CACHE_LINE_SIZE; 14426ce84bd8SYongseok Koh unsigned int mpw_room = 0; 14436ce84bd8SYongseok Koh unsigned int inl_pad = 0; 14446ce84bd8SYongseok Koh uint32_t inl_hdr; 14456ce84bd8SYongseok Koh struct mlx5_mpw mpw = { 14466ce84bd8SYongseok Koh .state = MLX5_MPW_STATE_CLOSED, 14476ce84bd8SYongseok Koh }; 14486ce84bd8SYongseok Koh 14496ce84bd8SYongseok Koh if (unlikely(!pkts_n)) 14506ce84bd8SYongseok Koh return 0; 14516ce84bd8SYongseok Koh /* Start processing. */ 14526cb559d6SYongseok Koh mlx5_tx_complete(txq); 14536ce84bd8SYongseok Koh max_elts = (elts_n - (elts_head - txq->elts_tail)); 14546ce84bd8SYongseok Koh /* A CQE slot must always be available. */ 14556ce84bd8SYongseok Koh assert((1u << txq->cqe_n) - (txq->cq_pi - txq->cq_ci)); 14566ce84bd8SYongseok Koh max_wqe = (1u << txq->wqe_n) - (txq->wqe_ci - txq->wqe_pi); 14576ce84bd8SYongseok Koh if (unlikely(!max_wqe)) 14586ce84bd8SYongseok Koh return 0; 14596ce84bd8SYongseok Koh do { 14606ce84bd8SYongseok Koh struct rte_mbuf *buf = *(pkts++); 14616ce84bd8SYongseok Koh uintptr_t addr; 14626ce84bd8SYongseok Koh unsigned int do_inline = 0; /* Whether inline is possible. */ 14636ce84bd8SYongseok Koh uint32_t length; 14644aa15eb1SNélio Laranjeiro uint8_t cs_flags; 14656ce84bd8SYongseok Koh 146648642ec5SYongseok Koh /* Multi-segmented packet is handled in slow-path outside. */ 146748642ec5SYongseok Koh assert(NB_SEGS(buf) == 1); 146848642ec5SYongseok Koh /* Make sure there is enough room to store this packet. */ 146948642ec5SYongseok Koh if (max_elts - j == 0) 14706ce84bd8SYongseok Koh break; 14715f8ba81cSXueming Li cs_flags = txq_ol_cksum_to_cs(buf); 14726ce84bd8SYongseok Koh /* Retrieve packet information. */ 14736ce84bd8SYongseok Koh length = PKT_LEN(buf); 14746ce84bd8SYongseok Koh /* Start new session if: 14756ce84bd8SYongseok Koh * - multi-segment packet 14766ce84bd8SYongseok Koh * - no space left even for a dseg 14776ce84bd8SYongseok Koh * - next packet can be inlined with a new WQE 14786ce84bd8SYongseok Koh * - cs_flag differs 14796ce84bd8SYongseok Koh */ 14806ce84bd8SYongseok Koh if (mpw.state == MLX5_MPW_ENHANCED_STATE_OPENED) { 148148642ec5SYongseok Koh if ((inl_pad + sizeof(struct mlx5_wqe_data_seg) > 14826ce84bd8SYongseok Koh mpw_room) || 14836ce84bd8SYongseok Koh (length <= txq->inline_max_packet_sz && 14846ce84bd8SYongseok Koh inl_pad + sizeof(inl_hdr) + length > 14856ce84bd8SYongseok Koh mpw_room) || 14866ce84bd8SYongseok Koh (mpw.wqe->eseg.cs_flags != cs_flags)) 14876ce84bd8SYongseok Koh max_wqe -= mlx5_empw_close(txq, &mpw); 14886ce84bd8SYongseok Koh } 14896ce84bd8SYongseok Koh if (unlikely(mpw.state == MLX5_MPW_STATE_CLOSED)) { 149048642ec5SYongseok Koh /* In Enhanced MPW, inline as much as the budget is 149148642ec5SYongseok Koh * allowed. The remaining space is to be filled with 149248642ec5SYongseok Koh * dsegs. If the title WQEBB isn't padded, it will have 149348642ec5SYongseok Koh * 2 dsegs there. 14946ce84bd8SYongseok Koh */ 14956ce84bd8SYongseok Koh mpw_room = RTE_MIN(MLX5_WQE_SIZE_MAX, 14966ce84bd8SYongseok Koh (max_inline ? max_inline : 14976ce84bd8SYongseok Koh pkts_n * MLX5_WQE_DWORD_SIZE) + 14986ce84bd8SYongseok Koh MLX5_WQE_SIZE); 149948642ec5SYongseok Koh if (unlikely(max_wqe * MLX5_WQE_SIZE < mpw_room)) 15006ce84bd8SYongseok Koh break; 15016ce84bd8SYongseok Koh /* Don't pad the title WQEBB to not waste WQ. */ 15026ce84bd8SYongseok Koh mlx5_empw_new(txq, &mpw, 0); 15036ce84bd8SYongseok Koh mpw_room -= mpw.total_len; 15046ce84bd8SYongseok Koh inl_pad = 0; 150548642ec5SYongseok Koh do_inline = length <= txq->inline_max_packet_sz && 15066ce84bd8SYongseok Koh sizeof(inl_hdr) + length <= mpw_room && 15076ce84bd8SYongseok Koh !txq->mpw_hdr_dseg; 15086ce84bd8SYongseok Koh mpw.wqe->eseg.cs_flags = cs_flags; 15096ce84bd8SYongseok Koh } else { 15106ce84bd8SYongseok Koh /* Evaluate whether the next packet can be inlined. 15116ce84bd8SYongseok Koh * Inlininig is possible when: 15126ce84bd8SYongseok Koh * - length is less than configured value 15136ce84bd8SYongseok Koh * - length fits for remaining space 15146ce84bd8SYongseok Koh * - not required to fill the title WQEBB with dsegs 15156ce84bd8SYongseok Koh */ 15166ce84bd8SYongseok Koh do_inline = 15176ce84bd8SYongseok Koh length <= txq->inline_max_packet_sz && 15186ce84bd8SYongseok Koh inl_pad + sizeof(inl_hdr) + length <= 15196ce84bd8SYongseok Koh mpw_room && 15206ce84bd8SYongseok Koh (!txq->mpw_hdr_dseg || 15216ce84bd8SYongseok Koh mpw.total_len >= MLX5_WQE_SIZE); 15226ce84bd8SYongseok Koh } 152324a8f524SYongseok Koh if (max_inline && do_inline) { 15246ce84bd8SYongseok Koh /* Inline packet into WQE. */ 15256ce84bd8SYongseok Koh unsigned int max; 15266ce84bd8SYongseok Koh 15276ce84bd8SYongseok Koh assert(mpw.state == MLX5_MPW_ENHANCED_STATE_OPENED); 15286ce84bd8SYongseok Koh assert(length == DATA_LEN(buf)); 15296b30a6a8SShachar Beiser inl_hdr = rte_cpu_to_be_32(length | MLX5_INLINE_SEG); 15306ce84bd8SYongseok Koh addr = rte_pktmbuf_mtod(buf, uintptr_t); 15316ce84bd8SYongseok Koh mpw.data.raw = (volatile void *) 15326ce84bd8SYongseok Koh ((uintptr_t)mpw.data.raw + inl_pad); 15336ce84bd8SYongseok Koh max = tx_mlx5_wq_tailroom(txq, 15346ce84bd8SYongseok Koh (void *)(uintptr_t)mpw.data.raw); 15356ce84bd8SYongseok Koh /* Copy inline header. */ 15366ce84bd8SYongseok Koh mpw.data.raw = (volatile void *) 15376ce84bd8SYongseok Koh mlx5_copy_to_wq( 15386ce84bd8SYongseok Koh (void *)(uintptr_t)mpw.data.raw, 15396ce84bd8SYongseok Koh &inl_hdr, 15406ce84bd8SYongseok Koh sizeof(inl_hdr), 15416ce84bd8SYongseok Koh (void *)(uintptr_t)txq->wqes, 15426ce84bd8SYongseok Koh max); 15436ce84bd8SYongseok Koh max = tx_mlx5_wq_tailroom(txq, 15446ce84bd8SYongseok Koh (void *)(uintptr_t)mpw.data.raw); 15456ce84bd8SYongseok Koh /* Copy packet data. */ 15466ce84bd8SYongseok Koh mpw.data.raw = (volatile void *) 15476ce84bd8SYongseok Koh mlx5_copy_to_wq( 15486ce84bd8SYongseok Koh (void *)(uintptr_t)mpw.data.raw, 15496ce84bd8SYongseok Koh (void *)addr, 15506ce84bd8SYongseok Koh length, 15516ce84bd8SYongseok Koh (void *)(uintptr_t)txq->wqes, 15526ce84bd8SYongseok Koh max); 15536ce84bd8SYongseok Koh ++mpw.pkts_n; 15546ce84bd8SYongseok Koh mpw.total_len += (inl_pad + sizeof(inl_hdr) + length); 15556ce84bd8SYongseok Koh /* No need to get completion as the entire packet is 15566ce84bd8SYongseok Koh * copied to WQ. Free the buf right away. 15576ce84bd8SYongseok Koh */ 15586ce84bd8SYongseok Koh rte_pktmbuf_free_seg(buf); 15596ce84bd8SYongseok Koh mpw_room -= (inl_pad + sizeof(inl_hdr) + length); 15606ce84bd8SYongseok Koh /* Add pad in the next packet if any. */ 15616ce84bd8SYongseok Koh inl_pad = (((uintptr_t)mpw.data.raw + 15626ce84bd8SYongseok Koh (MLX5_WQE_DWORD_SIZE - 1)) & 15636ce84bd8SYongseok Koh ~(MLX5_WQE_DWORD_SIZE - 1)) - 15646ce84bd8SYongseok Koh (uintptr_t)mpw.data.raw; 15656ce84bd8SYongseok Koh } else { 15666ce84bd8SYongseok Koh /* No inline. Load a dseg of packet pointer. */ 15676ce84bd8SYongseok Koh volatile rte_v128u32_t *dseg; 15686ce84bd8SYongseok Koh 15696ce84bd8SYongseok Koh assert(mpw.state == MLX5_MPW_ENHANCED_STATE_OPENED); 15706ce84bd8SYongseok Koh assert((inl_pad + sizeof(*dseg)) <= mpw_room); 15716ce84bd8SYongseok Koh assert(length == DATA_LEN(buf)); 15726ce84bd8SYongseok Koh if (!tx_mlx5_wq_tailroom(txq, 15736ce84bd8SYongseok Koh (void *)((uintptr_t)mpw.data.raw 15746ce84bd8SYongseok Koh + inl_pad))) 15756ce84bd8SYongseok Koh dseg = (volatile void *)txq->wqes; 15766ce84bd8SYongseok Koh else 15776ce84bd8SYongseok Koh dseg = (volatile void *) 15786ce84bd8SYongseok Koh ((uintptr_t)mpw.data.raw + 15796ce84bd8SYongseok Koh inl_pad); 15808c819a69SYongseok Koh (*txq->elts)[elts_head++ & elts_m] = buf; 1581f84411beSYongseok Koh addr = rte_cpu_to_be_64(rte_pktmbuf_mtod(buf, 1582f84411beSYongseok Koh uintptr_t)); 15836ce84bd8SYongseok Koh *dseg = (rte_v128u32_t) { 15846b30a6a8SShachar Beiser rte_cpu_to_be_32(length), 15856cb559d6SYongseok Koh mlx5_tx_mb2mr(txq, buf), 1586ebbb81ebSNélio Laranjeiro addr, 1587ebbb81ebSNélio Laranjeiro addr >> 32, 15886ce84bd8SYongseok Koh }; 15896ce84bd8SYongseok Koh mpw.data.raw = (volatile void *)(dseg + 1); 15906ce84bd8SYongseok Koh mpw.total_len += (inl_pad + sizeof(*dseg)); 15916ce84bd8SYongseok Koh ++j; 15926ce84bd8SYongseok Koh ++mpw.pkts_n; 15936ce84bd8SYongseok Koh mpw_room -= (inl_pad + sizeof(*dseg)); 15946ce84bd8SYongseok Koh inl_pad = 0; 15956ce84bd8SYongseok Koh } 15966ce84bd8SYongseok Koh #ifdef MLX5_PMD_SOFT_COUNTERS 15976ce84bd8SYongseok Koh /* Increment sent bytes counter. */ 15986ce84bd8SYongseok Koh txq->stats.obytes += length; 15996ce84bd8SYongseok Koh #endif 16006ce84bd8SYongseok Koh ++i; 16016ce84bd8SYongseok Koh } while (i < pkts_n); 16026ce84bd8SYongseok Koh /* Take a shortcut if nothing must be sent. */ 16036ce84bd8SYongseok Koh if (unlikely(i == 0)) 16046ce84bd8SYongseok Koh return 0; 16056ce84bd8SYongseok Koh /* Check whether completion threshold has been reached. */ 16066ce84bd8SYongseok Koh if (txq->elts_comp + j >= MLX5_TX_COMP_THRESH || 16076ce84bd8SYongseok Koh (uint16_t)(txq->wqe_ci - txq->mpw_comp) >= 16086ce84bd8SYongseok Koh (1 << txq->wqe_n) / MLX5_TX_COMP_THRESH_INLINE_DIV) { 16096ce84bd8SYongseok Koh volatile struct mlx5_wqe *wqe = mpw.wqe; 16106ce84bd8SYongseok Koh 16116ce84bd8SYongseok Koh /* Request completion on last WQE. */ 16126b30a6a8SShachar Beiser wqe->ctrl[2] = rte_cpu_to_be_32(8); 16136ce84bd8SYongseok Koh /* Save elts_head in unused "immediate" field of WQE. */ 16146ce84bd8SYongseok Koh wqe->ctrl[3] = elts_head; 16156ce84bd8SYongseok Koh txq->elts_comp = 0; 16166ce84bd8SYongseok Koh txq->mpw_comp = txq->wqe_ci; 16172eefbec5SYongseok Koh #ifndef NDEBUG 16182eefbec5SYongseok Koh ++txq->cq_pi; 16192eefbec5SYongseok Koh #endif 16206ce84bd8SYongseok Koh } else { 16216ce84bd8SYongseok Koh txq->elts_comp += j; 16226ce84bd8SYongseok Koh } 16236ce84bd8SYongseok Koh #ifdef MLX5_PMD_SOFT_COUNTERS 16246ce84bd8SYongseok Koh /* Increment sent packets counter. */ 16256ce84bd8SYongseok Koh txq->stats.opackets += i; 16266ce84bd8SYongseok Koh #endif 16276ce84bd8SYongseok Koh if (mpw.state == MLX5_MPW_ENHANCED_STATE_OPENED) 16286ce84bd8SYongseok Koh mlx5_empw_close(txq, &mpw); 16296ce84bd8SYongseok Koh /* Ring QP doorbell. */ 16306ce84bd8SYongseok Koh mlx5_tx_dbrec(txq, mpw.wqe); 16316ce84bd8SYongseok Koh txq->elts_head = elts_head; 16326ce84bd8SYongseok Koh return i; 16336ce84bd8SYongseok Koh } 16346ce84bd8SYongseok Koh 16356ce84bd8SYongseok Koh /** 16364b0d7b7fSYongseok Koh * DPDK callback for TX with Enhanced MPW support. 16374b0d7b7fSYongseok Koh * 16384b0d7b7fSYongseok Koh * @param dpdk_txq 16394b0d7b7fSYongseok Koh * Generic pointer to TX queue structure. 16404b0d7b7fSYongseok Koh * @param[in] pkts 16414b0d7b7fSYongseok Koh * Packets to transmit. 16424b0d7b7fSYongseok Koh * @param pkts_n 16434b0d7b7fSYongseok Koh * Number of packets in array. 16444b0d7b7fSYongseok Koh * 16454b0d7b7fSYongseok Koh * @return 16464b0d7b7fSYongseok Koh * Number of packets successfully transmitted (<= pkts_n). 16474b0d7b7fSYongseok Koh */ 16484b0d7b7fSYongseok Koh uint16_t 16494b0d7b7fSYongseok Koh mlx5_tx_burst_empw(void *dpdk_txq, struct rte_mbuf **pkts, uint16_t pkts_n) 16504b0d7b7fSYongseok Koh { 16514b0d7b7fSYongseok Koh struct mlx5_txq_data *txq = (struct mlx5_txq_data *)dpdk_txq; 16524b0d7b7fSYongseok Koh uint16_t nb_tx = 0; 16534b0d7b7fSYongseok Koh 16544b0d7b7fSYongseok Koh while (pkts_n > nb_tx) { 16554b0d7b7fSYongseok Koh uint16_t n; 16564b0d7b7fSYongseok Koh uint16_t ret; 16574b0d7b7fSYongseok Koh 16584b0d7b7fSYongseok Koh n = txq_count_contig_multi_seg(&pkts[nb_tx], pkts_n - nb_tx); 16594b0d7b7fSYongseok Koh if (n) { 16604b0d7b7fSYongseok Koh ret = mlx5_tx_burst(dpdk_txq, &pkts[nb_tx], n); 16614b0d7b7fSYongseok Koh if (!ret) 16624b0d7b7fSYongseok Koh break; 16634b0d7b7fSYongseok Koh nb_tx += ret; 16644b0d7b7fSYongseok Koh } 16654b0d7b7fSYongseok Koh n = txq_count_contig_single_seg(&pkts[nb_tx], pkts_n - nb_tx); 16664b0d7b7fSYongseok Koh if (n) { 16674b0d7b7fSYongseok Koh ret = txq_burst_empw(txq, &pkts[nb_tx], n); 16684b0d7b7fSYongseok Koh if (!ret) 16694b0d7b7fSYongseok Koh break; 16704b0d7b7fSYongseok Koh nb_tx += ret; 16714b0d7b7fSYongseok Koh } 16724b0d7b7fSYongseok Koh } 16734b0d7b7fSYongseok Koh return nb_tx; 16744b0d7b7fSYongseok Koh } 16754b0d7b7fSYongseok Koh 16764b0d7b7fSYongseok Koh /** 167767fa62bcSAdrien Mazarguil * Translate RX completion flags to packet type. 167867fa62bcSAdrien Mazarguil * 16793cc08bc6SXueming Li * @param[in] rxq 16803cc08bc6SXueming Li * Pointer to RX queue structure. 16816218063bSNélio Laranjeiro * @param[in] cqe 16826218063bSNélio Laranjeiro * Pointer to CQE. 168367fa62bcSAdrien Mazarguil * 168478a38edfSJianfeng Tan * @note: fix mlx5_dev_supported_ptypes_get() if any change here. 168578a38edfSJianfeng Tan * 168667fa62bcSAdrien Mazarguil * @return 168767fa62bcSAdrien Mazarguil * Packet type for struct rte_mbuf. 168867fa62bcSAdrien Mazarguil */ 168967fa62bcSAdrien Mazarguil static inline uint32_t 16903cc08bc6SXueming Li rxq_cq_to_pkt_type(struct mlx5_rxq_data *rxq, volatile struct mlx5_cqe *cqe) 169167fa62bcSAdrien Mazarguil { 1692ea16068cSYongseok Koh uint8_t idx; 1693ea16068cSYongseok Koh uint8_t pinfo = cqe->pkt_info; 1694ea16068cSYongseok Koh uint16_t ptype = cqe->hdr_type_etc; 169567fa62bcSAdrien Mazarguil 1696ea16068cSYongseok Koh /* 1697ea16068cSYongseok Koh * The index to the array should have: 1698ea16068cSYongseok Koh * bit[1:0] = l3_hdr_type 1699ea16068cSYongseok Koh * bit[4:2] = l4_hdr_type 1700ea16068cSYongseok Koh * bit[5] = ip_frag 1701ea16068cSYongseok Koh * bit[6] = tunneled 1702ea16068cSYongseok Koh * bit[7] = outer_l3_type 1703ea16068cSYongseok Koh */ 1704ea16068cSYongseok Koh idx = ((pinfo & 0x3) << 6) | ((ptype & 0xfc00) >> 10); 17053cc08bc6SXueming Li return mlx5_ptype_table[idx] | rxq->tunnel * !!(idx & (1 << 6)); 170667fa62bcSAdrien Mazarguil } 170767fa62bcSAdrien Mazarguil 170867fa62bcSAdrien Mazarguil /** 170999c12dccSNélio Laranjeiro * Get size of the next packet for a given CQE. For compressed CQEs, the 171099c12dccSNélio Laranjeiro * consumer index is updated only once all packets of the current one have 171199c12dccSNélio Laranjeiro * been processed. 171299c12dccSNélio Laranjeiro * 171399c12dccSNélio Laranjeiro * @param rxq 171499c12dccSNélio Laranjeiro * Pointer to RX queue. 171599c12dccSNélio Laranjeiro * @param cqe 171699c12dccSNélio Laranjeiro * CQE to process. 1717ecf60761SNélio Laranjeiro * @param[out] rss_hash 1718ecf60761SNélio Laranjeiro * Packet RSS Hash result. 171999c12dccSNélio Laranjeiro * 172099c12dccSNélio Laranjeiro * @return 172199c12dccSNélio Laranjeiro * Packet size in bytes (0 if there is none), -1 in case of completion 172299c12dccSNélio Laranjeiro * with error. 172399c12dccSNélio Laranjeiro */ 172499c12dccSNélio Laranjeiro static inline int 172578142aacSNélio Laranjeiro mlx5_rx_poll_len(struct mlx5_rxq_data *rxq, volatile struct mlx5_cqe *cqe, 1726ecf60761SNélio Laranjeiro uint16_t cqe_cnt, uint32_t *rss_hash) 172799c12dccSNélio Laranjeiro { 172899c12dccSNélio Laranjeiro struct rxq_zip *zip = &rxq->zip; 172999c12dccSNélio Laranjeiro uint16_t cqe_n = cqe_cnt + 1; 173099c12dccSNélio Laranjeiro int len = 0; 1731d2e842d0SYongseok Koh uint16_t idx, end; 173299c12dccSNélio Laranjeiro 173399c12dccSNélio Laranjeiro /* Process compressed data in the CQE and mini arrays. */ 173499c12dccSNélio Laranjeiro if (zip->ai) { 173599c12dccSNélio Laranjeiro volatile struct mlx5_mini_cqe8 (*mc)[8] = 173699c12dccSNélio Laranjeiro (volatile struct mlx5_mini_cqe8 (*)[8]) 17374aff4bcbSYongseok Koh (uintptr_t)(&(*rxq->cqes)[zip->ca & cqe_cnt].pkt_info); 173899c12dccSNélio Laranjeiro 17396b30a6a8SShachar Beiser len = rte_be_to_cpu_32((*mc)[zip->ai & 7].byte_cnt); 17406b30a6a8SShachar Beiser *rss_hash = rte_be_to_cpu_32((*mc)[zip->ai & 7].rx_hash_result); 174199c12dccSNélio Laranjeiro if ((++zip->ai & 7) == 0) { 1742d2e842d0SYongseok Koh /* Invalidate consumed CQEs */ 1743d2e842d0SYongseok Koh idx = zip->ca; 1744d2e842d0SYongseok Koh end = zip->na; 1745d2e842d0SYongseok Koh while (idx != end) { 1746d2e842d0SYongseok Koh (*rxq->cqes)[idx & cqe_cnt].op_own = 1747d2e842d0SYongseok Koh MLX5_CQE_INVALIDATE; 1748d2e842d0SYongseok Koh ++idx; 1749d2e842d0SYongseok Koh } 175099c12dccSNélio Laranjeiro /* 175199c12dccSNélio Laranjeiro * Increment consumer index to skip the number of 175299c12dccSNélio Laranjeiro * CQEs consumed. Hardware leaves holes in the CQ 175399c12dccSNélio Laranjeiro * ring for software use. 175499c12dccSNélio Laranjeiro */ 175599c12dccSNélio Laranjeiro zip->ca = zip->na; 175699c12dccSNélio Laranjeiro zip->na += 8; 175799c12dccSNélio Laranjeiro } 175899c12dccSNélio Laranjeiro if (unlikely(rxq->zip.ai == rxq->zip.cqe_cnt)) { 1759d2e842d0SYongseok Koh /* Invalidate the rest */ 1760d2e842d0SYongseok Koh idx = zip->ca; 1761d2e842d0SYongseok Koh end = zip->cq_ci; 176299c12dccSNélio Laranjeiro 176399c12dccSNélio Laranjeiro while (idx != end) { 176497267b8eSNelio Laranjeiro (*rxq->cqes)[idx & cqe_cnt].op_own = 176599c12dccSNélio Laranjeiro MLX5_CQE_INVALIDATE; 176699c12dccSNélio Laranjeiro ++idx; 176799c12dccSNélio Laranjeiro } 176899c12dccSNélio Laranjeiro rxq->cq_ci = zip->cq_ci; 176999c12dccSNélio Laranjeiro zip->ai = 0; 177099c12dccSNélio Laranjeiro } 177199c12dccSNélio Laranjeiro /* No compressed data, get next CQE and verify if it is compressed. */ 177299c12dccSNélio Laranjeiro } else { 177399c12dccSNélio Laranjeiro int ret; 177499c12dccSNélio Laranjeiro int8_t op_own; 177599c12dccSNélio Laranjeiro 177697267b8eSNelio Laranjeiro ret = check_cqe(cqe, cqe_n, rxq->cq_ci); 177799c12dccSNélio Laranjeiro if (unlikely(ret == 1)) 177899c12dccSNélio Laranjeiro return 0; 177999c12dccSNélio Laranjeiro ++rxq->cq_ci; 178099c12dccSNélio Laranjeiro op_own = cqe->op_own; 17811742c2d9SYongseok Koh rte_cio_rmb(); 178299c12dccSNélio Laranjeiro if (MLX5_CQE_FORMAT(op_own) == MLX5_COMPRESSED) { 178399c12dccSNélio Laranjeiro volatile struct mlx5_mini_cqe8 (*mc)[8] = 178499c12dccSNélio Laranjeiro (volatile struct mlx5_mini_cqe8 (*)[8]) 178599c12dccSNélio Laranjeiro (uintptr_t)(&(*rxq->cqes)[rxq->cq_ci & 17864aff4bcbSYongseok Koh cqe_cnt].pkt_info); 178799c12dccSNélio Laranjeiro 178899c12dccSNélio Laranjeiro /* Fix endianness. */ 17896b30a6a8SShachar Beiser zip->cqe_cnt = rte_be_to_cpu_32(cqe->byte_cnt); 179099c12dccSNélio Laranjeiro /* 179199c12dccSNélio Laranjeiro * Current mini array position is the one returned by 179299c12dccSNélio Laranjeiro * check_cqe64(). 179399c12dccSNélio Laranjeiro * 179499c12dccSNélio Laranjeiro * If completion comprises several mini arrays, as a 179599c12dccSNélio Laranjeiro * special case the second one is located 7 CQEs after 179699c12dccSNélio Laranjeiro * the initial CQE instead of 8 for subsequent ones. 179799c12dccSNélio Laranjeiro */ 1798d2e842d0SYongseok Koh zip->ca = rxq->cq_ci; 179999c12dccSNélio Laranjeiro zip->na = zip->ca + 7; 180099c12dccSNélio Laranjeiro /* Compute the next non compressed CQE. */ 180199c12dccSNélio Laranjeiro --rxq->cq_ci; 180299c12dccSNélio Laranjeiro zip->cq_ci = rxq->cq_ci + zip->cqe_cnt; 180399c12dccSNélio Laranjeiro /* Get packet size to return. */ 18046b30a6a8SShachar Beiser len = rte_be_to_cpu_32((*mc)[0].byte_cnt); 18056b30a6a8SShachar Beiser *rss_hash = rte_be_to_cpu_32((*mc)[0].rx_hash_result); 180699c12dccSNélio Laranjeiro zip->ai = 1; 1807d2e842d0SYongseok Koh /* Prefetch all the entries to be invalidated */ 1808d2e842d0SYongseok Koh idx = zip->ca; 1809d2e842d0SYongseok Koh end = zip->cq_ci; 1810d2e842d0SYongseok Koh while (idx != end) { 1811d2e842d0SYongseok Koh rte_prefetch0(&(*rxq->cqes)[(idx) & cqe_cnt]); 1812d2e842d0SYongseok Koh ++idx; 1813d2e842d0SYongseok Koh } 181499c12dccSNélio Laranjeiro } else { 18156b30a6a8SShachar Beiser len = rte_be_to_cpu_32(cqe->byte_cnt); 18166b30a6a8SShachar Beiser *rss_hash = rte_be_to_cpu_32(cqe->rx_hash_res); 181799c12dccSNélio Laranjeiro } 181899c12dccSNélio Laranjeiro /* Error while receiving packet. */ 181999c12dccSNélio Laranjeiro if (unlikely(MLX5_CQE_OPCODE(op_own) == MLX5_CQE_RESP_ERR)) 182099c12dccSNélio Laranjeiro return -1; 182199c12dccSNélio Laranjeiro } 182299c12dccSNélio Laranjeiro return len; 182399c12dccSNélio Laranjeiro } 182499c12dccSNélio Laranjeiro 182599c12dccSNélio Laranjeiro /** 182667fa62bcSAdrien Mazarguil * Translate RX completion flags to offload flags. 182767fa62bcSAdrien Mazarguil * 18286218063bSNélio Laranjeiro * @param[in] cqe 18296218063bSNélio Laranjeiro * Pointer to CQE. 183067fa62bcSAdrien Mazarguil * 183167fa62bcSAdrien Mazarguil * @return 183267fa62bcSAdrien Mazarguil * Offload flags (ol_flags) for struct rte_mbuf. 183367fa62bcSAdrien Mazarguil */ 183467fa62bcSAdrien Mazarguil static inline uint32_t 18356ba07449SXueming Li rxq_cq_to_ol_flags(volatile struct mlx5_cqe *cqe) 183667fa62bcSAdrien Mazarguil { 183767fa62bcSAdrien Mazarguil uint32_t ol_flags = 0; 18386b30a6a8SShachar Beiser uint16_t flags = rte_be_to_cpu_16(cqe->hdr_type_etc); 183967fa62bcSAdrien Mazarguil 18400603df73SNélio Laranjeiro ol_flags = 18410603df73SNélio Laranjeiro TRANSPOSE(flags, 18420603df73SNélio Laranjeiro MLX5_CQE_RX_L3_HDR_VALID, 18430603df73SNélio Laranjeiro PKT_RX_IP_CKSUM_GOOD) | 18440603df73SNélio Laranjeiro TRANSPOSE(flags, 18450603df73SNélio Laranjeiro MLX5_CQE_RX_L4_HDR_VALID, 184683e9d9a3SNelio Laranjeiro PKT_RX_L4_CKSUM_GOOD); 184767fa62bcSAdrien Mazarguil return ol_flags; 184867fa62bcSAdrien Mazarguil } 184967fa62bcSAdrien Mazarguil 185067fa62bcSAdrien Mazarguil /** 18513e1f82a1SYongseok Koh * Fill in mbuf fields from RX completion flags. 18523e1f82a1SYongseok Koh * Note that pkt->ol_flags should be initialized outside of this function. 18533e1f82a1SYongseok Koh * 18543e1f82a1SYongseok Koh * @param rxq 18553e1f82a1SYongseok Koh * Pointer to RX queue. 18563e1f82a1SYongseok Koh * @param pkt 18573e1f82a1SYongseok Koh * mbuf to fill. 18583e1f82a1SYongseok Koh * @param cqe 18593e1f82a1SYongseok Koh * CQE to process. 18603e1f82a1SYongseok Koh * @param rss_hash_res 18613e1f82a1SYongseok Koh * Packet RSS Hash result. 18623e1f82a1SYongseok Koh */ 18633e1f82a1SYongseok Koh static inline void 18643e1f82a1SYongseok Koh rxq_cq_to_mbuf(struct mlx5_rxq_data *rxq, struct rte_mbuf *pkt, 18653e1f82a1SYongseok Koh volatile struct mlx5_cqe *cqe, uint32_t rss_hash_res) 18663e1f82a1SYongseok Koh { 18673e1f82a1SYongseok Koh /* Update packet information. */ 18683e1f82a1SYongseok Koh pkt->packet_type = rxq_cq_to_pkt_type(rxq, cqe); 18693e1f82a1SYongseok Koh if (rss_hash_res && rxq->rss_hash) { 18703e1f82a1SYongseok Koh pkt->hash.rss = rss_hash_res; 18713e1f82a1SYongseok Koh pkt->ol_flags |= PKT_RX_RSS_HASH; 18723e1f82a1SYongseok Koh } 18733e1f82a1SYongseok Koh if (rxq->mark && MLX5_FLOW_MARK_IS_VALID(cqe->sop_drop_qpn)) { 18743e1f82a1SYongseok Koh pkt->ol_flags |= PKT_RX_FDIR; 18753e1f82a1SYongseok Koh if (cqe->sop_drop_qpn != 18763e1f82a1SYongseok Koh rte_cpu_to_be_32(MLX5_FLOW_MARK_DEFAULT)) { 18773e1f82a1SYongseok Koh uint32_t mark = cqe->sop_drop_qpn; 18783e1f82a1SYongseok Koh 18793e1f82a1SYongseok Koh pkt->ol_flags |= PKT_RX_FDIR_ID; 18803e1f82a1SYongseok Koh pkt->hash.fdir.hi = mlx5_flow_mark_get(mark); 18813e1f82a1SYongseok Koh } 18823e1f82a1SYongseok Koh } 18833e1f82a1SYongseok Koh if (rxq->csum) 18843e1f82a1SYongseok Koh pkt->ol_flags |= rxq_cq_to_ol_flags(cqe); 18853e1f82a1SYongseok Koh if (rxq->vlan_strip && 18863e1f82a1SYongseok Koh (cqe->hdr_type_etc & rte_cpu_to_be_16(MLX5_CQE_VLAN_STRIPPED))) { 18873e1f82a1SYongseok Koh pkt->ol_flags |= PKT_RX_VLAN | PKT_RX_VLAN_STRIPPED; 18883e1f82a1SYongseok Koh pkt->vlan_tci = rte_be_to_cpu_16(cqe->vlan_info); 18893e1f82a1SYongseok Koh } 18903e1f82a1SYongseok Koh if (rxq->hw_timestamp) { 18913e1f82a1SYongseok Koh pkt->timestamp = rte_be_to_cpu_64(cqe->timestamp); 18923e1f82a1SYongseok Koh pkt->ol_flags |= PKT_RX_TIMESTAMP; 18933e1f82a1SYongseok Koh } 18943e1f82a1SYongseok Koh } 18953e1f82a1SYongseok Koh 18963e1f82a1SYongseok Koh /** 18972e22920bSAdrien Mazarguil * DPDK callback for RX. 18982e22920bSAdrien Mazarguil * 18992e22920bSAdrien Mazarguil * @param dpdk_rxq 19002e22920bSAdrien Mazarguil * Generic pointer to RX queue structure. 19012e22920bSAdrien Mazarguil * @param[out] pkts 19022e22920bSAdrien Mazarguil * Array to store received packets. 19032e22920bSAdrien Mazarguil * @param pkts_n 19042e22920bSAdrien Mazarguil * Maximum number of packets in array. 19052e22920bSAdrien Mazarguil * 19062e22920bSAdrien Mazarguil * @return 19072e22920bSAdrien Mazarguil * Number of packets successfully received (<= pkts_n). 19082e22920bSAdrien Mazarguil */ 19092e22920bSAdrien Mazarguil uint16_t 19102e22920bSAdrien Mazarguil mlx5_rx_burst(void *dpdk_rxq, struct rte_mbuf **pkts, uint16_t pkts_n) 19112e22920bSAdrien Mazarguil { 191278142aacSNélio Laranjeiro struct mlx5_rxq_data *rxq = dpdk_rxq; 1913b4b12e55SNélio Laranjeiro const unsigned int wqe_cnt = (1 << rxq->elts_n) - 1; 1914e2f116eeSNélio Laranjeiro const unsigned int cqe_cnt = (1 << rxq->cqe_n) - 1; 19159964b965SNélio Laranjeiro const unsigned int sges_n = rxq->sges_n; 19169964b965SNélio Laranjeiro struct rte_mbuf *pkt = NULL; 19179964b965SNélio Laranjeiro struct rte_mbuf *seg = NULL; 191897267b8eSNelio Laranjeiro volatile struct mlx5_cqe *cqe = 191997267b8eSNelio Laranjeiro &(*rxq->cqes)[rxq->cq_ci & cqe_cnt]; 19209964b965SNélio Laranjeiro unsigned int i = 0; 19219964b965SNélio Laranjeiro unsigned int rq_ci = rxq->rq_ci << sges_n; 19224e66a6feSNelio Laranjeiro int len = 0; /* keep its value across iterations. */ 19232e22920bSAdrien Mazarguil 19249964b965SNélio Laranjeiro while (pkts_n) { 19259964b965SNélio Laranjeiro unsigned int idx = rq_ci & wqe_cnt; 1926*7d6bf6b8SYongseok Koh volatile struct mlx5_wqe_data_seg *wqe = 1927*7d6bf6b8SYongseok Koh &((volatile struct mlx5_wqe_data_seg *)rxq->wqes)[idx]; 19289964b965SNélio Laranjeiro struct rte_mbuf *rep = (*rxq->elts)[idx]; 1929ecf60761SNélio Laranjeiro uint32_t rss_hash_res = 0; 19309964b965SNélio Laranjeiro 19319964b965SNélio Laranjeiro if (pkt) 19329964b965SNélio Laranjeiro NEXT(seg) = rep; 19339964b965SNélio Laranjeiro seg = rep; 19349964b965SNélio Laranjeiro rte_prefetch0(seg); 19356218063bSNélio Laranjeiro rte_prefetch0(cqe); 19369964b965SNélio Laranjeiro rte_prefetch0(wqe); 1937fbfd9955SOlivier Matz rep = rte_mbuf_raw_alloc(rxq->mp); 19382e22920bSAdrien Mazarguil if (unlikely(rep == NULL)) { 193915a756b6SSagi Grimberg ++rxq->stats.rx_nombuf; 194015a756b6SSagi Grimberg if (!pkt) { 194115a756b6SSagi Grimberg /* 194215a756b6SSagi Grimberg * no buffers before we even started, 194315a756b6SSagi Grimberg * bail out silently. 194415a756b6SSagi Grimberg */ 194515a756b6SSagi Grimberg break; 194615a756b6SSagi Grimberg } 1947a1bdb71aSNélio Laranjeiro while (pkt != seg) { 1948a1bdb71aSNélio Laranjeiro assert(pkt != (*rxq->elts)[idx]); 1949fe5fe382SNélio Laranjeiro rep = NEXT(pkt); 19508f094a9aSOlivier Matz NEXT(pkt) = NULL; 19518f094a9aSOlivier Matz NB_SEGS(pkt) = 1; 19521f88c0a2SOlivier Matz rte_mbuf_raw_free(pkt); 1953fe5fe382SNélio Laranjeiro pkt = rep; 19549964b965SNélio Laranjeiro } 19556218063bSNélio Laranjeiro break; 19562e22920bSAdrien Mazarguil } 19579964b965SNélio Laranjeiro if (!pkt) { 195897267b8eSNelio Laranjeiro cqe = &(*rxq->cqes)[rxq->cq_ci & cqe_cnt]; 1959ecf60761SNélio Laranjeiro len = mlx5_rx_poll_len(rxq, cqe, cqe_cnt, 1960ecf60761SNélio Laranjeiro &rss_hash_res); 1961ecf60761SNélio Laranjeiro if (!len) { 19621f88c0a2SOlivier Matz rte_mbuf_raw_free(rep); 19636218063bSNélio Laranjeiro break; 19646218063bSNélio Laranjeiro } 196599c12dccSNélio Laranjeiro if (unlikely(len == -1)) { 196699c12dccSNélio Laranjeiro /* RX error, packet is likely too large. */ 19671f88c0a2SOlivier Matz rte_mbuf_raw_free(rep); 196899c12dccSNélio Laranjeiro ++rxq->stats.idropped; 196999c12dccSNélio Laranjeiro goto skip; 197099c12dccSNélio Laranjeiro } 19719964b965SNélio Laranjeiro pkt = seg; 19729964b965SNélio Laranjeiro assert(len >= (rxq->crc_present << 2)); 19730ac64846SMaxime Leroy pkt->ol_flags = 0; 19743e1f82a1SYongseok Koh rxq_cq_to_mbuf(rxq, pkt, cqe, rss_hash_res); 19756218063bSNélio Laranjeiro if (rxq->crc_present) 19766218063bSNélio Laranjeiro len -= ETHER_CRC_LEN; 19776218063bSNélio Laranjeiro PKT_LEN(pkt) = len; 19789964b965SNélio Laranjeiro } 19799964b965SNélio Laranjeiro DATA_LEN(rep) = DATA_LEN(seg); 19809964b965SNélio Laranjeiro PKT_LEN(rep) = PKT_LEN(seg); 19819964b965SNélio Laranjeiro SET_DATA_OFF(rep, DATA_OFF(seg)); 19829964b965SNélio Laranjeiro PORT(rep) = PORT(seg); 19839964b965SNélio Laranjeiro (*rxq->elts)[idx] = rep; 19849964b965SNélio Laranjeiro /* 19859964b965SNélio Laranjeiro * Fill NIC descriptor with the new buffer. The lkey and size 19869964b965SNélio Laranjeiro * of the buffers are already known, only the buffer address 19879964b965SNélio Laranjeiro * changes. 19889964b965SNélio Laranjeiro */ 19896b30a6a8SShachar Beiser wqe->addr = rte_cpu_to_be_64(rte_pktmbuf_mtod(rep, uintptr_t)); 1990974f1e7eSYongseok Koh /* If there's only one MR, no need to replace LKey in WQE. */ 1991974f1e7eSYongseok Koh if (unlikely(mlx5_mr_btree_len(&rxq->mr_ctrl.cache_bh) > 1)) 1992974f1e7eSYongseok Koh wqe->lkey = mlx5_rx_mb2mr(rxq, rep); 19939964b965SNélio Laranjeiro if (len > DATA_LEN(seg)) { 19949964b965SNélio Laranjeiro len -= DATA_LEN(seg); 19959964b965SNélio Laranjeiro ++NB_SEGS(pkt); 19969964b965SNélio Laranjeiro ++rq_ci; 19979964b965SNélio Laranjeiro continue; 19989964b965SNélio Laranjeiro } 19999964b965SNélio Laranjeiro DATA_LEN(seg) = len; 200087011737SAdrien Mazarguil #ifdef MLX5_PMD_SOFT_COUNTERS 200187011737SAdrien Mazarguil /* Increment bytes counter. */ 20029964b965SNélio Laranjeiro rxq->stats.ibytes += PKT_LEN(pkt); 200387011737SAdrien Mazarguil #endif 20046218063bSNélio Laranjeiro /* Return packet. */ 20056218063bSNélio Laranjeiro *(pkts++) = pkt; 20069964b965SNélio Laranjeiro pkt = NULL; 20079964b965SNélio Laranjeiro --pkts_n; 20089964b965SNélio Laranjeiro ++i; 200999c12dccSNélio Laranjeiro skip: 20109964b965SNélio Laranjeiro /* Align consumer index to the next stride. */ 20119964b965SNélio Laranjeiro rq_ci >>= sges_n; 20126218063bSNélio Laranjeiro ++rq_ci; 20139964b965SNélio Laranjeiro rq_ci <<= sges_n; 20142e22920bSAdrien Mazarguil } 20159964b965SNélio Laranjeiro if (unlikely((i == 0) && ((rq_ci >> sges_n) == rxq->rq_ci))) 20162e22920bSAdrien Mazarguil return 0; 20176218063bSNélio Laranjeiro /* Update the consumer index. */ 20189964b965SNélio Laranjeiro rxq->rq_ci = rq_ci >> sges_n; 20194fe7f662SYongseok Koh rte_cio_wmb(); 20206b30a6a8SShachar Beiser *rxq->cq_db = rte_cpu_to_be_32(rxq->cq_ci); 20214fe7f662SYongseok Koh rte_cio_wmb(); 20226b30a6a8SShachar Beiser *rxq->rq_db = rte_cpu_to_be_32(rxq->rq_ci); 202387011737SAdrien Mazarguil #ifdef MLX5_PMD_SOFT_COUNTERS 202487011737SAdrien Mazarguil /* Increment packets counter. */ 20259964b965SNélio Laranjeiro rxq->stats.ipackets += i; 202687011737SAdrien Mazarguil #endif 20279964b965SNélio Laranjeiro return i; 20282e22920bSAdrien Mazarguil } 20292e22920bSAdrien Mazarguil 2030*7d6bf6b8SYongseok Koh void 2031*7d6bf6b8SYongseok Koh mlx5_mprq_buf_free_cb(void *addr __rte_unused, void *opaque) 2032*7d6bf6b8SYongseok Koh { 2033*7d6bf6b8SYongseok Koh struct mlx5_mprq_buf *buf = opaque; 2034*7d6bf6b8SYongseok Koh 2035*7d6bf6b8SYongseok Koh if (rte_atomic16_read(&buf->refcnt) == 1) { 2036*7d6bf6b8SYongseok Koh rte_mempool_put(buf->mp, buf); 2037*7d6bf6b8SYongseok Koh } else if (rte_atomic16_add_return(&buf->refcnt, -1) == 0) { 2038*7d6bf6b8SYongseok Koh rte_atomic16_set(&buf->refcnt, 1); 2039*7d6bf6b8SYongseok Koh rte_mempool_put(buf->mp, buf); 2040*7d6bf6b8SYongseok Koh } 2041*7d6bf6b8SYongseok Koh } 2042*7d6bf6b8SYongseok Koh 2043*7d6bf6b8SYongseok Koh void 2044*7d6bf6b8SYongseok Koh mlx5_mprq_buf_free(struct mlx5_mprq_buf *buf) 2045*7d6bf6b8SYongseok Koh { 2046*7d6bf6b8SYongseok Koh mlx5_mprq_buf_free_cb(NULL, buf); 2047*7d6bf6b8SYongseok Koh } 2048*7d6bf6b8SYongseok Koh 2049*7d6bf6b8SYongseok Koh static inline void 2050*7d6bf6b8SYongseok Koh mprq_buf_replace(struct mlx5_rxq_data *rxq, uint16_t rq_idx) 2051*7d6bf6b8SYongseok Koh { 2052*7d6bf6b8SYongseok Koh struct mlx5_mprq_buf *rep = rxq->mprq_repl; 2053*7d6bf6b8SYongseok Koh volatile struct mlx5_wqe_data_seg *wqe = 2054*7d6bf6b8SYongseok Koh &((volatile struct mlx5_wqe_mprq *)rxq->wqes)[rq_idx].dseg; 2055*7d6bf6b8SYongseok Koh void *addr; 2056*7d6bf6b8SYongseok Koh 2057*7d6bf6b8SYongseok Koh assert(rep != NULL); 2058*7d6bf6b8SYongseok Koh /* Replace MPRQ buf. */ 2059*7d6bf6b8SYongseok Koh (*rxq->mprq_bufs)[rq_idx] = rep; 2060*7d6bf6b8SYongseok Koh /* Replace WQE. */ 2061*7d6bf6b8SYongseok Koh addr = mlx5_mprq_buf_addr(rep); 2062*7d6bf6b8SYongseok Koh wqe->addr = rte_cpu_to_be_64((uintptr_t)addr); 2063*7d6bf6b8SYongseok Koh /* If there's only one MR, no need to replace LKey in WQE. */ 2064*7d6bf6b8SYongseok Koh if (unlikely(mlx5_mr_btree_len(&rxq->mr_ctrl.cache_bh) > 1)) 2065*7d6bf6b8SYongseok Koh wqe->lkey = mlx5_rx_addr2mr(rxq, (uintptr_t)addr); 2066*7d6bf6b8SYongseok Koh /* Stash a mbuf for next replacement. */ 2067*7d6bf6b8SYongseok Koh if (likely(!rte_mempool_get(rxq->mprq_mp, (void **)&rep))) 2068*7d6bf6b8SYongseok Koh rxq->mprq_repl = rep; 2069*7d6bf6b8SYongseok Koh else 2070*7d6bf6b8SYongseok Koh rxq->mprq_repl = NULL; 2071*7d6bf6b8SYongseok Koh } 2072*7d6bf6b8SYongseok Koh 2073*7d6bf6b8SYongseok Koh /** 2074*7d6bf6b8SYongseok Koh * DPDK callback for RX with Multi-Packet RQ support. 2075*7d6bf6b8SYongseok Koh * 2076*7d6bf6b8SYongseok Koh * @param dpdk_rxq 2077*7d6bf6b8SYongseok Koh * Generic pointer to RX queue structure. 2078*7d6bf6b8SYongseok Koh * @param[out] pkts 2079*7d6bf6b8SYongseok Koh * Array to store received packets. 2080*7d6bf6b8SYongseok Koh * @param pkts_n 2081*7d6bf6b8SYongseok Koh * Maximum number of packets in array. 2082*7d6bf6b8SYongseok Koh * 2083*7d6bf6b8SYongseok Koh * @return 2084*7d6bf6b8SYongseok Koh * Number of packets successfully received (<= pkts_n). 2085*7d6bf6b8SYongseok Koh */ 2086*7d6bf6b8SYongseok Koh uint16_t 2087*7d6bf6b8SYongseok Koh mlx5_rx_burst_mprq(void *dpdk_rxq, struct rte_mbuf **pkts, uint16_t pkts_n) 2088*7d6bf6b8SYongseok Koh { 2089*7d6bf6b8SYongseok Koh struct mlx5_rxq_data *rxq = dpdk_rxq; 2090*7d6bf6b8SYongseok Koh const unsigned int strd_n = 1 << rxq->strd_num_n; 2091*7d6bf6b8SYongseok Koh const unsigned int strd_sz = 1 << rxq->strd_sz_n; 2092*7d6bf6b8SYongseok Koh const unsigned int strd_shift = 2093*7d6bf6b8SYongseok Koh MLX5_MPRQ_STRIDE_SHIFT_BYTE * rxq->strd_shift_en; 2094*7d6bf6b8SYongseok Koh const unsigned int cq_mask = (1 << rxq->cqe_n) - 1; 2095*7d6bf6b8SYongseok Koh const unsigned int wq_mask = (1 << rxq->elts_n) - 1; 2096*7d6bf6b8SYongseok Koh volatile struct mlx5_cqe *cqe = &(*rxq->cqes)[rxq->cq_ci & cq_mask]; 2097*7d6bf6b8SYongseok Koh unsigned int i = 0; 2098*7d6bf6b8SYongseok Koh uint16_t rq_ci = rxq->rq_ci; 2099*7d6bf6b8SYongseok Koh uint16_t strd_idx = rxq->strd_ci; 2100*7d6bf6b8SYongseok Koh struct mlx5_mprq_buf *buf = (*rxq->mprq_bufs)[rq_ci & wq_mask]; 2101*7d6bf6b8SYongseok Koh 2102*7d6bf6b8SYongseok Koh while (i < pkts_n) { 2103*7d6bf6b8SYongseok Koh struct rte_mbuf *pkt; 2104*7d6bf6b8SYongseok Koh void *addr; 2105*7d6bf6b8SYongseok Koh int ret; 2106*7d6bf6b8SYongseok Koh unsigned int len; 2107*7d6bf6b8SYongseok Koh uint16_t consumed_strd; 2108*7d6bf6b8SYongseok Koh uint32_t offset; 2109*7d6bf6b8SYongseok Koh uint32_t byte_cnt; 2110*7d6bf6b8SYongseok Koh uint32_t rss_hash_res = 0; 2111*7d6bf6b8SYongseok Koh 2112*7d6bf6b8SYongseok Koh if (strd_idx == strd_n) { 2113*7d6bf6b8SYongseok Koh /* Replace WQE only if the buffer is still in use. */ 2114*7d6bf6b8SYongseok Koh if (rte_atomic16_read(&buf->refcnt) > 1) { 2115*7d6bf6b8SYongseok Koh mprq_buf_replace(rxq, rq_ci & wq_mask); 2116*7d6bf6b8SYongseok Koh /* Release the old buffer. */ 2117*7d6bf6b8SYongseok Koh mlx5_mprq_buf_free(buf); 2118*7d6bf6b8SYongseok Koh } else if (unlikely(rxq->mprq_repl == NULL)) { 2119*7d6bf6b8SYongseok Koh struct mlx5_mprq_buf *rep; 2120*7d6bf6b8SYongseok Koh 2121*7d6bf6b8SYongseok Koh /* 2122*7d6bf6b8SYongseok Koh * Currently, the MPRQ mempool is out of buffer 2123*7d6bf6b8SYongseok Koh * and doing memcpy regardless of the size of Rx 2124*7d6bf6b8SYongseok Koh * packet. Retry allocation to get back to 2125*7d6bf6b8SYongseok Koh * normal. 2126*7d6bf6b8SYongseok Koh */ 2127*7d6bf6b8SYongseok Koh if (!rte_mempool_get(rxq->mprq_mp, 2128*7d6bf6b8SYongseok Koh (void **)&rep)) 2129*7d6bf6b8SYongseok Koh rxq->mprq_repl = rep; 2130*7d6bf6b8SYongseok Koh } 2131*7d6bf6b8SYongseok Koh /* Advance to the next WQE. */ 2132*7d6bf6b8SYongseok Koh strd_idx = 0; 2133*7d6bf6b8SYongseok Koh ++rq_ci; 2134*7d6bf6b8SYongseok Koh buf = (*rxq->mprq_bufs)[rq_ci & wq_mask]; 2135*7d6bf6b8SYongseok Koh } 2136*7d6bf6b8SYongseok Koh cqe = &(*rxq->cqes)[rxq->cq_ci & cq_mask]; 2137*7d6bf6b8SYongseok Koh ret = mlx5_rx_poll_len(rxq, cqe, cq_mask, &rss_hash_res); 2138*7d6bf6b8SYongseok Koh if (!ret) 2139*7d6bf6b8SYongseok Koh break; 2140*7d6bf6b8SYongseok Koh if (unlikely(ret == -1)) { 2141*7d6bf6b8SYongseok Koh /* RX error, packet is likely too large. */ 2142*7d6bf6b8SYongseok Koh ++rxq->stats.idropped; 2143*7d6bf6b8SYongseok Koh continue; 2144*7d6bf6b8SYongseok Koh } 2145*7d6bf6b8SYongseok Koh byte_cnt = ret; 2146*7d6bf6b8SYongseok Koh consumed_strd = (byte_cnt & MLX5_MPRQ_STRIDE_NUM_MASK) >> 2147*7d6bf6b8SYongseok Koh MLX5_MPRQ_STRIDE_NUM_SHIFT; 2148*7d6bf6b8SYongseok Koh assert(consumed_strd); 2149*7d6bf6b8SYongseok Koh /* Calculate offset before adding up stride index. */ 2150*7d6bf6b8SYongseok Koh offset = strd_idx * strd_sz + strd_shift; 2151*7d6bf6b8SYongseok Koh strd_idx += consumed_strd; 2152*7d6bf6b8SYongseok Koh if (byte_cnt & MLX5_MPRQ_FILLER_MASK) 2153*7d6bf6b8SYongseok Koh continue; 2154*7d6bf6b8SYongseok Koh /* 2155*7d6bf6b8SYongseok Koh * Currently configured to receive a packet per a stride. But if 2156*7d6bf6b8SYongseok Koh * MTU is adjusted through kernel interface, device could 2157*7d6bf6b8SYongseok Koh * consume multiple strides without raising an error. In this 2158*7d6bf6b8SYongseok Koh * case, the packet should be dropped because it is bigger than 2159*7d6bf6b8SYongseok Koh * the max_rx_pkt_len. 2160*7d6bf6b8SYongseok Koh */ 2161*7d6bf6b8SYongseok Koh if (unlikely(consumed_strd > 1)) { 2162*7d6bf6b8SYongseok Koh ++rxq->stats.idropped; 2163*7d6bf6b8SYongseok Koh continue; 2164*7d6bf6b8SYongseok Koh } 2165*7d6bf6b8SYongseok Koh pkt = rte_pktmbuf_alloc(rxq->mp); 2166*7d6bf6b8SYongseok Koh if (unlikely(pkt == NULL)) { 2167*7d6bf6b8SYongseok Koh ++rxq->stats.rx_nombuf; 2168*7d6bf6b8SYongseok Koh break; 2169*7d6bf6b8SYongseok Koh } 2170*7d6bf6b8SYongseok Koh len = (byte_cnt & MLX5_MPRQ_LEN_MASK) >> MLX5_MPRQ_LEN_SHIFT; 2171*7d6bf6b8SYongseok Koh assert((int)len >= (rxq->crc_present << 2)); 2172*7d6bf6b8SYongseok Koh if (rxq->crc_present) 2173*7d6bf6b8SYongseok Koh len -= ETHER_CRC_LEN; 2174*7d6bf6b8SYongseok Koh addr = RTE_PTR_ADD(mlx5_mprq_buf_addr(buf), offset); 2175*7d6bf6b8SYongseok Koh /* Initialize the offload flag. */ 2176*7d6bf6b8SYongseok Koh pkt->ol_flags = 0; 2177*7d6bf6b8SYongseok Koh /* 2178*7d6bf6b8SYongseok Koh * Memcpy packets to the target mbuf if: 2179*7d6bf6b8SYongseok Koh * - The size of packet is smaller than mprq_max_memcpy_len. 2180*7d6bf6b8SYongseok Koh * - Out of buffer in the Mempool for Multi-Packet RQ. 2181*7d6bf6b8SYongseok Koh */ 2182*7d6bf6b8SYongseok Koh if (len <= rxq->mprq_max_memcpy_len || rxq->mprq_repl == NULL) { 2183*7d6bf6b8SYongseok Koh /* 2184*7d6bf6b8SYongseok Koh * When memcpy'ing packet due to out-of-buffer, the 2185*7d6bf6b8SYongseok Koh * packet must be smaller than the target mbuf. 2186*7d6bf6b8SYongseok Koh */ 2187*7d6bf6b8SYongseok Koh if (unlikely(rte_pktmbuf_tailroom(pkt) < len)) { 2188*7d6bf6b8SYongseok Koh rte_pktmbuf_free_seg(pkt); 2189*7d6bf6b8SYongseok Koh ++rxq->stats.idropped; 2190*7d6bf6b8SYongseok Koh continue; 2191*7d6bf6b8SYongseok Koh } 2192*7d6bf6b8SYongseok Koh rte_memcpy(rte_pktmbuf_mtod(pkt, void *), addr, len); 2193*7d6bf6b8SYongseok Koh } else { 2194*7d6bf6b8SYongseok Koh rte_iova_t buf_iova; 2195*7d6bf6b8SYongseok Koh struct rte_mbuf_ext_shared_info *shinfo; 2196*7d6bf6b8SYongseok Koh uint16_t buf_len = consumed_strd * strd_sz; 2197*7d6bf6b8SYongseok Koh 2198*7d6bf6b8SYongseok Koh /* Increment the refcnt of the whole chunk. */ 2199*7d6bf6b8SYongseok Koh rte_atomic16_add_return(&buf->refcnt, 1); 2200*7d6bf6b8SYongseok Koh assert((uint16_t)rte_atomic16_read(&buf->refcnt) <= 2201*7d6bf6b8SYongseok Koh strd_n + 1); 2202*7d6bf6b8SYongseok Koh addr = RTE_PTR_SUB(addr, RTE_PKTMBUF_HEADROOM); 2203*7d6bf6b8SYongseok Koh /* 2204*7d6bf6b8SYongseok Koh * MLX5 device doesn't use iova but it is necessary in a 2205*7d6bf6b8SYongseok Koh * case where the Rx packet is transmitted via a 2206*7d6bf6b8SYongseok Koh * different PMD. 2207*7d6bf6b8SYongseok Koh */ 2208*7d6bf6b8SYongseok Koh buf_iova = rte_mempool_virt2iova(buf) + 2209*7d6bf6b8SYongseok Koh RTE_PTR_DIFF(addr, buf); 2210*7d6bf6b8SYongseok Koh shinfo = rte_pktmbuf_ext_shinfo_init_helper(addr, 2211*7d6bf6b8SYongseok Koh &buf_len, mlx5_mprq_buf_free_cb, buf); 2212*7d6bf6b8SYongseok Koh /* 2213*7d6bf6b8SYongseok Koh * EXT_ATTACHED_MBUF will be set to pkt->ol_flags when 2214*7d6bf6b8SYongseok Koh * attaching the stride to mbuf and more offload flags 2215*7d6bf6b8SYongseok Koh * will be added below by calling rxq_cq_to_mbuf(). 2216*7d6bf6b8SYongseok Koh * Other fields will be overwritten. 2217*7d6bf6b8SYongseok Koh */ 2218*7d6bf6b8SYongseok Koh rte_pktmbuf_attach_extbuf(pkt, addr, buf_iova, buf_len, 2219*7d6bf6b8SYongseok Koh shinfo); 2220*7d6bf6b8SYongseok Koh rte_pktmbuf_reset_headroom(pkt); 2221*7d6bf6b8SYongseok Koh assert(pkt->ol_flags == EXT_ATTACHED_MBUF); 2222*7d6bf6b8SYongseok Koh /* 2223*7d6bf6b8SYongseok Koh * Prevent potential overflow due to MTU change through 2224*7d6bf6b8SYongseok Koh * kernel interface. 2225*7d6bf6b8SYongseok Koh */ 2226*7d6bf6b8SYongseok Koh if (unlikely(rte_pktmbuf_tailroom(pkt) < len)) { 2227*7d6bf6b8SYongseok Koh rte_pktmbuf_free_seg(pkt); 2228*7d6bf6b8SYongseok Koh ++rxq->stats.idropped; 2229*7d6bf6b8SYongseok Koh continue; 2230*7d6bf6b8SYongseok Koh } 2231*7d6bf6b8SYongseok Koh } 2232*7d6bf6b8SYongseok Koh rxq_cq_to_mbuf(rxq, pkt, cqe, rss_hash_res); 2233*7d6bf6b8SYongseok Koh PKT_LEN(pkt) = len; 2234*7d6bf6b8SYongseok Koh DATA_LEN(pkt) = len; 2235*7d6bf6b8SYongseok Koh PORT(pkt) = rxq->port_id; 2236*7d6bf6b8SYongseok Koh #ifdef MLX5_PMD_SOFT_COUNTERS 2237*7d6bf6b8SYongseok Koh /* Increment bytes counter. */ 2238*7d6bf6b8SYongseok Koh rxq->stats.ibytes += PKT_LEN(pkt); 2239*7d6bf6b8SYongseok Koh #endif 2240*7d6bf6b8SYongseok Koh /* Return packet. */ 2241*7d6bf6b8SYongseok Koh *(pkts++) = pkt; 2242*7d6bf6b8SYongseok Koh ++i; 2243*7d6bf6b8SYongseok Koh } 2244*7d6bf6b8SYongseok Koh /* Update the consumer indexes. */ 2245*7d6bf6b8SYongseok Koh rxq->strd_ci = strd_idx; 2246*7d6bf6b8SYongseok Koh rte_io_wmb(); 2247*7d6bf6b8SYongseok Koh *rxq->cq_db = rte_cpu_to_be_32(rxq->cq_ci); 2248*7d6bf6b8SYongseok Koh if (rq_ci != rxq->rq_ci) { 2249*7d6bf6b8SYongseok Koh rxq->rq_ci = rq_ci; 2250*7d6bf6b8SYongseok Koh rte_io_wmb(); 2251*7d6bf6b8SYongseok Koh *rxq->rq_db = rte_cpu_to_be_32(rxq->rq_ci); 2252*7d6bf6b8SYongseok Koh } 2253*7d6bf6b8SYongseok Koh #ifdef MLX5_PMD_SOFT_COUNTERS 2254*7d6bf6b8SYongseok Koh /* Increment packets counter. */ 2255*7d6bf6b8SYongseok Koh rxq->stats.ipackets += i; 2256*7d6bf6b8SYongseok Koh #endif 2257*7d6bf6b8SYongseok Koh return i; 2258*7d6bf6b8SYongseok Koh } 2259*7d6bf6b8SYongseok Koh 22602e22920bSAdrien Mazarguil /** 22612e22920bSAdrien Mazarguil * Dummy DPDK callback for TX. 22622e22920bSAdrien Mazarguil * 22632e22920bSAdrien Mazarguil * This function is used to temporarily replace the real callback during 22642e22920bSAdrien Mazarguil * unsafe control operations on the queue, or in case of error. 22652e22920bSAdrien Mazarguil * 22662e22920bSAdrien Mazarguil * @param dpdk_txq 22672e22920bSAdrien Mazarguil * Generic pointer to TX queue structure. 22682e22920bSAdrien Mazarguil * @param[in] pkts 22692e22920bSAdrien Mazarguil * Packets to transmit. 22702e22920bSAdrien Mazarguil * @param pkts_n 22712e22920bSAdrien Mazarguil * Number of packets in array. 22722e22920bSAdrien Mazarguil * 22732e22920bSAdrien Mazarguil * @return 22742e22920bSAdrien Mazarguil * Number of packets successfully transmitted (<= pkts_n). 22752e22920bSAdrien Mazarguil */ 22762e22920bSAdrien Mazarguil uint16_t 227756f08e16SNélio Laranjeiro removed_tx_burst(void *dpdk_txq __rte_unused, 227856f08e16SNélio Laranjeiro struct rte_mbuf **pkts __rte_unused, 227956f08e16SNélio Laranjeiro uint16_t pkts_n __rte_unused) 22802e22920bSAdrien Mazarguil { 22812e22920bSAdrien Mazarguil return 0; 22822e22920bSAdrien Mazarguil } 22832e22920bSAdrien Mazarguil 22842e22920bSAdrien Mazarguil /** 22852e22920bSAdrien Mazarguil * Dummy DPDK callback for RX. 22862e22920bSAdrien Mazarguil * 22872e22920bSAdrien Mazarguil * This function is used to temporarily replace the real callback during 22882e22920bSAdrien Mazarguil * unsafe control operations on the queue, or in case of error. 22892e22920bSAdrien Mazarguil * 22902e22920bSAdrien Mazarguil * @param dpdk_rxq 22912e22920bSAdrien Mazarguil * Generic pointer to RX queue structure. 22922e22920bSAdrien Mazarguil * @param[out] pkts 22932e22920bSAdrien Mazarguil * Array to store received packets. 22942e22920bSAdrien Mazarguil * @param pkts_n 22952e22920bSAdrien Mazarguil * Maximum number of packets in array. 22962e22920bSAdrien Mazarguil * 22972e22920bSAdrien Mazarguil * @return 22982e22920bSAdrien Mazarguil * Number of packets successfully received (<= pkts_n). 22992e22920bSAdrien Mazarguil */ 23002e22920bSAdrien Mazarguil uint16_t 230156f08e16SNélio Laranjeiro removed_rx_burst(void *dpdk_txq __rte_unused, 230256f08e16SNélio Laranjeiro struct rte_mbuf **pkts __rte_unused, 230356f08e16SNélio Laranjeiro uint16_t pkts_n __rte_unused) 23042e22920bSAdrien Mazarguil { 23052e22920bSAdrien Mazarguil return 0; 23062e22920bSAdrien Mazarguil } 23076cb559d6SYongseok Koh 23086cb559d6SYongseok Koh /* 23096cb559d6SYongseok Koh * Vectorized Rx/Tx routines are not compiled in when required vector 23106cb559d6SYongseok Koh * instructions are not supported on a target architecture. The following null 23116cb559d6SYongseok Koh * stubs are needed for linkage when those are not included outside of this file 23126cb559d6SYongseok Koh * (e.g. mlx5_rxtx_vec_sse.c for x86). 23136cb559d6SYongseok Koh */ 23146cb559d6SYongseok Koh 23156cb559d6SYongseok Koh uint16_t __attribute__((weak)) 231656f08e16SNélio Laranjeiro mlx5_tx_burst_raw_vec(void *dpdk_txq __rte_unused, 231756f08e16SNélio Laranjeiro struct rte_mbuf **pkts __rte_unused, 231856f08e16SNélio Laranjeiro uint16_t pkts_n __rte_unused) 23196cb559d6SYongseok Koh { 23206cb559d6SYongseok Koh return 0; 23216cb559d6SYongseok Koh } 23226cb559d6SYongseok Koh 23236cb559d6SYongseok Koh uint16_t __attribute__((weak)) 232456f08e16SNélio Laranjeiro mlx5_tx_burst_vec(void *dpdk_txq __rte_unused, 232556f08e16SNélio Laranjeiro struct rte_mbuf **pkts __rte_unused, 232656f08e16SNélio Laranjeiro uint16_t pkts_n __rte_unused) 23276cb559d6SYongseok Koh { 23286cb559d6SYongseok Koh return 0; 23296cb559d6SYongseok Koh } 23306cb559d6SYongseok Koh 23316cb559d6SYongseok Koh uint16_t __attribute__((weak)) 233256f08e16SNélio Laranjeiro mlx5_rx_burst_vec(void *dpdk_txq __rte_unused, 233356f08e16SNélio Laranjeiro struct rte_mbuf **pkts __rte_unused, 233456f08e16SNélio Laranjeiro uint16_t pkts_n __rte_unused) 23356cb559d6SYongseok Koh { 23366cb559d6SYongseok Koh return 0; 23376cb559d6SYongseok Koh } 23386cb559d6SYongseok Koh 23396cb559d6SYongseok Koh int __attribute__((weak)) 2340af4f09f2SNélio Laranjeiro mlx5_check_raw_vec_tx_support(struct rte_eth_dev *dev __rte_unused) 23416cb559d6SYongseok Koh { 23426cb559d6SYongseok Koh return -ENOTSUP; 23436cb559d6SYongseok Koh } 23446cb559d6SYongseok Koh 23456cb559d6SYongseok Koh int __attribute__((weak)) 2346af4f09f2SNélio Laranjeiro mlx5_check_vec_tx_support(struct rte_eth_dev *dev __rte_unused) 23476cb559d6SYongseok Koh { 23486cb559d6SYongseok Koh return -ENOTSUP; 23496cb559d6SYongseok Koh } 23506cb559d6SYongseok Koh 23516cb559d6SYongseok Koh int __attribute__((weak)) 2352af4f09f2SNélio Laranjeiro mlx5_rxq_check_vec_support(struct mlx5_rxq_data *rxq __rte_unused) 23536cb559d6SYongseok Koh { 23546cb559d6SYongseok Koh return -ENOTSUP; 23556cb559d6SYongseok Koh } 23566cb559d6SYongseok Koh 23576cb559d6SYongseok Koh int __attribute__((weak)) 2358af4f09f2SNélio Laranjeiro mlx5_check_vec_rx_support(struct rte_eth_dev *dev __rte_unused) 23596cb559d6SYongseok Koh { 23606cb559d6SYongseok Koh return -ENOTSUP; 23616cb559d6SYongseok Koh } 2362