xref: /dpdk/drivers/common/mlx5/mlx5_common.h (revision daa9003984eff2b8b794deaefc5fba1f3f9f5bcf)
17b4f1e6bSMatan Azrad /* SPDX-License-Identifier: BSD-3-Clause
27b4f1e6bSMatan Azrad  * Copyright 2019 Mellanox Technologies, Ltd
37b4f1e6bSMatan Azrad  */
47b4f1e6bSMatan Azrad 
57b4f1e6bSMatan Azrad #ifndef RTE_PMD_MLX5_COMMON_H_
67b4f1e6bSMatan Azrad #define RTE_PMD_MLX5_COMMON_H_
77b4f1e6bSMatan Azrad 
893e30982SMatan Azrad #include <stdio.h>
97b4f1e6bSMatan Azrad 
101094dd94SDavid Marchand #include <rte_compat.h>
1193e30982SMatan Azrad #include <rte_pci.h>
121f37cb2bSDavid Marchand #include <bus_pci_driver.h>
130dcba525SBruce Richardson #include <rte_debug.h>
148fc74217SMatan Azrad #include <rte_atomic.h>
15d3c52126SXueming Li #include <rte_rwlock.h>
167b4f1e6bSMatan Azrad #include <rte_log.h>
17d768f324SMatan Azrad #include <rte_kvargs.h>
18d768f324SMatan Azrad #include <rte_devargs.h>
1915008f56SParav Pandit #include <rte_bitops.h>
209c373c52SSuanming Mou #include <rte_lcore.h>
219c373c52SSuanming Mou #include <rte_spinlock.h>
2245d62067SDmitry Kozlyuk #include <rte_os_shim.h>
237b4f1e6bSMatan Azrad 
248fc74217SMatan Azrad #include "mlx5_prm.h"
25262c7ad0SOri Kam #include "mlx5_devx_cmds.h"
26c31f3f7fSShiri Kuzin #include "mlx5_common_os.h"
27fb690f71SMichael Baum #include "mlx5_common_mr.h"
288fc74217SMatan Azrad 
29e4b7b8d0SBing Zhao /* Reported driver name. */
30188773a2SAsaf Penso #define MLX5_PCI_DRIVER_NAME "mlx5_pci"
31777b72a9SXueming Li #define MLX5_AUXILIARY_DRIVER_NAME "mlx5_auxiliary"
327b4f1e6bSMatan Azrad 
33654810b5SMatan Azrad /* Bit-field manipulation. */
34654810b5SMatan Azrad #define BITFIELD_DECLARE(bf, type, size) \
35654810b5SMatan Azrad 	type bf[(((size_t)(size) / (sizeof(type) * CHAR_BIT)) + \
36654810b5SMatan Azrad 		!!((size_t)(size) % (sizeof(type) * CHAR_BIT)))]
37654810b5SMatan Azrad #define BITFIELD_DEFINE(bf, type, size) \
38654810b5SMatan Azrad 	BITFIELD_DECLARE((bf), type, (size)) = { 0 }
39654810b5SMatan Azrad #define BITFIELD_SET(bf, b) \
40654810b5SMatan Azrad 	(void)((bf)[((b) / (sizeof((bf)[0]) * CHAR_BIT))] |= \
418e46d4e1SAlexander Kozyrev 		((size_t)1 << ((b) % (sizeof((bf)[0]) * CHAR_BIT))))
42654810b5SMatan Azrad #define BITFIELD_RESET(bf, b) \
43654810b5SMatan Azrad 	(void)((bf)[((b) / (sizeof((bf)[0]) * CHAR_BIT))] &= \
448e46d4e1SAlexander Kozyrev 		~((size_t)1 << ((b) % (sizeof((bf)[0]) * CHAR_BIT))))
45654810b5SMatan Azrad #define BITFIELD_ISSET(bf, b) \
46654810b5SMatan Azrad 	!!(((bf)[((b) / (sizeof((bf)[0]) * CHAR_BIT))] & \
478e46d4e1SAlexander Kozyrev 		((size_t)1 << ((b) % (sizeof((bf)[0]) * CHAR_BIT)))))
48654810b5SMatan Azrad 
49654810b5SMatan Azrad /*
507b4f1e6bSMatan Azrad  * Helper macros to work around __VA_ARGS__ limitations in a C99 compliant
517b4f1e6bSMatan Azrad  * manner.
527b4f1e6bSMatan Azrad  */
537b4f1e6bSMatan Azrad #define PMD_DRV_LOG_STRIP(a, b) a
547b4f1e6bSMatan Azrad #define PMD_DRV_LOG_OPAREN (
557b4f1e6bSMatan Azrad #define PMD_DRV_LOG_CPAREN )
567b4f1e6bSMatan Azrad #define PMD_DRV_LOG_COMMA ,
577b4f1e6bSMatan Azrad 
587b4f1e6bSMatan Azrad /* Return the file name part of a path. */
597b4f1e6bSMatan Azrad static inline const char *
pmd_drv_log_basename(const char * s)607b4f1e6bSMatan Azrad pmd_drv_log_basename(const char *s)
617b4f1e6bSMatan Azrad {
627b4f1e6bSMatan Azrad 	const char *n = s;
637b4f1e6bSMatan Azrad 
647b4f1e6bSMatan Azrad 	while (*n)
657b4f1e6bSMatan Azrad 		if (*(n++) == '/')
667b4f1e6bSMatan Azrad 			s = n;
677b4f1e6bSMatan Azrad 	return s;
687b4f1e6bSMatan Azrad }
697b4f1e6bSMatan Azrad 
707b4f1e6bSMatan Azrad #define PMD_DRV_LOG___(level, type, name, ...) \
717b4f1e6bSMatan Azrad 	rte_log(RTE_LOG_ ## level, \
727b4f1e6bSMatan Azrad 		type, \
737b4f1e6bSMatan Azrad 		RTE_FMT(name ": " \
747b4f1e6bSMatan Azrad 			RTE_FMT_HEAD(__VA_ARGS__,), \
757b4f1e6bSMatan Azrad 		RTE_FMT_TAIL(__VA_ARGS__,)))
767b4f1e6bSMatan Azrad 
770afacb04SAlexander Kozyrev #ifdef RTE_LIBRTE_MLX5_DEBUG
787b4f1e6bSMatan Azrad 
797b4f1e6bSMatan Azrad #define PMD_DRV_LOG__(level, type, name, ...) \
807b4f1e6bSMatan Azrad 	PMD_DRV_LOG___(level, type, name, "%s:%u: %s(): " __VA_ARGS__)
817b4f1e6bSMatan Azrad #define PMD_DRV_LOG_(level, type, name, s, ...) \
827b4f1e6bSMatan Azrad 	PMD_DRV_LOG__(level, type, name,\
837b4f1e6bSMatan Azrad 		s "\n" PMD_DRV_LOG_COMMA \
847b4f1e6bSMatan Azrad 		pmd_drv_log_basename(__FILE__) PMD_DRV_LOG_COMMA \
857b4f1e6bSMatan Azrad 		__LINE__ PMD_DRV_LOG_COMMA \
867b4f1e6bSMatan Azrad 		__func__, \
877b4f1e6bSMatan Azrad 		__VA_ARGS__)
887b4f1e6bSMatan Azrad 
890afacb04SAlexander Kozyrev #else /* RTE_LIBRTE_MLX5_DEBUG */
907b4f1e6bSMatan Azrad #define PMD_DRV_LOG__(level, type, name, ...) \
917b4f1e6bSMatan Azrad 	PMD_DRV_LOG___(level, type, name, __VA_ARGS__)
927b4f1e6bSMatan Azrad #define PMD_DRV_LOG_(level, type, name, s, ...) \
937b4f1e6bSMatan Azrad 	PMD_DRV_LOG__(level, type, name, s "\n", __VA_ARGS__)
947b4f1e6bSMatan Azrad 
950afacb04SAlexander Kozyrev #endif /* RTE_LIBRTE_MLX5_DEBUG */
967b4f1e6bSMatan Azrad 
977b4f1e6bSMatan Azrad /* claim_zero() does not perform any check when debugging is disabled. */
980afacb04SAlexander Kozyrev #ifdef RTE_LIBRTE_MLX5_DEBUG
997b4f1e6bSMatan Azrad 
1008e46d4e1SAlexander Kozyrev #define MLX5_ASSERT(exp) RTE_VERIFY(exp)
1018e46d4e1SAlexander Kozyrev #define claim_zero(...) MLX5_ASSERT((__VA_ARGS__) == 0)
1028e46d4e1SAlexander Kozyrev #define claim_nonzero(...) MLX5_ASSERT((__VA_ARGS__) != 0)
1037b4f1e6bSMatan Azrad 
1040afacb04SAlexander Kozyrev #else /* RTE_LIBRTE_MLX5_DEBUG */
1057b4f1e6bSMatan Azrad 
1068e46d4e1SAlexander Kozyrev #define MLX5_ASSERT(exp) RTE_ASSERT(exp)
1077b4f1e6bSMatan Azrad #define claim_zero(...) (__VA_ARGS__)
1087b4f1e6bSMatan Azrad #define claim_nonzero(...) (__VA_ARGS__)
1097b4f1e6bSMatan Azrad 
1100afacb04SAlexander Kozyrev #endif /* RTE_LIBRTE_MLX5_DEBUG */
1117b4f1e6bSMatan Azrad 
112*daa90039SDariusz Sosnowski /**
113*daa90039SDariusz Sosnowski  * Returns true if debug mode is enabled for fast path operations.
114*daa90039SDariusz Sosnowski  */
115*daa90039SDariusz Sosnowski static inline bool
mlx5_fp_debug_enabled(void)116*daa90039SDariusz Sosnowski mlx5_fp_debug_enabled(void)
117*daa90039SDariusz Sosnowski {
118*daa90039SDariusz Sosnowski #ifdef RTE_LIBRTE_MLX5_DEBUG
119*daa90039SDariusz Sosnowski 	return true;
120*daa90039SDariusz Sosnowski #else
121*daa90039SDariusz Sosnowski 	return false;
122*daa90039SDariusz Sosnowski #endif
123*daa90039SDariusz Sosnowski }
124*daa90039SDariusz Sosnowski 
1257b4f1e6bSMatan Azrad /* Allocate a buffer on the stack and fill it with a printf format string. */
1267b4f1e6bSMatan Azrad #define MKSTR(name, ...) \
1277b4f1e6bSMatan Azrad 	int mkstr_size_##name = snprintf(NULL, 0, "" __VA_ARGS__); \
1287b4f1e6bSMatan Azrad 	char name[mkstr_size_##name + 1]; \
1297b4f1e6bSMatan Azrad 	\
130777b72a9SXueming Li 	memset(name, 0, mkstr_size_##name + 1); \
1317b4f1e6bSMatan Azrad 	snprintf(name, sizeof(name), "" __VA_ARGS__)
1327b4f1e6bSMatan Azrad 
133e415f348SMatan Azrad enum {
134e415f348SMatan Azrad 	PCI_VENDOR_ID_MELLANOX = 0x15b3,
135e415f348SMatan Azrad };
136e415f348SMatan Azrad 
137e415f348SMatan Azrad enum {
138e415f348SMatan Azrad 	PCI_DEVICE_ID_MELLANOX_CONNECTX4 = 0x1013,
139e415f348SMatan Azrad 	PCI_DEVICE_ID_MELLANOX_CONNECTX4VF = 0x1014,
140e415f348SMatan Azrad 	PCI_DEVICE_ID_MELLANOX_CONNECTX4LX = 0x1015,
141e415f348SMatan Azrad 	PCI_DEVICE_ID_MELLANOX_CONNECTX4LXVF = 0x1016,
142e415f348SMatan Azrad 	PCI_DEVICE_ID_MELLANOX_CONNECTX5 = 0x1017,
143e415f348SMatan Azrad 	PCI_DEVICE_ID_MELLANOX_CONNECTX5VF = 0x1018,
144e415f348SMatan Azrad 	PCI_DEVICE_ID_MELLANOX_CONNECTX5EX = 0x1019,
145e415f348SMatan Azrad 	PCI_DEVICE_ID_MELLANOX_CONNECTX5EXVF = 0x101a,
1460a9fff95SRaslan Darawsheh 	PCI_DEVICE_ID_MELLANOX_BLUEFIELD = 0xa2d2,
1470a9fff95SRaslan Darawsheh 	PCI_DEVICE_ID_MELLANOX_BLUEFIELDVF = 0xa2d3,
148e415f348SMatan Azrad 	PCI_DEVICE_ID_MELLANOX_CONNECTX6 = 0x101b,
149e415f348SMatan Azrad 	PCI_DEVICE_ID_MELLANOX_CONNECTX6VF = 0x101c,
150e415f348SMatan Azrad 	PCI_DEVICE_ID_MELLANOX_CONNECTX6DX = 0x101d,
1513ea12cadSRaslan Darawsheh 	PCI_DEVICE_ID_MELLANOX_CONNECTXVF = 0x101e,
1520a9fff95SRaslan Darawsheh 	PCI_DEVICE_ID_MELLANOX_BLUEFIELD2 = 0xa2d6,
15328c9a7d7SAli Alnubani 	PCI_DEVICE_ID_MELLANOX_CONNECTX6LX = 0x101f,
1546ca37b06SRaslan Darawsheh 	PCI_DEVICE_ID_MELLANOX_CONNECTX7 = 0x1021,
1550a9fff95SRaslan Darawsheh 	PCI_DEVICE_ID_MELLANOX_BLUEFIELD3 = 0Xa2dc,
156e415f348SMatan Azrad };
157e415f348SMatan Azrad 
158654810b5SMatan Azrad /* Maximum number of simultaneous unicast MAC addresses. */
159654810b5SMatan Azrad #define MLX5_MAX_UC_MAC_ADDRESSES 128
160654810b5SMatan Azrad /* Maximum number of simultaneous Multicast MAC addresses. */
161654810b5SMatan Azrad #define MLX5_MAX_MC_MAC_ADDRESSES 128
162654810b5SMatan Azrad /* Maximum number of simultaneous MAC addresses. */
163654810b5SMatan Azrad #define MLX5_MAX_MAC_ADDRESSES \
164654810b5SMatan Azrad 	(MLX5_MAX_UC_MAC_ADDRESSES + MLX5_MAX_MC_MAC_ADDRESSES)
165654810b5SMatan Azrad 
166654810b5SMatan Azrad /* Recognized Infiniband device physical port name types. */
167654810b5SMatan Azrad enum mlx5_nl_phys_port_name_type {
168654810b5SMatan Azrad 	MLX5_PHYS_PORT_NAME_TYPE_NOTSET = 0, /* Not set. */
169654810b5SMatan Azrad 	MLX5_PHYS_PORT_NAME_TYPE_LEGACY, /* before kernel ver < 5.0 */
170654810b5SMatan Azrad 	MLX5_PHYS_PORT_NAME_TYPE_UPLINK, /* p0, kernel ver >= 5.0 */
171654810b5SMatan Azrad 	MLX5_PHYS_PORT_NAME_TYPE_PFVF, /* pf0vf0, kernel ver >= 5.0 */
172420bbdaeSViacheslav Ovsiienko 	MLX5_PHYS_PORT_NAME_TYPE_PFHPF, /* pf0, kernel ver >= 5.7, HPF rep */
17359df97f1SXueming Li 	MLX5_PHYS_PORT_NAME_TYPE_PFSF, /* pf0sf0, kernel ver >= 5.0 */
174654810b5SMatan Azrad 	MLX5_PHYS_PORT_NAME_TYPE_UNKNOWN, /* Unrecognized. */
175654810b5SMatan Azrad };
176654810b5SMatan Azrad 
177654810b5SMatan Azrad /** Switch information returned by mlx5_nl_switch_info(). */
178654810b5SMatan Azrad struct mlx5_switch_info {
179654810b5SMatan Azrad 	uint32_t master:1; /**< Master device. */
180654810b5SMatan Azrad 	uint32_t representor:1; /**< Representor device. */
181654810b5SMatan Azrad 	enum mlx5_nl_phys_port_name_type name_type; /** < Port name type. */
18259df97f1SXueming Li 	int32_t ctrl_num; /**< Controller number (valid for c#pf#vf# format). */
183654810b5SMatan Azrad 	int32_t pf_num; /**< PF number (valid for pfxvfx format only). */
184654810b5SMatan Azrad 	int32_t port_name; /**< Representor port name. */
18511c73de9SDariusz Sosnowski 	int32_t mpesw_owner; /**< MPESW owner port number. */
186654810b5SMatan Azrad 	uint64_t switch_id; /**< Switch identifier. */
187654810b5SMatan Azrad };
188654810b5SMatan Azrad 
1898fc74217SMatan Azrad /* CQE status. */
1908fc74217SMatan Azrad enum mlx5_cqe_status {
1918fc74217SMatan Azrad 	MLX5_CQE_STATUS_SW_OWN = -1,
1928fc74217SMatan Azrad 	MLX5_CQE_STATUS_HW_OWN = -2,
1938fc74217SMatan Azrad 	MLX5_CQE_STATUS_ERR = -3,
1948fc74217SMatan Azrad };
1958fc74217SMatan Azrad 
1968fc74217SMatan Azrad /**
197a7da07e5SAlexander Kozyrev  * Check whether CQE has an error opcode.
198a7da07e5SAlexander Kozyrev  *
199a7da07e5SAlexander Kozyrev  * @param op_code
200a7da07e5SAlexander Kozyrev  *   Opcode to check.
201a7da07e5SAlexander Kozyrev  *
202a7da07e5SAlexander Kozyrev  * @return
203a7da07e5SAlexander Kozyrev  *   The CQE status.
204a7da07e5SAlexander Kozyrev  */
205a7da07e5SAlexander Kozyrev static __rte_always_inline enum mlx5_cqe_status
check_cqe_error(const uint8_t op_code)206a7da07e5SAlexander Kozyrev check_cqe_error(const uint8_t op_code)
207a7da07e5SAlexander Kozyrev {
2087be74edbSHonnappa Nagarahalli 	/* Prevent speculative reading of other fields in CQE until
2097be74edbSHonnappa Nagarahalli 	 * CQE is valid.
2107be74edbSHonnappa Nagarahalli 	 */
211e12a0166STyler Retzlaff 	rte_atomic_thread_fence(rte_memory_order_acquire);
2127be74edbSHonnappa Nagarahalli 
213a7da07e5SAlexander Kozyrev 	if (unlikely(op_code == MLX5_CQE_RESP_ERR ||
214a7da07e5SAlexander Kozyrev 		     op_code == MLX5_CQE_REQ_ERR))
215a7da07e5SAlexander Kozyrev 		return MLX5_CQE_STATUS_ERR;
216a7da07e5SAlexander Kozyrev 	return MLX5_CQE_STATUS_SW_OWN;
217a7da07e5SAlexander Kozyrev }
218a7da07e5SAlexander Kozyrev 
219a7da07e5SAlexander Kozyrev /**
220a7da07e5SAlexander Kozyrev  * Check whether CQE is valid using owner bit.
2218fc74217SMatan Azrad  *
2228fc74217SMatan Azrad  * @param cqe
2238fc74217SMatan Azrad  *   Pointer to CQE.
2248fc74217SMatan Azrad  * @param cqes_n
2258fc74217SMatan Azrad  *   Size of completion queue.
2268fc74217SMatan Azrad  * @param ci
2278fc74217SMatan Azrad  *   Consumer index.
2288fc74217SMatan Azrad  *
2298fc74217SMatan Azrad  * @return
2308fc74217SMatan Azrad  *   The CQE status.
2318fc74217SMatan Azrad  */
2328fc74217SMatan Azrad static __rte_always_inline enum mlx5_cqe_status
check_cqe(volatile struct mlx5_cqe * cqe,const uint16_t cqes_n,const uint16_t ci)2338fc74217SMatan Azrad check_cqe(volatile struct mlx5_cqe *cqe, const uint16_t cqes_n,
2348fc74217SMatan Azrad 	  const uint16_t ci)
2358fc74217SMatan Azrad {
2368fc74217SMatan Azrad 	const uint16_t idx = ci & cqes_n;
2378fc74217SMatan Azrad 	const uint8_t op_own = cqe->op_own;
2388fc74217SMatan Azrad 	const uint8_t op_owner = MLX5_CQE_OWNER(op_own);
2398fc74217SMatan Azrad 	const uint8_t op_code = MLX5_CQE_OPCODE(op_own);
2408fc74217SMatan Azrad 
241a7da07e5SAlexander Kozyrev 	if (unlikely((op_owner != (!!(idx))) ||
242a7da07e5SAlexander Kozyrev 		     (op_code == MLX5_CQE_INVALID)))
2438fc74217SMatan Azrad 		return MLX5_CQE_STATUS_HW_OWN;
244a7da07e5SAlexander Kozyrev 	return check_cqe_error(op_code);
245a7da07e5SAlexander Kozyrev }
246a7da07e5SAlexander Kozyrev 
247a7da07e5SAlexander Kozyrev /**
248a7da07e5SAlexander Kozyrev  * Check whether CQE is valid using validity iteration count.
249a7da07e5SAlexander Kozyrev  *
250a7da07e5SAlexander Kozyrev  * @param cqe
251a7da07e5SAlexander Kozyrev  *   Pointer to CQE.
252a7da07e5SAlexander Kozyrev  * @param cqes_n
253a7da07e5SAlexander Kozyrev  *   Log 2 of completion queue size.
254a7da07e5SAlexander Kozyrev  * @param ci
255a7da07e5SAlexander Kozyrev  *   Consumer index.
256a7da07e5SAlexander Kozyrev  *
257a7da07e5SAlexander Kozyrev  * @return
258a7da07e5SAlexander Kozyrev  *   The CQE status.
259a7da07e5SAlexander Kozyrev  */
260a7da07e5SAlexander Kozyrev static __rte_always_inline enum mlx5_cqe_status
check_cqe_iteration(volatile struct mlx5_cqe * cqe,const uint16_t cqes_n,const uint32_t ci)261a7da07e5SAlexander Kozyrev check_cqe_iteration(volatile struct mlx5_cqe *cqe, const uint16_t cqes_n,
262a7da07e5SAlexander Kozyrev 		    const uint32_t ci)
263a7da07e5SAlexander Kozyrev {
264a7da07e5SAlexander Kozyrev 	const uint8_t op_own = cqe->op_own;
265a7da07e5SAlexander Kozyrev 	const uint8_t op_code = MLX5_CQE_OPCODE(op_own);
266a7da07e5SAlexander Kozyrev 	const uint8_t vic = ci >> cqes_n;
267a7da07e5SAlexander Kozyrev 
268a7da07e5SAlexander Kozyrev 	if (unlikely((cqe->validity_iteration_count != vic) ||
269a7da07e5SAlexander Kozyrev 		     (op_code == MLX5_CQE_INVALID)))
270a7da07e5SAlexander Kozyrev 		return MLX5_CQE_STATUS_HW_OWN;
271a7da07e5SAlexander Kozyrev 	return check_cqe_error(op_code);
2728fc74217SMatan Azrad }
2738fc74217SMatan Azrad 
27467350881SThomas Monjalon /*
2754d567938SThomas Monjalon  * Get PCI address <DBDF> string from EAL device.
2764d567938SThomas Monjalon  *
2774d567938SThomas Monjalon  * @param[out] addr
2784d567938SThomas Monjalon  *	The output address buffer string
2794d567938SThomas Monjalon  * @param[in] size
2804d567938SThomas Monjalon  *	The output buffer size
2814d567938SThomas Monjalon  * @return
2824d567938SThomas Monjalon  *   - 0 on success.
2834d567938SThomas Monjalon  *   - Negative value and rte_errno is set otherwise.
2844d567938SThomas Monjalon  */
2859b31fc90SViacheslav Ovsiienko __rte_internal
2864d567938SThomas Monjalon int mlx5_dev_to_pci_str(const struct rte_device *dev, char *addr, size_t size);
2874d567938SThomas Monjalon 
2884d567938SThomas Monjalon /*
28967350881SThomas Monjalon  * Get PCI address from sysfs of a PCI-related device.
29067350881SThomas Monjalon  *
29167350881SThomas Monjalon  * @param[in] dev_path
29267350881SThomas Monjalon  *   The sysfs path should not point to the direct plain PCI device.
29367350881SThomas Monjalon  *   Instead, the node "/device/" is used to access the real device.
29467350881SThomas Monjalon  * @param[out] pci_addr
29567350881SThomas Monjalon  *   Parsed PCI address.
29667350881SThomas Monjalon  *
29767350881SThomas Monjalon  * @return
29867350881SThomas Monjalon  *   - 0 on success.
29967350881SThomas Monjalon  *   - Negative value and rte_errno is set otherwise.
30067350881SThomas Monjalon  */
30164c563f8SOphir Munk __rte_internal
3024d567938SThomas Monjalon int mlx5_get_pci_addr(const char *dev_path, struct rte_pci_addr *pci_addr);
30367350881SThomas Monjalon 
30467350881SThomas Monjalon /*
30567350881SThomas Monjalon  * Get kernel network interface name from sysfs IB device path.
30667350881SThomas Monjalon  *
30767350881SThomas Monjalon  * @param[in] ibdev_path
30867350881SThomas Monjalon  *   The sysfs path to IB device.
30967350881SThomas Monjalon  * @param[out] ifname
31067350881SThomas Monjalon  *   Interface name output of size IF_NAMESIZE.
31167350881SThomas Monjalon  *
31267350881SThomas Monjalon  * @return
31367350881SThomas Monjalon  *   - 0 on success.
31467350881SThomas Monjalon  *   - Negative value and rte_errno is set otherwise.
31567350881SThomas Monjalon  */
316aec086c9SMatan Azrad __rte_internal
317aec086c9SMatan Azrad int mlx5_get_ifname_sysfs(const char *ibdev_path, char *ifname);
318aec086c9SMatan Azrad 
319777b72a9SXueming Li __rte_internal
320777b72a9SXueming Li int mlx5_auxiliary_get_child_name(const char *dev, const char *node,
321777b72a9SXueming Li 				  char *child, size_t size);
32293e30982SMatan Azrad 
323d768f324SMatan Azrad enum mlx5_class {
324d768f324SMatan Azrad 	MLX5_CLASS_INVALID,
325a99f2f90SXueming Li 	MLX5_CLASS_ETH = RTE_BIT64(0),
32615008f56SParav Pandit 	MLX5_CLASS_VDPA = RTE_BIT64(1),
32715008f56SParav Pandit 	MLX5_CLASS_REGEX = RTE_BIT64(2),
328832a4cf1SMatan Azrad 	MLX5_CLASS_COMPRESS = RTE_BIT64(3),
329a7c86884SShiri Kuzin 	MLX5_CLASS_CRYPTO = RTE_BIT64(4),
330d768f324SMatan Azrad };
331654810b5SMatan Azrad 
33244c1b52bSViacheslav Ovsiienko #define MLX5_DBR_SIZE RTE_CACHE_LINE_SIZE
333262c7ad0SOri Kam 
334262c7ad0SOri Kam /* devX creation object */
335262c7ad0SOri Kam struct mlx5_devx_obj {
336262c7ad0SOri Kam 	void *obj; /* The DV object. */
337262c7ad0SOri Kam 	int id; /* The object ID. */
338262c7ad0SOri Kam };
339262c7ad0SOri Kam 
340262c7ad0SOri Kam /* UMR memory buffer used to define 1 entry in indirect mkey. */
341262c7ad0SOri Kam struct mlx5_klm {
342262c7ad0SOri Kam 	uint32_t byte_count;
343262c7ad0SOri Kam 	uint32_t mkey;
344262c7ad0SOri Kam 	uint64_t address;
345262c7ad0SOri Kam };
346262c7ad0SOri Kam 
347a729d2f0SMichael Baum /** Control for key/values list. */
348a729d2f0SMichael Baum struct mlx5_kvargs_ctrl {
349a729d2f0SMichael Baum 	struct rte_kvargs *kvlist; /* Structure containing list of key/values.*/
350a729d2f0SMichael Baum 	bool is_used[RTE_KVARGS_MAX]; /* Indicator which devargs were used. */
351a729d2f0SMichael Baum };
352a729d2f0SMichael Baum 
353a729d2f0SMichael Baum /**
354a729d2f0SMichael Baum  * Call a handler function for each key/value in the list of keys.
355a729d2f0SMichael Baum  *
356a729d2f0SMichael Baum  * For each key/value association that matches the given key, calls the
357a729d2f0SMichael Baum  * handler function with the for a given arg_name passing the value on the
358a729d2f0SMichael Baum  * dictionary for that key and a given extra argument.
359a729d2f0SMichael Baum  *
360a729d2f0SMichael Baum  * @param mkvlist
361a729d2f0SMichael Baum  *   The mlx5_kvargs structure.
362a729d2f0SMichael Baum  * @param keys
363a729d2f0SMichael Baum  *   A list of keys to process (table of const char *, the last must be NULL).
364a729d2f0SMichael Baum  * @param handler
365a729d2f0SMichael Baum  *   The function to call for each matching key.
366a729d2f0SMichael Baum  * @param opaque_arg
367a729d2f0SMichael Baum  *   A pointer passed unchanged to the handler.
368a729d2f0SMichael Baum  *
369a729d2f0SMichael Baum  * @return
370a729d2f0SMichael Baum  *   - 0 on success
371a729d2f0SMichael Baum  *   - Negative on error
372a729d2f0SMichael Baum  */
373a729d2f0SMichael Baum __rte_internal
374a729d2f0SMichael Baum int
375a729d2f0SMichael Baum mlx5_kvargs_process(struct mlx5_kvargs_ctrl *mkvlist, const char *const keys[],
376a729d2f0SMichael Baum 		    arg_handler_t handler, void *opaque_arg);
377a729d2f0SMichael Baum 
3785dfa003dSMichael Baum /* All UAR arguments using doorbell register in datapath. */
3795dfa003dSMichael Baum struct mlx5_uar_data {
3805dfa003dSMichael Baum 	uint64_t *db;
3815dfa003dSMichael Baum 	/* The doorbell's virtual address mapped to the relevant HW UAR space.*/
3825dfa003dSMichael Baum #ifndef RTE_ARCH_64
3835dfa003dSMichael Baum 	rte_spinlock_t *sl_p;
3845dfa003dSMichael Baum 	/* Pointer to UAR access lock required for 32bit implementations. */
3855dfa003dSMichael Baum #endif /* RTE_ARCH_64 */
3865dfa003dSMichael Baum };
3875dfa003dSMichael Baum 
3885dfa003dSMichael Baum /* DevX UAR control structure. */
3895dfa003dSMichael Baum struct mlx5_uar {
3905dfa003dSMichael Baum 	struct mlx5_uar_data bf_db; /* UAR data for Blueflame register. */
3915dfa003dSMichael Baum 	struct mlx5_uar_data cq_db; /* UAR data for CQ arm db register. */
3925dfa003dSMichael Baum 	void *obj; /* DevX UAR object. */
3935dfa003dSMichael Baum 	bool dbnc; /* Doorbell mapped to non-cached region. */
3945dfa003dSMichael Baum #ifndef RTE_ARCH_64
3955dfa003dSMichael Baum 	rte_spinlock_t bf_sl;
3965dfa003dSMichael Baum 	rte_spinlock_t cq_sl;
3975dfa003dSMichael Baum 	/* UAR access locks required for 32bit implementations. */
3985dfa003dSMichael Baum #endif /* RTE_ARCH_64 */
3995dfa003dSMichael Baum };
4005dfa003dSMichael Baum 
4015dfa003dSMichael Baum /**
4025dfa003dSMichael Baum  * Ring a doorbell and flush the update if requested.
4035dfa003dSMichael Baum  *
4045dfa003dSMichael Baum  * @param uar
4055dfa003dSMichael Baum  *   Pointer to UAR data structure.
4065dfa003dSMichael Baum  * @param val
4075dfa003dSMichael Baum  *   value to write in big endian format.
4085dfa003dSMichael Baum  * @param index
4095dfa003dSMichael Baum  *   Index of doorbell record.
4105dfa003dSMichael Baum  * @param db_rec
4115dfa003dSMichael Baum  *   Address of doorbell record.
4125dfa003dSMichael Baum  * @param flash
4135dfa003dSMichael Baum  *   Decide whether to flush the DB writing using a memory barrier.
4145dfa003dSMichael Baum  */
4155dfa003dSMichael Baum static __rte_always_inline void
mlx5_doorbell_ring(struct mlx5_uar_data * uar,uint64_t val,uint32_t index,volatile uint32_t * db_rec,bool flash)4165dfa003dSMichael Baum mlx5_doorbell_ring(struct mlx5_uar_data *uar, uint64_t val, uint32_t index,
4175dfa003dSMichael Baum 		   volatile uint32_t *db_rec, bool flash)
4185dfa003dSMichael Baum {
4195dfa003dSMichael Baum 	rte_io_wmb();
4205dfa003dSMichael Baum 	*db_rec = rte_cpu_to_be_32(index);
4215dfa003dSMichael Baum 	/* Ensure ordering between DB record actual update and UAR access. */
4225dfa003dSMichael Baum 	rte_wmb();
4235dfa003dSMichael Baum #ifdef RTE_ARCH_64
4245dfa003dSMichael Baum 	*uar->db = val;
4255dfa003dSMichael Baum #else /* !RTE_ARCH_64 */
4265dfa003dSMichael Baum 	rte_spinlock_lock(uar->sl_p);
4275dfa003dSMichael Baum 	*(volatile uint32_t *)uar->db = val;
4285dfa003dSMichael Baum 	rte_io_wmb();
4295dfa003dSMichael Baum 	*((volatile uint32_t *)uar->db + 1) = val >> 32;
4305dfa003dSMichael Baum 	rte_spinlock_unlock(uar->sl_p);
4315dfa003dSMichael Baum #endif
4325dfa003dSMichael Baum 	if (flash)
4335dfa003dSMichael Baum 		rte_wmb();
4345dfa003dSMichael Baum }
4355dfa003dSMichael Baum 
4365dfa003dSMichael Baum /**
4375dfa003dSMichael Baum  * Get the doorbell register mapping type.
4385dfa003dSMichael Baum  *
4395dfa003dSMichael Baum  * @param uar_mmap_offset
4405dfa003dSMichael Baum  *   Mmap offset of Verbs/DevX UAR.
4415dfa003dSMichael Baum  * @param page_size
4425dfa003dSMichael Baum  *   System page size
4435dfa003dSMichael Baum  *
4445dfa003dSMichael Baum  * @return
4455dfa003dSMichael Baum  *   1 for non-cached, 0 otherwise.
4465dfa003dSMichael Baum  */
4475dfa003dSMichael Baum static inline uint16_t
mlx5_db_map_type_get(off_t uar_mmap_offset,size_t page_size)4485dfa003dSMichael Baum mlx5_db_map_type_get(off_t uar_mmap_offset, size_t page_size)
4495dfa003dSMichael Baum {
4505dfa003dSMichael Baum 	off_t cmd = uar_mmap_offset / page_size;
4515dfa003dSMichael Baum 
4525dfa003dSMichael Baum 	cmd >>= MLX5_UAR_MMAP_CMD_SHIFT;
4535dfa003dSMichael Baum 	cmd &= MLX5_UAR_MMAP_CMD_MASK;
4545dfa003dSMichael Baum 	if (cmd == MLX5_MMAP_GET_NC_PAGES_CMD)
4555dfa003dSMichael Baum 		return 1;
4565dfa003dSMichael Baum 	return 0;
4575dfa003dSMichael Baum }
4585dfa003dSMichael Baum 
45964c563f8SOphir Munk __rte_internal
460654810b5SMatan Azrad void mlx5_translate_port_name(const char *port_name_in,
461654810b5SMatan Azrad 			      struct mlx5_switch_info *port_info_out);
46279aa4307SOphir Munk void mlx5_glue_constructor(void);
4634c204fe5SShiri Kuzin extern uint8_t haswell_broadwell_cpu;
4644c204fe5SShiri Kuzin 
46582088001SParav Pandit __rte_internal
46682088001SParav Pandit void mlx5_common_init(void);
46782088001SParav Pandit 
468ad435d32SXueming Li /*
469ad435d32SXueming Li  * Common Driver Interface
470ad435d32SXueming Li  *
471ad435d32SXueming Li  * ConnectX common driver supports multiple classes: net, vDPA, regex, crypto
472ad435d32SXueming Li  * and compress devices. This layer enables creating such multiple classes
473ad435d32SXueming Li  * on a single device by allowing to bind multiple class-specific device
474ad435d32SXueming Li  * drivers to attach to the common driver.
475ad435d32SXueming Li  *
476ad435d32SXueming Li  * ------------  -------------  --------------  -----------------  ------------
477ad435d32SXueming Li  * | mlx5 net |  | mlx5 vdpa |  | mlx5 regex |  | mlx5 compress |  | mlx5 ... |
478ad435d32SXueming Li  * |  driver  |  |  driver   |  |   driver   |  |     driver    |  |  drivers |
479ad435d32SXueming Li  * ------------  -------------  --------------  -----------------  ------------
480ad435d32SXueming Li  *                               ||
481ad435d32SXueming Li  *                        -----------------
482ad435d32SXueming Li  *                        |     mlx5      |
483ad435d32SXueming Li  *                        | common driver |
484ad435d32SXueming Li  *                        -----------------
485ad435d32SXueming Li  *                          |          |
486ad435d32SXueming Li  *                 -----------        -----------------
487ad435d32SXueming Li  *                 |   mlx5  |        |   mlx5        |
488ad435d32SXueming Li  *                 | pci dev |        | auxiliary dev |
489ad435d32SXueming Li  *                 -----------        -----------------
490ad435d32SXueming Li  *
491ad435d32SXueming Li  * - mlx5 PCI bus driver binds to mlx5 PCI devices defined by PCI ID table
492ad435d32SXueming Li  *   of all related devices.
493ad435d32SXueming Li  * - mlx5 class driver such as net, vDPA, regex defines its specific
494ad435d32SXueming Li  *   PCI ID table and mlx5 bus driver probes matching class drivers.
495ad435d32SXueming Li  * - mlx5 common driver is central place that validates supported
496ad435d32SXueming Li  *   class combinations.
497ad435d32SXueming Li  * - mlx5 common driver hides bus difference by resolving device address
498ad435d32SXueming Li  *   from devargs, locating target RDMA device and probing with it.
499ad435d32SXueming Li  */
500ad435d32SXueming Li 
50185209924SMichael Baum /*
50285209924SMichael Baum  * Device configuration structure.
50385209924SMichael Baum  *
50485209924SMichael Baum  * Merged configuration from:
50585209924SMichael Baum  *
50685209924SMichael Baum  *  - Device capabilities,
50785209924SMichael Baum  *  - User device parameters disabled features.
50885209924SMichael Baum  */
50985209924SMichael Baum struct mlx5_common_dev_config {
510fe46b20cSMichael Baum 	struct mlx5_hca_attr hca_attr; /* HCA attributes. */
51185209924SMichael Baum 	int dbnc; /* Skip doorbell register write barrier. */
5129d936f4fSMichael Baum 	int device_fd; /* Device file descriptor for importation. */
5139d936f4fSMichael Baum 	int pd_handle; /* Protection Domain handle for importation.  */
514887183efSMichael Baum 	unsigned int devx:1; /* Whether devx interface is available or not. */
51585209924SMichael Baum 	unsigned int sys_mem_en:1; /* The default memory allocator. */
51685209924SMichael Baum 	unsigned int mr_mempool_reg_en:1;
51785209924SMichael Baum 	/* Allow/prevent implicit mempool memory registration. */
51885209924SMichael Baum 	unsigned int mr_ext_memseg_en:1;
51985209924SMichael Baum 	/* Whether memseg should be extended for MR creation. */
52085209924SMichael Baum };
52185209924SMichael Baum 
5227af08c8fSMichael Baum struct mlx5_common_device {
5237af08c8fSMichael Baum 	struct rte_device *dev;
5247af08c8fSMichael Baum 	TAILQ_ENTRY(mlx5_common_device) next;
5257af08c8fSMichael Baum 	uint32_t classes_loaded;
526ca1418ceSMichael Baum 	void *ctx; /* Verbs/DV/DevX context. */
527e35ccf24SMichael Baum 	void *pd; /* Protection Domain. */
528e35ccf24SMichael Baum 	uint32_t pdn; /* Protection Domain Number. */
5299f1d636fSMichael Baum 	struct mlx5_mr_share_cache mr_scache; /* Global shared MR cache. */
53085209924SMichael Baum 	struct mlx5_common_dev_config config; /* Device configuration. */
5317af08c8fSMichael Baum };
5327af08c8fSMichael Baum 
533ad435d32SXueming Li /**
5349d936f4fSMichael Baum  * Indicates whether PD and CTX are imported from another process,
5359d936f4fSMichael Baum  * or created by this process.
5369d936f4fSMichael Baum  *
5379d936f4fSMichael Baum  * @param cdev
5389d936f4fSMichael Baum  *   Pointer to common device.
5399d936f4fSMichael Baum  *
5409d936f4fSMichael Baum  * @return
5419d936f4fSMichael Baum  *   True if PD and CTX are imported from another process, False otherwise.
5429d936f4fSMichael Baum  */
5439d936f4fSMichael Baum static inline bool
mlx5_imported_pd_and_ctx(struct mlx5_common_device * cdev)5449d936f4fSMichael Baum mlx5_imported_pd_and_ctx(struct mlx5_common_device *cdev)
5459d936f4fSMichael Baum {
5469d936f4fSMichael Baum 	return cdev->config.device_fd != MLX5_ARG_UNSET &&
5479d936f4fSMichael Baum 	       cdev->config.pd_handle != MLX5_ARG_UNSET;
5489d936f4fSMichael Baum }
5499d936f4fSMichael Baum 
5509d936f4fSMichael Baum /**
551ad435d32SXueming Li  * Initialization function for the driver called during device probing.
552ad435d32SXueming Li  */
553a729d2f0SMichael Baum typedef int (mlx5_class_driver_probe_t)(struct mlx5_common_device *cdev,
554a729d2f0SMichael Baum 					struct mlx5_kvargs_ctrl *mkvlist);
555ad435d32SXueming Li 
556ad435d32SXueming Li /**
557ad435d32SXueming Li  * Uninitialization function for the driver called during hot-unplugging.
558ad435d32SXueming Li  */
559a729d2f0SMichael Baum typedef int (mlx5_class_driver_remove_t)(struct mlx5_common_device *cdev);
560ad435d32SXueming Li 
561ad435d32SXueming Li /** Device already probed can be probed again to check for new ports. */
562ad435d32SXueming Li #define MLX5_DRV_PROBE_AGAIN 0x0004
563ad435d32SXueming Li 
564ad435d32SXueming Li /**
565ad435d32SXueming Li  * A structure describing a mlx5 common class driver.
566ad435d32SXueming Li  */
567ad435d32SXueming Li struct mlx5_class_driver {
568ad435d32SXueming Li 	TAILQ_ENTRY(mlx5_class_driver) next;
569ad435d32SXueming Li 	enum mlx5_class drv_class;            /**< Class of this driver. */
570ad435d32SXueming Li 	const char *name;                     /**< Driver name. */
571ad435d32SXueming Li 	mlx5_class_driver_probe_t *probe;     /**< Device probe function. */
572ad435d32SXueming Li 	mlx5_class_driver_remove_t *remove;   /**< Device remove function. */
573ad435d32SXueming Li 	const struct rte_pci_id *id_table;    /**< ID table, NULL terminated. */
574ad435d32SXueming Li 	uint32_t probe_again:1;
575ad435d32SXueming Li 	/**< Device already probed can be probed again to check new device. */
576ad435d32SXueming Li 	uint32_t intr_lsc:1; /**< Supports link state interrupt. */
577ad435d32SXueming Li 	uint32_t intr_rmv:1; /**< Supports device remove interrupt. */
578ad435d32SXueming Li };
579ad435d32SXueming Li 
580ad435d32SXueming Li /**
581ad435d32SXueming Li  * Register a mlx5 device driver.
582ad435d32SXueming Li  *
583ad435d32SXueming Li  * @param driver
584ad435d32SXueming Li  *   A pointer to a mlx5_driver structure describing the driver
585ad435d32SXueming Li  *   to be registered.
586ad435d32SXueming Li  */
587ad435d32SXueming Li __rte_internal
588ad435d32SXueming Li void
589ad435d32SXueming Li mlx5_class_driver_register(struct mlx5_class_driver *driver);
590ad435d32SXueming Li 
591ad435d32SXueming Li /**
592ad435d32SXueming Li  * Test device is a PCI bus device.
593ad435d32SXueming Li  *
594ad435d32SXueming Li  * @param dev
595ad435d32SXueming Li  *   Pointer to device.
596ad435d32SXueming Li  *
597ad435d32SXueming Li  * @return
598ad435d32SXueming Li  *   - True on device devargs is a PCI bus device.
599ad435d32SXueming Li  *   - False otherwise.
600ad435d32SXueming Li  */
601ad435d32SXueming Li __rte_internal
602ad435d32SXueming Li bool
603ad435d32SXueming Li mlx5_dev_is_pci(const struct rte_device *dev);
604ad435d32SXueming Li 
605c4c3e8afSMichael Baum /**
606c4c3e8afSMichael Baum  * Test PCI device is a VF device.
607c4c3e8afSMichael Baum  *
608c4c3e8afSMichael Baum  * @param pci_dev
609c4c3e8afSMichael Baum  *   Pointer to PCI device.
610c4c3e8afSMichael Baum  *
611c4c3e8afSMichael Baum  * @return
612c4c3e8afSMichael Baum  *   - True on PCI device is a VF device.
613c4c3e8afSMichael Baum  *   - False otherwise.
614c4c3e8afSMichael Baum  */
615c4c3e8afSMichael Baum __rte_internal
616c4c3e8afSMichael Baum bool
61737ca457dSBing Zhao mlx5_dev_is_vf_pci(const struct rte_pci_device *pci_dev);
618c4c3e8afSMichael Baum 
619fc59a1ecSMichael Baum __rte_internal
620fc59a1ecSMichael Baum int
621fc59a1ecSMichael Baum mlx5_dev_mempool_subscribe(struct mlx5_common_device *cdev);
622fc59a1ecSMichael Baum 
623fc59a1ecSMichael Baum __rte_internal
624fc59a1ecSMichael Baum void
625fc59a1ecSMichael Baum mlx5_dev_mempool_unregister(struct mlx5_common_device *cdev,
626fc59a1ecSMichael Baum 			    struct rte_mempool *mp);
627fc59a1ecSMichael Baum 
628b4371d3dSMichael Baum __rte_internal
6295dfa003dSMichael Baum int
6305dfa003dSMichael Baum mlx5_devx_uar_prepare(struct mlx5_common_device *cdev, struct mlx5_uar *uar);
6315dfa003dSMichael Baum 
6325dfa003dSMichael Baum __rte_internal
6335dfa003dSMichael Baum void
6345dfa003dSMichael Baum mlx5_devx_uar_release(struct mlx5_uar *uar);
635b4371d3dSMichael Baum 
636887183efSMichael Baum /* mlx5_common_os.c */
637887183efSMichael Baum 
638ca1418ceSMichael Baum int mlx5_os_open_device(struct mlx5_common_device *cdev, uint32_t classes);
6399d936f4fSMichael Baum int mlx5_os_pd_prepare(struct mlx5_common_device *cdev);
6409d936f4fSMichael Baum int mlx5_os_pd_release(struct mlx5_common_device *cdev);
6419d936f4fSMichael Baum int mlx5_os_remote_pd_and_ctx_validate(struct mlx5_common_dev_config *config);
642887183efSMichael Baum 
64376b5bdf8SMatan Azrad /* mlx5 PMD wrapped MR struct. */
64476b5bdf8SMatan Azrad struct mlx5_pmd_wrapped_mr {
64576b5bdf8SMatan Azrad 	uint32_t	     lkey;
64676b5bdf8SMatan Azrad 	void		     *addr;
64776b5bdf8SMatan Azrad 	size_t		     len;
64876b5bdf8SMatan Azrad 	void		     *obj; /* verbs mr object or devx umem object. */
64976b5bdf8SMatan Azrad 	void		     *imkey; /* DevX indirect mkey object. */
65076b5bdf8SMatan Azrad };
65176b5bdf8SMatan Azrad 
65276b5bdf8SMatan Azrad __rte_internal
65376b5bdf8SMatan Azrad int
65476b5bdf8SMatan Azrad mlx5_os_wrapped_mkey_create(void *ctx, void *pd, uint32_t pdn, void *addr,
65576b5bdf8SMatan Azrad 			    size_t length, struct mlx5_pmd_wrapped_mr *pmd_mr);
65676b5bdf8SMatan Azrad 
65776b5bdf8SMatan Azrad __rte_internal
65876b5bdf8SMatan Azrad void
65976b5bdf8SMatan Azrad mlx5_os_wrapped_mkey_destroy(struct mlx5_pmd_wrapped_mr *pmd_mr);
66076b5bdf8SMatan Azrad 
6617b4f1e6bSMatan Azrad #endif /* RTE_PMD_MLX5_COMMON_H_ */
662