11256805dSOphir Munk /* SPDX-License-Identifier: BSD-3-Clause 21256805dSOphir Munk * Copyright 2015 6WIND S.A. 31256805dSOphir Munk * Copyright 2015 Mellanox Technologies, Ltd 41256805dSOphir Munk */ 51256805dSOphir Munk 61256805dSOphir Munk #include <stddef.h> 71256805dSOphir Munk #include <inttypes.h> 81256805dSOphir Munk #include <unistd.h> 91256805dSOphir Munk #include <stdbool.h> 101256805dSOphir Munk #include <stdint.h> 111256805dSOphir Munk #include <stdio.h> 121256805dSOphir Munk #include <string.h> 131256805dSOphir Munk #include <stdlib.h> 141256805dSOphir Munk #include <errno.h> 151256805dSOphir Munk #include <dirent.h> 161256805dSOphir Munk #include <net/if.h> 171256805dSOphir Munk #include <sys/ioctl.h> 181256805dSOphir Munk #include <sys/socket.h> 191256805dSOphir Munk #include <netinet/in.h> 201256805dSOphir Munk #include <linux/ethtool.h> 211256805dSOphir Munk #include <linux/sockios.h> 221256805dSOphir Munk #include <fcntl.h> 231256805dSOphir Munk #include <stdalign.h> 241256805dSOphir Munk #include <sys/un.h> 251256805dSOphir Munk #include <time.h> 261256805dSOphir Munk 271256805dSOphir Munk #include <rte_atomic.h> 281256805dSOphir Munk #include <rte_ethdev_driver.h> 291256805dSOphir Munk #include <rte_bus_pci.h> 301256805dSOphir Munk #include <rte_mbuf.h> 311256805dSOphir Munk #include <rte_common.h> 321256805dSOphir Munk #include <rte_interrupts.h> 331256805dSOphir Munk #include <rte_malloc.h> 341256805dSOphir Munk #include <rte_string_fns.h> 351256805dSOphir Munk #include <rte_rwlock.h> 361256805dSOphir Munk #include <rte_cycles.h> 371256805dSOphir Munk 381256805dSOphir Munk #include <mlx5_glue.h> 391256805dSOphir Munk #include <mlx5_devx_cmds.h> 401256805dSOphir Munk #include <mlx5_common.h> 412175c4dcSSuanming Mou #include <mlx5_malloc.h> 421256805dSOphir Munk 431256805dSOphir Munk #include "mlx5.h" 441256805dSOphir Munk #include "mlx5_rxtx.h" 451256805dSOphir Munk #include "mlx5_utils.h" 461256805dSOphir Munk 471256805dSOphir Munk /* Supported speed values found in /usr/include/linux/ethtool.h */ 481256805dSOphir Munk #ifndef HAVE_SUPPORTED_40000baseKR4_Full 491256805dSOphir Munk #define SUPPORTED_40000baseKR4_Full (1 << 23) 501256805dSOphir Munk #endif 511256805dSOphir Munk #ifndef HAVE_SUPPORTED_40000baseCR4_Full 521256805dSOphir Munk #define SUPPORTED_40000baseCR4_Full (1 << 24) 531256805dSOphir Munk #endif 541256805dSOphir Munk #ifndef HAVE_SUPPORTED_40000baseSR4_Full 551256805dSOphir Munk #define SUPPORTED_40000baseSR4_Full (1 << 25) 561256805dSOphir Munk #endif 571256805dSOphir Munk #ifndef HAVE_SUPPORTED_40000baseLR4_Full 581256805dSOphir Munk #define SUPPORTED_40000baseLR4_Full (1 << 26) 591256805dSOphir Munk #endif 601256805dSOphir Munk #ifndef HAVE_SUPPORTED_56000baseKR4_Full 611256805dSOphir Munk #define SUPPORTED_56000baseKR4_Full (1 << 27) 621256805dSOphir Munk #endif 631256805dSOphir Munk #ifndef HAVE_SUPPORTED_56000baseCR4_Full 641256805dSOphir Munk #define SUPPORTED_56000baseCR4_Full (1 << 28) 651256805dSOphir Munk #endif 661256805dSOphir Munk #ifndef HAVE_SUPPORTED_56000baseSR4_Full 671256805dSOphir Munk #define SUPPORTED_56000baseSR4_Full (1 << 29) 681256805dSOphir Munk #endif 691256805dSOphir Munk #ifndef HAVE_SUPPORTED_56000baseLR4_Full 701256805dSOphir Munk #define SUPPORTED_56000baseLR4_Full (1 << 30) 711256805dSOphir Munk #endif 721256805dSOphir Munk 731256805dSOphir Munk /* Add defines in case the running kernel is not the same as user headers. */ 741256805dSOphir Munk #ifndef ETHTOOL_GLINKSETTINGS 751256805dSOphir Munk struct ethtool_link_settings { 761256805dSOphir Munk uint32_t cmd; 771256805dSOphir Munk uint32_t speed; 781256805dSOphir Munk uint8_t duplex; 791256805dSOphir Munk uint8_t port; 801256805dSOphir Munk uint8_t phy_address; 811256805dSOphir Munk uint8_t autoneg; 821256805dSOphir Munk uint8_t mdio_support; 831256805dSOphir Munk uint8_t eth_to_mdix; 841256805dSOphir Munk uint8_t eth_tp_mdix_ctrl; 851256805dSOphir Munk int8_t link_mode_masks_nwords; 861256805dSOphir Munk uint32_t reserved[8]; 871256805dSOphir Munk uint32_t link_mode_masks[]; 881256805dSOphir Munk }; 891256805dSOphir Munk 901256805dSOphir Munk /* The kernel values can be found in /include/uapi/linux/ethtool.h */ 911256805dSOphir Munk #define ETHTOOL_GLINKSETTINGS 0x0000004c 921256805dSOphir Munk #define ETHTOOL_LINK_MODE_1000baseT_Full_BIT 5 931256805dSOphir Munk #define ETHTOOL_LINK_MODE_Autoneg_BIT 6 941256805dSOphir Munk #define ETHTOOL_LINK_MODE_1000baseKX_Full_BIT 17 951256805dSOphir Munk #define ETHTOOL_LINK_MODE_10000baseKX4_Full_BIT 18 961256805dSOphir Munk #define ETHTOOL_LINK_MODE_10000baseKR_Full_BIT 19 971256805dSOphir Munk #define ETHTOOL_LINK_MODE_10000baseR_FEC_BIT 20 981256805dSOphir Munk #define ETHTOOL_LINK_MODE_20000baseMLD2_Full_BIT 21 991256805dSOphir Munk #define ETHTOOL_LINK_MODE_20000baseKR2_Full_BIT 22 1001256805dSOphir Munk #define ETHTOOL_LINK_MODE_40000baseKR4_Full_BIT 23 1011256805dSOphir Munk #define ETHTOOL_LINK_MODE_40000baseCR4_Full_BIT 24 1021256805dSOphir Munk #define ETHTOOL_LINK_MODE_40000baseSR4_Full_BIT 25 1031256805dSOphir Munk #define ETHTOOL_LINK_MODE_40000baseLR4_Full_BIT 26 1041256805dSOphir Munk #define ETHTOOL_LINK_MODE_56000baseKR4_Full_BIT 27 1051256805dSOphir Munk #define ETHTOOL_LINK_MODE_56000baseCR4_Full_BIT 28 1061256805dSOphir Munk #define ETHTOOL_LINK_MODE_56000baseSR4_Full_BIT 29 1071256805dSOphir Munk #define ETHTOOL_LINK_MODE_56000baseLR4_Full_BIT 30 1081256805dSOphir Munk #endif 1091256805dSOphir Munk #ifndef HAVE_ETHTOOL_LINK_MODE_25G 1101256805dSOphir Munk #define ETHTOOL_LINK_MODE_25000baseCR_Full_BIT 31 1111256805dSOphir Munk #define ETHTOOL_LINK_MODE_25000baseKR_Full_BIT 32 1121256805dSOphir Munk #define ETHTOOL_LINK_MODE_25000baseSR_Full_BIT 33 1131256805dSOphir Munk #endif 1141256805dSOphir Munk #ifndef HAVE_ETHTOOL_LINK_MODE_50G 1151256805dSOphir Munk #define ETHTOOL_LINK_MODE_50000baseCR2_Full_BIT 34 1161256805dSOphir Munk #define ETHTOOL_LINK_MODE_50000baseKR2_Full_BIT 35 1171256805dSOphir Munk #endif 1181256805dSOphir Munk #ifndef HAVE_ETHTOOL_LINK_MODE_100G 1191256805dSOphir Munk #define ETHTOOL_LINK_MODE_100000baseKR4_Full_BIT 36 1201256805dSOphir Munk #define ETHTOOL_LINK_MODE_100000baseSR4_Full_BIT 37 1211256805dSOphir Munk #define ETHTOOL_LINK_MODE_100000baseCR4_Full_BIT 38 1221256805dSOphir Munk #define ETHTOOL_LINK_MODE_100000baseLR4_ER4_Full_BIT 39 1231256805dSOphir Munk #endif 1241256805dSOphir Munk #ifndef HAVE_ETHTOOL_LINK_MODE_200G 1251256805dSOphir Munk #define ETHTOOL_LINK_MODE_200000baseKR4_Full_BIT 62 1261256805dSOphir Munk #define ETHTOOL_LINK_MODE_200000baseSR4_Full_BIT 63 1271256805dSOphir Munk #define ETHTOOL_LINK_MODE_200000baseLR4_ER4_FR4_Full_BIT 0 /* 64 - 64 */ 1281256805dSOphir Munk #define ETHTOOL_LINK_MODE_200000baseDR4_Full_BIT 1 /* 65 - 64 */ 1291256805dSOphir Munk #define ETHTOOL_LINK_MODE_200000baseCR4_Full_BIT 2 /* 66 - 64 */ 1301256805dSOphir Munk #endif 1311256805dSOphir Munk 1321256805dSOphir Munk 1331256805dSOphir Munk /** 1341256805dSOphir Munk * Get interface name from private structure. 1351256805dSOphir Munk * 136aec086c9SMatan Azrad * This is a port representor-aware version of mlx5_get_ifname_sysfs(). 1371256805dSOphir Munk * 1381256805dSOphir Munk * @param[in] dev 1391256805dSOphir Munk * Pointer to Ethernet device. 1401256805dSOphir Munk * @param[out] ifname 1411256805dSOphir Munk * Interface name output buffer. 1421256805dSOphir Munk * 1431256805dSOphir Munk * @return 1441256805dSOphir Munk * 0 on success, a negative errno value otherwise and rte_errno is set. 1451256805dSOphir Munk */ 1461256805dSOphir Munk int 1471256805dSOphir Munk mlx5_get_ifname(const struct rte_eth_dev *dev, char (*ifname)[IF_NAMESIZE]) 1481256805dSOphir Munk { 1491256805dSOphir Munk struct mlx5_priv *priv = dev->data->dev_private; 1501256805dSOphir Munk unsigned int ifindex; 1511256805dSOphir Munk 1521256805dSOphir Munk MLX5_ASSERT(priv); 1531256805dSOphir Munk MLX5_ASSERT(priv->sh); 1541256805dSOphir Munk ifindex = mlx5_ifindex(dev); 1551256805dSOphir Munk if (!ifindex) { 1561256805dSOphir Munk if (!priv->representor) 157aec086c9SMatan Azrad return mlx5_get_ifname_sysfs(priv->sh->ibdev_path, 158aec086c9SMatan Azrad *ifname); 1591256805dSOphir Munk rte_errno = ENXIO; 1601256805dSOphir Munk return -rte_errno; 1611256805dSOphir Munk } 1621256805dSOphir Munk if (if_indextoname(ifindex, &(*ifname)[0])) 1631256805dSOphir Munk return 0; 1641256805dSOphir Munk rte_errno = errno; 1651256805dSOphir Munk return -rte_errno; 1661256805dSOphir Munk } 1671256805dSOphir Munk 1681256805dSOphir Munk /** 1691256805dSOphir Munk * Perform ifreq ioctl() on associated Ethernet device. 1701256805dSOphir Munk * 1711256805dSOphir Munk * @param[in] dev 1721256805dSOphir Munk * Pointer to Ethernet device. 1731256805dSOphir Munk * @param req 1741256805dSOphir Munk * Request number to pass to ioctl(). 1751256805dSOphir Munk * @param[out] ifr 1761256805dSOphir Munk * Interface request structure output buffer. 1771256805dSOphir Munk * 1781256805dSOphir Munk * @return 1791256805dSOphir Munk * 0 on success, a negative errno value otherwise and rte_errno is set. 1801256805dSOphir Munk */ 18198c4b12aSOphir Munk static int 1821256805dSOphir Munk mlx5_ifreq(const struct rte_eth_dev *dev, int req, struct ifreq *ifr) 1831256805dSOphir Munk { 1841256805dSOphir Munk int sock = socket(PF_INET, SOCK_DGRAM, IPPROTO_IP); 1851256805dSOphir Munk int ret = 0; 1861256805dSOphir Munk 1871256805dSOphir Munk if (sock == -1) { 1881256805dSOphir Munk rte_errno = errno; 1891256805dSOphir Munk return -rte_errno; 1901256805dSOphir Munk } 1911256805dSOphir Munk ret = mlx5_get_ifname(dev, &ifr->ifr_name); 1921256805dSOphir Munk if (ret) 1931256805dSOphir Munk goto error; 1941256805dSOphir Munk ret = ioctl(sock, req, ifr); 1951256805dSOphir Munk if (ret == -1) { 1961256805dSOphir Munk rte_errno = errno; 1971256805dSOphir Munk goto error; 1981256805dSOphir Munk } 1991256805dSOphir Munk close(sock); 2001256805dSOphir Munk return 0; 2011256805dSOphir Munk error: 2021256805dSOphir Munk close(sock); 2031256805dSOphir Munk return -rte_errno; 2041256805dSOphir Munk } 2051256805dSOphir Munk 2061256805dSOphir Munk /** 2071256805dSOphir Munk * Get device MTU. 2081256805dSOphir Munk * 2091256805dSOphir Munk * @param dev 2101256805dSOphir Munk * Pointer to Ethernet device. 2111256805dSOphir Munk * @param[out] mtu 2121256805dSOphir Munk * MTU value output buffer. 2131256805dSOphir Munk * 2141256805dSOphir Munk * @return 2151256805dSOphir Munk * 0 on success, a negative errno value otherwise and rte_errno is set. 2161256805dSOphir Munk */ 2171256805dSOphir Munk int 2181256805dSOphir Munk mlx5_get_mtu(struct rte_eth_dev *dev, uint16_t *mtu) 2191256805dSOphir Munk { 2201256805dSOphir Munk struct ifreq request; 2211256805dSOphir Munk int ret = mlx5_ifreq(dev, SIOCGIFMTU, &request); 2221256805dSOphir Munk 2231256805dSOphir Munk if (ret) 2241256805dSOphir Munk return ret; 2251256805dSOphir Munk *mtu = request.ifr_mtu; 2261256805dSOphir Munk return 0; 2271256805dSOphir Munk } 2281256805dSOphir Munk 2291256805dSOphir Munk /** 2301256805dSOphir Munk * Set device MTU. 2311256805dSOphir Munk * 2321256805dSOphir Munk * @param dev 2331256805dSOphir Munk * Pointer to Ethernet device. 2341256805dSOphir Munk * @param mtu 2351256805dSOphir Munk * MTU value to set. 2361256805dSOphir Munk * 2371256805dSOphir Munk * @return 2381256805dSOphir Munk * 0 on success, a negative errno value otherwise and rte_errno is set. 2391256805dSOphir Munk */ 2401256805dSOphir Munk int 2411256805dSOphir Munk mlx5_set_mtu(struct rte_eth_dev *dev, uint16_t mtu) 2421256805dSOphir Munk { 2431256805dSOphir Munk struct ifreq request = { .ifr_mtu = mtu, }; 2441256805dSOphir Munk 2451256805dSOphir Munk return mlx5_ifreq(dev, SIOCSIFMTU, &request); 2461256805dSOphir Munk } 2471256805dSOphir Munk 2481256805dSOphir Munk /** 2491256805dSOphir Munk * Set device flags. 2501256805dSOphir Munk * 2511256805dSOphir Munk * @param dev 2521256805dSOphir Munk * Pointer to Ethernet device. 2531256805dSOphir Munk * @param keep 2541256805dSOphir Munk * Bitmask for flags that must remain untouched. 2551256805dSOphir Munk * @param flags 2561256805dSOphir Munk * Bitmask for flags to modify. 2571256805dSOphir Munk * 2581256805dSOphir Munk * @return 2591256805dSOphir Munk * 0 on success, a negative errno value otherwise and rte_errno is set. 2601256805dSOphir Munk */ 26198c4b12aSOphir Munk static int 2621256805dSOphir Munk mlx5_set_flags(struct rte_eth_dev *dev, unsigned int keep, unsigned int flags) 2631256805dSOphir Munk { 2641256805dSOphir Munk struct ifreq request; 2651256805dSOphir Munk int ret = mlx5_ifreq(dev, SIOCGIFFLAGS, &request); 2661256805dSOphir Munk 2671256805dSOphir Munk if (ret) 2681256805dSOphir Munk return ret; 2691256805dSOphir Munk request.ifr_flags &= keep; 2701256805dSOphir Munk request.ifr_flags |= flags & ~keep; 2711256805dSOphir Munk return mlx5_ifreq(dev, SIOCSIFFLAGS, &request); 2721256805dSOphir Munk } 2731256805dSOphir Munk 2741256805dSOphir Munk /** 2751256805dSOphir Munk * Get device current raw clock counter 2761256805dSOphir Munk * 2771256805dSOphir Munk * @param dev 2781256805dSOphir Munk * Pointer to Ethernet device structure. 2791256805dSOphir Munk * @param[out] time 2801256805dSOphir Munk * Current raw clock counter of the device. 2811256805dSOphir Munk * 2821256805dSOphir Munk * @return 2831256805dSOphir Munk * 0 if the clock has correctly been read 2841256805dSOphir Munk * The value of errno in case of error 2851256805dSOphir Munk */ 2861256805dSOphir Munk int 2871256805dSOphir Munk mlx5_read_clock(struct rte_eth_dev *dev, uint64_t *clock) 2881256805dSOphir Munk { 2891256805dSOphir Munk struct mlx5_priv *priv = dev->data->dev_private; 2901256805dSOphir Munk struct ibv_context *ctx = priv->sh->ctx; 2911256805dSOphir Munk struct ibv_values_ex values; 2921256805dSOphir Munk int err = 0; 2931256805dSOphir Munk 2941256805dSOphir Munk values.comp_mask = IBV_VALUES_MASK_RAW_CLOCK; 2951256805dSOphir Munk err = mlx5_glue->query_rt_values_ex(ctx, &values); 2961256805dSOphir Munk if (err != 0) { 2971256805dSOphir Munk DRV_LOG(WARNING, "Could not query the clock !"); 2981256805dSOphir Munk return err; 2991256805dSOphir Munk } 3001256805dSOphir Munk *clock = values.raw_clock.tv_nsec; 3011256805dSOphir Munk return 0; 3021256805dSOphir Munk } 3031256805dSOphir Munk 3041256805dSOphir Munk /** 3051256805dSOphir Munk * Retrieve the master device for representor in the same switch domain. 3061256805dSOphir Munk * 3071256805dSOphir Munk * @param dev 3081256805dSOphir Munk * Pointer to representor Ethernet device structure. 3091256805dSOphir Munk * 3101256805dSOphir Munk * @return 3111256805dSOphir Munk * Master device structure on success, NULL otherwise. 3121256805dSOphir Munk */ 3131256805dSOphir Munk static struct rte_eth_dev * 3141256805dSOphir Munk mlx5_find_master_dev(struct rte_eth_dev *dev) 3151256805dSOphir Munk { 3161256805dSOphir Munk struct mlx5_priv *priv; 3171256805dSOphir Munk uint16_t port_id; 3181256805dSOphir Munk uint16_t domain_id; 3191256805dSOphir Munk 3201256805dSOphir Munk priv = dev->data->dev_private; 3211256805dSOphir Munk domain_id = priv->domain_id; 3221256805dSOphir Munk MLX5_ASSERT(priv->representor); 3231256805dSOphir Munk MLX5_ETH_FOREACH_DEV(port_id, priv->pci_dev) { 3241256805dSOphir Munk struct mlx5_priv *opriv = 3251256805dSOphir Munk rte_eth_devices[port_id].data->dev_private; 3261256805dSOphir Munk if (opriv && 3271256805dSOphir Munk opriv->master && 3281256805dSOphir Munk opriv->domain_id == domain_id && 3291256805dSOphir Munk opriv->sh == priv->sh) 3301256805dSOphir Munk return &rte_eth_devices[port_id]; 3311256805dSOphir Munk } 3321256805dSOphir Munk return NULL; 3331256805dSOphir Munk } 3341256805dSOphir Munk 3351256805dSOphir Munk /** 3361256805dSOphir Munk * DPDK callback to retrieve physical link information. 3371256805dSOphir Munk * 3381256805dSOphir Munk * @param dev 3391256805dSOphir Munk * Pointer to Ethernet device structure. 3401256805dSOphir Munk * @param[out] link 3411256805dSOphir Munk * Storage for current link status. 3421256805dSOphir Munk * 3431256805dSOphir Munk * @return 3441256805dSOphir Munk * 0 on success, a negative errno value otherwise and rte_errno is set. 3451256805dSOphir Munk */ 3461256805dSOphir Munk static int 3471256805dSOphir Munk mlx5_link_update_unlocked_gset(struct rte_eth_dev *dev, 3481256805dSOphir Munk struct rte_eth_link *link) 3491256805dSOphir Munk { 3501256805dSOphir Munk struct mlx5_priv *priv = dev->data->dev_private; 3511256805dSOphir Munk struct ethtool_cmd edata = { 3521256805dSOphir Munk .cmd = ETHTOOL_GSET /* Deprecated since Linux v4.5. */ 3531256805dSOphir Munk }; 3541256805dSOphir Munk struct ifreq ifr; 3551256805dSOphir Munk struct rte_eth_link dev_link; 3561256805dSOphir Munk int link_speed = 0; 3571256805dSOphir Munk int ret; 3581256805dSOphir Munk 3591256805dSOphir Munk ret = mlx5_ifreq(dev, SIOCGIFFLAGS, &ifr); 3601256805dSOphir Munk if (ret) { 3611256805dSOphir Munk DRV_LOG(WARNING, "port %u ioctl(SIOCGIFFLAGS) failed: %s", 3621256805dSOphir Munk dev->data->port_id, strerror(rte_errno)); 3631256805dSOphir Munk return ret; 3641256805dSOphir Munk } 3651256805dSOphir Munk dev_link = (struct rte_eth_link) { 3661256805dSOphir Munk .link_status = ((ifr.ifr_flags & IFF_UP) && 3671256805dSOphir Munk (ifr.ifr_flags & IFF_RUNNING)), 3681256805dSOphir Munk }; 3691256805dSOphir Munk ifr = (struct ifreq) { 3701256805dSOphir Munk .ifr_data = (void *)&edata, 3711256805dSOphir Munk }; 3721256805dSOphir Munk ret = mlx5_ifreq(dev, SIOCETHTOOL, &ifr); 3731256805dSOphir Munk if (ret) { 3741256805dSOphir Munk if (ret == -ENOTSUP && priv->representor) { 3751256805dSOphir Munk struct rte_eth_dev *master; 3761256805dSOphir Munk 3771256805dSOphir Munk /* 3781256805dSOphir Munk * For representors we can try to inherit link 3791256805dSOphir Munk * settings from the master device. Actually 3801256805dSOphir Munk * link settings do not make a lot of sense 3811256805dSOphir Munk * for representors due to missing physical 3821256805dSOphir Munk * link. The old kernel drivers supported 3831256805dSOphir Munk * emulated settings query for representors, 3841256805dSOphir Munk * the new ones do not, so we have to add 3851256805dSOphir Munk * this code for compatibility issues. 3861256805dSOphir Munk */ 3871256805dSOphir Munk master = mlx5_find_master_dev(dev); 3881256805dSOphir Munk if (master) { 3891256805dSOphir Munk ifr = (struct ifreq) { 3901256805dSOphir Munk .ifr_data = (void *)&edata, 3911256805dSOphir Munk }; 3921256805dSOphir Munk ret = mlx5_ifreq(master, SIOCETHTOOL, &ifr); 3931256805dSOphir Munk } 3941256805dSOphir Munk } 3951256805dSOphir Munk if (ret) { 3961256805dSOphir Munk DRV_LOG(WARNING, 3971256805dSOphir Munk "port %u ioctl(SIOCETHTOOL," 3981256805dSOphir Munk " ETHTOOL_GSET) failed: %s", 3991256805dSOphir Munk dev->data->port_id, strerror(rte_errno)); 4001256805dSOphir Munk return ret; 4011256805dSOphir Munk } 4021256805dSOphir Munk } 4031256805dSOphir Munk link_speed = ethtool_cmd_speed(&edata); 4041256805dSOphir Munk if (link_speed == -1) 4051256805dSOphir Munk dev_link.link_speed = ETH_SPEED_NUM_NONE; 4061256805dSOphir Munk else 4071256805dSOphir Munk dev_link.link_speed = link_speed; 4081256805dSOphir Munk priv->link_speed_capa = 0; 4091256805dSOphir Munk if (edata.supported & SUPPORTED_Autoneg) 4101256805dSOphir Munk priv->link_speed_capa |= ETH_LINK_SPEED_AUTONEG; 4111256805dSOphir Munk if (edata.supported & (SUPPORTED_1000baseT_Full | 4121256805dSOphir Munk SUPPORTED_1000baseKX_Full)) 4131256805dSOphir Munk priv->link_speed_capa |= ETH_LINK_SPEED_1G; 4141256805dSOphir Munk if (edata.supported & SUPPORTED_10000baseKR_Full) 4151256805dSOphir Munk priv->link_speed_capa |= ETH_LINK_SPEED_10G; 4161256805dSOphir Munk if (edata.supported & (SUPPORTED_40000baseKR4_Full | 4171256805dSOphir Munk SUPPORTED_40000baseCR4_Full | 4181256805dSOphir Munk SUPPORTED_40000baseSR4_Full | 4191256805dSOphir Munk SUPPORTED_40000baseLR4_Full)) 4201256805dSOphir Munk priv->link_speed_capa |= ETH_LINK_SPEED_40G; 4211256805dSOphir Munk dev_link.link_duplex = ((edata.duplex == DUPLEX_HALF) ? 4221256805dSOphir Munk ETH_LINK_HALF_DUPLEX : ETH_LINK_FULL_DUPLEX); 4231256805dSOphir Munk dev_link.link_autoneg = !(dev->data->dev_conf.link_speeds & 4241256805dSOphir Munk ETH_LINK_SPEED_FIXED); 4251256805dSOphir Munk if (((dev_link.link_speed && !dev_link.link_status) || 4261256805dSOphir Munk (!dev_link.link_speed && dev_link.link_status))) { 4271256805dSOphir Munk rte_errno = EAGAIN; 4281256805dSOphir Munk return -rte_errno; 4291256805dSOphir Munk } 4301256805dSOphir Munk *link = dev_link; 4311256805dSOphir Munk return 0; 4321256805dSOphir Munk } 4331256805dSOphir Munk 4341256805dSOphir Munk /** 4351256805dSOphir Munk * Retrieve physical link information (unlocked version using new ioctl). 4361256805dSOphir Munk * 4371256805dSOphir Munk * @param dev 4381256805dSOphir Munk * Pointer to Ethernet device structure. 4391256805dSOphir Munk * @param[out] link 4401256805dSOphir Munk * Storage for current link status. 4411256805dSOphir Munk * 4421256805dSOphir Munk * @return 4431256805dSOphir Munk * 0 on success, a negative errno value otherwise and rte_errno is set. 4441256805dSOphir Munk */ 4451256805dSOphir Munk static int 4461256805dSOphir Munk mlx5_link_update_unlocked_gs(struct rte_eth_dev *dev, 4471256805dSOphir Munk struct rte_eth_link *link) 4481256805dSOphir Munk 4491256805dSOphir Munk { 4501256805dSOphir Munk struct mlx5_priv *priv = dev->data->dev_private; 4511256805dSOphir Munk struct ethtool_link_settings gcmd = { .cmd = ETHTOOL_GLINKSETTINGS }; 4521256805dSOphir Munk struct ifreq ifr; 4531256805dSOphir Munk struct rte_eth_link dev_link; 4541256805dSOphir Munk struct rte_eth_dev *master = NULL; 4551256805dSOphir Munk uint64_t sc; 4561256805dSOphir Munk int ret; 4571256805dSOphir Munk 4581256805dSOphir Munk ret = mlx5_ifreq(dev, SIOCGIFFLAGS, &ifr); 4591256805dSOphir Munk if (ret) { 4601256805dSOphir Munk DRV_LOG(WARNING, "port %u ioctl(SIOCGIFFLAGS) failed: %s", 4611256805dSOphir Munk dev->data->port_id, strerror(rte_errno)); 4621256805dSOphir Munk return ret; 4631256805dSOphir Munk } 4641256805dSOphir Munk dev_link = (struct rte_eth_link) { 4651256805dSOphir Munk .link_status = ((ifr.ifr_flags & IFF_UP) && 4661256805dSOphir Munk (ifr.ifr_flags & IFF_RUNNING)), 4671256805dSOphir Munk }; 4681256805dSOphir Munk ifr = (struct ifreq) { 4691256805dSOphir Munk .ifr_data = (void *)&gcmd, 4701256805dSOphir Munk }; 4711256805dSOphir Munk ret = mlx5_ifreq(dev, SIOCETHTOOL, &ifr); 4721256805dSOphir Munk if (ret) { 4731256805dSOphir Munk if (ret == -ENOTSUP && priv->representor) { 4741256805dSOphir Munk /* 4751256805dSOphir Munk * For representors we can try to inherit link 4761256805dSOphir Munk * settings from the master device. Actually 4771256805dSOphir Munk * link settings do not make a lot of sense 4781256805dSOphir Munk * for representors due to missing physical 4791256805dSOphir Munk * link. The old kernel drivers supported 4801256805dSOphir Munk * emulated settings query for representors, 4811256805dSOphir Munk * the new ones do not, so we have to add 4821256805dSOphir Munk * this code for compatibility issues. 4831256805dSOphir Munk */ 4841256805dSOphir Munk master = mlx5_find_master_dev(dev); 4851256805dSOphir Munk if (master) { 4861256805dSOphir Munk ifr = (struct ifreq) { 4871256805dSOphir Munk .ifr_data = (void *)&gcmd, 4881256805dSOphir Munk }; 4891256805dSOphir Munk ret = mlx5_ifreq(master, SIOCETHTOOL, &ifr); 4901256805dSOphir Munk } 4911256805dSOphir Munk } 4921256805dSOphir Munk if (ret) { 4931256805dSOphir Munk DRV_LOG(DEBUG, 4941256805dSOphir Munk "port %u ioctl(SIOCETHTOOL," 4951256805dSOphir Munk " ETHTOOL_GLINKSETTINGS) failed: %s", 4961256805dSOphir Munk dev->data->port_id, strerror(rte_errno)); 4971256805dSOphir Munk return ret; 4981256805dSOphir Munk } 4991256805dSOphir Munk } 5001256805dSOphir Munk gcmd.link_mode_masks_nwords = -gcmd.link_mode_masks_nwords; 5011256805dSOphir Munk 5021256805dSOphir Munk alignas(struct ethtool_link_settings) 5031256805dSOphir Munk uint8_t data[offsetof(struct ethtool_link_settings, link_mode_masks) + 5041256805dSOphir Munk sizeof(uint32_t) * gcmd.link_mode_masks_nwords * 3]; 5051256805dSOphir Munk struct ethtool_link_settings *ecmd = (void *)data; 5061256805dSOphir Munk 5071256805dSOphir Munk *ecmd = gcmd; 5081256805dSOphir Munk ifr.ifr_data = (void *)ecmd; 5091256805dSOphir Munk ret = mlx5_ifreq(master ? master : dev, SIOCETHTOOL, &ifr); 5101256805dSOphir Munk if (ret) { 5111256805dSOphir Munk DRV_LOG(DEBUG, 5121256805dSOphir Munk "port %u ioctl(SIOCETHTOOL," 5131256805dSOphir Munk "ETHTOOL_GLINKSETTINGS) failed: %s", 5141256805dSOphir Munk dev->data->port_id, strerror(rte_errno)); 5151256805dSOphir Munk return ret; 5161256805dSOphir Munk } 5171256805dSOphir Munk dev_link.link_speed = (ecmd->speed == UINT32_MAX) ? ETH_SPEED_NUM_NONE : 5181256805dSOphir Munk ecmd->speed; 5191256805dSOphir Munk sc = ecmd->link_mode_masks[0] | 5201256805dSOphir Munk ((uint64_t)ecmd->link_mode_masks[1] << 32); 5211256805dSOphir Munk priv->link_speed_capa = 0; 5221256805dSOphir Munk if (sc & MLX5_BITSHIFT(ETHTOOL_LINK_MODE_Autoneg_BIT)) 5231256805dSOphir Munk priv->link_speed_capa |= ETH_LINK_SPEED_AUTONEG; 5241256805dSOphir Munk if (sc & (MLX5_BITSHIFT(ETHTOOL_LINK_MODE_1000baseT_Full_BIT) | 5251256805dSOphir Munk MLX5_BITSHIFT(ETHTOOL_LINK_MODE_1000baseKX_Full_BIT))) 5261256805dSOphir Munk priv->link_speed_capa |= ETH_LINK_SPEED_1G; 5271256805dSOphir Munk if (sc & (MLX5_BITSHIFT(ETHTOOL_LINK_MODE_10000baseKX4_Full_BIT) | 5281256805dSOphir Munk MLX5_BITSHIFT(ETHTOOL_LINK_MODE_10000baseKR_Full_BIT) | 5291256805dSOphir Munk MLX5_BITSHIFT(ETHTOOL_LINK_MODE_10000baseR_FEC_BIT))) 5301256805dSOphir Munk priv->link_speed_capa |= ETH_LINK_SPEED_10G; 5311256805dSOphir Munk if (sc & (MLX5_BITSHIFT(ETHTOOL_LINK_MODE_20000baseMLD2_Full_BIT) | 5321256805dSOphir Munk MLX5_BITSHIFT(ETHTOOL_LINK_MODE_20000baseKR2_Full_BIT))) 5331256805dSOphir Munk priv->link_speed_capa |= ETH_LINK_SPEED_20G; 5341256805dSOphir Munk if (sc & (MLX5_BITSHIFT(ETHTOOL_LINK_MODE_40000baseKR4_Full_BIT) | 5351256805dSOphir Munk MLX5_BITSHIFT(ETHTOOL_LINK_MODE_40000baseCR4_Full_BIT) | 5361256805dSOphir Munk MLX5_BITSHIFT(ETHTOOL_LINK_MODE_40000baseSR4_Full_BIT) | 5371256805dSOphir Munk MLX5_BITSHIFT(ETHTOOL_LINK_MODE_40000baseLR4_Full_BIT))) 5381256805dSOphir Munk priv->link_speed_capa |= ETH_LINK_SPEED_40G; 5391256805dSOphir Munk if (sc & (MLX5_BITSHIFT(ETHTOOL_LINK_MODE_56000baseKR4_Full_BIT) | 5401256805dSOphir Munk MLX5_BITSHIFT(ETHTOOL_LINK_MODE_56000baseCR4_Full_BIT) | 5411256805dSOphir Munk MLX5_BITSHIFT(ETHTOOL_LINK_MODE_56000baseSR4_Full_BIT) | 5421256805dSOphir Munk MLX5_BITSHIFT(ETHTOOL_LINK_MODE_56000baseLR4_Full_BIT))) 5431256805dSOphir Munk priv->link_speed_capa |= ETH_LINK_SPEED_56G; 5441256805dSOphir Munk if (sc & (MLX5_BITSHIFT(ETHTOOL_LINK_MODE_25000baseCR_Full_BIT) | 5451256805dSOphir Munk MLX5_BITSHIFT(ETHTOOL_LINK_MODE_25000baseKR_Full_BIT) | 5461256805dSOphir Munk MLX5_BITSHIFT(ETHTOOL_LINK_MODE_25000baseSR_Full_BIT))) 5471256805dSOphir Munk priv->link_speed_capa |= ETH_LINK_SPEED_25G; 5481256805dSOphir Munk if (sc & (MLX5_BITSHIFT(ETHTOOL_LINK_MODE_50000baseCR2_Full_BIT) | 5491256805dSOphir Munk MLX5_BITSHIFT(ETHTOOL_LINK_MODE_50000baseKR2_Full_BIT))) 5501256805dSOphir Munk priv->link_speed_capa |= ETH_LINK_SPEED_50G; 5511256805dSOphir Munk if (sc & (MLX5_BITSHIFT(ETHTOOL_LINK_MODE_100000baseKR4_Full_BIT) | 5521256805dSOphir Munk MLX5_BITSHIFT(ETHTOOL_LINK_MODE_100000baseSR4_Full_BIT) | 5531256805dSOphir Munk MLX5_BITSHIFT(ETHTOOL_LINK_MODE_100000baseCR4_Full_BIT) | 5541256805dSOphir Munk MLX5_BITSHIFT(ETHTOOL_LINK_MODE_100000baseLR4_ER4_Full_BIT))) 5551256805dSOphir Munk priv->link_speed_capa |= ETH_LINK_SPEED_100G; 5561256805dSOphir Munk if (sc & (MLX5_BITSHIFT(ETHTOOL_LINK_MODE_200000baseKR4_Full_BIT) | 5571256805dSOphir Munk MLX5_BITSHIFT(ETHTOOL_LINK_MODE_200000baseSR4_Full_BIT))) 5581256805dSOphir Munk priv->link_speed_capa |= ETH_LINK_SPEED_200G; 5591256805dSOphir Munk 5601256805dSOphir Munk sc = ecmd->link_mode_masks[2] | 5611256805dSOphir Munk ((uint64_t)ecmd->link_mode_masks[3] << 32); 5621256805dSOphir Munk if (sc & (MLX5_BITSHIFT(ETHTOOL_LINK_MODE_200000baseCR4_Full_BIT) | 5631256805dSOphir Munk MLX5_BITSHIFT 5641256805dSOphir Munk (ETHTOOL_LINK_MODE_200000baseLR4_ER4_FR4_Full_BIT) | 5651256805dSOphir Munk MLX5_BITSHIFT(ETHTOOL_LINK_MODE_200000baseDR4_Full_BIT))) 5661256805dSOphir Munk priv->link_speed_capa |= ETH_LINK_SPEED_200G; 5671256805dSOphir Munk dev_link.link_duplex = ((ecmd->duplex == DUPLEX_HALF) ? 5681256805dSOphir Munk ETH_LINK_HALF_DUPLEX : ETH_LINK_FULL_DUPLEX); 5691256805dSOphir Munk dev_link.link_autoneg = !(dev->data->dev_conf.link_speeds & 5701256805dSOphir Munk ETH_LINK_SPEED_FIXED); 5711256805dSOphir Munk if (((dev_link.link_speed && !dev_link.link_status) || 5721256805dSOphir Munk (!dev_link.link_speed && dev_link.link_status))) { 5731256805dSOphir Munk rte_errno = EAGAIN; 5741256805dSOphir Munk return -rte_errno; 5751256805dSOphir Munk } 5761256805dSOphir Munk *link = dev_link; 5771256805dSOphir Munk return 0; 5781256805dSOphir Munk } 5791256805dSOphir Munk 5801256805dSOphir Munk /** 5811256805dSOphir Munk * DPDK callback to retrieve physical link information. 5821256805dSOphir Munk * 5831256805dSOphir Munk * @param dev 5841256805dSOphir Munk * Pointer to Ethernet device structure. 5851256805dSOphir Munk * @param wait_to_complete 5861256805dSOphir Munk * Wait for request completion. 5871256805dSOphir Munk * 5881256805dSOphir Munk * @return 5891256805dSOphir Munk * 0 if link status was not updated, positive if it was, a negative errno 5901256805dSOphir Munk * value otherwise and rte_errno is set. 5911256805dSOphir Munk */ 5921256805dSOphir Munk int 5931256805dSOphir Munk mlx5_link_update(struct rte_eth_dev *dev, int wait_to_complete) 5941256805dSOphir Munk { 5951256805dSOphir Munk int ret; 5961256805dSOphir Munk struct rte_eth_link dev_link; 5971256805dSOphir Munk time_t start_time = time(NULL); 5981256805dSOphir Munk int retry = MLX5_GET_LINK_STATUS_RETRY_COUNT; 5991256805dSOphir Munk 6001256805dSOphir Munk do { 6011256805dSOphir Munk ret = mlx5_link_update_unlocked_gs(dev, &dev_link); 6021256805dSOphir Munk if (ret == -ENOTSUP) 6031256805dSOphir Munk ret = mlx5_link_update_unlocked_gset(dev, &dev_link); 6041256805dSOphir Munk if (ret == 0) 6051256805dSOphir Munk break; 6061256805dSOphir Munk /* Handle wait to complete situation. */ 6071256805dSOphir Munk if ((wait_to_complete || retry) && ret == -EAGAIN) { 6081256805dSOphir Munk if (abs((int)difftime(time(NULL), start_time)) < 6091256805dSOphir Munk MLX5_LINK_STATUS_TIMEOUT) { 6101256805dSOphir Munk usleep(0); 6111256805dSOphir Munk continue; 6121256805dSOphir Munk } else { 6131256805dSOphir Munk rte_errno = EBUSY; 6141256805dSOphir Munk return -rte_errno; 6151256805dSOphir Munk } 6161256805dSOphir Munk } else if (ret < 0) { 6171256805dSOphir Munk return ret; 6181256805dSOphir Munk } 6191256805dSOphir Munk } while (wait_to_complete || retry-- > 0); 6201256805dSOphir Munk ret = !!memcmp(&dev->data->dev_link, &dev_link, 6211256805dSOphir Munk sizeof(struct rte_eth_link)); 6221256805dSOphir Munk dev->data->dev_link = dev_link; 6231256805dSOphir Munk return ret; 6241256805dSOphir Munk } 6251256805dSOphir Munk 6261256805dSOphir Munk /** 6271256805dSOphir Munk * DPDK callback to get flow control status. 6281256805dSOphir Munk * 6291256805dSOphir Munk * @param dev 6301256805dSOphir Munk * Pointer to Ethernet device structure. 6311256805dSOphir Munk * @param[out] fc_conf 6321256805dSOphir Munk * Flow control output buffer. 6331256805dSOphir Munk * 6341256805dSOphir Munk * @return 6351256805dSOphir Munk * 0 on success, a negative errno value otherwise and rte_errno is set. 6361256805dSOphir Munk */ 6371256805dSOphir Munk int 6381256805dSOphir Munk mlx5_dev_get_flow_ctrl(struct rte_eth_dev *dev, struct rte_eth_fc_conf *fc_conf) 6391256805dSOphir Munk { 6401256805dSOphir Munk struct ifreq ifr; 6411256805dSOphir Munk struct ethtool_pauseparam ethpause = { 6421256805dSOphir Munk .cmd = ETHTOOL_GPAUSEPARAM 6431256805dSOphir Munk }; 6441256805dSOphir Munk int ret; 6451256805dSOphir Munk 6461256805dSOphir Munk ifr.ifr_data = (void *)ðpause; 6471256805dSOphir Munk ret = mlx5_ifreq(dev, SIOCETHTOOL, &ifr); 6481256805dSOphir Munk if (ret) { 6491256805dSOphir Munk DRV_LOG(WARNING, 6501256805dSOphir Munk "port %u ioctl(SIOCETHTOOL, ETHTOOL_GPAUSEPARAM) failed:" 6511256805dSOphir Munk " %s", 6521256805dSOphir Munk dev->data->port_id, strerror(rte_errno)); 6531256805dSOphir Munk return ret; 6541256805dSOphir Munk } 6551256805dSOphir Munk fc_conf->autoneg = ethpause.autoneg; 6561256805dSOphir Munk if (ethpause.rx_pause && ethpause.tx_pause) 6571256805dSOphir Munk fc_conf->mode = RTE_FC_FULL; 6581256805dSOphir Munk else if (ethpause.rx_pause) 6591256805dSOphir Munk fc_conf->mode = RTE_FC_RX_PAUSE; 6601256805dSOphir Munk else if (ethpause.tx_pause) 6611256805dSOphir Munk fc_conf->mode = RTE_FC_TX_PAUSE; 6621256805dSOphir Munk else 6631256805dSOphir Munk fc_conf->mode = RTE_FC_NONE; 6641256805dSOphir Munk return 0; 6651256805dSOphir Munk } 6661256805dSOphir Munk 6671256805dSOphir Munk /** 6681256805dSOphir Munk * DPDK callback to modify flow control parameters. 6691256805dSOphir Munk * 6701256805dSOphir Munk * @param dev 6711256805dSOphir Munk * Pointer to Ethernet device structure. 6721256805dSOphir Munk * @param[in] fc_conf 6731256805dSOphir Munk * Flow control parameters. 6741256805dSOphir Munk * 6751256805dSOphir Munk * @return 6761256805dSOphir Munk * 0 on success, a negative errno value otherwise and rte_errno is set. 6771256805dSOphir Munk */ 6781256805dSOphir Munk int 6791256805dSOphir Munk mlx5_dev_set_flow_ctrl(struct rte_eth_dev *dev, struct rte_eth_fc_conf *fc_conf) 6801256805dSOphir Munk { 6811256805dSOphir Munk struct ifreq ifr; 6821256805dSOphir Munk struct ethtool_pauseparam ethpause = { 6831256805dSOphir Munk .cmd = ETHTOOL_SPAUSEPARAM 6841256805dSOphir Munk }; 6851256805dSOphir Munk int ret; 6861256805dSOphir Munk 6871256805dSOphir Munk ifr.ifr_data = (void *)ðpause; 6881256805dSOphir Munk ethpause.autoneg = fc_conf->autoneg; 6891256805dSOphir Munk if (((fc_conf->mode & RTE_FC_FULL) == RTE_FC_FULL) || 6901256805dSOphir Munk (fc_conf->mode & RTE_FC_RX_PAUSE)) 6911256805dSOphir Munk ethpause.rx_pause = 1; 6921256805dSOphir Munk else 6931256805dSOphir Munk ethpause.rx_pause = 0; 6941256805dSOphir Munk 6951256805dSOphir Munk if (((fc_conf->mode & RTE_FC_FULL) == RTE_FC_FULL) || 6961256805dSOphir Munk (fc_conf->mode & RTE_FC_TX_PAUSE)) 6971256805dSOphir Munk ethpause.tx_pause = 1; 6981256805dSOphir Munk else 6991256805dSOphir Munk ethpause.tx_pause = 0; 7001256805dSOphir Munk ret = mlx5_ifreq(dev, SIOCETHTOOL, &ifr); 7011256805dSOphir Munk if (ret) { 7021256805dSOphir Munk DRV_LOG(WARNING, 7031256805dSOphir Munk "port %u ioctl(SIOCETHTOOL, ETHTOOL_SPAUSEPARAM)" 7041256805dSOphir Munk " failed: %s", 7051256805dSOphir Munk dev->data->port_id, strerror(rte_errno)); 7061256805dSOphir Munk return ret; 7071256805dSOphir Munk } 7081256805dSOphir Munk return 0; 7091256805dSOphir Munk } 7101256805dSOphir Munk 7111256805dSOphir Munk /** 7121256805dSOphir Munk * Handle asynchronous removal event for entire multiport device. 7131256805dSOphir Munk * 7141256805dSOphir Munk * @param sh 7151256805dSOphir Munk * Infiniband device shared context. 7161256805dSOphir Munk */ 7171256805dSOphir Munk static void 7181256805dSOphir Munk mlx5_dev_interrupt_device_fatal(struct mlx5_dev_ctx_shared *sh) 7191256805dSOphir Munk { 7201256805dSOphir Munk uint32_t i; 7211256805dSOphir Munk 7221256805dSOphir Munk for (i = 0; i < sh->max_port; ++i) { 7231256805dSOphir Munk struct rte_eth_dev *dev; 7241256805dSOphir Munk 7251256805dSOphir Munk if (sh->port[i].ih_port_id >= RTE_MAX_ETHPORTS) { 7261256805dSOphir Munk /* 7271256805dSOphir Munk * Or not existing port either no 7281256805dSOphir Munk * handler installed for this port. 7291256805dSOphir Munk */ 7301256805dSOphir Munk continue; 7311256805dSOphir Munk } 7321256805dSOphir Munk dev = &rte_eth_devices[sh->port[i].ih_port_id]; 7331256805dSOphir Munk MLX5_ASSERT(dev); 7341256805dSOphir Munk if (dev->data->dev_conf.intr_conf.rmv) 7351256805dSOphir Munk _rte_eth_dev_callback_process 7361256805dSOphir Munk (dev, RTE_ETH_EVENT_INTR_RMV, NULL); 7371256805dSOphir Munk } 7381256805dSOphir Munk } 7391256805dSOphir Munk 7401256805dSOphir Munk /** 7411256805dSOphir Munk * Handle shared asynchronous events the NIC (removal event 7421256805dSOphir Munk * and link status change). Supports multiport IB device. 7431256805dSOphir Munk * 7441256805dSOphir Munk * @param cb_arg 7451256805dSOphir Munk * Callback argument. 7461256805dSOphir Munk */ 7471256805dSOphir Munk void 7481256805dSOphir Munk mlx5_dev_interrupt_handler(void *cb_arg) 7491256805dSOphir Munk { 7501256805dSOphir Munk struct mlx5_dev_ctx_shared *sh = cb_arg; 7511256805dSOphir Munk struct ibv_async_event event; 7521256805dSOphir Munk 7531256805dSOphir Munk /* Read all message from the IB device and acknowledge them. */ 7541256805dSOphir Munk for (;;) { 7551256805dSOphir Munk struct rte_eth_dev *dev; 7561256805dSOphir Munk uint32_t tmp; 7571256805dSOphir Munk 7581256805dSOphir Munk if (mlx5_glue->get_async_event(sh->ctx, &event)) 7591256805dSOphir Munk break; 7601256805dSOphir Munk /* Retrieve and check IB port index. */ 7611256805dSOphir Munk tmp = (uint32_t)event.element.port_num; 7621256805dSOphir Munk if (!tmp && event.event_type == IBV_EVENT_DEVICE_FATAL) { 7631256805dSOphir Munk /* 7641256805dSOphir Munk * The DEVICE_FATAL event is called once for 7651256805dSOphir Munk * entire device without port specifying. 7661256805dSOphir Munk * We should notify all existing ports. 7671256805dSOphir Munk */ 7681256805dSOphir Munk mlx5_glue->ack_async_event(&event); 7691256805dSOphir Munk mlx5_dev_interrupt_device_fatal(sh); 7701256805dSOphir Munk continue; 7711256805dSOphir Munk } 7721256805dSOphir Munk MLX5_ASSERT(tmp && (tmp <= sh->max_port)); 7731256805dSOphir Munk if (!tmp) { 7741256805dSOphir Munk /* Unsupported device level event. */ 7751256805dSOphir Munk mlx5_glue->ack_async_event(&event); 7761256805dSOphir Munk DRV_LOG(DEBUG, 7771256805dSOphir Munk "unsupported common event (type %d)", 7781256805dSOphir Munk event.event_type); 7791256805dSOphir Munk continue; 7801256805dSOphir Munk } 7811256805dSOphir Munk if (tmp > sh->max_port) { 7821256805dSOphir Munk /* Invalid IB port index. */ 7831256805dSOphir Munk mlx5_glue->ack_async_event(&event); 7841256805dSOphir Munk DRV_LOG(DEBUG, 7851256805dSOphir Munk "cannot handle an event (type %d)" 7861256805dSOphir Munk "due to invalid IB port index (%u)", 7871256805dSOphir Munk event.event_type, tmp); 7881256805dSOphir Munk continue; 7891256805dSOphir Munk } 7901256805dSOphir Munk if (sh->port[tmp - 1].ih_port_id >= RTE_MAX_ETHPORTS) { 7911256805dSOphir Munk /* No handler installed. */ 7921256805dSOphir Munk mlx5_glue->ack_async_event(&event); 7931256805dSOphir Munk DRV_LOG(DEBUG, 7941256805dSOphir Munk "cannot handle an event (type %d)" 7951256805dSOphir Munk "due to no handler installed for port %u", 7961256805dSOphir Munk event.event_type, tmp); 7971256805dSOphir Munk continue; 7981256805dSOphir Munk } 7991256805dSOphir Munk /* Retrieve ethernet device descriptor. */ 8001256805dSOphir Munk tmp = sh->port[tmp - 1].ih_port_id; 8011256805dSOphir Munk dev = &rte_eth_devices[tmp]; 8021256805dSOphir Munk MLX5_ASSERT(dev); 8031256805dSOphir Munk if ((event.event_type == IBV_EVENT_PORT_ACTIVE || 8041256805dSOphir Munk event.event_type == IBV_EVENT_PORT_ERR) && 8051256805dSOphir Munk dev->data->dev_conf.intr_conf.lsc) { 8061256805dSOphir Munk mlx5_glue->ack_async_event(&event); 8071256805dSOphir Munk if (mlx5_link_update(dev, 0) == -EAGAIN) { 8081256805dSOphir Munk usleep(0); 8091256805dSOphir Munk continue; 8101256805dSOphir Munk } 8111256805dSOphir Munk _rte_eth_dev_callback_process 8121256805dSOphir Munk (dev, RTE_ETH_EVENT_INTR_LSC, NULL); 8131256805dSOphir Munk continue; 8141256805dSOphir Munk } 8151256805dSOphir Munk DRV_LOG(DEBUG, 8161256805dSOphir Munk "port %u cannot handle an unknown event (type %d)", 8171256805dSOphir Munk dev->data->port_id, event.event_type); 8181256805dSOphir Munk mlx5_glue->ack_async_event(&event); 8191256805dSOphir Munk } 8201256805dSOphir Munk } 8211256805dSOphir Munk 8221256805dSOphir Munk /* 8231256805dSOphir Munk * Unregister callback handler safely. The handler may be active 8241256805dSOphir Munk * while we are trying to unregister it, in this case code -EAGAIN 8251256805dSOphir Munk * is returned by rte_intr_callback_unregister(). This routine checks 8261256805dSOphir Munk * the return code and tries to unregister handler again. 8271256805dSOphir Munk * 8281256805dSOphir Munk * @param handle 8291256805dSOphir Munk * interrupt handle 8301256805dSOphir Munk * @param cb_fn 8311256805dSOphir Munk * pointer to callback routine 8321256805dSOphir Munk * @cb_arg 8331256805dSOphir Munk * opaque callback parameter 8341256805dSOphir Munk */ 8351256805dSOphir Munk void 8361256805dSOphir Munk mlx5_intr_callback_unregister(const struct rte_intr_handle *handle, 8371256805dSOphir Munk rte_intr_callback_fn cb_fn, void *cb_arg) 8381256805dSOphir Munk { 8391256805dSOphir Munk /* 8401256805dSOphir Munk * Try to reduce timeout management overhead by not calling 8411256805dSOphir Munk * the timer related routines on the first iteration. If the 8421256805dSOphir Munk * unregistering succeeds on first call there will be no 8431256805dSOphir Munk * timer calls at all. 8441256805dSOphir Munk */ 8451256805dSOphir Munk uint64_t twait = 0; 8461256805dSOphir Munk uint64_t start = 0; 8471256805dSOphir Munk 8481256805dSOphir Munk do { 8491256805dSOphir Munk int ret; 8501256805dSOphir Munk 8511256805dSOphir Munk ret = rte_intr_callback_unregister(handle, cb_fn, cb_arg); 8521256805dSOphir Munk if (ret >= 0) 8531256805dSOphir Munk return; 8541256805dSOphir Munk if (ret != -EAGAIN) { 8551256805dSOphir Munk DRV_LOG(INFO, "failed to unregister interrupt" 8561256805dSOphir Munk " handler (error: %d)", ret); 8571256805dSOphir Munk MLX5_ASSERT(false); 8581256805dSOphir Munk return; 8591256805dSOphir Munk } 8601256805dSOphir Munk if (twait) { 8611256805dSOphir Munk struct timespec onems; 8621256805dSOphir Munk 8631256805dSOphir Munk /* Wait one millisecond and try again. */ 8641256805dSOphir Munk onems.tv_sec = 0; 8651256805dSOphir Munk onems.tv_nsec = NS_PER_S / MS_PER_S; 8661256805dSOphir Munk nanosleep(&onems, 0); 8671256805dSOphir Munk /* Check whether one second elapsed. */ 8681256805dSOphir Munk if ((rte_get_timer_cycles() - start) <= twait) 8691256805dSOphir Munk continue; 8701256805dSOphir Munk } else { 8711256805dSOphir Munk /* 8721256805dSOphir Munk * We get the amount of timer ticks for one second. 8731256805dSOphir Munk * If this amount elapsed it means we spent one 8741256805dSOphir Munk * second in waiting. This branch is executed once 8751256805dSOphir Munk * on first iteration. 8761256805dSOphir Munk */ 8771256805dSOphir Munk twait = rte_get_timer_hz(); 8781256805dSOphir Munk MLX5_ASSERT(twait); 8791256805dSOphir Munk } 8801256805dSOphir Munk /* 8811256805dSOphir Munk * Timeout elapsed, show message (once a second) and retry. 8821256805dSOphir Munk * We have no other acceptable option here, if we ignore 8831256805dSOphir Munk * the unregistering return code the handler will not 8841256805dSOphir Munk * be unregistered, fd will be closed and we may get the 8851256805dSOphir Munk * crush. Hanging and messaging in the loop seems not to be 8861256805dSOphir Munk * the worst choice. 8871256805dSOphir Munk */ 8881256805dSOphir Munk DRV_LOG(INFO, "Retrying to unregister interrupt handler"); 8891256805dSOphir Munk start = rte_get_timer_cycles(); 8901256805dSOphir Munk } while (true); 8911256805dSOphir Munk } 8921256805dSOphir Munk 8931256805dSOphir Munk /** 8941256805dSOphir Munk * Handle DEVX interrupts from the NIC. 8951256805dSOphir Munk * This function is probably called from the DPDK host thread. 8961256805dSOphir Munk * 8971256805dSOphir Munk * @param cb_arg 8981256805dSOphir Munk * Callback argument. 8991256805dSOphir Munk */ 9001256805dSOphir Munk void 9011256805dSOphir Munk mlx5_dev_interrupt_handler_devx(void *cb_arg) 9021256805dSOphir Munk { 9031256805dSOphir Munk #ifndef HAVE_IBV_DEVX_ASYNC 9041256805dSOphir Munk (void)cb_arg; 9051256805dSOphir Munk return; 9061256805dSOphir Munk #else 9071256805dSOphir Munk struct mlx5_dev_ctx_shared *sh = cb_arg; 9081256805dSOphir Munk union { 9091256805dSOphir Munk struct mlx5dv_devx_async_cmd_hdr cmd_resp; 9101256805dSOphir Munk uint8_t buf[MLX5_ST_SZ_BYTES(query_flow_counter_out) + 9111256805dSOphir Munk MLX5_ST_SZ_BYTES(traffic_counter) + 9121256805dSOphir Munk sizeof(struct mlx5dv_devx_async_cmd_hdr)]; 9131256805dSOphir Munk } out; 9141256805dSOphir Munk uint8_t *buf = out.buf + sizeof(out.cmd_resp); 9151256805dSOphir Munk 9161256805dSOphir Munk while (!mlx5_glue->devx_get_async_cmd_comp(sh->devx_comp, 9171256805dSOphir Munk &out.cmd_resp, 9181256805dSOphir Munk sizeof(out.buf))) 9191256805dSOphir Munk mlx5_flow_async_pool_query_handle 9201256805dSOphir Munk (sh, (uint64_t)out.cmd_resp.wr_id, 9211256805dSOphir Munk mlx5_devx_get_out_command_status(buf)); 9221256805dSOphir Munk #endif /* HAVE_IBV_DEVX_ASYNC */ 9231256805dSOphir Munk } 9241256805dSOphir Munk 9251256805dSOphir Munk /** 9261256805dSOphir Munk * DPDK callback to bring the link DOWN. 9271256805dSOphir Munk * 9281256805dSOphir Munk * @param dev 9291256805dSOphir Munk * Pointer to Ethernet device structure. 9301256805dSOphir Munk * 9311256805dSOphir Munk * @return 9321256805dSOphir Munk * 0 on success, a negative errno value otherwise and rte_errno is set. 9331256805dSOphir Munk */ 9341256805dSOphir Munk int 9351256805dSOphir Munk mlx5_set_link_down(struct rte_eth_dev *dev) 9361256805dSOphir Munk { 9371256805dSOphir Munk return mlx5_set_flags(dev, ~IFF_UP, ~IFF_UP); 9381256805dSOphir Munk } 9391256805dSOphir Munk 9401256805dSOphir Munk /** 9411256805dSOphir Munk * DPDK callback to bring the link UP. 9421256805dSOphir Munk * 9431256805dSOphir Munk * @param dev 9441256805dSOphir Munk * Pointer to Ethernet device structure. 9451256805dSOphir Munk * 9461256805dSOphir Munk * @return 9471256805dSOphir Munk * 0 on success, a negative errno value otherwise and rte_errno is set. 9481256805dSOphir Munk */ 9491256805dSOphir Munk int 9501256805dSOphir Munk mlx5_set_link_up(struct rte_eth_dev *dev) 9511256805dSOphir Munk { 9521256805dSOphir Munk return mlx5_set_flags(dev, ~IFF_UP, IFF_UP); 9531256805dSOphir Munk } 9541256805dSOphir Munk 9551256805dSOphir Munk /** 9561256805dSOphir Munk * Check if mlx5 device was removed. 9571256805dSOphir Munk * 9581256805dSOphir Munk * @param dev 9591256805dSOphir Munk * Pointer to Ethernet device structure. 9601256805dSOphir Munk * 9611256805dSOphir Munk * @return 9621256805dSOphir Munk * 1 when device is removed, otherwise 0. 9631256805dSOphir Munk */ 9641256805dSOphir Munk int 9651256805dSOphir Munk mlx5_is_removed(struct rte_eth_dev *dev) 9661256805dSOphir Munk { 9671256805dSOphir Munk struct ibv_device_attr device_attr; 9681256805dSOphir Munk struct mlx5_priv *priv = dev->data->dev_private; 9691256805dSOphir Munk 9701256805dSOphir Munk if (mlx5_glue->query_device(priv->sh->ctx, &device_attr) == EIO) 9711256805dSOphir Munk return 1; 9721256805dSOphir Munk return 0; 9731256805dSOphir Munk } 9741256805dSOphir Munk 9751256805dSOphir Munk /** 976ef9ee13fSOphir Munk * Analyze gathered port parameters via sysfs to recognize master 977ef9ee13fSOphir Munk * and representor devices for E-Switch configuration. 978ef9ee13fSOphir Munk * 979ef9ee13fSOphir Munk * @param[in] device_dir 980ef9ee13fSOphir Munk * flag of presence of "device" directory under port device key. 981ef9ee13fSOphir Munk * @param[inout] switch_info 982ef9ee13fSOphir Munk * Port information, including port name as a number and port name 983ef9ee13fSOphir Munk * type if recognized 984ef9ee13fSOphir Munk * 985ef9ee13fSOphir Munk * @return 986ef9ee13fSOphir Munk * master and representor flags are set in switch_info according to 987ef9ee13fSOphir Munk * recognized parameters (if any). 988ef9ee13fSOphir Munk */ 989ef9ee13fSOphir Munk static void 990ef9ee13fSOphir Munk mlx5_sysfs_check_switch_info(bool device_dir, 991ef9ee13fSOphir Munk struct mlx5_switch_info *switch_info) 992ef9ee13fSOphir Munk { 993ef9ee13fSOphir Munk switch (switch_info->name_type) { 994ef9ee13fSOphir Munk case MLX5_PHYS_PORT_NAME_TYPE_UNKNOWN: 995ef9ee13fSOphir Munk /* 996ef9ee13fSOphir Munk * Name is not recognized, assume the master, 997ef9ee13fSOphir Munk * check the device directory presence. 998ef9ee13fSOphir Munk */ 999ef9ee13fSOphir Munk switch_info->master = device_dir; 1000ef9ee13fSOphir Munk break; 1001ef9ee13fSOphir Munk case MLX5_PHYS_PORT_NAME_TYPE_NOTSET: 1002ef9ee13fSOphir Munk /* 1003ef9ee13fSOphir Munk * Name is not set, this assumes the legacy naming 1004ef9ee13fSOphir Munk * schema for master, just check if there is 1005ef9ee13fSOphir Munk * a device directory. 1006ef9ee13fSOphir Munk */ 1007ef9ee13fSOphir Munk switch_info->master = device_dir; 1008ef9ee13fSOphir Munk break; 1009ef9ee13fSOphir Munk case MLX5_PHYS_PORT_NAME_TYPE_UPLINK: 1010ef9ee13fSOphir Munk /* New uplink naming schema recognized. */ 1011ef9ee13fSOphir Munk switch_info->master = 1; 1012ef9ee13fSOphir Munk break; 1013ef9ee13fSOphir Munk case MLX5_PHYS_PORT_NAME_TYPE_LEGACY: 1014ef9ee13fSOphir Munk /* Legacy representors naming schema. */ 1015ef9ee13fSOphir Munk switch_info->representor = !device_dir; 1016ef9ee13fSOphir Munk break; 1017ef9ee13fSOphir Munk case MLX5_PHYS_PORT_NAME_TYPE_PFHPF: 1018ef9ee13fSOphir Munk /* Fallthrough */ 1019ef9ee13fSOphir Munk case MLX5_PHYS_PORT_NAME_TYPE_PFVF: 1020ef9ee13fSOphir Munk /* New representors naming schema. */ 1021ef9ee13fSOphir Munk switch_info->representor = 1; 1022ef9ee13fSOphir Munk break; 1023ef9ee13fSOphir Munk } 1024ef9ee13fSOphir Munk } 1025ef9ee13fSOphir Munk 1026ef9ee13fSOphir Munk /** 10271256805dSOphir Munk * Get switch information associated with network interface. 10281256805dSOphir Munk * 10291256805dSOphir Munk * @param ifindex 10301256805dSOphir Munk * Network interface index. 10311256805dSOphir Munk * @param[out] info 10321256805dSOphir Munk * Switch information object, populated in case of success. 10331256805dSOphir Munk * 10341256805dSOphir Munk * @return 10351256805dSOphir Munk * 0 on success, a negative errno value otherwise and rte_errno is set. 10361256805dSOphir Munk */ 10371256805dSOphir Munk int 10381256805dSOphir Munk mlx5_sysfs_switch_info(unsigned int ifindex, struct mlx5_switch_info *info) 10391256805dSOphir Munk { 10401256805dSOphir Munk char ifname[IF_NAMESIZE]; 10411256805dSOphir Munk char port_name[IF_NAMESIZE]; 10421256805dSOphir Munk FILE *file; 10431256805dSOphir Munk struct mlx5_switch_info data = { 10441256805dSOphir Munk .master = 0, 10451256805dSOphir Munk .representor = 0, 10461256805dSOphir Munk .name_type = MLX5_PHYS_PORT_NAME_TYPE_NOTSET, 10471256805dSOphir Munk .port_name = 0, 10481256805dSOphir Munk .switch_id = 0, 10491256805dSOphir Munk }; 10501256805dSOphir Munk DIR *dir; 10511256805dSOphir Munk bool port_switch_id_set = false; 10521256805dSOphir Munk bool device_dir = false; 10531256805dSOphir Munk char c; 10541256805dSOphir Munk int ret; 10551256805dSOphir Munk 10561256805dSOphir Munk if (!if_indextoname(ifindex, ifname)) { 10571256805dSOphir Munk rte_errno = errno; 10581256805dSOphir Munk return -rte_errno; 10591256805dSOphir Munk } 10601256805dSOphir Munk 10611256805dSOphir Munk MKSTR(phys_port_name, "/sys/class/net/%s/phys_port_name", 10621256805dSOphir Munk ifname); 10631256805dSOphir Munk MKSTR(phys_switch_id, "/sys/class/net/%s/phys_switch_id", 10641256805dSOphir Munk ifname); 10651256805dSOphir Munk MKSTR(pci_device, "/sys/class/net/%s/device", 10661256805dSOphir Munk ifname); 10671256805dSOphir Munk 10681256805dSOphir Munk file = fopen(phys_port_name, "rb"); 10691256805dSOphir Munk if (file != NULL) { 1070*63d1db71SMichael Baum ret = fscanf(file, "%" RTE_STR(IF_NAMESIZE) "s", port_name); 10711256805dSOphir Munk fclose(file); 10721256805dSOphir Munk if (ret == 1) 10731256805dSOphir Munk mlx5_translate_port_name(port_name, &data); 10741256805dSOphir Munk } 10751256805dSOphir Munk file = fopen(phys_switch_id, "rb"); 10761256805dSOphir Munk if (file == NULL) { 10771256805dSOphir Munk rte_errno = errno; 10781256805dSOphir Munk return -rte_errno; 10791256805dSOphir Munk } 10801256805dSOphir Munk port_switch_id_set = 10811256805dSOphir Munk fscanf(file, "%" SCNx64 "%c", &data.switch_id, &c) == 2 && 10821256805dSOphir Munk c == '\n'; 10831256805dSOphir Munk fclose(file); 10841256805dSOphir Munk dir = opendir(pci_device); 10851256805dSOphir Munk if (dir != NULL) { 10861256805dSOphir Munk closedir(dir); 10871256805dSOphir Munk device_dir = true; 10881256805dSOphir Munk } 10891256805dSOphir Munk if (port_switch_id_set) { 10901256805dSOphir Munk /* We have some E-Switch configuration. */ 10911256805dSOphir Munk mlx5_sysfs_check_switch_info(device_dir, &data); 10921256805dSOphir Munk } 10931256805dSOphir Munk *info = data; 10941256805dSOphir Munk MLX5_ASSERT(!(data.master && data.representor)); 10951256805dSOphir Munk if (data.master && data.representor) { 10961256805dSOphir Munk DRV_LOG(ERR, "ifindex %u device is recognized as master" 10971256805dSOphir Munk " and as representor", ifindex); 10981256805dSOphir Munk rte_errno = ENODEV; 10991256805dSOphir Munk return -rte_errno; 11001256805dSOphir Munk } 11011256805dSOphir Munk return 0; 11021256805dSOphir Munk } 11031256805dSOphir Munk 11041256805dSOphir Munk /** 11051256805dSOphir Munk * DPDK callback to retrieve plug-in module EEPROM information (type and size). 11061256805dSOphir Munk * 11071256805dSOphir Munk * @param dev 11081256805dSOphir Munk * Pointer to Ethernet device structure. 11091256805dSOphir Munk * @param[out] modinfo 11101256805dSOphir Munk * Storage for plug-in module EEPROM information. 11111256805dSOphir Munk * 11121256805dSOphir Munk * @return 11131256805dSOphir Munk * 0 on success, a negative errno value otherwise and rte_errno is set. 11141256805dSOphir Munk */ 11151256805dSOphir Munk int 11161256805dSOphir Munk mlx5_get_module_info(struct rte_eth_dev *dev, 11171256805dSOphir Munk struct rte_eth_dev_module_info *modinfo) 11181256805dSOphir Munk { 11191256805dSOphir Munk struct ethtool_modinfo info = { 11201256805dSOphir Munk .cmd = ETHTOOL_GMODULEINFO, 11211256805dSOphir Munk }; 11221256805dSOphir Munk struct ifreq ifr = (struct ifreq) { 11231256805dSOphir Munk .ifr_data = (void *)&info, 11241256805dSOphir Munk }; 11251256805dSOphir Munk int ret = 0; 11261256805dSOphir Munk 11271256805dSOphir Munk if (!dev || !modinfo) { 11281256805dSOphir Munk DRV_LOG(WARNING, "missing argument, cannot get module info"); 11291256805dSOphir Munk rte_errno = EINVAL; 11301256805dSOphir Munk return -rte_errno; 11311256805dSOphir Munk } 11321256805dSOphir Munk ret = mlx5_ifreq(dev, SIOCETHTOOL, &ifr); 11331256805dSOphir Munk if (ret) { 11341256805dSOphir Munk DRV_LOG(WARNING, "port %u ioctl(SIOCETHTOOL) failed: %s", 11351256805dSOphir Munk dev->data->port_id, strerror(rte_errno)); 11361256805dSOphir Munk return ret; 11371256805dSOphir Munk } 11381256805dSOphir Munk modinfo->type = info.type; 11391256805dSOphir Munk modinfo->eeprom_len = info.eeprom_len; 11401256805dSOphir Munk return ret; 11411256805dSOphir Munk } 11421256805dSOphir Munk 11431256805dSOphir Munk /** 11441256805dSOphir Munk * DPDK callback to retrieve plug-in module EEPROM data. 11451256805dSOphir Munk * 11461256805dSOphir Munk * @param dev 11471256805dSOphir Munk * Pointer to Ethernet device structure. 11481256805dSOphir Munk * @param[out] info 11491256805dSOphir Munk * Storage for plug-in module EEPROM data. 11501256805dSOphir Munk * 11511256805dSOphir Munk * @return 11521256805dSOphir Munk * 0 on success, a negative errno value otherwise and rte_errno is set. 11531256805dSOphir Munk */ 11541256805dSOphir Munk int mlx5_get_module_eeprom(struct rte_eth_dev *dev, 11551256805dSOphir Munk struct rte_dev_eeprom_info *info) 11561256805dSOphir Munk { 11571256805dSOphir Munk struct ethtool_eeprom *eeprom; 11581256805dSOphir Munk struct ifreq ifr; 11591256805dSOphir Munk int ret = 0; 11601256805dSOphir Munk 11611256805dSOphir Munk if (!dev || !info) { 11621256805dSOphir Munk DRV_LOG(WARNING, "missing argument, cannot get module eeprom"); 11631256805dSOphir Munk rte_errno = EINVAL; 11641256805dSOphir Munk return -rte_errno; 11651256805dSOphir Munk } 11662175c4dcSSuanming Mou eeprom = mlx5_malloc(MLX5_MEM_ZERO, 11672175c4dcSSuanming Mou (sizeof(struct ethtool_eeprom) + info->length), 0, 11682175c4dcSSuanming Mou SOCKET_ID_ANY); 11691256805dSOphir Munk if (!eeprom) { 11701256805dSOphir Munk DRV_LOG(WARNING, "port %u cannot allocate memory for " 11711256805dSOphir Munk "eeprom data", dev->data->port_id); 11721256805dSOphir Munk rte_errno = ENOMEM; 11731256805dSOphir Munk return -rte_errno; 11741256805dSOphir Munk } 11751256805dSOphir Munk eeprom->cmd = ETHTOOL_GMODULEEEPROM; 11761256805dSOphir Munk eeprom->offset = info->offset; 11771256805dSOphir Munk eeprom->len = info->length; 11781256805dSOphir Munk ifr = (struct ifreq) { 11791256805dSOphir Munk .ifr_data = (void *)eeprom, 11801256805dSOphir Munk }; 11811256805dSOphir Munk ret = mlx5_ifreq(dev, SIOCETHTOOL, &ifr); 11821256805dSOphir Munk if (ret) 11831256805dSOphir Munk DRV_LOG(WARNING, "port %u ioctl(SIOCETHTOOL) failed: %s", 11841256805dSOphir Munk dev->data->port_id, strerror(rte_errno)); 11851256805dSOphir Munk else 11861256805dSOphir Munk rte_memcpy(info->data, eeprom->data, info->length); 11872175c4dcSSuanming Mou mlx5_free(eeprom); 11881256805dSOphir Munk return ret; 11891256805dSOphir Munk } 119098c4b12aSOphir Munk 119198c4b12aSOphir Munk /** 119298c4b12aSOphir Munk * Read device counters table. 119398c4b12aSOphir Munk * 119498c4b12aSOphir Munk * @param dev 119598c4b12aSOphir Munk * Pointer to Ethernet device. 119698c4b12aSOphir Munk * @param[out] stats 119798c4b12aSOphir Munk * Counters table output buffer. 119898c4b12aSOphir Munk * 119998c4b12aSOphir Munk * @return 120098c4b12aSOphir Munk * 0 on success and stats is filled, negative errno value otherwise and 120198c4b12aSOphir Munk * rte_errno is set. 120298c4b12aSOphir Munk */ 120398c4b12aSOphir Munk int 120498c4b12aSOphir Munk mlx5_os_read_dev_counters(struct rte_eth_dev *dev, uint64_t *stats) 120598c4b12aSOphir Munk { 120698c4b12aSOphir Munk struct mlx5_priv *priv = dev->data->dev_private; 120798c4b12aSOphir Munk struct mlx5_xstats_ctrl *xstats_ctrl = &priv->xstats_ctrl; 120898c4b12aSOphir Munk unsigned int i; 120998c4b12aSOphir Munk struct ifreq ifr; 121098c4b12aSOphir Munk unsigned int stats_sz = xstats_ctrl->stats_n * sizeof(uint64_t); 121198c4b12aSOphir Munk unsigned char et_stat_buf[sizeof(struct ethtool_stats) + stats_sz]; 121298c4b12aSOphir Munk struct ethtool_stats *et_stats = (struct ethtool_stats *)et_stat_buf; 121398c4b12aSOphir Munk int ret; 121498c4b12aSOphir Munk 121598c4b12aSOphir Munk et_stats->cmd = ETHTOOL_GSTATS; 121698c4b12aSOphir Munk et_stats->n_stats = xstats_ctrl->stats_n; 121798c4b12aSOphir Munk ifr.ifr_data = (caddr_t)et_stats; 121898c4b12aSOphir Munk ret = mlx5_ifreq(dev, SIOCETHTOOL, &ifr); 121998c4b12aSOphir Munk if (ret) { 122098c4b12aSOphir Munk DRV_LOG(WARNING, 122198c4b12aSOphir Munk "port %u unable to read statistic values from device", 122298c4b12aSOphir Munk dev->data->port_id); 122398c4b12aSOphir Munk return ret; 122498c4b12aSOphir Munk } 122598c4b12aSOphir Munk for (i = 0; i != xstats_ctrl->mlx5_stats_n; ++i) { 122698c4b12aSOphir Munk if (xstats_ctrl->info[i].dev) { 122798c4b12aSOphir Munk ret = mlx5_os_read_dev_stat(priv, 122898c4b12aSOphir Munk xstats_ctrl->info[i].ctr_name, 122998c4b12aSOphir Munk &stats[i]); 123098c4b12aSOphir Munk /* return last xstats counter if fail to read. */ 123198c4b12aSOphir Munk if (ret == 0) 123298c4b12aSOphir Munk xstats_ctrl->xstats[i] = stats[i]; 123398c4b12aSOphir Munk else 123498c4b12aSOphir Munk stats[i] = xstats_ctrl->xstats[i]; 123598c4b12aSOphir Munk } else { 123698c4b12aSOphir Munk stats[i] = (uint64_t) 123798c4b12aSOphir Munk et_stats->data[xstats_ctrl->dev_table_idx[i]]; 123898c4b12aSOphir Munk } 123998c4b12aSOphir Munk } 124098c4b12aSOphir Munk return 0; 124198c4b12aSOphir Munk } 124298c4b12aSOphir Munk 124398c4b12aSOphir Munk /** 124498c4b12aSOphir Munk * Query the number of statistics provided by ETHTOOL. 124598c4b12aSOphir Munk * 124698c4b12aSOphir Munk * @param dev 124798c4b12aSOphir Munk * Pointer to Ethernet device. 124898c4b12aSOphir Munk * 124998c4b12aSOphir Munk * @return 125098c4b12aSOphir Munk * Number of statistics on success, negative errno value otherwise and 125198c4b12aSOphir Munk * rte_errno is set. 125298c4b12aSOphir Munk */ 125398c4b12aSOphir Munk int 125498c4b12aSOphir Munk mlx5_os_get_stats_n(struct rte_eth_dev *dev) 125598c4b12aSOphir Munk { 125698c4b12aSOphir Munk struct ethtool_drvinfo drvinfo; 125798c4b12aSOphir Munk struct ifreq ifr; 125898c4b12aSOphir Munk int ret; 125998c4b12aSOphir Munk 126098c4b12aSOphir Munk drvinfo.cmd = ETHTOOL_GDRVINFO; 126198c4b12aSOphir Munk ifr.ifr_data = (caddr_t)&drvinfo; 126298c4b12aSOphir Munk ret = mlx5_ifreq(dev, SIOCETHTOOL, &ifr); 126398c4b12aSOphir Munk if (ret) { 126498c4b12aSOphir Munk DRV_LOG(WARNING, "port %u unable to query number of statistics", 126598c4b12aSOphir Munk dev->data->port_id); 126698c4b12aSOphir Munk return ret; 126798c4b12aSOphir Munk } 126898c4b12aSOphir Munk return drvinfo.n_stats; 126998c4b12aSOphir Munk } 127098c4b12aSOphir Munk 127198c4b12aSOphir Munk static const struct mlx5_counter_ctrl mlx5_counters_init[] = { 127298c4b12aSOphir Munk { 127398c4b12aSOphir Munk .dpdk_name = "rx_port_unicast_bytes", 127498c4b12aSOphir Munk .ctr_name = "rx_vport_unicast_bytes", 127598c4b12aSOphir Munk }, 127698c4b12aSOphir Munk { 127798c4b12aSOphir Munk .dpdk_name = "rx_port_multicast_bytes", 127898c4b12aSOphir Munk .ctr_name = "rx_vport_multicast_bytes", 127998c4b12aSOphir Munk }, 128098c4b12aSOphir Munk { 128198c4b12aSOphir Munk .dpdk_name = "rx_port_broadcast_bytes", 128298c4b12aSOphir Munk .ctr_name = "rx_vport_broadcast_bytes", 128398c4b12aSOphir Munk }, 128498c4b12aSOphir Munk { 128598c4b12aSOphir Munk .dpdk_name = "rx_port_unicast_packets", 128698c4b12aSOphir Munk .ctr_name = "rx_vport_unicast_packets", 128798c4b12aSOphir Munk }, 128898c4b12aSOphir Munk { 128998c4b12aSOphir Munk .dpdk_name = "rx_port_multicast_packets", 129098c4b12aSOphir Munk .ctr_name = "rx_vport_multicast_packets", 129198c4b12aSOphir Munk }, 129298c4b12aSOphir Munk { 129398c4b12aSOphir Munk .dpdk_name = "rx_port_broadcast_packets", 129498c4b12aSOphir Munk .ctr_name = "rx_vport_broadcast_packets", 129598c4b12aSOphir Munk }, 129698c4b12aSOphir Munk { 129798c4b12aSOphir Munk .dpdk_name = "tx_port_unicast_bytes", 129898c4b12aSOphir Munk .ctr_name = "tx_vport_unicast_bytes", 129998c4b12aSOphir Munk }, 130098c4b12aSOphir Munk { 130198c4b12aSOphir Munk .dpdk_name = "tx_port_multicast_bytes", 130298c4b12aSOphir Munk .ctr_name = "tx_vport_multicast_bytes", 130398c4b12aSOphir Munk }, 130498c4b12aSOphir Munk { 130598c4b12aSOphir Munk .dpdk_name = "tx_port_broadcast_bytes", 130698c4b12aSOphir Munk .ctr_name = "tx_vport_broadcast_bytes", 130798c4b12aSOphir Munk }, 130898c4b12aSOphir Munk { 130998c4b12aSOphir Munk .dpdk_name = "tx_port_unicast_packets", 131098c4b12aSOphir Munk .ctr_name = "tx_vport_unicast_packets", 131198c4b12aSOphir Munk }, 131298c4b12aSOphir Munk { 131398c4b12aSOphir Munk .dpdk_name = "tx_port_multicast_packets", 131498c4b12aSOphir Munk .ctr_name = "tx_vport_multicast_packets", 131598c4b12aSOphir Munk }, 131698c4b12aSOphir Munk { 131798c4b12aSOphir Munk .dpdk_name = "tx_port_broadcast_packets", 131898c4b12aSOphir Munk .ctr_name = "tx_vport_broadcast_packets", 131998c4b12aSOphir Munk }, 132098c4b12aSOphir Munk { 132198c4b12aSOphir Munk .dpdk_name = "rx_wqe_err", 132298c4b12aSOphir Munk .ctr_name = "rx_wqe_err", 132398c4b12aSOphir Munk }, 132498c4b12aSOphir Munk { 132598c4b12aSOphir Munk .dpdk_name = "rx_crc_errors_phy", 132698c4b12aSOphir Munk .ctr_name = "rx_crc_errors_phy", 132798c4b12aSOphir Munk }, 132898c4b12aSOphir Munk { 132998c4b12aSOphir Munk .dpdk_name = "rx_in_range_len_errors_phy", 133098c4b12aSOphir Munk .ctr_name = "rx_in_range_len_errors_phy", 133198c4b12aSOphir Munk }, 133298c4b12aSOphir Munk { 133398c4b12aSOphir Munk .dpdk_name = "rx_symbol_err_phy", 133498c4b12aSOphir Munk .ctr_name = "rx_symbol_err_phy", 133598c4b12aSOphir Munk }, 133698c4b12aSOphir Munk { 133798c4b12aSOphir Munk .dpdk_name = "tx_errors_phy", 133898c4b12aSOphir Munk .ctr_name = "tx_errors_phy", 133998c4b12aSOphir Munk }, 134098c4b12aSOphir Munk { 134198c4b12aSOphir Munk .dpdk_name = "rx_out_of_buffer", 134298c4b12aSOphir Munk .ctr_name = "out_of_buffer", 134398c4b12aSOphir Munk .dev = 1, 134498c4b12aSOphir Munk }, 134598c4b12aSOphir Munk { 134698c4b12aSOphir Munk .dpdk_name = "tx_packets_phy", 134798c4b12aSOphir Munk .ctr_name = "tx_packets_phy", 134898c4b12aSOphir Munk }, 134998c4b12aSOphir Munk { 135098c4b12aSOphir Munk .dpdk_name = "rx_packets_phy", 135198c4b12aSOphir Munk .ctr_name = "rx_packets_phy", 135298c4b12aSOphir Munk }, 135398c4b12aSOphir Munk { 135498c4b12aSOphir Munk .dpdk_name = "tx_discards_phy", 135598c4b12aSOphir Munk .ctr_name = "tx_discards_phy", 135698c4b12aSOphir Munk }, 135798c4b12aSOphir Munk { 135898c4b12aSOphir Munk .dpdk_name = "rx_discards_phy", 135998c4b12aSOphir Munk .ctr_name = "rx_discards_phy", 136098c4b12aSOphir Munk }, 136198c4b12aSOphir Munk { 136298c4b12aSOphir Munk .dpdk_name = "tx_bytes_phy", 136398c4b12aSOphir Munk .ctr_name = "tx_bytes_phy", 136498c4b12aSOphir Munk }, 136598c4b12aSOphir Munk { 136698c4b12aSOphir Munk .dpdk_name = "rx_bytes_phy", 136798c4b12aSOphir Munk .ctr_name = "rx_bytes_phy", 136898c4b12aSOphir Munk }, 136998c4b12aSOphir Munk /* Representor only */ 137098c4b12aSOphir Munk { 137198c4b12aSOphir Munk .dpdk_name = "rx_packets", 137298c4b12aSOphir Munk .ctr_name = "vport_rx_packets", 137398c4b12aSOphir Munk }, 137498c4b12aSOphir Munk { 137598c4b12aSOphir Munk .dpdk_name = "rx_bytes", 137698c4b12aSOphir Munk .ctr_name = "vport_rx_bytes", 137798c4b12aSOphir Munk }, 137898c4b12aSOphir Munk { 137998c4b12aSOphir Munk .dpdk_name = "tx_packets", 138098c4b12aSOphir Munk .ctr_name = "vport_tx_packets", 138198c4b12aSOphir Munk }, 138298c4b12aSOphir Munk { 138398c4b12aSOphir Munk .dpdk_name = "tx_bytes", 138498c4b12aSOphir Munk .ctr_name = "vport_tx_bytes", 138598c4b12aSOphir Munk }, 138698c4b12aSOphir Munk }; 138798c4b12aSOphir Munk 138898c4b12aSOphir Munk static const unsigned int xstats_n = RTE_DIM(mlx5_counters_init); 138998c4b12aSOphir Munk 139098c4b12aSOphir Munk /** 139198c4b12aSOphir Munk * Init the structures to read device counters. 139298c4b12aSOphir Munk * 139398c4b12aSOphir Munk * @param dev 139498c4b12aSOphir Munk * Pointer to Ethernet device. 139598c4b12aSOphir Munk */ 139698c4b12aSOphir Munk void 139798c4b12aSOphir Munk mlx5_os_stats_init(struct rte_eth_dev *dev) 139898c4b12aSOphir Munk { 139998c4b12aSOphir Munk struct mlx5_priv *priv = dev->data->dev_private; 140098c4b12aSOphir Munk struct mlx5_xstats_ctrl *xstats_ctrl = &priv->xstats_ctrl; 140198c4b12aSOphir Munk struct mlx5_stats_ctrl *stats_ctrl = &priv->stats_ctrl; 140298c4b12aSOphir Munk unsigned int i; 140398c4b12aSOphir Munk unsigned int j; 140498c4b12aSOphir Munk struct ifreq ifr; 140598c4b12aSOphir Munk struct ethtool_gstrings *strings = NULL; 140698c4b12aSOphir Munk unsigned int dev_stats_n; 140798c4b12aSOphir Munk unsigned int str_sz; 140898c4b12aSOphir Munk int ret; 140998c4b12aSOphir Munk 141098c4b12aSOphir Munk /* So that it won't aggregate for each init. */ 141198c4b12aSOphir Munk xstats_ctrl->mlx5_stats_n = 0; 141298c4b12aSOphir Munk ret = mlx5_os_get_stats_n(dev); 141398c4b12aSOphir Munk if (ret < 0) { 141498c4b12aSOphir Munk DRV_LOG(WARNING, "port %u no extended statistics available", 141598c4b12aSOphir Munk dev->data->port_id); 141698c4b12aSOphir Munk return; 141798c4b12aSOphir Munk } 141898c4b12aSOphir Munk dev_stats_n = ret; 141998c4b12aSOphir Munk /* Allocate memory to grab stat names and values. */ 142098c4b12aSOphir Munk str_sz = dev_stats_n * ETH_GSTRING_LEN; 142198c4b12aSOphir Munk strings = (struct ethtool_gstrings *) 142298c4b12aSOphir Munk mlx5_malloc(0, str_sz + sizeof(struct ethtool_gstrings), 0, 142398c4b12aSOphir Munk SOCKET_ID_ANY); 142498c4b12aSOphir Munk if (!strings) { 142598c4b12aSOphir Munk DRV_LOG(WARNING, "port %u unable to allocate memory for xstats", 142698c4b12aSOphir Munk dev->data->port_id); 142798c4b12aSOphir Munk return; 142898c4b12aSOphir Munk } 142998c4b12aSOphir Munk strings->cmd = ETHTOOL_GSTRINGS; 143098c4b12aSOphir Munk strings->string_set = ETH_SS_STATS; 143198c4b12aSOphir Munk strings->len = dev_stats_n; 143298c4b12aSOphir Munk ifr.ifr_data = (caddr_t)strings; 143398c4b12aSOphir Munk ret = mlx5_ifreq(dev, SIOCETHTOOL, &ifr); 143498c4b12aSOphir Munk if (ret) { 143598c4b12aSOphir Munk DRV_LOG(WARNING, "port %u unable to get statistic names", 143698c4b12aSOphir Munk dev->data->port_id); 143798c4b12aSOphir Munk goto free; 143898c4b12aSOphir Munk } 143998c4b12aSOphir Munk for (i = 0; i != dev_stats_n; ++i) { 144098c4b12aSOphir Munk const char *curr_string = (const char *) 144198c4b12aSOphir Munk &strings->data[i * ETH_GSTRING_LEN]; 144298c4b12aSOphir Munk 144398c4b12aSOphir Munk for (j = 0; j != xstats_n; ++j) { 144498c4b12aSOphir Munk if (!strcmp(mlx5_counters_init[j].ctr_name, 144598c4b12aSOphir Munk curr_string)) { 144698c4b12aSOphir Munk unsigned int idx = xstats_ctrl->mlx5_stats_n++; 144798c4b12aSOphir Munk 144898c4b12aSOphir Munk xstats_ctrl->dev_table_idx[idx] = i; 144998c4b12aSOphir Munk xstats_ctrl->info[idx] = mlx5_counters_init[j]; 145098c4b12aSOphir Munk break; 145198c4b12aSOphir Munk } 145298c4b12aSOphir Munk } 145398c4b12aSOphir Munk } 145498c4b12aSOphir Munk /* Add dev counters. */ 145598c4b12aSOphir Munk for (i = 0; i != xstats_n; ++i) { 145698c4b12aSOphir Munk if (mlx5_counters_init[i].dev) { 145798c4b12aSOphir Munk unsigned int idx = xstats_ctrl->mlx5_stats_n++; 145898c4b12aSOphir Munk 145998c4b12aSOphir Munk xstats_ctrl->info[idx] = mlx5_counters_init[i]; 146098c4b12aSOphir Munk xstats_ctrl->hw_stats[idx] = 0; 146198c4b12aSOphir Munk } 146298c4b12aSOphir Munk } 146398c4b12aSOphir Munk MLX5_ASSERT(xstats_ctrl->mlx5_stats_n <= MLX5_MAX_XSTATS); 146498c4b12aSOphir Munk xstats_ctrl->stats_n = dev_stats_n; 146598c4b12aSOphir Munk /* Copy to base at first time. */ 146698c4b12aSOphir Munk ret = mlx5_os_read_dev_counters(dev, xstats_ctrl->base); 146798c4b12aSOphir Munk if (ret) 146898c4b12aSOphir Munk DRV_LOG(ERR, "port %u cannot read device counters: %s", 146998c4b12aSOphir Munk dev->data->port_id, strerror(rte_errno)); 147098c4b12aSOphir Munk mlx5_os_read_dev_stat(priv, "out_of_buffer", &stats_ctrl->imissed_base); 147198c4b12aSOphir Munk stats_ctrl->imissed = 0; 147298c4b12aSOphir Munk free: 147398c4b12aSOphir Munk mlx5_free(strings); 147498c4b12aSOphir Munk } 147598c4b12aSOphir Munk 147698c4b12aSOphir Munk /** 147798c4b12aSOphir Munk * Get MAC address by querying netdevice. 147898c4b12aSOphir Munk * 147998c4b12aSOphir Munk * @param[in] dev 148098c4b12aSOphir Munk * Pointer to Ethernet device. 148198c4b12aSOphir Munk * @param[out] mac 148298c4b12aSOphir Munk * MAC address output buffer. 148398c4b12aSOphir Munk * 148498c4b12aSOphir Munk * @return 148598c4b12aSOphir Munk * 0 on success, a negative errno value otherwise and rte_errno is set. 148698c4b12aSOphir Munk */ 148798c4b12aSOphir Munk int 148898c4b12aSOphir Munk mlx5_get_mac(struct rte_eth_dev *dev, uint8_t (*mac)[RTE_ETHER_ADDR_LEN]) 148998c4b12aSOphir Munk { 149098c4b12aSOphir Munk struct ifreq request; 149198c4b12aSOphir Munk int ret; 149298c4b12aSOphir Munk 149398c4b12aSOphir Munk ret = mlx5_ifreq(dev, SIOCGIFHWADDR, &request); 149498c4b12aSOphir Munk if (ret) 149598c4b12aSOphir Munk return ret; 149698c4b12aSOphir Munk memcpy(mac, request.ifr_hwaddr.sa_data, RTE_ETHER_ADDR_LEN); 149798c4b12aSOphir Munk return 0; 149898c4b12aSOphir Munk } 149998c4b12aSOphir Munk 1500