176668754SAndrew Boyer /* SPDX-License-Identifier: BSD-3-Clause 2a5205992SAndrew Boyer * Copyright 2018-2022 Advanced Micro Devices, Inc. 3a27d9013SAlfredo Cardigliano */ 4a27d9013SAlfredo Cardigliano 5a27d9013SAlfredo Cardigliano #include <stdio.h> 6a27d9013SAlfredo Cardigliano #include <string.h> 7a27d9013SAlfredo Cardigliano #include <errno.h> 8a27d9013SAlfredo Cardigliano #include <stdint.h> 9a27d9013SAlfredo Cardigliano 10a27d9013SAlfredo Cardigliano #include <rte_common.h> 11e86a6fccSAndrew Boyer #include <rte_byteorder.h> 12e86a6fccSAndrew Boyer #include <rte_errno.h> 13a27d9013SAlfredo Cardigliano #include <rte_log.h> 14a27d9013SAlfredo Cardigliano #include <rte_mbuf.h> 15a27d9013SAlfredo Cardigliano #include <rte_ether.h> 16a27d9013SAlfredo Cardigliano #include <rte_ip.h> 17e86a6fccSAndrew Boyer #include <rte_tcp.h> 18e86a6fccSAndrew Boyer #include <rte_ethdev.h> 19e86a6fccSAndrew Boyer #include <ethdev_driver.h> 20a27d9013SAlfredo Cardigliano 21e86a6fccSAndrew Boyer #include "ionic.h" 22e86a6fccSAndrew Boyer #include "ionic_dev.h" 23a27d9013SAlfredo Cardigliano #include "ionic_lif.h" 24e86a6fccSAndrew Boyer #include "ionic_ethdev.h" 25a27d9013SAlfredo Cardigliano #include "ionic_rxtx.h" 26e86a6fccSAndrew Boyer #include "ionic_logs.h" 27a27d9013SAlfredo Cardigliano 28e7222f94SAndrew Boyer static void 29e7222f94SAndrew Boyer ionic_empty_array(void **array, uint32_t cnt, uint16_t idx) 30e7222f94SAndrew Boyer { 31e7222f94SAndrew Boyer uint32_t i; 32e7222f94SAndrew Boyer 33e7222f94SAndrew Boyer for (i = idx; i < cnt; i++) 34e7222f94SAndrew Boyer if (array[i]) 35e7222f94SAndrew Boyer rte_pktmbuf_free_seg(array[i]); 36e7222f94SAndrew Boyer 37e7222f94SAndrew Boyer memset(array, 0, sizeof(void *) * cnt); 38e7222f94SAndrew Boyer } 39e7222f94SAndrew Boyer 40e7222f94SAndrew Boyer static void __rte_cold 41e7222f94SAndrew Boyer ionic_tx_empty(struct ionic_tx_qcq *txq) 42e7222f94SAndrew Boyer { 43e7222f94SAndrew Boyer struct ionic_queue *q = &txq->qcq.q; 44e7222f94SAndrew Boyer 45b4beb84aSAndrew Boyer ionic_empty_array(q->info, q->num_descs * q->num_segs, 0); 46e7222f94SAndrew Boyer } 47e7222f94SAndrew Boyer 48e7222f94SAndrew Boyer static void __rte_cold 49e7222f94SAndrew Boyer ionic_rx_empty(struct ionic_rx_qcq *rxq) 50e7222f94SAndrew Boyer { 51e7222f94SAndrew Boyer struct ionic_queue *q = &rxq->qcq.q; 52e7222f94SAndrew Boyer 537b20fc2fSAndrew Boyer /* 547b20fc2fSAndrew Boyer * Walk the full info array so that the clean up includes any 557b20fc2fSAndrew Boyer * fragments that were left dangling for later reuse 567b20fc2fSAndrew Boyer */ 577b20fc2fSAndrew Boyer ionic_empty_array(q->info, q->num_descs * q->num_segs, 0); 58218afd82SAndrew Boyer 59218afd82SAndrew Boyer ionic_empty_array((void **)rxq->mbs, 60218afd82SAndrew Boyer IONIC_MBUF_BULK_ALLOC, rxq->mb_idx); 61218afd82SAndrew Boyer rxq->mb_idx = 0; 62e7222f94SAndrew Boyer } 63e7222f94SAndrew Boyer 64a27d9013SAlfredo Cardigliano /********************************************************************* 65a27d9013SAlfredo Cardigliano * 66a27d9013SAlfredo Cardigliano * TX functions 67a27d9013SAlfredo Cardigliano * 68a27d9013SAlfredo Cardigliano **********************************************************************/ 69a27d9013SAlfredo Cardigliano 70a27d9013SAlfredo Cardigliano void 71a27d9013SAlfredo Cardigliano ionic_txq_info_get(struct rte_eth_dev *dev, uint16_t queue_id, 72a27d9013SAlfredo Cardigliano struct rte_eth_txq_info *qinfo) 73a27d9013SAlfredo Cardigliano { 74be39f75cSAndrew Boyer struct ionic_tx_qcq *txq = dev->data->tx_queues[queue_id]; 75be39f75cSAndrew Boyer struct ionic_queue *q = &txq->qcq.q; 76a27d9013SAlfredo Cardigliano 77a27d9013SAlfredo Cardigliano qinfo->nb_desc = q->num_descs; 7868591087SAndrew Boyer qinfo->conf.offloads = dev->data->dev_conf.txmode.offloads; 799ac234eeSAndrew Boyer if (txq->flags & IONIC_QCQ_F_FAST_FREE) 809ac234eeSAndrew Boyer qinfo->conf.offloads |= RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE; 8102eabf57SAndrew Boyer qinfo->conf.tx_deferred_start = txq->flags & IONIC_QCQ_F_DEFERRED; 82a27d9013SAlfredo Cardigliano } 83a27d9013SAlfredo Cardigliano 84ce6427ddSThomas Monjalon void __rte_cold 857483341aSXueming Li ionic_dev_tx_queue_release(struct rte_eth_dev *dev, uint16_t qid) 86a27d9013SAlfredo Cardigliano { 877483341aSXueming Li struct ionic_tx_qcq *txq = dev->data->tx_queues[qid]; 88a27d9013SAlfredo Cardigliano 89a27d9013SAlfredo Cardigliano IONIC_PRINT_CALL(); 90a27d9013SAlfredo Cardigliano 91be39f75cSAndrew Boyer ionic_qcq_free(&txq->qcq); 92a27d9013SAlfredo Cardigliano } 93a27d9013SAlfredo Cardigliano 94ce6427ddSThomas Monjalon int __rte_cold 95a27d9013SAlfredo Cardigliano ionic_dev_tx_queue_stop(struct rte_eth_dev *eth_dev, uint16_t tx_queue_id) 96a27d9013SAlfredo Cardigliano { 97e7222f94SAndrew Boyer struct ionic_tx_stats *stats; 98be39f75cSAndrew Boyer struct ionic_tx_qcq *txq; 99a27d9013SAlfredo Cardigliano 1004ae96cb8SAndrew Boyer IONIC_PRINT(DEBUG, "Stopping TX queue %u", tx_queue_id); 101a27d9013SAlfredo Cardigliano 102a27d9013SAlfredo Cardigliano txq = eth_dev->data->tx_queues[tx_queue_id]; 103a27d9013SAlfredo Cardigliano 1049fdf11c4SAndrew Boyer eth_dev->data->tx_queue_state[tx_queue_id] = 1059fdf11c4SAndrew Boyer RTE_ETH_QUEUE_STATE_STOPPED; 1069fdf11c4SAndrew Boyer 107a27d9013SAlfredo Cardigliano /* 108a27d9013SAlfredo Cardigliano * Note: we should better post NOP Tx desc and wait for its completion 109a27d9013SAlfredo Cardigliano * before disabling Tx queue 110a27d9013SAlfredo Cardigliano */ 111a27d9013SAlfredo Cardigliano 112e7222f94SAndrew Boyer ionic_lif_txq_deinit(txq); 113a27d9013SAlfredo Cardigliano 114e7222f94SAndrew Boyer /* Free all buffers from descriptor ring */ 115e7222f94SAndrew Boyer ionic_tx_empty(txq); 116e7222f94SAndrew Boyer 117e7222f94SAndrew Boyer stats = &txq->stats; 118e7222f94SAndrew Boyer IONIC_PRINT(DEBUG, "TX queue %u pkts %ju tso %ju", 119e7222f94SAndrew Boyer txq->qcq.q.index, stats->packets, stats->tso); 120ea81e9f2SAndrew Boyer IONIC_PRINT(DEBUG, "TX queue %u comps %ju (%ju per)", 121ea81e9f2SAndrew Boyer txq->qcq.q.index, stats->comps, 122ea81e9f2SAndrew Boyer stats->comps ? stats->packets / stats->comps : 0); 123a27d9013SAlfredo Cardigliano 124a27d9013SAlfredo Cardigliano return 0; 125a27d9013SAlfredo Cardigliano } 126a27d9013SAlfredo Cardigliano 127ce6427ddSThomas Monjalon int __rte_cold 128a27d9013SAlfredo Cardigliano ionic_dev_tx_queue_setup(struct rte_eth_dev *eth_dev, uint16_t tx_queue_id, 1294ae96cb8SAndrew Boyer uint16_t nb_desc, uint32_t socket_id, 130a27d9013SAlfredo Cardigliano const struct rte_eth_txconf *tx_conf) 131a27d9013SAlfredo Cardigliano { 132a27d9013SAlfredo Cardigliano struct ionic_lif *lif = IONIC_ETH_DEV_TO_LIF(eth_dev); 133be39f75cSAndrew Boyer struct ionic_tx_qcq *txq; 134a27d9013SAlfredo Cardigliano uint64_t offloads; 135a27d9013SAlfredo Cardigliano int err; 136a27d9013SAlfredo Cardigliano 137a27d9013SAlfredo Cardigliano if (tx_queue_id >= lif->ntxqcqs) { 138a27d9013SAlfredo Cardigliano IONIC_PRINT(DEBUG, "Queue index %u not available " 139a27d9013SAlfredo Cardigliano "(max %u queues)", 140a27d9013SAlfredo Cardigliano tx_queue_id, lif->ntxqcqs); 141a27d9013SAlfredo Cardigliano return -EINVAL; 142a27d9013SAlfredo Cardigliano } 143a27d9013SAlfredo Cardigliano 144a27d9013SAlfredo Cardigliano offloads = tx_conf->offloads | eth_dev->data->dev_conf.txmode.offloads; 1454ae96cb8SAndrew Boyer IONIC_PRINT(DEBUG, 1464ae96cb8SAndrew Boyer "Configuring skt %u TX queue %u with %u buffers, offloads %jx", 1474ae96cb8SAndrew Boyer socket_id, tx_queue_id, nb_desc, offloads); 148a27d9013SAlfredo Cardigliano 149a27d9013SAlfredo Cardigliano /* Validate number of receive descriptors */ 150a27d9013SAlfredo Cardigliano if (!rte_is_power_of_2(nb_desc) || nb_desc < IONIC_MIN_RING_DESC) 151a27d9013SAlfredo Cardigliano return -EINVAL; /* or use IONIC_DEFAULT_RING_DESC */ 152a27d9013SAlfredo Cardigliano 15307512941SAndrew Boyer if (tx_conf->tx_free_thresh > nb_desc) { 15407512941SAndrew Boyer IONIC_PRINT(ERR, 15507512941SAndrew Boyer "tx_free_thresh must be less than nb_desc (%u)", 15607512941SAndrew Boyer nb_desc); 15707512941SAndrew Boyer return -EINVAL; 15807512941SAndrew Boyer } 15907512941SAndrew Boyer 160a27d9013SAlfredo Cardigliano /* Free memory prior to re-allocation if needed... */ 161a27d9013SAlfredo Cardigliano if (eth_dev->data->tx_queues[tx_queue_id] != NULL) { 1627483341aSXueming Li ionic_dev_tx_queue_release(eth_dev, tx_queue_id); 163a27d9013SAlfredo Cardigliano eth_dev->data->tx_queues[tx_queue_id] = NULL; 164a27d9013SAlfredo Cardigliano } 165a27d9013SAlfredo Cardigliano 1669fdf11c4SAndrew Boyer eth_dev->data->tx_queue_state[tx_queue_id] = 1679fdf11c4SAndrew Boyer RTE_ETH_QUEUE_STATE_STOPPED; 1689fdf11c4SAndrew Boyer 1698ec5ad7fSAndrew Boyer err = ionic_tx_qcq_alloc(lif, socket_id, tx_queue_id, nb_desc, &txq); 170a27d9013SAlfredo Cardigliano if (err) { 171a27d9013SAlfredo Cardigliano IONIC_PRINT(DEBUG, "Queue allocation failure"); 172a27d9013SAlfredo Cardigliano return -EINVAL; 173a27d9013SAlfredo Cardigliano } 174a27d9013SAlfredo Cardigliano 175a27d9013SAlfredo Cardigliano /* Do not start queue with rte_eth_dev_start() */ 17602eabf57SAndrew Boyer if (tx_conf->tx_deferred_start) 17702eabf57SAndrew Boyer txq->flags |= IONIC_QCQ_F_DEFERRED; 178a27d9013SAlfredo Cardigliano 17968591087SAndrew Boyer /* Convert the offload flags into queue flags */ 180295968d1SFerruh Yigit if (offloads & RTE_ETH_TX_OFFLOAD_IPV4_CKSUM) 18168591087SAndrew Boyer txq->flags |= IONIC_QCQ_F_CSUM_L3; 182295968d1SFerruh Yigit if (offloads & RTE_ETH_TX_OFFLOAD_TCP_CKSUM) 18368591087SAndrew Boyer txq->flags |= IONIC_QCQ_F_CSUM_TCP; 184295968d1SFerruh Yigit if (offloads & RTE_ETH_TX_OFFLOAD_UDP_CKSUM) 18568591087SAndrew Boyer txq->flags |= IONIC_QCQ_F_CSUM_UDP; 1869ac234eeSAndrew Boyer if (offloads & RTE_ETH_TX_OFFLOAD_MBUF_FAST_FREE) 1879ac234eeSAndrew Boyer txq->flags |= IONIC_QCQ_F_FAST_FREE; 188a27d9013SAlfredo Cardigliano 18907512941SAndrew Boyer txq->free_thresh = 19007512941SAndrew Boyer tx_conf->tx_free_thresh ? tx_conf->tx_free_thresh : 19107512941SAndrew Boyer nb_desc - IONIC_DEF_TXRX_BURST; 19207512941SAndrew Boyer 193a27d9013SAlfredo Cardigliano eth_dev->data->tx_queues[tx_queue_id] = txq; 194a27d9013SAlfredo Cardigliano 195a27d9013SAlfredo Cardigliano return 0; 196a27d9013SAlfredo Cardigliano } 197a27d9013SAlfredo Cardigliano 198a27d9013SAlfredo Cardigliano /* 199a27d9013SAlfredo Cardigliano * Start Transmit Units for specified queue. 200a27d9013SAlfredo Cardigliano */ 201ce6427ddSThomas Monjalon int __rte_cold 202a27d9013SAlfredo Cardigliano ionic_dev_tx_queue_start(struct rte_eth_dev *eth_dev, uint16_t tx_queue_id) 203a27d9013SAlfredo Cardigliano { 2049fdf11c4SAndrew Boyer uint8_t *tx_queue_state = eth_dev->data->tx_queue_state; 205be39f75cSAndrew Boyer struct ionic_tx_qcq *txq; 206a27d9013SAlfredo Cardigliano int err; 207a27d9013SAlfredo Cardigliano 2089fdf11c4SAndrew Boyer if (tx_queue_state[tx_queue_id] == RTE_ETH_QUEUE_STATE_STARTED) { 2099fdf11c4SAndrew Boyer IONIC_PRINT(DEBUG, "TX queue %u already started", 2109fdf11c4SAndrew Boyer tx_queue_id); 2119fdf11c4SAndrew Boyer return 0; 2129fdf11c4SAndrew Boyer } 2139fdf11c4SAndrew Boyer 214a27d9013SAlfredo Cardigliano txq = eth_dev->data->tx_queues[tx_queue_id]; 215a27d9013SAlfredo Cardigliano 2164ae96cb8SAndrew Boyer IONIC_PRINT(DEBUG, "Starting TX queue %u, %u descs", 217be39f75cSAndrew Boyer tx_queue_id, txq->qcq.q.num_descs); 2184ae96cb8SAndrew Boyer 219a27d9013SAlfredo Cardigliano err = ionic_lif_txq_init(txq); 220a27d9013SAlfredo Cardigliano if (err) 221a27d9013SAlfredo Cardigliano return err; 222a27d9013SAlfredo Cardigliano 2239fdf11c4SAndrew Boyer tx_queue_state[tx_queue_id] = RTE_ETH_QUEUE_STATE_STARTED; 224a27d9013SAlfredo Cardigliano 225a27d9013SAlfredo Cardigliano return 0; 226a27d9013SAlfredo Cardigliano } 227a27d9013SAlfredo Cardigliano 228a27d9013SAlfredo Cardigliano static void 22964b08152SAlfredo Cardigliano ionic_tx_tcp_pseudo_csum(struct rte_mbuf *txm) 23064b08152SAlfredo Cardigliano { 23164b08152SAlfredo Cardigliano struct ether_hdr *eth_hdr = rte_pktmbuf_mtod(txm, struct ether_hdr *); 23264b08152SAlfredo Cardigliano char *l3_hdr = ((char *)eth_hdr) + txm->l2_len; 23364b08152SAlfredo Cardigliano struct rte_tcp_hdr *tcp_hdr = (struct rte_tcp_hdr *) 23464b08152SAlfredo Cardigliano (l3_hdr + txm->l3_len); 23564b08152SAlfredo Cardigliano 236daa02b5cSOlivier Matz if (txm->ol_flags & RTE_MBUF_F_TX_IP_CKSUM) { 23764b08152SAlfredo Cardigliano struct rte_ipv4_hdr *ipv4_hdr = (struct rte_ipv4_hdr *)l3_hdr; 23864b08152SAlfredo Cardigliano ipv4_hdr->hdr_checksum = 0; 23964b08152SAlfredo Cardigliano tcp_hdr->cksum = 0; 24064b08152SAlfredo Cardigliano tcp_hdr->cksum = rte_ipv4_udptcp_cksum(ipv4_hdr, tcp_hdr); 24164b08152SAlfredo Cardigliano } else { 24264b08152SAlfredo Cardigliano struct rte_ipv6_hdr *ipv6_hdr = (struct rte_ipv6_hdr *)l3_hdr; 24364b08152SAlfredo Cardigliano tcp_hdr->cksum = 0; 24464b08152SAlfredo Cardigliano tcp_hdr->cksum = rte_ipv6_udptcp_cksum(ipv6_hdr, tcp_hdr); 24564b08152SAlfredo Cardigliano } 24664b08152SAlfredo Cardigliano } 24764b08152SAlfredo Cardigliano 24864b08152SAlfredo Cardigliano static void 24964b08152SAlfredo Cardigliano ionic_tx_tcp_inner_pseudo_csum(struct rte_mbuf *txm) 25064b08152SAlfredo Cardigliano { 25164b08152SAlfredo Cardigliano struct ether_hdr *eth_hdr = rte_pktmbuf_mtod(txm, struct ether_hdr *); 25264b08152SAlfredo Cardigliano char *l3_hdr = ((char *)eth_hdr) + txm->outer_l2_len + 25364b08152SAlfredo Cardigliano txm->outer_l3_len + txm->l2_len; 25464b08152SAlfredo Cardigliano struct rte_tcp_hdr *tcp_hdr = (struct rte_tcp_hdr *) 25564b08152SAlfredo Cardigliano (l3_hdr + txm->l3_len); 25664b08152SAlfredo Cardigliano 257daa02b5cSOlivier Matz if (txm->ol_flags & RTE_MBUF_F_TX_IPV4) { 25864b08152SAlfredo Cardigliano struct rte_ipv4_hdr *ipv4_hdr = (struct rte_ipv4_hdr *)l3_hdr; 25964b08152SAlfredo Cardigliano ipv4_hdr->hdr_checksum = 0; 26064b08152SAlfredo Cardigliano tcp_hdr->cksum = 0; 26164b08152SAlfredo Cardigliano tcp_hdr->cksum = rte_ipv4_udptcp_cksum(ipv4_hdr, tcp_hdr); 26264b08152SAlfredo Cardigliano } else { 26364b08152SAlfredo Cardigliano struct rte_ipv6_hdr *ipv6_hdr = (struct rte_ipv6_hdr *)l3_hdr; 26464b08152SAlfredo Cardigliano tcp_hdr->cksum = 0; 26564b08152SAlfredo Cardigliano tcp_hdr->cksum = rte_ipv6_udptcp_cksum(ipv6_hdr, tcp_hdr); 26664b08152SAlfredo Cardigliano } 26764b08152SAlfredo Cardigliano } 26864b08152SAlfredo Cardigliano 26964b08152SAlfredo Cardigliano static void 270a27d9013SAlfredo Cardigliano ionic_tx_tso_post(struct ionic_queue *q, struct ionic_txq_desc *desc, 271a27d9013SAlfredo Cardigliano struct rte_mbuf *txm, 272a27d9013SAlfredo Cardigliano rte_iova_t addr, uint8_t nsge, uint16_t len, 273a27d9013SAlfredo Cardigliano uint32_t hdrlen, uint32_t mss, 27464b08152SAlfredo Cardigliano bool encap, 275a27d9013SAlfredo Cardigliano uint16_t vlan_tci, bool has_vlan, 276a27d9013SAlfredo Cardigliano bool start, bool done) 277a27d9013SAlfredo Cardigliano { 278b4beb84aSAndrew Boyer struct rte_mbuf *txm_seg; 279dd10c5b4SAndrew Boyer void **info; 2804a735599SAndrew Boyer uint64_t cmd; 281a27d9013SAlfredo Cardigliano uint8_t flags = 0; 282b4beb84aSAndrew Boyer int i; 283b4beb84aSAndrew Boyer 284a27d9013SAlfredo Cardigliano flags |= has_vlan ? IONIC_TXQ_DESC_FLAG_VLAN : 0; 28564b08152SAlfredo Cardigliano flags |= encap ? IONIC_TXQ_DESC_FLAG_ENCAP : 0; 286a27d9013SAlfredo Cardigliano flags |= start ? IONIC_TXQ_DESC_FLAG_TSO_SOT : 0; 287a27d9013SAlfredo Cardigliano flags |= done ? IONIC_TXQ_DESC_FLAG_TSO_EOT : 0; 288a27d9013SAlfredo Cardigliano 2894a735599SAndrew Boyer cmd = encode_txq_desc_cmd(IONIC_TXQ_DESC_OPCODE_TSO, 290a27d9013SAlfredo Cardigliano flags, nsge, addr); 2914a735599SAndrew Boyer desc->cmd = rte_cpu_to_le_64(cmd); 2924a735599SAndrew Boyer desc->len = rte_cpu_to_le_16(len); 2934a735599SAndrew Boyer desc->vlan_tci = rte_cpu_to_le_16(vlan_tci); 2944a735599SAndrew Boyer desc->hdr_len = rte_cpu_to_le_16(hdrlen); 2954a735599SAndrew Boyer desc->mss = rte_cpu_to_le_16(mss); 296a27d9013SAlfredo Cardigliano 297dd10c5b4SAndrew Boyer if (done) { 298dd10c5b4SAndrew Boyer info = IONIC_INFO_PTR(q, q->head_idx); 299b4beb84aSAndrew Boyer 300b4beb84aSAndrew Boyer /* Walk the mbuf chain to stash pointers in the array */ 301b4beb84aSAndrew Boyer txm_seg = txm; 302b4beb84aSAndrew Boyer for (i = 0; i < txm->nb_segs; i++) { 303b4beb84aSAndrew Boyer info[i] = txm_seg; 304b4beb84aSAndrew Boyer txm_seg = txm_seg->next; 305b4beb84aSAndrew Boyer } 306dd10c5b4SAndrew Boyer } 307dd10c5b4SAndrew Boyer 308dd10c5b4SAndrew Boyer q->head_idx = Q_NEXT_TO_POST(q, 1); 309a27d9013SAlfredo Cardigliano } 310a27d9013SAlfredo Cardigliano 311a27d9013SAlfredo Cardigliano static struct ionic_txq_desc * 312be39f75cSAndrew Boyer ionic_tx_tso_next(struct ionic_tx_qcq *txq, struct ionic_txq_sg_elem **elem) 313a27d9013SAlfredo Cardigliano { 314be39f75cSAndrew Boyer struct ionic_queue *q = &txq->qcq.q; 315a27d9013SAlfredo Cardigliano struct ionic_txq_desc *desc_base = q->base; 31656117636SAndrew Boyer struct ionic_txq_sg_desc_v1 *sg_desc_base = q->sg_base; 317a27d9013SAlfredo Cardigliano struct ionic_txq_desc *desc = &desc_base[q->head_idx]; 31856117636SAndrew Boyer struct ionic_txq_sg_desc_v1 *sg_desc = &sg_desc_base[q->head_idx]; 319a27d9013SAlfredo Cardigliano 320a27d9013SAlfredo Cardigliano *elem = sg_desc->elems; 321a27d9013SAlfredo Cardigliano return desc; 322a27d9013SAlfredo Cardigliano } 323a27d9013SAlfredo Cardigliano 324e86a6fccSAndrew Boyer int 32577c60793SAndrew Boyer ionic_tx_tso(struct ionic_tx_qcq *txq, struct rte_mbuf *txm) 326a27d9013SAlfredo Cardigliano { 327be39f75cSAndrew Boyer struct ionic_queue *q = &txq->qcq.q; 328be39f75cSAndrew Boyer struct ionic_tx_stats *stats = &txq->stats; 329a27d9013SAlfredo Cardigliano struct ionic_txq_desc *desc; 330a27d9013SAlfredo Cardigliano struct ionic_txq_sg_elem *elem; 331a27d9013SAlfredo Cardigliano struct rte_mbuf *txm_seg; 3327c3a867bSAndrew Boyer rte_iova_t data_iova; 3337c3a867bSAndrew Boyer uint64_t desc_addr = 0, next_addr; 334a27d9013SAlfredo Cardigliano uint16_t desc_len = 0; 335e86a6fccSAndrew Boyer uint8_t desc_nsge = 0; 336a27d9013SAlfredo Cardigliano uint32_t hdrlen; 337a27d9013SAlfredo Cardigliano uint32_t mss = txm->tso_segsz; 338a27d9013SAlfredo Cardigliano uint32_t frag_left = 0; 339a27d9013SAlfredo Cardigliano uint32_t left; 340a27d9013SAlfredo Cardigliano uint32_t seglen; 341a27d9013SAlfredo Cardigliano uint32_t len; 342a27d9013SAlfredo Cardigliano uint32_t offset = 0; 343a27d9013SAlfredo Cardigliano bool start, done; 34464b08152SAlfredo Cardigliano bool encap; 345daa02b5cSOlivier Matz bool has_vlan = !!(txm->ol_flags & RTE_MBUF_F_TX_VLAN); 346e86a6fccSAndrew Boyer bool use_sgl = !!(txq->flags & IONIC_QCQ_F_SG); 347a27d9013SAlfredo Cardigliano uint16_t vlan_tci = txm->vlan_tci; 34864b08152SAlfredo Cardigliano uint64_t ol_flags = txm->ol_flags; 349a27d9013SAlfredo Cardigliano 350daa02b5cSOlivier Matz encap = ((ol_flags & RTE_MBUF_F_TX_OUTER_IP_CKSUM) || 351daa02b5cSOlivier Matz (ol_flags & RTE_MBUF_F_TX_OUTER_UDP_CKSUM)) && 352daa02b5cSOlivier Matz ((ol_flags & RTE_MBUF_F_TX_OUTER_IPV4) || 353daa02b5cSOlivier Matz (ol_flags & RTE_MBUF_F_TX_OUTER_IPV6)); 35464b08152SAlfredo Cardigliano 35564b08152SAlfredo Cardigliano /* Preload inner-most TCP csum field with IP pseudo hdr 35664b08152SAlfredo Cardigliano * calculated with IP length set to zero. HW will later 35764b08152SAlfredo Cardigliano * add in length to each TCP segment resulting from the TSO. 35864b08152SAlfredo Cardigliano */ 35964b08152SAlfredo Cardigliano 36064b08152SAlfredo Cardigliano if (encap) { 36164b08152SAlfredo Cardigliano ionic_tx_tcp_inner_pseudo_csum(txm); 36264b08152SAlfredo Cardigliano hdrlen = txm->outer_l2_len + txm->outer_l3_len + 36364b08152SAlfredo Cardigliano txm->l2_len + txm->l3_len + txm->l4_len; 36464b08152SAlfredo Cardigliano } else { 36564b08152SAlfredo Cardigliano ionic_tx_tcp_pseudo_csum(txm); 36664b08152SAlfredo Cardigliano hdrlen = txm->l2_len + txm->l3_len + txm->l4_len; 36764b08152SAlfredo Cardigliano } 368a27d9013SAlfredo Cardigliano 369be39f75cSAndrew Boyer desc = ionic_tx_tso_next(txq, &elem); 370e86a6fccSAndrew Boyer txm_seg = txm; 371a27d9013SAlfredo Cardigliano start = true; 372e86a6fccSAndrew Boyer seglen = hdrlen + mss; 373a27d9013SAlfredo Cardigliano 374e86a6fccSAndrew Boyer /* Walk the chain of mbufs */ 375a27d9013SAlfredo Cardigliano while (txm_seg != NULL) { 376a27d9013SAlfredo Cardigliano offset = 0; 3777c3a867bSAndrew Boyer data_iova = rte_mbuf_data_iova(txm_seg); 378a27d9013SAlfredo Cardigliano left = txm_seg->data_len; 379a27d9013SAlfredo Cardigliano 380e86a6fccSAndrew Boyer /* Split the mbuf data up into multiple descriptors */ 381a27d9013SAlfredo Cardigliano while (left > 0) { 3827c3a867bSAndrew Boyer next_addr = rte_cpu_to_le_64(data_iova + offset); 383e86a6fccSAndrew Boyer if (frag_left > 0 && use_sgl) { 384e86a6fccSAndrew Boyer /* Fill previous descriptor's SGE */ 385a27d9013SAlfredo Cardigliano len = RTE_MIN(frag_left, left); 386a27d9013SAlfredo Cardigliano frag_left -= len; 3877c3a867bSAndrew Boyer elem->addr = next_addr; 3884a735599SAndrew Boyer elem->len = rte_cpu_to_le_16(len); 389a27d9013SAlfredo Cardigliano elem++; 390a27d9013SAlfredo Cardigliano desc_nsge++; 391a27d9013SAlfredo Cardigliano } else { 392e86a6fccSAndrew Boyer /* Fill new descriptor's data field */ 393e86a6fccSAndrew Boyer len = RTE_MIN(seglen, left); 394e86a6fccSAndrew Boyer frag_left = seglen - len; 3957c3a867bSAndrew Boyer desc_addr = next_addr; 396a27d9013SAlfredo Cardigliano desc_len = len; 397a27d9013SAlfredo Cardigliano desc_nsge = 0; 398a27d9013SAlfredo Cardigliano } 399a27d9013SAlfredo Cardigliano left -= len; 400a27d9013SAlfredo Cardigliano offset += len; 401e86a6fccSAndrew Boyer 402e86a6fccSAndrew Boyer /* Pack the next mbuf's data into the descriptor */ 403e86a6fccSAndrew Boyer if (txm_seg->next != NULL && frag_left > 0 && use_sgl) 404e86a6fccSAndrew Boyer break; 4057c3a867bSAndrew Boyer 406a27d9013SAlfredo Cardigliano done = (txm_seg->next == NULL && left == 0); 407a27d9013SAlfredo Cardigliano ionic_tx_tso_post(q, desc, txm_seg, 408a27d9013SAlfredo Cardigliano desc_addr, desc_nsge, desc_len, 409a27d9013SAlfredo Cardigliano hdrlen, mss, 41064b08152SAlfredo Cardigliano encap, 411a27d9013SAlfredo Cardigliano vlan_tci, has_vlan, 41277c60793SAndrew Boyer start, done); 413be39f75cSAndrew Boyer desc = ionic_tx_tso_next(txq, &elem); 414a27d9013SAlfredo Cardigliano start = false; 415e86a6fccSAndrew Boyer seglen = mss; 416a27d9013SAlfredo Cardigliano } 417a27d9013SAlfredo Cardigliano 418a27d9013SAlfredo Cardigliano txm_seg = txm_seg->next; 419a27d9013SAlfredo Cardigliano } 420a27d9013SAlfredo Cardigliano 421a27d9013SAlfredo Cardigliano stats->tso++; 422a27d9013SAlfredo Cardigliano 423a27d9013SAlfredo Cardigliano return 0; 424a27d9013SAlfredo Cardigliano } 425a27d9013SAlfredo Cardigliano 426a27d9013SAlfredo Cardigliano /********************************************************************* 427a27d9013SAlfredo Cardigliano * 428a27d9013SAlfredo Cardigliano * TX prep functions 429a27d9013SAlfredo Cardigliano * 430a27d9013SAlfredo Cardigliano **********************************************************************/ 431a27d9013SAlfredo Cardigliano 432daa02b5cSOlivier Matz #define IONIC_TX_OFFLOAD_MASK (RTE_MBUF_F_TX_IPV4 | \ 433daa02b5cSOlivier Matz RTE_MBUF_F_TX_IPV6 | \ 434daa02b5cSOlivier Matz RTE_MBUF_F_TX_VLAN | \ 435daa02b5cSOlivier Matz RTE_MBUF_F_TX_IP_CKSUM | \ 436daa02b5cSOlivier Matz RTE_MBUF_F_TX_TCP_SEG | \ 437daa02b5cSOlivier Matz RTE_MBUF_F_TX_L4_MASK) 438a27d9013SAlfredo Cardigliano 439a27d9013SAlfredo Cardigliano #define IONIC_TX_OFFLOAD_NOTSUP_MASK \ 440daa02b5cSOlivier Matz (RTE_MBUF_F_TX_OFFLOAD_MASK ^ IONIC_TX_OFFLOAD_MASK) 441a27d9013SAlfredo Cardigliano 442a27d9013SAlfredo Cardigliano uint16_t 443e19eea1eSAndrew Boyer ionic_prep_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, uint16_t nb_pkts) 444a27d9013SAlfredo Cardigliano { 445e19eea1eSAndrew Boyer struct ionic_tx_qcq *txq = tx_queue; 446a27d9013SAlfredo Cardigliano struct rte_mbuf *txm; 447a27d9013SAlfredo Cardigliano uint64_t offloads; 448a27d9013SAlfredo Cardigliano int i = 0; 449a27d9013SAlfredo Cardigliano 450a27d9013SAlfredo Cardigliano for (i = 0; i < nb_pkts; i++) { 451a27d9013SAlfredo Cardigliano txm = tx_pkts[i]; 452a27d9013SAlfredo Cardigliano 453e19eea1eSAndrew Boyer if (txm->nb_segs > txq->num_segs_fw) { 454a27d9013SAlfredo Cardigliano rte_errno = -EINVAL; 455a27d9013SAlfredo Cardigliano break; 456a27d9013SAlfredo Cardigliano } 457a27d9013SAlfredo Cardigliano 458a27d9013SAlfredo Cardigliano offloads = txm->ol_flags; 459a27d9013SAlfredo Cardigliano 460a27d9013SAlfredo Cardigliano if (offloads & IONIC_TX_OFFLOAD_NOTSUP_MASK) { 461a27d9013SAlfredo Cardigliano rte_errno = -ENOTSUP; 462a27d9013SAlfredo Cardigliano break; 463a27d9013SAlfredo Cardigliano } 464a27d9013SAlfredo Cardigliano } 465a27d9013SAlfredo Cardigliano 466a27d9013SAlfredo Cardigliano return i; 467a27d9013SAlfredo Cardigliano } 468a27d9013SAlfredo Cardigliano 469a27d9013SAlfredo Cardigliano /********************************************************************* 470a27d9013SAlfredo Cardigliano * 471a27d9013SAlfredo Cardigliano * RX functions 472a27d9013SAlfredo Cardigliano * 473a27d9013SAlfredo Cardigliano **********************************************************************/ 474a27d9013SAlfredo Cardigliano 475a27d9013SAlfredo Cardigliano void 476a27d9013SAlfredo Cardigliano ionic_rxq_info_get(struct rte_eth_dev *dev, uint16_t queue_id, 477a27d9013SAlfredo Cardigliano struct rte_eth_rxq_info *qinfo) 478a27d9013SAlfredo Cardigliano { 479be39f75cSAndrew Boyer struct ionic_rx_qcq *rxq = dev->data->rx_queues[queue_id]; 480be39f75cSAndrew Boyer struct ionic_queue *q = &rxq->qcq.q; 481a27d9013SAlfredo Cardigliano 482a27d9013SAlfredo Cardigliano qinfo->mp = rxq->mb_pool; 483a27d9013SAlfredo Cardigliano qinfo->scattered_rx = dev->data->scattered_rx; 484a27d9013SAlfredo Cardigliano qinfo->nb_desc = q->num_descs; 48502eabf57SAndrew Boyer qinfo->conf.rx_deferred_start = rxq->flags & IONIC_QCQ_F_DEFERRED; 48668591087SAndrew Boyer qinfo->conf.offloads = dev->data->dev_conf.rxmode.offloads; 487a27d9013SAlfredo Cardigliano } 488a27d9013SAlfredo Cardigliano 489ce6427ddSThomas Monjalon void __rte_cold 4907483341aSXueming Li ionic_dev_rx_queue_release(struct rte_eth_dev *dev, uint16_t qid) 491a27d9013SAlfredo Cardigliano { 4927483341aSXueming Li struct ionic_rx_qcq *rxq = dev->data->rx_queues[qid]; 493be39f75cSAndrew Boyer 494be39f75cSAndrew Boyer if (!rxq) 495be39f75cSAndrew Boyer return; 496a27d9013SAlfredo Cardigliano 497a27d9013SAlfredo Cardigliano IONIC_PRINT_CALL(); 498a27d9013SAlfredo Cardigliano 499be39f75cSAndrew Boyer ionic_qcq_free(&rxq->qcq); 500a27d9013SAlfredo Cardigliano } 501a27d9013SAlfredo Cardigliano 502ce6427ddSThomas Monjalon int __rte_cold 503a27d9013SAlfredo Cardigliano ionic_dev_rx_queue_setup(struct rte_eth_dev *eth_dev, 504a27d9013SAlfredo Cardigliano uint16_t rx_queue_id, 505a27d9013SAlfredo Cardigliano uint16_t nb_desc, 5064ae96cb8SAndrew Boyer uint32_t socket_id, 507a27d9013SAlfredo Cardigliano const struct rte_eth_rxconf *rx_conf, 508a27d9013SAlfredo Cardigliano struct rte_mempool *mp) 509a27d9013SAlfredo Cardigliano { 510a27d9013SAlfredo Cardigliano struct ionic_lif *lif = IONIC_ETH_DEV_TO_LIF(eth_dev); 511be39f75cSAndrew Boyer struct ionic_rx_qcq *rxq; 512a27d9013SAlfredo Cardigliano uint64_t offloads; 513a27d9013SAlfredo Cardigliano int err; 514a27d9013SAlfredo Cardigliano 515a27d9013SAlfredo Cardigliano if (rx_queue_id >= lif->nrxqcqs) { 516a27d9013SAlfredo Cardigliano IONIC_PRINT(ERR, 517a27d9013SAlfredo Cardigliano "Queue index %u not available (max %u queues)", 518a27d9013SAlfredo Cardigliano rx_queue_id, lif->nrxqcqs); 519a27d9013SAlfredo Cardigliano return -EINVAL; 520a27d9013SAlfredo Cardigliano } 521a27d9013SAlfredo Cardigliano 522a27d9013SAlfredo Cardigliano offloads = rx_conf->offloads | eth_dev->data->dev_conf.rxmode.offloads; 5234ae96cb8SAndrew Boyer IONIC_PRINT(DEBUG, 5244ae96cb8SAndrew Boyer "Configuring skt %u RX queue %u with %u buffers, offloads %jx", 5254ae96cb8SAndrew Boyer socket_id, rx_queue_id, nb_desc, offloads); 526a27d9013SAlfredo Cardigliano 52718a44465SAndrew Boyer if (!rx_conf->rx_drop_en) 52818a44465SAndrew Boyer IONIC_PRINT(WARNING, "No-drop mode is not supported"); 52918a44465SAndrew Boyer 530a27d9013SAlfredo Cardigliano /* Validate number of receive descriptors */ 531a27d9013SAlfredo Cardigliano if (!rte_is_power_of_2(nb_desc) || 532a27d9013SAlfredo Cardigliano nb_desc < IONIC_MIN_RING_DESC || 533a27d9013SAlfredo Cardigliano nb_desc > IONIC_MAX_RING_DESC) { 534a27d9013SAlfredo Cardigliano IONIC_PRINT(ERR, 5354ae96cb8SAndrew Boyer "Bad descriptor count (%u) for queue %u (min: %u)", 536a27d9013SAlfredo Cardigliano nb_desc, rx_queue_id, IONIC_MIN_RING_DESC); 537a27d9013SAlfredo Cardigliano return -EINVAL; /* or use IONIC_DEFAULT_RING_DESC */ 538a27d9013SAlfredo Cardigliano } 539a27d9013SAlfredo Cardigliano 540a27d9013SAlfredo Cardigliano /* Free memory prior to re-allocation if needed... */ 541a27d9013SAlfredo Cardigliano if (eth_dev->data->rx_queues[rx_queue_id] != NULL) { 5427483341aSXueming Li ionic_dev_rx_queue_release(eth_dev, rx_queue_id); 543a27d9013SAlfredo Cardigliano eth_dev->data->rx_queues[rx_queue_id] = NULL; 544a27d9013SAlfredo Cardigliano } 545a27d9013SAlfredo Cardigliano 5469fdf11c4SAndrew Boyer eth_dev->data->rx_queue_state[rx_queue_id] = 5479fdf11c4SAndrew Boyer RTE_ETH_QUEUE_STATE_STOPPED; 5489fdf11c4SAndrew Boyer 549d5850081SAndrew Boyer err = ionic_rx_qcq_alloc(lif, socket_id, rx_queue_id, nb_desc, mp, 550be39f75cSAndrew Boyer &rxq); 551a27d9013SAlfredo Cardigliano if (err) { 5524ae96cb8SAndrew Boyer IONIC_PRINT(ERR, "Queue %d allocation failure", rx_queue_id); 553a27d9013SAlfredo Cardigliano return -EINVAL; 554a27d9013SAlfredo Cardigliano } 555a27d9013SAlfredo Cardigliano 556a27d9013SAlfredo Cardigliano rxq->mb_pool = mp; 557a5b1ffd8SAndrew Boyer rxq->wdog_ms = IONIC_Q_WDOG_MS; 558a27d9013SAlfredo Cardigliano 559a27d9013SAlfredo Cardigliano /* 560a27d9013SAlfredo Cardigliano * Note: the interface does not currently support 561295968d1SFerruh Yigit * RTE_ETH_RX_OFFLOAD_KEEP_CRC, please also consider ETHER_CRC_LEN 562a27d9013SAlfredo Cardigliano * when the adapter will be able to keep the CRC and subtract 563a27d9013SAlfredo Cardigliano * it to the length for all received packets: 564a27d9013SAlfredo Cardigliano * if (eth_dev->data->dev_conf.rxmode.offloads & 565295968d1SFerruh Yigit * RTE_ETH_RX_OFFLOAD_KEEP_CRC) 566a27d9013SAlfredo Cardigliano * rxq->crc_len = ETHER_CRC_LEN; 567a27d9013SAlfredo Cardigliano */ 568a27d9013SAlfredo Cardigliano 569a27d9013SAlfredo Cardigliano /* Do not start queue with rte_eth_dev_start() */ 57002eabf57SAndrew Boyer if (rx_conf->rx_deferred_start) 57102eabf57SAndrew Boyer rxq->flags |= IONIC_QCQ_F_DEFERRED; 572a27d9013SAlfredo Cardigliano 573a27d9013SAlfredo Cardigliano eth_dev->data->rx_queues[rx_queue_id] = rxq; 574a27d9013SAlfredo Cardigliano 575a27d9013SAlfredo Cardigliano return 0; 576a27d9013SAlfredo Cardigliano } 577a27d9013SAlfredo Cardigliano 578bbdf955dSAndrew Boyer #define IONIC_CSUM_FLAG_MASK (IONIC_RXQ_COMP_CSUM_F_VLAN - 1) 579e86a6fccSAndrew Boyer const uint64_t ionic_csum_flags[IONIC_CSUM_FLAG_MASK] 580bbdf955dSAndrew Boyer __rte_cache_aligned = { 581bbdf955dSAndrew Boyer /* IP_BAD set */ 582bbdf955dSAndrew Boyer [IONIC_RXQ_COMP_CSUM_F_IP_BAD] = RTE_MBUF_F_RX_IP_CKSUM_BAD, 583bbdf955dSAndrew Boyer [IONIC_RXQ_COMP_CSUM_F_IP_BAD | IONIC_RXQ_COMP_CSUM_F_TCP_OK] = 584bbdf955dSAndrew Boyer RTE_MBUF_F_RX_IP_CKSUM_BAD | RTE_MBUF_F_RX_L4_CKSUM_GOOD, 585bbdf955dSAndrew Boyer [IONIC_RXQ_COMP_CSUM_F_IP_BAD | IONIC_RXQ_COMP_CSUM_F_TCP_BAD] = 586bbdf955dSAndrew Boyer RTE_MBUF_F_RX_IP_CKSUM_BAD | RTE_MBUF_F_RX_L4_CKSUM_BAD, 587bbdf955dSAndrew Boyer [IONIC_RXQ_COMP_CSUM_F_IP_BAD | IONIC_RXQ_COMP_CSUM_F_UDP_OK] = 588bbdf955dSAndrew Boyer RTE_MBUF_F_RX_IP_CKSUM_BAD | RTE_MBUF_F_RX_L4_CKSUM_GOOD, 589bbdf955dSAndrew Boyer [IONIC_RXQ_COMP_CSUM_F_IP_BAD | IONIC_RXQ_COMP_CSUM_F_UDP_BAD] = 590bbdf955dSAndrew Boyer RTE_MBUF_F_RX_IP_CKSUM_BAD | RTE_MBUF_F_RX_L4_CKSUM_BAD, 591bbdf955dSAndrew Boyer /* IP_OK set */ 592bbdf955dSAndrew Boyer [IONIC_RXQ_COMP_CSUM_F_IP_OK] = RTE_MBUF_F_RX_IP_CKSUM_GOOD, 593bbdf955dSAndrew Boyer [IONIC_RXQ_COMP_CSUM_F_IP_OK | IONIC_RXQ_COMP_CSUM_F_TCP_OK] = 594bbdf955dSAndrew Boyer RTE_MBUF_F_RX_IP_CKSUM_GOOD | RTE_MBUF_F_RX_L4_CKSUM_GOOD, 595bbdf955dSAndrew Boyer [IONIC_RXQ_COMP_CSUM_F_IP_OK | IONIC_RXQ_COMP_CSUM_F_TCP_BAD] = 596bbdf955dSAndrew Boyer RTE_MBUF_F_RX_IP_CKSUM_GOOD | RTE_MBUF_F_RX_L4_CKSUM_BAD, 597bbdf955dSAndrew Boyer [IONIC_RXQ_COMP_CSUM_F_IP_OK | IONIC_RXQ_COMP_CSUM_F_UDP_OK] = 598bbdf955dSAndrew Boyer RTE_MBUF_F_RX_IP_CKSUM_GOOD | RTE_MBUF_F_RX_L4_CKSUM_GOOD, 599bbdf955dSAndrew Boyer [IONIC_RXQ_COMP_CSUM_F_IP_OK | IONIC_RXQ_COMP_CSUM_F_UDP_BAD] = 600bbdf955dSAndrew Boyer RTE_MBUF_F_RX_IP_CKSUM_GOOD | RTE_MBUF_F_RX_L4_CKSUM_BAD, 601bbdf955dSAndrew Boyer /* No IP flag set */ 602bbdf955dSAndrew Boyer [IONIC_RXQ_COMP_CSUM_F_TCP_OK] = RTE_MBUF_F_RX_L4_CKSUM_GOOD, 603bbdf955dSAndrew Boyer [IONIC_RXQ_COMP_CSUM_F_TCP_BAD] = RTE_MBUF_F_RX_L4_CKSUM_BAD, 604bbdf955dSAndrew Boyer [IONIC_RXQ_COMP_CSUM_F_UDP_OK] = RTE_MBUF_F_RX_L4_CKSUM_GOOD, 605bbdf955dSAndrew Boyer [IONIC_RXQ_COMP_CSUM_F_UDP_BAD] = RTE_MBUF_F_RX_L4_CKSUM_BAD, 606bbdf955dSAndrew Boyer }; 607bbdf955dSAndrew Boyer 60873b1c67eSAndrew Boyer /* RTE_PTYPE_UNKNOWN is 0x0 */ 609e86a6fccSAndrew Boyer const uint32_t ionic_ptype_table[IONIC_RXQ_COMP_PKT_TYPE_MASK] 61073b1c67eSAndrew Boyer __rte_cache_aligned = { 61173b1c67eSAndrew Boyer [IONIC_PKT_TYPE_NON_IP] = RTE_PTYPE_UNKNOWN, 61273b1c67eSAndrew Boyer [IONIC_PKT_TYPE_IPV4] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4, 61373b1c67eSAndrew Boyer [IONIC_PKT_TYPE_IPV4_TCP] = 61473b1c67eSAndrew Boyer RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4 | RTE_PTYPE_L4_TCP, 61573b1c67eSAndrew Boyer [IONIC_PKT_TYPE_IPV4_UDP] = 61673b1c67eSAndrew Boyer RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4 | RTE_PTYPE_L4_UDP, 61773b1c67eSAndrew Boyer [IONIC_PKT_TYPE_IPV6] = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6, 61873b1c67eSAndrew Boyer [IONIC_PKT_TYPE_IPV6_TCP] = 61973b1c67eSAndrew Boyer RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6 | RTE_PTYPE_L4_TCP, 62073b1c67eSAndrew Boyer [IONIC_PKT_TYPE_IPV6_UDP] = 62173b1c67eSAndrew Boyer RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6 | RTE_PTYPE_L4_UDP, 62273b1c67eSAndrew Boyer }; 62373b1c67eSAndrew Boyer 624b5b56afdSAndrew Boyer const uint32_t * 625ba6a168aSSivaramakrishnan Venkat ionic_dev_supported_ptypes_get(struct rte_eth_dev *dev __rte_unused, 626ba6a168aSSivaramakrishnan Venkat size_t *no_of_elements) 627b5b56afdSAndrew Boyer { 628b5b56afdSAndrew Boyer /* See ionic_ptype_table[] */ 629b5b56afdSAndrew Boyer static const uint32_t ptypes[] = { 630b5b56afdSAndrew Boyer RTE_PTYPE_L2_ETHER, 631b5b56afdSAndrew Boyer RTE_PTYPE_L2_ETHER_TIMESYNC, 632b5b56afdSAndrew Boyer RTE_PTYPE_L2_ETHER_LLDP, 633b5b56afdSAndrew Boyer RTE_PTYPE_L2_ETHER_ARP, 634b5b56afdSAndrew Boyer RTE_PTYPE_L3_IPV4, 635b5b56afdSAndrew Boyer RTE_PTYPE_L3_IPV6, 636b5b56afdSAndrew Boyer RTE_PTYPE_L4_TCP, 637b5b56afdSAndrew Boyer RTE_PTYPE_L4_UDP, 638b5b56afdSAndrew Boyer }; 639b5b56afdSAndrew Boyer 640ba6a168aSSivaramakrishnan Venkat *no_of_elements = RTE_DIM(ptypes); 641b5b56afdSAndrew Boyer return ptypes; 642b5b56afdSAndrew Boyer } 643b5b56afdSAndrew Boyer 6447b20fc2fSAndrew Boyer /* 6457b2eb674SAndrew Boyer * Perform one-time initialization of descriptor fields 6467b2eb674SAndrew Boyer * which will not change for the life of the queue. 6477b2eb674SAndrew Boyer */ 6487b2eb674SAndrew Boyer static void __rte_cold 6497b2eb674SAndrew Boyer ionic_rx_init_descriptors(struct ionic_rx_qcq *rxq) 6507b2eb674SAndrew Boyer { 6517b2eb674SAndrew Boyer struct ionic_queue *q = &rxq->qcq.q; 6527b2eb674SAndrew Boyer struct ionic_rxq_desc *desc, *desc_base = q->base; 6537b2eb674SAndrew Boyer struct ionic_rxq_sg_desc *sg_desc, *sg_desc_base = q->sg_base; 6547b2eb674SAndrew Boyer uint32_t i, j; 6557b2eb674SAndrew Boyer uint8_t opcode; 6567b2eb674SAndrew Boyer 6577b2eb674SAndrew Boyer opcode = (q->num_segs > 1) ? 6587b2eb674SAndrew Boyer IONIC_RXQ_DESC_OPCODE_SG : IONIC_RXQ_DESC_OPCODE_SIMPLE; 6597b2eb674SAndrew Boyer 6607b2eb674SAndrew Boyer /* 6617b2eb674SAndrew Boyer * NB: Only the first segment needs to leave headroom (hdr_seg_size). 6627b2eb674SAndrew Boyer * Later segments (seg_size) do not. 6637b2eb674SAndrew Boyer */ 6647b2eb674SAndrew Boyer for (i = 0; i < q->num_descs; i++) { 6657b2eb674SAndrew Boyer desc = &desc_base[i]; 6667b2eb674SAndrew Boyer desc->len = rte_cpu_to_le_16(rxq->hdr_seg_size); 6677b2eb674SAndrew Boyer desc->opcode = opcode; 6687b2eb674SAndrew Boyer 6697b2eb674SAndrew Boyer sg_desc = &sg_desc_base[i]; 6707b2eb674SAndrew Boyer for (j = 0; j < q->num_segs - 1u; j++) 6717b2eb674SAndrew Boyer sg_desc->elems[j].len = 6727b2eb674SAndrew Boyer rte_cpu_to_le_16(rxq->seg_size); 6737b2eb674SAndrew Boyer } 6747b2eb674SAndrew Boyer } 6757b2eb674SAndrew Boyer 6767b2eb674SAndrew Boyer /* 677a27d9013SAlfredo Cardigliano * Start Receive Units for specified queue. 678a27d9013SAlfredo Cardigliano */ 679ce6427ddSThomas Monjalon int __rte_cold 680a27d9013SAlfredo Cardigliano ionic_dev_rx_queue_start(struct rte_eth_dev *eth_dev, uint16_t rx_queue_id) 681a27d9013SAlfredo Cardigliano { 6829fdf11c4SAndrew Boyer uint8_t *rx_queue_state = eth_dev->data->rx_queue_state; 683be39f75cSAndrew Boyer struct ionic_rx_qcq *rxq; 684d5850081SAndrew Boyer struct ionic_queue *q; 685a27d9013SAlfredo Cardigliano int err; 686a27d9013SAlfredo Cardigliano 6879fdf11c4SAndrew Boyer if (rx_queue_state[rx_queue_id] == RTE_ETH_QUEUE_STATE_STARTED) { 6889fdf11c4SAndrew Boyer IONIC_PRINT(DEBUG, "RX queue %u already started", 6899fdf11c4SAndrew Boyer rx_queue_id); 6909fdf11c4SAndrew Boyer return 0; 6919fdf11c4SAndrew Boyer } 6929fdf11c4SAndrew Boyer 693a27d9013SAlfredo Cardigliano rxq = eth_dev->data->rx_queues[rx_queue_id]; 694d5850081SAndrew Boyer q = &rxq->qcq.q; 695a27d9013SAlfredo Cardigliano 696b671e69aSAndrew Boyer rxq->frame_size = rxq->qcq.lif->frame_size - RTE_ETHER_CRC_LEN; 697b671e69aSAndrew Boyer 698d5850081SAndrew Boyer /* Recalculate segment count based on MTU */ 699d5850081SAndrew Boyer q->num_segs = 1 + 700d5850081SAndrew Boyer (rxq->frame_size + RTE_PKTMBUF_HEADROOM - 1) / rxq->seg_size; 701d5850081SAndrew Boyer 702d5850081SAndrew Boyer IONIC_PRINT(DEBUG, "Starting RX queue %u, %u descs, size %u segs %u", 703d5850081SAndrew Boyer rx_queue_id, q->num_descs, rxq->frame_size, q->num_segs); 7044ae96cb8SAndrew Boyer 7057b2eb674SAndrew Boyer ionic_rx_init_descriptors(rxq); 7067b2eb674SAndrew Boyer 707a27d9013SAlfredo Cardigliano err = ionic_lif_rxq_init(rxq); 708a27d9013SAlfredo Cardigliano if (err) 709a27d9013SAlfredo Cardigliano return err; 710a27d9013SAlfredo Cardigliano 711e86a6fccSAndrew Boyer /* Allocate buffers for descriptor ring */ 712e86a6fccSAndrew Boyer if (rxq->flags & IONIC_QCQ_F_SG) 713e86a6fccSAndrew Boyer err = ionic_rx_fill_sg(rxq); 714e86a6fccSAndrew Boyer else 715e86a6fccSAndrew Boyer err = ionic_rx_fill(rxq); 716e86a6fccSAndrew Boyer if (err != 0) { 717e86a6fccSAndrew Boyer IONIC_PRINT(ERR, "Could not fill queue %d", rx_queue_id); 718a27d9013SAlfredo Cardigliano return -1; 719a27d9013SAlfredo Cardigliano } 720a27d9013SAlfredo Cardigliano 7219fdf11c4SAndrew Boyer rx_queue_state[rx_queue_id] = RTE_ETH_QUEUE_STATE_STARTED; 722a27d9013SAlfredo Cardigliano 723a27d9013SAlfredo Cardigliano return 0; 724a27d9013SAlfredo Cardigliano } 725a27d9013SAlfredo Cardigliano 7267b20fc2fSAndrew Boyer /* 727a27d9013SAlfredo Cardigliano * Stop Receive Units for specified queue. 728a27d9013SAlfredo Cardigliano */ 729ce6427ddSThomas Monjalon int __rte_cold 730a27d9013SAlfredo Cardigliano ionic_dev_rx_queue_stop(struct rte_eth_dev *eth_dev, uint16_t rx_queue_id) 731a27d9013SAlfredo Cardigliano { 732e7222f94SAndrew Boyer uint8_t *rx_queue_state = eth_dev->data->rx_queue_state; 733e7222f94SAndrew Boyer struct ionic_rx_stats *stats; 734be39f75cSAndrew Boyer struct ionic_rx_qcq *rxq; 735a27d9013SAlfredo Cardigliano 7364ae96cb8SAndrew Boyer IONIC_PRINT(DEBUG, "Stopping RX queue %u", rx_queue_id); 737a27d9013SAlfredo Cardigliano 738a27d9013SAlfredo Cardigliano rxq = eth_dev->data->rx_queues[rx_queue_id]; 739a27d9013SAlfredo Cardigliano 740e7222f94SAndrew Boyer rx_queue_state[rx_queue_id] = RTE_ETH_QUEUE_STATE_STOPPED; 7419fdf11c4SAndrew Boyer 742e7222f94SAndrew Boyer ionic_lif_rxq_deinit(rxq); 743a27d9013SAlfredo Cardigliano 744e7222f94SAndrew Boyer /* Free all buffers from descriptor ring */ 745e7222f94SAndrew Boyer ionic_rx_empty(rxq); 746e7222f94SAndrew Boyer 747e7222f94SAndrew Boyer stats = &rxq->stats; 748e7222f94SAndrew Boyer IONIC_PRINT(DEBUG, "RX queue %u pkts %ju mtod %ju", 749e7222f94SAndrew Boyer rxq->qcq.q.index, stats->packets, stats->mtods); 750a27d9013SAlfredo Cardigliano 751a27d9013SAlfredo Cardigliano return 0; 752a27d9013SAlfredo Cardigliano } 753a27d9013SAlfredo Cardigliano 7540983a74aSAndrew Boyer int 7550983a74aSAndrew Boyer ionic_dev_rx_descriptor_status(void *rx_queue, uint16_t offset) 7560983a74aSAndrew Boyer { 7570983a74aSAndrew Boyer struct ionic_rx_qcq *rxq = rx_queue; 7580983a74aSAndrew Boyer struct ionic_qcq *qcq = &rxq->qcq; 759*463ad260SNeel Patel volatile struct ionic_rxq_comp *cq_desc; 7600983a74aSAndrew Boyer uint16_t mask, head, tail, pos; 7610983a74aSAndrew Boyer bool done_color; 7620983a74aSAndrew Boyer 7630983a74aSAndrew Boyer mask = qcq->q.size_mask; 7640983a74aSAndrew Boyer 7650983a74aSAndrew Boyer /* offset must be within the size of the ring */ 7660983a74aSAndrew Boyer if (offset > mask) 7670983a74aSAndrew Boyer return -EINVAL; 7680983a74aSAndrew Boyer 7690983a74aSAndrew Boyer head = qcq->q.head_idx; 7700983a74aSAndrew Boyer tail = qcq->q.tail_idx; 7710983a74aSAndrew Boyer 7720983a74aSAndrew Boyer /* offset is beyond what is posted */ 7730983a74aSAndrew Boyer if (offset >= ((head - tail) & mask)) 7740983a74aSAndrew Boyer return RTE_ETH_RX_DESC_UNAVAIL; 7750983a74aSAndrew Boyer 7760983a74aSAndrew Boyer /* interested in this absolute position in the rxq */ 7770983a74aSAndrew Boyer pos = (tail + offset) & mask; 7780983a74aSAndrew Boyer 7790983a74aSAndrew Boyer /* rx cq position == rx q position */ 7800983a74aSAndrew Boyer cq_desc = qcq->cq.base; 7810983a74aSAndrew Boyer cq_desc = &cq_desc[pos]; 7820983a74aSAndrew Boyer 7830983a74aSAndrew Boyer /* expected done color at this position */ 7840983a74aSAndrew Boyer done_color = qcq->cq.done_color != (pos < tail); 7850983a74aSAndrew Boyer 7860983a74aSAndrew Boyer /* has the hw indicated the done color at this position? */ 7870983a74aSAndrew Boyer if (color_match(cq_desc->pkt_type_color, done_color)) 7880983a74aSAndrew Boyer return RTE_ETH_RX_DESC_DONE; 7890983a74aSAndrew Boyer 7900983a74aSAndrew Boyer return RTE_ETH_RX_DESC_AVAIL; 7910983a74aSAndrew Boyer } 79260625147SAndrew Boyer 79360625147SAndrew Boyer int 79460625147SAndrew Boyer ionic_dev_tx_descriptor_status(void *tx_queue, uint16_t offset) 79560625147SAndrew Boyer { 79660625147SAndrew Boyer struct ionic_tx_qcq *txq = tx_queue; 79760625147SAndrew Boyer struct ionic_qcq *qcq = &txq->qcq; 798*463ad260SNeel Patel volatile struct ionic_txq_comp *cq_desc; 79960625147SAndrew Boyer uint16_t mask, head, tail, pos, cq_pos; 80060625147SAndrew Boyer bool done_color; 80160625147SAndrew Boyer 80260625147SAndrew Boyer mask = qcq->q.size_mask; 80360625147SAndrew Boyer 80460625147SAndrew Boyer /* offset must be within the size of the ring */ 80560625147SAndrew Boyer if (offset > mask) 80660625147SAndrew Boyer return -EINVAL; 80760625147SAndrew Boyer 80860625147SAndrew Boyer head = qcq->q.head_idx; 80960625147SAndrew Boyer tail = qcq->q.tail_idx; 81060625147SAndrew Boyer 81160625147SAndrew Boyer /* offset is beyond what is posted */ 81260625147SAndrew Boyer if (offset >= ((head - tail) & mask)) 81360625147SAndrew Boyer return RTE_ETH_TX_DESC_DONE; 81460625147SAndrew Boyer 81560625147SAndrew Boyer /* interested in this absolute position in the txq */ 81660625147SAndrew Boyer pos = (tail + offset) & mask; 81760625147SAndrew Boyer 81860625147SAndrew Boyer /* tx cq position != tx q position, need to walk cq */ 81960625147SAndrew Boyer cq_pos = qcq->cq.tail_idx; 82060625147SAndrew Boyer cq_desc = qcq->cq.base; 82160625147SAndrew Boyer cq_desc = &cq_desc[cq_pos]; 82260625147SAndrew Boyer 82360625147SAndrew Boyer /* how far behind is pos from head? */ 82460625147SAndrew Boyer offset = (head - pos) & mask; 82560625147SAndrew Boyer 82660625147SAndrew Boyer /* walk cq descriptors that match the expected done color */ 82760625147SAndrew Boyer done_color = qcq->cq.done_color; 82860625147SAndrew Boyer while (color_match(cq_desc->color, done_color)) { 82960625147SAndrew Boyer /* is comp index no further behind than pos? */ 83060625147SAndrew Boyer tail = rte_cpu_to_le_16(cq_desc->comp_index); 83160625147SAndrew Boyer if (((head - tail) & mask) <= offset) 83260625147SAndrew Boyer return RTE_ETH_TX_DESC_DONE; 83360625147SAndrew Boyer 83460625147SAndrew Boyer cq_pos = (cq_pos + 1) & mask; 83560625147SAndrew Boyer cq_desc = qcq->cq.base; 83660625147SAndrew Boyer cq_desc = &cq_desc[cq_pos]; 83760625147SAndrew Boyer 83860625147SAndrew Boyer done_color = done_color != (cq_pos == 0); 83960625147SAndrew Boyer } 84060625147SAndrew Boyer 84160625147SAndrew Boyer return RTE_ETH_TX_DESC_FULL; 84260625147SAndrew Boyer } 843