1a27d9013SAlfredo Cardigliano /* SPDX-License-Identifier: (BSD-3-Clause OR GPL-2.0) 2a27d9013SAlfredo Cardigliano * Copyright(c) 2018-2019 Pensando Systems, Inc. All rights reserved. 3a27d9013SAlfredo Cardigliano */ 4a27d9013SAlfredo Cardigliano 5a27d9013SAlfredo Cardigliano #include <sys/queue.h> 6a27d9013SAlfredo Cardigliano #include <stdio.h> 7a27d9013SAlfredo Cardigliano #include <stdlib.h> 8a27d9013SAlfredo Cardigliano #include <string.h> 9a27d9013SAlfredo Cardigliano #include <errno.h> 10a27d9013SAlfredo Cardigliano #include <stdint.h> 11a27d9013SAlfredo Cardigliano #include <stdarg.h> 12a27d9013SAlfredo Cardigliano #include <unistd.h> 13a27d9013SAlfredo Cardigliano #include <inttypes.h> 14a27d9013SAlfredo Cardigliano 15a27d9013SAlfredo Cardigliano #include <rte_byteorder.h> 16a27d9013SAlfredo Cardigliano #include <rte_common.h> 17a27d9013SAlfredo Cardigliano #include <rte_cycles.h> 18a27d9013SAlfredo Cardigliano #include <rte_log.h> 19a27d9013SAlfredo Cardigliano #include <rte_debug.h> 20a27d9013SAlfredo Cardigliano #include <rte_interrupts.h> 21a27d9013SAlfredo Cardigliano #include <rte_pci.h> 22a27d9013SAlfredo Cardigliano #include <rte_memory.h> 23a27d9013SAlfredo Cardigliano #include <rte_memzone.h> 24a27d9013SAlfredo Cardigliano #include <rte_launch.h> 25a27d9013SAlfredo Cardigliano #include <rte_eal.h> 26a27d9013SAlfredo Cardigliano #include <rte_per_lcore.h> 27a27d9013SAlfredo Cardigliano #include <rte_lcore.h> 28a27d9013SAlfredo Cardigliano #include <rte_atomic.h> 29a27d9013SAlfredo Cardigliano #include <rte_branch_prediction.h> 30a27d9013SAlfredo Cardigliano #include <rte_mempool.h> 31a27d9013SAlfredo Cardigliano #include <rte_malloc.h> 32a27d9013SAlfredo Cardigliano #include <rte_mbuf.h> 33a27d9013SAlfredo Cardigliano #include <rte_ether.h> 34df96fd0dSBruce Richardson #include <ethdev_driver.h> 35a27d9013SAlfredo Cardigliano #include <rte_prefetch.h> 36a27d9013SAlfredo Cardigliano #include <rte_udp.h> 37a27d9013SAlfredo Cardigliano #include <rte_tcp.h> 38a27d9013SAlfredo Cardigliano #include <rte_sctp.h> 39a27d9013SAlfredo Cardigliano #include <rte_string_fns.h> 40a27d9013SAlfredo Cardigliano #include <rte_errno.h> 41a27d9013SAlfredo Cardigliano #include <rte_ip.h> 42a27d9013SAlfredo Cardigliano #include <rte_net.h> 43a27d9013SAlfredo Cardigliano 44a27d9013SAlfredo Cardigliano #include "ionic_logs.h" 45a27d9013SAlfredo Cardigliano #include "ionic_mac_api.h" 46a27d9013SAlfredo Cardigliano #include "ionic_ethdev.h" 47a27d9013SAlfredo Cardigliano #include "ionic_lif.h" 48a27d9013SAlfredo Cardigliano #include "ionic_rxtx.h" 49a27d9013SAlfredo Cardigliano 50a27d9013SAlfredo Cardigliano #define IONIC_RX_RING_DOORBELL_STRIDE (32 - 1) 51a27d9013SAlfredo Cardigliano 52a27d9013SAlfredo Cardigliano /********************************************************************* 53a27d9013SAlfredo Cardigliano * 54a27d9013SAlfredo Cardigliano * TX functions 55a27d9013SAlfredo Cardigliano * 56a27d9013SAlfredo Cardigliano **********************************************************************/ 57a27d9013SAlfredo Cardigliano 58a27d9013SAlfredo Cardigliano void 59a27d9013SAlfredo Cardigliano ionic_txq_info_get(struct rte_eth_dev *dev, uint16_t queue_id, 60a27d9013SAlfredo Cardigliano struct rte_eth_txq_info *qinfo) 61a27d9013SAlfredo Cardigliano { 62*be39f75cSAndrew Boyer struct ionic_tx_qcq *txq = dev->data->tx_queues[queue_id]; 63*be39f75cSAndrew Boyer struct ionic_queue *q = &txq->qcq.q; 64a27d9013SAlfredo Cardigliano 65a27d9013SAlfredo Cardigliano qinfo->nb_desc = q->num_descs; 6668591087SAndrew Boyer qinfo->conf.offloads = dev->data->dev_conf.txmode.offloads; 6702eabf57SAndrew Boyer qinfo->conf.tx_deferred_start = txq->flags & IONIC_QCQ_F_DEFERRED; 68a27d9013SAlfredo Cardigliano } 69a27d9013SAlfredo Cardigliano 700de3e209SAndrew Boyer static __rte_always_inline void 71*be39f75cSAndrew Boyer ionic_tx_flush(struct ionic_tx_qcq *txq) 72a27d9013SAlfredo Cardigliano { 73*be39f75cSAndrew Boyer struct ionic_cq *cq = &txq->qcq.cq; 74*be39f75cSAndrew Boyer struct ionic_queue *q = &txq->qcq.q; 75a27d9013SAlfredo Cardigliano struct rte_mbuf *txm, *next; 76a27d9013SAlfredo Cardigliano struct ionic_txq_comp *cq_desc_base = cq->base; 77a27d9013SAlfredo Cardigliano struct ionic_txq_comp *cq_desc; 78700f974dSAndrew Boyer void **info; 79a27d9013SAlfredo Cardigliano u_int32_t comp_index = (u_int32_t)-1; 80a27d9013SAlfredo Cardigliano 81a27d9013SAlfredo Cardigliano cq_desc = &cq_desc_base[cq->tail_idx]; 82a27d9013SAlfredo Cardigliano while (color_match(cq_desc->color, cq->done_color)) { 832aed9865SAndrew Boyer cq->tail_idx = Q_NEXT_TO_SRVC(cq, 1); 84a27d9013SAlfredo Cardigliano 85a27d9013SAlfredo Cardigliano /* Prefetch the next 4 descriptors (not really useful here) */ 86a27d9013SAlfredo Cardigliano if ((cq->tail_idx & 0x3) == 0) 87a27d9013SAlfredo Cardigliano rte_prefetch0(&cq_desc_base[cq->tail_idx]); 88a27d9013SAlfredo Cardigliano 89a27d9013SAlfredo Cardigliano if (cq->tail_idx == 0) 90a27d9013SAlfredo Cardigliano cq->done_color = !cq->done_color; 91a27d9013SAlfredo Cardigliano 92a27d9013SAlfredo Cardigliano comp_index = cq_desc->comp_index; 93a27d9013SAlfredo Cardigliano 94a27d9013SAlfredo Cardigliano cq_desc = &cq_desc_base[cq->tail_idx]; 95a27d9013SAlfredo Cardigliano } 96a27d9013SAlfredo Cardigliano 97a27d9013SAlfredo Cardigliano if (comp_index != (u_int32_t)-1) { 98a27d9013SAlfredo Cardigliano while (q->tail_idx != comp_index) { 99700f974dSAndrew Boyer info = IONIC_INFO_PTR(q, q->tail_idx); 100a27d9013SAlfredo Cardigliano 1014ad56b7aSAndrew Boyer q->tail_idx = Q_NEXT_TO_SRVC(q, 1); 102a27d9013SAlfredo Cardigliano 103a27d9013SAlfredo Cardigliano /* Prefetch the next 4 descriptors */ 104a27d9013SAlfredo Cardigliano if ((q->tail_idx & 0x3) == 0) 105a27d9013SAlfredo Cardigliano /* q desc info */ 106a27d9013SAlfredo Cardigliano rte_prefetch0(&q->info[q->tail_idx]); 107a27d9013SAlfredo Cardigliano 108a27d9013SAlfredo Cardigliano /* 109a27d9013SAlfredo Cardigliano * Note: you can just use rte_pktmbuf_free, 110a27d9013SAlfredo Cardigliano * but this loop is faster 111a27d9013SAlfredo Cardigliano */ 112700f974dSAndrew Boyer txm = info[0]; 113a27d9013SAlfredo Cardigliano while (txm != NULL) { 114a27d9013SAlfredo Cardigliano next = txm->next; 115a27d9013SAlfredo Cardigliano rte_pktmbuf_free_seg(txm); 116a27d9013SAlfredo Cardigliano txm = next; 117a27d9013SAlfredo Cardigliano } 118a27d9013SAlfredo Cardigliano } 119a27d9013SAlfredo Cardigliano } 120a27d9013SAlfredo Cardigliano } 121a27d9013SAlfredo Cardigliano 122ce6427ddSThomas Monjalon void __rte_cold 123a27d9013SAlfredo Cardigliano ionic_dev_tx_queue_release(void *tx_queue) 124a27d9013SAlfredo Cardigliano { 125*be39f75cSAndrew Boyer struct ionic_tx_qcq *txq = tx_queue; 126a27d9013SAlfredo Cardigliano 127a27d9013SAlfredo Cardigliano IONIC_PRINT_CALL(); 128a27d9013SAlfredo Cardigliano 1299fdf11c4SAndrew Boyer ionic_lif_txq_deinit(txq); 1309fdf11c4SAndrew Boyer 131*be39f75cSAndrew Boyer ionic_qcq_free(&txq->qcq); 132a27d9013SAlfredo Cardigliano } 133a27d9013SAlfredo Cardigliano 134ce6427ddSThomas Monjalon int __rte_cold 135a27d9013SAlfredo Cardigliano ionic_dev_tx_queue_stop(struct rte_eth_dev *eth_dev, uint16_t tx_queue_id) 136a27d9013SAlfredo Cardigliano { 137*be39f75cSAndrew Boyer struct ionic_tx_qcq *txq; 138a27d9013SAlfredo Cardigliano 1394ae96cb8SAndrew Boyer IONIC_PRINT(DEBUG, "Stopping TX queue %u", tx_queue_id); 140a27d9013SAlfredo Cardigliano 141a27d9013SAlfredo Cardigliano txq = eth_dev->data->tx_queues[tx_queue_id]; 142a27d9013SAlfredo Cardigliano 1439fdf11c4SAndrew Boyer eth_dev->data->tx_queue_state[tx_queue_id] = 1449fdf11c4SAndrew Boyer RTE_ETH_QUEUE_STATE_STOPPED; 1459fdf11c4SAndrew Boyer 146a27d9013SAlfredo Cardigliano /* 147a27d9013SAlfredo Cardigliano * Note: we should better post NOP Tx desc and wait for its completion 148a27d9013SAlfredo Cardigliano * before disabling Tx queue 149a27d9013SAlfredo Cardigliano */ 150a27d9013SAlfredo Cardigliano 151*be39f75cSAndrew Boyer ionic_qcq_disable(&txq->qcq); 152a27d9013SAlfredo Cardigliano 1532aed9865SAndrew Boyer ionic_tx_flush(txq); 154a27d9013SAlfredo Cardigliano 155a27d9013SAlfredo Cardigliano return 0; 156a27d9013SAlfredo Cardigliano } 157a27d9013SAlfredo Cardigliano 158ce6427ddSThomas Monjalon int __rte_cold 159a27d9013SAlfredo Cardigliano ionic_dev_tx_queue_setup(struct rte_eth_dev *eth_dev, uint16_t tx_queue_id, 1604ae96cb8SAndrew Boyer uint16_t nb_desc, uint32_t socket_id, 161a27d9013SAlfredo Cardigliano const struct rte_eth_txconf *tx_conf) 162a27d9013SAlfredo Cardigliano { 163a27d9013SAlfredo Cardigliano struct ionic_lif *lif = IONIC_ETH_DEV_TO_LIF(eth_dev); 164*be39f75cSAndrew Boyer struct ionic_tx_qcq *txq; 165a27d9013SAlfredo Cardigliano uint64_t offloads; 166a27d9013SAlfredo Cardigliano int err; 167a27d9013SAlfredo Cardigliano 168a27d9013SAlfredo Cardigliano if (tx_queue_id >= lif->ntxqcqs) { 169a27d9013SAlfredo Cardigliano IONIC_PRINT(DEBUG, "Queue index %u not available " 170a27d9013SAlfredo Cardigliano "(max %u queues)", 171a27d9013SAlfredo Cardigliano tx_queue_id, lif->ntxqcqs); 172a27d9013SAlfredo Cardigliano return -EINVAL; 173a27d9013SAlfredo Cardigliano } 174a27d9013SAlfredo Cardigliano 175a27d9013SAlfredo Cardigliano offloads = tx_conf->offloads | eth_dev->data->dev_conf.txmode.offloads; 1764ae96cb8SAndrew Boyer IONIC_PRINT(DEBUG, 1774ae96cb8SAndrew Boyer "Configuring skt %u TX queue %u with %u buffers, offloads %jx", 1784ae96cb8SAndrew Boyer socket_id, tx_queue_id, nb_desc, offloads); 179a27d9013SAlfredo Cardigliano 180a27d9013SAlfredo Cardigliano /* Validate number of receive descriptors */ 181a27d9013SAlfredo Cardigliano if (!rte_is_power_of_2(nb_desc) || nb_desc < IONIC_MIN_RING_DESC) 182a27d9013SAlfredo Cardigliano return -EINVAL; /* or use IONIC_DEFAULT_RING_DESC */ 183a27d9013SAlfredo Cardigliano 184a27d9013SAlfredo Cardigliano /* Free memory prior to re-allocation if needed... */ 185a27d9013SAlfredo Cardigliano if (eth_dev->data->tx_queues[tx_queue_id] != NULL) { 186a27d9013SAlfredo Cardigliano void *tx_queue = eth_dev->data->tx_queues[tx_queue_id]; 187a27d9013SAlfredo Cardigliano ionic_dev_tx_queue_release(tx_queue); 188a27d9013SAlfredo Cardigliano eth_dev->data->tx_queues[tx_queue_id] = NULL; 189a27d9013SAlfredo Cardigliano } 190a27d9013SAlfredo Cardigliano 1919fdf11c4SAndrew Boyer eth_dev->data->tx_queue_state[tx_queue_id] = 1929fdf11c4SAndrew Boyer RTE_ETH_QUEUE_STATE_STOPPED; 1939fdf11c4SAndrew Boyer 194a27d9013SAlfredo Cardigliano err = ionic_tx_qcq_alloc(lif, tx_queue_id, nb_desc, &txq); 195a27d9013SAlfredo Cardigliano if (err) { 196a27d9013SAlfredo Cardigliano IONIC_PRINT(DEBUG, "Queue allocation failure"); 197a27d9013SAlfredo Cardigliano return -EINVAL; 198a27d9013SAlfredo Cardigliano } 199a27d9013SAlfredo Cardigliano 200a27d9013SAlfredo Cardigliano /* Do not start queue with rte_eth_dev_start() */ 20102eabf57SAndrew Boyer if (tx_conf->tx_deferred_start) 20202eabf57SAndrew Boyer txq->flags |= IONIC_QCQ_F_DEFERRED; 203a27d9013SAlfredo Cardigliano 20468591087SAndrew Boyer /* Convert the offload flags into queue flags */ 20568591087SAndrew Boyer if (offloads & DEV_TX_OFFLOAD_IPV4_CKSUM) 20668591087SAndrew Boyer txq->flags |= IONIC_QCQ_F_CSUM_L3; 20768591087SAndrew Boyer if (offloads & DEV_TX_OFFLOAD_TCP_CKSUM) 20868591087SAndrew Boyer txq->flags |= IONIC_QCQ_F_CSUM_TCP; 20968591087SAndrew Boyer if (offloads & DEV_TX_OFFLOAD_UDP_CKSUM) 21068591087SAndrew Boyer txq->flags |= IONIC_QCQ_F_CSUM_UDP; 211a27d9013SAlfredo Cardigliano 212a27d9013SAlfredo Cardigliano eth_dev->data->tx_queues[tx_queue_id] = txq; 213a27d9013SAlfredo Cardigliano 214a27d9013SAlfredo Cardigliano return 0; 215a27d9013SAlfredo Cardigliano } 216a27d9013SAlfredo Cardigliano 217a27d9013SAlfredo Cardigliano /* 218a27d9013SAlfredo Cardigliano * Start Transmit Units for specified queue. 219a27d9013SAlfredo Cardigliano */ 220ce6427ddSThomas Monjalon int __rte_cold 221a27d9013SAlfredo Cardigliano ionic_dev_tx_queue_start(struct rte_eth_dev *eth_dev, uint16_t tx_queue_id) 222a27d9013SAlfredo Cardigliano { 2239fdf11c4SAndrew Boyer uint8_t *tx_queue_state = eth_dev->data->tx_queue_state; 224*be39f75cSAndrew Boyer struct ionic_tx_qcq *txq; 225a27d9013SAlfredo Cardigliano int err; 226a27d9013SAlfredo Cardigliano 2279fdf11c4SAndrew Boyer if (tx_queue_state[tx_queue_id] == RTE_ETH_QUEUE_STATE_STARTED) { 2289fdf11c4SAndrew Boyer IONIC_PRINT(DEBUG, "TX queue %u already started", 2299fdf11c4SAndrew Boyer tx_queue_id); 2309fdf11c4SAndrew Boyer return 0; 2319fdf11c4SAndrew Boyer } 2329fdf11c4SAndrew Boyer 233a27d9013SAlfredo Cardigliano txq = eth_dev->data->tx_queues[tx_queue_id]; 234a27d9013SAlfredo Cardigliano 2354ae96cb8SAndrew Boyer IONIC_PRINT(DEBUG, "Starting TX queue %u, %u descs", 236*be39f75cSAndrew Boyer tx_queue_id, txq->qcq.q.num_descs); 2374ae96cb8SAndrew Boyer 2389fdf11c4SAndrew Boyer if (!(txq->flags & IONIC_QCQ_F_INITED)) { 239a27d9013SAlfredo Cardigliano err = ionic_lif_txq_init(txq); 240a27d9013SAlfredo Cardigliano if (err) 241a27d9013SAlfredo Cardigliano return err; 242b5d9a4f0SAndrew Boyer } else { 243*be39f75cSAndrew Boyer ionic_qcq_enable(&txq->qcq); 244b5d9a4f0SAndrew Boyer } 245a27d9013SAlfredo Cardigliano 2469fdf11c4SAndrew Boyer tx_queue_state[tx_queue_id] = RTE_ETH_QUEUE_STATE_STARTED; 247a27d9013SAlfredo Cardigliano 248a27d9013SAlfredo Cardigliano return 0; 249a27d9013SAlfredo Cardigliano } 250a27d9013SAlfredo Cardigliano 251a27d9013SAlfredo Cardigliano static void 25264b08152SAlfredo Cardigliano ionic_tx_tcp_pseudo_csum(struct rte_mbuf *txm) 25364b08152SAlfredo Cardigliano { 25464b08152SAlfredo Cardigliano struct ether_hdr *eth_hdr = rte_pktmbuf_mtod(txm, struct ether_hdr *); 25564b08152SAlfredo Cardigliano char *l3_hdr = ((char *)eth_hdr) + txm->l2_len; 25664b08152SAlfredo Cardigliano struct rte_tcp_hdr *tcp_hdr = (struct rte_tcp_hdr *) 25764b08152SAlfredo Cardigliano (l3_hdr + txm->l3_len); 25864b08152SAlfredo Cardigliano 25964b08152SAlfredo Cardigliano if (txm->ol_flags & PKT_TX_IP_CKSUM) { 26064b08152SAlfredo Cardigliano struct rte_ipv4_hdr *ipv4_hdr = (struct rte_ipv4_hdr *)l3_hdr; 26164b08152SAlfredo Cardigliano ipv4_hdr->hdr_checksum = 0; 26264b08152SAlfredo Cardigliano tcp_hdr->cksum = 0; 26364b08152SAlfredo Cardigliano tcp_hdr->cksum = rte_ipv4_udptcp_cksum(ipv4_hdr, tcp_hdr); 26464b08152SAlfredo Cardigliano } else { 26564b08152SAlfredo Cardigliano struct rte_ipv6_hdr *ipv6_hdr = (struct rte_ipv6_hdr *)l3_hdr; 26664b08152SAlfredo Cardigliano tcp_hdr->cksum = 0; 26764b08152SAlfredo Cardigliano tcp_hdr->cksum = rte_ipv6_udptcp_cksum(ipv6_hdr, tcp_hdr); 26864b08152SAlfredo Cardigliano } 26964b08152SAlfredo Cardigliano } 27064b08152SAlfredo Cardigliano 27164b08152SAlfredo Cardigliano static void 27264b08152SAlfredo Cardigliano ionic_tx_tcp_inner_pseudo_csum(struct rte_mbuf *txm) 27364b08152SAlfredo Cardigliano { 27464b08152SAlfredo Cardigliano struct ether_hdr *eth_hdr = rte_pktmbuf_mtod(txm, struct ether_hdr *); 27564b08152SAlfredo Cardigliano char *l3_hdr = ((char *)eth_hdr) + txm->outer_l2_len + 27664b08152SAlfredo Cardigliano txm->outer_l3_len + txm->l2_len; 27764b08152SAlfredo Cardigliano struct rte_tcp_hdr *tcp_hdr = (struct rte_tcp_hdr *) 27864b08152SAlfredo Cardigliano (l3_hdr + txm->l3_len); 27964b08152SAlfredo Cardigliano 28064b08152SAlfredo Cardigliano if (txm->ol_flags & PKT_TX_IPV4) { 28164b08152SAlfredo Cardigliano struct rte_ipv4_hdr *ipv4_hdr = (struct rte_ipv4_hdr *)l3_hdr; 28264b08152SAlfredo Cardigliano ipv4_hdr->hdr_checksum = 0; 28364b08152SAlfredo Cardigliano tcp_hdr->cksum = 0; 28464b08152SAlfredo Cardigliano tcp_hdr->cksum = rte_ipv4_udptcp_cksum(ipv4_hdr, tcp_hdr); 28564b08152SAlfredo Cardigliano } else { 28664b08152SAlfredo Cardigliano struct rte_ipv6_hdr *ipv6_hdr = (struct rte_ipv6_hdr *)l3_hdr; 28764b08152SAlfredo Cardigliano tcp_hdr->cksum = 0; 28864b08152SAlfredo Cardigliano tcp_hdr->cksum = rte_ipv6_udptcp_cksum(ipv6_hdr, tcp_hdr); 28964b08152SAlfredo Cardigliano } 29064b08152SAlfredo Cardigliano } 29164b08152SAlfredo Cardigliano 29264b08152SAlfredo Cardigliano static void 293a27d9013SAlfredo Cardigliano ionic_tx_tso_post(struct ionic_queue *q, struct ionic_txq_desc *desc, 294a27d9013SAlfredo Cardigliano struct rte_mbuf *txm, 295a27d9013SAlfredo Cardigliano rte_iova_t addr, uint8_t nsge, uint16_t len, 296a27d9013SAlfredo Cardigliano uint32_t hdrlen, uint32_t mss, 29764b08152SAlfredo Cardigliano bool encap, 298a27d9013SAlfredo Cardigliano uint16_t vlan_tci, bool has_vlan, 299a27d9013SAlfredo Cardigliano bool start, bool done) 300a27d9013SAlfredo Cardigliano { 301a27d9013SAlfredo Cardigliano uint8_t flags = 0; 302a27d9013SAlfredo Cardigliano flags |= has_vlan ? IONIC_TXQ_DESC_FLAG_VLAN : 0; 30364b08152SAlfredo Cardigliano flags |= encap ? IONIC_TXQ_DESC_FLAG_ENCAP : 0; 304a27d9013SAlfredo Cardigliano flags |= start ? IONIC_TXQ_DESC_FLAG_TSO_SOT : 0; 305a27d9013SAlfredo Cardigliano flags |= done ? IONIC_TXQ_DESC_FLAG_TSO_EOT : 0; 306a27d9013SAlfredo Cardigliano 307a27d9013SAlfredo Cardigliano desc->cmd = encode_txq_desc_cmd(IONIC_TXQ_DESC_OPCODE_TSO, 308a27d9013SAlfredo Cardigliano flags, nsge, addr); 309a27d9013SAlfredo Cardigliano desc->len = len; 310a27d9013SAlfredo Cardigliano desc->vlan_tci = vlan_tci; 311a27d9013SAlfredo Cardigliano desc->hdr_len = hdrlen; 312a27d9013SAlfredo Cardigliano desc->mss = mss; 313a27d9013SAlfredo Cardigliano 314700f974dSAndrew Boyer ionic_q_post(q, done, done ? txm : NULL); 315a27d9013SAlfredo Cardigliano } 316a27d9013SAlfredo Cardigliano 317a27d9013SAlfredo Cardigliano static struct ionic_txq_desc * 318*be39f75cSAndrew Boyer ionic_tx_tso_next(struct ionic_tx_qcq *txq, struct ionic_txq_sg_elem **elem) 319a27d9013SAlfredo Cardigliano { 320*be39f75cSAndrew Boyer struct ionic_queue *q = &txq->qcq.q; 321a27d9013SAlfredo Cardigliano struct ionic_txq_desc *desc_base = q->base; 32256117636SAndrew Boyer struct ionic_txq_sg_desc_v1 *sg_desc_base = q->sg_base; 323a27d9013SAlfredo Cardigliano struct ionic_txq_desc *desc = &desc_base[q->head_idx]; 32456117636SAndrew Boyer struct ionic_txq_sg_desc_v1 *sg_desc = &sg_desc_base[q->head_idx]; 325a27d9013SAlfredo Cardigliano 326a27d9013SAlfredo Cardigliano *elem = sg_desc->elems; 327a27d9013SAlfredo Cardigliano return desc; 328a27d9013SAlfredo Cardigliano } 329a27d9013SAlfredo Cardigliano 330a27d9013SAlfredo Cardigliano static int 331*be39f75cSAndrew Boyer ionic_tx_tso(struct ionic_tx_qcq *txq, struct rte_mbuf *txm, 33268591087SAndrew Boyer bool not_xmit_more) 333a27d9013SAlfredo Cardigliano { 334*be39f75cSAndrew Boyer struct ionic_queue *q = &txq->qcq.q; 335*be39f75cSAndrew Boyer struct ionic_tx_stats *stats = &txq->stats; 336a27d9013SAlfredo Cardigliano struct ionic_txq_desc *desc; 337a27d9013SAlfredo Cardigliano struct ionic_txq_sg_elem *elem; 338a27d9013SAlfredo Cardigliano struct rte_mbuf *txm_seg; 3397c3a867bSAndrew Boyer rte_iova_t data_iova; 3407c3a867bSAndrew Boyer uint64_t desc_addr = 0, next_addr; 341a27d9013SAlfredo Cardigliano uint16_t desc_len = 0; 342a27d9013SAlfredo Cardigliano uint8_t desc_nsge; 343a27d9013SAlfredo Cardigliano uint32_t hdrlen; 344a27d9013SAlfredo Cardigliano uint32_t mss = txm->tso_segsz; 345a27d9013SAlfredo Cardigliano uint32_t frag_left = 0; 346a27d9013SAlfredo Cardigliano uint32_t left; 347a27d9013SAlfredo Cardigliano uint32_t seglen; 348a27d9013SAlfredo Cardigliano uint32_t len; 349a27d9013SAlfredo Cardigliano uint32_t offset = 0; 350a27d9013SAlfredo Cardigliano bool start, done; 35164b08152SAlfredo Cardigliano bool encap; 352a27d9013SAlfredo Cardigliano bool has_vlan = !!(txm->ol_flags & PKT_TX_VLAN_PKT); 353a27d9013SAlfredo Cardigliano uint16_t vlan_tci = txm->vlan_tci; 35464b08152SAlfredo Cardigliano uint64_t ol_flags = txm->ol_flags; 355a27d9013SAlfredo Cardigliano 35664b08152SAlfredo Cardigliano encap = ((ol_flags & PKT_TX_OUTER_IP_CKSUM) || 35764b08152SAlfredo Cardigliano (ol_flags & PKT_TX_OUTER_UDP_CKSUM)) && 35864b08152SAlfredo Cardigliano ((ol_flags & PKT_TX_OUTER_IPV4) || 35964b08152SAlfredo Cardigliano (ol_flags & PKT_TX_OUTER_IPV6)); 36064b08152SAlfredo Cardigliano 36164b08152SAlfredo Cardigliano /* Preload inner-most TCP csum field with IP pseudo hdr 36264b08152SAlfredo Cardigliano * calculated with IP length set to zero. HW will later 36364b08152SAlfredo Cardigliano * add in length to each TCP segment resulting from the TSO. 36464b08152SAlfredo Cardigliano */ 36564b08152SAlfredo Cardigliano 36664b08152SAlfredo Cardigliano if (encap) { 36764b08152SAlfredo Cardigliano ionic_tx_tcp_inner_pseudo_csum(txm); 36864b08152SAlfredo Cardigliano hdrlen = txm->outer_l2_len + txm->outer_l3_len + 36964b08152SAlfredo Cardigliano txm->l2_len + txm->l3_len + txm->l4_len; 37064b08152SAlfredo Cardigliano } else { 37164b08152SAlfredo Cardigliano ionic_tx_tcp_pseudo_csum(txm); 37264b08152SAlfredo Cardigliano hdrlen = txm->l2_len + txm->l3_len + txm->l4_len; 37364b08152SAlfredo Cardigliano } 374a27d9013SAlfredo Cardigliano 375a27d9013SAlfredo Cardigliano seglen = hdrlen + mss; 376a27d9013SAlfredo Cardigliano left = txm->data_len; 3777c3a867bSAndrew Boyer data_iova = rte_mbuf_data_iova(txm); 378a27d9013SAlfredo Cardigliano 379*be39f75cSAndrew Boyer desc = ionic_tx_tso_next(txq, &elem); 380a27d9013SAlfredo Cardigliano start = true; 381a27d9013SAlfredo Cardigliano 382a27d9013SAlfredo Cardigliano /* Chop data up into desc segments */ 383a27d9013SAlfredo Cardigliano 384a27d9013SAlfredo Cardigliano while (left > 0) { 385a27d9013SAlfredo Cardigliano len = RTE_MIN(seglen, left); 386a27d9013SAlfredo Cardigliano frag_left = seglen - len; 3877c3a867bSAndrew Boyer desc_addr = rte_cpu_to_le_64(data_iova + offset); 388a27d9013SAlfredo Cardigliano desc_len = len; 389a27d9013SAlfredo Cardigliano desc_nsge = 0; 390a27d9013SAlfredo Cardigliano left -= len; 391a27d9013SAlfredo Cardigliano offset += len; 392a27d9013SAlfredo Cardigliano if (txm->nb_segs > 1 && frag_left > 0) 393a27d9013SAlfredo Cardigliano continue; 394a27d9013SAlfredo Cardigliano done = (txm->nb_segs == 1 && left == 0); 395a27d9013SAlfredo Cardigliano ionic_tx_tso_post(q, desc, txm, 396a27d9013SAlfredo Cardigliano desc_addr, desc_nsge, desc_len, 397a27d9013SAlfredo Cardigliano hdrlen, mss, 39864b08152SAlfredo Cardigliano encap, 399a27d9013SAlfredo Cardigliano vlan_tci, has_vlan, 400a27d9013SAlfredo Cardigliano start, done && not_xmit_more); 401*be39f75cSAndrew Boyer desc = ionic_tx_tso_next(txq, &elem); 402a27d9013SAlfredo Cardigliano start = false; 403a27d9013SAlfredo Cardigliano seglen = mss; 404a27d9013SAlfredo Cardigliano } 405a27d9013SAlfredo Cardigliano 406a27d9013SAlfredo Cardigliano /* Chop frags into desc segments */ 407a27d9013SAlfredo Cardigliano 408a27d9013SAlfredo Cardigliano txm_seg = txm->next; 409a27d9013SAlfredo Cardigliano while (txm_seg != NULL) { 410a27d9013SAlfredo Cardigliano offset = 0; 4117c3a867bSAndrew Boyer data_iova = rte_mbuf_data_iova(txm_seg); 412a27d9013SAlfredo Cardigliano left = txm_seg->data_len; 413a27d9013SAlfredo Cardigliano stats->frags++; 414a27d9013SAlfredo Cardigliano 415a27d9013SAlfredo Cardigliano while (left > 0) { 4167c3a867bSAndrew Boyer next_addr = rte_cpu_to_le_64(data_iova + offset); 417a27d9013SAlfredo Cardigliano if (frag_left > 0) { 418a27d9013SAlfredo Cardigliano len = RTE_MIN(frag_left, left); 419a27d9013SAlfredo Cardigliano frag_left -= len; 4207c3a867bSAndrew Boyer elem->addr = next_addr; 421a27d9013SAlfredo Cardigliano elem->len = len; 422a27d9013SAlfredo Cardigliano elem++; 423a27d9013SAlfredo Cardigliano desc_nsge++; 424a27d9013SAlfredo Cardigliano } else { 425a27d9013SAlfredo Cardigliano len = RTE_MIN(mss, left); 426a27d9013SAlfredo Cardigliano frag_left = mss - len; 4277c3a867bSAndrew Boyer desc_addr = next_addr; 428a27d9013SAlfredo Cardigliano desc_len = len; 429a27d9013SAlfredo Cardigliano desc_nsge = 0; 430a27d9013SAlfredo Cardigliano } 431a27d9013SAlfredo Cardigliano left -= len; 432a27d9013SAlfredo Cardigliano offset += len; 433a27d9013SAlfredo Cardigliano if (txm_seg->next != NULL && frag_left > 0) 434a27d9013SAlfredo Cardigliano continue; 4357c3a867bSAndrew Boyer 436a27d9013SAlfredo Cardigliano done = (txm_seg->next == NULL && left == 0); 437a27d9013SAlfredo Cardigliano ionic_tx_tso_post(q, desc, txm_seg, 438a27d9013SAlfredo Cardigliano desc_addr, desc_nsge, desc_len, 439a27d9013SAlfredo Cardigliano hdrlen, mss, 44064b08152SAlfredo Cardigliano encap, 441a27d9013SAlfredo Cardigliano vlan_tci, has_vlan, 442a27d9013SAlfredo Cardigliano start, done && not_xmit_more); 443*be39f75cSAndrew Boyer desc = ionic_tx_tso_next(txq, &elem); 444a27d9013SAlfredo Cardigliano start = false; 445a27d9013SAlfredo Cardigliano } 446a27d9013SAlfredo Cardigliano 447a27d9013SAlfredo Cardigliano txm_seg = txm_seg->next; 448a27d9013SAlfredo Cardigliano } 449a27d9013SAlfredo Cardigliano 450a27d9013SAlfredo Cardigliano stats->tso++; 451a27d9013SAlfredo Cardigliano 452a27d9013SAlfredo Cardigliano return 0; 453a27d9013SAlfredo Cardigliano } 454a27d9013SAlfredo Cardigliano 4550de3e209SAndrew Boyer static __rte_always_inline int 456*be39f75cSAndrew Boyer ionic_tx(struct ionic_tx_qcq *txq, struct rte_mbuf *txm, 45768591087SAndrew Boyer bool not_xmit_more) 458a27d9013SAlfredo Cardigliano { 459*be39f75cSAndrew Boyer struct ionic_queue *q = &txq->qcq.q; 460*be39f75cSAndrew Boyer struct ionic_txq_desc *desc, *desc_base = q->base; 46156117636SAndrew Boyer struct ionic_txq_sg_desc_v1 *sg_desc_base = q->sg_base; 462*be39f75cSAndrew Boyer struct ionic_txq_sg_elem *elem; 463*be39f75cSAndrew Boyer struct ionic_tx_stats *stats = &txq->stats; 464a27d9013SAlfredo Cardigliano struct rte_mbuf *txm_seg; 46564b08152SAlfredo Cardigliano bool encap; 466a27d9013SAlfredo Cardigliano bool has_vlan; 467a27d9013SAlfredo Cardigliano uint64_t ol_flags = txm->ol_flags; 4687c3a867bSAndrew Boyer uint64_t addr; 469a27d9013SAlfredo Cardigliano uint8_t opcode = IONIC_TXQ_DESC_OPCODE_CSUM_NONE; 470a27d9013SAlfredo Cardigliano uint8_t flags = 0; 471a27d9013SAlfredo Cardigliano 472*be39f75cSAndrew Boyer desc = &desc_base[q->head_idx]; 473*be39f75cSAndrew Boyer 47464b08152SAlfredo Cardigliano if ((ol_flags & PKT_TX_IP_CKSUM) && 47568591087SAndrew Boyer (txq->flags & IONIC_QCQ_F_CSUM_L3)) { 47664b08152SAlfredo Cardigliano opcode = IONIC_TXQ_DESC_OPCODE_CSUM_HW; 47764b08152SAlfredo Cardigliano flags |= IONIC_TXQ_DESC_FLAG_CSUM_L3; 478f603eebcSAndrew Boyer } 479f603eebcSAndrew Boyer 48064b08152SAlfredo Cardigliano if (((ol_flags & PKT_TX_TCP_CKSUM) && 48168591087SAndrew Boyer (txq->flags & IONIC_QCQ_F_CSUM_TCP)) || 48264b08152SAlfredo Cardigliano ((ol_flags & PKT_TX_UDP_CKSUM) && 48368591087SAndrew Boyer (txq->flags & IONIC_QCQ_F_CSUM_UDP))) { 484f603eebcSAndrew Boyer opcode = IONIC_TXQ_DESC_OPCODE_CSUM_HW; 48564b08152SAlfredo Cardigliano flags |= IONIC_TXQ_DESC_FLAG_CSUM_L4; 48664b08152SAlfredo Cardigliano } 48764b08152SAlfredo Cardigliano 488f603eebcSAndrew Boyer if (opcode == IONIC_TXQ_DESC_OPCODE_CSUM_NONE) 489f603eebcSAndrew Boyer stats->no_csum++; 490f603eebcSAndrew Boyer 491a27d9013SAlfredo Cardigliano has_vlan = (ol_flags & PKT_TX_VLAN_PKT); 49264b08152SAlfredo Cardigliano encap = ((ol_flags & PKT_TX_OUTER_IP_CKSUM) || 49364b08152SAlfredo Cardigliano (ol_flags & PKT_TX_OUTER_UDP_CKSUM)) && 49464b08152SAlfredo Cardigliano ((ol_flags & PKT_TX_OUTER_IPV4) || 49564b08152SAlfredo Cardigliano (ol_flags & PKT_TX_OUTER_IPV6)); 496a27d9013SAlfredo Cardigliano 497a27d9013SAlfredo Cardigliano flags |= has_vlan ? IONIC_TXQ_DESC_FLAG_VLAN : 0; 49864b08152SAlfredo Cardigliano flags |= encap ? IONIC_TXQ_DESC_FLAG_ENCAP : 0; 499a27d9013SAlfredo Cardigliano 5007c3a867bSAndrew Boyer addr = rte_cpu_to_le_64(rte_mbuf_data_iova(txm)); 5017c3a867bSAndrew Boyer 502a27d9013SAlfredo Cardigliano desc->cmd = encode_txq_desc_cmd(opcode, flags, txm->nb_segs - 1, addr); 503a27d9013SAlfredo Cardigliano desc->len = txm->data_len; 504a27d9013SAlfredo Cardigliano desc->vlan_tci = txm->vlan_tci; 505a27d9013SAlfredo Cardigliano 506*be39f75cSAndrew Boyer elem = sg_desc_base[q->head_idx].elems; 507a27d9013SAlfredo Cardigliano txm_seg = txm->next; 508a27d9013SAlfredo Cardigliano while (txm_seg != NULL) { 509a27d9013SAlfredo Cardigliano elem->len = txm_seg->data_len; 510a27d9013SAlfredo Cardigliano elem->addr = rte_cpu_to_le_64(rte_mbuf_data_iova(txm_seg)); 511a27d9013SAlfredo Cardigliano stats->frags++; 512a27d9013SAlfredo Cardigliano elem++; 513a27d9013SAlfredo Cardigliano txm_seg = txm_seg->next; 514a27d9013SAlfredo Cardigliano } 515a27d9013SAlfredo Cardigliano 516700f974dSAndrew Boyer ionic_q_post(q, not_xmit_more, txm); 517a27d9013SAlfredo Cardigliano 518a27d9013SAlfredo Cardigliano return 0; 519a27d9013SAlfredo Cardigliano } 520a27d9013SAlfredo Cardigliano 521a27d9013SAlfredo Cardigliano uint16_t 522a27d9013SAlfredo Cardigliano ionic_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, 523a27d9013SAlfredo Cardigliano uint16_t nb_pkts) 524a27d9013SAlfredo Cardigliano { 525*be39f75cSAndrew Boyer struct ionic_tx_qcq *txq = tx_queue; 526*be39f75cSAndrew Boyer struct ionic_queue *q = &txq->qcq.q; 527*be39f75cSAndrew Boyer struct ionic_tx_stats *stats = &txq->stats; 528a27d9013SAlfredo Cardigliano uint32_t next_q_head_idx; 529a27d9013SAlfredo Cardigliano uint32_t bytes_tx = 0; 530a27d9013SAlfredo Cardigliano uint16_t nb_tx = 0; 531a27d9013SAlfredo Cardigliano int err; 532a27d9013SAlfredo Cardigliano bool last; 533a27d9013SAlfredo Cardigliano 534a27d9013SAlfredo Cardigliano /* Cleaning old buffers */ 5352aed9865SAndrew Boyer ionic_tx_flush(txq); 536a27d9013SAlfredo Cardigliano 537a27d9013SAlfredo Cardigliano if (unlikely(ionic_q_space_avail(q) < nb_pkts)) { 538a27d9013SAlfredo Cardigliano stats->stop += nb_pkts; 539a27d9013SAlfredo Cardigliano return 0; 540a27d9013SAlfredo Cardigliano } 541a27d9013SAlfredo Cardigliano 542a27d9013SAlfredo Cardigliano while (nb_tx < nb_pkts) { 543a27d9013SAlfredo Cardigliano last = (nb_tx == (nb_pkts - 1)); 544a27d9013SAlfredo Cardigliano 5454ad56b7aSAndrew Boyer next_q_head_idx = Q_NEXT_TO_POST(q, 1); 546a27d9013SAlfredo Cardigliano if ((next_q_head_idx & 0x3) == 0) { 547a27d9013SAlfredo Cardigliano struct ionic_txq_desc *desc_base = q->base; 548a27d9013SAlfredo Cardigliano rte_prefetch0(&desc_base[next_q_head_idx]); 549a27d9013SAlfredo Cardigliano rte_prefetch0(&q->info[next_q_head_idx]); 550a27d9013SAlfredo Cardigliano } 551a27d9013SAlfredo Cardigliano 552a27d9013SAlfredo Cardigliano if (tx_pkts[nb_tx]->ol_flags & PKT_TX_TCP_SEG) 55368591087SAndrew Boyer err = ionic_tx_tso(txq, tx_pkts[nb_tx], last); 554a27d9013SAlfredo Cardigliano else 55568591087SAndrew Boyer err = ionic_tx(txq, tx_pkts[nb_tx], last); 556a27d9013SAlfredo Cardigliano if (err) { 557a27d9013SAlfredo Cardigliano stats->drop += nb_pkts - nb_tx; 558a27d9013SAlfredo Cardigliano if (nb_tx > 0) 559a27d9013SAlfredo Cardigliano ionic_q_flush(q); 560a27d9013SAlfredo Cardigliano break; 561a27d9013SAlfredo Cardigliano } 562a27d9013SAlfredo Cardigliano 563a27d9013SAlfredo Cardigliano bytes_tx += tx_pkts[nb_tx]->pkt_len; 564a27d9013SAlfredo Cardigliano nb_tx++; 565a27d9013SAlfredo Cardigliano } 566a27d9013SAlfredo Cardigliano 567a27d9013SAlfredo Cardigliano stats->packets += nb_tx; 568a27d9013SAlfredo Cardigliano stats->bytes += bytes_tx; 569a27d9013SAlfredo Cardigliano 570a27d9013SAlfredo Cardigliano return nb_tx; 571a27d9013SAlfredo Cardigliano } 572a27d9013SAlfredo Cardigliano 573a27d9013SAlfredo Cardigliano /********************************************************************* 574a27d9013SAlfredo Cardigliano * 575a27d9013SAlfredo Cardigliano * TX prep functions 576a27d9013SAlfredo Cardigliano * 577a27d9013SAlfredo Cardigliano **********************************************************************/ 578a27d9013SAlfredo Cardigliano 579a27d9013SAlfredo Cardigliano #define IONIC_TX_OFFLOAD_MASK ( \ 580a27d9013SAlfredo Cardigliano PKT_TX_IPV4 | \ 581a27d9013SAlfredo Cardigliano PKT_TX_IPV6 | \ 582a27d9013SAlfredo Cardigliano PKT_TX_VLAN | \ 58364b08152SAlfredo Cardigliano PKT_TX_IP_CKSUM | \ 584a27d9013SAlfredo Cardigliano PKT_TX_TCP_SEG | \ 585a27d9013SAlfredo Cardigliano PKT_TX_L4_MASK) 586a27d9013SAlfredo Cardigliano 587a27d9013SAlfredo Cardigliano #define IONIC_TX_OFFLOAD_NOTSUP_MASK \ 588a27d9013SAlfredo Cardigliano (PKT_TX_OFFLOAD_MASK ^ IONIC_TX_OFFLOAD_MASK) 589a27d9013SAlfredo Cardigliano 590a27d9013SAlfredo Cardigliano uint16_t 591a27d9013SAlfredo Cardigliano ionic_prep_pkts(void *tx_queue __rte_unused, struct rte_mbuf **tx_pkts, 592a27d9013SAlfredo Cardigliano uint16_t nb_pkts) 593a27d9013SAlfredo Cardigliano { 594a27d9013SAlfredo Cardigliano struct rte_mbuf *txm; 595a27d9013SAlfredo Cardigliano uint64_t offloads; 596a27d9013SAlfredo Cardigliano int i = 0; 597a27d9013SAlfredo Cardigliano 598a27d9013SAlfredo Cardigliano for (i = 0; i < nb_pkts; i++) { 599a27d9013SAlfredo Cardigliano txm = tx_pkts[i]; 600a27d9013SAlfredo Cardigliano 601d13d7829SAndrew Boyer if (txm->nb_segs > IONIC_TX_MAX_SG_ELEMS_V1 + 1) { 602a27d9013SAlfredo Cardigliano rte_errno = -EINVAL; 603a27d9013SAlfredo Cardigliano break; 604a27d9013SAlfredo Cardigliano } 605a27d9013SAlfredo Cardigliano 606a27d9013SAlfredo Cardigliano offloads = txm->ol_flags; 607a27d9013SAlfredo Cardigliano 608a27d9013SAlfredo Cardigliano if (offloads & IONIC_TX_OFFLOAD_NOTSUP_MASK) { 609a27d9013SAlfredo Cardigliano rte_errno = -ENOTSUP; 610a27d9013SAlfredo Cardigliano break; 611a27d9013SAlfredo Cardigliano } 612a27d9013SAlfredo Cardigliano } 613a27d9013SAlfredo Cardigliano 614a27d9013SAlfredo Cardigliano return i; 615a27d9013SAlfredo Cardigliano } 616a27d9013SAlfredo Cardigliano 617a27d9013SAlfredo Cardigliano /********************************************************************* 618a27d9013SAlfredo Cardigliano * 619a27d9013SAlfredo Cardigliano * RX functions 620a27d9013SAlfredo Cardigliano * 621a27d9013SAlfredo Cardigliano **********************************************************************/ 622a27d9013SAlfredo Cardigliano 623a27d9013SAlfredo Cardigliano static void ionic_rx_recycle(struct ionic_queue *q, uint32_t q_desc_index, 624a27d9013SAlfredo Cardigliano struct rte_mbuf *mbuf); 625a27d9013SAlfredo Cardigliano 626a27d9013SAlfredo Cardigliano void 627a27d9013SAlfredo Cardigliano ionic_rxq_info_get(struct rte_eth_dev *dev, uint16_t queue_id, 628a27d9013SAlfredo Cardigliano struct rte_eth_rxq_info *qinfo) 629a27d9013SAlfredo Cardigliano { 630*be39f75cSAndrew Boyer struct ionic_rx_qcq *rxq = dev->data->rx_queues[queue_id]; 631*be39f75cSAndrew Boyer struct ionic_queue *q = &rxq->qcq.q; 632a27d9013SAlfredo Cardigliano 633a27d9013SAlfredo Cardigliano qinfo->mp = rxq->mb_pool; 634a27d9013SAlfredo Cardigliano qinfo->scattered_rx = dev->data->scattered_rx; 635a27d9013SAlfredo Cardigliano qinfo->nb_desc = q->num_descs; 63602eabf57SAndrew Boyer qinfo->conf.rx_deferred_start = rxq->flags & IONIC_QCQ_F_DEFERRED; 63768591087SAndrew Boyer qinfo->conf.offloads = dev->data->dev_conf.rxmode.offloads; 638a27d9013SAlfredo Cardigliano } 639a27d9013SAlfredo Cardigliano 640ce6427ddSThomas Monjalon static void __rte_cold 641*be39f75cSAndrew Boyer ionic_rx_empty(struct ionic_rx_qcq *rxq) 642a27d9013SAlfredo Cardigliano { 643*be39f75cSAndrew Boyer struct ionic_queue *q = &rxq->qcq.q; 644a27d9013SAlfredo Cardigliano struct rte_mbuf *mbuf; 645700f974dSAndrew Boyer void **info; 646a27d9013SAlfredo Cardigliano 647a27d9013SAlfredo Cardigliano while (q->tail_idx != q->head_idx) { 648700f974dSAndrew Boyer info = IONIC_INFO_PTR(q, q->tail_idx); 649700f974dSAndrew Boyer mbuf = info[0]; 650a27d9013SAlfredo Cardigliano rte_mempool_put(rxq->mb_pool, mbuf); 651a27d9013SAlfredo Cardigliano 6524ad56b7aSAndrew Boyer q->tail_idx = Q_NEXT_TO_SRVC(q, 1); 653a27d9013SAlfredo Cardigliano } 654a27d9013SAlfredo Cardigliano } 655a27d9013SAlfredo Cardigliano 656ce6427ddSThomas Monjalon void __rte_cold 657a27d9013SAlfredo Cardigliano ionic_dev_rx_queue_release(void *rx_queue) 658a27d9013SAlfredo Cardigliano { 659*be39f75cSAndrew Boyer struct ionic_rx_qcq *rxq = rx_queue; 660*be39f75cSAndrew Boyer 661*be39f75cSAndrew Boyer if (!rxq) 662*be39f75cSAndrew Boyer return; 663a27d9013SAlfredo Cardigliano 664a27d9013SAlfredo Cardigliano IONIC_PRINT_CALL(); 665a27d9013SAlfredo Cardigliano 666*be39f75cSAndrew Boyer ionic_rx_empty(rxq); 667a27d9013SAlfredo Cardigliano 6689fdf11c4SAndrew Boyer ionic_lif_rxq_deinit(rxq); 6699fdf11c4SAndrew Boyer 670*be39f75cSAndrew Boyer ionic_qcq_free(&rxq->qcq); 671a27d9013SAlfredo Cardigliano } 672a27d9013SAlfredo Cardigliano 673ce6427ddSThomas Monjalon int __rte_cold 674a27d9013SAlfredo Cardigliano ionic_dev_rx_queue_setup(struct rte_eth_dev *eth_dev, 675a27d9013SAlfredo Cardigliano uint16_t rx_queue_id, 676a27d9013SAlfredo Cardigliano uint16_t nb_desc, 6774ae96cb8SAndrew Boyer uint32_t socket_id, 678a27d9013SAlfredo Cardigliano const struct rte_eth_rxconf *rx_conf, 679a27d9013SAlfredo Cardigliano struct rte_mempool *mp) 680a27d9013SAlfredo Cardigliano { 681a27d9013SAlfredo Cardigliano struct ionic_lif *lif = IONIC_ETH_DEV_TO_LIF(eth_dev); 682*be39f75cSAndrew Boyer struct ionic_rx_qcq *rxq; 683a27d9013SAlfredo Cardigliano uint64_t offloads; 684a27d9013SAlfredo Cardigliano int err; 685a27d9013SAlfredo Cardigliano 686a27d9013SAlfredo Cardigliano if (rx_queue_id >= lif->nrxqcqs) { 687a27d9013SAlfredo Cardigliano IONIC_PRINT(ERR, 688a27d9013SAlfredo Cardigliano "Queue index %u not available (max %u queues)", 689a27d9013SAlfredo Cardigliano rx_queue_id, lif->nrxqcqs); 690a27d9013SAlfredo Cardigliano return -EINVAL; 691a27d9013SAlfredo Cardigliano } 692a27d9013SAlfredo Cardigliano 693a27d9013SAlfredo Cardigliano offloads = rx_conf->offloads | eth_dev->data->dev_conf.rxmode.offloads; 6944ae96cb8SAndrew Boyer IONIC_PRINT(DEBUG, 6954ae96cb8SAndrew Boyer "Configuring skt %u RX queue %u with %u buffers, offloads %jx", 6964ae96cb8SAndrew Boyer socket_id, rx_queue_id, nb_desc, offloads); 697a27d9013SAlfredo Cardigliano 69818a44465SAndrew Boyer if (!rx_conf->rx_drop_en) 69918a44465SAndrew Boyer IONIC_PRINT(WARNING, "No-drop mode is not supported"); 70018a44465SAndrew Boyer 701a27d9013SAlfredo Cardigliano /* Validate number of receive descriptors */ 702a27d9013SAlfredo Cardigliano if (!rte_is_power_of_2(nb_desc) || 703a27d9013SAlfredo Cardigliano nb_desc < IONIC_MIN_RING_DESC || 704a27d9013SAlfredo Cardigliano nb_desc > IONIC_MAX_RING_DESC) { 705a27d9013SAlfredo Cardigliano IONIC_PRINT(ERR, 7064ae96cb8SAndrew Boyer "Bad descriptor count (%u) for queue %u (min: %u)", 707a27d9013SAlfredo Cardigliano nb_desc, rx_queue_id, IONIC_MIN_RING_DESC); 708a27d9013SAlfredo Cardigliano return -EINVAL; /* or use IONIC_DEFAULT_RING_DESC */ 709a27d9013SAlfredo Cardigliano } 710a27d9013SAlfredo Cardigliano 711a27d9013SAlfredo Cardigliano /* Free memory prior to re-allocation if needed... */ 712a27d9013SAlfredo Cardigliano if (eth_dev->data->rx_queues[rx_queue_id] != NULL) { 713a27d9013SAlfredo Cardigliano void *rx_queue = eth_dev->data->rx_queues[rx_queue_id]; 714a27d9013SAlfredo Cardigliano ionic_dev_rx_queue_release(rx_queue); 715a27d9013SAlfredo Cardigliano eth_dev->data->rx_queues[rx_queue_id] = NULL; 716a27d9013SAlfredo Cardigliano } 717a27d9013SAlfredo Cardigliano 7189fdf11c4SAndrew Boyer eth_dev->data->rx_queue_state[rx_queue_id] = 7199fdf11c4SAndrew Boyer RTE_ETH_QUEUE_STATE_STOPPED; 7209fdf11c4SAndrew Boyer 721*be39f75cSAndrew Boyer err = ionic_rx_qcq_alloc(lif, rx_queue_id, nb_desc, 722*be39f75cSAndrew Boyer &rxq); 723a27d9013SAlfredo Cardigliano if (err) { 7244ae96cb8SAndrew Boyer IONIC_PRINT(ERR, "Queue %d allocation failure", rx_queue_id); 725a27d9013SAlfredo Cardigliano return -EINVAL; 726a27d9013SAlfredo Cardigliano } 727a27d9013SAlfredo Cardigliano 728a27d9013SAlfredo Cardigliano rxq->mb_pool = mp; 729a27d9013SAlfredo Cardigliano 730a27d9013SAlfredo Cardigliano /* 731a27d9013SAlfredo Cardigliano * Note: the interface does not currently support 732a27d9013SAlfredo Cardigliano * DEV_RX_OFFLOAD_KEEP_CRC, please also consider ETHER_CRC_LEN 733a27d9013SAlfredo Cardigliano * when the adapter will be able to keep the CRC and subtract 734a27d9013SAlfredo Cardigliano * it to the length for all received packets: 735a27d9013SAlfredo Cardigliano * if (eth_dev->data->dev_conf.rxmode.offloads & 736a27d9013SAlfredo Cardigliano * DEV_RX_OFFLOAD_KEEP_CRC) 737a27d9013SAlfredo Cardigliano * rxq->crc_len = ETHER_CRC_LEN; 738a27d9013SAlfredo Cardigliano */ 739a27d9013SAlfredo Cardigliano 740a27d9013SAlfredo Cardigliano /* Do not start queue with rte_eth_dev_start() */ 74102eabf57SAndrew Boyer if (rx_conf->rx_deferred_start) 74202eabf57SAndrew Boyer rxq->flags |= IONIC_QCQ_F_DEFERRED; 743a27d9013SAlfredo Cardigliano 744a27d9013SAlfredo Cardigliano eth_dev->data->rx_queues[rx_queue_id] = rxq; 745a27d9013SAlfredo Cardigliano 746a27d9013SAlfredo Cardigliano return 0; 747a27d9013SAlfredo Cardigliano } 748a27d9013SAlfredo Cardigliano 7490de3e209SAndrew Boyer static __rte_always_inline void 750*be39f75cSAndrew Boyer ionic_rx_clean(struct ionic_rx_qcq *rxq, 751a27d9013SAlfredo Cardigliano uint32_t q_desc_index, uint32_t cq_desc_index, 752700f974dSAndrew Boyer void *service_cb_arg) 753a27d9013SAlfredo Cardigliano { 754*be39f75cSAndrew Boyer struct ionic_queue *q = &rxq->qcq.q; 755*be39f75cSAndrew Boyer struct ionic_cq *cq = &rxq->qcq.cq; 756c6a9a6fbSAndrew Boyer struct ionic_rxq_comp *cq_desc_base = cq->base; 757a27d9013SAlfredo Cardigliano struct ionic_rxq_comp *cq_desc = &cq_desc_base[cq_desc_index]; 758700f974dSAndrew Boyer struct rte_mbuf *rxm, *rxm_seg; 759a27d9013SAlfredo Cardigliano uint32_t max_frame_size = 760*be39f75cSAndrew Boyer rxq->qcq.lif->eth_dev->data->dev_conf.rxmode.max_rx_pkt_len; 761a27d9013SAlfredo Cardigliano uint64_t pkt_flags = 0; 762a27d9013SAlfredo Cardigliano uint32_t pkt_type; 763*be39f75cSAndrew Boyer struct ionic_rx_stats *stats = &rxq->stats; 764a27d9013SAlfredo Cardigliano struct ionic_rx_service *recv_args = (struct ionic_rx_service *) 765a27d9013SAlfredo Cardigliano service_cb_arg; 766a27d9013SAlfredo Cardigliano uint32_t buf_size = (uint16_t) 767a27d9013SAlfredo Cardigliano (rte_pktmbuf_data_room_size(rxq->mb_pool) - 768a27d9013SAlfredo Cardigliano RTE_PKTMBUF_HEADROOM); 769a27d9013SAlfredo Cardigliano uint32_t left; 770700f974dSAndrew Boyer void **info; 771700f974dSAndrew Boyer 772700f974dSAndrew Boyer assert(q_desc_index == cq_desc->comp_index); 773700f974dSAndrew Boyer 774700f974dSAndrew Boyer info = IONIC_INFO_PTR(q, cq_desc->comp_index); 775700f974dSAndrew Boyer 776700f974dSAndrew Boyer rxm = info[0]; 777a27d9013SAlfredo Cardigliano 778a27d9013SAlfredo Cardigliano if (!recv_args) { 779a27d9013SAlfredo Cardigliano stats->no_cb_arg++; 780a27d9013SAlfredo Cardigliano /* Flush */ 781a27d9013SAlfredo Cardigliano rte_pktmbuf_free(rxm); 782a27d9013SAlfredo Cardigliano /* 783a27d9013SAlfredo Cardigliano * Note: rte_mempool_put is faster with no segs 784a27d9013SAlfredo Cardigliano * rte_mempool_put(rxq->mb_pool, rxm); 785a27d9013SAlfredo Cardigliano */ 786a27d9013SAlfredo Cardigliano return; 787a27d9013SAlfredo Cardigliano } 788a27d9013SAlfredo Cardigliano 789a27d9013SAlfredo Cardigliano if (cq_desc->status) { 790a27d9013SAlfredo Cardigliano stats->bad_cq_status++; 791a27d9013SAlfredo Cardigliano ionic_rx_recycle(q, q_desc_index, rxm); 792a27d9013SAlfredo Cardigliano return; 793a27d9013SAlfredo Cardigliano } 794a27d9013SAlfredo Cardigliano 795a27d9013SAlfredo Cardigliano if (recv_args->nb_rx >= recv_args->nb_pkts) { 796a27d9013SAlfredo Cardigliano stats->no_room++; 797a27d9013SAlfredo Cardigliano ionic_rx_recycle(q, q_desc_index, rxm); 798a27d9013SAlfredo Cardigliano return; 799a27d9013SAlfredo Cardigliano } 800a27d9013SAlfredo Cardigliano 801a27d9013SAlfredo Cardigliano if (cq_desc->len > max_frame_size || 802a27d9013SAlfredo Cardigliano cq_desc->len == 0) { 803a27d9013SAlfredo Cardigliano stats->bad_len++; 804a27d9013SAlfredo Cardigliano ionic_rx_recycle(q, q_desc_index, rxm); 805a27d9013SAlfredo Cardigliano return; 806a27d9013SAlfredo Cardigliano } 807a27d9013SAlfredo Cardigliano 808a27d9013SAlfredo Cardigliano rxm->data_off = RTE_PKTMBUF_HEADROOM; 809a27d9013SAlfredo Cardigliano rte_prefetch1((char *)rxm->buf_addr + rxm->data_off); 810a27d9013SAlfredo Cardigliano rxm->nb_segs = 1; /* cq_desc->num_sg_elems */ 811a27d9013SAlfredo Cardigliano rxm->pkt_len = cq_desc->len; 812*be39f75cSAndrew Boyer rxm->port = rxq->qcq.lif->port_id; 813a27d9013SAlfredo Cardigliano 814a27d9013SAlfredo Cardigliano left = cq_desc->len; 815a27d9013SAlfredo Cardigliano 816a27d9013SAlfredo Cardigliano rxm->data_len = RTE_MIN(buf_size, left); 817a27d9013SAlfredo Cardigliano left -= rxm->data_len; 818a27d9013SAlfredo Cardigliano 819a27d9013SAlfredo Cardigliano rxm_seg = rxm->next; 820a27d9013SAlfredo Cardigliano while (rxm_seg && left) { 821a27d9013SAlfredo Cardigliano rxm_seg->data_len = RTE_MIN(buf_size, left); 822a27d9013SAlfredo Cardigliano left -= rxm_seg->data_len; 823a27d9013SAlfredo Cardigliano 824a27d9013SAlfredo Cardigliano rxm_seg = rxm_seg->next; 825a27d9013SAlfredo Cardigliano rxm->nb_segs++; 826a27d9013SAlfredo Cardigliano } 827a27d9013SAlfredo Cardigliano 82822e7171bSAlfredo Cardigliano /* RSS */ 82922e7171bSAlfredo Cardigliano pkt_flags |= PKT_RX_RSS_HASH; 83022e7171bSAlfredo Cardigliano rxm->hash.rss = cq_desc->rss_hash; 83122e7171bSAlfredo Cardigliano 832a27d9013SAlfredo Cardigliano /* Vlan Strip */ 833a27d9013SAlfredo Cardigliano if (cq_desc->csum_flags & IONIC_RXQ_COMP_CSUM_F_VLAN) { 834a27d9013SAlfredo Cardigliano pkt_flags |= PKT_RX_VLAN | PKT_RX_VLAN_STRIPPED; 835a27d9013SAlfredo Cardigliano rxm->vlan_tci = cq_desc->vlan_tci; 836a27d9013SAlfredo Cardigliano } 837a27d9013SAlfredo Cardigliano 838a27d9013SAlfredo Cardigliano /* Checksum */ 839a27d9013SAlfredo Cardigliano if (cq_desc->csum_flags & IONIC_RXQ_COMP_CSUM_F_CALC) { 840a27d9013SAlfredo Cardigliano if (cq_desc->csum_flags & IONIC_RXQ_COMP_CSUM_F_IP_OK) 841a27d9013SAlfredo Cardigliano pkt_flags |= PKT_RX_IP_CKSUM_GOOD; 842a27d9013SAlfredo Cardigliano else if (cq_desc->csum_flags & IONIC_RXQ_COMP_CSUM_F_IP_BAD) 843a27d9013SAlfredo Cardigliano pkt_flags |= PKT_RX_IP_CKSUM_BAD; 844a27d9013SAlfredo Cardigliano 845a27d9013SAlfredo Cardigliano if ((cq_desc->csum_flags & IONIC_RXQ_COMP_CSUM_F_TCP_OK) || 846a27d9013SAlfredo Cardigliano (cq_desc->csum_flags & IONIC_RXQ_COMP_CSUM_F_UDP_OK)) 847a27d9013SAlfredo Cardigliano pkt_flags |= PKT_RX_L4_CKSUM_GOOD; 848a27d9013SAlfredo Cardigliano else if ((cq_desc->csum_flags & 849a27d9013SAlfredo Cardigliano IONIC_RXQ_COMP_CSUM_F_TCP_BAD) || 850a27d9013SAlfredo Cardigliano (cq_desc->csum_flags & 851a27d9013SAlfredo Cardigliano IONIC_RXQ_COMP_CSUM_F_UDP_BAD)) 852a27d9013SAlfredo Cardigliano pkt_flags |= PKT_RX_L4_CKSUM_BAD; 853a27d9013SAlfredo Cardigliano } 854a27d9013SAlfredo Cardigliano 855a27d9013SAlfredo Cardigliano rxm->ol_flags = pkt_flags; 856a27d9013SAlfredo Cardigliano 857a27d9013SAlfredo Cardigliano /* Packet Type */ 858a27d9013SAlfredo Cardigliano switch (cq_desc->pkt_type_color & IONIC_RXQ_COMP_PKT_TYPE_MASK) { 859a27d9013SAlfredo Cardigliano case IONIC_PKT_TYPE_IPV4: 860a27d9013SAlfredo Cardigliano pkt_type = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4; 861a27d9013SAlfredo Cardigliano break; 862a27d9013SAlfredo Cardigliano case IONIC_PKT_TYPE_IPV6: 863a27d9013SAlfredo Cardigliano pkt_type = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6; 864a27d9013SAlfredo Cardigliano break; 865a27d9013SAlfredo Cardigliano case IONIC_PKT_TYPE_IPV4_TCP: 866a27d9013SAlfredo Cardigliano pkt_type = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4 | 867a27d9013SAlfredo Cardigliano RTE_PTYPE_L4_TCP; 868a27d9013SAlfredo Cardigliano break; 869a27d9013SAlfredo Cardigliano case IONIC_PKT_TYPE_IPV6_TCP: 870a27d9013SAlfredo Cardigliano pkt_type = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6 | 871a27d9013SAlfredo Cardigliano RTE_PTYPE_L4_TCP; 872a27d9013SAlfredo Cardigliano break; 873a27d9013SAlfredo Cardigliano case IONIC_PKT_TYPE_IPV4_UDP: 874a27d9013SAlfredo Cardigliano pkt_type = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4 | 875a27d9013SAlfredo Cardigliano RTE_PTYPE_L4_UDP; 876a27d9013SAlfredo Cardigliano break; 877a27d9013SAlfredo Cardigliano case IONIC_PKT_TYPE_IPV6_UDP: 878a27d9013SAlfredo Cardigliano pkt_type = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6 | 879a27d9013SAlfredo Cardigliano RTE_PTYPE_L4_UDP; 880a27d9013SAlfredo Cardigliano break; 881a27d9013SAlfredo Cardigliano default: 882a27d9013SAlfredo Cardigliano { 883a27d9013SAlfredo Cardigliano struct rte_ether_hdr *eth_h = rte_pktmbuf_mtod(rxm, 884a27d9013SAlfredo Cardigliano struct rte_ether_hdr *); 885a27d9013SAlfredo Cardigliano uint16_t ether_type = eth_h->ether_type; 886a27d9013SAlfredo Cardigliano if (ether_type == rte_cpu_to_be_16(RTE_ETHER_TYPE_ARP)) 887a27d9013SAlfredo Cardigliano pkt_type = RTE_PTYPE_L2_ETHER_ARP; 888a27d9013SAlfredo Cardigliano else 889a27d9013SAlfredo Cardigliano pkt_type = RTE_PTYPE_UNKNOWN; 890a27d9013SAlfredo Cardigliano break; 891a27d9013SAlfredo Cardigliano } 892a27d9013SAlfredo Cardigliano } 893a27d9013SAlfredo Cardigliano 894a27d9013SAlfredo Cardigliano rxm->packet_type = pkt_type; 895a27d9013SAlfredo Cardigliano 896a27d9013SAlfredo Cardigliano recv_args->rx_pkts[recv_args->nb_rx] = rxm; 897a27d9013SAlfredo Cardigliano recv_args->nb_rx++; 898a27d9013SAlfredo Cardigliano 899a27d9013SAlfredo Cardigliano stats->packets++; 900a27d9013SAlfredo Cardigliano stats->bytes += rxm->pkt_len; 901a27d9013SAlfredo Cardigliano } 902a27d9013SAlfredo Cardigliano 903a27d9013SAlfredo Cardigliano static void 904a27d9013SAlfredo Cardigliano ionic_rx_recycle(struct ionic_queue *q, uint32_t q_desc_index, 905a27d9013SAlfredo Cardigliano struct rte_mbuf *mbuf) 906a27d9013SAlfredo Cardigliano { 907a27d9013SAlfredo Cardigliano struct ionic_rxq_desc *desc_base = q->base; 908a27d9013SAlfredo Cardigliano struct ionic_rxq_desc *old = &desc_base[q_desc_index]; 909a27d9013SAlfredo Cardigliano struct ionic_rxq_desc *new = &desc_base[q->head_idx]; 910a27d9013SAlfredo Cardigliano 911a27d9013SAlfredo Cardigliano new->addr = old->addr; 912a27d9013SAlfredo Cardigliano new->len = old->len; 913a27d9013SAlfredo Cardigliano 914700f974dSAndrew Boyer ionic_q_post(q, true, mbuf); 915a27d9013SAlfredo Cardigliano } 916a27d9013SAlfredo Cardigliano 9170de3e209SAndrew Boyer static __rte_always_inline int 918*be39f75cSAndrew Boyer ionic_rx_fill(struct ionic_rx_qcq *rxq, uint32_t len) 919a27d9013SAlfredo Cardigliano { 920*be39f75cSAndrew Boyer struct ionic_queue *q = &rxq->qcq.q; 921*be39f75cSAndrew Boyer struct ionic_rxq_desc *desc, *desc_base = q->base; 922*be39f75cSAndrew Boyer struct ionic_rxq_sg_desc *sg_desc, *sg_desc_base = q->sg_base; 923a27d9013SAlfredo Cardigliano struct ionic_rxq_sg_elem *elem; 924a27d9013SAlfredo Cardigliano rte_iova_t dma_addr; 925a27d9013SAlfredo Cardigliano uint32_t i, j, nsegs, buf_size, size; 926a27d9013SAlfredo Cardigliano bool ring_doorbell; 927a27d9013SAlfredo Cardigliano 928a27d9013SAlfredo Cardigliano buf_size = (uint16_t)(rte_pktmbuf_data_room_size(rxq->mb_pool) - 929a27d9013SAlfredo Cardigliano RTE_PKTMBUF_HEADROOM); 930a27d9013SAlfredo Cardigliano 931a27d9013SAlfredo Cardigliano /* Initialize software ring entries */ 932a27d9013SAlfredo Cardigliano for (i = ionic_q_space_avail(q); i; i--) { 933a27d9013SAlfredo Cardigliano struct rte_mbuf *rxm = rte_mbuf_raw_alloc(rxq->mb_pool); 934a27d9013SAlfredo Cardigliano struct rte_mbuf *prev_rxm_seg; 935a27d9013SAlfredo Cardigliano 936a27d9013SAlfredo Cardigliano if (rxm == NULL) { 937a27d9013SAlfredo Cardigliano IONIC_PRINT(ERR, "RX mbuf alloc failed"); 938a27d9013SAlfredo Cardigliano return -ENOMEM; 939a27d9013SAlfredo Cardigliano } 940a27d9013SAlfredo Cardigliano 941a27d9013SAlfredo Cardigliano nsegs = (len + buf_size - 1) / buf_size; 942a27d9013SAlfredo Cardigliano 943a27d9013SAlfredo Cardigliano desc = &desc_base[q->head_idx]; 944a27d9013SAlfredo Cardigliano dma_addr = rte_cpu_to_le_64(rte_mbuf_data_iova_default(rxm)); 945a27d9013SAlfredo Cardigliano desc->addr = dma_addr; 946a27d9013SAlfredo Cardigliano desc->len = buf_size; 947a27d9013SAlfredo Cardigliano size = buf_size; 948a27d9013SAlfredo Cardigliano desc->opcode = (nsegs > 1) ? IONIC_RXQ_DESC_OPCODE_SG : 949a27d9013SAlfredo Cardigliano IONIC_RXQ_DESC_OPCODE_SIMPLE; 950a27d9013SAlfredo Cardigliano rxm->next = NULL; 951a27d9013SAlfredo Cardigliano 952a27d9013SAlfredo Cardigliano prev_rxm_seg = rxm; 953a27d9013SAlfredo Cardigliano sg_desc = &sg_desc_base[q->head_idx]; 954a27d9013SAlfredo Cardigliano elem = sg_desc->elems; 955a27d9013SAlfredo Cardigliano for (j = 0; j < nsegs - 1 && j < IONIC_RX_MAX_SG_ELEMS; j++) { 956a27d9013SAlfredo Cardigliano struct rte_mbuf *rxm_seg; 957a27d9013SAlfredo Cardigliano rte_iova_t data_iova; 958a27d9013SAlfredo Cardigliano 959a27d9013SAlfredo Cardigliano rxm_seg = rte_mbuf_raw_alloc(rxq->mb_pool); 960a27d9013SAlfredo Cardigliano if (rxm_seg == NULL) { 961a27d9013SAlfredo Cardigliano IONIC_PRINT(ERR, "RX mbuf alloc failed"); 962a27d9013SAlfredo Cardigliano return -ENOMEM; 963a27d9013SAlfredo Cardigliano } 964a27d9013SAlfredo Cardigliano 965a27d9013SAlfredo Cardigliano data_iova = rte_mbuf_data_iova(rxm_seg); 966a27d9013SAlfredo Cardigliano dma_addr = rte_cpu_to_le_64(data_iova); 967a27d9013SAlfredo Cardigliano elem->addr = dma_addr; 968a27d9013SAlfredo Cardigliano elem->len = buf_size; 969a27d9013SAlfredo Cardigliano size += buf_size; 970a27d9013SAlfredo Cardigliano elem++; 971a27d9013SAlfredo Cardigliano rxm_seg->next = NULL; 972a27d9013SAlfredo Cardigliano prev_rxm_seg->next = rxm_seg; 973a27d9013SAlfredo Cardigliano prev_rxm_seg = rxm_seg; 974a27d9013SAlfredo Cardigliano } 975a27d9013SAlfredo Cardigliano 976a27d9013SAlfredo Cardigliano if (size < len) 977a27d9013SAlfredo Cardigliano IONIC_PRINT(ERR, "Rx SG size is not sufficient (%d < %d)", 978a27d9013SAlfredo Cardigliano size, len); 979a27d9013SAlfredo Cardigliano 980a27d9013SAlfredo Cardigliano ring_doorbell = ((q->head_idx + 1) & 981a27d9013SAlfredo Cardigliano IONIC_RX_RING_DOORBELL_STRIDE) == 0; 982a27d9013SAlfredo Cardigliano 983700f974dSAndrew Boyer ionic_q_post(q, ring_doorbell, rxm); 984a27d9013SAlfredo Cardigliano } 985a27d9013SAlfredo Cardigliano 986a27d9013SAlfredo Cardigliano return 0; 987a27d9013SAlfredo Cardigliano } 988a27d9013SAlfredo Cardigliano 989a27d9013SAlfredo Cardigliano /* 990a27d9013SAlfredo Cardigliano * Start Receive Units for specified queue. 991a27d9013SAlfredo Cardigliano */ 992ce6427ddSThomas Monjalon int __rte_cold 993a27d9013SAlfredo Cardigliano ionic_dev_rx_queue_start(struct rte_eth_dev *eth_dev, uint16_t rx_queue_id) 994a27d9013SAlfredo Cardigliano { 995a27d9013SAlfredo Cardigliano uint32_t frame_size = eth_dev->data->dev_conf.rxmode.max_rx_pkt_len; 9969fdf11c4SAndrew Boyer uint8_t *rx_queue_state = eth_dev->data->rx_queue_state; 997*be39f75cSAndrew Boyer struct ionic_rx_qcq *rxq; 998a27d9013SAlfredo Cardigliano int err; 999a27d9013SAlfredo Cardigliano 10009fdf11c4SAndrew Boyer if (rx_queue_state[rx_queue_id] == RTE_ETH_QUEUE_STATE_STARTED) { 10019fdf11c4SAndrew Boyer IONIC_PRINT(DEBUG, "RX queue %u already started", 10029fdf11c4SAndrew Boyer rx_queue_id); 10039fdf11c4SAndrew Boyer return 0; 10049fdf11c4SAndrew Boyer } 10059fdf11c4SAndrew Boyer 1006a27d9013SAlfredo Cardigliano rxq = eth_dev->data->rx_queues[rx_queue_id]; 1007a27d9013SAlfredo Cardigliano 10084ae96cb8SAndrew Boyer IONIC_PRINT(DEBUG, "Starting RX queue %u, %u descs (size: %u)", 1009*be39f75cSAndrew Boyer rx_queue_id, rxq->qcq.q.num_descs, frame_size); 10104ae96cb8SAndrew Boyer 10119fdf11c4SAndrew Boyer if (!(rxq->flags & IONIC_QCQ_F_INITED)) { 1012a27d9013SAlfredo Cardigliano err = ionic_lif_rxq_init(rxq); 1013a27d9013SAlfredo Cardigliano if (err) 1014a27d9013SAlfredo Cardigliano return err; 1015b5d9a4f0SAndrew Boyer } else { 1016*be39f75cSAndrew Boyer ionic_qcq_enable(&rxq->qcq); 10179fdf11c4SAndrew Boyer } 1018a27d9013SAlfredo Cardigliano 1019a27d9013SAlfredo Cardigliano /* Allocate buffers for descriptor rings */ 1020a27d9013SAlfredo Cardigliano if (ionic_rx_fill(rxq, frame_size) != 0) { 1021a27d9013SAlfredo Cardigliano IONIC_PRINT(ERR, "Could not alloc mbuf for queue:%d", 1022a27d9013SAlfredo Cardigliano rx_queue_id); 1023a27d9013SAlfredo Cardigliano return -1; 1024a27d9013SAlfredo Cardigliano } 1025a27d9013SAlfredo Cardigliano 10269fdf11c4SAndrew Boyer rx_queue_state[rx_queue_id] = RTE_ETH_QUEUE_STATE_STARTED; 1027a27d9013SAlfredo Cardigliano 1028a27d9013SAlfredo Cardigliano return 0; 1029a27d9013SAlfredo Cardigliano } 1030a27d9013SAlfredo Cardigliano 10310de3e209SAndrew Boyer static __rte_always_inline void 1032*be39f75cSAndrew Boyer ionic_rxq_service(struct ionic_rx_qcq *rxq, uint32_t work_to_do, 1033a27d9013SAlfredo Cardigliano void *service_cb_arg) 1034a27d9013SAlfredo Cardigliano { 1035*be39f75cSAndrew Boyer struct ionic_cq *cq = &rxq->qcq.cq; 1036*be39f75cSAndrew Boyer struct ionic_queue *q = &rxq->qcq.q; 1037*be39f75cSAndrew Boyer struct ionic_rxq_comp *cq_desc, *cq_desc_base = cq->base; 1038a27d9013SAlfredo Cardigliano bool more; 1039a27d9013SAlfredo Cardigliano uint32_t curr_q_tail_idx, curr_cq_tail_idx; 1040a27d9013SAlfredo Cardigliano uint32_t work_done = 0; 1041a27d9013SAlfredo Cardigliano 1042a27d9013SAlfredo Cardigliano if (work_to_do == 0) 1043a27d9013SAlfredo Cardigliano return; 1044a27d9013SAlfredo Cardigliano 1045a27d9013SAlfredo Cardigliano cq_desc = &cq_desc_base[cq->tail_idx]; 1046a27d9013SAlfredo Cardigliano while (color_match(cq_desc->pkt_type_color, cq->done_color)) { 1047a27d9013SAlfredo Cardigliano curr_cq_tail_idx = cq->tail_idx; 10482aed9865SAndrew Boyer cq->tail_idx = Q_NEXT_TO_SRVC(cq, 1); 1049a27d9013SAlfredo Cardigliano 1050a27d9013SAlfredo Cardigliano if (cq->tail_idx == 0) 1051a27d9013SAlfredo Cardigliano cq->done_color = !cq->done_color; 1052a27d9013SAlfredo Cardigliano 1053a27d9013SAlfredo Cardigliano /* Prefetch the next 4 descriptors */ 1054a27d9013SAlfredo Cardigliano if ((cq->tail_idx & 0x3) == 0) 1055a27d9013SAlfredo Cardigliano rte_prefetch0(&cq_desc_base[cq->tail_idx]); 1056a27d9013SAlfredo Cardigliano 1057a27d9013SAlfredo Cardigliano do { 1058a27d9013SAlfredo Cardigliano more = (q->tail_idx != cq_desc->comp_index); 1059a27d9013SAlfredo Cardigliano 1060a27d9013SAlfredo Cardigliano curr_q_tail_idx = q->tail_idx; 10614ad56b7aSAndrew Boyer q->tail_idx = Q_NEXT_TO_SRVC(q, 1); 1062a27d9013SAlfredo Cardigliano 1063a27d9013SAlfredo Cardigliano /* Prefetch the next 4 descriptors */ 1064a27d9013SAlfredo Cardigliano if ((q->tail_idx & 0x3) == 0) 1065a27d9013SAlfredo Cardigliano /* q desc info */ 1066a27d9013SAlfredo Cardigliano rte_prefetch0(&q->info[q->tail_idx]); 1067a27d9013SAlfredo Cardigliano 1068c6a9a6fbSAndrew Boyer ionic_rx_clean(rxq, curr_q_tail_idx, curr_cq_tail_idx, 1069700f974dSAndrew Boyer service_cb_arg); 1070a27d9013SAlfredo Cardigliano 1071a27d9013SAlfredo Cardigliano } while (more); 1072a27d9013SAlfredo Cardigliano 1073a27d9013SAlfredo Cardigliano if (++work_done == work_to_do) 1074a27d9013SAlfredo Cardigliano break; 1075a27d9013SAlfredo Cardigliano 1076a27d9013SAlfredo Cardigliano cq_desc = &cq_desc_base[cq->tail_idx]; 1077a27d9013SAlfredo Cardigliano } 1078a27d9013SAlfredo Cardigliano } 1079a27d9013SAlfredo Cardigliano 1080a27d9013SAlfredo Cardigliano /* 1081a27d9013SAlfredo Cardigliano * Stop Receive Units for specified queue. 1082a27d9013SAlfredo Cardigliano */ 1083ce6427ddSThomas Monjalon int __rte_cold 1084a27d9013SAlfredo Cardigliano ionic_dev_rx_queue_stop(struct rte_eth_dev *eth_dev, uint16_t rx_queue_id) 1085a27d9013SAlfredo Cardigliano { 1086*be39f75cSAndrew Boyer struct ionic_rx_qcq *rxq; 1087a27d9013SAlfredo Cardigliano 10884ae96cb8SAndrew Boyer IONIC_PRINT(DEBUG, "Stopping RX queue %u", rx_queue_id); 1089a27d9013SAlfredo Cardigliano 1090a27d9013SAlfredo Cardigliano rxq = eth_dev->data->rx_queues[rx_queue_id]; 1091a27d9013SAlfredo Cardigliano 10929fdf11c4SAndrew Boyer eth_dev->data->rx_queue_state[rx_queue_id] = 10939fdf11c4SAndrew Boyer RTE_ETH_QUEUE_STATE_STOPPED; 10949fdf11c4SAndrew Boyer 1095*be39f75cSAndrew Boyer ionic_qcq_disable(&rxq->qcq); 1096a27d9013SAlfredo Cardigliano 1097a27d9013SAlfredo Cardigliano /* Flush */ 10982aed9865SAndrew Boyer ionic_rxq_service(rxq, -1, NULL); 1099a27d9013SAlfredo Cardigliano 1100a27d9013SAlfredo Cardigliano return 0; 1101a27d9013SAlfredo Cardigliano } 1102a27d9013SAlfredo Cardigliano 1103a27d9013SAlfredo Cardigliano uint16_t 1104a27d9013SAlfredo Cardigliano ionic_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts, 1105a27d9013SAlfredo Cardigliano uint16_t nb_pkts) 1106a27d9013SAlfredo Cardigliano { 1107*be39f75cSAndrew Boyer struct ionic_rx_qcq *rxq = rx_queue; 1108a27d9013SAlfredo Cardigliano uint32_t frame_size = 1109*be39f75cSAndrew Boyer rxq->qcq.lif->eth_dev->data->dev_conf.rxmode.max_rx_pkt_len; 1110a27d9013SAlfredo Cardigliano struct ionic_rx_service service_cb_arg; 1111a27d9013SAlfredo Cardigliano 1112a27d9013SAlfredo Cardigliano service_cb_arg.rx_pkts = rx_pkts; 1113a27d9013SAlfredo Cardigliano service_cb_arg.nb_pkts = nb_pkts; 1114a27d9013SAlfredo Cardigliano service_cb_arg.nb_rx = 0; 1115a27d9013SAlfredo Cardigliano 11162aed9865SAndrew Boyer ionic_rxq_service(rxq, nb_pkts, &service_cb_arg); 1117a27d9013SAlfredo Cardigliano 1118a27d9013SAlfredo Cardigliano ionic_rx_fill(rxq, frame_size); 1119a27d9013SAlfredo Cardigliano 1120a27d9013SAlfredo Cardigliano return service_cb_arg.nb_rx; 1121a27d9013SAlfredo Cardigliano } 1122