1a27d9013SAlfredo Cardigliano /* SPDX-License-Identifier: (BSD-3-Clause OR GPL-2.0) 2a27d9013SAlfredo Cardigliano * Copyright(c) 2018-2019 Pensando Systems, Inc. All rights reserved. 3a27d9013SAlfredo Cardigliano */ 4a27d9013SAlfredo Cardigliano 5a27d9013SAlfredo Cardigliano #include <sys/queue.h> 6a27d9013SAlfredo Cardigliano #include <stdio.h> 7a27d9013SAlfredo Cardigliano #include <stdlib.h> 8a27d9013SAlfredo Cardigliano #include <string.h> 9a27d9013SAlfredo Cardigliano #include <errno.h> 10a27d9013SAlfredo Cardigliano #include <stdint.h> 11a27d9013SAlfredo Cardigliano #include <stdarg.h> 12a27d9013SAlfredo Cardigliano #include <unistd.h> 13a27d9013SAlfredo Cardigliano #include <inttypes.h> 14a27d9013SAlfredo Cardigliano 15a27d9013SAlfredo Cardigliano #include <rte_byteorder.h> 16a27d9013SAlfredo Cardigliano #include <rte_common.h> 17a27d9013SAlfredo Cardigliano #include <rte_cycles.h> 18a27d9013SAlfredo Cardigliano #include <rte_log.h> 19a27d9013SAlfredo Cardigliano #include <rte_debug.h> 20a27d9013SAlfredo Cardigliano #include <rte_interrupts.h> 21a27d9013SAlfredo Cardigliano #include <rte_pci.h> 22a27d9013SAlfredo Cardigliano #include <rte_memory.h> 23a27d9013SAlfredo Cardigliano #include <rte_memzone.h> 24a27d9013SAlfredo Cardigliano #include <rte_launch.h> 25a27d9013SAlfredo Cardigliano #include <rte_eal.h> 26a27d9013SAlfredo Cardigliano #include <rte_per_lcore.h> 27a27d9013SAlfredo Cardigliano #include <rte_lcore.h> 28a27d9013SAlfredo Cardigliano #include <rte_atomic.h> 29a27d9013SAlfredo Cardigliano #include <rte_branch_prediction.h> 30a27d9013SAlfredo Cardigliano #include <rte_mempool.h> 31a27d9013SAlfredo Cardigliano #include <rte_malloc.h> 32a27d9013SAlfredo Cardigliano #include <rte_mbuf.h> 33a27d9013SAlfredo Cardigliano #include <rte_ether.h> 34df96fd0dSBruce Richardson #include <ethdev_driver.h> 35a27d9013SAlfredo Cardigliano #include <rte_prefetch.h> 36a27d9013SAlfredo Cardigliano #include <rte_udp.h> 37a27d9013SAlfredo Cardigliano #include <rte_tcp.h> 38a27d9013SAlfredo Cardigliano #include <rte_sctp.h> 39a27d9013SAlfredo Cardigliano #include <rte_string_fns.h> 40a27d9013SAlfredo Cardigliano #include <rte_errno.h> 41a27d9013SAlfredo Cardigliano #include <rte_ip.h> 42a27d9013SAlfredo Cardigliano #include <rte_net.h> 43a27d9013SAlfredo Cardigliano 44a27d9013SAlfredo Cardigliano #include "ionic_logs.h" 45a27d9013SAlfredo Cardigliano #include "ionic_mac_api.h" 46a27d9013SAlfredo Cardigliano #include "ionic_ethdev.h" 47a27d9013SAlfredo Cardigliano #include "ionic_lif.h" 48a27d9013SAlfredo Cardigliano #include "ionic_rxtx.h" 49a27d9013SAlfredo Cardigliano 50a27d9013SAlfredo Cardigliano #define IONIC_RX_RING_DOORBELL_STRIDE (32 - 1) 51a27d9013SAlfredo Cardigliano 52a27d9013SAlfredo Cardigliano /********************************************************************* 53a27d9013SAlfredo Cardigliano * 54a27d9013SAlfredo Cardigliano * TX functions 55a27d9013SAlfredo Cardigliano * 56a27d9013SAlfredo Cardigliano **********************************************************************/ 57a27d9013SAlfredo Cardigliano 58a27d9013SAlfredo Cardigliano void 59a27d9013SAlfredo Cardigliano ionic_txq_info_get(struct rte_eth_dev *dev, uint16_t queue_id, 60a27d9013SAlfredo Cardigliano struct rte_eth_txq_info *qinfo) 61a27d9013SAlfredo Cardigliano { 62a27d9013SAlfredo Cardigliano struct ionic_qcq *txq = dev->data->tx_queues[queue_id]; 63a27d9013SAlfredo Cardigliano struct ionic_queue *q = &txq->q; 64a27d9013SAlfredo Cardigliano 65a27d9013SAlfredo Cardigliano qinfo->nb_desc = q->num_descs; 6668591087SAndrew Boyer qinfo->conf.offloads = dev->data->dev_conf.txmode.offloads; 6702eabf57SAndrew Boyer qinfo->conf.tx_deferred_start = txq->flags & IONIC_QCQ_F_DEFERRED; 68a27d9013SAlfredo Cardigliano } 69a27d9013SAlfredo Cardigliano 700de3e209SAndrew Boyer static __rte_always_inline void 712aed9865SAndrew Boyer ionic_tx_flush(struct ionic_qcq *txq) 72a27d9013SAlfredo Cardigliano { 732aed9865SAndrew Boyer struct ionic_cq *cq = &txq->cq; 742aed9865SAndrew Boyer struct ionic_queue *q = &txq->q; 75a27d9013SAlfredo Cardigliano struct rte_mbuf *txm, *next; 76a27d9013SAlfredo Cardigliano struct ionic_txq_comp *cq_desc_base = cq->base; 77a27d9013SAlfredo Cardigliano struct ionic_txq_comp *cq_desc; 78700f974dSAndrew Boyer void **info; 79a27d9013SAlfredo Cardigliano u_int32_t comp_index = (u_int32_t)-1; 80a27d9013SAlfredo Cardigliano 81a27d9013SAlfredo Cardigliano cq_desc = &cq_desc_base[cq->tail_idx]; 82a27d9013SAlfredo Cardigliano while (color_match(cq_desc->color, cq->done_color)) { 832aed9865SAndrew Boyer cq->tail_idx = Q_NEXT_TO_SRVC(cq, 1); 84a27d9013SAlfredo Cardigliano 85a27d9013SAlfredo Cardigliano /* Prefetch the next 4 descriptors (not really useful here) */ 86a27d9013SAlfredo Cardigliano if ((cq->tail_idx & 0x3) == 0) 87a27d9013SAlfredo Cardigliano rte_prefetch0(&cq_desc_base[cq->tail_idx]); 88a27d9013SAlfredo Cardigliano 89a27d9013SAlfredo Cardigliano if (cq->tail_idx == 0) 90a27d9013SAlfredo Cardigliano cq->done_color = !cq->done_color; 91a27d9013SAlfredo Cardigliano 92a27d9013SAlfredo Cardigliano comp_index = cq_desc->comp_index; 93a27d9013SAlfredo Cardigliano 94a27d9013SAlfredo Cardigliano cq_desc = &cq_desc_base[cq->tail_idx]; 95a27d9013SAlfredo Cardigliano } 96a27d9013SAlfredo Cardigliano 97a27d9013SAlfredo Cardigliano if (comp_index != (u_int32_t)-1) { 98a27d9013SAlfredo Cardigliano while (q->tail_idx != comp_index) { 99700f974dSAndrew Boyer info = IONIC_INFO_PTR(q, q->tail_idx); 100a27d9013SAlfredo Cardigliano 101a27d9013SAlfredo Cardigliano q->tail_idx = (q->tail_idx + 1) & (q->num_descs - 1); 102a27d9013SAlfredo Cardigliano 103a27d9013SAlfredo Cardigliano /* Prefetch the next 4 descriptors */ 104a27d9013SAlfredo Cardigliano if ((q->tail_idx & 0x3) == 0) 105a27d9013SAlfredo Cardigliano /* q desc info */ 106a27d9013SAlfredo Cardigliano rte_prefetch0(&q->info[q->tail_idx]); 107a27d9013SAlfredo Cardigliano 108a27d9013SAlfredo Cardigliano /* 109a27d9013SAlfredo Cardigliano * Note: you can just use rte_pktmbuf_free, 110a27d9013SAlfredo Cardigliano * but this loop is faster 111a27d9013SAlfredo Cardigliano */ 112700f974dSAndrew Boyer txm = info[0]; 113a27d9013SAlfredo Cardigliano while (txm != NULL) { 114a27d9013SAlfredo Cardigliano next = txm->next; 115a27d9013SAlfredo Cardigliano rte_pktmbuf_free_seg(txm); 116a27d9013SAlfredo Cardigliano txm = next; 117a27d9013SAlfredo Cardigliano } 118a27d9013SAlfredo Cardigliano } 119a27d9013SAlfredo Cardigliano } 120a27d9013SAlfredo Cardigliano } 121a27d9013SAlfredo Cardigliano 122ce6427ddSThomas Monjalon void __rte_cold 123a27d9013SAlfredo Cardigliano ionic_dev_tx_queue_release(void *tx_queue) 124a27d9013SAlfredo Cardigliano { 125a27d9013SAlfredo Cardigliano struct ionic_qcq *txq = (struct ionic_qcq *)tx_queue; 126a27d9013SAlfredo Cardigliano 127a27d9013SAlfredo Cardigliano IONIC_PRINT_CALL(); 128a27d9013SAlfredo Cardigliano 1299fdf11c4SAndrew Boyer ionic_lif_txq_deinit(txq); 1309fdf11c4SAndrew Boyer 131a27d9013SAlfredo Cardigliano ionic_qcq_free(txq); 132a27d9013SAlfredo Cardigliano } 133a27d9013SAlfredo Cardigliano 134ce6427ddSThomas Monjalon int __rte_cold 135a27d9013SAlfredo Cardigliano ionic_dev_tx_queue_stop(struct rte_eth_dev *eth_dev, uint16_t tx_queue_id) 136a27d9013SAlfredo Cardigliano { 137a27d9013SAlfredo Cardigliano struct ionic_qcq *txq; 138a27d9013SAlfredo Cardigliano 1394ae96cb8SAndrew Boyer IONIC_PRINT(DEBUG, "Stopping TX queue %u", tx_queue_id); 140a27d9013SAlfredo Cardigliano 141a27d9013SAlfredo Cardigliano txq = eth_dev->data->tx_queues[tx_queue_id]; 142a27d9013SAlfredo Cardigliano 1439fdf11c4SAndrew Boyer eth_dev->data->tx_queue_state[tx_queue_id] = 1449fdf11c4SAndrew Boyer RTE_ETH_QUEUE_STATE_STOPPED; 1459fdf11c4SAndrew Boyer 146a27d9013SAlfredo Cardigliano /* 147a27d9013SAlfredo Cardigliano * Note: we should better post NOP Tx desc and wait for its completion 148a27d9013SAlfredo Cardigliano * before disabling Tx queue 149a27d9013SAlfredo Cardigliano */ 150a27d9013SAlfredo Cardigliano 151a27d9013SAlfredo Cardigliano ionic_qcq_disable(txq); 152a27d9013SAlfredo Cardigliano 1532aed9865SAndrew Boyer ionic_tx_flush(txq); 154a27d9013SAlfredo Cardigliano 155a27d9013SAlfredo Cardigliano return 0; 156a27d9013SAlfredo Cardigliano } 157a27d9013SAlfredo Cardigliano 158ce6427ddSThomas Monjalon int __rte_cold 159a27d9013SAlfredo Cardigliano ionic_dev_tx_queue_setup(struct rte_eth_dev *eth_dev, uint16_t tx_queue_id, 1604ae96cb8SAndrew Boyer uint16_t nb_desc, uint32_t socket_id, 161a27d9013SAlfredo Cardigliano const struct rte_eth_txconf *tx_conf) 162a27d9013SAlfredo Cardigliano { 163a27d9013SAlfredo Cardigliano struct ionic_lif *lif = IONIC_ETH_DEV_TO_LIF(eth_dev); 164a27d9013SAlfredo Cardigliano struct ionic_qcq *txq; 165a27d9013SAlfredo Cardigliano uint64_t offloads; 166a27d9013SAlfredo Cardigliano int err; 167a27d9013SAlfredo Cardigliano 168a27d9013SAlfredo Cardigliano if (tx_queue_id >= lif->ntxqcqs) { 169a27d9013SAlfredo Cardigliano IONIC_PRINT(DEBUG, "Queue index %u not available " 170a27d9013SAlfredo Cardigliano "(max %u queues)", 171a27d9013SAlfredo Cardigliano tx_queue_id, lif->ntxqcqs); 172a27d9013SAlfredo Cardigliano return -EINVAL; 173a27d9013SAlfredo Cardigliano } 174a27d9013SAlfredo Cardigliano 175a27d9013SAlfredo Cardigliano offloads = tx_conf->offloads | eth_dev->data->dev_conf.txmode.offloads; 1764ae96cb8SAndrew Boyer IONIC_PRINT(DEBUG, 1774ae96cb8SAndrew Boyer "Configuring skt %u TX queue %u with %u buffers, offloads %jx", 1784ae96cb8SAndrew Boyer socket_id, tx_queue_id, nb_desc, offloads); 179a27d9013SAlfredo Cardigliano 180a27d9013SAlfredo Cardigliano /* Validate number of receive descriptors */ 181a27d9013SAlfredo Cardigliano if (!rte_is_power_of_2(nb_desc) || nb_desc < IONIC_MIN_RING_DESC) 182a27d9013SAlfredo Cardigliano return -EINVAL; /* or use IONIC_DEFAULT_RING_DESC */ 183a27d9013SAlfredo Cardigliano 184a27d9013SAlfredo Cardigliano /* Free memory prior to re-allocation if needed... */ 185a27d9013SAlfredo Cardigliano if (eth_dev->data->tx_queues[tx_queue_id] != NULL) { 186a27d9013SAlfredo Cardigliano void *tx_queue = eth_dev->data->tx_queues[tx_queue_id]; 187a27d9013SAlfredo Cardigliano ionic_dev_tx_queue_release(tx_queue); 188a27d9013SAlfredo Cardigliano eth_dev->data->tx_queues[tx_queue_id] = NULL; 189a27d9013SAlfredo Cardigliano } 190a27d9013SAlfredo Cardigliano 1919fdf11c4SAndrew Boyer eth_dev->data->tx_queue_state[tx_queue_id] = 1929fdf11c4SAndrew Boyer RTE_ETH_QUEUE_STATE_STOPPED; 1939fdf11c4SAndrew Boyer 194a27d9013SAlfredo Cardigliano err = ionic_tx_qcq_alloc(lif, tx_queue_id, nb_desc, &txq); 195a27d9013SAlfredo Cardigliano if (err) { 196a27d9013SAlfredo Cardigliano IONIC_PRINT(DEBUG, "Queue allocation failure"); 197a27d9013SAlfredo Cardigliano return -EINVAL; 198a27d9013SAlfredo Cardigliano } 199a27d9013SAlfredo Cardigliano 200a27d9013SAlfredo Cardigliano /* Do not start queue with rte_eth_dev_start() */ 20102eabf57SAndrew Boyer if (tx_conf->tx_deferred_start) 20202eabf57SAndrew Boyer txq->flags |= IONIC_QCQ_F_DEFERRED; 203a27d9013SAlfredo Cardigliano 20468591087SAndrew Boyer /* Convert the offload flags into queue flags */ 20568591087SAndrew Boyer if (offloads & DEV_TX_OFFLOAD_IPV4_CKSUM) 20668591087SAndrew Boyer txq->flags |= IONIC_QCQ_F_CSUM_L3; 20768591087SAndrew Boyer if (offloads & DEV_TX_OFFLOAD_TCP_CKSUM) 20868591087SAndrew Boyer txq->flags |= IONIC_QCQ_F_CSUM_TCP; 20968591087SAndrew Boyer if (offloads & DEV_TX_OFFLOAD_UDP_CKSUM) 21068591087SAndrew Boyer txq->flags |= IONIC_QCQ_F_CSUM_UDP; 211a27d9013SAlfredo Cardigliano 212a27d9013SAlfredo Cardigliano eth_dev->data->tx_queues[tx_queue_id] = txq; 213a27d9013SAlfredo Cardigliano 214a27d9013SAlfredo Cardigliano return 0; 215a27d9013SAlfredo Cardigliano } 216a27d9013SAlfredo Cardigliano 217a27d9013SAlfredo Cardigliano /* 218a27d9013SAlfredo Cardigliano * Start Transmit Units for specified queue. 219a27d9013SAlfredo Cardigliano */ 220ce6427ddSThomas Monjalon int __rte_cold 221a27d9013SAlfredo Cardigliano ionic_dev_tx_queue_start(struct rte_eth_dev *eth_dev, uint16_t tx_queue_id) 222a27d9013SAlfredo Cardigliano { 2239fdf11c4SAndrew Boyer uint8_t *tx_queue_state = eth_dev->data->tx_queue_state; 224a27d9013SAlfredo Cardigliano struct ionic_qcq *txq; 225a27d9013SAlfredo Cardigliano int err; 226a27d9013SAlfredo Cardigliano 2279fdf11c4SAndrew Boyer if (tx_queue_state[tx_queue_id] == RTE_ETH_QUEUE_STATE_STARTED) { 2289fdf11c4SAndrew Boyer IONIC_PRINT(DEBUG, "TX queue %u already started", 2299fdf11c4SAndrew Boyer tx_queue_id); 2309fdf11c4SAndrew Boyer return 0; 2319fdf11c4SAndrew Boyer } 2329fdf11c4SAndrew Boyer 233a27d9013SAlfredo Cardigliano txq = eth_dev->data->tx_queues[tx_queue_id]; 234a27d9013SAlfredo Cardigliano 2354ae96cb8SAndrew Boyer IONIC_PRINT(DEBUG, "Starting TX queue %u, %u descs", 2364ae96cb8SAndrew Boyer tx_queue_id, txq->q.num_descs); 2374ae96cb8SAndrew Boyer 2389fdf11c4SAndrew Boyer if (!(txq->flags & IONIC_QCQ_F_INITED)) { 239a27d9013SAlfredo Cardigliano err = ionic_lif_txq_init(txq); 240a27d9013SAlfredo Cardigliano if (err) 241a27d9013SAlfredo Cardigliano return err; 242b5d9a4f0SAndrew Boyer } else { 243a27d9013SAlfredo Cardigliano ionic_qcq_enable(txq); 244b5d9a4f0SAndrew Boyer } 245a27d9013SAlfredo Cardigliano 2469fdf11c4SAndrew Boyer tx_queue_state[tx_queue_id] = RTE_ETH_QUEUE_STATE_STARTED; 247a27d9013SAlfredo Cardigliano 248a27d9013SAlfredo Cardigliano return 0; 249a27d9013SAlfredo Cardigliano } 250a27d9013SAlfredo Cardigliano 251a27d9013SAlfredo Cardigliano static void 25264b08152SAlfredo Cardigliano ionic_tx_tcp_pseudo_csum(struct rte_mbuf *txm) 25364b08152SAlfredo Cardigliano { 25464b08152SAlfredo Cardigliano struct ether_hdr *eth_hdr = rte_pktmbuf_mtod(txm, struct ether_hdr *); 25564b08152SAlfredo Cardigliano char *l3_hdr = ((char *)eth_hdr) + txm->l2_len; 25664b08152SAlfredo Cardigliano struct rte_tcp_hdr *tcp_hdr = (struct rte_tcp_hdr *) 25764b08152SAlfredo Cardigliano (l3_hdr + txm->l3_len); 25864b08152SAlfredo Cardigliano 25964b08152SAlfredo Cardigliano if (txm->ol_flags & PKT_TX_IP_CKSUM) { 26064b08152SAlfredo Cardigliano struct rte_ipv4_hdr *ipv4_hdr = (struct rte_ipv4_hdr *)l3_hdr; 26164b08152SAlfredo Cardigliano ipv4_hdr->hdr_checksum = 0; 26264b08152SAlfredo Cardigliano tcp_hdr->cksum = 0; 26364b08152SAlfredo Cardigliano tcp_hdr->cksum = rte_ipv4_udptcp_cksum(ipv4_hdr, tcp_hdr); 26464b08152SAlfredo Cardigliano } else { 26564b08152SAlfredo Cardigliano struct rte_ipv6_hdr *ipv6_hdr = (struct rte_ipv6_hdr *)l3_hdr; 26664b08152SAlfredo Cardigliano tcp_hdr->cksum = 0; 26764b08152SAlfredo Cardigliano tcp_hdr->cksum = rte_ipv6_udptcp_cksum(ipv6_hdr, tcp_hdr); 26864b08152SAlfredo Cardigliano } 26964b08152SAlfredo Cardigliano } 27064b08152SAlfredo Cardigliano 27164b08152SAlfredo Cardigliano static void 27264b08152SAlfredo Cardigliano ionic_tx_tcp_inner_pseudo_csum(struct rte_mbuf *txm) 27364b08152SAlfredo Cardigliano { 27464b08152SAlfredo Cardigliano struct ether_hdr *eth_hdr = rte_pktmbuf_mtod(txm, struct ether_hdr *); 27564b08152SAlfredo Cardigliano char *l3_hdr = ((char *)eth_hdr) + txm->outer_l2_len + 27664b08152SAlfredo Cardigliano txm->outer_l3_len + txm->l2_len; 27764b08152SAlfredo Cardigliano struct rte_tcp_hdr *tcp_hdr = (struct rte_tcp_hdr *) 27864b08152SAlfredo Cardigliano (l3_hdr + txm->l3_len); 27964b08152SAlfredo Cardigliano 28064b08152SAlfredo Cardigliano if (txm->ol_flags & PKT_TX_IPV4) { 28164b08152SAlfredo Cardigliano struct rte_ipv4_hdr *ipv4_hdr = (struct rte_ipv4_hdr *)l3_hdr; 28264b08152SAlfredo Cardigliano ipv4_hdr->hdr_checksum = 0; 28364b08152SAlfredo Cardigliano tcp_hdr->cksum = 0; 28464b08152SAlfredo Cardigliano tcp_hdr->cksum = rte_ipv4_udptcp_cksum(ipv4_hdr, tcp_hdr); 28564b08152SAlfredo Cardigliano } else { 28664b08152SAlfredo Cardigliano struct rte_ipv6_hdr *ipv6_hdr = (struct rte_ipv6_hdr *)l3_hdr; 28764b08152SAlfredo Cardigliano tcp_hdr->cksum = 0; 28864b08152SAlfredo Cardigliano tcp_hdr->cksum = rte_ipv6_udptcp_cksum(ipv6_hdr, tcp_hdr); 28964b08152SAlfredo Cardigliano } 29064b08152SAlfredo Cardigliano } 29164b08152SAlfredo Cardigliano 29264b08152SAlfredo Cardigliano static void 293a27d9013SAlfredo Cardigliano ionic_tx_tso_post(struct ionic_queue *q, struct ionic_txq_desc *desc, 294a27d9013SAlfredo Cardigliano struct rte_mbuf *txm, 295a27d9013SAlfredo Cardigliano rte_iova_t addr, uint8_t nsge, uint16_t len, 296a27d9013SAlfredo Cardigliano uint32_t hdrlen, uint32_t mss, 29764b08152SAlfredo Cardigliano bool encap, 298a27d9013SAlfredo Cardigliano uint16_t vlan_tci, bool has_vlan, 299a27d9013SAlfredo Cardigliano bool start, bool done) 300a27d9013SAlfredo Cardigliano { 301a27d9013SAlfredo Cardigliano uint8_t flags = 0; 302a27d9013SAlfredo Cardigliano flags |= has_vlan ? IONIC_TXQ_DESC_FLAG_VLAN : 0; 30364b08152SAlfredo Cardigliano flags |= encap ? IONIC_TXQ_DESC_FLAG_ENCAP : 0; 304a27d9013SAlfredo Cardigliano flags |= start ? IONIC_TXQ_DESC_FLAG_TSO_SOT : 0; 305a27d9013SAlfredo Cardigliano flags |= done ? IONIC_TXQ_DESC_FLAG_TSO_EOT : 0; 306a27d9013SAlfredo Cardigliano 307a27d9013SAlfredo Cardigliano desc->cmd = encode_txq_desc_cmd(IONIC_TXQ_DESC_OPCODE_TSO, 308a27d9013SAlfredo Cardigliano flags, nsge, addr); 309a27d9013SAlfredo Cardigliano desc->len = len; 310a27d9013SAlfredo Cardigliano desc->vlan_tci = vlan_tci; 311a27d9013SAlfredo Cardigliano desc->hdr_len = hdrlen; 312a27d9013SAlfredo Cardigliano desc->mss = mss; 313a27d9013SAlfredo Cardigliano 314700f974dSAndrew Boyer ionic_q_post(q, done, done ? txm : NULL); 315a27d9013SAlfredo Cardigliano } 316a27d9013SAlfredo Cardigliano 317a27d9013SAlfredo Cardigliano static struct ionic_txq_desc * 318a27d9013SAlfredo Cardigliano ionic_tx_tso_next(struct ionic_queue *q, struct ionic_txq_sg_elem **elem) 319a27d9013SAlfredo Cardigliano { 320a27d9013SAlfredo Cardigliano struct ionic_txq_desc *desc_base = q->base; 32156117636SAndrew Boyer struct ionic_txq_sg_desc_v1 *sg_desc_base = q->sg_base; 322a27d9013SAlfredo Cardigliano struct ionic_txq_desc *desc = &desc_base[q->head_idx]; 32356117636SAndrew Boyer struct ionic_txq_sg_desc_v1 *sg_desc = &sg_desc_base[q->head_idx]; 324a27d9013SAlfredo Cardigliano 325a27d9013SAlfredo Cardigliano *elem = sg_desc->elems; 326a27d9013SAlfredo Cardigliano return desc; 327a27d9013SAlfredo Cardigliano } 328a27d9013SAlfredo Cardigliano 329a27d9013SAlfredo Cardigliano static int 33068591087SAndrew Boyer ionic_tx_tso(struct ionic_qcq *txq, struct rte_mbuf *txm, 33168591087SAndrew Boyer bool not_xmit_more) 332a27d9013SAlfredo Cardigliano { 33368591087SAndrew Boyer struct ionic_queue *q = &txq->q; 334a27d9013SAlfredo Cardigliano struct ionic_tx_stats *stats = IONIC_Q_TO_TX_STATS(q); 335a27d9013SAlfredo Cardigliano struct ionic_txq_desc *desc; 336a27d9013SAlfredo Cardigliano struct ionic_txq_sg_elem *elem; 337a27d9013SAlfredo Cardigliano struct rte_mbuf *txm_seg; 3387c3a867bSAndrew Boyer rte_iova_t data_iova; 3397c3a867bSAndrew Boyer uint64_t desc_addr = 0, next_addr; 340a27d9013SAlfredo Cardigliano uint16_t desc_len = 0; 341a27d9013SAlfredo Cardigliano uint8_t desc_nsge; 342a27d9013SAlfredo Cardigliano uint32_t hdrlen; 343a27d9013SAlfredo Cardigliano uint32_t mss = txm->tso_segsz; 344a27d9013SAlfredo Cardigliano uint32_t frag_left = 0; 345a27d9013SAlfredo Cardigliano uint32_t left; 346a27d9013SAlfredo Cardigliano uint32_t seglen; 347a27d9013SAlfredo Cardigliano uint32_t len; 348a27d9013SAlfredo Cardigliano uint32_t offset = 0; 349a27d9013SAlfredo Cardigliano bool start, done; 35064b08152SAlfredo Cardigliano bool encap; 351a27d9013SAlfredo Cardigliano bool has_vlan = !!(txm->ol_flags & PKT_TX_VLAN_PKT); 352a27d9013SAlfredo Cardigliano uint16_t vlan_tci = txm->vlan_tci; 35364b08152SAlfredo Cardigliano uint64_t ol_flags = txm->ol_flags; 354a27d9013SAlfredo Cardigliano 35564b08152SAlfredo Cardigliano encap = ((ol_flags & PKT_TX_OUTER_IP_CKSUM) || 35664b08152SAlfredo Cardigliano (ol_flags & PKT_TX_OUTER_UDP_CKSUM)) && 35764b08152SAlfredo Cardigliano ((ol_flags & PKT_TX_OUTER_IPV4) || 35864b08152SAlfredo Cardigliano (ol_flags & PKT_TX_OUTER_IPV6)); 35964b08152SAlfredo Cardigliano 36064b08152SAlfredo Cardigliano /* Preload inner-most TCP csum field with IP pseudo hdr 36164b08152SAlfredo Cardigliano * calculated with IP length set to zero. HW will later 36264b08152SAlfredo Cardigliano * add in length to each TCP segment resulting from the TSO. 36364b08152SAlfredo Cardigliano */ 36464b08152SAlfredo Cardigliano 36564b08152SAlfredo Cardigliano if (encap) { 36664b08152SAlfredo Cardigliano ionic_tx_tcp_inner_pseudo_csum(txm); 36764b08152SAlfredo Cardigliano hdrlen = txm->outer_l2_len + txm->outer_l3_len + 36864b08152SAlfredo Cardigliano txm->l2_len + txm->l3_len + txm->l4_len; 36964b08152SAlfredo Cardigliano } else { 37064b08152SAlfredo Cardigliano ionic_tx_tcp_pseudo_csum(txm); 37164b08152SAlfredo Cardigliano hdrlen = txm->l2_len + txm->l3_len + txm->l4_len; 37264b08152SAlfredo Cardigliano } 373a27d9013SAlfredo Cardigliano 374a27d9013SAlfredo Cardigliano seglen = hdrlen + mss; 375a27d9013SAlfredo Cardigliano left = txm->data_len; 3767c3a867bSAndrew Boyer data_iova = rte_mbuf_data_iova(txm); 377a27d9013SAlfredo Cardigliano 378a27d9013SAlfredo Cardigliano desc = ionic_tx_tso_next(q, &elem); 379a27d9013SAlfredo Cardigliano start = true; 380a27d9013SAlfredo Cardigliano 381a27d9013SAlfredo Cardigliano /* Chop data up into desc segments */ 382a27d9013SAlfredo Cardigliano 383a27d9013SAlfredo Cardigliano while (left > 0) { 384a27d9013SAlfredo Cardigliano len = RTE_MIN(seglen, left); 385a27d9013SAlfredo Cardigliano frag_left = seglen - len; 3867c3a867bSAndrew Boyer desc_addr = rte_cpu_to_le_64(data_iova + offset); 387a27d9013SAlfredo Cardigliano desc_len = len; 388a27d9013SAlfredo Cardigliano desc_nsge = 0; 389a27d9013SAlfredo Cardigliano left -= len; 390a27d9013SAlfredo Cardigliano offset += len; 391a27d9013SAlfredo Cardigliano if (txm->nb_segs > 1 && frag_left > 0) 392a27d9013SAlfredo Cardigliano continue; 393a27d9013SAlfredo Cardigliano done = (txm->nb_segs == 1 && left == 0); 394a27d9013SAlfredo Cardigliano ionic_tx_tso_post(q, desc, txm, 395a27d9013SAlfredo Cardigliano desc_addr, desc_nsge, desc_len, 396a27d9013SAlfredo Cardigliano hdrlen, mss, 39764b08152SAlfredo Cardigliano encap, 398a27d9013SAlfredo Cardigliano vlan_tci, has_vlan, 399a27d9013SAlfredo Cardigliano start, done && not_xmit_more); 400a27d9013SAlfredo Cardigliano desc = ionic_tx_tso_next(q, &elem); 401a27d9013SAlfredo Cardigliano start = false; 402a27d9013SAlfredo Cardigliano seglen = mss; 403a27d9013SAlfredo Cardigliano } 404a27d9013SAlfredo Cardigliano 405a27d9013SAlfredo Cardigliano /* Chop frags into desc segments */ 406a27d9013SAlfredo Cardigliano 407a27d9013SAlfredo Cardigliano txm_seg = txm->next; 408a27d9013SAlfredo Cardigliano while (txm_seg != NULL) { 409a27d9013SAlfredo Cardigliano offset = 0; 4107c3a867bSAndrew Boyer data_iova = rte_mbuf_data_iova(txm_seg); 411a27d9013SAlfredo Cardigliano left = txm_seg->data_len; 412a27d9013SAlfredo Cardigliano stats->frags++; 413a27d9013SAlfredo Cardigliano 414a27d9013SAlfredo Cardigliano while (left > 0) { 4157c3a867bSAndrew Boyer next_addr = rte_cpu_to_le_64(data_iova + offset); 416a27d9013SAlfredo Cardigliano if (frag_left > 0) { 417a27d9013SAlfredo Cardigliano len = RTE_MIN(frag_left, left); 418a27d9013SAlfredo Cardigliano frag_left -= len; 4197c3a867bSAndrew Boyer elem->addr = next_addr; 420a27d9013SAlfredo Cardigliano elem->len = len; 421a27d9013SAlfredo Cardigliano elem++; 422a27d9013SAlfredo Cardigliano desc_nsge++; 423a27d9013SAlfredo Cardigliano } else { 424a27d9013SAlfredo Cardigliano len = RTE_MIN(mss, left); 425a27d9013SAlfredo Cardigliano frag_left = mss - len; 4267c3a867bSAndrew Boyer desc_addr = next_addr; 427a27d9013SAlfredo Cardigliano desc_len = len; 428a27d9013SAlfredo Cardigliano desc_nsge = 0; 429a27d9013SAlfredo Cardigliano } 430a27d9013SAlfredo Cardigliano left -= len; 431a27d9013SAlfredo Cardigliano offset += len; 432a27d9013SAlfredo Cardigliano if (txm_seg->next != NULL && frag_left > 0) 433a27d9013SAlfredo Cardigliano continue; 4347c3a867bSAndrew Boyer 435a27d9013SAlfredo Cardigliano done = (txm_seg->next == NULL && left == 0); 436a27d9013SAlfredo Cardigliano ionic_tx_tso_post(q, desc, txm_seg, 437a27d9013SAlfredo Cardigliano desc_addr, desc_nsge, desc_len, 438a27d9013SAlfredo Cardigliano hdrlen, mss, 43964b08152SAlfredo Cardigliano encap, 440a27d9013SAlfredo Cardigliano vlan_tci, has_vlan, 441a27d9013SAlfredo Cardigliano start, done && not_xmit_more); 442a27d9013SAlfredo Cardigliano desc = ionic_tx_tso_next(q, &elem); 443a27d9013SAlfredo Cardigliano start = false; 444a27d9013SAlfredo Cardigliano } 445a27d9013SAlfredo Cardigliano 446a27d9013SAlfredo Cardigliano txm_seg = txm_seg->next; 447a27d9013SAlfredo Cardigliano } 448a27d9013SAlfredo Cardigliano 449a27d9013SAlfredo Cardigliano stats->tso++; 450a27d9013SAlfredo Cardigliano 451a27d9013SAlfredo Cardigliano return 0; 452a27d9013SAlfredo Cardigliano } 453a27d9013SAlfredo Cardigliano 4540de3e209SAndrew Boyer static __rte_always_inline int 45568591087SAndrew Boyer ionic_tx(struct ionic_qcq *txq, struct rte_mbuf *txm, 45668591087SAndrew Boyer bool not_xmit_more) 457a27d9013SAlfredo Cardigliano { 45868591087SAndrew Boyer struct ionic_queue *q = &txq->q; 459a27d9013SAlfredo Cardigliano struct ionic_txq_desc *desc_base = q->base; 46056117636SAndrew Boyer struct ionic_txq_sg_desc_v1 *sg_desc_base = q->sg_base; 461a27d9013SAlfredo Cardigliano struct ionic_txq_desc *desc = &desc_base[q->head_idx]; 46256117636SAndrew Boyer struct ionic_txq_sg_desc_v1 *sg_desc = &sg_desc_base[q->head_idx]; 463a27d9013SAlfredo Cardigliano struct ionic_txq_sg_elem *elem = sg_desc->elems; 464a27d9013SAlfredo Cardigliano struct ionic_tx_stats *stats = IONIC_Q_TO_TX_STATS(q); 465a27d9013SAlfredo Cardigliano struct rte_mbuf *txm_seg; 46664b08152SAlfredo Cardigliano bool encap; 467a27d9013SAlfredo Cardigliano bool has_vlan; 468a27d9013SAlfredo Cardigliano uint64_t ol_flags = txm->ol_flags; 4697c3a867bSAndrew Boyer uint64_t addr; 470a27d9013SAlfredo Cardigliano uint8_t opcode = IONIC_TXQ_DESC_OPCODE_CSUM_NONE; 471a27d9013SAlfredo Cardigliano uint8_t flags = 0; 472a27d9013SAlfredo Cardigliano 47364b08152SAlfredo Cardigliano if ((ol_flags & PKT_TX_IP_CKSUM) && 47468591087SAndrew Boyer (txq->flags & IONIC_QCQ_F_CSUM_L3)) { 47564b08152SAlfredo Cardigliano opcode = IONIC_TXQ_DESC_OPCODE_CSUM_HW; 47664b08152SAlfredo Cardigliano flags |= IONIC_TXQ_DESC_FLAG_CSUM_L3; 477f603eebcSAndrew Boyer } 478f603eebcSAndrew Boyer 47964b08152SAlfredo Cardigliano if (((ol_flags & PKT_TX_TCP_CKSUM) && 48068591087SAndrew Boyer (txq->flags & IONIC_QCQ_F_CSUM_TCP)) || 48164b08152SAlfredo Cardigliano ((ol_flags & PKT_TX_UDP_CKSUM) && 48268591087SAndrew Boyer (txq->flags & IONIC_QCQ_F_CSUM_UDP))) { 483f603eebcSAndrew Boyer opcode = IONIC_TXQ_DESC_OPCODE_CSUM_HW; 48464b08152SAlfredo Cardigliano flags |= IONIC_TXQ_DESC_FLAG_CSUM_L4; 48564b08152SAlfredo Cardigliano } 48664b08152SAlfredo Cardigliano 487f603eebcSAndrew Boyer if (opcode == IONIC_TXQ_DESC_OPCODE_CSUM_NONE) 488f603eebcSAndrew Boyer stats->no_csum++; 489f603eebcSAndrew Boyer 490a27d9013SAlfredo Cardigliano has_vlan = (ol_flags & PKT_TX_VLAN_PKT); 49164b08152SAlfredo Cardigliano encap = ((ol_flags & PKT_TX_OUTER_IP_CKSUM) || 49264b08152SAlfredo Cardigliano (ol_flags & PKT_TX_OUTER_UDP_CKSUM)) && 49364b08152SAlfredo Cardigliano ((ol_flags & PKT_TX_OUTER_IPV4) || 49464b08152SAlfredo Cardigliano (ol_flags & PKT_TX_OUTER_IPV6)); 495a27d9013SAlfredo Cardigliano 496a27d9013SAlfredo Cardigliano flags |= has_vlan ? IONIC_TXQ_DESC_FLAG_VLAN : 0; 49764b08152SAlfredo Cardigliano flags |= encap ? IONIC_TXQ_DESC_FLAG_ENCAP : 0; 498a27d9013SAlfredo Cardigliano 4997c3a867bSAndrew Boyer addr = rte_cpu_to_le_64(rte_mbuf_data_iova(txm)); 5007c3a867bSAndrew Boyer 501a27d9013SAlfredo Cardigliano desc->cmd = encode_txq_desc_cmd(opcode, flags, txm->nb_segs - 1, addr); 502a27d9013SAlfredo Cardigliano desc->len = txm->data_len; 503a27d9013SAlfredo Cardigliano desc->vlan_tci = txm->vlan_tci; 504a27d9013SAlfredo Cardigliano 505a27d9013SAlfredo Cardigliano txm_seg = txm->next; 506a27d9013SAlfredo Cardigliano while (txm_seg != NULL) { 507a27d9013SAlfredo Cardigliano elem->len = txm_seg->data_len; 508a27d9013SAlfredo Cardigliano elem->addr = rte_cpu_to_le_64(rte_mbuf_data_iova(txm_seg)); 509a27d9013SAlfredo Cardigliano stats->frags++; 510a27d9013SAlfredo Cardigliano elem++; 511a27d9013SAlfredo Cardigliano txm_seg = txm_seg->next; 512a27d9013SAlfredo Cardigliano } 513a27d9013SAlfredo Cardigliano 514700f974dSAndrew Boyer ionic_q_post(q, not_xmit_more, txm); 515a27d9013SAlfredo Cardigliano 516a27d9013SAlfredo Cardigliano return 0; 517a27d9013SAlfredo Cardigliano } 518a27d9013SAlfredo Cardigliano 519a27d9013SAlfredo Cardigliano uint16_t 520a27d9013SAlfredo Cardigliano ionic_xmit_pkts(void *tx_queue, struct rte_mbuf **tx_pkts, 521a27d9013SAlfredo Cardigliano uint16_t nb_pkts) 522a27d9013SAlfredo Cardigliano { 523a27d9013SAlfredo Cardigliano struct ionic_qcq *txq = (struct ionic_qcq *)tx_queue; 524a27d9013SAlfredo Cardigliano struct ionic_queue *q = &txq->q; 525a27d9013SAlfredo Cardigliano struct ionic_tx_stats *stats = IONIC_Q_TO_TX_STATS(q); 526a27d9013SAlfredo Cardigliano uint32_t next_q_head_idx; 527a27d9013SAlfredo Cardigliano uint32_t bytes_tx = 0; 528a27d9013SAlfredo Cardigliano uint16_t nb_tx = 0; 529a27d9013SAlfredo Cardigliano int err; 530a27d9013SAlfredo Cardigliano bool last; 531a27d9013SAlfredo Cardigliano 532a27d9013SAlfredo Cardigliano /* Cleaning old buffers */ 5332aed9865SAndrew Boyer ionic_tx_flush(txq); 534a27d9013SAlfredo Cardigliano 535a27d9013SAlfredo Cardigliano if (unlikely(ionic_q_space_avail(q) < nb_pkts)) { 536a27d9013SAlfredo Cardigliano stats->stop += nb_pkts; 537a27d9013SAlfredo Cardigliano return 0; 538a27d9013SAlfredo Cardigliano } 539a27d9013SAlfredo Cardigliano 540a27d9013SAlfredo Cardigliano while (nb_tx < nb_pkts) { 541a27d9013SAlfredo Cardigliano last = (nb_tx == (nb_pkts - 1)); 542a27d9013SAlfredo Cardigliano 543a27d9013SAlfredo Cardigliano next_q_head_idx = (q->head_idx + 1) & (q->num_descs - 1); 544a27d9013SAlfredo Cardigliano if ((next_q_head_idx & 0x3) == 0) { 545a27d9013SAlfredo Cardigliano struct ionic_txq_desc *desc_base = q->base; 546a27d9013SAlfredo Cardigliano rte_prefetch0(&desc_base[next_q_head_idx]); 547a27d9013SAlfredo Cardigliano rte_prefetch0(&q->info[next_q_head_idx]); 548a27d9013SAlfredo Cardigliano } 549a27d9013SAlfredo Cardigliano 550a27d9013SAlfredo Cardigliano if (tx_pkts[nb_tx]->ol_flags & PKT_TX_TCP_SEG) 55168591087SAndrew Boyer err = ionic_tx_tso(txq, tx_pkts[nb_tx], last); 552a27d9013SAlfredo Cardigliano else 55368591087SAndrew Boyer err = ionic_tx(txq, tx_pkts[nb_tx], last); 554a27d9013SAlfredo Cardigliano if (err) { 555a27d9013SAlfredo Cardigliano stats->drop += nb_pkts - nb_tx; 556a27d9013SAlfredo Cardigliano if (nb_tx > 0) 557a27d9013SAlfredo Cardigliano ionic_q_flush(q); 558a27d9013SAlfredo Cardigliano break; 559a27d9013SAlfredo Cardigliano } 560a27d9013SAlfredo Cardigliano 561a27d9013SAlfredo Cardigliano bytes_tx += tx_pkts[nb_tx]->pkt_len; 562a27d9013SAlfredo Cardigliano nb_tx++; 563a27d9013SAlfredo Cardigliano } 564a27d9013SAlfredo Cardigliano 565a27d9013SAlfredo Cardigliano stats->packets += nb_tx; 566a27d9013SAlfredo Cardigliano stats->bytes += bytes_tx; 567a27d9013SAlfredo Cardigliano 568a27d9013SAlfredo Cardigliano return nb_tx; 569a27d9013SAlfredo Cardigliano } 570a27d9013SAlfredo Cardigliano 571a27d9013SAlfredo Cardigliano /********************************************************************* 572a27d9013SAlfredo Cardigliano * 573a27d9013SAlfredo Cardigliano * TX prep functions 574a27d9013SAlfredo Cardigliano * 575a27d9013SAlfredo Cardigliano **********************************************************************/ 576a27d9013SAlfredo Cardigliano 577a27d9013SAlfredo Cardigliano #define IONIC_TX_OFFLOAD_MASK ( \ 578a27d9013SAlfredo Cardigliano PKT_TX_IPV4 | \ 579a27d9013SAlfredo Cardigliano PKT_TX_IPV6 | \ 580a27d9013SAlfredo Cardigliano PKT_TX_VLAN | \ 58164b08152SAlfredo Cardigliano PKT_TX_IP_CKSUM | \ 582a27d9013SAlfredo Cardigliano PKT_TX_TCP_SEG | \ 583a27d9013SAlfredo Cardigliano PKT_TX_L4_MASK) 584a27d9013SAlfredo Cardigliano 585a27d9013SAlfredo Cardigliano #define IONIC_TX_OFFLOAD_NOTSUP_MASK \ 586a27d9013SAlfredo Cardigliano (PKT_TX_OFFLOAD_MASK ^ IONIC_TX_OFFLOAD_MASK) 587a27d9013SAlfredo Cardigliano 588a27d9013SAlfredo Cardigliano uint16_t 589a27d9013SAlfredo Cardigliano ionic_prep_pkts(void *tx_queue __rte_unused, struct rte_mbuf **tx_pkts, 590a27d9013SAlfredo Cardigliano uint16_t nb_pkts) 591a27d9013SAlfredo Cardigliano { 592a27d9013SAlfredo Cardigliano struct rte_mbuf *txm; 593a27d9013SAlfredo Cardigliano uint64_t offloads; 594a27d9013SAlfredo Cardigliano int i = 0; 595a27d9013SAlfredo Cardigliano 596a27d9013SAlfredo Cardigliano for (i = 0; i < nb_pkts; i++) { 597a27d9013SAlfredo Cardigliano txm = tx_pkts[i]; 598a27d9013SAlfredo Cardigliano 599d13d7829SAndrew Boyer if (txm->nb_segs > IONIC_TX_MAX_SG_ELEMS_V1 + 1) { 600a27d9013SAlfredo Cardigliano rte_errno = -EINVAL; 601a27d9013SAlfredo Cardigliano break; 602a27d9013SAlfredo Cardigliano } 603a27d9013SAlfredo Cardigliano 604a27d9013SAlfredo Cardigliano offloads = txm->ol_flags; 605a27d9013SAlfredo Cardigliano 606a27d9013SAlfredo Cardigliano if (offloads & IONIC_TX_OFFLOAD_NOTSUP_MASK) { 607a27d9013SAlfredo Cardigliano rte_errno = -ENOTSUP; 608a27d9013SAlfredo Cardigliano break; 609a27d9013SAlfredo Cardigliano } 610a27d9013SAlfredo Cardigliano } 611a27d9013SAlfredo Cardigliano 612a27d9013SAlfredo Cardigliano return i; 613a27d9013SAlfredo Cardigliano } 614a27d9013SAlfredo Cardigliano 615a27d9013SAlfredo Cardigliano /********************************************************************* 616a27d9013SAlfredo Cardigliano * 617a27d9013SAlfredo Cardigliano * RX functions 618a27d9013SAlfredo Cardigliano * 619a27d9013SAlfredo Cardigliano **********************************************************************/ 620a27d9013SAlfredo Cardigliano 621a27d9013SAlfredo Cardigliano static void ionic_rx_recycle(struct ionic_queue *q, uint32_t q_desc_index, 622a27d9013SAlfredo Cardigliano struct rte_mbuf *mbuf); 623a27d9013SAlfredo Cardigliano 624a27d9013SAlfredo Cardigliano void 625a27d9013SAlfredo Cardigliano ionic_rxq_info_get(struct rte_eth_dev *dev, uint16_t queue_id, 626a27d9013SAlfredo Cardigliano struct rte_eth_rxq_info *qinfo) 627a27d9013SAlfredo Cardigliano { 628a27d9013SAlfredo Cardigliano struct ionic_qcq *rxq = dev->data->rx_queues[queue_id]; 629a27d9013SAlfredo Cardigliano struct ionic_queue *q = &rxq->q; 630a27d9013SAlfredo Cardigliano 631a27d9013SAlfredo Cardigliano qinfo->mp = rxq->mb_pool; 632a27d9013SAlfredo Cardigliano qinfo->scattered_rx = dev->data->scattered_rx; 633a27d9013SAlfredo Cardigliano qinfo->nb_desc = q->num_descs; 63402eabf57SAndrew Boyer qinfo->conf.rx_deferred_start = rxq->flags & IONIC_QCQ_F_DEFERRED; 63568591087SAndrew Boyer qinfo->conf.offloads = dev->data->dev_conf.rxmode.offloads; 636a27d9013SAlfredo Cardigliano } 637a27d9013SAlfredo Cardigliano 638ce6427ddSThomas Monjalon static void __rte_cold 639a27d9013SAlfredo Cardigliano ionic_rx_empty(struct ionic_queue *q) 640a27d9013SAlfredo Cardigliano { 641a27d9013SAlfredo Cardigliano struct ionic_qcq *rxq = IONIC_Q_TO_QCQ(q); 642a27d9013SAlfredo Cardigliano struct rte_mbuf *mbuf; 643700f974dSAndrew Boyer void **info; 644a27d9013SAlfredo Cardigliano 645a27d9013SAlfredo Cardigliano while (q->tail_idx != q->head_idx) { 646700f974dSAndrew Boyer info = IONIC_INFO_PTR(q, q->tail_idx); 647700f974dSAndrew Boyer mbuf = info[0]; 648a27d9013SAlfredo Cardigliano rte_mempool_put(rxq->mb_pool, mbuf); 649a27d9013SAlfredo Cardigliano 650a27d9013SAlfredo Cardigliano q->tail_idx = (q->tail_idx + 1) & (q->num_descs - 1); 651a27d9013SAlfredo Cardigliano } 652a27d9013SAlfredo Cardigliano } 653a27d9013SAlfredo Cardigliano 654ce6427ddSThomas Monjalon void __rte_cold 655a27d9013SAlfredo Cardigliano ionic_dev_rx_queue_release(void *rx_queue) 656a27d9013SAlfredo Cardigliano { 657a27d9013SAlfredo Cardigliano struct ionic_qcq *rxq = (struct ionic_qcq *)rx_queue; 658a27d9013SAlfredo Cardigliano 659a27d9013SAlfredo Cardigliano IONIC_PRINT_CALL(); 660a27d9013SAlfredo Cardigliano 661a27d9013SAlfredo Cardigliano ionic_rx_empty(&rxq->q); 662a27d9013SAlfredo Cardigliano 6639fdf11c4SAndrew Boyer ionic_lif_rxq_deinit(rxq); 6649fdf11c4SAndrew Boyer 665a27d9013SAlfredo Cardigliano ionic_qcq_free(rxq); 666a27d9013SAlfredo Cardigliano } 667a27d9013SAlfredo Cardigliano 668ce6427ddSThomas Monjalon int __rte_cold 669a27d9013SAlfredo Cardigliano ionic_dev_rx_queue_setup(struct rte_eth_dev *eth_dev, 670a27d9013SAlfredo Cardigliano uint16_t rx_queue_id, 671a27d9013SAlfredo Cardigliano uint16_t nb_desc, 6724ae96cb8SAndrew Boyer uint32_t socket_id, 673a27d9013SAlfredo Cardigliano const struct rte_eth_rxconf *rx_conf, 674a27d9013SAlfredo Cardigliano struct rte_mempool *mp) 675a27d9013SAlfredo Cardigliano { 676a27d9013SAlfredo Cardigliano struct ionic_lif *lif = IONIC_ETH_DEV_TO_LIF(eth_dev); 677a27d9013SAlfredo Cardigliano struct ionic_qcq *rxq; 678a27d9013SAlfredo Cardigliano uint64_t offloads; 679a27d9013SAlfredo Cardigliano int err; 680a27d9013SAlfredo Cardigliano 681a27d9013SAlfredo Cardigliano if (rx_queue_id >= lif->nrxqcqs) { 682a27d9013SAlfredo Cardigliano IONIC_PRINT(ERR, 683a27d9013SAlfredo Cardigliano "Queue index %u not available (max %u queues)", 684a27d9013SAlfredo Cardigliano rx_queue_id, lif->nrxqcqs); 685a27d9013SAlfredo Cardigliano return -EINVAL; 686a27d9013SAlfredo Cardigliano } 687a27d9013SAlfredo Cardigliano 688a27d9013SAlfredo Cardigliano offloads = rx_conf->offloads | eth_dev->data->dev_conf.rxmode.offloads; 6894ae96cb8SAndrew Boyer IONIC_PRINT(DEBUG, 6904ae96cb8SAndrew Boyer "Configuring skt %u RX queue %u with %u buffers, offloads %jx", 6914ae96cb8SAndrew Boyer socket_id, rx_queue_id, nb_desc, offloads); 692a27d9013SAlfredo Cardigliano 69318a44465SAndrew Boyer if (!rx_conf->rx_drop_en) 69418a44465SAndrew Boyer IONIC_PRINT(WARNING, "No-drop mode is not supported"); 69518a44465SAndrew Boyer 696a27d9013SAlfredo Cardigliano /* Validate number of receive descriptors */ 697a27d9013SAlfredo Cardigliano if (!rte_is_power_of_2(nb_desc) || 698a27d9013SAlfredo Cardigliano nb_desc < IONIC_MIN_RING_DESC || 699a27d9013SAlfredo Cardigliano nb_desc > IONIC_MAX_RING_DESC) { 700a27d9013SAlfredo Cardigliano IONIC_PRINT(ERR, 7014ae96cb8SAndrew Boyer "Bad descriptor count (%u) for queue %u (min: %u)", 702a27d9013SAlfredo Cardigliano nb_desc, rx_queue_id, IONIC_MIN_RING_DESC); 703a27d9013SAlfredo Cardigliano return -EINVAL; /* or use IONIC_DEFAULT_RING_DESC */ 704a27d9013SAlfredo Cardigliano } 705a27d9013SAlfredo Cardigliano 706a27d9013SAlfredo Cardigliano /* Free memory prior to re-allocation if needed... */ 707a27d9013SAlfredo Cardigliano if (eth_dev->data->rx_queues[rx_queue_id] != NULL) { 708a27d9013SAlfredo Cardigliano void *rx_queue = eth_dev->data->rx_queues[rx_queue_id]; 709a27d9013SAlfredo Cardigliano ionic_dev_rx_queue_release(rx_queue); 710a27d9013SAlfredo Cardigliano eth_dev->data->rx_queues[rx_queue_id] = NULL; 711a27d9013SAlfredo Cardigliano } 712a27d9013SAlfredo Cardigliano 7139fdf11c4SAndrew Boyer eth_dev->data->rx_queue_state[rx_queue_id] = 7149fdf11c4SAndrew Boyer RTE_ETH_QUEUE_STATE_STOPPED; 7159fdf11c4SAndrew Boyer 716a27d9013SAlfredo Cardigliano err = ionic_rx_qcq_alloc(lif, rx_queue_id, nb_desc, &rxq); 717a27d9013SAlfredo Cardigliano if (err) { 7184ae96cb8SAndrew Boyer IONIC_PRINT(ERR, "Queue %d allocation failure", rx_queue_id); 719a27d9013SAlfredo Cardigliano return -EINVAL; 720a27d9013SAlfredo Cardigliano } 721a27d9013SAlfredo Cardigliano 722a27d9013SAlfredo Cardigliano rxq->mb_pool = mp; 723a27d9013SAlfredo Cardigliano 724a27d9013SAlfredo Cardigliano /* 725a27d9013SAlfredo Cardigliano * Note: the interface does not currently support 726a27d9013SAlfredo Cardigliano * DEV_RX_OFFLOAD_KEEP_CRC, please also consider ETHER_CRC_LEN 727a27d9013SAlfredo Cardigliano * when the adapter will be able to keep the CRC and subtract 728a27d9013SAlfredo Cardigliano * it to the length for all received packets: 729a27d9013SAlfredo Cardigliano * if (eth_dev->data->dev_conf.rxmode.offloads & 730a27d9013SAlfredo Cardigliano * DEV_RX_OFFLOAD_KEEP_CRC) 731a27d9013SAlfredo Cardigliano * rxq->crc_len = ETHER_CRC_LEN; 732a27d9013SAlfredo Cardigliano */ 733a27d9013SAlfredo Cardigliano 734a27d9013SAlfredo Cardigliano /* Do not start queue with rte_eth_dev_start() */ 73502eabf57SAndrew Boyer if (rx_conf->rx_deferred_start) 73602eabf57SAndrew Boyer rxq->flags |= IONIC_QCQ_F_DEFERRED; 737a27d9013SAlfredo Cardigliano 738a27d9013SAlfredo Cardigliano eth_dev->data->rx_queues[rx_queue_id] = rxq; 739a27d9013SAlfredo Cardigliano 740a27d9013SAlfredo Cardigliano return 0; 741a27d9013SAlfredo Cardigliano } 742a27d9013SAlfredo Cardigliano 7430de3e209SAndrew Boyer static __rte_always_inline void 744*c6a9a6fbSAndrew Boyer ionic_rx_clean(struct ionic_qcq *rxq, 745a27d9013SAlfredo Cardigliano uint32_t q_desc_index, uint32_t cq_desc_index, 746700f974dSAndrew Boyer void *service_cb_arg) 747a27d9013SAlfredo Cardigliano { 748*c6a9a6fbSAndrew Boyer struct ionic_queue *q = &rxq->q; 749*c6a9a6fbSAndrew Boyer struct ionic_cq *cq = &rxq->cq; 750*c6a9a6fbSAndrew Boyer struct ionic_rxq_comp *cq_desc_base = cq->base; 751a27d9013SAlfredo Cardigliano struct ionic_rxq_comp *cq_desc = &cq_desc_base[cq_desc_index]; 752700f974dSAndrew Boyer struct rte_mbuf *rxm, *rxm_seg; 753a27d9013SAlfredo Cardigliano uint32_t max_frame_size = 754a27d9013SAlfredo Cardigliano rxq->lif->eth_dev->data->dev_conf.rxmode.max_rx_pkt_len; 755a27d9013SAlfredo Cardigliano uint64_t pkt_flags = 0; 756a27d9013SAlfredo Cardigliano uint32_t pkt_type; 757a27d9013SAlfredo Cardigliano struct ionic_rx_stats *stats = IONIC_Q_TO_RX_STATS(q); 758a27d9013SAlfredo Cardigliano struct ionic_rx_service *recv_args = (struct ionic_rx_service *) 759a27d9013SAlfredo Cardigliano service_cb_arg; 760a27d9013SAlfredo Cardigliano uint32_t buf_size = (uint16_t) 761a27d9013SAlfredo Cardigliano (rte_pktmbuf_data_room_size(rxq->mb_pool) - 762a27d9013SAlfredo Cardigliano RTE_PKTMBUF_HEADROOM); 763a27d9013SAlfredo Cardigliano uint32_t left; 764700f974dSAndrew Boyer void **info; 765700f974dSAndrew Boyer 766700f974dSAndrew Boyer assert(q_desc_index == cq_desc->comp_index); 767700f974dSAndrew Boyer 768700f974dSAndrew Boyer info = IONIC_INFO_PTR(q, cq_desc->comp_index); 769700f974dSAndrew Boyer 770700f974dSAndrew Boyer rxm = info[0]; 771a27d9013SAlfredo Cardigliano 772a27d9013SAlfredo Cardigliano if (!recv_args) { 773a27d9013SAlfredo Cardigliano stats->no_cb_arg++; 774a27d9013SAlfredo Cardigliano /* Flush */ 775a27d9013SAlfredo Cardigliano rte_pktmbuf_free(rxm); 776a27d9013SAlfredo Cardigliano /* 777a27d9013SAlfredo Cardigliano * Note: rte_mempool_put is faster with no segs 778a27d9013SAlfredo Cardigliano * rte_mempool_put(rxq->mb_pool, rxm); 779a27d9013SAlfredo Cardigliano */ 780a27d9013SAlfredo Cardigliano return; 781a27d9013SAlfredo Cardigliano } 782a27d9013SAlfredo Cardigliano 783a27d9013SAlfredo Cardigliano if (cq_desc->status) { 784a27d9013SAlfredo Cardigliano stats->bad_cq_status++; 785a27d9013SAlfredo Cardigliano ionic_rx_recycle(q, q_desc_index, rxm); 786a27d9013SAlfredo Cardigliano return; 787a27d9013SAlfredo Cardigliano } 788a27d9013SAlfredo Cardigliano 789a27d9013SAlfredo Cardigliano if (recv_args->nb_rx >= recv_args->nb_pkts) { 790a27d9013SAlfredo Cardigliano stats->no_room++; 791a27d9013SAlfredo Cardigliano ionic_rx_recycle(q, q_desc_index, rxm); 792a27d9013SAlfredo Cardigliano return; 793a27d9013SAlfredo Cardigliano } 794a27d9013SAlfredo Cardigliano 795a27d9013SAlfredo Cardigliano if (cq_desc->len > max_frame_size || 796a27d9013SAlfredo Cardigliano cq_desc->len == 0) { 797a27d9013SAlfredo Cardigliano stats->bad_len++; 798a27d9013SAlfredo Cardigliano ionic_rx_recycle(q, q_desc_index, rxm); 799a27d9013SAlfredo Cardigliano return; 800a27d9013SAlfredo Cardigliano } 801a27d9013SAlfredo Cardigliano 802a27d9013SAlfredo Cardigliano rxm->data_off = RTE_PKTMBUF_HEADROOM; 803a27d9013SAlfredo Cardigliano rte_prefetch1((char *)rxm->buf_addr + rxm->data_off); 804a27d9013SAlfredo Cardigliano rxm->nb_segs = 1; /* cq_desc->num_sg_elems */ 805a27d9013SAlfredo Cardigliano rxm->pkt_len = cq_desc->len; 806a27d9013SAlfredo Cardigliano rxm->port = rxq->lif->port_id; 807a27d9013SAlfredo Cardigliano 808a27d9013SAlfredo Cardigliano left = cq_desc->len; 809a27d9013SAlfredo Cardigliano 810a27d9013SAlfredo Cardigliano rxm->data_len = RTE_MIN(buf_size, left); 811a27d9013SAlfredo Cardigliano left -= rxm->data_len; 812a27d9013SAlfredo Cardigliano 813a27d9013SAlfredo Cardigliano rxm_seg = rxm->next; 814a27d9013SAlfredo Cardigliano while (rxm_seg && left) { 815a27d9013SAlfredo Cardigliano rxm_seg->data_len = RTE_MIN(buf_size, left); 816a27d9013SAlfredo Cardigliano left -= rxm_seg->data_len; 817a27d9013SAlfredo Cardigliano 818a27d9013SAlfredo Cardigliano rxm_seg = rxm_seg->next; 819a27d9013SAlfredo Cardigliano rxm->nb_segs++; 820a27d9013SAlfredo Cardigliano } 821a27d9013SAlfredo Cardigliano 82222e7171bSAlfredo Cardigliano /* RSS */ 82322e7171bSAlfredo Cardigliano pkt_flags |= PKT_RX_RSS_HASH; 82422e7171bSAlfredo Cardigliano rxm->hash.rss = cq_desc->rss_hash; 82522e7171bSAlfredo Cardigliano 826a27d9013SAlfredo Cardigliano /* Vlan Strip */ 827a27d9013SAlfredo Cardigliano if (cq_desc->csum_flags & IONIC_RXQ_COMP_CSUM_F_VLAN) { 828a27d9013SAlfredo Cardigliano pkt_flags |= PKT_RX_VLAN | PKT_RX_VLAN_STRIPPED; 829a27d9013SAlfredo Cardigliano rxm->vlan_tci = cq_desc->vlan_tci; 830a27d9013SAlfredo Cardigliano } 831a27d9013SAlfredo Cardigliano 832a27d9013SAlfredo Cardigliano /* Checksum */ 833a27d9013SAlfredo Cardigliano if (cq_desc->csum_flags & IONIC_RXQ_COMP_CSUM_F_CALC) { 834a27d9013SAlfredo Cardigliano if (cq_desc->csum_flags & IONIC_RXQ_COMP_CSUM_F_IP_OK) 835a27d9013SAlfredo Cardigliano pkt_flags |= PKT_RX_IP_CKSUM_GOOD; 836a27d9013SAlfredo Cardigliano else if (cq_desc->csum_flags & IONIC_RXQ_COMP_CSUM_F_IP_BAD) 837a27d9013SAlfredo Cardigliano pkt_flags |= PKT_RX_IP_CKSUM_BAD; 838a27d9013SAlfredo Cardigliano 839a27d9013SAlfredo Cardigliano if ((cq_desc->csum_flags & IONIC_RXQ_COMP_CSUM_F_TCP_OK) || 840a27d9013SAlfredo Cardigliano (cq_desc->csum_flags & IONIC_RXQ_COMP_CSUM_F_UDP_OK)) 841a27d9013SAlfredo Cardigliano pkt_flags |= PKT_RX_L4_CKSUM_GOOD; 842a27d9013SAlfredo Cardigliano else if ((cq_desc->csum_flags & 843a27d9013SAlfredo Cardigliano IONIC_RXQ_COMP_CSUM_F_TCP_BAD) || 844a27d9013SAlfredo Cardigliano (cq_desc->csum_flags & 845a27d9013SAlfredo Cardigliano IONIC_RXQ_COMP_CSUM_F_UDP_BAD)) 846a27d9013SAlfredo Cardigliano pkt_flags |= PKT_RX_L4_CKSUM_BAD; 847a27d9013SAlfredo Cardigliano } 848a27d9013SAlfredo Cardigliano 849a27d9013SAlfredo Cardigliano rxm->ol_flags = pkt_flags; 850a27d9013SAlfredo Cardigliano 851a27d9013SAlfredo Cardigliano /* Packet Type */ 852a27d9013SAlfredo Cardigliano switch (cq_desc->pkt_type_color & IONIC_RXQ_COMP_PKT_TYPE_MASK) { 853a27d9013SAlfredo Cardigliano case IONIC_PKT_TYPE_IPV4: 854a27d9013SAlfredo Cardigliano pkt_type = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4; 855a27d9013SAlfredo Cardigliano break; 856a27d9013SAlfredo Cardigliano case IONIC_PKT_TYPE_IPV6: 857a27d9013SAlfredo Cardigliano pkt_type = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6; 858a27d9013SAlfredo Cardigliano break; 859a27d9013SAlfredo Cardigliano case IONIC_PKT_TYPE_IPV4_TCP: 860a27d9013SAlfredo Cardigliano pkt_type = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4 | 861a27d9013SAlfredo Cardigliano RTE_PTYPE_L4_TCP; 862a27d9013SAlfredo Cardigliano break; 863a27d9013SAlfredo Cardigliano case IONIC_PKT_TYPE_IPV6_TCP: 864a27d9013SAlfredo Cardigliano pkt_type = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6 | 865a27d9013SAlfredo Cardigliano RTE_PTYPE_L4_TCP; 866a27d9013SAlfredo Cardigliano break; 867a27d9013SAlfredo Cardigliano case IONIC_PKT_TYPE_IPV4_UDP: 868a27d9013SAlfredo Cardigliano pkt_type = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV4 | 869a27d9013SAlfredo Cardigliano RTE_PTYPE_L4_UDP; 870a27d9013SAlfredo Cardigliano break; 871a27d9013SAlfredo Cardigliano case IONIC_PKT_TYPE_IPV6_UDP: 872a27d9013SAlfredo Cardigliano pkt_type = RTE_PTYPE_L2_ETHER | RTE_PTYPE_L3_IPV6 | 873a27d9013SAlfredo Cardigliano RTE_PTYPE_L4_UDP; 874a27d9013SAlfredo Cardigliano break; 875a27d9013SAlfredo Cardigliano default: 876a27d9013SAlfredo Cardigliano { 877a27d9013SAlfredo Cardigliano struct rte_ether_hdr *eth_h = rte_pktmbuf_mtod(rxm, 878a27d9013SAlfredo Cardigliano struct rte_ether_hdr *); 879a27d9013SAlfredo Cardigliano uint16_t ether_type = eth_h->ether_type; 880a27d9013SAlfredo Cardigliano if (ether_type == rte_cpu_to_be_16(RTE_ETHER_TYPE_ARP)) 881a27d9013SAlfredo Cardigliano pkt_type = RTE_PTYPE_L2_ETHER_ARP; 882a27d9013SAlfredo Cardigliano else 883a27d9013SAlfredo Cardigliano pkt_type = RTE_PTYPE_UNKNOWN; 884a27d9013SAlfredo Cardigliano break; 885a27d9013SAlfredo Cardigliano } 886a27d9013SAlfredo Cardigliano } 887a27d9013SAlfredo Cardigliano 888a27d9013SAlfredo Cardigliano rxm->packet_type = pkt_type; 889a27d9013SAlfredo Cardigliano 890a27d9013SAlfredo Cardigliano recv_args->rx_pkts[recv_args->nb_rx] = rxm; 891a27d9013SAlfredo Cardigliano recv_args->nb_rx++; 892a27d9013SAlfredo Cardigliano 893a27d9013SAlfredo Cardigliano stats->packets++; 894a27d9013SAlfredo Cardigliano stats->bytes += rxm->pkt_len; 895a27d9013SAlfredo Cardigliano } 896a27d9013SAlfredo Cardigliano 897a27d9013SAlfredo Cardigliano static void 898a27d9013SAlfredo Cardigliano ionic_rx_recycle(struct ionic_queue *q, uint32_t q_desc_index, 899a27d9013SAlfredo Cardigliano struct rte_mbuf *mbuf) 900a27d9013SAlfredo Cardigliano { 901a27d9013SAlfredo Cardigliano struct ionic_rxq_desc *desc_base = q->base; 902a27d9013SAlfredo Cardigliano struct ionic_rxq_desc *old = &desc_base[q_desc_index]; 903a27d9013SAlfredo Cardigliano struct ionic_rxq_desc *new = &desc_base[q->head_idx]; 904a27d9013SAlfredo Cardigliano 905a27d9013SAlfredo Cardigliano new->addr = old->addr; 906a27d9013SAlfredo Cardigliano new->len = old->len; 907a27d9013SAlfredo Cardigliano 908700f974dSAndrew Boyer ionic_q_post(q, true, mbuf); 909a27d9013SAlfredo Cardigliano } 910a27d9013SAlfredo Cardigliano 9110de3e209SAndrew Boyer static __rte_always_inline int 912a27d9013SAlfredo Cardigliano ionic_rx_fill(struct ionic_qcq *rxq, uint32_t len) 913a27d9013SAlfredo Cardigliano { 914a27d9013SAlfredo Cardigliano struct ionic_queue *q = &rxq->q; 915a27d9013SAlfredo Cardigliano struct ionic_rxq_desc *desc_base = q->base; 916a27d9013SAlfredo Cardigliano struct ionic_rxq_sg_desc *sg_desc_base = q->sg_base; 917a27d9013SAlfredo Cardigliano struct ionic_rxq_desc *desc; 918a27d9013SAlfredo Cardigliano struct ionic_rxq_sg_desc *sg_desc; 919a27d9013SAlfredo Cardigliano struct ionic_rxq_sg_elem *elem; 920a27d9013SAlfredo Cardigliano rte_iova_t dma_addr; 921a27d9013SAlfredo Cardigliano uint32_t i, j, nsegs, buf_size, size; 922a27d9013SAlfredo Cardigliano bool ring_doorbell; 923a27d9013SAlfredo Cardigliano 924a27d9013SAlfredo Cardigliano buf_size = (uint16_t)(rte_pktmbuf_data_room_size(rxq->mb_pool) - 925a27d9013SAlfredo Cardigliano RTE_PKTMBUF_HEADROOM); 926a27d9013SAlfredo Cardigliano 927a27d9013SAlfredo Cardigliano /* Initialize software ring entries */ 928a27d9013SAlfredo Cardigliano for (i = ionic_q_space_avail(q); i; i--) { 929a27d9013SAlfredo Cardigliano struct rte_mbuf *rxm = rte_mbuf_raw_alloc(rxq->mb_pool); 930a27d9013SAlfredo Cardigliano struct rte_mbuf *prev_rxm_seg; 931a27d9013SAlfredo Cardigliano 932a27d9013SAlfredo Cardigliano if (rxm == NULL) { 933a27d9013SAlfredo Cardigliano IONIC_PRINT(ERR, "RX mbuf alloc failed"); 934a27d9013SAlfredo Cardigliano return -ENOMEM; 935a27d9013SAlfredo Cardigliano } 936a27d9013SAlfredo Cardigliano 937a27d9013SAlfredo Cardigliano nsegs = (len + buf_size - 1) / buf_size; 938a27d9013SAlfredo Cardigliano 939a27d9013SAlfredo Cardigliano desc = &desc_base[q->head_idx]; 940a27d9013SAlfredo Cardigliano dma_addr = rte_cpu_to_le_64(rte_mbuf_data_iova_default(rxm)); 941a27d9013SAlfredo Cardigliano desc->addr = dma_addr; 942a27d9013SAlfredo Cardigliano desc->len = buf_size; 943a27d9013SAlfredo Cardigliano size = buf_size; 944a27d9013SAlfredo Cardigliano desc->opcode = (nsegs > 1) ? IONIC_RXQ_DESC_OPCODE_SG : 945a27d9013SAlfredo Cardigliano IONIC_RXQ_DESC_OPCODE_SIMPLE; 946a27d9013SAlfredo Cardigliano rxm->next = NULL; 947a27d9013SAlfredo Cardigliano 948a27d9013SAlfredo Cardigliano prev_rxm_seg = rxm; 949a27d9013SAlfredo Cardigliano sg_desc = &sg_desc_base[q->head_idx]; 950a27d9013SAlfredo Cardigliano elem = sg_desc->elems; 951a27d9013SAlfredo Cardigliano for (j = 0; j < nsegs - 1 && j < IONIC_RX_MAX_SG_ELEMS; j++) { 952a27d9013SAlfredo Cardigliano struct rte_mbuf *rxm_seg; 953a27d9013SAlfredo Cardigliano rte_iova_t data_iova; 954a27d9013SAlfredo Cardigliano 955a27d9013SAlfredo Cardigliano rxm_seg = rte_mbuf_raw_alloc(rxq->mb_pool); 956a27d9013SAlfredo Cardigliano if (rxm_seg == NULL) { 957a27d9013SAlfredo Cardigliano IONIC_PRINT(ERR, "RX mbuf alloc failed"); 958a27d9013SAlfredo Cardigliano return -ENOMEM; 959a27d9013SAlfredo Cardigliano } 960a27d9013SAlfredo Cardigliano 961a27d9013SAlfredo Cardigliano data_iova = rte_mbuf_data_iova(rxm_seg); 962a27d9013SAlfredo Cardigliano dma_addr = rte_cpu_to_le_64(data_iova); 963a27d9013SAlfredo Cardigliano elem->addr = dma_addr; 964a27d9013SAlfredo Cardigliano elem->len = buf_size; 965a27d9013SAlfredo Cardigliano size += buf_size; 966a27d9013SAlfredo Cardigliano elem++; 967a27d9013SAlfredo Cardigliano rxm_seg->next = NULL; 968a27d9013SAlfredo Cardigliano prev_rxm_seg->next = rxm_seg; 969a27d9013SAlfredo Cardigliano prev_rxm_seg = rxm_seg; 970a27d9013SAlfredo Cardigliano } 971a27d9013SAlfredo Cardigliano 972a27d9013SAlfredo Cardigliano if (size < len) 973a27d9013SAlfredo Cardigliano IONIC_PRINT(ERR, "Rx SG size is not sufficient (%d < %d)", 974a27d9013SAlfredo Cardigliano size, len); 975a27d9013SAlfredo Cardigliano 976a27d9013SAlfredo Cardigliano ring_doorbell = ((q->head_idx + 1) & 977a27d9013SAlfredo Cardigliano IONIC_RX_RING_DOORBELL_STRIDE) == 0; 978a27d9013SAlfredo Cardigliano 979700f974dSAndrew Boyer ionic_q_post(q, ring_doorbell, rxm); 980a27d9013SAlfredo Cardigliano } 981a27d9013SAlfredo Cardigliano 982a27d9013SAlfredo Cardigliano return 0; 983a27d9013SAlfredo Cardigliano } 984a27d9013SAlfredo Cardigliano 985a27d9013SAlfredo Cardigliano /* 986a27d9013SAlfredo Cardigliano * Start Receive Units for specified queue. 987a27d9013SAlfredo Cardigliano */ 988ce6427ddSThomas Monjalon int __rte_cold 989a27d9013SAlfredo Cardigliano ionic_dev_rx_queue_start(struct rte_eth_dev *eth_dev, uint16_t rx_queue_id) 990a27d9013SAlfredo Cardigliano { 991a27d9013SAlfredo Cardigliano uint32_t frame_size = eth_dev->data->dev_conf.rxmode.max_rx_pkt_len; 9929fdf11c4SAndrew Boyer uint8_t *rx_queue_state = eth_dev->data->rx_queue_state; 993a27d9013SAlfredo Cardigliano struct ionic_qcq *rxq; 994a27d9013SAlfredo Cardigliano int err; 995a27d9013SAlfredo Cardigliano 9969fdf11c4SAndrew Boyer if (rx_queue_state[rx_queue_id] == RTE_ETH_QUEUE_STATE_STARTED) { 9979fdf11c4SAndrew Boyer IONIC_PRINT(DEBUG, "RX queue %u already started", 9989fdf11c4SAndrew Boyer rx_queue_id); 9999fdf11c4SAndrew Boyer return 0; 10009fdf11c4SAndrew Boyer } 10019fdf11c4SAndrew Boyer 1002a27d9013SAlfredo Cardigliano rxq = eth_dev->data->rx_queues[rx_queue_id]; 1003a27d9013SAlfredo Cardigliano 10044ae96cb8SAndrew Boyer IONIC_PRINT(DEBUG, "Starting RX queue %u, %u descs (size: %u)", 10054ae96cb8SAndrew Boyer rx_queue_id, rxq->q.num_descs, frame_size); 10064ae96cb8SAndrew Boyer 10079fdf11c4SAndrew Boyer if (!(rxq->flags & IONIC_QCQ_F_INITED)) { 1008a27d9013SAlfredo Cardigliano err = ionic_lif_rxq_init(rxq); 1009a27d9013SAlfredo Cardigliano if (err) 1010a27d9013SAlfredo Cardigliano return err; 1011b5d9a4f0SAndrew Boyer } else { 1012b5d9a4f0SAndrew Boyer ionic_qcq_enable(rxq); 10139fdf11c4SAndrew Boyer } 1014a27d9013SAlfredo Cardigliano 1015a27d9013SAlfredo Cardigliano /* Allocate buffers for descriptor rings */ 1016a27d9013SAlfredo Cardigliano if (ionic_rx_fill(rxq, frame_size) != 0) { 1017a27d9013SAlfredo Cardigliano IONIC_PRINT(ERR, "Could not alloc mbuf for queue:%d", 1018a27d9013SAlfredo Cardigliano rx_queue_id); 1019a27d9013SAlfredo Cardigliano return -1; 1020a27d9013SAlfredo Cardigliano } 1021a27d9013SAlfredo Cardigliano 10229fdf11c4SAndrew Boyer rx_queue_state[rx_queue_id] = RTE_ETH_QUEUE_STATE_STARTED; 1023a27d9013SAlfredo Cardigliano 1024a27d9013SAlfredo Cardigliano return 0; 1025a27d9013SAlfredo Cardigliano } 1026a27d9013SAlfredo Cardigliano 10270de3e209SAndrew Boyer static __rte_always_inline void 10282aed9865SAndrew Boyer ionic_rxq_service(struct ionic_qcq *rxq, uint32_t work_to_do, 1029a27d9013SAlfredo Cardigliano void *service_cb_arg) 1030a27d9013SAlfredo Cardigliano { 10312aed9865SAndrew Boyer struct ionic_cq *cq = &rxq->cq; 10322aed9865SAndrew Boyer struct ionic_queue *q = &rxq->q; 1033a27d9013SAlfredo Cardigliano struct ionic_rxq_comp *cq_desc_base = cq->base; 1034a27d9013SAlfredo Cardigliano struct ionic_rxq_comp *cq_desc; 1035a27d9013SAlfredo Cardigliano bool more; 1036a27d9013SAlfredo Cardigliano uint32_t curr_q_tail_idx, curr_cq_tail_idx; 1037a27d9013SAlfredo Cardigliano uint32_t work_done = 0; 1038a27d9013SAlfredo Cardigliano 1039a27d9013SAlfredo Cardigliano if (work_to_do == 0) 1040a27d9013SAlfredo Cardigliano return; 1041a27d9013SAlfredo Cardigliano 1042a27d9013SAlfredo Cardigliano cq_desc = &cq_desc_base[cq->tail_idx]; 1043a27d9013SAlfredo Cardigliano while (color_match(cq_desc->pkt_type_color, cq->done_color)) { 1044a27d9013SAlfredo Cardigliano curr_cq_tail_idx = cq->tail_idx; 10452aed9865SAndrew Boyer cq->tail_idx = Q_NEXT_TO_SRVC(cq, 1); 1046a27d9013SAlfredo Cardigliano 1047a27d9013SAlfredo Cardigliano if (cq->tail_idx == 0) 1048a27d9013SAlfredo Cardigliano cq->done_color = !cq->done_color; 1049a27d9013SAlfredo Cardigliano 1050a27d9013SAlfredo Cardigliano /* Prefetch the next 4 descriptors */ 1051a27d9013SAlfredo Cardigliano if ((cq->tail_idx & 0x3) == 0) 1052a27d9013SAlfredo Cardigliano rte_prefetch0(&cq_desc_base[cq->tail_idx]); 1053a27d9013SAlfredo Cardigliano 1054a27d9013SAlfredo Cardigliano do { 1055a27d9013SAlfredo Cardigliano more = (q->tail_idx != cq_desc->comp_index); 1056a27d9013SAlfredo Cardigliano 1057a27d9013SAlfredo Cardigliano curr_q_tail_idx = q->tail_idx; 1058a27d9013SAlfredo Cardigliano q->tail_idx = (q->tail_idx + 1) & (q->num_descs - 1); 1059a27d9013SAlfredo Cardigliano 1060a27d9013SAlfredo Cardigliano /* Prefetch the next 4 descriptors */ 1061a27d9013SAlfredo Cardigliano if ((q->tail_idx & 0x3) == 0) 1062a27d9013SAlfredo Cardigliano /* q desc info */ 1063a27d9013SAlfredo Cardigliano rte_prefetch0(&q->info[q->tail_idx]); 1064a27d9013SAlfredo Cardigliano 1065*c6a9a6fbSAndrew Boyer ionic_rx_clean(rxq, curr_q_tail_idx, curr_cq_tail_idx, 1066700f974dSAndrew Boyer service_cb_arg); 1067a27d9013SAlfredo Cardigliano 1068a27d9013SAlfredo Cardigliano } while (more); 1069a27d9013SAlfredo Cardigliano 1070a27d9013SAlfredo Cardigliano if (++work_done == work_to_do) 1071a27d9013SAlfredo Cardigliano break; 1072a27d9013SAlfredo Cardigliano 1073a27d9013SAlfredo Cardigliano cq_desc = &cq_desc_base[cq->tail_idx]; 1074a27d9013SAlfredo Cardigliano } 1075a27d9013SAlfredo Cardigliano } 1076a27d9013SAlfredo Cardigliano 1077a27d9013SAlfredo Cardigliano /* 1078a27d9013SAlfredo Cardigliano * Stop Receive Units for specified queue. 1079a27d9013SAlfredo Cardigliano */ 1080ce6427ddSThomas Monjalon int __rte_cold 1081a27d9013SAlfredo Cardigliano ionic_dev_rx_queue_stop(struct rte_eth_dev *eth_dev, uint16_t rx_queue_id) 1082a27d9013SAlfredo Cardigliano { 1083a27d9013SAlfredo Cardigliano struct ionic_qcq *rxq; 1084a27d9013SAlfredo Cardigliano 10854ae96cb8SAndrew Boyer IONIC_PRINT(DEBUG, "Stopping RX queue %u", rx_queue_id); 1086a27d9013SAlfredo Cardigliano 1087a27d9013SAlfredo Cardigliano rxq = eth_dev->data->rx_queues[rx_queue_id]; 1088a27d9013SAlfredo Cardigliano 10899fdf11c4SAndrew Boyer eth_dev->data->rx_queue_state[rx_queue_id] = 10909fdf11c4SAndrew Boyer RTE_ETH_QUEUE_STATE_STOPPED; 10919fdf11c4SAndrew Boyer 1092a27d9013SAlfredo Cardigliano ionic_qcq_disable(rxq); 1093a27d9013SAlfredo Cardigliano 1094a27d9013SAlfredo Cardigliano /* Flush */ 10952aed9865SAndrew Boyer ionic_rxq_service(rxq, -1, NULL); 1096a27d9013SAlfredo Cardigliano 1097a27d9013SAlfredo Cardigliano return 0; 1098a27d9013SAlfredo Cardigliano } 1099a27d9013SAlfredo Cardigliano 1100a27d9013SAlfredo Cardigliano uint16_t 1101a27d9013SAlfredo Cardigliano ionic_recv_pkts(void *rx_queue, struct rte_mbuf **rx_pkts, 1102a27d9013SAlfredo Cardigliano uint16_t nb_pkts) 1103a27d9013SAlfredo Cardigliano { 1104a27d9013SAlfredo Cardigliano struct ionic_qcq *rxq = (struct ionic_qcq *)rx_queue; 1105a27d9013SAlfredo Cardigliano uint32_t frame_size = 1106a27d9013SAlfredo Cardigliano rxq->lif->eth_dev->data->dev_conf.rxmode.max_rx_pkt_len; 1107a27d9013SAlfredo Cardigliano struct ionic_rx_service service_cb_arg; 1108a27d9013SAlfredo Cardigliano 1109a27d9013SAlfredo Cardigliano service_cb_arg.rx_pkts = rx_pkts; 1110a27d9013SAlfredo Cardigliano service_cb_arg.nb_pkts = nb_pkts; 1111a27d9013SAlfredo Cardigliano service_cb_arg.nb_rx = 0; 1112a27d9013SAlfredo Cardigliano 11132aed9865SAndrew Boyer ionic_rxq_service(rxq, nb_pkts, &service_cb_arg); 1114a27d9013SAlfredo Cardigliano 1115a27d9013SAlfredo Cardigliano ionic_rx_fill(rxq, frame_size); 1116a27d9013SAlfredo Cardigliano 1117a27d9013SAlfredo Cardigliano return service_cb_arg.nb_rx; 1118a27d9013SAlfredo Cardigliano } 1119