1 /* SPDX-License-Identifier: BSD-3-Clause 2 * Copyright 2018-2019 NXP 3 */ 4 5 #include <stdio.h> 6 #include <errno.h> 7 #include <stdint.h> 8 9 #include <rte_bus_vdev.h> 10 #include <rte_atomic.h> 11 #include <rte_interrupts.h> 12 #include <rte_branch_prediction.h> 13 #include <rte_lcore.h> 14 15 #include <rte_rawdev.h> 16 #include <rte_rawdev_pmd.h> 17 18 #include <portal/dpaa2_hw_pvt.h> 19 #include <portal/dpaa2_hw_dpio.h> 20 #include "dpaa2_cmdif_logs.h" 21 #include "rte_pmd_dpaa2_cmdif.h" 22 23 /* Dynamic log type identifier */ 24 int dpaa2_cmdif_logtype; 25 26 /* CMDIF driver name */ 27 #define DPAA2_CMDIF_PMD_NAME dpaa2_dpci 28 29 /* 30 * This API provides the DPCI device ID in 'attr_value'. 31 * The device ID shall be passed by GPP to the AIOP using CMDIF commands. 32 */ 33 static int 34 dpaa2_cmdif_get_attr(struct rte_rawdev *dev, 35 const char *attr_name, 36 uint64_t *attr_value) 37 { 38 struct dpaa2_dpci_dev *cidev = dev->dev_private; 39 40 DPAA2_CMDIF_FUNC_TRACE(); 41 42 RTE_SET_USED(attr_name); 43 44 if (!attr_value) { 45 DPAA2_CMDIF_ERR("Invalid arguments for getting attributes"); 46 return -EINVAL; 47 } 48 *attr_value = cidev->dpci_id; 49 50 return 0; 51 } 52 53 static int 54 dpaa2_cmdif_enqueue_bufs(struct rte_rawdev *dev, 55 struct rte_rawdev_buf **buffers, 56 unsigned int count, 57 rte_rawdev_obj_t context) 58 { 59 struct dpaa2_dpci_dev *cidev = dev->dev_private; 60 struct rte_dpaa2_cmdif_context *cmdif_send_cnxt; 61 struct dpaa2_queue *txq; 62 struct qbman_fd fd; 63 struct qbman_eq_desc eqdesc; 64 struct qbman_swp *swp; 65 uint32_t retry_count = 0; 66 int ret; 67 68 RTE_SET_USED(count); 69 70 if (unlikely(!DPAA2_PER_LCORE_DPIO)) { 71 ret = dpaa2_affine_qbman_swp(); 72 if (ret) { 73 DPAA2_CMDIF_ERR("Failure in affining portal\n"); 74 return 0; 75 } 76 } 77 swp = DPAA2_PER_LCORE_PORTAL; 78 79 cmdif_send_cnxt = (struct rte_dpaa2_cmdif_context *)(context); 80 txq = &(cidev->tx_queue[cmdif_send_cnxt->priority]); 81 82 /* Prepare enqueue descriptor */ 83 qbman_eq_desc_clear(&eqdesc); 84 qbman_eq_desc_set_fq(&eqdesc, txq->fqid); 85 qbman_eq_desc_set_no_orp(&eqdesc, 0); 86 qbman_eq_desc_set_response(&eqdesc, 0, 0); 87 88 /* Set some of the FD parameters to i. 89 * For performance reasons do not memset 90 */ 91 fd.simple.bpid_offset = 0; 92 fd.simple.ctrl = 0; 93 94 DPAA2_SET_FD_ADDR(&fd, DPAA2_VADDR_TO_IOVA(buffers[0]->buf_addr)); 95 DPAA2_SET_FD_LEN(&fd, cmdif_send_cnxt->size); 96 DPAA2_SET_FD_FRC(&fd, cmdif_send_cnxt->frc); 97 DPAA2_SET_FD_FLC(&fd, cmdif_send_cnxt->flc); 98 99 /* Enqueue a packet to the QBMAN */ 100 do { 101 ret = qbman_swp_enqueue_multiple(swp, &eqdesc, &fd, NULL, 1); 102 if (ret < 0 && ret != -EBUSY) 103 DPAA2_CMDIF_ERR("Transmit failure with err: %d\n", ret); 104 retry_count++; 105 } while ((ret == -EBUSY) && (retry_count < DPAA2_MAX_TX_RETRY_COUNT)); 106 107 if (ret < 0) 108 return ret; 109 110 DPAA2_CMDIF_DP_DEBUG("Successfully transmitted a packet\n"); 111 112 return 1; 113 } 114 115 static int 116 dpaa2_cmdif_dequeue_bufs(struct rte_rawdev *dev, 117 struct rte_rawdev_buf **buffers, 118 unsigned int count, 119 rte_rawdev_obj_t context) 120 { 121 struct dpaa2_dpci_dev *cidev = dev->dev_private; 122 struct rte_dpaa2_cmdif_context *cmdif_rcv_cnxt; 123 struct dpaa2_queue *rxq; 124 struct qbman_swp *swp; 125 struct qbman_result *dq_storage; 126 const struct qbman_fd *fd; 127 struct qbman_pull_desc pulldesc; 128 uint8_t status; 129 int ret; 130 131 RTE_SET_USED(count); 132 133 if (unlikely(!DPAA2_PER_LCORE_DPIO)) { 134 ret = dpaa2_affine_qbman_swp(); 135 if (ret) { 136 DPAA2_CMDIF_ERR("Failure in affining portal\n"); 137 return 0; 138 } 139 } 140 swp = DPAA2_PER_LCORE_PORTAL; 141 142 cmdif_rcv_cnxt = (struct rte_dpaa2_cmdif_context *)(context); 143 rxq = &(cidev->rx_queue[cmdif_rcv_cnxt->priority]); 144 dq_storage = rxq->q_storage->dq_storage[0]; 145 146 qbman_pull_desc_clear(&pulldesc); 147 qbman_pull_desc_set_fq(&pulldesc, rxq->fqid); 148 qbman_pull_desc_set_numframes(&pulldesc, 1); 149 qbman_pull_desc_set_storage(&pulldesc, dq_storage, 150 (uint64_t)(DPAA2_VADDR_TO_IOVA(dq_storage)), 1); 151 152 while (1) { 153 if (qbman_swp_pull(swp, &pulldesc)) { 154 DPAA2_CMDIF_DP_WARN("VDQ cmd not issued. QBMAN is busy\n"); 155 /* Portal was busy, try again */ 156 continue; 157 } 158 break; 159 } 160 161 /* Check if previous issued command is completed. */ 162 while (!qbman_check_command_complete(dq_storage)) 163 ; 164 /* Loop until the dq_storage is updated with new token by QBMAN */ 165 while (!qbman_result_has_new_result(swp, dq_storage)) 166 ; 167 168 /* Check for valid frame. */ 169 status = (uint8_t)qbman_result_DQ_flags(dq_storage); 170 if (unlikely((status & QBMAN_DQ_STAT_VALIDFRAME) == 0)) { 171 DPAA2_CMDIF_DP_DEBUG("No frame is delivered\n"); 172 return 0; 173 } 174 175 fd = qbman_result_DQ_fd(dq_storage); 176 177 buffers[0]->buf_addr = (void *)DPAA2_IOVA_TO_VADDR( 178 DPAA2_GET_FD_ADDR(fd) + DPAA2_GET_FD_OFFSET(fd)); 179 cmdif_rcv_cnxt->size = DPAA2_GET_FD_LEN(fd); 180 cmdif_rcv_cnxt->flc = DPAA2_GET_FD_FLC(fd); 181 cmdif_rcv_cnxt->frc = DPAA2_GET_FD_FRC(fd); 182 183 DPAA2_CMDIF_DP_DEBUG("packet received\n"); 184 185 return 1; 186 } 187 188 static const struct rte_rawdev_ops dpaa2_cmdif_ops = { 189 .attr_get = dpaa2_cmdif_get_attr, 190 .enqueue_bufs = dpaa2_cmdif_enqueue_bufs, 191 .dequeue_bufs = dpaa2_cmdif_dequeue_bufs, 192 }; 193 194 static int 195 dpaa2_cmdif_create(const char *name, 196 struct rte_vdev_device *vdev, 197 int socket_id) 198 { 199 struct rte_rawdev *rawdev; 200 struct dpaa2_dpci_dev *cidev; 201 202 /* Allocate device structure */ 203 rawdev = rte_rawdev_pmd_allocate(name, sizeof(struct dpaa2_dpci_dev), 204 socket_id); 205 if (!rawdev) { 206 DPAA2_CMDIF_ERR("Unable to allocate rawdevice"); 207 return -EINVAL; 208 } 209 210 rawdev->dev_ops = &dpaa2_cmdif_ops; 211 rawdev->device = &vdev->device; 212 213 /* For secondary processes, the primary has done all the work */ 214 if (rte_eal_process_type() != RTE_PROC_PRIMARY) 215 return 0; 216 217 cidev = rte_dpaa2_alloc_dpci_dev(); 218 if (!cidev) { 219 DPAA2_CMDIF_ERR("Unable to allocate CI device"); 220 rte_rawdev_pmd_release(rawdev); 221 return -ENODEV; 222 } 223 224 rawdev->dev_private = cidev; 225 226 return 0; 227 } 228 229 static int 230 dpaa2_cmdif_destroy(const char *name) 231 { 232 int ret; 233 struct rte_rawdev *rdev; 234 235 rdev = rte_rawdev_pmd_get_named_dev(name); 236 if (!rdev) { 237 DPAA2_CMDIF_ERR("Invalid device name (%s)", name); 238 return -EINVAL; 239 } 240 241 /* The primary process will only free the DPCI device */ 242 if (rte_eal_process_type() == RTE_PROC_PRIMARY) 243 rte_dpaa2_free_dpci_dev(rdev->dev_private); 244 245 ret = rte_rawdev_pmd_release(rdev); 246 if (ret) 247 DPAA2_CMDIF_DEBUG("Device cleanup failed"); 248 249 return 0; 250 } 251 252 static int 253 dpaa2_cmdif_probe(struct rte_vdev_device *vdev) 254 { 255 const char *name; 256 int ret = 0; 257 258 name = rte_vdev_device_name(vdev); 259 260 DPAA2_CMDIF_INFO("Init %s on NUMA node %d", name, rte_socket_id()); 261 262 ret = dpaa2_cmdif_create(name, vdev, rte_socket_id()); 263 264 return ret; 265 } 266 267 static int 268 dpaa2_cmdif_remove(struct rte_vdev_device *vdev) 269 { 270 const char *name; 271 int ret; 272 273 name = rte_vdev_device_name(vdev); 274 if (name == NULL) 275 return -1; 276 277 DPAA2_CMDIF_INFO("Closing %s on NUMA node %d", name, rte_socket_id()); 278 279 ret = dpaa2_cmdif_destroy(name); 280 281 return ret; 282 } 283 284 static struct rte_vdev_driver dpaa2_cmdif_drv = { 285 .probe = dpaa2_cmdif_probe, 286 .remove = dpaa2_cmdif_remove 287 }; 288 289 RTE_PMD_REGISTER_VDEV(DPAA2_CMDIF_PMD_NAME, dpaa2_cmdif_drv); 290 291 RTE_INIT(dpaa2_cmdif_init_log) 292 { 293 dpaa2_cmdif_logtype = rte_log_register("pmd.raw.dpaa2.cmdif"); 294 if (dpaa2_cmdif_logtype >= 0) 295 rte_log_set_level(dpaa2_cmdif_logtype, RTE_LOG_INFO); 296 } 297