1 /* SPDX-License-Identifier: BSD-3-Clause 2 * Copyright 2018 NXP 3 */ 4 5 #include <stdio.h> 6 #include <errno.h> 7 #include <stdint.h> 8 9 #include <rte_bus_vdev.h> 10 #include <rte_atomic.h> 11 #include <rte_interrupts.h> 12 #include <rte_branch_prediction.h> 13 #include <rte_lcore.h> 14 15 #include <rte_rawdev.h> 16 #include <rte_rawdev_pmd.h> 17 18 #include <portal/dpaa2_hw_pvt.h> 19 #include <portal/dpaa2_hw_dpio.h> 20 #include "dpaa2_cmdif_logs.h" 21 #include "rte_pmd_dpaa2_cmdif.h" 22 23 /* Dynamic log type identifier */ 24 int dpaa2_cmdif_logtype; 25 26 /* CMDIF driver name */ 27 #define DPAA2_CMDIF_PMD_NAME dpaa2_dpci 28 29 /* CMDIF driver object */ 30 static struct rte_vdev_driver dpaa2_cmdif_drv; 31 32 /* 33 * This API provides the DPCI device ID in 'attr_value'. 34 * The device ID shall be passed by GPP to the AIOP using CMDIF commands. 35 */ 36 static int 37 dpaa2_cmdif_get_attr(struct rte_rawdev *dev, 38 const char *attr_name, 39 uint64_t *attr_value) 40 { 41 struct dpaa2_dpci_dev *cidev = dev->dev_private; 42 43 DPAA2_CMDIF_FUNC_TRACE(); 44 45 RTE_SET_USED(attr_name); 46 47 if (!attr_value) { 48 DPAA2_CMDIF_ERR("Invalid arguments for getting attributes"); 49 return -EINVAL; 50 } 51 *attr_value = cidev->dpci_id; 52 53 return 0; 54 } 55 56 static int 57 dpaa2_cmdif_enqueue_bufs(struct rte_rawdev *dev, 58 struct rte_rawdev_buf **buffers, 59 unsigned int count, 60 rte_rawdev_obj_t context) 61 { 62 struct dpaa2_dpci_dev *cidev = dev->dev_private; 63 struct rte_dpaa2_cmdif_context *cmdif_send_cnxt; 64 struct dpaa2_queue *txq; 65 struct qbman_fd fd; 66 struct qbman_eq_desc eqdesc; 67 struct qbman_swp *swp; 68 int ret; 69 70 RTE_SET_USED(count); 71 72 if (unlikely(!DPAA2_PER_LCORE_DPIO)) { 73 ret = dpaa2_affine_qbman_swp(); 74 if (ret) { 75 DPAA2_CMDIF_ERR("Failure in affining portal\n"); 76 return 0; 77 } 78 } 79 swp = DPAA2_PER_LCORE_PORTAL; 80 81 cmdif_send_cnxt = (struct rte_dpaa2_cmdif_context *)(context); 82 txq = &(cidev->tx_queue[cmdif_send_cnxt->priority]); 83 84 /* Prepare enqueue descriptor */ 85 qbman_eq_desc_clear(&eqdesc); 86 qbman_eq_desc_set_fq(&eqdesc, txq->fqid); 87 qbman_eq_desc_set_no_orp(&eqdesc, 0); 88 qbman_eq_desc_set_response(&eqdesc, 0, 0); 89 90 /* Set some of the FD parameters to i. 91 * For performance reasons do not memset 92 */ 93 fd.simple.bpid_offset = 0; 94 fd.simple.ctrl = 0; 95 96 DPAA2_SET_FD_ADDR(&fd, DPAA2_VADDR_TO_IOVA(buffers[0]->buf_addr)); 97 DPAA2_SET_FD_LEN(&fd, cmdif_send_cnxt->size); 98 DPAA2_SET_FD_FRC(&fd, cmdif_send_cnxt->frc); 99 DPAA2_SET_FD_FLC(&fd, cmdif_send_cnxt->flc); 100 101 /* Enqueue a packet to the QBMAN */ 102 do { 103 ret = qbman_swp_enqueue_multiple(swp, &eqdesc, &fd, NULL, 1); 104 if (ret < 0 && ret != -EBUSY) 105 DPAA2_CMDIF_ERR("Transmit failure with err: %d\n", ret); 106 } while (ret == -EBUSY); 107 108 DPAA2_CMDIF_DP_DEBUG("Successfully transmitted a packet\n"); 109 110 return 0; 111 } 112 113 static int 114 dpaa2_cmdif_dequeue_bufs(struct rte_rawdev *dev, 115 struct rte_rawdev_buf **buffers, 116 unsigned int count, 117 rte_rawdev_obj_t context) 118 { 119 struct dpaa2_dpci_dev *cidev = dev->dev_private; 120 struct rte_dpaa2_cmdif_context *cmdif_rcv_cnxt; 121 struct dpaa2_queue *rxq; 122 struct qbman_swp *swp; 123 struct qbman_result *dq_storage; 124 const struct qbman_fd *fd; 125 struct qbman_pull_desc pulldesc; 126 uint8_t status; 127 int ret; 128 129 RTE_SET_USED(count); 130 131 if (unlikely(!DPAA2_PER_LCORE_DPIO)) { 132 ret = dpaa2_affine_qbman_swp(); 133 if (ret) { 134 DPAA2_CMDIF_ERR("Failure in affining portal\n"); 135 return 0; 136 } 137 } 138 swp = DPAA2_PER_LCORE_PORTAL; 139 140 cmdif_rcv_cnxt = (struct rte_dpaa2_cmdif_context *)(context); 141 rxq = &(cidev->rx_queue[cmdif_rcv_cnxt->priority]); 142 dq_storage = rxq->q_storage->dq_storage[0]; 143 144 qbman_pull_desc_clear(&pulldesc); 145 qbman_pull_desc_set_fq(&pulldesc, rxq->fqid); 146 qbman_pull_desc_set_numframes(&pulldesc, 1); 147 qbman_pull_desc_set_storage(&pulldesc, dq_storage, 148 (uint64_t)(DPAA2_VADDR_TO_IOVA(dq_storage)), 1); 149 150 while (1) { 151 if (qbman_swp_pull(swp, &pulldesc)) { 152 DPAA2_CMDIF_DP_WARN("VDQ cmd not issued. QBMAN is busy\n"); 153 /* Portal was busy, try again */ 154 continue; 155 } 156 break; 157 } 158 159 /* Check if previous issued command is completed. */ 160 while (!qbman_check_command_complete(dq_storage)) 161 ; 162 /* Loop until the dq_storage is updated with new token by QBMAN */ 163 while (!qbman_result_has_new_result(swp, dq_storage)) 164 ; 165 166 /* Check for valid frame. */ 167 status = (uint8_t)qbman_result_DQ_flags(dq_storage); 168 if (unlikely((status & QBMAN_DQ_STAT_VALIDFRAME) == 0)) { 169 DPAA2_CMDIF_DP_DEBUG("No frame is delivered\n"); 170 return 0; 171 } 172 173 fd = qbman_result_DQ_fd(dq_storage); 174 175 buffers[0]->buf_addr = (void *)DPAA2_IOVA_TO_VADDR( 176 DPAA2_GET_FD_ADDR(fd) + DPAA2_GET_FD_OFFSET(fd)); 177 cmdif_rcv_cnxt->size = DPAA2_GET_FD_LEN(fd); 178 cmdif_rcv_cnxt->flc = DPAA2_GET_FD_FLC(fd); 179 cmdif_rcv_cnxt->frc = DPAA2_GET_FD_FRC(fd); 180 181 DPAA2_CMDIF_DP_DEBUG("packet received\n"); 182 183 return 1; 184 } 185 186 static const struct rte_rawdev_ops dpaa2_cmdif_ops = { 187 .attr_get = dpaa2_cmdif_get_attr, 188 .enqueue_bufs = dpaa2_cmdif_enqueue_bufs, 189 .dequeue_bufs = dpaa2_cmdif_dequeue_bufs, 190 }; 191 192 static int 193 dpaa2_cmdif_create(const char *name, 194 struct rte_vdev_device *vdev, 195 int socket_id) 196 { 197 struct rte_rawdev *rawdev; 198 struct dpaa2_dpci_dev *cidev; 199 200 /* Allocate device structure */ 201 rawdev = rte_rawdev_pmd_allocate(name, sizeof(struct dpaa2_dpci_dev), 202 socket_id); 203 if (!rawdev) { 204 DPAA2_CMDIF_ERR("Unable to allocate rawdevice"); 205 return -EINVAL; 206 } 207 208 rawdev->dev_ops = &dpaa2_cmdif_ops; 209 rawdev->device = &vdev->device; 210 rawdev->driver_name = vdev->device.driver->name; 211 212 /* For secondary processes, the primary has done all the work */ 213 if (rte_eal_process_type() != RTE_PROC_PRIMARY) 214 return 0; 215 216 cidev = rte_dpaa2_alloc_dpci_dev(); 217 if (!cidev) { 218 DPAA2_CMDIF_ERR("Unable to allocate CI device"); 219 rte_rawdev_pmd_release(rawdev); 220 return -ENODEV; 221 } 222 223 rawdev->dev_private = cidev; 224 225 return 0; 226 } 227 228 static int 229 dpaa2_cmdif_destroy(const char *name) 230 { 231 int ret; 232 struct rte_rawdev *rdev; 233 234 rdev = rte_rawdev_pmd_get_named_dev(name); 235 if (!rdev) { 236 DPAA2_CMDIF_ERR("Invalid device name (%s)", name); 237 return -EINVAL; 238 } 239 240 /* The primary process will only free the DPCI device */ 241 if (rte_eal_process_type() == RTE_PROC_PRIMARY) 242 rte_dpaa2_free_dpci_dev(rdev->dev_private); 243 244 ret = rte_rawdev_pmd_release(rdev); 245 if (ret) 246 DPAA2_CMDIF_DEBUG("Device cleanup failed"); 247 248 return 0; 249 } 250 251 static int 252 dpaa2_cmdif_probe(struct rte_vdev_device *vdev) 253 { 254 const char *name; 255 int ret = 0; 256 257 name = rte_vdev_device_name(vdev); 258 259 DPAA2_CMDIF_INFO("Init %s on NUMA node %d", name, rte_socket_id()); 260 261 ret = dpaa2_cmdif_create(name, vdev, rte_socket_id()); 262 263 return ret; 264 } 265 266 static int 267 dpaa2_cmdif_remove(struct rte_vdev_device *vdev) 268 { 269 const char *name; 270 int ret; 271 272 name = rte_vdev_device_name(vdev); 273 274 DPAA2_CMDIF_INFO("Closing %s on NUMA node %d", name, rte_socket_id()); 275 276 ret = dpaa2_cmdif_destroy(name); 277 278 return ret; 279 } 280 281 static struct rte_vdev_driver dpaa2_cmdif_drv = { 282 .probe = dpaa2_cmdif_probe, 283 .remove = dpaa2_cmdif_remove 284 }; 285 286 RTE_PMD_REGISTER_VDEV(DPAA2_CMDIF_PMD_NAME, dpaa2_cmdif_drv); 287 288 RTE_INIT(dpaa2_cmdif_init_log) 289 { 290 dpaa2_cmdif_logtype = rte_log_register("pmd.raw.dpaa2.cmdif"); 291 if (dpaa2_cmdif_logtype >= 0) 292 rte_log_set_level(dpaa2_cmdif_logtype, RTE_LOG_INFO); 293 } 294