/* SPDX-License-Identifier: BSD-3-Clause * Copyright(c) 2001-2023 Intel Corporation */ #include "cpfl_controlq.h" #include "base/idpf_controlq.h" #include "rte_common.h" /** * cpfl_check_dma_mem_parameters - verify DMA memory params from CP * @qinfo: pointer to create control queue info struct * * Verify that DMA parameter of each DMA memory struct is present and * consistent with control queue parameters */ static inline int cpfl_check_dma_mem_parameters(struct cpfl_ctlq_create_info *qinfo) { struct idpf_dma_mem *ring = &qinfo->ring_mem; struct idpf_dma_mem *buf = &qinfo->buf_mem; if (!ring->va || !ring->size) return -EINVAL; if (ring->size != qinfo->len * sizeof(struct idpf_ctlq_desc)) return -EINVAL; /* no need for buffer checks for TX queues */ if (qinfo->type == IDPF_CTLQ_TYPE_MAILBOX_TX || qinfo->type == IDPF_CTLQ_TYPE_CONFIG_TX || qinfo->type == IDPF_CTLQ_TYPE_RDMA_TX) return 0; if (!buf->va || !buf->size) return -EINVAL; /* accommodate different types of rx ring buffer sizes */ if ((qinfo->type == IDPF_CTLQ_TYPE_MAILBOX_RX && buf->size != CPFL_CTLQ_MAILBOX_BUFFER_SIZE * qinfo->len) || (qinfo->type == IDPF_CTLQ_TYPE_CONFIG_RX && buf->size != CPFL_CFGQ_RING_LEN * CPFL_CTLQ_CFGQ_BUFFER_SIZE)) return -EINVAL; return 0; } /** * cpfl_ctlq_alloc_ring_res - store memory for descriptor ring and bufs * @hw: pointer to hw struct * @cq: pointer to control queue struct * @qinfo: pointer to create queue info struct * * The CP takes care of all DMA memory allocations. Store the allocated memory * information for the descriptor ring and buffers. If the memory for either the * descriptor ring or the buffers is not allocated properly and/or inconsistent * with the control queue parameters, this routine will free the memory for * both the descriptors and the buffers */ int cpfl_ctlq_alloc_ring_res(struct idpf_hw *hw __rte_unused, struct idpf_ctlq_info *cq, struct cpfl_ctlq_create_info *qinfo) { int ret_code = 0; unsigned int elem_size; int i = 0; ret_code = cpfl_check_dma_mem_parameters(qinfo); if (ret_code) /* TODO: Log an error message per CP */ goto err; cq->desc_ring.va = qinfo->ring_mem.va; cq->desc_ring.pa = qinfo->ring_mem.pa; cq->desc_ring.size = qinfo->ring_mem.size; switch (cq->cq_type) { case IDPF_CTLQ_TYPE_MAILBOX_RX: case IDPF_CTLQ_TYPE_CONFIG_RX: case IDPF_CTLQ_TYPE_EVENT_RX: case IDPF_CTLQ_TYPE_RDMA_RX: /* Only receive queues will have allocated buffers * during init. CP allocates one big chunk of DMA * region who size is equal to ring_len * buff_size. * In CPFLib, the block gets broken down to multiple * smaller blocks that actually gets programmed in the hardware. */ cq->bi.rx_buff = (struct idpf_dma_mem **) idpf_calloc(hw, cq->ring_size, sizeof(struct idpf_dma_mem *)); if (!cq->bi.rx_buff) { ret_code = -ENOMEM; /* TODO: Log an error message per CP */ goto err; } elem_size = qinfo->buf_size; for (i = 0; i < cq->ring_size; i++) { cq->bi.rx_buff[i] = (struct idpf_dma_mem *)idpf_calloc (hw, 1, sizeof(struct idpf_dma_mem)); if (!cq->bi.rx_buff[i]) { ret_code = -ENOMEM; goto free_rx_buffs; } cq->bi.rx_buff[i]->va = (uint64_t *)((char *)qinfo->buf_mem.va + (i * elem_size)); cq->bi.rx_buff[i]->pa = qinfo->buf_mem.pa + (i * elem_size); cq->bi.rx_buff[i]->size = elem_size; } break; case IDPF_CTLQ_TYPE_MAILBOX_TX: case IDPF_CTLQ_TYPE_CONFIG_TX: case IDPF_CTLQ_TYPE_RDMA_TX: case IDPF_CTLQ_TYPE_RDMA_COMPL: break; default: ret_code = -EINVAL; } return ret_code; free_rx_buffs: i--; for (; i >= 0; i--) idpf_free(hw, cq->bi.rx_buff[i]); if (!cq->bi.rx_buff) idpf_free(hw, cq->bi.rx_buff); err: return ret_code; } /** * cpfl_ctlq_init_rxq_bufs - populate receive queue descriptors with buf * @cq: pointer to the specific Control queue * * Record the address of the receive queue DMA buffers in the descriptors. * The buffers must have been previously allocated. */ static void cpfl_ctlq_init_rxq_bufs(struct idpf_ctlq_info *cq) { int i = 0; for (i = 0; i < cq->ring_size; i++) { struct idpf_ctlq_desc *desc = IDPF_CTLQ_DESC(cq, i); struct idpf_dma_mem *bi = cq->bi.rx_buff[i]; /* No buffer to post to descriptor, continue */ if (!bi) continue; desc->flags = CPU_TO_LE16(IDPF_CTLQ_FLAG_BUF | IDPF_CTLQ_FLAG_RD); desc->opcode = 0; desc->datalen = CPU_TO_LE16(bi->size); desc->ret_val = 0; desc->cookie_high = 0; desc->cookie_low = 0; desc->params.indirect.addr_high = CPU_TO_LE32(IDPF_HI_DWORD(bi->pa)); desc->params.indirect.addr_low = CPU_TO_LE32(IDPF_LO_DWORD(bi->pa)); desc->params.indirect.param0 = 0; desc->params.indirect.param1 = 0; } } /** * cpfl_ctlq_setup_regs - initialize control queue registers * @cq: pointer to the specific control queue * @q_create_info: structs containing info for each queue to be initialized */ static void cpfl_ctlq_setup_regs(struct idpf_ctlq_info *cq, struct cpfl_ctlq_create_info *q_create_info) { /* set control queue registers in our local struct */ cq->reg.head = q_create_info->reg.head; cq->reg.tail = q_create_info->reg.tail; cq->reg.len = q_create_info->reg.len; cq->reg.bah = q_create_info->reg.bah; cq->reg.bal = q_create_info->reg.bal; cq->reg.len_mask = q_create_info->reg.len_mask; cq->reg.len_ena_mask = q_create_info->reg.len_ena_mask; cq->reg.head_mask = q_create_info->reg.head_mask; } /** * cpfl_ctlq_init_regs - Initialize control queue registers * @hw: pointer to hw struct * @cq: pointer to the specific Control queue * @is_rxq: true if receive control queue, false otherwise * * Initialize registers. The caller is expected to have already initialized the * descriptor ring memory and buffer memory */ static void cpfl_ctlq_init_regs(struct idpf_hw *hw, struct idpf_ctlq_info *cq, bool is_rxq) { /* Update tail to post pre-allocated buffers for rx queues */ if (is_rxq) wr32(hw, cq->reg.tail, (uint32_t)(cq->ring_size - 1)); /* For non-Mailbox control queues only TAIL need to be set */ if (cq->q_id != -1) return; /* Clear Head for both send or receive */ wr32(hw, cq->reg.head, 0); /* set starting point */ wr32(hw, cq->reg.bal, IDPF_LO_DWORD(cq->desc_ring.pa)); wr32(hw, cq->reg.bah, IDPF_HI_DWORD(cq->desc_ring.pa)); wr32(hw, cq->reg.len, (cq->ring_size | cq->reg.len_ena_mask)); } /** * cpfl_ctlq_dealloc_ring_res - free up the descriptor buffer structure * @hw: context info for the callback * @cq: pointer to the specific control queue * * DMA buffers are released by the CP itself */ static void cpfl_ctlq_dealloc_ring_res(struct idpf_hw *hw __rte_unused, struct idpf_ctlq_info *cq) { int i; if (cq->cq_type == IDPF_CTLQ_TYPE_MAILBOX_RX || cq->cq_type == IDPF_CTLQ_TYPE_CONFIG_RX) { for (i = 0; i < cq->ring_size; i++) idpf_free(hw, cq->bi.rx_buff[i]); /* free the buffer header */ idpf_free(hw, cq->bi.rx_buff); } else { idpf_free(hw, cq->bi.tx_msg); } } /** * cpfl_ctlq_add - add one control queue * @hw: pointer to hardware struct * @qinfo: info for queue to be created * @cq_out: (output) double pointer to control queue to be created * * Allocate and initialize a control queue and add it to the control queue list. * The cq parameter will be allocated/initialized and passed back to the caller * if no errors occur. */ int cpfl_ctlq_add(struct idpf_hw *hw, struct cpfl_ctlq_create_info *qinfo, struct idpf_ctlq_info **cq_out) { struct idpf_ctlq_info *cq; bool is_rxq = false; int status = 0; if (!qinfo->len || !qinfo->buf_size || qinfo->len > IDPF_CTLQ_MAX_RING_SIZE || qinfo->buf_size > IDPF_CTLQ_MAX_BUF_LEN) return -EINVAL; cq = (struct idpf_ctlq_info *) idpf_calloc(hw, 1, sizeof(struct idpf_ctlq_info)); if (!cq) return -ENOMEM; cq->cq_type = qinfo->type; cq->q_id = qinfo->id; cq->buf_size = qinfo->buf_size; cq->ring_size = qinfo->len; cq->next_to_use = 0; cq->next_to_clean = 0; cq->next_to_post = cq->ring_size - 1; switch (qinfo->type) { case IDPF_CTLQ_TYPE_EVENT_RX: case IDPF_CTLQ_TYPE_CONFIG_RX: case IDPF_CTLQ_TYPE_MAILBOX_RX: is_rxq = true; /* fallthrough */ case IDPF_CTLQ_TYPE_CONFIG_TX: case IDPF_CTLQ_TYPE_MAILBOX_TX: status = cpfl_ctlq_alloc_ring_res(hw, cq, qinfo); break; default: status = -EINVAL; break; } if (status) goto init_free_q; if (is_rxq) { cpfl_ctlq_init_rxq_bufs(cq); } else { /* Allocate the array of msg pointers for TX queues */ cq->bi.tx_msg = (struct idpf_ctlq_msg **) idpf_calloc(hw, qinfo->len, sizeof(struct idpf_ctlq_msg *)); if (!cq->bi.tx_msg) { status = -ENOMEM; goto init_dealloc_q_mem; } } cpfl_ctlq_setup_regs(cq, qinfo); cpfl_ctlq_init_regs(hw, cq, is_rxq); idpf_init_lock(&cq->cq_lock); LIST_INSERT_HEAD(&hw->cq_list_head, cq, cq_list); *cq_out = cq; return status; init_dealloc_q_mem: /* free ring buffers and the ring itself */ cpfl_ctlq_dealloc_ring_res(hw, cq); init_free_q: idpf_free(hw, cq); cq = NULL; return status; } /** * cpfl_ctlq_send - send command to Control Queue (CTQ) * @hw: pointer to hw struct * @cq: handle to control queue struct to send on * @num_q_msg: number of messages to send on control queue * @q_msg: pointer to array of queue messages to be sent * * The caller is expected to allocate DMAable buffers and pass them to the * send routine via the q_msg struct / control queue specific data struct. * The control queue will hold a reference to each send message until * the completion for that message has been cleaned. */ int cpfl_ctlq_send(struct idpf_hw *hw, struct idpf_ctlq_info *cq, uint16_t num_q_msg, struct idpf_ctlq_msg q_msg[]) { struct idpf_ctlq_desc *desc; int num_desc_avail = 0; int status = 0; int i = 0; if (!cq || !cq->ring_size) return -ENOBUFS; idpf_acquire_lock(&cq->cq_lock); /* Ensure there are enough descriptors to send all messages */ num_desc_avail = IDPF_CTLQ_DESC_UNUSED(cq); if (num_desc_avail == 0 || num_desc_avail < num_q_msg) { status = -ENOSPC; goto sq_send_command_out; } for (i = 0; i < num_q_msg; i++) { struct idpf_ctlq_msg *msg = &q_msg[i]; desc = IDPF_CTLQ_DESC(cq, cq->next_to_use); desc->opcode = CPU_TO_LE16(msg->opcode); desc->pfid_vfid = CPU_TO_LE16(msg->func_id); desc->cookie_high = CPU_TO_LE32(msg->cookie.mbx.chnl_opcode); desc->cookie_low = CPU_TO_LE32(msg->cookie.mbx.chnl_retval); desc->flags = CPU_TO_LE16((msg->host_id & IDPF_HOST_ID_MASK) << IDPF_CTLQ_FLAG_HOST_ID_S); if (msg->data_len) { struct idpf_dma_mem *buff = msg->ctx.indirect.payload; desc->datalen |= CPU_TO_LE16(msg->data_len); desc->flags |= CPU_TO_LE16(IDPF_CTLQ_FLAG_BUF); desc->flags |= CPU_TO_LE16(IDPF_CTLQ_FLAG_RD); /* Update the address values in the desc with the pa * value for respective buffer */ desc->params.indirect.addr_high = CPU_TO_LE32(IDPF_HI_DWORD(buff->pa)); desc->params.indirect.addr_low = CPU_TO_LE32(IDPF_LO_DWORD(buff->pa)); idpf_memcpy(&desc->params, msg->ctx.indirect.context, IDPF_INDIRECT_CTX_SIZE, IDPF_NONDMA_TO_DMA); } else { idpf_memcpy(&desc->params, msg->ctx.direct, IDPF_DIRECT_CTX_SIZE, IDPF_NONDMA_TO_DMA); } /* Store buffer info */ cq->bi.tx_msg[cq->next_to_use] = msg; (cq->next_to_use)++; if (cq->next_to_use == cq->ring_size) cq->next_to_use = 0; } /* Force memory write to complete before letting hardware * know that there are new descriptors to fetch. */ idpf_wmb(); wr32(hw, cq->reg.tail, cq->next_to_use); sq_send_command_out: idpf_release_lock(&cq->cq_lock); return status; } /** * __cpfl_ctlq_clean_sq - helper function to reclaim descriptors on HW write * back for the requested queue * @cq: pointer to the specific Control queue * @clean_count: (input|output) number of descriptors to clean as input, and * number of descriptors actually cleaned as output * @msg_status: (output) pointer to msg pointer array to be populated; needs * to be allocated by caller * @force: (input) clean descriptors which were not done yet. Use with caution * in kernel mode only * * Returns an array of message pointers associated with the cleaned * descriptors. The pointers are to the original ctlq_msgs sent on the cleaned * descriptors. The status will be returned for each; any messages that failed * to send will have a non-zero status. The caller is expected to free original * ctlq_msgs and free or reuse the DMA buffers. */ static int __cpfl_ctlq_clean_sq(struct idpf_ctlq_info *cq, uint16_t *clean_count, struct idpf_ctlq_msg *msg_status[], bool force) { struct idpf_ctlq_desc *desc; uint16_t i = 0, num_to_clean; uint16_t ntc, desc_err; int ret = 0; if (!cq || !cq->ring_size) return -ENOBUFS; if (*clean_count == 0) return 0; if (*clean_count > cq->ring_size) return -EINVAL; idpf_acquire_lock(&cq->cq_lock); ntc = cq->next_to_clean; num_to_clean = *clean_count; for (i = 0; i < num_to_clean; i++) { /* Fetch next descriptor and check if marked as done */ desc = IDPF_CTLQ_DESC(cq, ntc); if (!force && !(LE16_TO_CPU(desc->flags) & IDPF_CTLQ_FLAG_DD)) break; desc_err = LE16_TO_CPU(desc->ret_val); if (desc_err) { /* strip off FW internal code */ desc_err &= 0xff; } msg_status[i] = cq->bi.tx_msg[ntc]; if (!msg_status[i]) break; msg_status[i]->status = desc_err; cq->bi.tx_msg[ntc] = NULL; /* Zero out any stale data */ idpf_memset(desc, 0, sizeof(*desc), IDPF_DMA_MEM); ntc++; if (ntc == cq->ring_size) ntc = 0; } cq->next_to_clean = ntc; idpf_release_lock(&cq->cq_lock); /* Return number of descriptors actually cleaned */ *clean_count = i; return ret; } /** * cpfl_ctlq_clean_sq - reclaim send descriptors on HW write back for the * requested queue * @cq: pointer to the specific Control queue * @clean_count: (input|output) number of descriptors to clean as input, and * number of descriptors actually cleaned as output * @msg_status: (output) pointer to msg pointer array to be populated; needs * to be allocated by caller * * Returns an array of message pointers associated with the cleaned * descriptors. The pointers are to the original ctlq_msgs sent on the cleaned * descriptors. The status will be returned for each; any messages that failed * to send will have a non-zero status. The caller is expected to free original * ctlq_msgs and free or reuse the DMA buffers. */ int cpfl_ctlq_clean_sq(struct idpf_ctlq_info *cq, uint16_t *clean_count, struct idpf_ctlq_msg *msg_status[]) { return __cpfl_ctlq_clean_sq(cq, clean_count, msg_status, false); } /** * cpfl_ctlq_post_rx_buffs - post buffers to descriptor ring * @hw: pointer to hw struct * @cq: pointer to control queue handle * @buff_count: (input|output) input is number of buffers caller is trying to * return; output is number of buffers that were not posted * @buffs: array of pointers to dma mem structs to be given to hardware * * Caller uses this function to return DMA buffers to the descriptor ring after * consuming them; buff_count will be the number of buffers. * * Note: this function needs to be called after a receive call even * if there are no DMA buffers to be returned, i.e. buff_count = 0, * buffs = NULL to support direct commands */ int cpfl_ctlq_post_rx_buffs(struct idpf_hw *hw, struct idpf_ctlq_info *cq, uint16_t *buff_count, struct idpf_dma_mem **buffs) { struct idpf_ctlq_desc *desc; uint16_t ntp = cq->next_to_post; bool buffs_avail = false; uint16_t tbp = ntp + 1; int status = 0; int i = 0; if (*buff_count > cq->ring_size) return -EINVAL; if (*buff_count > 0) buffs_avail = true; idpf_acquire_lock(&cq->cq_lock); if (tbp >= cq->ring_size) tbp = 0; if (tbp == cq->next_to_clean) /* Nothing to do */ goto post_buffs_out; /* Post buffers for as many as provided or up until the last one used */ while (ntp != cq->next_to_clean) { desc = IDPF_CTLQ_DESC(cq, ntp); if (cq->bi.rx_buff[ntp]) goto fill_desc; if (!buffs_avail) { /* If the caller hasn't given us any buffers or * there are none left, search the ring itself * for an available buffer to move to this * entry starting at the next entry in the ring */ tbp = ntp + 1; /* Wrap ring if necessary */ if (tbp >= cq->ring_size) tbp = 0; while (tbp != cq->next_to_clean) { if (cq->bi.rx_buff[tbp]) { cq->bi.rx_buff[ntp] = cq->bi.rx_buff[tbp]; cq->bi.rx_buff[tbp] = NULL; /* Found a buffer, no need to * search anymore */ break; } /* Wrap ring if necessary */ tbp++; if (tbp >= cq->ring_size) tbp = 0; } if (tbp == cq->next_to_clean) goto post_buffs_out; } else { /* Give back pointer to DMA buffer */ cq->bi.rx_buff[ntp] = buffs[i]; i++; if (i >= *buff_count) buffs_avail = false; } fill_desc: desc->flags = CPU_TO_LE16(IDPF_CTLQ_FLAG_BUF | IDPF_CTLQ_FLAG_RD); /* Post buffers to descriptor */ desc->datalen = CPU_TO_LE16(cq->bi.rx_buff[ntp]->size); desc->params.indirect.addr_high = CPU_TO_LE32(IDPF_HI_DWORD(cq->bi.rx_buff[ntp]->pa)); desc->params.indirect.addr_low = CPU_TO_LE32(IDPF_LO_DWORD(cq->bi.rx_buff[ntp]->pa)); ntp++; if (ntp == cq->ring_size) ntp = 0; } post_buffs_out: /* Only update tail if buffers were actually posted */ if (cq->next_to_post != ntp) { if (ntp) /* Update next_to_post to ntp - 1 since current ntp * will not have a buffer */ cq->next_to_post = ntp - 1; else /* Wrap to end of end ring since current ntp is 0 */ cq->next_to_post = cq->ring_size - 1; wr32(hw, cq->reg.tail, cq->next_to_post); } idpf_release_lock(&cq->cq_lock); /* return the number of buffers that were not posted */ *buff_count = *buff_count - i; return status; } /** * cpfl_ctlq_recv - receive control queue message call back * @cq: pointer to control queue handle to receive on * @num_q_msg: (input|output) input number of messages that should be received; * output number of messages actually received * @q_msg: (output) array of received control queue messages on this q; * needs to be pre-allocated by caller for as many messages as requested * * Called by interrupt handler or polling mechanism. Caller is expected * to free buffers */ int cpfl_ctlq_recv(struct idpf_ctlq_info *cq, uint16_t *num_q_msg, struct idpf_ctlq_msg *q_msg) { uint16_t num_to_clean, ntc, ret_val, flags; struct idpf_ctlq_desc *desc; int ret_code = 0; uint16_t i = 0; if (!cq || !cq->ring_size) return -ENOBUFS; if (*num_q_msg == 0) return 0; else if (*num_q_msg > cq->ring_size) return -EINVAL; /* take the lock before we start messing with the ring */ idpf_acquire_lock(&cq->cq_lock); ntc = cq->next_to_clean; num_to_clean = *num_q_msg; for (i = 0; i < num_to_clean; i++) { /* Fetch next descriptor and check if marked as done */ desc = IDPF_CTLQ_DESC(cq, ntc); flags = LE16_TO_CPU(desc->flags); if (!(flags & IDPF_CTLQ_FLAG_DD)) break; ret_val = LE16_TO_CPU(desc->ret_val); q_msg[i].vmvf_type = (flags & (IDPF_CTLQ_FLAG_FTYPE_VM | IDPF_CTLQ_FLAG_FTYPE_PF)) >> IDPF_CTLQ_FLAG_FTYPE_S; if (flags & IDPF_CTLQ_FLAG_ERR) ret_code = -EBADMSG; q_msg[i].cookie.mbx.chnl_opcode = LE32_TO_CPU(desc->cookie_high); q_msg[i].cookie.mbx.chnl_retval = LE32_TO_CPU(desc->cookie_low); q_msg[i].opcode = LE16_TO_CPU(desc->opcode); q_msg[i].data_len = LE16_TO_CPU(desc->datalen); q_msg[i].status = ret_val; if (desc->datalen) { idpf_memcpy(q_msg[i].ctx.indirect.context, &desc->params.indirect, IDPF_INDIRECT_CTX_SIZE, IDPF_DMA_TO_NONDMA); /* Assign pointer to dma buffer to ctlq_msg array * to be given to upper layer */ q_msg[i].ctx.indirect.payload = cq->bi.rx_buff[ntc]; /* Zero out pointer to DMA buffer info; * will be repopulated by post buffers API */ cq->bi.rx_buff[ntc] = NULL; } else { idpf_memcpy(q_msg[i].ctx.direct, desc->params.raw, IDPF_DIRECT_CTX_SIZE, IDPF_DMA_TO_NONDMA); } /* Zero out stale data in descriptor */ idpf_memset(desc, 0, sizeof(struct idpf_ctlq_desc), IDPF_DMA_MEM); ntc++; if (ntc == cq->ring_size) ntc = 0; }; cq->next_to_clean = ntc; idpf_release_lock(&cq->cq_lock); *num_q_msg = i; if (*num_q_msg == 0) ret_code = -ENOMSG; return ret_code; } int cpfl_vport_ctlq_add(struct idpf_hw *hw, struct cpfl_ctlq_create_info *qinfo, struct idpf_ctlq_info **cq) { return cpfl_ctlq_add(hw, qinfo, cq); } /** * cpfl_ctlq_shutdown - shutdown the CQ * The main shutdown routine for any controq queue */ static void cpfl_ctlq_shutdown(struct idpf_hw *hw, struct idpf_ctlq_info *cq) { idpf_acquire_lock(&cq->cq_lock); if (!cq->ring_size) goto shutdown_sq_out; /* free ring buffers and the ring itself */ cpfl_ctlq_dealloc_ring_res(hw, cq); /* Set ring_size to 0 to indicate uninitialized queue */ cq->ring_size = 0; shutdown_sq_out: idpf_release_lock(&cq->cq_lock); idpf_destroy_lock(&cq->cq_lock); } /** * cpfl_ctlq_remove - deallocate and remove specified control queue */ static void cpfl_ctlq_remove(struct idpf_hw *hw, struct idpf_ctlq_info *cq) { LIST_REMOVE(cq, cq_list); cpfl_ctlq_shutdown(hw, cq); idpf_free(hw, cq); } void cpfl_vport_ctlq_remove(struct idpf_hw *hw, struct idpf_ctlq_info *cq) { cpfl_ctlq_remove(hw, cq); } int cpfl_vport_ctlq_send(struct idpf_hw *hw, struct idpf_ctlq_info *cq, uint16_t num_q_msg, struct idpf_ctlq_msg q_msg[]) { return cpfl_ctlq_send(hw, cq, num_q_msg, q_msg); } int cpfl_vport_ctlq_recv(struct idpf_ctlq_info *cq, uint16_t *num_q_msg, struct idpf_ctlq_msg q_msg[]) { return cpfl_ctlq_recv(cq, num_q_msg, q_msg); } int cpfl_vport_ctlq_post_rx_buffs(struct idpf_hw *hw, struct idpf_ctlq_info *cq, uint16_t *buff_count, struct idpf_dma_mem **buffs) { return cpfl_ctlq_post_rx_buffs(hw, cq, buff_count, buffs); } int cpfl_vport_ctlq_clean_sq(struct idpf_ctlq_info *cq, uint16_t *clean_count, struct idpf_ctlq_msg *msg_status[]) { return cpfl_ctlq_clean_sq(cq, clean_count, msg_status); }