Lines Matching defs:buf_vec

720 		struct buf_vector *buf_vec, uint16_t *vec_idx,
742 buf_vec[vec_id].buf_iova = desc_iova;
743 buf_vec[vec_id].buf_addr = desc_addr;
744 buf_vec[vec_id].buf_len = desc_chunck_len;
758 struct buf_vector *buf_vec, uint16_t *desc_chain_head,
814 if (unlikely(map_one_desc(dev, vq, buf_vec, &vec_id,
841 uint64_t size, struct buf_vector *buf_vec,
873 &vec_idx, buf_vec,
894 struct buf_vector *buf_vec, uint32_t *len, uint8_t perm)
936 if (unlikely(map_one_desc(dev, vq, buf_vec, &vec_id,
952 struct buf_vector *buf_vec, uint16_t *vec_idx,
988 &vec_id, buf_vec,
995 if (unlikely(map_one_desc(dev, vq, buf_vec, &vec_id,
1018 struct buf_vector *buf_vec,
1025 uint64_t iova = buf_vec->buf_iova;
1029 buf_vec->buf_len);
1030 dst = buf_vec->buf_addr;
1043 buf_vec++;
1201 struct rte_mbuf *m, struct buf_vector *buf_vec,
1219 buf_addr = buf_vec[vec_idx].buf_addr;
1220 buf_iova = buf_vec[vec_idx].buf_iova;
1221 buf_len = buf_vec[vec_idx].buf_len;
1239 buf_addr = buf_vec[vec_idx].buf_addr;
1240 buf_iova = buf_vec[vec_idx].buf_iova;
1241 buf_len = buf_vec[vec_idx].buf_len;
1263 buf_addr = buf_vec[vec_idx].buf_addr;
1264 buf_iova = buf_vec[vec_idx].buf_iova;
1265 buf_len = buf_vec[vec_idx].buf_len;
1286 copy_vnet_hdr_to_desc(dev, vq, buf_vec, hdr);
1291 buf_vec[0].buf_iova,
1331 struct buf_vector *buf_vec,
1364 buf_vec, &nr_vec,
1383 if (mbuf_to_desc(dev, vq, pkt, buf_vec, nr_vec, num_buffers, false) < 0)
1400 struct buf_vector buf_vec[BUF_VECTOR_MAX];
1417 pkt_len, buf_vec, &num_buffers,
1429 if (mbuf_to_desc(dev, vq, pkts[pkt_idx], buf_vec, nr_vec,
1632 struct buf_vector buf_vec[BUF_VECTOR_MAX];
1635 if (unlikely(vhost_enqueue_single_packed(dev, vq, pkt, buf_vec,
1802 struct buf_vector buf_vec[BUF_VECTOR_MAX];
1827 if (unlikely(reserve_avail_buf_split(dev, vq, pkt_len, buf_vec,
1839 if (mbuf_to_desc(dev, vq, pkts[pkt_idx], buf_vec, nr_vec, num_buffers, true) < 0) {
1907 struct buf_vector *buf_vec,
1940 buf_vec, &nr_vec,
1958 if (unlikely(mbuf_to_desc(dev, vq, pkt, buf_vec, nr_vec, *nr_buffers, true) < 0))
1973 struct buf_vector buf_vec[BUF_VECTOR_MAX];
1975 if (unlikely(vhost_enqueue_async_packed(dev, vq, pkt, buf_vec,
2869 struct buf_vector *buf_vec)
2877 len = RTE_MIN(remain, buf_vec->buf_len);
2878 src = buf_vec->buf_addr;
2884 buf_vec++;
2890 struct buf_vector *buf_vec, uint16_t nr_vec,
2914 if (unlikely(buf_vec[0].buf_len < sizeof(struct virtio_net_hdr))) {
2919 copy_vnet_hdr_from_desc(&tmp_hdr, buf_vec);
2922 hdr = (struct virtio_net_hdr *)((uintptr_t)buf_vec[0].buf_addr);
2927 if (buf_vec[vec_idx].buf_len > hdr_remain)
2930 hdr_remain -= buf_vec[vec_idx].buf_len;
2933 buf_addr = buf_vec[vec_idx].buf_addr;
2934 buf_iova = buf_vec[vec_idx].buf_iova;
2935 buf_len = buf_vec[vec_idx].buf_len;
2937 buf_avail = buf_vec[vec_idx].buf_len - hdr_remain;
2979 buf_addr = buf_vec[vec_idx].buf_addr;
2980 buf_iova = buf_vec[vec_idx].buf_iova;
2981 buf_len = buf_vec[vec_idx].buf_len;
3131 struct buf_vector buf_vec[BUF_VECTOR_MAX];
3139 &nr_vec, buf_vec,
3167 err = desc_to_mbuf(dev, vq, buf_vec, nr_vec, pkts[i],
3415 struct buf_vector buf_vec[BUF_VECTOR_MAX];
3423 buf_vec, &nr_vec,
3443 err = desc_to_mbuf(dev, vq, buf_vec, nr_vec, pkts,
3791 struct buf_vector buf_vec[BUF_VECTOR_MAX];
3795 &nr_vec, buf_vec,
3828 err = desc_to_mbuf(dev, vq, buf_vec, nr_vec, pkt, mbuf_pool,
3953 struct buf_vector buf_vec[BUF_VECTOR_MAX];
3959 buf_vec, &nr_vec, &buf_id, &buf_len,
3973 err = desc_to_mbuf(dev, vq, buf_vec, nr_vec, pkts, mbuf_pool,