/dpdk/drivers/crypto/scheduler/ |
H A D | scheduler_failover.c | 27 struct rte_crypto_op **ops, uint16_t nb_ops, uint8_t index) in failover_worker_enqueue() 41 schedule_enqueue(void *qp, struct rte_crypto_op **ops, uint16_t nb_ops) in schedule_enqueue() 72 uint16_t nb_ops) in schedule_enqueue_ordering() 87 schedule_dequeue(void *qp, struct rte_crypto_op **ops, uint16_t nb_ops) in schedule_dequeue() 123 uint16_t nb_ops) in schedule_dequeue_ordering()
|
H A D | scheduler_roundrobin.c | 20 schedule_enqueue(void *qp, struct rte_crypto_op **ops, uint16_t nb_ops) in schedule_enqueue() 48 uint16_t nb_ops) in schedule_enqueue_ordering() 64 schedule_dequeue(void *qp, struct rte_crypto_op **ops, uint16_t nb_ops) in schedule_dequeue() 103 uint16_t nb_ops) in schedule_dequeue_ordering()
|
H A D | scheduler_pkt_size_distr.c | 37 schedule_enqueue(void *qp, struct rte_crypto_op **ops, uint16_t nb_ops) in schedule_enqueue() 182 uint16_t nb_ops) in schedule_enqueue_ordering() 197 schedule_dequeue(void *qp, struct rte_crypto_op **ops, uint16_t nb_ops) in schedule_dequeue() 237 uint16_t nb_ops) in schedule_dequeue_ordering()
|
H A D | scheduler_multicore.c | 39 schedule_enqueue(void *qp, struct rte_crypto_op **ops, uint16_t nb_ops) in schedule_enqueue() 68 uint16_t nb_ops) in schedule_enqueue_ordering() 84 schedule_dequeue(void *qp, struct rte_crypto_op **ops, uint16_t nb_ops) in schedule_dequeue() 111 uint16_t nb_ops) in schedule_dequeue_ordering()
|
/dpdk/lib/compressdev/ |
H A D | rte_comp.c | 102 struct rte_comp_op **ops, uint16_t nb_ops) in rte_comp_op_raw_bulk_alloc() 198 struct rte_comp_op **ops, uint16_t nb_ops) in rte_comp_op_bulk_alloc() 229 rte_comp_op_bulk_free(struct rte_comp_op **ops, uint16_t nb_ops) in rte_comp_op_bulk_free()
|
H A D | rte_compressdev.c | 583 struct rte_comp_op **ops, uint16_t nb_ops) in rte_compressdev_dequeue_burst() 595 struct rte_comp_op **ops, uint16_t nb_ops) in rte_compressdev_enqueue_burst()
|
/dpdk/drivers/crypto/null/ |
H A D | null_crypto_pmd.c | 103 uint16_t nb_ops) in null_crypto_pmd_enqueue_burst() 134 uint16_t nb_ops) in null_crypto_pmd_dequeue_burst()
|
/dpdk/lib/cryptodev/ |
H A D | cryptodev_pmd.c | 193 __rte_unused uint16_t nb_ops) in dummy_crypto_enqueue_burst() 204 __rte_unused uint16_t nb_ops) in dummy_crypto_dequeue_burst()
|
H A D | rte_crypto.h | 282 struct rte_crypto_op **ops, uint16_t nb_ops) in __rte_crypto_op_raw_bulk_alloc() argument 340 struct rte_crypto_op **ops, uint16_t nb_ops) in rte_crypto_op_bulk_alloc() argument [all...] |
H A D | rte_cryptodev.h | 1900 rte_cryptodev_dequeue_burst(uint8_t dev_id,uint16_t qp_id,struct rte_crypto_op ** ops,uint16_t nb_ops) rte_cryptodev_dequeue_burst() argument 1972 rte_cryptodev_enqueue_burst(uint8_t dev_id,uint16_t qp_id,struct rte_crypto_op ** ops,uint16_t nb_ops) rte_cryptodev_enqueue_burst() argument [all...] |
/dpdk/drivers/regex/mlx5/ |
H A D | mlx5_regex_fastpath.c | 61 validate_ops(struct rte_regex_ops **ops, uint16_t nb_ops) in validate_ops() 310 size_t nb_ops) in prep_regex_umr_wqe_set() 408 struct rte_regex_ops **ops, uint16_t nb_ops) in mlx5_regexdev_enqueue_gga() 447 struct rte_regex_ops **ops, uint16_t nb_ops) in mlx5_regexdev_enqueue() 572 struct rte_regex_ops **ops, uint16_t nb_ops) in mlx5_regexdev_dequeue()
|
/dpdk/drivers/compress/uadk/ |
H A D | uadk_compress_pmd.c | 296 struct rte_comp_op **ops, uint16_t nb_ops) in uadk_compress_pmd_enqueue_burst_sync() 366 uint16_t nb_ops) in uadk_compress_pmd_dequeue_burst_sync()
|
/dpdk/drivers/crypto/armv8/ |
H A D | rte_armv8_pmd.c | 688 uint16_t nb_ops) in armv8_crypto_pmd_enqueue_burst() 723 uint16_t nb_ops) in armv8_crypto_pmd_dequeue_burst()
|
/dpdk/drivers/crypto/ionic/ |
H A D | ionic_crypto_ops.c | 365 iocpt_enqueue_sym(void *qp, struct rte_crypto_op **ops, uint16_t nb_ops) in iocpt_enqueue_sym() 472 iocpt_dequeue_sym(void *qp, struct rte_crypto_op **ops, uint16_t nb_ops) in iocpt_dequeue_sym()
|
/dpdk/drivers/crypto/mvsam/ |
H A D | rte_mrvl_pmd.c | 857 uint16_t nb_ops) in mrvl_crypto_pmd_enqueue_burst() 987 uint16_t nb_ops) in mrvl_crypto_pmd_dequeue_burst()
|
/dpdk/drivers/compress/qat/ |
H A D | qat_comp_pmd.c | 627 qat_comp_dequeue_burst(void *qp, struct rte_comp_op **ops, uint16_t nb_ops) in qat_comp_dequeue_burst() 635 uint16_t nb_ops) in qat_comp_pmd_dequeue_first_op_burst()
|
H A D | qat_comp.c | 1161 qat_enqueue_comp_op_burst(void *qp, void **ops, uint16_t nb_ops) in qat_enqueue_comp_op_burst()
|
/dpdk/drivers/crypto/ipsec_mb/ |
H A D | pmd_zuc.c | 253 struct rte_crypto_op **c_ops, uint16_t nb_ops) in zuc_pmd_dequeue_burst()
|
H A D | pmd_aesni_gcm.c | 487 struct rte_crypto_op **ops, uint16_t nb_ops) in aesni_gcm_pmd_dequeue_burst()
|
/dpdk/lib/mldev/ |
H A D | rte_mldev.c | 808 rte_ml_enqueue_burst(int16_t dev_id,uint16_t qp_id,struct rte_ml_op ** ops,uint16_t nb_ops) rte_ml_enqueue_burst() argument 844 rte_ml_dequeue_burst(int16_t dev_id,uint16_t qp_id,struct rte_ml_op ** ops,uint16_t nb_ops) rte_ml_dequeue_burst() argument
|
/dpdk/lib/vhost/ |
H A D | vhost_crypto.c | 1334 uint16_t nb_ops, int *callfd) in vhost_crypto_complete_one_vm_requests() 1553 struct rte_crypto_op **ops, uint16_t nb_ops) in rte_vhost_crypto_fetch_requests() 1664 uint16_t nb_ops, int *callfds, uint16_t *nb_callfds) in rte_vhost_crypto_finalize_requests()
|
/dpdk/lib/regexdev/ |
H A D | rte_regexdev.h | 1471 struct rte_regex_ops **ops, uint16_t nb_ops) in rte_regexdev_enqueue_burst() argument 1531 rte_regexdev_dequeue_burst(uint8_t dev_id,uint16_t qp_id,struct rte_regex_ops ** ops,uint16_t nb_ops) rte_regexdev_dequeue_burst() argument [all...] |
/dpdk/drivers/crypto/ccp/ |
H A D | ccp_crypto.c | 2656 uint16_t nb_ops, in process_ops_to_enqueue() 2825 uint16_t nb_ops) in ccp_prepare_ops() 2879 uint16_t nb_ops, in process_ops_to_dequeue()
|