Home
last modified time | relevance | path

Searched refs:dev_id (Results 1 – 25 of 155) sorted by relevance

1234567

/dpdk/lib/mldev/
H A Drte_mldev.c27 rte_ml_dev_pmd_get_dev(int16_t dev_id) in rte_ml_dev_pmd_get_dev() argument
29 return &ml_dev_globals.devs[dev_id]; in rte_ml_dev_pmd_get_dev()
36 int16_t dev_id; in rte_ml_dev_pmd_get_named_dev() local
41 for (dev_id = 0; dev_id < ml_dev_globals.max_devs; dev_id++) { in rte_ml_dev_pmd_get_named_dev()
42 dev = rte_ml_dev_pmd_get_dev(dev_id); in rte_ml_dev_pmd_get_named_dev()
56 int16_t dev_id; in rte_ml_dev_pmd_allocate() local
70 for (dev_id = 0; dev_id < ml_dev_global in rte_ml_dev_pmd_allocate()
128 int16_t dev_id; rte_ml_dev_pmd_release() local
200 rte_ml_dev_is_valid_dev(int16_t dev_id) rte_ml_dev_is_valid_dev() argument
215 rte_ml_dev_socket_id(int16_t dev_id) rte_ml_dev_socket_id() argument
230 rte_ml_dev_info_get(int16_t dev_id,struct rte_ml_dev_info * dev_info) rte_ml_dev_info_get() argument
253 rte_ml_dev_configure(int16_t dev_id,const struct rte_ml_dev_config * config) rte_ml_dev_configure() argument
292 rte_ml_dev_close(int16_t dev_id) rte_ml_dev_close() argument
315 rte_ml_dev_start(int16_t dev_id) rte_ml_dev_start() argument
342 rte_ml_dev_stop(int16_t dev_id) rte_ml_dev_stop() argument
369 rte_ml_dev_queue_pair_setup(int16_t dev_id,uint16_t queue_pair_id,const struct rte_ml_dev_qp_conf * qp_conf,int socket_id) rte_ml_dev_queue_pair_setup() argument
402 rte_ml_dev_stats_get(int16_t dev_id,struct rte_ml_dev_stats * stats) rte_ml_dev_stats_get() argument
425 rte_ml_dev_stats_reset(int16_t dev_id) rte_ml_dev_stats_reset() argument
442 rte_ml_dev_xstats_names_get(int16_t dev_id,enum rte_ml_dev_xstats_mode mode,int32_t model_id,struct rte_ml_dev_xstats_map * xstats_map,uint32_t size) rte_ml_dev_xstats_names_get() argument
460 rte_ml_dev_xstats_by_name_get(int16_t dev_id,const char * name,uint16_t * stat_id,uint64_t * value) rte_ml_dev_xstats_by_name_get() argument
487 rte_ml_dev_xstats_get(int16_t dev_id,enum rte_ml_dev_xstats_mode mode,int32_t model_id,const uint16_t stat_ids[],uint64_t values[],uint16_t nb_ids) rte_ml_dev_xstats_get() argument
515 rte_ml_dev_xstats_reset(int16_t dev_id,enum rte_ml_dev_xstats_mode mode,int32_t model_id,const uint16_t stat_ids[],uint16_t nb_ids) rte_ml_dev_xstats_reset() argument
533 rte_ml_dev_dump(int16_t dev_id,FILE * fd) rte_ml_dev_dump() argument
555 rte_ml_dev_selftest(int16_t dev_id) rte_ml_dev_selftest() argument
572 rte_ml_model_load(int16_t dev_id,struct rte_ml_model_params * params,uint16_t * model_id) rte_ml_model_load() argument
599 rte_ml_model_unload(int16_t dev_id,uint16_t model_id) rte_ml_model_unload() argument
616 rte_ml_model_start(int16_t dev_id,uint16_t model_id) rte_ml_model_start() argument
633 rte_ml_model_stop(int16_t dev_id,uint16_t model_id) rte_ml_model_stop() argument
650 rte_ml_model_info_get(int16_t dev_id,uint16_t model_id,struct rte_ml_model_info * model_info) rte_ml_model_info_get() argument
673 rte_ml_model_params_update(int16_t dev_id,uint16_t model_id,void * buffer) rte_ml_model_params_update() argument
695 rte_ml_io_quantize(int16_t dev_id,uint16_t model_id,struct rte_ml_buff_seg ** dbuffer,struct rte_ml_buff_seg ** qbuffer) rte_ml_io_quantize() argument
723 rte_ml_io_dequantize(int16_t dev_id,uint16_t model_id,struct rte_ml_buff_seg ** qbuffer,struct rte_ml_buff_seg ** dbuffer) rte_ml_io_dequantize() argument
808 rte_ml_enqueue_burst(int16_t dev_id,uint16_t qp_id,struct rte_ml_op ** ops,uint16_t nb_ops) rte_ml_enqueue_burst() argument
844 rte_ml_dequeue_burst(int16_t dev_id,uint16_t qp_id,struct rte_ml_op ** ops,uint16_t nb_ops) rte_ml_dequeue_burst() argument
880 rte_ml_op_error_get(int16_t dev_id,struct rte_ml_op * op,struct rte_ml_op_error * error) rte_ml_op_error_get() argument
[all...]
H A Drte_mldev.h63 * integer >= 0 named as device identifier *dev_id*.
190 * @param dev_id
199 rte_ml_dev_is_valid_dev(int16_t dev_id);
204 * @param dev_id
210 * - -EINVAL: if the dev_id value is not valid.
214 rte_ml_dev_socket_id(int16_t dev_id);
245 * @param dev_id
256 rte_ml_dev_info_get(int16_t dev_id, struct rte_ml_dev_info *dev_info);
285 * @param dev_id
296 rte_ml_dev_configure(int16_t dev_id, cons
[all...]
/dpdk/app/test-bbdev/
H A Dtest_bbdev.c93 uint8_t dev_id; in test_bbdev_configure_invalid_dev_id() local
97 for (dev_id = 0; dev_id < RTE_BBDEV_MAX_DEVS; dev_id++) { in test_bbdev_configure_invalid_dev_id()
98 if (!rte_bbdev_is_valid(dev_id)) { in test_bbdev_configure_invalid_dev_id()
99 TEST_ASSERT_FAIL(rte_bbdev_setup_queues(dev_id, in test_bbdev_configure_invalid_dev_id()
102 "invalid dev_num %u", dev_id); in test_bbdev_configure_invalid_dev_id()
103 TEST_ASSERT(rte_bbdev_intr_enable(dev_id) == -ENODEV, in test_bbdev_configure_invalid_dev_id()
105 "invalid dev_num %u", dev_id); in test_bbdev_configure_invalid_dev_id()
117 uint8_t dev_id, num_devs; in test_bbdev_configure_invalid_num_queues() local
128 dev_id = null_dev_id; in test_bbdev_configure_invalid_num_queues()
131 rte_bbdev_stop(dev_id); in test_bbdev_configure_invalid_num_queues()
[all …]
/dpdk/lib/rawdev/
H A Drte_rawdev.c56 rte_rawdev_socket_id(uint16_t dev_id) in rte_rawdev_socket_id() argument
60 RTE_RAWDEV_VALID_DEVID_OR_ERR_RET(dev_id, -EINVAL); in rte_rawdev_socket_id()
61 dev = &rte_rawdevs[dev_id]; in rte_rawdev_socket_id()
67 rte_rawdev_info_get(uint16_t dev_id, struct rte_rawdev_info *dev_info, in rte_rawdev_info_get() argument
73 RTE_RAWDEV_VALID_DEVID_OR_ERR_RET(dev_id, -EINVAL); in rte_rawdev_info_get()
77 rawdev = &rte_rawdevs[dev_id]; in rte_rawdev_info_get()
95 rte_rawdev_configure(uint16_t dev_id, struct rte_rawdev_info *dev_conf, in rte_rawdev_configure() argument
101 RTE_RAWDEV_VALID_DEVID_OR_ERR_RET(dev_id, -EINVAL); in rte_rawdev_configure()
105 dev = &rte_rawdevs[dev_id]; in rte_rawdev_configure()
112 "device %d must be stopped to allow configuration", dev_id); in rte_rawdev_configure()
[all …]
H A Drte_rawdev.h53 * @param dev_id
58 * -(-EINVAL) dev_id value is out of range.
61 rte_rawdev_socket_id(uint16_t dev_id);
71 * @param dev_id
95 rte_rawdev_info_get(uint16_t dev_id, struct rte_rawdev_info *dev_info,
108 * @param dev_id
128 rte_rawdev_configure(uint16_t dev_id, struct rte_rawdev_info *dev_conf,
139 * @param dev_id
156 rte_rawdev_queue_conf_get(uint16_t dev_id,
164 * @param dev_id
286 uint16_t dev_id; global() member
[all...]
/dpdk/lib/regexdev/
H A Drte_regexdev.c95 uint16_t dev_id; in rte_regexdev_register() local
107 dev_id = regexdev_find_free_dev(); in rte_regexdev_register()
108 if (dev_id == RTE_MAX_REGEXDEV_DEVS) { in rte_regexdev_register()
118 dev = &rte_regex_devices[dev_id]; in rte_regexdev_register()
121 dev->data = &rte_regexdev_shared_data->data[dev_id]; in rte_regexdev_register()
124 dev->data->dev_id = dev_id; in rte_regexdev_register()
167 id = rte_regex_devices[i].data->dev_id; in rte_regexdev_get_dev_id()
175 rte_regexdev_is_valid_dev(uint16_t dev_id) in rte_regexdev_is_valid_dev() argument
177 if (dev_id >= RTE_MAX_REGEXDEV_DEVS || in rte_regexdev_is_valid_dev()
178 rte_regex_devices[dev_id].state != RTE_REGEXDEV_READY) in rte_regexdev_is_valid_dev()
[all …]
H A Drte_regexdev.h111 * designated by an integer >= 0 named the device identifier *dev_id*
211 #define RTE_REGEXDEV_VALID_DEV_ID_OR_ERR_RET(dev_id, retval) do { \
212 if (!rte_regexdev_is_valid_dev(dev_id)) { \
213 RTE_REGEXDEV_LOG_LINE(ERR, "Invalid dev_id=%u", dev_id); \
218 #define RTE_REGEXDEV_VALID_DEV_ID_OR_RET(dev_id) do { \
219 if (!rte_regexdev_is_valid_dev(dev_id)) { \
220 RTE_REGEXDEV_LOG_LINE(ERR, "Invalid dev_id=%u", dev_id); \
229 * Check if dev_id i
215 RTE_REGEXDEV_VALID_DEV_ID_OR_ERR_RET(dev_id,retval) global() argument
222 RTE_REGEXDEV_VALID_DEV_ID_OR_RET(dev_id) global() argument
1470 rte_regexdev_enqueue_burst(uint8_t dev_id,uint16_t qp_id,struct rte_regex_ops ** ops,uint16_t nb_ops) rte_regexdev_enqueue_burst() argument
1530 rte_regexdev_dequeue_burst(uint8_t dev_id,uint16_t qp_id,struct rte_regex_ops ** ops,uint16_t nb_ops) rte_regexdev_dequeue_burst() argument
[all...]
/dpdk/lib/compressdev/
H A Drte_compressdev.c32 rte_compressdev_capability_get(uint8_t dev_id, in rte_compressdev_capability_get() argument
39 if (dev_id >= compressdev_globals.nb_devs) { in rte_compressdev_capability_get()
40 COMPRESSDEV_LOG(ERR, "Invalid dev_id=%d", dev_id); in rte_compressdev_capability_get()
43 rte_compressdev_info_get(dev_id, &dev_info); in rte_compressdev_capability_get()
78 rte_compressdev_get_dev(uint8_t dev_id) in rte_compressdev_get_dev() argument
80 return &compressdev_globals.devs[dev_id]; in rte_compressdev_get_dev()
104 rte_compressdev_is_valid_dev(uint8_t dev_id) in rte_compressdev_is_valid_dev() argument
108 if (dev_id >= compressdev_globals.nb_devs) in rte_compressdev_is_valid_dev()
111 dev = rte_compressdev_get_dev(dev_id); in rte_compressdev_is_valid_dev()
161 devices[count++] = devs[i].data->dev_id; in rte_compressdev_devices_get()
[all …]
H A Drte_compressdev.h52 rte_compressdev_capability_get(uint8_t dev_id,
134 * @param dev_id
141 rte_compressdev_name_get(uint8_t dev_id);
174 * @param dev_id
179 * -1 if returned is the dev_id value is out of range.
182 rte_compressdev_socket_id(uint8_t dev_id);
203 * @param dev_id
212 rte_compressdev_configure(uint8_t dev_id,
223 * @param dev_id
230 rte_compressdev_start(uint8_t dev_id);
[all...]
/dpdk/lib/bbdev/
H A Drte_bbdev.c41 /* Helper macro to check dev_id is valid */
42 #define VALID_DEV_OR_RET_ERR(dev, dev_id) do { \ argument
44 rte_bbdev_log(ERR, "device %u is invalid", dev_id); \
50 #define VALID_DEV_OPS_OR_RET_ERR(dev, dev_id) do { \ argument
53 dev_id); \
59 #define VALID_FUNC_OR_RET_ERR(func, dev_id) do { \ argument
62 dev_id, #func); \
71 queue_id, dev->data->dev_id); \
106 get_dev(uint16_t dev_id) in get_dev() argument
108 if (rte_bbdev_is_valid(dev_id)) in get_dev()
179 uint16_t dev_id; rte_bbdev_allocate() local
236 uint16_t dev_id; rte_bbdev_release() local
295 rte_bbdev_is_valid(uint16_t dev_id) rte_bbdev_is_valid() argument
304 rte_bbdev_find_next(uint16_t dev_id) rte_bbdev_find_next() argument
314 rte_bbdev_setup_queues(uint16_t dev_id,uint16_t num_queues,int socket_id) rte_bbdev_setup_queues() argument
404 rte_bbdev_intr_enable(uint16_t dev_id) rte_bbdev_intr_enable() argument
436 rte_bbdev_queue_configure(uint16_t dev_id,uint16_t queue_id,const struct rte_bbdev_queue_conf * conf) rte_bbdev_queue_configure() argument
552 rte_bbdev_start(uint16_t dev_id) rte_bbdev_start() argument
584 rte_bbdev_stop(uint16_t dev_id) rte_bbdev_stop() argument
605 rte_bbdev_close(uint16_t dev_id) rte_bbdev_close() argument
650 rte_bbdev_queue_start(uint16_t dev_id,uint16_t queue_id) rte_bbdev_queue_start() argument
680 rte_bbdev_queue_stop(uint16_t dev_id,uint16_t queue_id) rte_bbdev_queue_stop() argument
742 rte_bbdev_stats_get(uint16_t dev_id,struct rte_bbdev_stats * stats) rte_bbdev_stats_get() argument
765 rte_bbdev_stats_reset(uint16_t dev_id) rte_bbdev_stats_reset() argument
782 rte_bbdev_info_get(uint16_t dev_id,struct rte_bbdev_info * dev_info) rte_bbdev_info_get() argument
810 rte_bbdev_queue_info_get(uint16_t dev_id,uint16_t queue_id,struct rte_bbdev_queue_info * queue_info) rte_bbdev_queue_info_get() argument
943 rte_bbdev_callback_register(uint16_t dev_id,enum rte_bbdev_event_type event,rte_bbdev_cb_fn cb_fn,void * cb_arg) rte_bbdev_callback_register() argument
988 rte_bbdev_callback_unregister(uint16_t dev_id,enum rte_bbdev_event_type event,rte_bbdev_cb_fn cb_fn,void * cb_arg) rte_bbdev_callback_unregister() argument
1075 rte_bbdev_queue_intr_enable(uint16_t dev_id,uint16_t queue_id) rte_bbdev_queue_intr_enable() argument
1086 rte_bbdev_queue_intr_disable(uint16_t dev_id,uint16_t queue_id) rte_bbdev_queue_intr_disable() argument
1097 rte_bbdev_queue_intr_ctl(uint16_t dev_id,uint16_t queue_id,int epfd,int op,void * data) rte_bbdev_queue_intr_ctl() argument
[all...]
H A Drte_bbdev.h64 * @param dev_id
71 rte_bbdev_is_valid(uint16_t dev_id);
76 * @param dev_id
84 rte_bbdev_find_next(uint16_t dev_id);
98 * @param dev_id
107 * - -ENODEV if dev_id is invalid or the device is corrupted
113 rte_bbdev_setup_queues(uint16_t dev_id, uint16_t num_queues, int socket_id);
120 * @param dev_id
125 * - -ENODEV if dev_id is invalid or the device is corrupted
130 rte_bbdev_intr_enable(uint16_t dev_id);
488 uint16_t dev_id; /**< Device ID */ global() member
565 rte_bbdev_enqueue_enc_ops(uint16_t dev_id,uint16_t queue_id,struct rte_bbdev_enc_op ** ops,uint16_t num_ops) rte_bbdev_enqueue_enc_ops() argument
595 rte_bbdev_enqueue_dec_ops(uint16_t dev_id,uint16_t queue_id,struct rte_bbdev_dec_op ** ops,uint16_t num_ops) rte_bbdev_enqueue_dec_ops() argument
625 rte_bbdev_enqueue_ldpc_enc_ops(uint16_t dev_id,uint16_t queue_id,struct rte_bbdev_enc_op ** ops,uint16_t num_ops) rte_bbdev_enqueue_ldpc_enc_ops() argument
655 rte_bbdev_enqueue_ldpc_dec_ops(uint16_t dev_id,uint16_t queue_id,struct rte_bbdev_dec_op ** ops,uint16_t num_ops) rte_bbdev_enqueue_ldpc_dec_ops() argument
685 rte_bbdev_enqueue_fft_ops(uint16_t dev_id,uint16_t queue_id,struct rte_bbdev_fft_op ** ops,uint16_t num_ops) rte_bbdev_enqueue_fft_ops() argument
715 rte_bbdev_enqueue_mldts_ops(uint16_t dev_id,uint16_t queue_id,struct rte_bbdev_mldts_op ** ops,uint16_t num_ops) rte_bbdev_enqueue_mldts_ops() argument
746 rte_bbdev_dequeue_enc_ops(uint16_t dev_id,uint16_t queue_id,struct rte_bbdev_enc_op ** ops,uint16_t num_ops) rte_bbdev_dequeue_enc_ops() argument
778 rte_bbdev_dequeue_dec_ops(uint16_t dev_id,uint16_t queue_id,struct rte_bbdev_dec_op ** ops,uint16_t num_ops) rte_bbdev_dequeue_dec_ops() argument
809 rte_bbdev_dequeue_ldpc_enc_ops(uint16_t dev_id,uint16_t queue_id,struct rte_bbdev_enc_op ** ops,uint16_t num_ops) rte_bbdev_dequeue_ldpc_enc_ops() argument
839 rte_bbdev_dequeue_ldpc_dec_ops(uint16_t dev_id,uint16_t queue_id,struct rte_bbdev_dec_op ** ops,uint16_t num_ops) rte_bbdev_dequeue_ldpc_dec_ops() argument
869 rte_bbdev_dequeue_fft_ops(uint16_t dev_id,uint16_t queue_id,struct rte_bbdev_fft_op ** ops,uint16_t num_ops) rte_bbdev_dequeue_fft_ops() argument
900 rte_bbdev_dequeue_mldts_ops(uint16_t dev_id,uint16_t queue_id,struct rte_bbdev_mldts_op ** ops,uint16_t num_ops) rte_bbdev_dequeue_mldts_ops() argument
[all...]
/dpdk/lib/eventdev/
H A Drte_eventdev.c78 rte_event_dev_socket_id(uint8_t dev_id) in rte_event_dev_socket_id() argument
82 RTE_EVENTDEV_VALID_DEVID_OR_ERR_RET(dev_id, -EINVAL); in rte_event_dev_socket_id()
83 dev = &rte_eventdevs[dev_id]; in rte_event_dev_socket_id()
85 rte_eventdev_trace_socket_id(dev_id, dev, dev->data->socket_id); in rte_event_dev_socket_id()
91 rte_event_dev_info_get(uint8_t dev_id, struct rte_event_dev_info *dev_info) in rte_event_dev_info_get() argument
95 RTE_EVENTDEV_VALID_DEVID_OR_ERR_RET(dev_id, -EINVAL); in rte_event_dev_info_get()
96 dev = &rte_eventdevs[dev_id]; in rte_event_dev_info_get()
113 rte_eventdev_trace_info_get(dev_id, dev_info, dev_info->dev); in rte_event_dev_info_get()
119 rte_event_eth_rx_adapter_caps_get(uint8_t dev_id, uint16_t eth_port_id, in rte_event_eth_rx_adapter_caps_get() argument
124 rte_eventdev_trace_eth_rx_adapter_caps_get(dev_id, eth_port_id); in rte_event_eth_rx_adapter_caps_get()
[all …]
H A Drte_eventdev.h125 * designated by an integer >= 0 named the device identifier *dev_id*
175 * id (dev_id) for application use.
574 * Event device identifier (dev_id >= 0) on success.
585 * @param dev_id
590 * -EINVAL on error, where the given dev_id value does not
594 rte_event_dev_socket_id(uint8_t dev_id);
689 * @param dev_id
703 rte_event_dev_info_get(uint8_t dev_id, struct rte_event_dev_info *dev_info);
721 * @param dev_id Eventdev id
731 rte_event_dev_attr_get(uint8_t dev_id, uint32_
2473 __rte_event_enqueue_burst(uint8_t dev_id,uint8_t port_id,const struct rte_event ev[],uint16_t nb_events,const event_enqueue_burst_t fn) __rte_event_enqueue_burst() argument
2549 rte_event_enqueue_burst(uint8_t dev_id,uint8_t port_id,const struct rte_event ev[],uint16_t nb_events) rte_event_enqueue_burst() argument
2601 rte_event_enqueue_new_burst(uint8_t dev_id,uint8_t port_id,const struct rte_event ev[],uint16_t nb_events) rte_event_enqueue_new_burst() argument
2653 rte_event_enqueue_forward_burst(uint8_t dev_id,uint8_t port_id,const struct rte_event ev[],uint16_t nb_events) rte_event_enqueue_forward_burst() argument
2730 rte_event_dequeue_burst(uint8_t dev_id,uint8_t port_id,struct rte_event ev[],uint16_t nb_events,uint64_t timeout_ticks) rte_event_dequeue_burst() argument
2805 rte_event_maintain(uint8_t dev_id,uint8_t port_id,int op) rte_event_maintain() argument
2853 rte_event_port_profile_switch(uint8_t dev_id,uint8_t port_id,uint8_t profile_id) rte_event_port_profile_switch() argument
[all...]
H A Deventdev_trace.h27 RTE_TRACE_POINT_ARGS(uint8_t dev_id,
29 rte_trace_point_emit_u8(dev_id);
44 RTE_TRACE_POINT_ARGS(uint8_t dev_id, uint8_t queue_id,
46 rte_trace_point_emit_u8(dev_id);
57 RTE_TRACE_POINT_ARGS(uint8_t dev_id, uint8_t port_id,
59 rte_trace_point_emit_u8(dev_id);
71 RTE_TRACE_POINT_ARGS(uint8_t dev_id, uint8_t port_id,
73 rte_trace_point_emit_u8(dev_id);
81 RTE_TRACE_POINT_ARGS(uint8_t dev_id, uint8_t port_id,
83 rte_trace_point_emit_u8(dev_id);
[all...]
/dpdk/lib/gpudev/
H A Dgpudev.c86 rte_gpu_is_valid(int16_t dev_id) in rte_gpu_is_valid() argument
88 if (dev_id >= 0 && dev_id < gpu_max && in rte_gpu_is_valid()
89 gpus[dev_id].process_state == RTE_GPU_STATE_INITIALIZED) in rte_gpu_is_valid()
95 gpu_match_parent(int16_t dev_id, int16_t parent) in gpu_match_parent() argument
99 return gpus[dev_id].mpshared->info.parent == parent; in gpu_match_parent()
103 rte_gpu_find_next(int16_t dev_id, int16_t parent) in rte_gpu_find_next() argument
105 if (dev_id < 0) in rte_gpu_find_next()
106 dev_id = 0; in rte_gpu_find_next()
107 while (dev_id < gpu_max && in rte_gpu_find_next()
108 (gpus[dev_id].process_state == RTE_GPU_STATE_UNUSED || in rte_gpu_find_next()
[all …]
H A Drte_gpudev.h55 int16_t dev_id; member
77 typedef void (rte_gpu_callback_t)(int16_t dev_id,
89 uint16_t dev_id; member
122 uint16_t dev_id; member
180 bool rte_gpu_is_valid(int16_t dev_id);
227 int16_t rte_gpu_find_next(int16_t dev_id, int16_t parent);
238 #define RTE_GPU_FOREACH(dev_id) \ argument
239 RTE_GPU_FOREACH_CHILD(dev_id, RTE_GPU_ID_ANY)
250 #define RTE_GPU_FOREACH_PARENT(dev_id) \ argument
251 RTE_GPU_FOREACH_CHILD(dev_id, RTE_GPU_ID_NONE)
[all …]
/dpdk/lib/cryptodev/
H A Dcryptodev_trace.h24 RTE_TRACE_POINT_ARGS(uint8_t dev_id,
26 rte_trace_point_emit_u8(dev_id);
33 RTE_TRACE_POINT_ARGS(uint8_t dev_id, int rc),
34 rte_trace_point_emit_u8(dev_id);
40 RTE_TRACE_POINT_ARGS(uint8_t dev_id),
41 rte_trace_point_emit_u8(dev_id);
46 RTE_TRACE_POINT_ARGS(uint8_t dev_id, int rc),
47 rte_trace_point_emit_u8(dev_id);
53 RTE_TRACE_POINT_ARGS(uint8_t dev_id, uint16_t queue_pair_id,
55 rte_trace_point_emit_u8(dev_id);
[all...]
H A Drte_cryptodev.c384 rte_cryptodev_sym_capability_get(uint8_t dev_id, in rte_cryptodev_sym_capability_get()
392 rte_cryptodev_info_get(dev_id, &dev_info); in rte_cryptodev_sym_capability_get()
421 rte_cryptodev_trace_sym_capability_get(dev_id, dev_info.driver_name, in rte_cryptodev_sym_capability_get()
453 rte_cryptodev_asym_capability_get(uint8_t dev_id, in rte_cryptodev_asym_capability_get()
462 rte_cryptodev_info_get(dev_id, &dev_info); in rte_cryptodev_asym_capability_get()
829 rte_cryptodev_pmd_get_dev(uint8_t dev_id) in rte_cryptodev_pmd_get_named_dev()
831 return &cryptodev_globals.devs[dev_id]; in rte_cryptodev_pmd_get_named_dev()
855 rte_cryptodev_is_valid_device_data(uint8_t dev_id) in rte_cryptodev_is_valid_dev()
857 if (dev_id >= RTE_CRYPTO_MAX_DEVS || in rte_cryptodev_is_valid_dev()
858 rte_crypto_devices[dev_id] in rte_cryptodev_is_valid_dev()
383 rte_cryptodev_sym_capability_get(uint8_t dev_id,const struct rte_cryptodev_sym_capability_idx * idx) rte_cryptodev_sym_capability_get() argument
452 rte_cryptodev_asym_capability_get(uint8_t dev_id,const struct rte_cryptodev_asym_capability_idx * idx) rte_cryptodev_asym_capability_get() argument
812 rte_cryptodev_pmd_get_dev(uint8_t dev_id) rte_cryptodev_pmd_get_dev() argument
838 rte_cryptodev_is_valid_device_data(uint8_t dev_id) rte_cryptodev_is_valid_device_data() argument
848 rte_cryptodev_is_valid_dev(uint8_t dev_id) rte_cryptodev_is_valid_dev() argument
947 rte_cryptodev_get_sec_ctx(uint8_t dev_id) rte_cryptodev_get_sec_ctx() argument
962 rte_cryptodev_socket_id(uint8_t dev_id) rte_cryptodev_socket_id() argument
977 rte_cryptodev_data_alloc(uint8_t dev_id,struct rte_cryptodev_data ** data,int socket_id) rte_cryptodev_data_alloc() argument
1012 rte_cryptodev_data_free(uint8_t dev_id,struct rte_cryptodev_data ** data) rte_cryptodev_data_free() argument
1045 uint8_t dev_id; rte_cryptodev_find_free_device_index() local
1059 uint8_t dev_id; rte_cryptodev_pmd_allocate() local
1118 uint8_t dev_id; rte_cryptodev_pmd_release_device() local
1144 rte_cryptodev_queue_pair_count(uint8_t dev_id) rte_cryptodev_queue_pair_count() argument
1226 rte_cryptodev_configure(uint8_t dev_id,struct rte_cryptodev_config * config) rte_cryptodev_configure() argument
1273 rte_cryptodev_start(uint8_t dev_id) rte_cryptodev_start() argument
1310 rte_cryptodev_stop(uint8_t dev_id) rte_cryptodev_stop() argument
1339 rte_cryptodev_close(uint8_t dev_id) rte_cryptodev_close() argument
1381 rte_cryptodev_get_qp_status(uint8_t dev_id,uint16_t queue_pair_id) rte_cryptodev_get_qp_status() argument
1435 rte_cryptodev_queue_pair_setup(uint8_t dev_id,uint16_t queue_pair_id,const struct rte_cryptodev_qp_conf * qp_conf,int socket_id) rte_cryptodev_queue_pair_setup() argument
1489 rte_cryptodev_add_enq_callback(uint8_t dev_id,uint16_t qp_id,rte_cryptodev_callback_fn cb_fn,void * cb_arg) rte_cryptodev_add_enq_callback() argument
1559 rte_cryptodev_remove_enq_callback(uint8_t dev_id,uint16_t qp_id,struct rte_cryptodev_cb * cb) rte_cryptodev_remove_enq_callback() argument
1635 rte_cryptodev_add_deq_callback(uint8_t dev_id,uint16_t qp_id,rte_cryptodev_callback_fn cb_fn,void * cb_arg) rte_cryptodev_add_deq_callback() argument
1706 rte_cryptodev_remove_deq_callback(uint8_t dev_id,uint16_t qp_id,struct rte_cryptodev_cb * cb) rte_cryptodev_remove_deq_callback() argument
1782 rte_cryptodev_stats_get(uint8_t dev_id,struct rte_cryptodev_stats * stats) rte_cryptodev_stats_get() argument
1808 rte_cryptodev_stats_reset(uint8_t dev_id) rte_cryptodev_stats_reset() argument
1827 rte_cryptodev_info_get(uint8_t dev_id,struct rte_cryptodev_info * dev_info) rte_cryptodev_info_get() argument
1852 rte_cryptodev_callback_register(uint8_t dev_id,enum rte_cryptodev_event_type event,rte_cryptodev_cb_fn cb_fn,void * cb_arg) rte_cryptodev_callback_register() argument
1897 rte_cryptodev_callback_unregister(uint8_t dev_id,enum rte_cryptodev_event_type event,rte_cryptodev_cb_fn cb_fn,void * cb_arg) rte_cryptodev_callback_unregister() argument
1967 rte_cryptodev_queue_pair_event_error_query(uint8_t dev_id,uint16_t qp_id) rte_cryptodev_queue_pair_event_error_query() argument
2030 uint8_t dev_id; rte_cryptodev_asym_session_pool_create() local
2074 rte_cryptodev_sym_session_create(uint8_t dev_id,struct rte_crypto_sym_xform * xforms,struct rte_mempool * mp) rte_cryptodev_sym_session_create() argument
2141 rte_cryptodev_asym_session_create(uint8_t dev_id,struct rte_crypto_asym_xform * xforms,struct rte_mempool * mp,void ** session) rte_cryptodev_asym_session_create() argument
2217 rte_cryptodev_sym_session_free(uint8_t dev_id,void * _sess) rte_cryptodev_sym_session_free() argument
2263 rte_cryptodev_asym_session_free(uint8_t dev_id,void * sess) rte_cryptodev_asym_session_free() argument
2300 rte_cryptodev_sym_get_private_session_size(uint8_t dev_id) rte_cryptodev_sym_get_private_session_size() argument
2322 rte_cryptodev_asym_get_private_session_size(uint8_t dev_id) rte_cryptodev_asym_get_private_session_size() argument
2422 rte_cryptodev_sym_cpu_crypto_process(uint8_t dev_id,void * _sess,union rte_crypto_sym_ofs ofs,struct rte_crypto_sym_vec * vec) rte_cryptodev_sym_cpu_crypto_process() argument
2448 rte_cryptodev_get_raw_dp_ctx_size(uint8_t dev_id) rte_cryptodev_get_raw_dp_ctx_size() argument
2474 rte_cryptodev_configure_raw_dp_ctx(uint8_t dev_id,uint16_t qp_id,struct rte_crypto_raw_dp_ctx * ctx,enum rte_crypto_op_sess_type sess_type,union rte_cryptodev_session_ctx session_ctx,uint8_t is_update) rte_cryptodev_configure_raw_dp_ctx() argument
2497 rte_cryptodev_session_event_mdata_set(uint8_t dev_id,void * sess,enum rte_crypto_op_type op_type,enum rte_crypto_op_sess_type sess_type,void * ev_mdata,uint16_t size) rte_cryptodev_session_event_mdata_set() argument
2719 rte_cryptodev_name_get(uint8_t dev_id) rte_cryptodev_name_get() argument
2779 int dev_id; cryptodev_handle_dev_list() local
2797 int dev_id; cryptodev_handle_dev_info() local
2828 int dev_id, ret; cryptodev_handle_dev_stats() local
2887 int dev_id; cryptodev_handle_dev_caps() local
[all...]
/dpdk/app/test/
H A Dtest_dmadev.c71 int (*test_fn)(int16_t dev_id, uint16_t vchan);
78 int (*test_fn)(int16_t dev_id, uint16_t vchan); in runtest()
83 int16_t dev_id; in runtest() local
89 dev_id = test_dev_id; in runtest()
91 rte_dma_stats_reset(dev_id, vchan); in runtest()
92 printf("DMA Dev %d: Running %s Tests %s\n", dev_id, printable, in runtest()
95 if (test_fn(dev_id, vchan) < 0) in runtest()
98 rte_dma_stats_get(dev_id, 0, &stats); in runtest()
113 await_hw(int16_t dev_id, uint16_t vchan) in await_hw() argument
117 if (rte_dma_vchan_status(dev_id, vchan, &st) < 0) { in await_hw()
[all …]
/dpdk/lib/dmadev/
H A Drte_dmadev.c60 int16_t dev_id = start_dev_id; in rte_dma_next_dev() local
61 while (dev_id < dma_devices_max && rte_dma_devices[dev_id].state == RTE_DMA_DEV_UNUSED) in rte_dma_next_dev()
62 dev_id++; in rte_dma_next_dev()
64 if (dev_id < dma_devices_max) in rte_dma_next_dev()
65 return dev_id; in rte_dma_next_dev()
252 int16_t dev_id; in dma_allocate_primary() local
274 dev_id = dma_find_free_id(); in dma_allocate_primary()
275 if (dev_id < 0) { in dma_allocate_primary()
281 dev = &rte_dma_devices[dev_id]; in dma_allocate_primary()
403 rte_dma_is_valid(int16_t dev_id) rte_dma_is_valid() argument
411 rte_dma_pmd_get_dev_by_id(int16_t dev_id) rte_dma_pmd_get_dev_by_id() argument
437 rte_dma_info_get(int16_t dev_id,struct rte_dma_info * dev_info) rte_dma_info_get() argument
463 rte_dma_configure(int16_t dev_id,const struct rte_dma_conf * dev_conf) rte_dma_configure() argument
514 rte_dma_start(int16_t dev_id) rte_dma_start() argument
546 rte_dma_stop(int16_t dev_id) rte_dma_stop() argument
573 rte_dma_close(int16_t dev_id) rte_dma_close() argument
600 rte_dma_vchan_setup(int16_t dev_id,uint16_t vchan,const struct rte_dma_vchan_conf * conf) rte_dma_vchan_setup() argument
694 rte_dma_stats_get(int16_t dev_id,uint16_t vchan,struct rte_dma_stats * stats) rte_dma_stats_get() argument
716 rte_dma_stats_reset(int16_t dev_id,uint16_t vchan) rte_dma_stats_reset() argument
740 rte_dma_vchan_status(int16_t dev_id,uint16_t vchan,enum rte_dma_vchan_status * status) rte_dma_vchan_status() argument
805 rte_dma_dump(int16_t dev_id,FILE * f) rte_dma_dump() argument
920 int dev_id; dmadev_handle_dev_list() local
938 int dev_id, ret; dmadev_handle_dev_info() local
994 int dev_id, ret, vchan_id; dmadev_handle_dev_stats() local
1042 int dev_id, ret; dmadev_handle_dev_dump() local
[all...]
H A Drte_dmadev.h48 * This framework uses 'int16_t dev_id' as the device identifier of a dmadev,
186 * Check whether the dev_id is valid.
188 * @param dev_id
194 bool rte_dma_is_valid(int16_t dev_id);
316 * @param dev_id
325 int rte_dma_info_get(int16_t dev_id, struct rte_dma_info *dev_info);
363 * @param dev_id
372 int rte_dma_configure(int16_t dev_id, const struct rte_dma_conf *dev_conf);
380 * @param dev_id
386 int rte_dma_start(int16_t dev_id);
836 rte_dma_copy(int16_t dev_id,uint16_t vchan,rte_iova_t src,rte_iova_t dst,uint32_t length,uint64_t flags) rte_dma_copy() argument
886 rte_dma_copy_sg(int16_t dev_id,uint16_t vchan,struct rte_dma_sge * src,struct rte_dma_sge * dst,uint16_t nb_src,uint16_t nb_dst,uint64_t flags) rte_dma_copy_sg() argument
936 rte_dma_fill(int16_t dev_id,uint16_t vchan,uint64_t pattern,rte_iova_t dst,uint32_t length,uint64_t flags) rte_dma_fill() argument
971 rte_dma_submit(int16_t dev_id,uint16_t vchan) rte_dma_submit() argument
1012 rte_dma_completed(int16_t dev_id,uint16_t vchan,const uint16_t nb_cpls,uint16_t * last_idx,bool * has_error) rte_dma_completed() argument
1075 rte_dma_completed_status(int16_t dev_id,uint16_t vchan,const uint16_t nb_cpls,uint16_t * last_idx,enum rte_dma_status_code * status) rte_dma_completed_status() argument
1113 rte_dma_burst_capacity(int16_t dev_id,uint16_t vchan) rte_dma_burst_capacity() argument
[all...]
/dpdk/drivers/raw/cnxk_gpio/
H A Drte_pmd_cnxk_gpio.h109 __rte_pmd_gpio_enq_deq(uint16_t dev_id, int gpio, void *req, void *rsp, in __rte_pmd_gpio_enq_deq() argument
121 ret = rte_rawdev_enqueue_buffers(dev_id, bufs, RTE_DIM(bufs), q); in __rte_pmd_gpio_enq_deq()
130 ret = rte_rawdev_dequeue_buffers(dev_id, bufs, RTE_DIM(bufs), q); in __rte_pmd_gpio_enq_deq()
157 rte_pmd_gpio_set_pin_value(uint16_t dev_id, int gpio, int val) in rte_pmd_gpio_set_pin_value() argument
164 return __rte_pmd_gpio_enq_deq(dev_id, gpio, &msg, NULL, 0); in rte_pmd_gpio_set_pin_value()
181 rte_pmd_gpio_set_pin_edge(uint16_t dev_id, int gpio, in rte_pmd_gpio_set_pin_edge() argument
189 return __rte_pmd_gpio_enq_deq(dev_id, gpio, &msg, NULL, 0); in rte_pmd_gpio_set_pin_edge()
206 rte_pmd_gpio_set_pin_dir(uint16_t dev_id, int gpio, enum cnxk_gpio_pin_dir dir) in rte_pmd_gpio_set_pin_dir() argument
213 return __rte_pmd_gpio_enq_deq(dev_id, gpio, &msg, NULL, 0); in rte_pmd_gpio_set_pin_dir()
236 rte_pmd_gpio_set_pin_active_low(uint16_t dev_id, int gpio, int val) in rte_pmd_gpio_set_pin_active_low() argument
[all …]
H A Dcnxk_gpio_selftest.c84 cnxk_gpio_test_input(uint16_t dev_id, int base, int gpio) in cnxk_gpio_test_input() argument
92 ret = rte_pmd_gpio_set_pin_dir(dev_id, gpio, CNXK_GPIO_PIN_DIR_IN); in cnxk_gpio_test_input()
97 ret = rte_pmd_gpio_set_pin_value(dev_id, gpio, 1) | in cnxk_gpio_test_input()
98 rte_pmd_gpio_set_pin_value(dev_id, gpio, 0); in cnxk_gpio_test_input()
106 ret = rte_pmd_gpio_set_pin_edge(dev_id, gpio, in cnxk_gpio_test_input()
112 ret = rte_pmd_gpio_set_pin_edge(dev_id, gpio, in cnxk_gpio_test_input()
118 ret = rte_pmd_gpio_set_pin_edge(dev_id, gpio, CNXK_GPIO_PIN_EDGE_BOTH); in cnxk_gpio_test_input()
123 ret = rte_pmd_gpio_set_pin_edge(dev_id, gpio, CNXK_GPIO_PIN_EDGE_NONE); in cnxk_gpio_test_input()
132 rte_pmd_gpio_set_pin_dir(dev_id, gpio, CNXK_GPIO_PIN_DIR_IN); in cnxk_gpio_test_input()
155 cnxk_gpio_test_irq(uint16_t dev_id, int gpio) in cnxk_gpio_test_irq() argument
[all …]
/dpdk/drivers/raw/cnxk_bphy/
H A Dcnxk_bphy_cgx_test.c15 cnxk_bphy_cgx_link_cond(uint16_t dev_id, unsigned int queue, int cond) in cnxk_bphy_cgx_link_cond() argument
21 ret = rte_pmd_bphy_cgx_get_link_info(dev_id, queue, &link_info); in cnxk_bphy_cgx_link_cond()
38 cnxk_bphy_cgx_dev_selftest(uint16_t dev_id) in cnxk_bphy_cgx_dev_selftest() argument
44 queues = rte_rawdev_queue_count(dev_id); in cnxk_bphy_cgx_dev_selftest()
48 ret = rte_rawdev_start(dev_id); in cnxk_bphy_cgx_dev_selftest()
56 ret = rte_rawdev_queue_conf_get(dev_id, i, &descs, in cnxk_bphy_cgx_dev_selftest()
68 ret = rte_pmd_bphy_cgx_stop_rxtx(dev_id, i); in cnxk_bphy_cgx_dev_selftest()
74 ret = rte_pmd_bphy_cgx_start_rxtx(dev_id, i); in cnxk_bphy_cgx_dev_selftest()
81 ret = rte_pmd_bphy_cgx_set_link_state(dev_id, i, &link_state); in cnxk_bphy_cgx_dev_selftest()
87 ret = cnxk_bphy_cgx_link_cond(dev_id, i, 0); in cnxk_bphy_cgx_dev_selftest()
[all …]
H A Dcnxk_bphy_irq.c14 cnxk_bphy_get_bphy_dev_by_dev_id(uint16_t dev_id) in cnxk_bphy_get_bphy_dev_by_dev_id() argument
18 if (!rte_rawdev_pmd_is_valid_dev(dev_id)) in cnxk_bphy_get_bphy_dev_by_dev_id()
21 rawdev = &rte_rawdevs[dev_id]; in cnxk_bphy_get_bphy_dev_by_dev_id()
27 cnxk_bphy_irq_max_get(uint16_t dev_id) in cnxk_bphy_irq_max_get() argument
32 bphy_dev = cnxk_bphy_get_bphy_dev_by_dev_id(dev_id); in cnxk_bphy_irq_max_get()
39 cnxk_bphy_intr_init(uint16_t dev_id) in cnxk_bphy_intr_init() argument
41 struct bphy_device *bphy_dev = cnxk_bphy_get_bphy_dev_by_dev_id(dev_id); in cnxk_bphy_intr_init()
51 cnxk_bphy_intr_fini(uint16_t dev_id) in cnxk_bphy_intr_fini() argument
53 struct bphy_device *bphy_dev = cnxk_bphy_get_bphy_dev_by_dev_id(dev_id); in cnxk_bphy_intr_fini()
61 cnxk_bphy_intr_register(uint16_t dev_id, int irq_num, in cnxk_bphy_intr_register() argument
[all …]

1234567