| /dpdk/drivers/event/dlb2/ |
| H A D | dlb2_selftest.c | 285 struct rte_event_queue_conf queue_conf; in test_single_link() local 340 if (rte_event_queue_default_conf_get(evdev, 0, &queue_conf)) { in test_single_link() 345 queue_conf.event_queue_cfg = RTE_EVENT_QUEUE_CFG_SINGLE_LINK; in test_single_link() 347 if (rte_event_queue_setup(evdev, 0, &queue_conf) < 0) { in test_single_link() 353 if (rte_event_queue_setup(evdev, 1, &queue_conf) == 0) { in test_single_link() 359 queue_conf.event_queue_cfg = 0; in test_single_link() 361 if (rte_event_queue_setup(evdev, 1, &queue_conf) < 0) { in test_single_link() 494 struct rte_event_queue_conf queue_conf; in test_reconfiguration_link() local 536 if (rte_event_queue_default_conf_get(evdev, 0, &queue_conf)) { in test_reconfiguration_link() 542 if (rte_event_queue_setup(evdev, i, &queue_conf) < 0) { in test_reconfiguration_link() [all …]
|
| /dpdk/lib/eventdev/ |
| H A D | eventdev_trace.h | 45 const struct rte_event_queue_conf *queue_conf), 48 rte_trace_point_emit_u32(queue_conf->nb_atomic_flows); 49 rte_trace_point_emit_u32(queue_conf->nb_atomic_order_sequences); 50 rte_trace_point_emit_u32(queue_conf->event_queue_cfg); 51 rte_trace_point_emit_u8(queue_conf->schedule_type); 52 rte_trace_point_emit_u8(queue_conf->priority); 150 const struct rte_event_eth_rx_adapter_queue_conf *queue_conf, 155 rte_trace_point_emit_u32(queue_conf->rx_queue_flags); 156 rte_trace_point_emit_u16(queue_conf->servicing_weight); 157 rte_trace_point_emit_u8(queue_conf [all...] |
| H A D | rte_eventdev.c | 552 struct rte_event_queue_conf *queue_conf) in rte_event_queue_default_conf_get() argument 559 if (queue_conf == NULL) in rte_event_queue_default_conf_get() 569 memset(queue_conf, 0, sizeof(struct rte_event_queue_conf)); in rte_event_queue_default_conf_get() 570 (*dev->dev_ops->queue_def_conf)(dev, queue_id, queue_conf); in rte_event_queue_default_conf_get() 572 rte_eventdev_trace_queue_default_conf_get(dev_id, dev, queue_id, queue_conf); in rte_event_queue_default_conf_get() 578 is_valid_atomic_queue_conf(const struct rte_event_queue_conf *queue_conf) in is_valid_atomic_queue_conf() argument 580 if (queue_conf && in is_valid_atomic_queue_conf() 581 !(queue_conf->event_queue_cfg & in is_valid_atomic_queue_conf() 583 ((queue_conf->event_queue_cfg & in is_valid_atomic_queue_conf() 585 (queue_conf->schedule_type in is_valid_atomic_queue_conf() [all …]
|
| H A D | rte_event_eth_rx_adapter.c | 2198 const struct rte_event_eth_rx_adapter_queue_conf *queue_conf) in rxa_sw_add() argument 2212 if (queue_conf->servicing_weight == 0) { in rxa_sw_add() 2215 temp_conf = *queue_conf; in rxa_sw_add() 2220 queue_conf = &temp_conf; in rxa_sw_add() 2222 if (queue_conf->servicing_weight == 0 && in rxa_sw_add() 2235 wt = queue_conf->servicing_weight; in rxa_sw_add() 2250 queue_conf->servicing_weight, in rxa_sw_add() 2310 ret = rxa_add_queue(rx_adapter, dev_info, rx_queue_id, queue_conf); in rxa_sw_add() 2646 const struct rte_event_eth_rx_adapter_queue_conf *queue_conf) in rte_event_eth_rx_adapter_queue_add() argument 2662 if ((rx_adapter == NULL) || (queue_conf in rte_event_eth_rx_adapter_queue_add() 3336 rte_event_eth_rx_adapter_queue_conf_get(uint8_t id,uint16_t eth_dev_id,uint16_t rx_queue_id,struct rte_event_eth_rx_adapter_queue_conf * queue_conf) rte_event_eth_rx_adapter_queue_conf_get() argument 3655 struct rte_event_eth_rx_adapter_queue_conf queue_conf; handle_rxa_get_queue_conf() local [all...] |
| H A D | eventdev_pmd.h | 315 * @param[out] queue_conf 319 uint8_t queue_id, struct rte_event_queue_conf *queue_conf); 328 * @param queue_conf 336 const struct rte_event_queue_conf *queue_conf); 698 * @param queue_conf 709 const struct rte_event_eth_rx_adapter_queue_conf *queue_conf); 747 * @param[out] queue_conf 758 struct rte_event_eth_rx_adapter_queue_conf *queue_conf); 1009 const struct rte_event_crypto_adapter_queue_conf *queue_conf);
|
| H A D | rte_event_eth_rx_adapter.h | 706 * @param[out] queue_conf 716 struct rte_event_eth_rx_adapter_queue_conf *queue_conf);
|
| H A D | rte_eventdev.h | 1002 * @param[out] queue_conf 1012 struct rte_event_queue_conf *queue_conf); 1023 * @param queue_conf 1035 const struct rte_event_queue_conf *queue_conf);
|
| /dpdk/app/test/ |
| H A D | test_event_eth_rx_adapter.c | 1059 struct rte_event_eth_rx_adapter_queue_conf queue_conf = {0}; in adapter_queue_conf() local 1064 0, &queue_conf); in adapter_queue_conf() 1068 queue_conf.ev.queue_id = 0; in adapter_queue_conf() 1069 queue_conf.ev.sched_type = RTE_SCHED_TYPE_ATOMIC; in adapter_queue_conf() 1070 queue_conf.ev.priority = RTE_EVENT_DEV_PRIORITY_NORMAL; in adapter_queue_conf() 1074 0, &queue_conf); in adapter_queue_conf() 1080 0, &queue_conf); in adapter_queue_conf() 1086 -1, &queue_conf); in adapter_queue_conf() 1111 struct rte_event_eth_rx_adapter_queue_conf queue_conf = {0}; in adapter_pollq_instance_get() local 1137 queue_conf.ev.queue_id = TEST_ETH_QUEUE_ID; in adapter_pollq_instance_get() [all …]
|
| /dpdk/lib/power/ |
| H A D | rte_power_pmd_mgmt.c | 256 struct queue_list_entry *queue_conf = arg; in clb_multiwait() local 265 queue_reset(lcore_conf, queue_conf); in clb_multiwait() 271 if (!queue_can_sleep(lcore_conf, queue_conf)) in clb_multiwait() 294 struct queue_list_entry *queue_conf = arg; in clb_umwait() local 298 queue_conf->n_empty_polls++; in clb_umwait() 299 if (unlikely(queue_conf->n_empty_polls > emptypoll_max)) { in clb_umwait() 310 queue_conf->n_empty_polls = 0; in clb_umwait() 320 struct queue_list_entry *queue_conf = arg; in clb_pause() 329 queue_reset(lcore_conf, queue_conf); in clb_pause() 332 if (!queue_can_sleep(lcore_conf, queue_conf)) in clb_pause() 321 struct queue_list_entry *queue_conf = arg; clb_pause() local 364 struct queue_list_entry *queue_conf = arg; clb_scale_freq() local [all...] |
| /dpdk/examples/pipeline/ |
| H A D | obj.c | 203 struct rte_cryptodev_qp_conf queue_conf; in cryptodev_config() local 238 memset(&queue_conf, 0, sizeof(queue_conf)); in cryptodev_config() 239 queue_conf.nb_descriptors = params->queue_size; in cryptodev_config() 240 queue_conf.mp_session = NULL; in cryptodev_config() 243 status = rte_cryptodev_queue_pair_setup(dev_id, i, &queue_conf, socket_id); in cryptodev_config()
|
| /dpdk/examples/ip_pipeline/ |
| H A D | cryptodev.c | 53 struct rte_cryptodev_qp_conf queue_conf; in cryptodev_create() local 101 queue_conf.nb_descriptors = params->queue_size; in cryptodev_create() 104 &queue_conf, socket_id); in cryptodev_create()
|
| /dpdk/examples/ipsec-secgw/ |
| H A D | event_helper.c | 801 struct rte_event_crypto_adapter_queue_conf queue_conf; in eh_initialize_crypto_adapter() 864 memset(&queue_conf, 0, sizeof(queue_conf)); in eh_initialize_crypto_adapter() 867 queue_conf.flags |= RTE_EVENT_CRYPTO_ADAPTER_EVENT_VECTOR; in eh_initialize_crypto_adapter() 868 queue_conf.vector_sz = em_conf->ext_params.vector_size; in eh_initialize_crypto_adapter() 875 nb_elem = (qp_desc_nb / queue_conf.vector_sz) + 1; in eh_initialize_crypto_adapter() 880 queue_conf.vector_mp = rte_event_vector_pool_create( in eh_initialize_crypto_adapter() 882 queue_conf.vector_sz, socket_id); in eh_initialize_crypto_adapter() 883 if (queue_conf.vector_mp == NULL) { in eh_initialize_crypto_adapter() 892 &queue_conf); in eh_initialize_crypto_adapter() 795 struct rte_event_crypto_adapter_queue_conf queue_conf; eh_initialize_crypto_adapter() local 983 struct rte_event_eth_rx_adapter_queue_conf queue_conf = {0}; eh_rx_adapter_configure() local [all...] |
| /dpdk/lib/rawdev/ |
| H A D | rte_rawdev.h | 145 * @param[out] queue_conf 148 * The size of the structure pointed to by queue_conf 158 rte_rawdev_obj_t queue_conf, 169 * @param queue_conf 173 * The size of the structure pointed to by queue_conf 184 rte_rawdev_obj_t queue_conf,
|
| H A D | rte_rawdev_pmd.h | 242 * @param[out] queue_conf 252 rte_rawdev_obj_t queue_conf, 262 * @param queue_conf 272 rte_rawdev_obj_t queue_conf,
|
| H A D | rte_rawdev.c | 130 rte_rawdev_obj_t queue_conf, in rte_rawdev_queue_conf_get() argument 140 return (*dev->dev_ops->queue_def_conf)(dev, queue_id, queue_conf, in rte_rawdev_queue_conf_get() 147 rte_rawdev_obj_t queue_conf, in rte_rawdev_queue_setup() argument 157 return (*dev->dev_ops->queue_setup)(dev, queue_id, queue_conf, in rte_rawdev_queue_setup()
|
| /dpdk/examples/eventdev_pipeline/ |
| H A D | pipeline_worker_generic.c | 460 struct rte_event_eth_rx_adapter_queue_conf queue_conf; in init_adapters() 461 memset(&queue_conf, 0, sizeof(queue_conf)); in init_adapters() 462 queue_conf.ev.sched_type = cdata.queue_type; in init_adapters() 463 queue_conf.ev.queue_id = cdata.qid[0]; in init_adapters() 467 -1, &queue_conf); in init_adapters() 450 struct rte_event_eth_rx_adapter_queue_conf queue_conf; init_adapters() local
|
| H A D | pipeline_worker_tx.c | 793 struct rte_event_eth_rx_adapter_queue_conf queue_conf; in init_adapters() 794 memset(&queue_conf, 0, sizeof(queue_conf)); in init_adapters() 795 queue_conf.ev.sched_type = cdata.queue_type; in init_adapters() 813 queue_conf.ev.queue_id = cdata.rx_stride ? in init_adapters() 817 ret = rte_event_eth_rx_adapter_queue_add(i, i, -1, &queue_conf); in init_adapters() 787 struct rte_event_eth_rx_adapter_queue_conf queue_conf; init_adapters() local
|
| /dpdk/drivers/raw/skeleton/ |
| H A D | skeleton_rawdev.c | 232 rte_rawdev_obj_t queue_conf, in skeleton_rawdev_queue_def_conf() argument 240 if (!dev || !queue_conf || in skeleton_rawdev_queue_def_conf() 248 rte_memcpy(queue_conf, skelq, in skeleton_rawdev_queue_def_conf() 266 rte_rawdev_obj_t queue_conf, in skeleton_rawdev_queue_setup() argument 275 if (!dev || !queue_conf || in skeleton_rawdev_queue_setup() 284 rte_memcpy(q, queue_conf, in skeleton_rawdev_queue_setup()
|
| /dpdk/drivers/event/dpaa2/ |
| H A D | dpaa2_eventdev_selftest.c | 165 struct rte_event_queue_conf queue_conf; in _eventdev_setup() local 168 &queue_conf); in _eventdev_setup() 171 queue_conf.priority = i * step; in _eventdev_setup() 172 ret = rte_event_queue_setup(evdev, i, &queue_conf); in _eventdev_setup()
|
| /dpdk/drivers/raw/ntb/ |
| H A D | ntb.c | 256 rte_rawdev_obj_t queue_conf, in ntb_queue_conf_get() argument 259 struct ntb_queue_conf *q_conf = queue_conf; in ntb_queue_conf_get() 307 rte_rawdev_obj_t queue_conf, in ntb_rxq_setup() argument 310 struct ntb_queue_conf *rxq_conf = queue_conf; in ntb_rxq_setup() 392 rte_rawdev_obj_t queue_conf, in ntb_txq_setup() argument 395 struct ntb_queue_conf *txq_conf = queue_conf; in ntb_txq_setup() 460 rte_rawdev_obj_t queue_conf, in ntb_queue_setup() argument 469 ret = ntb_txq_setup(dev, queue_id, queue_conf, conf_size); in ntb_queue_setup() 473 ret = ntb_rxq_setup(dev, queue_id, queue_conf, conf_size); in ntb_queue_setup()
|
| /dpdk/drivers/raw/cnxk_gpio/ |
| H A D | cnxk_gpio.c | 368 rte_rawdev_obj_t queue_conf, size_t queue_conf_size) in cnxk_gpio_queue_setup() argument 375 RTE_SET_USED(queue_conf); in cnxk_gpio_queue_setup() 438 rte_rawdev_obj_t queue_conf, size_t queue_conf_size) in cnxk_gpio_queue_def_conf() argument 441 struct cnxk_gpio_queue_conf *conf = queue_conf; in cnxk_gpio_queue_def_conf()
|
| /dpdk/drivers/raw/cnxk_bphy/ |
| H A D | cnxk_bphy_cgx.c | 37 rte_rawdev_obj_t queue_conf, in cnxk_bphy_cgx_queue_def_conf() argument 48 conf = (unsigned int *)queue_conf; in cnxk_bphy_cgx_queue_def_conf()
|
| /dpdk/drivers/event/octeontx/ |
| H A D | ssovf_evdev_selftest.c | 200 struct rte_event_queue_conf queue_conf; in _eventdev_setup() local 203 &queue_conf); in _eventdev_setup() 206 queue_conf.priority = i * step; in _eventdev_setup() 207 ret = rte_event_queue_setup(evdev, i, &queue_conf); in _eventdev_setup()
|