Lines Matching defs:lcore
369 /* The lcore service worker thread is the only writer, and
525 * lcore thread really is done in service cores code.
532 rte_service_lcore_may_be_active(uint32_t lcore)
534 struct core_state *cs = RTE_LCORE_VAR_LCORE(lcore, lcore_states);
536 if (lcore >= RTE_MAX_LCORE || !cs->is_service_core)
584 rte_service_lcore_count_services(uint32_t lcore)
586 if (lcore >= RTE_MAX_LCORE)
589 struct core_state *cs = RTE_LCORE_VAR_LCORE(lcore, lcore_states);
639 service_update(uint32_t sid, uint32_t lcore, uint32_t *set, uint32_t *enabled)
641 struct core_state *cs = RTE_LCORE_VAR_LCORE(lcore, lcore_states);
644 if (!service_valid(sid) || lcore >= RTE_MAX_LCORE ||
670 rte_service_map_lcore_set(uint32_t id, uint32_t lcore, uint32_t enabled)
673 rte_eal_trace_service_map_lcore(id, lcore, enabled);
674 return service_update(id, lcore, &on, 0);
678 rte_service_map_lcore_get(uint32_t id, uint32_t lcore)
681 int ret = service_update(id, lcore, 0, &enabled);
688 set_lcore_state(uint32_t lcore, int32_t state)
692 struct core_state *cs = RTE_LCORE_VAR_LCORE(lcore, lcore_states);
693 cfg->lcore_role[lcore] = state;
696 lcore_config[lcore].core_role = state;
698 /* update per-lcore optimized state tracking */
701 rte_eal_trace_service_lcore_state_change(lcore, state);
731 rte_service_lcore_add(uint32_t lcore)
733 if (lcore >= RTE_MAX_LCORE)
736 struct core_state *cs = RTE_LCORE_VAR_LCORE(lcore, lcore_states);
740 set_lcore_state(lcore, ROLE_SERVICE);
750 return rte_eal_wait_lcore(lcore);
754 rte_service_lcore_del(uint32_t lcore)
756 if (lcore >= RTE_MAX_LCORE)
759 struct core_state *cs = RTE_LCORE_VAR_LCORE(lcore, lcore_states);
771 set_lcore_state(lcore, ROLE_RTE);
778 rte_service_lcore_start(uint32_t lcore)
780 if (lcore >= RTE_MAX_LCORE)
783 struct core_state *cs = RTE_LCORE_VAR_LCORE(lcore, lcore_states);
803 rte_eal_trace_service_lcore_start(lcore);
805 int ret = rte_eal_remote_launch(service_runner_func, 0, lcore);
811 rte_service_lcore_stop(uint32_t lcore)
813 struct core_state *cs = RTE_LCORE_VAR_LCORE(lcore, lcore_states);
815 if (lcore >= RTE_MAX_LCORE)
849 rte_eal_trace_service_lcore_stop(lcore);
855 lcore_attr_get_loops(unsigned int lcore)
857 struct core_state *cs = RTE_LCORE_VAR_LCORE(lcore, lcore_states);
863 lcore_attr_get_cycles(unsigned int lcore)
865 struct core_state *cs = RTE_LCORE_VAR_LCORE(lcore, lcore_states);
871 lcore_attr_get_service_calls(uint32_t service_id, unsigned int lcore)
873 struct core_state *cs = RTE_LCORE_VAR_LCORE(lcore, lcore_states);
880 lcore_attr_get_service_idle_calls(uint32_t service_id, unsigned int lcore)
882 struct core_state *cs = &lcore_states[lcore];
889 lcore_attr_get_service_error_calls(uint32_t service_id, unsigned int lcore)
891 struct core_state *cs = &lcore_states[lcore];
898 lcore_attr_get_service_cycles(uint32_t service_id, unsigned int lcore)
900 struct core_state *cs = RTE_LCORE_VAR_LCORE(lcore, lcore_states);
907 unsigned int lcore);
912 unsigned int lcore;
915 for (lcore = 0; lcore < RTE_MAX_LCORE; lcore++) {
917 RTE_LCORE_VAR_LCORE(lcore, lcore_states);
920 sum += lcore_attr_get(id, lcore);
978 rte_service_lcore_attr_get(uint32_t lcore, uint32_t attr_id,
981 struct core_state *cs = RTE_LCORE_VAR_LCORE(lcore, lcore_states);
983 if (lcore >= RTE_MAX_LCORE || !attr_value)
991 *attr_value = lcore_attr_get_loops(lcore);
994 *attr_value = lcore_attr_get_cycles(lcore);
1004 unsigned int lcore;
1009 for (lcore = 0; lcore < RTE_MAX_LCORE; lcore++) {
1011 RTE_LCORE_VAR_LCORE(lcore, lcore_states);
1020 rte_service_lcore_attr_reset_all(uint32_t lcore)
1022 struct core_state *cs = RTE_LCORE_VAR_LCORE(lcore, lcore_states);
1024 if (lcore >= RTE_MAX_LCORE)
1058 service_dump_calls_per_lcore(FILE *f, uint32_t lcore)
1061 struct core_state *cs = RTE_LCORE_VAR_LCORE(lcore, lcore_states);
1063 fprintf(f, "%02d\t", lcore);