1 /* SPDX-License-Identifier: BSD-3-Clause 2 * Copyright 2019 Mellanox Technologies, Ltd 3 */ 4 5 #ifndef RTE_PMD_MLX5_DEVX_CMDS_H_ 6 #define RTE_PMD_MLX5_DEVX_CMDS_H_ 7 8 #include "mlx5_glue.h" 9 #include "mlx5_prm.h" 10 #include <rte_compat.h> 11 12 /* 13 * Defines the amount of retries to allocate the first UAR in the page. 14 * OFED 5.0.x and Upstream rdma_core before v29 returned the NULL as 15 * UAR base address if UAR was not the first object in the UAR page. 16 * It caused the PMD failure and we should try to get another UAR 17 * till we get the first one with non-NULL base address returned. 18 */ 19 #define MLX5_ALLOC_UAR_RETRY 32 20 21 /* This is limitation of libibverbs: in length variable type is u16. */ 22 #define MLX5_DEVX_MAX_KLM_ENTRIES ((UINT16_MAX - \ 23 MLX5_ST_SZ_DW(create_mkey_in) * 4) / (MLX5_ST_SZ_DW(klm) * 4)) 24 25 struct mlx5_devx_mkey_attr { 26 uint64_t addr; 27 uint64_t size; 28 uint32_t umem_id; 29 uint32_t pd; 30 uint32_t log_entity_size; 31 uint32_t pg_access:1; 32 uint32_t relaxed_ordering_write:1; 33 uint32_t relaxed_ordering_read:1; 34 struct mlx5_klm *klm_array; 35 int klm_num; 36 }; 37 38 /* HCA qos attributes. */ 39 struct mlx5_hca_qos_attr { 40 uint32_t sup:1; /* Whether QOS is supported. */ 41 uint32_t flow_meter_old:1; /* Flow meter is supported, old version. */ 42 uint32_t packet_pacing:1; /* Packet pacing is supported. */ 43 uint32_t wqe_rate_pp:1; /* Packet pacing WQE rate mode. */ 44 uint32_t flow_meter:1; 45 /* 46 * Flow meter is supported, updated version. 47 * When flow_meter is 1, it indicates that REG_C sharing is supported. 48 * If flow_meter is 1, flow_meter_old is also 1. 49 * Using older driver versions, flow_meter_old can be 1 50 * while flow_meter is 0. 51 */ 52 uint8_t log_max_flow_meter; 53 /* Power of the maximum supported meters. */ 54 uint8_t flow_meter_reg_c_ids; 55 /* Bitmap of the reg_Cs available for flow meter to use. */ 56 57 }; 58 59 struct mlx5_hca_vdpa_attr { 60 uint8_t virtio_queue_type; 61 uint32_t valid:1; 62 uint32_t desc_tunnel_offload_type:1; 63 uint32_t eth_frame_offload_type:1; 64 uint32_t virtio_version_1_0:1; 65 uint32_t tso_ipv4:1; 66 uint32_t tso_ipv6:1; 67 uint32_t tx_csum:1; 68 uint32_t rx_csum:1; 69 uint32_t event_mode:3; 70 uint32_t log_doorbell_stride:5; 71 uint32_t log_doorbell_bar_size:5; 72 uint32_t queue_counters_valid:1; 73 uint32_t max_num_virtio_queues; 74 struct { 75 uint32_t a; 76 uint32_t b; 77 } umems[3]; 78 uint64_t doorbell_bar_offset; 79 }; 80 81 /* HCA supports this number of time periods for LRO. */ 82 #define MLX5_LRO_NUM_SUPP_PERIODS 4 83 84 /* HCA attributes. */ 85 struct mlx5_hca_attr { 86 uint32_t eswitch_manager:1; 87 uint32_t flow_counters_dump:1; 88 uint32_t log_max_rqt_size:5; 89 uint32_t parse_graph_flex_node:1; 90 uint8_t flow_counter_bulk_alloc_bitmap; 91 uint32_t eth_net_offloads:1; 92 uint32_t eth_virt:1; 93 uint32_t wqe_vlan_insert:1; 94 uint32_t wqe_inline_mode:2; 95 uint32_t vport_inline_mode:3; 96 uint32_t tunnel_stateless_geneve_rx:1; 97 uint32_t geneve_max_opt_len:1; /* 0x0: 14DW, 0x1: 63DW */ 98 uint32_t tunnel_stateless_gtp:1; 99 uint32_t lro_cap:1; 100 uint32_t tunnel_lro_gre:1; 101 uint32_t tunnel_lro_vxlan:1; 102 uint32_t lro_max_msg_sz_mode:2; 103 uint32_t lro_timer_supported_periods[MLX5_LRO_NUM_SUPP_PERIODS]; 104 uint16_t lro_min_mss_size; 105 uint32_t flex_parser_protocols; 106 uint32_t max_geneve_tlv_options; 107 uint32_t max_geneve_tlv_option_data_len; 108 uint32_t hairpin:1; 109 uint32_t log_max_hairpin_queues:5; 110 uint32_t log_max_hairpin_wq_data_sz:5; 111 uint32_t log_max_hairpin_num_packets:5; 112 uint32_t vhca_id:16; 113 uint32_t relaxed_ordering_write:1; 114 uint32_t relaxed_ordering_read:1; 115 uint32_t access_register_user:1; 116 uint32_t wqe_index_ignore:1; 117 uint32_t cross_channel:1; 118 uint32_t non_wire_sq:1; /* SQ with non-wire ops is supported. */ 119 uint32_t log_max_static_sq_wq:5; /* Static WQE size SQ. */ 120 uint32_t num_lag_ports:4; /* Number of ports can be bonded. */ 121 uint32_t dev_freq_khz; /* Timestamp counter frequency, kHz. */ 122 uint32_t scatter_fcs_w_decap_disable:1; 123 uint32_t flow_hit_aso:1; /* General obj type FLOW_HIT_ASO supported. */ 124 uint32_t roce:1; 125 uint32_t rq_ts_format:2; 126 uint32_t sq_ts_format:2; 127 uint32_t qp_ts_format:2; 128 uint32_t regex:1; 129 uint32_t reg_c_preserve:1; 130 uint32_t regexp_num_of_engines; 131 uint32_t log_max_ft_sampler_num:8; 132 uint32_t geneve_tlv_opt; 133 uint32_t cqe_compression:1; 134 uint32_t mini_cqe_resp_flow_tag:1; 135 uint32_t mini_cqe_resp_l3_l4_tag:1; 136 struct mlx5_hca_qos_attr qos; 137 struct mlx5_hca_vdpa_attr vdpa; 138 int log_max_qp_sz; 139 int log_max_cq_sz; 140 int log_max_qp; 141 int log_max_cq; 142 uint32_t log_max_pd; 143 uint32_t log_max_mrw_sz; 144 uint32_t log_max_srq; 145 uint32_t log_max_srq_sz; 146 uint32_t rss_ind_tbl_cap; 147 uint32_t mmo_dma_en:1; 148 uint32_t mmo_compress_en:1; 149 uint32_t mmo_decompress_en:1; 150 uint32_t compress_min_block_size:4; 151 uint32_t log_max_mmo_dma:5; 152 uint32_t log_max_mmo_compress:5; 153 uint32_t log_max_mmo_decompress:5; 154 }; 155 156 struct mlx5_devx_wq_attr { 157 uint32_t wq_type:4; 158 uint32_t wq_signature:1; 159 uint32_t end_padding_mode:2; 160 uint32_t cd_slave:1; 161 uint32_t hds_skip_first_sge:1; 162 uint32_t log2_hds_buf_size:3; 163 uint32_t page_offset:5; 164 uint32_t lwm:16; 165 uint32_t pd:24; 166 uint32_t uar_page:24; 167 uint64_t dbr_addr; 168 uint32_t hw_counter; 169 uint32_t sw_counter; 170 uint32_t log_wq_stride:4; 171 uint32_t log_wq_pg_sz:5; 172 uint32_t log_wq_sz:5; 173 uint32_t dbr_umem_valid:1; 174 uint32_t wq_umem_valid:1; 175 uint32_t log_hairpin_num_packets:5; 176 uint32_t log_hairpin_data_sz:5; 177 uint32_t single_wqe_log_num_of_strides:4; 178 uint32_t two_byte_shift_en:1; 179 uint32_t single_stride_log_num_of_bytes:3; 180 uint32_t dbr_umem_id; 181 uint32_t wq_umem_id; 182 uint64_t wq_umem_offset; 183 }; 184 185 /* Create RQ attributes structure, used by create RQ operation. */ 186 struct mlx5_devx_create_rq_attr { 187 uint32_t rlky:1; 188 uint32_t delay_drop_en:1; 189 uint32_t scatter_fcs:1; 190 uint32_t vsd:1; 191 uint32_t mem_rq_type:4; 192 uint32_t state:4; 193 uint32_t flush_in_error_en:1; 194 uint32_t hairpin:1; 195 uint32_t ts_format:2; 196 uint32_t user_index:24; 197 uint32_t cqn:24; 198 uint32_t counter_set_id:8; 199 uint32_t rmpn:24; 200 struct mlx5_devx_wq_attr wq_attr; 201 }; 202 203 /* Modify RQ attributes structure, used by modify RQ operation. */ 204 struct mlx5_devx_modify_rq_attr { 205 uint32_t rqn:24; 206 uint32_t rq_state:4; /* Current RQ state. */ 207 uint32_t state:4; /* Required RQ state. */ 208 uint32_t scatter_fcs:1; 209 uint32_t vsd:1; 210 uint32_t counter_set_id:8; 211 uint32_t hairpin_peer_sq:24; 212 uint32_t hairpin_peer_vhca:16; 213 uint64_t modify_bitmask; 214 uint32_t lwm:16; /* Contained WQ lwm. */ 215 }; 216 217 struct mlx5_rx_hash_field_select { 218 uint32_t l3_prot_type:1; 219 uint32_t l4_prot_type:1; 220 uint32_t selected_fields:30; 221 }; 222 223 /* TIR attributes structure, used by TIR operations. */ 224 struct mlx5_devx_tir_attr { 225 uint32_t disp_type:4; 226 uint32_t lro_timeout_period_usecs:16; 227 uint32_t lro_enable_mask:4; 228 uint32_t lro_max_msg_sz:8; 229 uint32_t inline_rqn:24; 230 uint32_t rx_hash_symmetric:1; 231 uint32_t tunneled_offload_en:1; 232 uint32_t indirect_table:24; 233 uint32_t rx_hash_fn:4; 234 uint32_t self_lb_block:2; 235 uint32_t transport_domain:24; 236 uint8_t rx_hash_toeplitz_key[MLX5_RSS_HASH_KEY_LEN]; 237 struct mlx5_rx_hash_field_select rx_hash_field_selector_outer; 238 struct mlx5_rx_hash_field_select rx_hash_field_selector_inner; 239 }; 240 241 /* TIR attributes structure, used by TIR modify. */ 242 struct mlx5_devx_modify_tir_attr { 243 uint32_t tirn:24; 244 uint64_t modify_bitmask; 245 struct mlx5_devx_tir_attr tir; 246 }; 247 248 /* RQT attributes structure, used by RQT operations. */ 249 struct mlx5_devx_rqt_attr { 250 uint8_t rq_type; 251 uint32_t rqt_max_size:16; 252 uint32_t rqt_actual_size:16; 253 uint32_t rq_list[]; 254 }; 255 256 /* TIS attributes structure. */ 257 struct mlx5_devx_tis_attr { 258 uint32_t strict_lag_tx_port_affinity:1; 259 uint32_t tls_en:1; 260 uint32_t lag_tx_port_affinity:4; 261 uint32_t prio:4; 262 uint32_t transport_domain:24; 263 }; 264 265 /* SQ attributes structure, used by SQ create operation. */ 266 struct mlx5_devx_create_sq_attr { 267 uint32_t rlky:1; 268 uint32_t cd_master:1; 269 uint32_t fre:1; 270 uint32_t flush_in_error_en:1; 271 uint32_t allow_multi_pkt_send_wqe:1; 272 uint32_t min_wqe_inline_mode:3; 273 uint32_t state:4; 274 uint32_t reg_umr:1; 275 uint32_t allow_swp:1; 276 uint32_t hairpin:1; 277 uint32_t non_wire:1; 278 uint32_t static_sq_wq:1; 279 uint32_t ts_format:2; 280 uint32_t user_index:24; 281 uint32_t cqn:24; 282 uint32_t packet_pacing_rate_limit_index:16; 283 uint32_t tis_lst_sz:16; 284 uint32_t tis_num:24; 285 struct mlx5_devx_wq_attr wq_attr; 286 }; 287 288 /* SQ attributes structure, used by SQ modify operation. */ 289 struct mlx5_devx_modify_sq_attr { 290 uint32_t sq_state:4; 291 uint32_t state:4; 292 uint32_t hairpin_peer_rq:24; 293 uint32_t hairpin_peer_vhca:16; 294 }; 295 296 297 /* CQ attributes structure, used by CQ operations. */ 298 struct mlx5_devx_cq_attr { 299 uint32_t q_umem_valid:1; 300 uint32_t db_umem_valid:1; 301 uint32_t use_first_only:1; 302 uint32_t overrun_ignore:1; 303 uint32_t cqe_comp_en:1; 304 uint32_t mini_cqe_res_format:2; 305 uint32_t mini_cqe_res_format_ext:2; 306 uint32_t log_cq_size:5; 307 uint32_t log_page_size:5; 308 uint32_t uar_page_id; 309 uint32_t q_umem_id; 310 uint64_t q_umem_offset; 311 uint32_t db_umem_id; 312 uint64_t db_umem_offset; 313 uint32_t eqn; 314 uint64_t db_addr; 315 }; 316 317 /* Virtq attributes structure, used by VIRTQ operations. */ 318 struct mlx5_devx_virtq_attr { 319 uint16_t hw_available_index; 320 uint16_t hw_used_index; 321 uint16_t q_size; 322 uint32_t pd:24; 323 uint32_t virtio_version_1_0:1; 324 uint32_t tso_ipv4:1; 325 uint32_t tso_ipv6:1; 326 uint32_t tx_csum:1; 327 uint32_t rx_csum:1; 328 uint32_t event_mode:3; 329 uint32_t state:4; 330 uint32_t hw_latency_mode:2; 331 uint32_t hw_max_latency_us:12; 332 uint32_t hw_max_pending_comp:16; 333 uint32_t dirty_bitmap_dump_enable:1; 334 uint32_t dirty_bitmap_mkey; 335 uint32_t dirty_bitmap_size; 336 uint32_t mkey; 337 uint32_t qp_id; 338 uint32_t queue_index; 339 uint32_t tis_id; 340 uint32_t counters_obj_id; 341 uint64_t dirty_bitmap_addr; 342 uint64_t type; 343 uint64_t desc_addr; 344 uint64_t used_addr; 345 uint64_t available_addr; 346 struct { 347 uint32_t id; 348 uint32_t size; 349 uint64_t offset; 350 } umems[3]; 351 uint8_t error_type; 352 }; 353 354 355 struct mlx5_devx_qp_attr { 356 uint32_t pd:24; 357 uint32_t uar_index:24; 358 uint32_t cqn:24; 359 uint32_t log_page_size:5; 360 uint32_t rq_size:17; /* Must be power of 2. */ 361 uint32_t log_rq_stride:3; 362 uint32_t sq_size:17; /* Must be power of 2. */ 363 uint32_t ts_format:2; 364 uint32_t dbr_umem_valid:1; 365 uint32_t dbr_umem_id; 366 uint64_t dbr_address; 367 uint32_t wq_umem_id; 368 uint64_t wq_umem_offset; 369 }; 370 371 struct mlx5_devx_virtio_q_couners_attr { 372 uint64_t received_desc; 373 uint64_t completed_desc; 374 uint32_t error_cqes; 375 uint32_t bad_desc_errors; 376 uint32_t exceed_max_chain; 377 uint32_t invalid_buffer; 378 }; 379 380 /* 381 * graph flow match sample attributes structure, 382 * used by flex parser operations. 383 */ 384 struct mlx5_devx_match_sample_attr { 385 uint32_t flow_match_sample_en:1; 386 uint32_t flow_match_sample_field_offset:16; 387 uint32_t flow_match_sample_offset_mode:4; 388 uint32_t flow_match_sample_field_offset_mask; 389 uint32_t flow_match_sample_field_offset_shift:4; 390 uint32_t flow_match_sample_field_base_offset:8; 391 uint32_t flow_match_sample_tunnel_mode:3; 392 uint32_t flow_match_sample_field_id; 393 }; 394 395 /* graph node arc attributes structure, used by flex parser operations. */ 396 struct mlx5_devx_graph_arc_attr { 397 uint32_t compare_condition_value:16; 398 uint32_t start_inner_tunnel:1; 399 uint32_t arc_parse_graph_node:8; 400 uint32_t parse_graph_node_handle; 401 }; 402 403 /* Maximal number of samples per graph node. */ 404 #define MLX5_GRAPH_NODE_SAMPLE_NUM 8 405 406 /* Maximal number of input/output arcs per graph node. */ 407 #define MLX5_GRAPH_NODE_ARC_NUM 8 408 409 /* parse graph node attributes structure, used by flex parser operations. */ 410 struct mlx5_devx_graph_node_attr { 411 uint32_t modify_field_select; 412 uint32_t header_length_mode:4; 413 uint32_t header_length_base_value:16; 414 uint32_t header_length_field_shift:4; 415 uint32_t header_length_field_offset:16; 416 uint32_t header_length_field_mask; 417 struct mlx5_devx_match_sample_attr sample[MLX5_GRAPH_NODE_SAMPLE_NUM]; 418 uint32_t next_header_field_offset:16; 419 uint32_t next_header_field_size:5; 420 struct mlx5_devx_graph_arc_attr in[MLX5_GRAPH_NODE_ARC_NUM]; 421 struct mlx5_devx_graph_arc_attr out[MLX5_GRAPH_NODE_ARC_NUM]; 422 }; 423 424 /* mlx5_devx_cmds.c */ 425 426 __rte_internal 427 struct mlx5_devx_obj *mlx5_devx_cmd_flow_counter_alloc(void *ctx, 428 uint32_t bulk_sz); 429 __rte_internal 430 int mlx5_devx_cmd_destroy(struct mlx5_devx_obj *obj); 431 __rte_internal 432 int mlx5_devx_cmd_flow_counter_query(struct mlx5_devx_obj *dcs, 433 int clear, uint32_t n_counters, 434 uint64_t *pkts, uint64_t *bytes, 435 uint32_t mkey, void *addr, 436 void *cmd_comp, 437 uint64_t async_id); 438 __rte_internal 439 int mlx5_devx_cmd_query_hca_attr(void *ctx, 440 struct mlx5_hca_attr *attr); 441 __rte_internal 442 struct mlx5_devx_obj *mlx5_devx_cmd_mkey_create(void *ctx, 443 struct mlx5_devx_mkey_attr *attr); 444 __rte_internal 445 int mlx5_devx_get_out_command_status(void *out); 446 __rte_internal 447 int mlx5_devx_cmd_qp_query_tis_td(void *qp, uint32_t tis_num, 448 uint32_t *tis_td); 449 __rte_internal 450 struct mlx5_devx_obj *mlx5_devx_cmd_create_rq(void *ctx, 451 struct mlx5_devx_create_rq_attr *rq_attr, 452 int socket); 453 __rte_internal 454 int mlx5_devx_cmd_modify_rq(struct mlx5_devx_obj *rq, 455 struct mlx5_devx_modify_rq_attr *rq_attr); 456 __rte_internal 457 struct mlx5_devx_obj *mlx5_devx_cmd_create_tir(void *ctx, 458 struct mlx5_devx_tir_attr *tir_attr); 459 __rte_internal 460 struct mlx5_devx_obj *mlx5_devx_cmd_create_rqt(void *ctx, 461 struct mlx5_devx_rqt_attr *rqt_attr); 462 __rte_internal 463 struct mlx5_devx_obj *mlx5_devx_cmd_create_sq(void *ctx, 464 struct mlx5_devx_create_sq_attr *sq_attr); 465 __rte_internal 466 int mlx5_devx_cmd_modify_sq(struct mlx5_devx_obj *sq, 467 struct mlx5_devx_modify_sq_attr *sq_attr); 468 __rte_internal 469 struct mlx5_devx_obj *mlx5_devx_cmd_create_tis(void *ctx, 470 struct mlx5_devx_tis_attr *tis_attr); 471 __rte_internal 472 struct mlx5_devx_obj *mlx5_devx_cmd_create_td(void *ctx); 473 __rte_internal 474 int mlx5_devx_cmd_flow_dump(void *fdb_domain, void *rx_domain, void *tx_domain, 475 FILE *file); 476 __rte_internal 477 struct mlx5_devx_obj *mlx5_devx_cmd_create_cq(void *ctx, 478 struct mlx5_devx_cq_attr *attr); 479 __rte_internal 480 struct mlx5_devx_obj *mlx5_devx_cmd_create_virtq(void *ctx, 481 struct mlx5_devx_virtq_attr *attr); 482 __rte_internal 483 int mlx5_devx_cmd_modify_virtq(struct mlx5_devx_obj *virtq_obj, 484 struct mlx5_devx_virtq_attr *attr); 485 __rte_internal 486 int mlx5_devx_cmd_query_virtq(struct mlx5_devx_obj *virtq_obj, 487 struct mlx5_devx_virtq_attr *attr); 488 __rte_internal 489 struct mlx5_devx_obj *mlx5_devx_cmd_create_qp(void *ctx, 490 struct mlx5_devx_qp_attr *attr); 491 __rte_internal 492 int mlx5_devx_cmd_modify_qp_state(struct mlx5_devx_obj *qp, 493 uint32_t qp_st_mod_op, uint32_t remote_qp_id); 494 __rte_internal 495 int mlx5_devx_cmd_modify_rqt(struct mlx5_devx_obj *rqt, 496 struct mlx5_devx_rqt_attr *rqt_attr); 497 __rte_internal 498 int mlx5_devx_cmd_modify_tir(struct mlx5_devx_obj *tir, 499 struct mlx5_devx_modify_tir_attr *tir_attr); 500 __rte_internal 501 int mlx5_devx_cmd_query_parse_samples(struct mlx5_devx_obj *flex_obj, 502 uint32_t ids[], uint32_t num); 503 504 __rte_internal 505 struct mlx5_devx_obj *mlx5_devx_cmd_create_flex_parser(void *ctx, 506 struct mlx5_devx_graph_node_attr *data); 507 508 __rte_internal 509 int mlx5_devx_cmd_register_read(void *ctx, uint16_t reg_id, 510 uint32_t arg, uint32_t *data, uint32_t dw_cnt); 511 512 __rte_internal 513 struct mlx5_devx_obj * 514 mlx5_devx_cmd_create_geneve_tlv_option(void *ctx, 515 uint16_t class, uint8_t type, uint8_t len); 516 517 /** 518 * Create virtio queue counters object DevX API. 519 * 520 * @param[in] ctx 521 * Device context. 522 523 * @return 524 * The DevX object created, NULL otherwise and rte_errno is set. 525 */ 526 __rte_internal 527 struct mlx5_devx_obj *mlx5_devx_cmd_create_virtio_q_counters(void *ctx); 528 529 /** 530 * Query virtio queue counters object using DevX API. 531 * 532 * @param[in] couners_obj 533 * Pointer to virtq object structure. 534 * @param [in/out] attr 535 * Pointer to virtio queue counters attributes structure. 536 * 537 * @return 538 * 0 on success, a negative errno value otherwise and rte_errno is set. 539 */ 540 __rte_internal 541 int mlx5_devx_cmd_query_virtio_q_counters(struct mlx5_devx_obj *couners_obj, 542 struct mlx5_devx_virtio_q_couners_attr *attr); 543 __rte_internal 544 struct mlx5_devx_obj *mlx5_devx_cmd_create_flow_hit_aso_obj(void *ctx, 545 uint32_t pd); 546 547 __rte_internal 548 struct mlx5_devx_obj *mlx5_devx_cmd_alloc_pd(void *ctx); 549 550 __rte_internal 551 int mlx5_devx_cmd_wq_query(void *wq, uint32_t *counter_set_id); 552 553 __rte_internal 554 struct mlx5_devx_obj *mlx5_devx_cmd_queue_counter_alloc(void *ctx); 555 __rte_internal 556 int mlx5_devx_cmd_queue_counter_query(struct mlx5_devx_obj *dcs, int clear, 557 uint32_t *out_of_buffers); 558 #endif /* RTE_PMD_MLX5_DEVX_CMDS_H_ */ 559