1 /* SPDX-License-Identifier: BSD-3-Clause 2 * Copyright (c) 2018-2019 Broadcom. All Rights Reserved. 3 * The term "Broadcom" refers to Broadcom Inc. and/or its subsidiaries. 4 */ 5 6 /* NVMF FC Transport Unit Test */ 7 8 #include "spdk/env.h" 9 #include "spdk_cunit.h" 10 #include "spdk/nvmf.h" 11 #include "spdk/endian.h" 12 #include "spdk/trace.h" 13 #include "spdk/log.h" 14 15 #include "ut_multithread.c" 16 17 #include "transport.h" 18 #include "nvmf_internal.h" 19 20 #include "nvmf_fc.h" 21 22 #include "json/json_util.c" 23 #include "json/json_write.c" 24 #include "nvmf/nvmf.c" 25 #include "nvmf/transport.c" 26 #include "spdk/bdev_module.h" 27 #include "nvmf/subsystem.c" 28 #include "nvmf/fc.c" 29 #include "nvmf/fc_ls.c" 30 31 /* 32 * SPDK Stuff 33 */ 34 35 #ifdef SPDK_CONFIG_RDMA 36 const struct spdk_nvmf_transport_ops spdk_nvmf_transport_rdma = { 37 .type = SPDK_NVME_TRANSPORT_RDMA, 38 .opts_init = NULL, 39 .create = NULL, 40 .destroy = NULL, 41 42 .listen = NULL, 43 .stop_listen = NULL, 44 .accept = NULL, 45 46 .listener_discover = NULL, 47 48 .poll_group_create = NULL, 49 .poll_group_destroy = NULL, 50 .poll_group_add = NULL, 51 .poll_group_poll = NULL, 52 53 .req_free = NULL, 54 .req_complete = NULL, 55 56 .qpair_fini = NULL, 57 .qpair_get_peer_trid = NULL, 58 .qpair_get_local_trid = NULL, 59 .qpair_get_listen_trid = NULL, 60 }; 61 #endif 62 63 const struct spdk_nvmf_transport_ops spdk_nvmf_transport_tcp = { 64 .type = SPDK_NVME_TRANSPORT_TCP, 65 }; 66 67 DEFINE_STUB(spdk_nvme_transport_id_compare, int, 68 (const struct spdk_nvme_transport_id *trid1, 69 const struct spdk_nvme_transport_id *trid2), 0); 70 DEFINE_STUB(spdk_bdev_get_name, const char *, (const struct spdk_bdev *bdev), "fc_ut_test"); 71 DEFINE_STUB_V(nvmf_ctrlr_destruct, (struct spdk_nvmf_ctrlr *ctrlr)); 72 DEFINE_STUB_V(nvmf_qpair_free_aer, (struct spdk_nvmf_qpair *qpair)); 73 DEFINE_STUB_V(nvmf_qpair_abort_pending_zcopy_reqs, (struct spdk_nvmf_qpair *qpair)); 74 DEFINE_STUB(spdk_bdev_get_io_channel, struct spdk_io_channel *, (struct spdk_bdev_desc *desc), 75 NULL); 76 DEFINE_STUB_V(spdk_nvmf_request_exec, (struct spdk_nvmf_request *req)); 77 DEFINE_STUB_V(nvmf_ctrlr_ns_changed, (struct spdk_nvmf_ctrlr *ctrlr, uint32_t nsid)); 78 DEFINE_STUB_V(spdk_bdev_close, (struct spdk_bdev_desc *desc)); 79 DEFINE_STUB(spdk_bdev_module_claim_bdev, int, 80 (struct spdk_bdev *bdev, struct spdk_bdev_desc *desc, 81 struct spdk_bdev_module *module), 0); 82 DEFINE_STUB_V(spdk_bdev_module_release_bdev, (struct spdk_bdev *bdev)); 83 DEFINE_STUB(spdk_bdev_get_block_size, uint32_t, (const struct spdk_bdev *bdev), 512); 84 DEFINE_STUB(spdk_bdev_get_num_blocks, uint64_t, (const struct spdk_bdev *bdev), 1024); 85 86 DEFINE_STUB(nvmf_ctrlr_async_event_ns_notice, int, (struct spdk_nvmf_ctrlr *ctrlr), 0); 87 DEFINE_STUB(nvmf_ctrlr_async_event_ana_change_notice, int, 88 (struct spdk_nvmf_ctrlr *ctrlr), 0); 89 DEFINE_STUB_V(spdk_nvme_trid_populate_transport, (struct spdk_nvme_transport_id *trid, 90 enum spdk_nvme_transport_type trtype)); 91 DEFINE_STUB_V(spdk_nvmf_ctrlr_data_init, (struct spdk_nvmf_transport_opts *opts, 92 struct spdk_nvmf_ctrlr_data *cdata)); 93 DEFINE_STUB(spdk_nvmf_request_complete, int, (struct spdk_nvmf_request *req), 94 -ENOSPC); 95 96 DEFINE_STUB_V(nvmf_update_discovery_log, 97 (struct spdk_nvmf_tgt *tgt, const char *hostnqn)); 98 99 DEFINE_STUB(rte_hash_create, struct rte_hash *, (const struct rte_hash_parameters *params), 100 (void *)1); 101 DEFINE_STUB(rte_hash_del_key, int32_t, (const struct rte_hash *h, const void *key), 0); 102 DEFINE_STUB(rte_hash_lookup_data, int, (const struct rte_hash *h, const void *key, void **data), 103 -ENOENT); 104 DEFINE_STUB(rte_hash_add_key_data, int, (const struct rte_hash *h, const void *key, void *data), 0); 105 DEFINE_STUB_V(rte_hash_free, (struct rte_hash *h)); 106 DEFINE_STUB(nvmf_fc_lld_port_add, int, (struct spdk_nvmf_fc_port *fc_port), 0); 107 DEFINE_STUB(nvmf_fc_lld_port_remove, int, (struct spdk_nvmf_fc_port *fc_port), 0); 108 109 DEFINE_STUB_V(spdk_nvmf_request_zcopy_start, (struct spdk_nvmf_request *req)); 110 DEFINE_STUB_V(spdk_nvmf_request_zcopy_end, (struct spdk_nvmf_request *req, bool commit)); 111 112 const char * 113 spdk_nvme_transport_id_trtype_str(enum spdk_nvme_transport_type trtype) 114 { 115 switch (trtype) { 116 case SPDK_NVME_TRANSPORT_PCIE: 117 return "PCIe"; 118 case SPDK_NVME_TRANSPORT_RDMA: 119 return "RDMA"; 120 case SPDK_NVME_TRANSPORT_FC: 121 return "FC"; 122 default: 123 return NULL; 124 } 125 } 126 127 const char * 128 spdk_nvme_transport_id_adrfam_str(enum spdk_nvmf_adrfam adrfam) 129 { 130 switch (adrfam) { 131 case SPDK_NVMF_ADRFAM_IPV4: 132 return "IPv4"; 133 case SPDK_NVMF_ADRFAM_IPV6: 134 return "IPv6"; 135 case SPDK_NVMF_ADRFAM_IB: 136 return "IB"; 137 case SPDK_NVMF_ADRFAM_FC: 138 return "FC"; 139 default: 140 return NULL; 141 } 142 } 143 144 const struct spdk_uuid * 145 spdk_bdev_get_uuid(const struct spdk_bdev *bdev) 146 { 147 return &bdev->uuid; 148 } 149 150 static bool g_lld_init_called = false; 151 152 int 153 nvmf_fc_lld_init(void) 154 { 155 g_lld_init_called = true; 156 return 0; 157 } 158 159 static bool g_lld_fini_called = false; 160 161 void 162 nvmf_fc_lld_fini(spdk_nvmf_transport_destroy_done_cb cb_fn, void *ctx) 163 { 164 g_lld_fini_called = true; 165 } 166 167 DEFINE_STUB_V(nvmf_fc_lld_start, (void)); 168 DEFINE_STUB(nvmf_fc_init_q, int, (struct spdk_nvmf_fc_hwqp *hwqp), 0); 169 DEFINE_STUB_V(nvmf_fc_reinit_q, (void *queues_prev, void *queues_curr)); 170 DEFINE_STUB(nvmf_fc_init_rqpair_buffers, int, (struct spdk_nvmf_fc_hwqp *hwqp), 0); 171 DEFINE_STUB(nvmf_fc_set_q_online_state, int, (struct spdk_nvmf_fc_hwqp *hwqp, bool online), 0); 172 DEFINE_STUB(nvmf_fc_put_xchg, int, (struct spdk_nvmf_fc_hwqp *hwqp, struct spdk_nvmf_fc_xchg *xri), 173 0); 174 DEFINE_STUB(nvmf_fc_recv_data, int, (struct spdk_nvmf_fc_request *fc_req), 0); 175 DEFINE_STUB(nvmf_fc_send_data, int, (struct spdk_nvmf_fc_request *fc_req), 0); 176 DEFINE_STUB_V(nvmf_fc_rqpair_buffer_release, (struct spdk_nvmf_fc_hwqp *hwqp, uint16_t buff_idx)); 177 DEFINE_STUB(nvmf_fc_xmt_rsp, int, (struct spdk_nvmf_fc_request *fc_req, uint8_t *ersp_buf, 178 uint32_t ersp_len), 0); 179 DEFINE_STUB(nvmf_fc_xmt_ls_rsp, int, (struct spdk_nvmf_fc_nport *tgtport, 180 struct spdk_nvmf_fc_ls_rqst *ls_rqst), 0); 181 DEFINE_STUB(nvmf_fc_issue_abort, int, (struct spdk_nvmf_fc_hwqp *hwqp, 182 struct spdk_nvmf_fc_xchg *xri, 183 spdk_nvmf_fc_caller_cb cb, void *cb_args), 0); 184 DEFINE_STUB(nvmf_fc_xmt_bls_rsp, int, (struct spdk_nvmf_fc_hwqp *hwqp, 185 uint16_t ox_id, uint16_t rx_id, 186 uint16_t rpi, bool rjt, uint8_t rjt_exp, 187 spdk_nvmf_fc_caller_cb cb, void *cb_args), 0); 188 DEFINE_STUB(nvmf_fc_alloc_srsr_bufs, struct spdk_nvmf_fc_srsr_bufs *, (size_t rqst_len, 189 size_t rsp_len), NULL); 190 DEFINE_STUB_V(nvmf_fc_free_srsr_bufs, (struct spdk_nvmf_fc_srsr_bufs *srsr_bufs)); 191 DEFINE_STUB(nvmf_fc_xmt_srsr_req, int, (struct spdk_nvmf_fc_hwqp *hwqp, 192 struct spdk_nvmf_fc_srsr_bufs *xmt_srsr_bufs, 193 spdk_nvmf_fc_caller_cb cb, void *cb_args), 0); 194 DEFINE_STUB(nvmf_fc_q_sync_available, bool, (void), true); 195 DEFINE_STUB(nvmf_fc_issue_q_sync, int, (struct spdk_nvmf_fc_hwqp *hwqp, uint64_t u_id, 196 uint16_t skip_rq), 0); 197 DEFINE_STUB(nvmf_fc_assign_conn_to_hwqp, bool, (struct spdk_nvmf_fc_hwqp *hwqp, 198 uint64_t *conn_id, uint32_t sq_size), true); 199 DEFINE_STUB(nvmf_fc_get_hwqp_from_conn_id, struct spdk_nvmf_fc_hwqp *, 200 (struct spdk_nvmf_fc_hwqp *queues, 201 uint32_t num_queues, uint64_t conn_id), NULL); 202 DEFINE_STUB_V(nvmf_fc_dump_all_queues, (struct spdk_nvmf_fc_hwqp *ls_queue, 203 struct spdk_nvmf_fc_hwqp *io_queues, 204 uint32_t num_io_queues, 205 struct spdk_nvmf_fc_queue_dump_info *dump_info)); 206 DEFINE_STUB_V(nvmf_fc_get_xri_info, (struct spdk_nvmf_fc_hwqp *hwqp, 207 struct spdk_nvmf_fc_xchg_info *info)); 208 DEFINE_STUB(nvmf_fc_get_rsvd_thread, struct spdk_thread *, (void), NULL); 209 210 uint32_t 211 nvmf_fc_process_queue(struct spdk_nvmf_fc_hwqp *hwqp) 212 { 213 hwqp->lcore_id++; 214 return 0; /* always return 0 or else it will poll forever */ 215 } 216 217 struct spdk_nvmf_fc_xchg * 218 nvmf_fc_get_xri(struct spdk_nvmf_fc_hwqp *hwqp) 219 { 220 static struct spdk_nvmf_fc_xchg xchg; 221 222 xchg.xchg_id = 1; 223 return &xchg; 224 } 225 226 #define MAX_FC_UT_POLL_THREADS 8 227 static struct spdk_nvmf_poll_group *g_poll_groups[MAX_FC_UT_POLL_THREADS] = {0}; 228 #define MAX_FC_UT_HWQPS MAX_FC_UT_POLL_THREADS 229 static struct spdk_nvmf_tgt *g_nvmf_tgt = NULL; 230 static struct spdk_nvmf_transport *g_nvmf_tprt = NULL; 231 uint8_t g_fc_port_handle = 0xff; 232 struct spdk_nvmf_fc_hwqp lld_q[MAX_FC_UT_HWQPS]; 233 234 static void 235 _add_transport_done(void *arg, int status) 236 { 237 CU_ASSERT(status == 0); 238 } 239 240 static void 241 _add_transport_done_dup_err(void *arg, int status) 242 { 243 CU_ASSERT(status == -EEXIST); 244 } 245 246 static void 247 create_transport_test(void) 248 { 249 const struct spdk_nvmf_transport_ops *ops = NULL; 250 struct spdk_nvmf_transport_opts opts = { 0 }; 251 struct spdk_nvmf_target_opts tgt_opts = { 252 .name = "nvmf_test_tgt", 253 .max_subsystems = 0 254 }; 255 256 allocate_threads(8); 257 set_thread(0); 258 259 g_nvmf_tgt = spdk_nvmf_tgt_create(&tgt_opts); 260 SPDK_CU_ASSERT_FATAL(g_nvmf_tgt != NULL); 261 262 ops = nvmf_get_transport_ops(SPDK_NVME_TRANSPORT_NAME_FC); 263 SPDK_CU_ASSERT_FATAL(ops != NULL); 264 265 ops->opts_init(&opts); 266 267 g_lld_init_called = false; 268 opts.opts_size = sizeof(opts); 269 g_nvmf_tprt = spdk_nvmf_transport_create("FC", &opts); 270 SPDK_CU_ASSERT_FATAL(g_nvmf_tprt != NULL); 271 272 CU_ASSERT(g_lld_init_called == true); 273 CU_ASSERT(opts.max_queue_depth == g_nvmf_tprt->opts.max_queue_depth); 274 CU_ASSERT(opts.max_qpairs_per_ctrlr == g_nvmf_tprt->opts.max_qpairs_per_ctrlr); 275 CU_ASSERT(opts.in_capsule_data_size == g_nvmf_tprt->opts.in_capsule_data_size); 276 CU_ASSERT(opts.max_io_size == g_nvmf_tprt->opts.max_io_size); 277 CU_ASSERT(opts.io_unit_size == g_nvmf_tprt->opts.io_unit_size); 278 CU_ASSERT(opts.max_aq_depth == g_nvmf_tprt->opts.max_aq_depth); 279 280 set_thread(0); 281 282 spdk_nvmf_tgt_add_transport(g_nvmf_tgt, g_nvmf_tprt, 283 _add_transport_done, 0); 284 poll_thread(0); 285 286 /* Add transport again - should get error */ 287 spdk_nvmf_tgt_add_transport(g_nvmf_tgt, g_nvmf_tprt, 288 _add_transport_done_dup_err, 0); 289 poll_thread(0); 290 291 /* create transport with bad args/options */ 292 #ifndef SPDK_CONFIG_RDMA 293 CU_ASSERT(spdk_nvmf_transport_create("RDMA", &opts) == NULL); 294 #endif 295 CU_ASSERT(spdk_nvmf_transport_create("Bogus Transport", &opts) == NULL); 296 opts.max_io_size = 1024 ^ 3; 297 CU_ASSERT(spdk_nvmf_transport_create("FC", &opts) == NULL); 298 opts.max_io_size = 999; 299 opts.io_unit_size = 1024; 300 CU_ASSERT(spdk_nvmf_transport_create("FC", &opts) == NULL); 301 } 302 303 static void 304 port_init_cb(uint8_t port_handle, enum spdk_fc_event event_type, void *arg, int err) 305 { 306 CU_ASSERT(err == 0); 307 CU_ASSERT(port_handle == 2); 308 g_fc_port_handle = port_handle; 309 } 310 311 static void 312 create_fc_port_test(void) 313 { 314 struct spdk_nvmf_fc_hw_port_init_args init_args = { 0 }; 315 struct spdk_nvmf_fc_port *fc_port = NULL; 316 int err; 317 318 SPDK_CU_ASSERT_FATAL(g_nvmf_tprt != NULL); 319 320 init_args.port_handle = 2; 321 init_args.io_queue_cnt = spdk_min(MAX_FC_UT_HWQPS, spdk_env_get_core_count()); 322 init_args.ls_queue_size = 100; 323 init_args.io_queue_size = 100; 324 init_args.io_queues = (void *)lld_q; 325 326 set_thread(0); 327 err = nvmf_fc_main_enqueue_event(SPDK_FC_HW_PORT_INIT, (void *)&init_args, port_init_cb); 328 CU_ASSERT(err == 0); 329 poll_thread(0); 330 331 fc_port = nvmf_fc_port_lookup(g_fc_port_handle); 332 CU_ASSERT(fc_port != NULL); 333 } 334 335 static void 336 online_fc_port_test(void) 337 { 338 struct spdk_nvmf_fc_port *fc_port; 339 struct spdk_nvmf_fc_hw_port_online_args args; 340 int err; 341 342 SPDK_CU_ASSERT_FATAL(g_nvmf_tprt != NULL); 343 344 fc_port = nvmf_fc_port_lookup(g_fc_port_handle); 345 SPDK_CU_ASSERT_FATAL(fc_port != NULL); 346 347 set_thread(0); 348 args.port_handle = g_fc_port_handle; 349 err = nvmf_fc_main_enqueue_event(SPDK_FC_HW_PORT_ONLINE, (void *)&args, port_init_cb); 350 CU_ASSERT(err == 0); 351 poll_threads(); 352 set_thread(0); 353 if (err == 0) { 354 uint32_t i; 355 for (i = 0; i < fc_port->num_io_queues; i++) { 356 CU_ASSERT(fc_port->io_queues[i].fgroup != 0); 357 CU_ASSERT(fc_port->io_queues[i].fgroup != 0); 358 CU_ASSERT(fc_port->io_queues[i].fgroup->hwqp_count != 0); 359 } 360 } 361 } 362 363 static void 364 create_poll_groups_test(void) 365 { 366 unsigned i; 367 368 SPDK_CU_ASSERT_FATAL(g_nvmf_tprt != NULL); 369 370 for (i = 0; i < MAX_FC_UT_POLL_THREADS; i++) { 371 set_thread(i); 372 g_poll_groups[i] = spdk_nvmf_poll_group_create(g_nvmf_tgt); 373 poll_thread(i); 374 CU_ASSERT(g_poll_groups[i] != NULL); 375 } 376 set_thread(0); 377 } 378 379 static void 380 poll_group_poll_test(void) 381 { 382 unsigned i; 383 unsigned poll_cnt = 10; 384 struct spdk_nvmf_fc_port *fc_port = NULL; 385 386 SPDK_CU_ASSERT_FATAL(g_nvmf_tprt != NULL); 387 388 set_thread(0); 389 fc_port = nvmf_fc_port_lookup(g_fc_port_handle); 390 SPDK_CU_ASSERT_FATAL(fc_port != NULL); 391 392 for (i = 0; i < fc_port->num_io_queues; i++) { 393 fc_port->io_queues[i].lcore_id = 0; 394 } 395 396 for (i = 0; i < poll_cnt; i++) { 397 /* this should cause spdk_nvmf_fc_poll_group_poll to be called() */ 398 poll_threads(); 399 } 400 401 /* check if hwqp's lcore_id has been updated */ 402 for (i = 0; i < fc_port->num_io_queues; i++) { 403 CU_ASSERT(fc_port->io_queues[i].lcore_id == poll_cnt); 404 } 405 } 406 407 static void 408 remove_hwqps_from_poll_groups_test(void) 409 { 410 unsigned i; 411 struct spdk_nvmf_fc_port *fc_port = NULL; 412 413 SPDK_CU_ASSERT_FATAL(g_nvmf_tprt != NULL); 414 415 fc_port = nvmf_fc_port_lookup(g_fc_port_handle); 416 SPDK_CU_ASSERT_FATAL(fc_port != NULL); 417 418 for (i = 0; i < fc_port->num_io_queues; i++) { 419 nvmf_fc_poll_group_remove_hwqp(&fc_port->io_queues[i], NULL, NULL); 420 poll_threads(); 421 CU_ASSERT(fc_port->io_queues[i].fgroup == 0); 422 } 423 } 424 425 static void 426 destroy_transport_test(void) 427 { 428 unsigned i; 429 430 SPDK_CU_ASSERT_FATAL(g_nvmf_tprt != NULL); 431 432 for (i = 0; i < MAX_FC_UT_POLL_THREADS; i++) { 433 set_thread(i); 434 spdk_nvmf_poll_group_destroy(g_poll_groups[i], NULL, NULL); 435 poll_thread(0); 436 } 437 438 set_thread(0); 439 SPDK_CU_ASSERT_FATAL(g_nvmf_tgt != NULL); 440 g_lld_fini_called = false; 441 spdk_nvmf_tgt_destroy(g_nvmf_tgt, NULL, NULL); 442 poll_threads(); 443 CU_ASSERT(g_lld_fini_called == true); 444 } 445 446 static int 447 nvmf_fc_tests_init(void) 448 { 449 return 0; 450 } 451 452 static int 453 nvmf_fc_tests_fini(void) 454 { 455 free_threads(); 456 return 0; 457 } 458 459 int 460 main(int argc, char **argv) 461 { 462 unsigned int num_failures = 0; 463 CU_pSuite suite = NULL; 464 465 CU_set_error_action(CUEA_ABORT); 466 CU_initialize_registry(); 467 468 suite = CU_add_suite("NVMf-FC", nvmf_fc_tests_init, nvmf_fc_tests_fini); 469 470 CU_ADD_TEST(suite, create_transport_test); 471 CU_ADD_TEST(suite, create_poll_groups_test); 472 CU_ADD_TEST(suite, create_fc_port_test); 473 CU_ADD_TEST(suite, online_fc_port_test); 474 CU_ADD_TEST(suite, poll_group_poll_test); 475 CU_ADD_TEST(suite, remove_hwqps_from_poll_groups_test); 476 CU_ADD_TEST(suite, destroy_transport_test); 477 478 CU_basic_set_mode(CU_BRM_VERBOSE); 479 CU_basic_run_tests(); 480 num_failures = CU_get_number_of_failures(); 481 CU_cleanup_registry(); 482 483 return num_failures; 484 } 485