xref: /spdk/test/unit/lib/bdev/crypto.c/crypto_ut.c (revision 1a00f5c09488e7466a331b8c75cde4969740357f)
1 /*-
2  *   BSD LICENSE
3  *
4  *   Copyright (c) Intel Corporation.
5  *   All rights reserved.
6  *
7  *   Redistribution and use in source and binary forms, with or without
8  *   modification, are permitted provided that the following conditions
9  *   are met:
10  *
11  *     * Redistributions of source code must retain the above copyright
12  *       notice, this list of conditions and the following disclaimer.
13  *     * Redistributions in binary form must reproduce the above copyright
14  *       notice, this list of conditions and the following disclaimer in
15  *       the documentation and/or other materials provided with the
16  *       distribution.
17  *     * Neither the name of Intel Corporation nor the names of its
18  *       contributors may be used to endorse or promote products derived
19  *       from this software without specific prior written permission.
20  *
21  *   THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
22  *   "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
23  *   LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
24  *   A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
25  *   OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
26  *   SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
27  *   LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
28  *   DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
29  *   THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
30  *   (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
31  *   OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
32  */
33 
34 #include "spdk_cunit.h"
35 
36 #include "common/lib/test_env.c"
37 #include "spdk_internal/mock.h"
38 #include "thread/thread_internal.h"
39 #include "unit/lib/json_mock.c"
40 
41 #include <rte_crypto.h>
42 #include <rte_cryptodev.h>
43 
44 #define MAX_TEST_BLOCKS 8192
45 struct rte_crypto_op *g_test_crypto_ops[MAX_TEST_BLOCKS];
46 struct rte_crypto_op *g_test_dev_full_ops[MAX_TEST_BLOCKS];
47 
48 uint16_t g_dequeue_mock;
49 uint16_t g_enqueue_mock;
50 unsigned ut_rte_crypto_op_bulk_alloc;
51 int ut_rte_crypto_op_attach_sym_session = 0;
52 #define MOCK_INFO_GET_1QP_AESNI 0
53 #define MOCK_INFO_GET_1QP_QAT 1
54 #define MOCK_INFO_GET_1QP_BOGUS_PMD 2
55 int ut_rte_cryptodev_info_get = 0;
56 bool ut_rte_cryptodev_info_get_mocked = false;
57 
58 void mock_rte_pktmbuf_free_bulk(struct rte_mbuf **m, unsigned int cnt);
59 #define rte_pktmbuf_free_bulk mock_rte_pktmbuf_free_bulk
60 void mock_rte_pktmbuf_free_bulk(struct rte_mbuf **m, unsigned int cnt)
61 {
62 	spdk_mempool_put_bulk((struct spdk_mempool *)m[0]->pool, (void **)m, cnt);
63 }
64 
65 void mock_rte_pktmbuf_free(struct rte_mbuf *m);
66 #define rte_pktmbuf_free mock_rte_pktmbuf_free
67 void mock_rte_pktmbuf_free(struct rte_mbuf *m)
68 {
69 	spdk_mempool_put((struct spdk_mempool *)m->pool, (void *)m);
70 }
71 
72 void rte_mempool_free(struct rte_mempool *mp)
73 {
74 	spdk_mempool_free((struct spdk_mempool *)mp);
75 }
76 
77 int mock_rte_pktmbuf_alloc_bulk(struct rte_mempool *pool, struct rte_mbuf **mbufs,
78 				unsigned count);
79 #define rte_pktmbuf_alloc_bulk mock_rte_pktmbuf_alloc_bulk
80 int mock_rte_pktmbuf_alloc_bulk(struct rte_mempool *pool, struct rte_mbuf **mbufs,
81 				unsigned count)
82 {
83 	int rc;
84 
85 	rc = spdk_mempool_get_bulk((struct spdk_mempool *)pool, (void **)mbufs, count);
86 	if (rc) {
87 		return rc;
88 	}
89 	for (unsigned i = 0; i < count; i++) {
90 		rte_pktmbuf_reset(mbufs[i]);
91 		mbufs[i]->pool = pool;
92 	}
93 	return rc;
94 }
95 
96 struct rte_mempool *
97 rte_cryptodev_sym_session_pool_create(const char *name, uint32_t nb_elts,
98 				      uint32_t elt_size, uint32_t cache_size,
99 				      uint16_t priv_size, int socket_id)
100 {
101 	struct spdk_mempool *tmp;
102 
103 	tmp = spdk_mempool_create(name, nb_elts, elt_size + priv_size,
104 				  cache_size, socket_id);
105 
106 	return (struct rte_mempool *)tmp;
107 
108 }
109 
110 struct rte_mempool *
111 rte_pktmbuf_pool_create(const char *name, unsigned n, unsigned cache_size,
112 			uint16_t priv_size, uint16_t data_room_size, int socket_id)
113 {
114 	struct spdk_mempool *tmp;
115 
116 	tmp = spdk_mempool_create(name, n, sizeof(struct rte_mbuf) + priv_size,
117 				  cache_size, socket_id);
118 
119 	return (struct rte_mempool *)tmp;
120 }
121 
122 struct rte_mempool *
123 rte_mempool_create(const char *name, unsigned n, unsigned elt_size,
124 		   unsigned cache_size, unsigned private_data_size,
125 		   rte_mempool_ctor_t *mp_init, void *mp_init_arg,
126 		   rte_mempool_obj_cb_t *obj_init, void *obj_init_arg,
127 		   int socket_id, unsigned flags)
128 {
129 	struct spdk_mempool *tmp;
130 
131 	tmp = spdk_mempool_create(name, n, elt_size + private_data_size,
132 				  cache_size, socket_id);
133 
134 	return (struct rte_mempool *)tmp;
135 }
136 
137 DEFINE_RETURN_MOCK(rte_crypto_op_pool_create, struct rte_mempool *);
138 struct rte_mempool *
139 rte_crypto_op_pool_create(const char *name, enum rte_crypto_op_type type,
140 			  unsigned nb_elts, unsigned cache_size,
141 			  uint16_t priv_size, int socket_id)
142 {
143 	struct spdk_mempool *tmp;
144 
145 	HANDLE_RETURN_MOCK(rte_crypto_op_pool_create);
146 
147 	tmp = spdk_mempool_create(name, nb_elts,
148 				  sizeof(struct rte_crypto_op) + priv_size,
149 				  cache_size, socket_id);
150 
151 	return (struct rte_mempool *)tmp;
152 
153 }
154 
155 /* Those functions are defined as static inline in DPDK, so we can't
156  * mock them straight away. We use defines to redirect them into
157  * our custom functions.
158  */
159 static bool g_resubmit_test = false;
160 #define rte_cryptodev_enqueue_burst mock_rte_cryptodev_enqueue_burst
161 static inline uint16_t
162 mock_rte_cryptodev_enqueue_burst(uint8_t dev_id, uint16_t qp_id,
163 				 struct rte_crypto_op **ops, uint16_t nb_ops)
164 {
165 	int i;
166 
167 	CU_ASSERT(nb_ops > 0);
168 
169 	for (i = 0; i < nb_ops; i++) {
170 		/* Use this empty (til now) array of pointers to store
171 		 * enqueued operations for assertion in dev_full test.
172 		 */
173 		g_test_dev_full_ops[i] = *ops++;
174 		if (g_resubmit_test == true) {
175 			CU_ASSERT(g_test_dev_full_ops[i] == (void *)0xDEADBEEF);
176 		}
177 	}
178 
179 	return g_enqueue_mock;
180 }
181 
182 #define rte_cryptodev_dequeue_burst mock_rte_cryptodev_dequeue_burst
183 static inline uint16_t
184 mock_rte_cryptodev_dequeue_burst(uint8_t dev_id, uint16_t qp_id,
185 				 struct rte_crypto_op **ops, uint16_t nb_ops)
186 {
187 	int i;
188 
189 	CU_ASSERT(nb_ops > 0);
190 
191 	for (i = 0; i < g_dequeue_mock; i++) {
192 		*ops++ = g_test_crypto_ops[i];
193 	}
194 
195 	return g_dequeue_mock;
196 }
197 
198 /* Instead of allocating real memory, assign the allocations to our
199  * test array for assertion in tests.
200  */
201 #define rte_crypto_op_bulk_alloc mock_rte_crypto_op_bulk_alloc
202 static inline unsigned
203 mock_rte_crypto_op_bulk_alloc(struct rte_mempool *mempool,
204 			      enum rte_crypto_op_type type,
205 			      struct rte_crypto_op **ops, uint16_t nb_ops)
206 {
207 	int i;
208 
209 	for (i = 0; i < nb_ops; i++) {
210 		*ops++ = g_test_crypto_ops[i];
211 	}
212 	return ut_rte_crypto_op_bulk_alloc;
213 }
214 
215 #define rte_mempool_put_bulk mock_rte_mempool_put_bulk
216 static __rte_always_inline void
217 mock_rte_mempool_put_bulk(struct rte_mempool *mp, void *const *obj_table,
218 			  unsigned int n)
219 {
220 	return;
221 }
222 
223 #define rte_crypto_op_attach_sym_session mock_rte_crypto_op_attach_sym_session
224 static inline int
225 mock_rte_crypto_op_attach_sym_session(struct rte_crypto_op *op,
226 				      struct rte_cryptodev_sym_session *sess)
227 {
228 	return ut_rte_crypto_op_attach_sym_session;
229 }
230 
231 #define rte_lcore_count mock_rte_lcore_count
232 static inline unsigned
233 mock_rte_lcore_count(void)
234 {
235 	return 1;
236 }
237 
238 #include "bdev/crypto/vbdev_crypto.c"
239 
240 /* SPDK stubs */
241 DEFINE_STUB(spdk_bdev_queue_io_wait, int, (struct spdk_bdev *bdev, struct spdk_io_channel *ch,
242 		struct spdk_bdev_io_wait_entry *entry), 0);
243 DEFINE_STUB_V(spdk_bdev_module_list_add, (struct spdk_bdev_module *bdev_module));
244 DEFINE_STUB_V(spdk_bdev_free_io, (struct spdk_bdev_io *g_bdev_io));
245 DEFINE_STUB_V(spdk_bdev_io_put_aux_buf, (struct spdk_bdev_io *bdev_io, void *aux_buf));
246 DEFINE_STUB(spdk_bdev_io_type_supported, bool, (struct spdk_bdev *bdev,
247 		enum spdk_bdev_io_type io_type), 0);
248 DEFINE_STUB_V(spdk_bdev_module_release_bdev, (struct spdk_bdev *bdev));
249 DEFINE_STUB_V(spdk_bdev_close, (struct spdk_bdev_desc *desc));
250 DEFINE_STUB(spdk_bdev_get_name, const char *, (const struct spdk_bdev *bdev), 0);
251 DEFINE_STUB(spdk_bdev_get_buf_align, size_t, (const struct spdk_bdev *bdev), 64);
252 DEFINE_STUB(spdk_bdev_get_io_channel, struct spdk_io_channel *, (struct spdk_bdev_desc *desc), 0);
253 DEFINE_STUB_V(spdk_bdev_unregister, (struct spdk_bdev *bdev, spdk_bdev_unregister_cb cb_fn,
254 				     void *cb_arg));
255 DEFINE_STUB(spdk_bdev_open_ext, int, (const char *bdev_name, bool write,
256 				      spdk_bdev_event_cb_t event_cb,
257 				      void *event_ctx, struct spdk_bdev_desc **_desc), 0);
258 DEFINE_STUB(spdk_bdev_desc_get_bdev, struct spdk_bdev *, (struct spdk_bdev_desc *desc), NULL);
259 DEFINE_STUB(spdk_bdev_module_claim_bdev, int, (struct spdk_bdev *bdev, struct spdk_bdev_desc *desc,
260 		struct spdk_bdev_module *module), 0);
261 DEFINE_STUB_V(spdk_bdev_module_examine_done, (struct spdk_bdev_module *module));
262 DEFINE_STUB(spdk_bdev_register, int, (struct spdk_bdev *vbdev), 0);
263 
264 /* DPDK stubs */
265 #define DPDK_DYNFIELD_OFFSET offsetof(struct rte_mbuf, dynfield1[1])
266 DEFINE_STUB(rte_mbuf_dynfield_register, int, (const struct rte_mbuf_dynfield *params),
267 	    DPDK_DYNFIELD_OFFSET);
268 DEFINE_STUB(rte_cryptodev_count, uint8_t, (void), 0);
269 DEFINE_STUB(rte_socket_id, unsigned, (void), 0);
270 DEFINE_STUB(rte_cryptodev_device_count_by_driver, uint8_t, (uint8_t driver_id), 0);
271 DEFINE_STUB(rte_cryptodev_configure, int, (uint8_t dev_id, struct rte_cryptodev_config *config), 0);
272 DEFINE_STUB(rte_cryptodev_queue_pair_setup, int, (uint8_t dev_id, uint16_t queue_pair_id,
273 		const struct rte_cryptodev_qp_conf *qp_conf, int socket_id), 0);
274 DEFINE_STUB(rte_cryptodev_start, int, (uint8_t dev_id), 0);
275 DEFINE_STUB_V(rte_cryptodev_stop, (uint8_t dev_id));
276 DEFINE_STUB(rte_cryptodev_close, int, (uint8_t dev_id), 0);
277 DEFINE_STUB(rte_cryptodev_sym_session_create, struct rte_cryptodev_sym_session *,
278 	    (struct rte_mempool *mempool), (struct rte_cryptodev_sym_session *)1);
279 DEFINE_STUB(rte_cryptodev_sym_session_init, int, (uint8_t dev_id,
280 		struct rte_cryptodev_sym_session *sess,
281 		struct rte_crypto_sym_xform *xforms, struct rte_mempool *mempool), 0);
282 DEFINE_STUB(rte_vdev_init, int, (const char *name, const char *args), 0);
283 DEFINE_STUB(rte_cryptodev_sym_session_free, int, (struct rte_cryptodev_sym_session *sess), 0);
284 DEFINE_STUB(rte_vdev_uninit, int, (const char *name), 0);
285 
286 struct rte_cryptodev *rte_cryptodevs;
287 
288 /* global vars and setup/cleanup functions used for all test functions */
289 struct spdk_bdev_io *g_bdev_io;
290 struct crypto_bdev_io *g_io_ctx;
291 struct crypto_io_channel *g_crypto_ch;
292 struct spdk_io_channel *g_io_ch;
293 struct vbdev_dev g_device;
294 struct vbdev_crypto g_crypto_bdev;
295 struct device_qp g_dev_qp;
296 
297 void
298 rte_cryptodev_info_get(uint8_t dev_id, struct rte_cryptodev_info *dev_info)
299 {
300 	dev_info->max_nb_queue_pairs = 1;
301 	if (ut_rte_cryptodev_info_get == MOCK_INFO_GET_1QP_AESNI) {
302 		dev_info->driver_name = g_driver_names[0];
303 	} else if (ut_rte_cryptodev_info_get == MOCK_INFO_GET_1QP_QAT) {
304 		dev_info->driver_name = g_driver_names[1];
305 	} else if (ut_rte_cryptodev_info_get == MOCK_INFO_GET_1QP_BOGUS_PMD) {
306 		dev_info->driver_name = "junk";
307 	}
308 }
309 
310 unsigned int
311 rte_cryptodev_sym_get_private_session_size(uint8_t dev_id)
312 {
313 	return (unsigned int)dev_id;
314 }
315 
316 void
317 spdk_bdev_io_get_aux_buf(struct spdk_bdev_io *bdev_io, spdk_bdev_io_get_aux_buf_cb cb)
318 {
319 	cb(g_io_ch, g_bdev_io, (void *)0xDEADBEEF);
320 }
321 
322 void
323 spdk_bdev_io_get_buf(struct spdk_bdev_io *bdev_io, spdk_bdev_io_get_buf_cb cb, uint64_t len)
324 {
325 	cb(g_io_ch, g_bdev_io, true);
326 }
327 
328 /* Mock these functions to call the callback and then return the value we require */
329 int ut_spdk_bdev_readv_blocks = 0;
330 bool ut_spdk_bdev_readv_blocks_mocked = false;
331 int
332 spdk_bdev_readv_blocks(struct spdk_bdev_desc *desc, struct spdk_io_channel *ch,
333 		       struct iovec *iov, int iovcnt,
334 		       uint64_t offset_blocks, uint64_t num_blocks,
335 		       spdk_bdev_io_completion_cb cb, void *cb_arg)
336 {
337 	cb(g_bdev_io, !ut_spdk_bdev_readv_blocks, cb_arg);
338 	return ut_spdk_bdev_readv_blocks;
339 }
340 
341 int ut_spdk_bdev_writev_blocks = 0;
342 bool ut_spdk_bdev_writev_blocks_mocked = false;
343 int
344 spdk_bdev_writev_blocks(struct spdk_bdev_desc *desc, struct spdk_io_channel *ch,
345 			struct iovec *iov, int iovcnt,
346 			uint64_t offset_blocks, uint64_t num_blocks,
347 			spdk_bdev_io_completion_cb cb, void *cb_arg)
348 {
349 	cb(g_bdev_io, !ut_spdk_bdev_writev_blocks, cb_arg);
350 	return ut_spdk_bdev_writev_blocks;
351 }
352 
353 int ut_spdk_bdev_unmap_blocks = 0;
354 bool ut_spdk_bdev_unmap_blocks_mocked = false;
355 int
356 spdk_bdev_unmap_blocks(struct spdk_bdev_desc *desc, struct spdk_io_channel *ch,
357 		       uint64_t offset_blocks, uint64_t num_blocks,
358 		       spdk_bdev_io_completion_cb cb, void *cb_arg)
359 {
360 	cb(g_bdev_io, !ut_spdk_bdev_unmap_blocks, cb_arg);
361 	return ut_spdk_bdev_unmap_blocks;
362 }
363 
364 int ut_spdk_bdev_flush_blocks = 0;
365 bool ut_spdk_bdev_flush_blocks_mocked = false;
366 int
367 spdk_bdev_flush_blocks(struct spdk_bdev_desc *desc, struct spdk_io_channel *ch,
368 		       uint64_t offset_blocks, uint64_t num_blocks, spdk_bdev_io_completion_cb cb,
369 		       void *cb_arg)
370 {
371 	cb(g_bdev_io, !ut_spdk_bdev_flush_blocks, cb_arg);
372 	return ut_spdk_bdev_flush_blocks;
373 }
374 
375 int ut_spdk_bdev_reset = 0;
376 bool ut_spdk_bdev_reset_mocked = false;
377 int
378 spdk_bdev_reset(struct spdk_bdev_desc *desc, struct spdk_io_channel *ch,
379 		spdk_bdev_io_completion_cb cb, void *cb_arg)
380 {
381 	cb(g_bdev_io, !ut_spdk_bdev_reset, cb_arg);
382 	return ut_spdk_bdev_reset;
383 }
384 
385 bool g_completion_called = false;
386 void
387 spdk_bdev_io_complete(struct spdk_bdev_io *bdev_io, enum spdk_bdev_io_status status)
388 {
389 	bdev_io->internal.status = status;
390 	g_completion_called = true;
391 }
392 
393 /* Global setup for all tests that share a bunch of preparation... */
394 static int
395 test_setup(void)
396 {
397 	int i, rc;
398 
399 	/* Prepare essential variables for test routines */
400 	g_bdev_io = calloc(1, sizeof(struct spdk_bdev_io) + sizeof(struct crypto_bdev_io));
401 	g_bdev_io->u.bdev.iovs = calloc(1, sizeof(struct iovec) * 128);
402 	g_bdev_io->bdev = &g_crypto_bdev.crypto_bdev;
403 	g_io_ch = calloc(1, sizeof(struct spdk_io_channel) + sizeof(struct crypto_io_channel));
404 	g_crypto_ch = (struct crypto_io_channel *)spdk_io_channel_get_ctx(g_io_ch);
405 	g_io_ctx = (struct crypto_bdev_io *)g_bdev_io->driver_ctx;
406 	memset(&g_device, 0, sizeof(struct vbdev_dev));
407 	memset(&g_crypto_bdev, 0, sizeof(struct vbdev_crypto));
408 	g_dev_qp.device = &g_device;
409 	g_io_ctx->crypto_ch = g_crypto_ch;
410 	g_io_ctx->crypto_bdev = &g_crypto_bdev;
411 	g_crypto_ch->device_qp = &g_dev_qp;
412 	TAILQ_INIT(&g_crypto_ch->pending_cry_ios);
413 	TAILQ_INIT(&g_crypto_ch->queued_cry_ops);
414 
415 	/* Allocate a real mbuf pool so we can test error paths */
416 	g_mbuf_mp = rte_pktmbuf_pool_create("mbuf_mp", NUM_MBUFS,
417 					    (unsigned)SPDK_MEMPOOL_DEFAULT_CACHE_SIZE,
418 					    0, 0, SPDK_ENV_SOCKET_ID_ANY);
419 	/* Instead of allocating real rte mempools for these, it's easier and provides the
420 	 * same coverage just calloc them here.
421 	 */
422 	for (i = 0; i < MAX_TEST_BLOCKS; i++) {
423 		rc = posix_memalign((void **)&g_test_crypto_ops[i], 64,
424 				    sizeof(struct rte_crypto_op) + sizeof(struct rte_crypto_sym_op) +
425 				    AES_CBC_IV_LENGTH + QUEUED_OP_LENGTH);
426 		if (rc != 0) {
427 			assert(false);
428 		}
429 		memset(g_test_crypto_ops[i], 0, sizeof(struct rte_crypto_op) +
430 		       sizeof(struct rte_crypto_sym_op) + QUEUED_OP_LENGTH);
431 	}
432 	g_mbuf_offset = DPDK_DYNFIELD_OFFSET;
433 
434 	return 0;
435 }
436 
437 /* Global teardown for all tests */
438 static int
439 test_cleanup(void)
440 {
441 	int i;
442 
443 	if (g_crypto_op_mp) {
444 		rte_mempool_free(g_crypto_op_mp);
445 		g_crypto_op_mp = NULL;
446 	}
447 	if (g_mbuf_mp) {
448 		rte_mempool_free(g_mbuf_mp);
449 		g_mbuf_mp = NULL;
450 	}
451 	if (g_session_mp) {
452 		rte_mempool_free(g_session_mp);
453 		g_session_mp = NULL;
454 	}
455 	if (g_session_mp_priv != NULL) {
456 		/* g_session_mp_priv may or may not be set depending on the DPDK version */
457 		rte_mempool_free(g_session_mp_priv);
458 		g_session_mp_priv = NULL;
459 	}
460 
461 	for (i = 0; i < MAX_TEST_BLOCKS; i++) {
462 		free(g_test_crypto_ops[i]);
463 	}
464 	free(g_bdev_io->u.bdev.iovs);
465 	free(g_bdev_io);
466 	free(g_io_ch);
467 	return 0;
468 }
469 
470 static void
471 test_error_paths(void)
472 {
473 	/* Single element block size write, just to test error paths
474 	 * in vbdev_crypto_submit_request().
475 	 */
476 	g_bdev_io->internal.status = SPDK_BDEV_IO_STATUS_SUCCESS;
477 	g_bdev_io->u.bdev.iovcnt = 1;
478 	g_bdev_io->u.bdev.num_blocks = 1;
479 	g_bdev_io->u.bdev.iovs[0].iov_len = 512;
480 	g_bdev_io->u.bdev.iovs[0].iov_base = (void *)0xDEADBEEF;
481 	g_crypto_bdev.crypto_bdev.blocklen = 512;
482 	g_bdev_io->type = SPDK_BDEV_IO_TYPE_WRITE;
483 	g_enqueue_mock = g_dequeue_mock = ut_rte_crypto_op_bulk_alloc = 1;
484 
485 	/* test failure of spdk_mempool_get_bulk(), will result in success because it
486 	 * will get queued.
487 	 */
488 	g_bdev_io->internal.status = SPDK_BDEV_IO_STATUS_SUCCESS;
489 	MOCK_SET(spdk_mempool_get, NULL);
490 	vbdev_crypto_submit_request(g_io_ch, g_bdev_io);
491 	CU_ASSERT(g_bdev_io->internal.status == SPDK_BDEV_IO_STATUS_SUCCESS);
492 
493 	/* same thing but switch to reads to test error path in _crypto_complete_io() */
494 	g_bdev_io->type = SPDK_BDEV_IO_TYPE_READ;
495 	g_bdev_io->internal.status = SPDK_BDEV_IO_STATUS_SUCCESS;
496 	TAILQ_INSERT_TAIL(&g_crypto_ch->pending_cry_ios, g_bdev_io, module_link);
497 	vbdev_crypto_submit_request(g_io_ch, g_bdev_io);
498 	CU_ASSERT(g_bdev_io->internal.status == SPDK_BDEV_IO_STATUS_FAILED);
499 	/* Now with the read_blocks failing */
500 	g_bdev_io->type = SPDK_BDEV_IO_TYPE_READ;
501 	g_bdev_io->internal.status = SPDK_BDEV_IO_STATUS_SUCCESS;
502 	MOCK_SET(spdk_bdev_readv_blocks, -1);
503 	vbdev_crypto_submit_request(g_io_ch, g_bdev_io);
504 	CU_ASSERT(g_bdev_io->internal.status == SPDK_BDEV_IO_STATUS_FAILED);
505 	MOCK_SET(spdk_bdev_readv_blocks, 0);
506 	MOCK_CLEAR(spdk_mempool_get);
507 
508 	/* test failure of rte_crypto_op_bulk_alloc() */
509 	g_bdev_io->internal.status = SPDK_BDEV_IO_STATUS_SUCCESS;
510 	ut_rte_crypto_op_bulk_alloc = 0;
511 	vbdev_crypto_submit_request(g_io_ch, g_bdev_io);
512 	CU_ASSERT(g_bdev_io->internal.status == SPDK_BDEV_IO_STATUS_FAILED);
513 	ut_rte_crypto_op_bulk_alloc = 1;
514 
515 	/* test failure of rte_crypto_op_attach_sym_session() */
516 	g_bdev_io->internal.status = SPDK_BDEV_IO_STATUS_SUCCESS;
517 	ut_rte_crypto_op_attach_sym_session = -1;
518 	vbdev_crypto_submit_request(g_io_ch, g_bdev_io);
519 	CU_ASSERT(g_bdev_io->internal.status == SPDK_BDEV_IO_STATUS_FAILED);
520 	ut_rte_crypto_op_attach_sym_session = 0;
521 }
522 
523 static void
524 test_simple_write(void)
525 {
526 	/* Single element block size write */
527 	g_bdev_io->internal.status = SPDK_BDEV_IO_STATUS_SUCCESS;
528 	g_bdev_io->u.bdev.iovcnt = 1;
529 	g_bdev_io->u.bdev.num_blocks = 1;
530 	g_bdev_io->u.bdev.offset_blocks = 0;
531 	g_bdev_io->u.bdev.iovs[0].iov_len = 512;
532 	g_bdev_io->u.bdev.iovs[0].iov_base = &test_simple_write;
533 	g_crypto_bdev.crypto_bdev.blocklen = 512;
534 	g_bdev_io->type = SPDK_BDEV_IO_TYPE_WRITE;
535 	g_enqueue_mock = g_dequeue_mock = ut_rte_crypto_op_bulk_alloc = 1;
536 
537 	vbdev_crypto_submit_request(g_io_ch, g_bdev_io);
538 	CU_ASSERT(g_bdev_io->internal.status == SPDK_BDEV_IO_STATUS_SUCCESS);
539 	CU_ASSERT(g_io_ctx->cryop_cnt_remaining == 1);
540 	CU_ASSERT(g_io_ctx->aux_buf_iov.iov_len == 512);
541 	CU_ASSERT(g_io_ctx->aux_buf_iov.iov_base != NULL);
542 	CU_ASSERT(g_io_ctx->aux_offset_blocks == 0);
543 	CU_ASSERT(g_io_ctx->aux_num_blocks == 1);
544 	CU_ASSERT(g_test_crypto_ops[0]->sym->m_src->buf_addr == &test_simple_write);
545 	CU_ASSERT(g_test_crypto_ops[0]->sym->m_src->data_len == 512);
546 	CU_ASSERT(g_test_crypto_ops[0]->sym->m_src->next == NULL);
547 	CU_ASSERT(g_test_crypto_ops[0]->sym->cipher.data.length == 512);
548 	CU_ASSERT(g_test_crypto_ops[0]->sym->cipher.data.offset == 0);
549 	CU_ASSERT(*RTE_MBUF_DYNFIELD(g_test_crypto_ops[0]->sym->m_src, g_mbuf_offset,
550 				     uint64_t *) == (uint64_t)g_bdev_io);
551 	CU_ASSERT(g_test_crypto_ops[0]->sym->m_dst->buf_addr != NULL);
552 	CU_ASSERT(g_test_crypto_ops[0]->sym->m_dst->data_len == 512);
553 
554 	rte_pktmbuf_free(g_test_crypto_ops[0]->sym->m_src);
555 	rte_pktmbuf_free(g_test_crypto_ops[0]->sym->m_dst);
556 }
557 
558 static void
559 test_simple_read(void)
560 {
561 	/* Single element block size read */
562 	g_bdev_io->internal.status = SPDK_BDEV_IO_STATUS_SUCCESS;
563 	g_bdev_io->u.bdev.iovcnt = 1;
564 	g_bdev_io->u.bdev.num_blocks = 1;
565 	g_bdev_io->u.bdev.iovs[0].iov_len = 512;
566 	g_bdev_io->u.bdev.iovs[0].iov_base = &test_simple_read;
567 	g_crypto_bdev.crypto_bdev.blocklen = 512;
568 	g_bdev_io->type = SPDK_BDEV_IO_TYPE_READ;
569 	g_enqueue_mock = g_dequeue_mock = ut_rte_crypto_op_bulk_alloc = 1;
570 
571 	vbdev_crypto_submit_request(g_io_ch, g_bdev_io);
572 	CU_ASSERT(g_bdev_io->internal.status == SPDK_BDEV_IO_STATUS_SUCCESS);
573 	CU_ASSERT(g_io_ctx->cryop_cnt_remaining == 1);
574 	CU_ASSERT(g_test_crypto_ops[0]->sym->m_src->buf_addr == &test_simple_read);
575 	CU_ASSERT(g_test_crypto_ops[0]->sym->m_src->data_len == 512);
576 	CU_ASSERT(g_test_crypto_ops[0]->sym->m_src->next == NULL);
577 	CU_ASSERT(g_test_crypto_ops[0]->sym->cipher.data.length == 512);
578 	CU_ASSERT(g_test_crypto_ops[0]->sym->cipher.data.offset == 0);
579 	CU_ASSERT(*RTE_MBUF_DYNFIELD(g_test_crypto_ops[0]->sym->m_src, g_mbuf_offset,
580 				     uint64_t *) == (uint64_t)g_bdev_io);
581 	CU_ASSERT(g_test_crypto_ops[0]->sym->m_dst == NULL);
582 
583 	rte_pktmbuf_free(g_test_crypto_ops[0]->sym->m_src);
584 }
585 
586 static void
587 test_large_rw(void)
588 {
589 	unsigned block_len = 512;
590 	unsigned num_blocks = CRYPTO_MAX_IO / block_len;
591 	unsigned io_len = block_len * num_blocks;
592 	unsigned i;
593 
594 	/* Multi block size read, multi-element */
595 	g_bdev_io->internal.status = SPDK_BDEV_IO_STATUS_SUCCESS;
596 	g_bdev_io->u.bdev.iovcnt = 1;
597 	g_bdev_io->u.bdev.num_blocks = num_blocks;
598 	g_bdev_io->u.bdev.iovs[0].iov_len = io_len;
599 	g_bdev_io->u.bdev.iovs[0].iov_base = &test_large_rw;
600 	g_crypto_bdev.crypto_bdev.blocklen = block_len;
601 	g_bdev_io->type = SPDK_BDEV_IO_TYPE_READ;
602 	g_enqueue_mock = g_dequeue_mock = ut_rte_crypto_op_bulk_alloc = num_blocks;
603 
604 	vbdev_crypto_submit_request(g_io_ch, g_bdev_io);
605 	CU_ASSERT(g_bdev_io->internal.status == SPDK_BDEV_IO_STATUS_SUCCESS);
606 	CU_ASSERT(g_io_ctx->cryop_cnt_remaining == (int)num_blocks);
607 
608 	for (i = 0; i < num_blocks; i++) {
609 		CU_ASSERT(g_test_crypto_ops[i]->sym->m_src->buf_addr == &test_large_rw + (i * block_len));
610 		CU_ASSERT(g_test_crypto_ops[i]->sym->m_src->data_len == block_len);
611 		CU_ASSERT(g_test_crypto_ops[i]->sym->m_src->next == NULL);
612 		CU_ASSERT(g_test_crypto_ops[i]->sym->cipher.data.length == block_len);
613 		CU_ASSERT(g_test_crypto_ops[i]->sym->cipher.data.offset == 0);
614 		CU_ASSERT(*RTE_MBUF_DYNFIELD(g_test_crypto_ops[i]->sym->m_src, g_mbuf_offset,
615 					     uint64_t *) == (uint64_t)g_bdev_io);
616 		CU_ASSERT(g_test_crypto_ops[i]->sym->m_dst == NULL);
617 		rte_pktmbuf_free(g_test_crypto_ops[i]->sym->m_src);
618 	}
619 
620 	/* Multi block size write, multi-element */
621 	g_bdev_io->internal.status = SPDK_BDEV_IO_STATUS_SUCCESS;
622 	g_bdev_io->u.bdev.iovcnt = 1;
623 	g_bdev_io->u.bdev.num_blocks = num_blocks;
624 	g_bdev_io->u.bdev.iovs[0].iov_len = io_len;
625 	g_bdev_io->u.bdev.iovs[0].iov_base = &test_large_rw;
626 	g_crypto_bdev.crypto_bdev.blocklen = block_len;
627 	g_bdev_io->type = SPDK_BDEV_IO_TYPE_WRITE;
628 	g_enqueue_mock = g_dequeue_mock = ut_rte_crypto_op_bulk_alloc = num_blocks;
629 
630 	vbdev_crypto_submit_request(g_io_ch, g_bdev_io);
631 	CU_ASSERT(g_bdev_io->internal.status == SPDK_BDEV_IO_STATUS_SUCCESS);
632 	CU_ASSERT(g_io_ctx->cryop_cnt_remaining == (int)num_blocks);
633 
634 	for (i = 0; i < num_blocks; i++) {
635 		CU_ASSERT(g_test_crypto_ops[i]->sym->m_src->buf_addr == &test_large_rw + (i * block_len));
636 		CU_ASSERT(g_test_crypto_ops[i]->sym->m_src->data_len == block_len);
637 		CU_ASSERT(g_test_crypto_ops[i]->sym->m_src->next == NULL);
638 		CU_ASSERT(g_test_crypto_ops[i]->sym->cipher.data.length == block_len);
639 		CU_ASSERT(g_test_crypto_ops[i]->sym->cipher.data.offset == 0);
640 		CU_ASSERT(*RTE_MBUF_DYNFIELD(g_test_crypto_ops[i]->sym->m_src, g_mbuf_offset,
641 					     uint64_t *) == (uint64_t)g_bdev_io);
642 		CU_ASSERT(g_io_ctx->aux_buf_iov.iov_len == io_len);
643 		CU_ASSERT(g_io_ctx->aux_buf_iov.iov_base != NULL);
644 		CU_ASSERT(g_io_ctx->aux_offset_blocks == 0);
645 		CU_ASSERT(g_io_ctx->aux_num_blocks == num_blocks);
646 		CU_ASSERT(g_test_crypto_ops[i]->sym->m_dst->buf_addr != NULL);
647 		CU_ASSERT(g_test_crypto_ops[i]->sym->m_dst->data_len == block_len);
648 		rte_pktmbuf_free(g_test_crypto_ops[i]->sym->m_src);
649 		rte_pktmbuf_free(g_test_crypto_ops[i]->sym->m_dst);
650 	}
651 }
652 
653 static void
654 test_dev_full(void)
655 {
656 	struct vbdev_crypto_op *queued_op;
657 	struct rte_crypto_sym_op *sym_op;
658 	struct crypto_bdev_io *io_ctx;
659 
660 	/* Two element block size read */
661 	g_bdev_io->internal.status = SPDK_BDEV_IO_STATUS_SUCCESS;
662 	g_bdev_io->u.bdev.iovcnt = 1;
663 	g_bdev_io->u.bdev.num_blocks = 2;
664 	g_bdev_io->u.bdev.iovs[0].iov_len = 512;
665 	g_bdev_io->u.bdev.iovs[0].iov_base = (void *)0xDEADBEEF;
666 	g_bdev_io->u.bdev.iovs[1].iov_len = 512;
667 	g_bdev_io->u.bdev.iovs[1].iov_base = (void *)0xFEEDBEEF;
668 	g_crypto_bdev.crypto_bdev.blocklen = 512;
669 	g_bdev_io->type = SPDK_BDEV_IO_TYPE_READ;
670 	g_enqueue_mock = g_dequeue_mock = 1;
671 	ut_rte_crypto_op_bulk_alloc = 2;
672 
673 	g_test_crypto_ops[1]->status = RTE_CRYPTO_OP_STATUS_NOT_PROCESSED;
674 	CU_ASSERT(TAILQ_EMPTY(&g_crypto_ch->queued_cry_ops) == true);
675 
676 	vbdev_crypto_submit_request(g_io_ch, g_bdev_io);
677 	CU_ASSERT(g_bdev_io->internal.status == SPDK_BDEV_IO_STATUS_SUCCESS);
678 	CU_ASSERT(g_io_ctx->cryop_cnt_remaining == 2);
679 	sym_op = g_test_crypto_ops[0]->sym;
680 	CU_ASSERT(sym_op->m_src->buf_addr == (void *)0xDEADBEEF);
681 	CU_ASSERT(sym_op->m_src->data_len == 512);
682 	CU_ASSERT(sym_op->m_src->next == NULL);
683 	CU_ASSERT(sym_op->cipher.data.length == 512);
684 	CU_ASSERT(sym_op->cipher.data.offset == 0);
685 	CU_ASSERT(*RTE_MBUF_DYNFIELD(sym_op->m_src, g_mbuf_offset, uint64_t *) == (uint64_t)g_bdev_io);
686 	CU_ASSERT(sym_op->m_dst == NULL);
687 
688 	/* make sure one got queued and confirm its values */
689 	CU_ASSERT(TAILQ_EMPTY(&g_crypto_ch->queued_cry_ops) == false);
690 	queued_op = TAILQ_FIRST(&g_crypto_ch->queued_cry_ops);
691 	sym_op = queued_op->crypto_op->sym;
692 	TAILQ_REMOVE(&g_crypto_ch->queued_cry_ops, queued_op, link);
693 	CU_ASSERT(queued_op->bdev_io == g_bdev_io);
694 	CU_ASSERT(queued_op->crypto_op == g_test_crypto_ops[1]);
695 	CU_ASSERT(sym_op->m_src->buf_addr == (void *)0xFEEDBEEF);
696 	CU_ASSERT(sym_op->m_src->data_len == 512);
697 	CU_ASSERT(sym_op->m_src->next == NULL);
698 	CU_ASSERT(sym_op->cipher.data.length == 512);
699 	CU_ASSERT(sym_op->cipher.data.offset == 0);
700 	CU_ASSERT(*RTE_MBUF_DYNFIELD(sym_op->m_src, g_mbuf_offset, uint64_t *) == (uint64_t)g_bdev_io);
701 	CU_ASSERT(sym_op->m_dst == NULL);
702 	CU_ASSERT(TAILQ_EMPTY(&g_crypto_ch->queued_cry_ops) == true);
703 	rte_pktmbuf_free(g_test_crypto_ops[0]->sym->m_src);
704 	rte_pktmbuf_free(g_test_crypto_ops[1]->sym->m_src);
705 
706 	/* Non-busy reason for enqueue failure, all were rejected. */
707 	g_enqueue_mock = 0;
708 	g_test_crypto_ops[0]->status = RTE_CRYPTO_OP_STATUS_ERROR;
709 	vbdev_crypto_submit_request(g_io_ch, g_bdev_io);
710 	io_ctx = (struct crypto_bdev_io *)g_bdev_io->driver_ctx;
711 	CU_ASSERT(io_ctx->bdev_io_status == SPDK_BDEV_IO_STATUS_FAILED);
712 }
713 
714 static void
715 test_crazy_rw(void)
716 {
717 	unsigned block_len = 512;
718 	int num_blocks = 4;
719 	int i;
720 
721 	/* Multi block size read, single element, strange IOV makeup */
722 	g_bdev_io->internal.status = SPDK_BDEV_IO_STATUS_SUCCESS;
723 	g_bdev_io->u.bdev.iovcnt = 3;
724 	g_bdev_io->u.bdev.num_blocks = num_blocks;
725 	g_bdev_io->u.bdev.iovs[0].iov_len = 512;
726 	g_bdev_io->u.bdev.iovs[0].iov_base = &test_crazy_rw;
727 	g_bdev_io->u.bdev.iovs[1].iov_len = 1024;
728 	g_bdev_io->u.bdev.iovs[1].iov_base = &test_crazy_rw + 512;
729 	g_bdev_io->u.bdev.iovs[2].iov_len = 512;
730 	g_bdev_io->u.bdev.iovs[2].iov_base = &test_crazy_rw + 512 + 1024;
731 
732 	g_crypto_bdev.crypto_bdev.blocklen = block_len;
733 	g_bdev_io->type = SPDK_BDEV_IO_TYPE_READ;
734 	g_enqueue_mock = g_dequeue_mock = ut_rte_crypto_op_bulk_alloc = num_blocks;
735 
736 	vbdev_crypto_submit_request(g_io_ch, g_bdev_io);
737 	CU_ASSERT(g_bdev_io->internal.status == SPDK_BDEV_IO_STATUS_SUCCESS);
738 	CU_ASSERT(g_io_ctx->cryop_cnt_remaining == num_blocks);
739 
740 	for (i = 0; i < num_blocks; i++) {
741 		CU_ASSERT(g_test_crypto_ops[i]->sym->m_src->buf_addr == &test_crazy_rw + (i * block_len));
742 		CU_ASSERT(g_test_crypto_ops[i]->sym->m_src->data_len == block_len);
743 		CU_ASSERT(g_test_crypto_ops[i]->sym->m_src->next == NULL);
744 		CU_ASSERT(g_test_crypto_ops[i]->sym->cipher.data.length == block_len);
745 		CU_ASSERT(g_test_crypto_ops[i]->sym->cipher.data.offset == 0);
746 		CU_ASSERT(*RTE_MBUF_DYNFIELD(g_test_crypto_ops[i]->sym->m_src, g_mbuf_offset,
747 					     uint64_t *) == (uint64_t)g_bdev_io);
748 		CU_ASSERT(g_test_crypto_ops[i]->sym->m_src == g_test_crypto_ops[i]->sym->m_src);
749 		CU_ASSERT(g_test_crypto_ops[i]->sym->m_dst == NULL);
750 		rte_pktmbuf_free(g_test_crypto_ops[i]->sym->m_src);
751 	}
752 
753 	/* Multi block size write, single element strange IOV makeup */
754 	num_blocks = 8;
755 	g_bdev_io->internal.status = SPDK_BDEV_IO_STATUS_SUCCESS;
756 	g_bdev_io->u.bdev.iovcnt = 4;
757 	g_bdev_io->u.bdev.num_blocks = num_blocks;
758 	g_bdev_io->u.bdev.iovs[0].iov_len = 2048;
759 	g_bdev_io->u.bdev.iovs[0].iov_base = &test_crazy_rw;
760 	g_bdev_io->u.bdev.iovs[1].iov_len = 512;
761 	g_bdev_io->u.bdev.iovs[1].iov_base = &test_crazy_rw + 2048;
762 	g_bdev_io->u.bdev.iovs[2].iov_len = 512;
763 	g_bdev_io->u.bdev.iovs[2].iov_base = &test_crazy_rw + 2048 + 512;
764 	g_bdev_io->u.bdev.iovs[3].iov_len = 1024;
765 	g_bdev_io->u.bdev.iovs[3].iov_base = &test_crazy_rw + 2048 + 512 + 512;
766 
767 	g_crypto_bdev.crypto_bdev.blocklen = block_len;
768 	g_bdev_io->type = SPDK_BDEV_IO_TYPE_WRITE;
769 	g_enqueue_mock = g_dequeue_mock = ut_rte_crypto_op_bulk_alloc = num_blocks;
770 
771 	vbdev_crypto_submit_request(g_io_ch, g_bdev_io);
772 	CU_ASSERT(g_bdev_io->internal.status == SPDK_BDEV_IO_STATUS_SUCCESS);
773 	CU_ASSERT(g_io_ctx->cryop_cnt_remaining == num_blocks);
774 
775 	for (i = 0; i < num_blocks; i++) {
776 		CU_ASSERT(g_test_crypto_ops[i]->sym->m_src->buf_addr == &test_crazy_rw + (i * block_len));
777 		CU_ASSERT(g_test_crypto_ops[i]->sym->m_src->data_len == block_len);
778 		CU_ASSERT(g_test_crypto_ops[i]->sym->m_src->next == NULL);
779 		CU_ASSERT(g_test_crypto_ops[i]->sym->cipher.data.length == block_len);
780 		CU_ASSERT(g_test_crypto_ops[i]->sym->cipher.data.offset == 0);
781 		CU_ASSERT(*RTE_MBUF_DYNFIELD(g_test_crypto_ops[i]->sym->m_src, g_mbuf_offset,
782 					     uint64_t *) == (uint64_t)g_bdev_io);
783 		CU_ASSERT(g_test_crypto_ops[i]->sym->m_src == g_test_crypto_ops[i]->sym->m_src);
784 		CU_ASSERT(g_test_crypto_ops[i]->sym->m_dst == g_test_crypto_ops[i]->sym->m_dst);
785 		rte_pktmbuf_free(g_test_crypto_ops[i]->sym->m_src);
786 		rte_pktmbuf_free(g_test_crypto_ops[i]->sym->m_dst);
787 	}
788 }
789 
790 static void
791 test_passthru(void)
792 {
793 	/* Make sure these follow our completion callback, test success & fail. */
794 	g_bdev_io->type = SPDK_BDEV_IO_TYPE_UNMAP;
795 	MOCK_SET(spdk_bdev_unmap_blocks, 0);
796 	vbdev_crypto_submit_request(g_io_ch, g_bdev_io);
797 	CU_ASSERT(g_bdev_io->internal.status == SPDK_BDEV_IO_STATUS_SUCCESS);
798 	MOCK_SET(spdk_bdev_unmap_blocks, -1);
799 	vbdev_crypto_submit_request(g_io_ch, g_bdev_io);
800 	CU_ASSERT(g_bdev_io->internal.status == SPDK_BDEV_IO_STATUS_FAILED);
801 	MOCK_CLEAR(spdk_bdev_unmap_blocks);
802 
803 	g_bdev_io->type = SPDK_BDEV_IO_TYPE_FLUSH;
804 	MOCK_SET(spdk_bdev_flush_blocks, 0);
805 	vbdev_crypto_submit_request(g_io_ch, g_bdev_io);
806 	CU_ASSERT(g_bdev_io->internal.status == SPDK_BDEV_IO_STATUS_SUCCESS);
807 	MOCK_SET(spdk_bdev_flush_blocks, -1);
808 	vbdev_crypto_submit_request(g_io_ch, g_bdev_io);
809 	CU_ASSERT(g_bdev_io->internal.status == SPDK_BDEV_IO_STATUS_FAILED);
810 	MOCK_CLEAR(spdk_bdev_flush_blocks);
811 
812 	/* We should never get a WZ command, we report that we don't support it. */
813 	g_bdev_io->type = SPDK_BDEV_IO_TYPE_WRITE_ZEROES;
814 	vbdev_crypto_submit_request(g_io_ch, g_bdev_io);
815 	CU_ASSERT(g_bdev_io->internal.status == SPDK_BDEV_IO_STATUS_FAILED);
816 }
817 
818 static void
819 test_reset(void)
820 {
821 	/* TODO: There are a few different ways to do this given that
822 	 * the code uses spdk_for_each_channel() to implement reset
823 	 * handling. Submitting w/o UT for this function for now and
824 	 * will follow up with something shortly.
825 	 */
826 }
827 
828 static void
829 init_cleanup(void)
830 {
831 	if (g_crypto_op_mp) {
832 		rte_mempool_free(g_crypto_op_mp);
833 		g_crypto_op_mp = NULL;
834 	}
835 	if (g_mbuf_mp) {
836 		rte_mempool_free(g_mbuf_mp);
837 		g_mbuf_mp = NULL;
838 	}
839 	if (g_session_mp) {
840 		rte_mempool_free(g_session_mp);
841 		g_session_mp = NULL;
842 	}
843 	if (g_session_mp_priv != NULL) {
844 		/* g_session_mp_priv may or may not be set depending on the DPDK version */
845 		rte_mempool_free(g_session_mp_priv);
846 		g_session_mp_priv = NULL;
847 	}
848 }
849 
850 static void
851 test_initdrivers(void)
852 {
853 	int rc;
854 	static struct rte_mempool *orig_mbuf_mp;
855 	static struct rte_mempool *orig_session_mp;
856 	static struct rte_mempool *orig_session_mp_priv;
857 
858 	/* These tests will alloc and free our g_mbuf_mp
859 	 * so save that off here and restore it after each test is over.
860 	 */
861 	orig_mbuf_mp = g_mbuf_mp;
862 	orig_session_mp = g_session_mp;
863 	orig_session_mp_priv = g_session_mp_priv;
864 
865 	g_session_mp_priv = NULL;
866 	g_session_mp = NULL;
867 	g_mbuf_mp = NULL;
868 
869 	/* No drivers available, not an error though */
870 	MOCK_SET(rte_cryptodev_count, 0);
871 	rc = vbdev_crypto_init_crypto_drivers();
872 	CU_ASSERT(rc == 0);
873 	CU_ASSERT(g_mbuf_mp == NULL);
874 	CU_ASSERT(g_session_mp == NULL);
875 	CU_ASSERT(g_session_mp_priv == NULL);
876 
877 	/* Can't create session pool. */
878 	MOCK_SET(rte_cryptodev_count, 2);
879 	MOCK_SET(spdk_mempool_create, NULL);
880 	rc = vbdev_crypto_init_crypto_drivers();
881 	CU_ASSERT(rc == -ENOMEM);
882 	CU_ASSERT(g_mbuf_mp == NULL);
883 	CU_ASSERT(g_session_mp == NULL);
884 	CU_ASSERT(g_session_mp_priv == NULL);
885 	MOCK_CLEAR(spdk_mempool_create);
886 
887 	/* Can't create op pool. */
888 	MOCK_SET(rte_crypto_op_pool_create, NULL);
889 	rc = vbdev_crypto_init_crypto_drivers();
890 	CU_ASSERT(rc == -ENOMEM);
891 	CU_ASSERT(g_mbuf_mp == NULL);
892 	CU_ASSERT(g_session_mp == NULL);
893 	CU_ASSERT(g_session_mp_priv == NULL);
894 	MOCK_CLEAR(rte_crypto_op_pool_create);
895 
896 	/* Check resources are not sufficient */
897 	MOCK_CLEARED_ASSERT(spdk_mempool_create);
898 	rc = vbdev_crypto_init_crypto_drivers();
899 	CU_ASSERT(rc == -EINVAL);
900 
901 	/* Test crypto dev configure failure. */
902 	MOCK_SET(rte_cryptodev_device_count_by_driver, 2);
903 	MOCK_SET(rte_cryptodev_info_get, MOCK_INFO_GET_1QP_AESNI);
904 	MOCK_SET(rte_cryptodev_configure, -1);
905 	MOCK_CLEARED_ASSERT(spdk_mempool_create);
906 	rc = vbdev_crypto_init_crypto_drivers();
907 	MOCK_SET(rte_cryptodev_configure, 0);
908 	CU_ASSERT(g_mbuf_mp == NULL);
909 	CU_ASSERT(g_session_mp == NULL);
910 	CU_ASSERT(g_session_mp_priv == NULL);
911 	CU_ASSERT(rc == -EINVAL);
912 
913 	/* Test failure of qp setup. */
914 	MOCK_SET(rte_cryptodev_queue_pair_setup, -1);
915 	MOCK_CLEARED_ASSERT(spdk_mempool_create);
916 	rc = vbdev_crypto_init_crypto_drivers();
917 	CU_ASSERT(rc == -EINVAL);
918 	CU_ASSERT(g_mbuf_mp == NULL);
919 	CU_ASSERT(g_session_mp == NULL);
920 	CU_ASSERT(g_session_mp_priv == NULL);
921 	MOCK_SET(rte_cryptodev_queue_pair_setup, 0);
922 
923 	/* Test failure of dev start. */
924 	MOCK_SET(rte_cryptodev_start, -1);
925 	MOCK_CLEARED_ASSERT(spdk_mempool_create);
926 	rc = vbdev_crypto_init_crypto_drivers();
927 	CU_ASSERT(rc == -EINVAL);
928 	CU_ASSERT(g_mbuf_mp == NULL);
929 	CU_ASSERT(g_session_mp == NULL);
930 	CU_ASSERT(g_session_mp_priv == NULL);
931 	MOCK_SET(rte_cryptodev_start, 0);
932 
933 	/* Test bogus PMD */
934 	MOCK_CLEARED_ASSERT(spdk_mempool_create);
935 	MOCK_SET(rte_cryptodev_info_get, MOCK_INFO_GET_1QP_BOGUS_PMD);
936 	rc = vbdev_crypto_init_crypto_drivers();
937 	CU_ASSERT(g_mbuf_mp == NULL);
938 	CU_ASSERT(g_session_mp == NULL);
939 	CU_ASSERT(rc == -EINVAL);
940 
941 	/* Test happy path QAT. */
942 	MOCK_CLEARED_ASSERT(spdk_mempool_create);
943 	MOCK_SET(rte_cryptodev_info_get, MOCK_INFO_GET_1QP_QAT);
944 	rc = vbdev_crypto_init_crypto_drivers();
945 	CU_ASSERT(g_mbuf_mp != NULL);
946 	CU_ASSERT(g_session_mp != NULL);
947 	init_cleanup();
948 	CU_ASSERT(rc == 0);
949 
950 	/* Test happy path AESNI. */
951 	MOCK_CLEARED_ASSERT(spdk_mempool_create);
952 	MOCK_SET(rte_cryptodev_info_get, MOCK_INFO_GET_1QP_AESNI);
953 	rc = vbdev_crypto_init_crypto_drivers();
954 	CU_ASSERT(g_mbuf_offset == DPDK_DYNFIELD_OFFSET);
955 	init_cleanup();
956 	CU_ASSERT(rc == 0);
957 
958 	/* Test failure of DPDK dev init. By now it is not longer an error
959 	 * situation for entire crypto framework. */
960 	MOCK_SET(rte_cryptodev_count, 2);
961 	MOCK_SET(rte_cryptodev_device_count_by_driver, 2);
962 	MOCK_SET(rte_vdev_init, -1);
963 	MOCK_CLEARED_ASSERT(spdk_mempool_create);
964 	MOCK_SET(rte_cryptodev_info_get, MOCK_INFO_GET_1QP_QAT);
965 	rc = vbdev_crypto_init_crypto_drivers();
966 	CU_ASSERT(rc == 0);
967 	CU_ASSERT(g_mbuf_mp != NULL);
968 	CU_ASSERT(g_session_mp != NULL);
969 	CU_ASSERT(g_session_mp_priv != NULL);
970 	init_cleanup();
971 	MOCK_SET(rte_vdev_init, 0);
972 	MOCK_CLEAR(rte_cryptodev_device_count_by_driver);
973 
974 	/* restore our initial values. */
975 	g_mbuf_mp = orig_mbuf_mp;
976 	g_session_mp = orig_session_mp;
977 	g_session_mp_priv = orig_session_mp_priv;
978 }
979 
980 static void
981 test_crypto_op_complete(void)
982 {
983 	/* Make sure completion code respects failure. */
984 	g_bdev_io->internal.status = SPDK_BDEV_IO_STATUS_FAILED;
985 	g_completion_called = false;
986 	_crypto_operation_complete(g_bdev_io);
987 	CU_ASSERT(g_bdev_io->internal.status == SPDK_BDEV_IO_STATUS_FAILED);
988 	CU_ASSERT(g_completion_called == true);
989 
990 	/* Test read completion. */
991 	g_bdev_io->internal.status = SPDK_BDEV_IO_STATUS_SUCCESS;
992 	g_bdev_io->type = SPDK_BDEV_IO_TYPE_READ;
993 	g_completion_called = false;
994 	_crypto_operation_complete(g_bdev_io);
995 	CU_ASSERT(g_bdev_io->internal.status == SPDK_BDEV_IO_STATUS_SUCCESS);
996 	CU_ASSERT(g_completion_called == true);
997 
998 	/* Test write completion success. */
999 	g_bdev_io->internal.status = SPDK_BDEV_IO_STATUS_SUCCESS;
1000 	g_bdev_io->type = SPDK_BDEV_IO_TYPE_WRITE;
1001 	g_completion_called = false;
1002 	MOCK_SET(spdk_bdev_writev_blocks, 0);
1003 	_crypto_operation_complete(g_bdev_io);
1004 	CU_ASSERT(g_bdev_io->internal.status == SPDK_BDEV_IO_STATUS_SUCCESS);
1005 	CU_ASSERT(g_completion_called == true);
1006 
1007 	/* Test write completion failed. */
1008 	g_bdev_io->internal.status = SPDK_BDEV_IO_STATUS_SUCCESS;
1009 	g_bdev_io->type = SPDK_BDEV_IO_TYPE_WRITE;
1010 	g_completion_called = false;
1011 	MOCK_SET(spdk_bdev_writev_blocks, -1);
1012 	_crypto_operation_complete(g_bdev_io);
1013 	CU_ASSERT(g_bdev_io->internal.status == SPDK_BDEV_IO_STATUS_FAILED);
1014 	CU_ASSERT(g_completion_called == true);
1015 
1016 	/* Test bogus type for this completion. */
1017 	g_bdev_io->internal.status = SPDK_BDEV_IO_STATUS_SUCCESS;
1018 	g_bdev_io->type = SPDK_BDEV_IO_TYPE_RESET;
1019 	g_completion_called = false;
1020 	_crypto_operation_complete(g_bdev_io);
1021 	CU_ASSERT(g_bdev_io->internal.status == SPDK_BDEV_IO_STATUS_FAILED);
1022 	CU_ASSERT(g_completion_called == true);
1023 }
1024 
1025 static void
1026 test_supported_io(void)
1027 {
1028 	void *ctx = NULL;
1029 	bool rc = true;
1030 
1031 	/* Make sure we always report false to WZ, we need the bdev layer to
1032 	 * send real 0's so we can encrypt/decrypt them.
1033 	 */
1034 	rc = vbdev_crypto_io_type_supported(ctx, SPDK_BDEV_IO_TYPE_WRITE_ZEROES);
1035 	CU_ASSERT(rc == false);
1036 }
1037 
1038 static void
1039 test_poller(void)
1040 {
1041 	int rc;
1042 	struct rte_mbuf *src_mbufs[2];
1043 	struct vbdev_crypto_op *op_to_resubmit;
1044 
1045 	/* test regular 1 op to dequeue and complete */
1046 	g_dequeue_mock = g_enqueue_mock = 1;
1047 	rte_pktmbuf_alloc_bulk(g_mbuf_mp, src_mbufs, 1);
1048 	g_test_crypto_ops[0]->sym->m_src = src_mbufs[0];
1049 	*RTE_MBUF_DYNFIELD(g_test_crypto_ops[0]->sym->m_src, g_mbuf_offset,
1050 			   uint64_t *) = (uintptr_t)g_bdev_io;
1051 	g_test_crypto_ops[0]->sym->m_dst = NULL;
1052 	g_io_ctx->cryop_cnt_remaining = 1;
1053 	g_bdev_io->type = SPDK_BDEV_IO_TYPE_READ;
1054 	rc = crypto_dev_poller(g_crypto_ch);
1055 	CU_ASSERT(rc == 1);
1056 
1057 	/* We have nothing dequeued but have some to resubmit */
1058 	g_dequeue_mock = 0;
1059 	CU_ASSERT(TAILQ_EMPTY(&g_crypto_ch->queued_cry_ops) == true);
1060 
1061 	/* add an op to the queued list. */
1062 	g_resubmit_test = true;
1063 	op_to_resubmit = (struct vbdev_crypto_op *)((uint8_t *)g_test_crypto_ops[0] + QUEUED_OP_OFFSET);
1064 	op_to_resubmit->crypto_op = (void *)0xDEADBEEF;
1065 	op_to_resubmit->bdev_io = g_bdev_io;
1066 	TAILQ_INSERT_TAIL(&g_crypto_ch->queued_cry_ops,
1067 			  op_to_resubmit,
1068 			  link);
1069 	CU_ASSERT(TAILQ_EMPTY(&g_crypto_ch->queued_cry_ops) == false);
1070 	rc = crypto_dev_poller(g_crypto_ch);
1071 	g_resubmit_test = false;
1072 	CU_ASSERT(rc == 0);
1073 	CU_ASSERT(TAILQ_EMPTY(&g_crypto_ch->queued_cry_ops) == true);
1074 
1075 	/* 2 to dequeue but 2nd one failed */
1076 	g_dequeue_mock = g_enqueue_mock = 2;
1077 	g_io_ctx->cryop_cnt_remaining = 2;
1078 	rte_pktmbuf_alloc_bulk(g_mbuf_mp, src_mbufs, 2);
1079 	g_test_crypto_ops[0]->sym->m_src = src_mbufs[0];
1080 	*RTE_MBUF_DYNFIELD(g_test_crypto_ops[0]->sym->m_src, g_mbuf_offset,
1081 			   uint64_t *) = (uint64_t)g_bdev_io;
1082 	g_test_crypto_ops[0]->sym->m_dst = NULL;
1083 	g_test_crypto_ops[0]->status =  RTE_CRYPTO_OP_STATUS_SUCCESS;
1084 	g_test_crypto_ops[1]->sym->m_src = src_mbufs[1];
1085 	*RTE_MBUF_DYNFIELD(g_test_crypto_ops[1]->sym->m_src, g_mbuf_offset,
1086 			   uint64_t *) = (uint64_t)g_bdev_io;
1087 	g_test_crypto_ops[1]->sym->m_dst = NULL;
1088 	g_test_crypto_ops[1]->status = RTE_CRYPTO_OP_STATUS_NOT_PROCESSED;
1089 	g_bdev_io->internal.status = SPDK_BDEV_IO_STATUS_SUCCESS;
1090 	rc = crypto_dev_poller(g_crypto_ch);
1091 	CU_ASSERT(g_bdev_io->internal.status == SPDK_BDEV_IO_STATUS_FAILED);
1092 	CU_ASSERT(rc == 2);
1093 }
1094 
1095 /* Helper function for test_assign_device_qp() */
1096 static void
1097 _clear_device_qp_lists(void)
1098 {
1099 	struct device_qp *device_qp = NULL;
1100 
1101 	while (!TAILQ_EMPTY(&g_device_qp_qat)) {
1102 		device_qp = TAILQ_FIRST(&g_device_qp_qat);
1103 		TAILQ_REMOVE(&g_device_qp_qat, device_qp, link);
1104 		free(device_qp);
1105 
1106 	}
1107 	CU_ASSERT(TAILQ_EMPTY(&g_device_qp_qat) == true);
1108 	while (!TAILQ_EMPTY(&g_device_qp_aesni_mb)) {
1109 		device_qp = TAILQ_FIRST(&g_device_qp_aesni_mb);
1110 		TAILQ_REMOVE(&g_device_qp_aesni_mb, device_qp, link);
1111 		free(device_qp);
1112 	}
1113 	CU_ASSERT(TAILQ_EMPTY(&g_device_qp_aesni_mb) == true);
1114 }
1115 
1116 /* Helper function for test_assign_device_qp() */
1117 static void
1118 _check_expected_values(struct vbdev_crypto *crypto_bdev, struct device_qp *device_qp,
1119 		       struct crypto_io_channel *crypto_ch, uint8_t expected_index,
1120 		       uint8_t current_index)
1121 {
1122 	_assign_device_qp(&g_crypto_bdev, device_qp, g_crypto_ch);
1123 	CU_ASSERT(g_crypto_ch->device_qp->index == expected_index);
1124 	CU_ASSERT(g_next_qat_index == current_index);
1125 }
1126 
1127 static void
1128 test_assign_device_qp(void)
1129 {
1130 	struct device_qp *device_qp = NULL;
1131 	int i;
1132 
1133 	/* start with a known state, clear the device/qp lists */
1134 	_clear_device_qp_lists();
1135 
1136 	/* make sure that one AESNI_MB qp is found */
1137 	device_qp = calloc(1, sizeof(struct device_qp));
1138 	TAILQ_INSERT_TAIL(&g_device_qp_aesni_mb, device_qp, link);
1139 	g_crypto_ch->device_qp = NULL;
1140 	g_crypto_bdev.drv_name = AESNI_MB;
1141 	_assign_device_qp(&g_crypto_bdev, device_qp, g_crypto_ch);
1142 	CU_ASSERT(g_crypto_ch->device_qp != NULL);
1143 
1144 	/* QAT testing is more complex as the code under test load balances by
1145 	 * assigning each subsequent device/qp to every QAT_VF_SPREAD modulo
1146 	 * g_qat_total_qp. For the current latest QAT we'll have 48 virtual functions
1147 	 * each with 2 qp so the "spread" between assignments is 32.
1148 	 */
1149 	g_qat_total_qp = 96;
1150 	for (i = 0; i < g_qat_total_qp; i++) {
1151 		device_qp = calloc(1, sizeof(struct device_qp));
1152 		device_qp->index = i;
1153 		TAILQ_INSERT_TAIL(&g_device_qp_qat, device_qp, link);
1154 	}
1155 	g_crypto_ch->device_qp = NULL;
1156 	g_crypto_bdev.drv_name = QAT;
1157 
1158 	/* First assignment will assign to 0 and next at 32. */
1159 	_check_expected_values(&g_crypto_bdev, device_qp, g_crypto_ch,
1160 			       0, QAT_VF_SPREAD);
1161 
1162 	/* Second assignment will assign to 32 and next at 64. */
1163 	_check_expected_values(&g_crypto_bdev, device_qp, g_crypto_ch,
1164 			       QAT_VF_SPREAD, QAT_VF_SPREAD * 2);
1165 
1166 	/* Third assignment will assign to 64 and next at 0. */
1167 	_check_expected_values(&g_crypto_bdev, device_qp, g_crypto_ch,
1168 			       QAT_VF_SPREAD * 2, 0);
1169 
1170 	/* Fourth assignment will assign to 1 and next at 33. */
1171 	_check_expected_values(&g_crypto_bdev, device_qp, g_crypto_ch,
1172 			       1, QAT_VF_SPREAD + 1);
1173 
1174 	_clear_device_qp_lists();
1175 }
1176 
1177 int
1178 main(int argc, char **argv)
1179 {
1180 	CU_pSuite	suite = NULL;
1181 	unsigned int	num_failures;
1182 
1183 	CU_set_error_action(CUEA_ABORT);
1184 	CU_initialize_registry();
1185 
1186 	suite = CU_add_suite("crypto", test_setup, test_cleanup);
1187 	CU_ADD_TEST(suite, test_error_paths);
1188 	CU_ADD_TEST(suite, test_simple_write);
1189 	CU_ADD_TEST(suite, test_simple_read);
1190 	CU_ADD_TEST(suite, test_large_rw);
1191 	CU_ADD_TEST(suite, test_dev_full);
1192 	CU_ADD_TEST(suite, test_crazy_rw);
1193 	CU_ADD_TEST(suite, test_passthru);
1194 	CU_ADD_TEST(suite, test_initdrivers);
1195 	CU_ADD_TEST(suite, test_crypto_op_complete);
1196 	CU_ADD_TEST(suite, test_supported_io);
1197 	CU_ADD_TEST(suite, test_reset);
1198 	CU_ADD_TEST(suite, test_poller);
1199 	CU_ADD_TEST(suite, test_assign_device_qp);
1200 
1201 	CU_basic_set_mode(CU_BRM_VERBOSE);
1202 	CU_basic_run_tests();
1203 	num_failures = CU_get_number_of_failures();
1204 	CU_cleanup_registry();
1205 	return num_failures;
1206 }
1207