xref: /spdk/module/bdev/compress/vbdev_compress.c (revision 57fd99b91e71a4baa5543e19ff83958dc99d4dac)
1 /*   SPDX-License-Identifier: BSD-3-Clause
2  *   Copyright (C) 2018 Intel Corporation.
3  *   All rights reserved.
4  *   Copyright (c) 2021, 2022 NVIDIA CORPORATION & AFFILIATES. All rights reserved.
5  */
6 
7 #include "vbdev_compress.h"
8 
9 #include "spdk/reduce.h"
10 #include "spdk/stdinc.h"
11 #include "spdk/rpc.h"
12 #include "spdk/env.h"
13 #include "spdk/endian.h"
14 #include "spdk/string.h"
15 #include "spdk/thread.h"
16 #include "spdk/util.h"
17 #include "spdk/bdev_module.h"
18 #include "spdk/likely.h"
19 #include "spdk/log.h"
20 #include "spdk/accel.h"
21 
22 #include "spdk/accel_module.h"
23 
24 
25 #define CHUNK_SIZE (1024 * 16)
26 #define COMP_BDEV_NAME "compress"
27 #define BACKING_IO_SZ (4 * 1024)
28 
29 /* This namespace UUID was generated using uuid_generate() method. */
30 #define BDEV_COMPRESS_NAMESPACE_UUID "c3fad6da-832f-4cc0-9cdc-5c552b225e7b"
31 
32 struct vbdev_comp_delete_ctx {
33 	spdk_delete_compress_complete	cb_fn;
34 	void				*cb_arg;
35 	int				cb_rc;
36 	struct spdk_thread		*orig_thread;
37 };
38 
39 /* List of virtual bdevs and associated info for each. */
40 struct vbdev_compress {
41 	struct spdk_bdev		*base_bdev;	/* the thing we're attaching to */
42 	struct spdk_bdev_desc		*base_desc;	/* its descriptor we get from open */
43 	struct spdk_io_channel		*base_ch;	/* IO channel of base device */
44 	struct spdk_bdev		comp_bdev;	/* the compression virtual bdev */
45 	struct comp_io_channel		*comp_ch;	/* channel associated with this bdev */
46 	struct spdk_io_channel		*accel_channel;	/* to communicate with the accel framework */
47 	struct spdk_thread		*reduce_thread;
48 	pthread_mutex_t			reduce_lock;
49 	uint32_t			ch_count;
50 	TAILQ_HEAD(, spdk_bdev_io)	pending_comp_ios;	/* outstanding operations to a comp library */
51 	struct spdk_poller		*poller;	/* completion poller */
52 	struct spdk_reduce_vol_params	params;		/* params for the reduce volume */
53 	struct spdk_reduce_backing_dev	backing_dev;	/* backing device info for the reduce volume */
54 	struct spdk_reduce_vol		*vol;		/* the reduce volume */
55 	struct vbdev_comp_delete_ctx	*delete_ctx;
56 	bool				orphaned;	/* base bdev claimed but comp_bdev not registered */
57 	int				reduce_errno;
58 	TAILQ_HEAD(, vbdev_comp_op)	queued_comp_ops;
59 	TAILQ_ENTRY(vbdev_compress)	link;
60 	struct spdk_thread		*thread;	/* thread where base device is opened */
61 };
62 static TAILQ_HEAD(, vbdev_compress) g_vbdev_comp = TAILQ_HEAD_INITIALIZER(g_vbdev_comp);
63 
64 /* The comp vbdev channel struct. It is allocated and freed on my behalf by the io channel code.
65  */
66 struct comp_io_channel {
67 	struct spdk_io_channel_iter	*iter;	/* used with for_each_channel in reset */
68 };
69 
70 /* Per I/O context for the compression vbdev. */
71 struct comp_bdev_io {
72 	struct comp_io_channel		*comp_ch;		/* used in completion handling */
73 	struct vbdev_compress		*comp_bdev;		/* vbdev associated with this IO */
74 	struct spdk_bdev_io_wait_entry	bdev_io_wait;		/* for bdev_io_wait */
75 	struct spdk_bdev_io		*orig_io;		/* the original IO */
76 	struct spdk_io_channel		*ch;			/* for resubmission */
77 	int				status;			/* save for completion on orig thread */
78 };
79 
80 static void vbdev_compress_examine(struct spdk_bdev *bdev);
81 static int vbdev_compress_claim(struct vbdev_compress *comp_bdev);
82 static void vbdev_compress_queue_io(struct spdk_bdev_io *bdev_io);
83 struct vbdev_compress *_prepare_for_load_init(struct spdk_bdev_desc *bdev_desc, uint32_t lb_size);
84 static void vbdev_compress_submit_request(struct spdk_io_channel *ch, struct spdk_bdev_io *bdev_io);
85 static void comp_bdev_ch_destroy_cb(void *io_device, void *ctx_buf);
86 static void vbdev_compress_delete_done(void *cb_arg, int bdeverrno);
87 
88 /* for completing rw requests on the orig IO thread. */
89 static void
90 _reduce_rw_blocks_cb(void *arg)
91 {
92 	struct comp_bdev_io *io_ctx = arg;
93 
94 	if (spdk_likely(io_ctx->status == 0)) {
95 		spdk_bdev_io_complete(io_ctx->orig_io, SPDK_BDEV_IO_STATUS_SUCCESS);
96 	} else if (io_ctx->status == -ENOMEM) {
97 		vbdev_compress_queue_io(spdk_bdev_io_from_ctx(io_ctx));
98 	} else {
99 		SPDK_ERRLOG("Failed to execute reduce api. %s\n", spdk_strerror(-io_ctx->status));
100 		spdk_bdev_io_complete(io_ctx->orig_io, SPDK_BDEV_IO_STATUS_FAILED);
101 	}
102 }
103 
104 /* Completion callback for r/w that were issued via reducelib. */
105 static void
106 reduce_rw_blocks_cb(void *arg, int reduce_errno)
107 {
108 	struct spdk_bdev_io *bdev_io = arg;
109 	struct comp_bdev_io *io_ctx = (struct comp_bdev_io *)bdev_io->driver_ctx;
110 	struct spdk_io_channel *ch = spdk_io_channel_from_ctx(io_ctx->comp_ch);
111 	struct spdk_thread *orig_thread;
112 
113 	/* TODO: need to decide which error codes are bdev_io success vs failure;
114 	 * example examine calls reading metadata */
115 
116 	io_ctx->status = reduce_errno;
117 
118 	/* Send this request to the orig IO thread. */
119 	orig_thread = spdk_io_channel_get_thread(ch);
120 
121 	spdk_thread_exec_msg(orig_thread, _reduce_rw_blocks_cb, io_ctx);
122 }
123 
124 static int
125 _compress_operation(struct spdk_reduce_backing_dev *backing_dev, struct iovec *src_iovs,
126 		    int src_iovcnt, struct iovec *dst_iovs,
127 		    int dst_iovcnt, bool compress, void *cb_arg)
128 {
129 	struct spdk_reduce_vol_cb_args *reduce_cb_arg = cb_arg;
130 	struct vbdev_compress *comp_bdev = SPDK_CONTAINEROF(backing_dev, struct vbdev_compress,
131 					   backing_dev);
132 	int rc;
133 
134 	if (compress) {
135 		assert(dst_iovcnt == 1);
136 		rc = spdk_accel_submit_compress(comp_bdev->accel_channel, dst_iovs[0].iov_base, dst_iovs[0].iov_len,
137 						src_iovs, src_iovcnt, &reduce_cb_arg->output_size,
138 						reduce_cb_arg->cb_fn, reduce_cb_arg->cb_arg);
139 	} else {
140 		rc = spdk_accel_submit_decompress(comp_bdev->accel_channel, dst_iovs, dst_iovcnt,
141 						  src_iovs, src_iovcnt, &reduce_cb_arg->output_size,
142 						  reduce_cb_arg->cb_fn, reduce_cb_arg->cb_arg);
143 	}
144 
145 	return rc;
146 }
147 
148 /* Entry point for reduce lib to issue a compress operation. */
149 static void
150 _comp_reduce_compress(struct spdk_reduce_backing_dev *dev,
151 		      struct iovec *src_iovs, int src_iovcnt,
152 		      struct iovec *dst_iovs, int dst_iovcnt,
153 		      struct spdk_reduce_vol_cb_args *cb_arg)
154 {
155 	int rc;
156 
157 	rc = _compress_operation(dev, src_iovs, src_iovcnt, dst_iovs, dst_iovcnt, true, cb_arg);
158 	if (rc) {
159 		SPDK_ERRLOG("with compress operation code %d (%s)\n", rc, spdk_strerror(-rc));
160 		cb_arg->cb_fn(cb_arg->cb_arg, rc);
161 	}
162 }
163 
164 /* Entry point for reduce lib to issue a decompress operation. */
165 static void
166 _comp_reduce_decompress(struct spdk_reduce_backing_dev *dev,
167 			struct iovec *src_iovs, int src_iovcnt,
168 			struct iovec *dst_iovs, int dst_iovcnt,
169 			struct spdk_reduce_vol_cb_args *cb_arg)
170 {
171 	int rc;
172 
173 	rc = _compress_operation(dev, src_iovs, src_iovcnt, dst_iovs, dst_iovcnt, false, cb_arg);
174 	if (rc) {
175 		SPDK_ERRLOG("with decompress operation code %d (%s)\n", rc, spdk_strerror(-rc));
176 		cb_arg->cb_fn(cb_arg->cb_arg, rc);
177 	}
178 }
179 
180 static void
181 _comp_submit_write(void *ctx)
182 {
183 	struct spdk_bdev_io *bdev_io = ctx;
184 	struct vbdev_compress *comp_bdev = SPDK_CONTAINEROF(bdev_io->bdev, struct vbdev_compress,
185 					   comp_bdev);
186 
187 	spdk_reduce_vol_writev(comp_bdev->vol, bdev_io->u.bdev.iovs, bdev_io->u.bdev.iovcnt,
188 			       bdev_io->u.bdev.offset_blocks, bdev_io->u.bdev.num_blocks,
189 			       reduce_rw_blocks_cb, bdev_io);
190 }
191 
192 static void
193 _comp_submit_read(void *ctx)
194 {
195 	struct spdk_bdev_io *bdev_io = ctx;
196 	struct vbdev_compress *comp_bdev = SPDK_CONTAINEROF(bdev_io->bdev, struct vbdev_compress,
197 					   comp_bdev);
198 
199 	spdk_reduce_vol_readv(comp_bdev->vol, bdev_io->u.bdev.iovs, bdev_io->u.bdev.iovcnt,
200 			      bdev_io->u.bdev.offset_blocks, bdev_io->u.bdev.num_blocks,
201 			      reduce_rw_blocks_cb, bdev_io);
202 }
203 
204 
205 /* Callback for getting a buf from the bdev pool in the event that the caller passed
206  * in NULL, we need to own the buffer so it doesn't get freed by another vbdev module
207  * beneath us before we're done with it.
208  */
209 static void
210 comp_read_get_buf_cb(struct spdk_io_channel *ch, struct spdk_bdev_io *bdev_io, bool success)
211 {
212 	struct vbdev_compress *comp_bdev = SPDK_CONTAINEROF(bdev_io->bdev, struct vbdev_compress,
213 					   comp_bdev);
214 
215 	if (spdk_unlikely(!success)) {
216 		SPDK_ERRLOG("Failed to get data buffer\n");
217 		reduce_rw_blocks_cb(bdev_io, -ENOMEM);
218 		return;
219 	}
220 
221 	spdk_thread_exec_msg(comp_bdev->reduce_thread, _comp_submit_read, bdev_io);
222 }
223 
224 /* Called when someone above submits IO to this vbdev. */
225 static void
226 vbdev_compress_submit_request(struct spdk_io_channel *ch, struct spdk_bdev_io *bdev_io)
227 {
228 	struct comp_bdev_io *io_ctx = (struct comp_bdev_io *)bdev_io->driver_ctx;
229 	struct vbdev_compress *comp_bdev = SPDK_CONTAINEROF(bdev_io->bdev, struct vbdev_compress,
230 					   comp_bdev);
231 	struct comp_io_channel *comp_ch = spdk_io_channel_get_ctx(ch);
232 
233 	memset(io_ctx, 0, sizeof(struct comp_bdev_io));
234 	io_ctx->comp_bdev = comp_bdev;
235 	io_ctx->comp_ch = comp_ch;
236 	io_ctx->orig_io = bdev_io;
237 
238 	switch (bdev_io->type) {
239 	case SPDK_BDEV_IO_TYPE_READ:
240 		spdk_bdev_io_get_buf(bdev_io, comp_read_get_buf_cb,
241 				     bdev_io->u.bdev.num_blocks * bdev_io->bdev->blocklen);
242 		return;
243 	case SPDK_BDEV_IO_TYPE_WRITE:
244 		spdk_thread_exec_msg(comp_bdev->reduce_thread, _comp_submit_write, bdev_io);
245 		return;
246 	/* TODO support RESET in future patch in the series */
247 	case SPDK_BDEV_IO_TYPE_RESET:
248 	case SPDK_BDEV_IO_TYPE_WRITE_ZEROES:
249 	case SPDK_BDEV_IO_TYPE_UNMAP:
250 	case SPDK_BDEV_IO_TYPE_FLUSH:
251 	default:
252 		SPDK_ERRLOG("Unknown I/O type %d\n", bdev_io->type);
253 		spdk_bdev_io_complete(io_ctx->orig_io, SPDK_BDEV_IO_STATUS_FAILED);
254 		break;
255 	}
256 }
257 
258 static bool
259 vbdev_compress_io_type_supported(void *ctx, enum spdk_bdev_io_type io_type)
260 {
261 	struct vbdev_compress *comp_bdev = (struct vbdev_compress *)ctx;
262 
263 	switch (io_type) {
264 	case SPDK_BDEV_IO_TYPE_READ:
265 	case SPDK_BDEV_IO_TYPE_WRITE:
266 		return spdk_bdev_io_type_supported(comp_bdev->base_bdev, io_type);
267 	case SPDK_BDEV_IO_TYPE_UNMAP:
268 	case SPDK_BDEV_IO_TYPE_RESET:
269 	case SPDK_BDEV_IO_TYPE_FLUSH:
270 	case SPDK_BDEV_IO_TYPE_WRITE_ZEROES:
271 	default:
272 		return false;
273 	}
274 }
275 
276 /* Resubmission function used by the bdev layer when a queued IO is ready to be
277  * submitted.
278  */
279 static void
280 vbdev_compress_resubmit_io(void *arg)
281 {
282 	struct spdk_bdev_io *bdev_io = (struct spdk_bdev_io *)arg;
283 	struct comp_bdev_io *io_ctx = (struct comp_bdev_io *)bdev_io->driver_ctx;
284 
285 	vbdev_compress_submit_request(io_ctx->ch, bdev_io);
286 }
287 
288 /* Used to queue an IO in the event of resource issues. */
289 static void
290 vbdev_compress_queue_io(struct spdk_bdev_io *bdev_io)
291 {
292 	struct comp_bdev_io *io_ctx = (struct comp_bdev_io *)bdev_io->driver_ctx;
293 	int rc;
294 
295 	io_ctx->bdev_io_wait.bdev = bdev_io->bdev;
296 	io_ctx->bdev_io_wait.cb_fn = vbdev_compress_resubmit_io;
297 	io_ctx->bdev_io_wait.cb_arg = bdev_io;
298 
299 	rc = spdk_bdev_queue_io_wait(bdev_io->bdev, io_ctx->comp_bdev->base_ch, &io_ctx->bdev_io_wait);
300 	if (rc) {
301 		SPDK_ERRLOG("Queue io failed in vbdev_compress_queue_io, rc=%d.\n", rc);
302 		assert(false);
303 		spdk_bdev_io_complete(bdev_io, SPDK_BDEV_IO_STATUS_FAILED);
304 	}
305 }
306 
307 /* Callback for unregistering the IO device. */
308 static void
309 _device_unregister_cb(void *io_device)
310 {
311 	struct vbdev_compress *comp_bdev = io_device;
312 
313 	/* Done with this comp_bdev. */
314 	pthread_mutex_destroy(&comp_bdev->reduce_lock);
315 	free(comp_bdev->comp_bdev.name);
316 	free(comp_bdev);
317 }
318 
319 static void
320 _vbdev_compress_destruct_cb(void *ctx)
321 {
322 	struct vbdev_compress *comp_bdev = ctx;
323 
324 	TAILQ_REMOVE(&g_vbdev_comp, comp_bdev, link);
325 	spdk_bdev_module_release_bdev(comp_bdev->base_bdev);
326 	/* Close the underlying bdev on its same opened thread. */
327 	spdk_bdev_close(comp_bdev->base_desc);
328 	comp_bdev->vol = NULL;
329 	if (comp_bdev->orphaned == false) {
330 		spdk_io_device_unregister(comp_bdev, _device_unregister_cb);
331 	} else {
332 		vbdev_compress_delete_done(comp_bdev->delete_ctx, 0);
333 		_device_unregister_cb(comp_bdev);
334 	}
335 }
336 
337 static void
338 vbdev_compress_destruct_cb(void *cb_arg, int reduce_errno)
339 {
340 	struct vbdev_compress *comp_bdev = (struct vbdev_compress *)cb_arg;
341 
342 	if (reduce_errno) {
343 		SPDK_ERRLOG("number %d\n", reduce_errno);
344 	} else {
345 		if (comp_bdev->thread && comp_bdev->thread != spdk_get_thread()) {
346 			spdk_thread_send_msg(comp_bdev->thread,
347 					     _vbdev_compress_destruct_cb, comp_bdev);
348 		} else {
349 			_vbdev_compress_destruct_cb(comp_bdev);
350 		}
351 	}
352 }
353 
354 static void
355 _reduce_destroy_cb(void *ctx, int reduce_errno)
356 {
357 	struct vbdev_compress *comp_bdev = (struct vbdev_compress *)ctx;
358 
359 	if (reduce_errno) {
360 		SPDK_ERRLOG("number %d\n", reduce_errno);
361 	}
362 
363 	comp_bdev->vol = NULL;
364 	spdk_put_io_channel(comp_bdev->base_ch);
365 	if (comp_bdev->orphaned == false) {
366 		spdk_bdev_unregister(&comp_bdev->comp_bdev, vbdev_compress_delete_done,
367 				     comp_bdev->delete_ctx);
368 	} else {
369 		vbdev_compress_destruct_cb((void *)comp_bdev, 0);
370 	}
371 
372 }
373 
374 static void
375 _delete_vol_unload_cb(void *ctx)
376 {
377 	struct vbdev_compress *comp_bdev = ctx;
378 
379 	/* FIXME: Assert if these conditions are not satisfied for now. */
380 	assert(!comp_bdev->reduce_thread ||
381 	       comp_bdev->reduce_thread == spdk_get_thread());
382 
383 	/* reducelib needs a channel to comm with the backing device */
384 	comp_bdev->base_ch = spdk_bdev_get_io_channel(comp_bdev->base_desc);
385 
386 	/* Clean the device before we free our resources. */
387 	spdk_reduce_vol_destroy(&comp_bdev->backing_dev, _reduce_destroy_cb, comp_bdev);
388 }
389 
390 /* Called by reduceLib after performing unload vol actions */
391 static void
392 delete_vol_unload_cb(void *cb_arg, int reduce_errno)
393 {
394 	struct vbdev_compress *comp_bdev = (struct vbdev_compress *)cb_arg;
395 
396 	if (reduce_errno) {
397 		SPDK_ERRLOG("number %d\n", reduce_errno);
398 		/* FIXME: callback should be executed. */
399 		return;
400 	}
401 
402 	pthread_mutex_lock(&comp_bdev->reduce_lock);
403 	if (comp_bdev->reduce_thread && comp_bdev->reduce_thread != spdk_get_thread()) {
404 		spdk_thread_send_msg(comp_bdev->reduce_thread,
405 				     _delete_vol_unload_cb, comp_bdev);
406 		pthread_mutex_unlock(&comp_bdev->reduce_lock);
407 	} else {
408 		pthread_mutex_unlock(&comp_bdev->reduce_lock);
409 
410 		_delete_vol_unload_cb(comp_bdev);
411 	}
412 }
413 
414 const char *
415 compress_get_name(const struct vbdev_compress *comp_bdev)
416 {
417 	return comp_bdev->comp_bdev.name;
418 }
419 
420 struct vbdev_compress *
421 compress_bdev_first(void)
422 {
423 	struct vbdev_compress *comp_bdev;
424 
425 	comp_bdev = TAILQ_FIRST(&g_vbdev_comp);
426 
427 	return comp_bdev;
428 }
429 
430 struct vbdev_compress *
431 compress_bdev_next(struct vbdev_compress *prev)
432 {
433 	struct vbdev_compress *comp_bdev;
434 
435 	comp_bdev = TAILQ_NEXT(prev, link);
436 
437 	return comp_bdev;
438 }
439 
440 bool
441 compress_has_orphan(const char *name)
442 {
443 	struct vbdev_compress *comp_bdev;
444 
445 	TAILQ_FOREACH(comp_bdev, &g_vbdev_comp, link) {
446 		if (comp_bdev->orphaned && strcmp(name, comp_bdev->comp_bdev.name) == 0) {
447 			return true;
448 		}
449 	}
450 	return false;
451 }
452 
453 /* Called after we've unregistered following a hot remove callback.
454  * Our finish entry point will be called next.
455  */
456 static int
457 vbdev_compress_destruct(void *ctx)
458 {
459 	struct vbdev_compress *comp_bdev = (struct vbdev_compress *)ctx;
460 
461 	if (comp_bdev->vol != NULL) {
462 		/* Tell reducelib that we're done with this volume. */
463 		spdk_reduce_vol_unload(comp_bdev->vol, vbdev_compress_destruct_cb, comp_bdev);
464 	} else {
465 		vbdev_compress_destruct_cb(comp_bdev, 0);
466 	}
467 
468 	return 0;
469 }
470 
471 /* We supplied this as an entry point for upper layers who want to communicate to this
472  * bdev.  This is how they get a channel.
473  */
474 static struct spdk_io_channel *
475 vbdev_compress_get_io_channel(void *ctx)
476 {
477 	struct vbdev_compress *comp_bdev = (struct vbdev_compress *)ctx;
478 
479 	/* The IO channel code will allocate a channel for us which consists of
480 	 * the SPDK channel structure plus the size of our comp_io_channel struct
481 	 * that we passed in when we registered our IO device. It will then call
482 	 * our channel create callback to populate any elements that we need to
483 	 * update.
484 	 */
485 	return spdk_get_io_channel(comp_bdev);
486 }
487 
488 /* This is the output for bdev_get_bdevs() for this vbdev */
489 static int
490 vbdev_compress_dump_info_json(void *ctx, struct spdk_json_write_ctx *w)
491 {
492 	struct vbdev_compress *comp_bdev = (struct vbdev_compress *)ctx;
493 
494 	spdk_json_write_name(w, "compress");
495 	spdk_json_write_object_begin(w);
496 	spdk_json_write_named_string(w, "name", spdk_bdev_get_name(&comp_bdev->comp_bdev));
497 	spdk_json_write_named_string(w, "base_bdev_name", spdk_bdev_get_name(comp_bdev->base_bdev));
498 	spdk_json_write_named_string(w, "pm_path", spdk_reduce_vol_get_pm_path(comp_bdev->vol));
499 	spdk_json_write_object_end(w);
500 
501 	return 0;
502 }
503 
504 static int
505 vbdev_compress_config_json(struct spdk_json_write_ctx *w)
506 {
507 	/* Nothing to dump as compress bdev configuration is saved on physical device. */
508 	return 0;
509 }
510 
511 static void
512 _vbdev_reduce_init_cb(void *ctx)
513 {
514 	struct vbdev_compress *comp_bdev = ctx;
515 	int rc;
516 
517 	assert(comp_bdev->base_desc != NULL);
518 
519 	/* We're done with metadata operations */
520 	spdk_put_io_channel(comp_bdev->base_ch);
521 
522 	if (comp_bdev->vol) {
523 		rc = vbdev_compress_claim(comp_bdev);
524 		if (rc == 0) {
525 			return;
526 		}
527 	}
528 
529 	/* Close the underlying bdev on its same opened thread. */
530 	spdk_bdev_close(comp_bdev->base_desc);
531 	free(comp_bdev);
532 }
533 
534 /* Callback from reduce for when init is complete. We'll pass the vbdev_comp struct
535  * used for initial metadata operations to claim where it will be further filled out
536  * and added to the global list.
537  */
538 static void
539 vbdev_reduce_init_cb(void *cb_arg, struct spdk_reduce_vol *vol, int reduce_errno)
540 {
541 	struct vbdev_compress *comp_bdev = cb_arg;
542 
543 	if (reduce_errno == 0) {
544 		comp_bdev->vol = vol;
545 	} else {
546 		SPDK_ERRLOG("for vol %s, error %u\n",
547 			    spdk_bdev_get_name(comp_bdev->base_bdev), reduce_errno);
548 	}
549 
550 	if (comp_bdev->thread && comp_bdev->thread != spdk_get_thread()) {
551 		spdk_thread_send_msg(comp_bdev->thread, _vbdev_reduce_init_cb, comp_bdev);
552 	} else {
553 		_vbdev_reduce_init_cb(comp_bdev);
554 	}
555 }
556 
557 /* Callback for the function used by reduceLib to perform IO to/from the backing device. We just
558  * call the callback provided by reduceLib when it called the read/write/unmap function and
559  * free the bdev_io.
560  */
561 static void
562 comp_reduce_io_cb(struct spdk_bdev_io *bdev_io, bool success, void *arg)
563 {
564 	struct spdk_reduce_vol_cb_args *cb_args = arg;
565 	int reduce_errno;
566 
567 	if (success) {
568 		reduce_errno = 0;
569 	} else {
570 		reduce_errno = -EIO;
571 	}
572 	spdk_bdev_free_io(bdev_io);
573 	cb_args->cb_fn(cb_args->cb_arg, reduce_errno);
574 }
575 
576 /* This is the function provided to the reduceLib for sending reads directly to
577  * the backing device.
578  */
579 static void
580 _comp_reduce_readv(struct spdk_reduce_backing_dev *dev, struct iovec *iov, int iovcnt,
581 		   uint64_t lba, uint32_t lba_count, struct spdk_reduce_vol_cb_args *args)
582 {
583 	struct vbdev_compress *comp_bdev = SPDK_CONTAINEROF(dev, struct vbdev_compress,
584 					   backing_dev);
585 	int rc;
586 
587 	rc = spdk_bdev_readv_blocks(comp_bdev->base_desc, comp_bdev->base_ch,
588 				    iov, iovcnt, lba, lba_count,
589 				    comp_reduce_io_cb,
590 				    args);
591 	if (rc) {
592 		if (rc == -ENOMEM) {
593 			SPDK_ERRLOG("No memory, start to queue io.\n");
594 			/* TODO: there's no bdev_io to queue */
595 		} else {
596 			SPDK_ERRLOG("submitting readv request\n");
597 		}
598 		args->cb_fn(args->cb_arg, rc);
599 	}
600 }
601 
602 /* This is the function provided to the reduceLib for sending writes directly to
603  * the backing device.
604  */
605 static void
606 _comp_reduce_writev(struct spdk_reduce_backing_dev *dev, struct iovec *iov, int iovcnt,
607 		    uint64_t lba, uint32_t lba_count, struct spdk_reduce_vol_cb_args *args)
608 {
609 	struct vbdev_compress *comp_bdev = SPDK_CONTAINEROF(dev, struct vbdev_compress,
610 					   backing_dev);
611 	int rc;
612 
613 	rc = spdk_bdev_writev_blocks(comp_bdev->base_desc, comp_bdev->base_ch,
614 				     iov, iovcnt, lba, lba_count,
615 				     comp_reduce_io_cb,
616 				     args);
617 	if (rc) {
618 		if (rc == -ENOMEM) {
619 			SPDK_ERRLOG("No memory, start to queue io.\n");
620 			/* TODO: there's no bdev_io to queue */
621 		} else {
622 			SPDK_ERRLOG("error submitting writev request\n");
623 		}
624 		args->cb_fn(args->cb_arg, rc);
625 	}
626 }
627 
628 /* This is the function provided to the reduceLib for sending unmaps directly to
629  * the backing device.
630  */
631 static void
632 _comp_reduce_unmap(struct spdk_reduce_backing_dev *dev,
633 		   uint64_t lba, uint32_t lba_count, struct spdk_reduce_vol_cb_args *args)
634 {
635 	struct vbdev_compress *comp_bdev = SPDK_CONTAINEROF(dev, struct vbdev_compress,
636 					   backing_dev);
637 	int rc;
638 
639 	rc = spdk_bdev_unmap_blocks(comp_bdev->base_desc, comp_bdev->base_ch,
640 				    lba, lba_count,
641 				    comp_reduce_io_cb,
642 				    args);
643 
644 	if (rc) {
645 		if (rc == -ENOMEM) {
646 			SPDK_ERRLOG("No memory, start to queue io.\n");
647 			/* TODO: there's no bdev_io to queue */
648 		} else {
649 			SPDK_ERRLOG("submitting unmap request\n");
650 		}
651 		args->cb_fn(args->cb_arg, rc);
652 	}
653 }
654 
655 /* Called by reduceLib after performing unload vol actions following base bdev hotremove */
656 static void
657 bdev_hotremove_vol_unload_cb(void *cb_arg, int reduce_errno)
658 {
659 	struct vbdev_compress *comp_bdev = (struct vbdev_compress *)cb_arg;
660 
661 	if (reduce_errno) {
662 		SPDK_ERRLOG("number %d\n", reduce_errno);
663 	}
664 
665 	comp_bdev->vol = NULL;
666 	spdk_bdev_unregister(&comp_bdev->comp_bdev, NULL, NULL);
667 }
668 
669 static void
670 vbdev_compress_base_bdev_hotremove_cb(struct spdk_bdev *bdev_find)
671 {
672 	struct vbdev_compress *comp_bdev, *tmp;
673 
674 	TAILQ_FOREACH_SAFE(comp_bdev, &g_vbdev_comp, link, tmp) {
675 		if (bdev_find == comp_bdev->base_bdev) {
676 			/* Tell reduceLib that we're done with this volume. */
677 			spdk_reduce_vol_unload(comp_bdev->vol, bdev_hotremove_vol_unload_cb, comp_bdev);
678 		}
679 	}
680 }
681 
682 /* Called when the underlying base bdev triggers asynchronous event such as bdev removal. */
683 static void
684 vbdev_compress_base_bdev_event_cb(enum spdk_bdev_event_type type, struct spdk_bdev *bdev,
685 				  void *event_ctx)
686 {
687 	switch (type) {
688 	case SPDK_BDEV_EVENT_REMOVE:
689 		vbdev_compress_base_bdev_hotremove_cb(bdev);
690 		break;
691 	default:
692 		SPDK_NOTICELOG("Unsupported bdev event: type %d\n", type);
693 		break;
694 	}
695 }
696 
697 /* TODO: determine which parms we want user configurable, HC for now
698  * params.vol_size
699  * params.chunk_size
700  * compression PMD, algorithm, window size, comp level, etc.
701  * DEV_MD_PATH
702  */
703 
704 /* Common function for init and load to allocate and populate the minimal
705  * information for reducelib to init or load.
706  */
707 struct vbdev_compress *
708 _prepare_for_load_init(struct spdk_bdev_desc *bdev_desc, uint32_t lb_size)
709 {
710 	struct vbdev_compress *comp_bdev;
711 	struct spdk_bdev *bdev;
712 
713 	comp_bdev = calloc(1, sizeof(struct vbdev_compress));
714 	if (comp_bdev == NULL) {
715 		SPDK_ERRLOG("failed to alloc comp_bdev\n");
716 		return NULL;
717 	}
718 
719 	comp_bdev->backing_dev.unmap = _comp_reduce_unmap;
720 	comp_bdev->backing_dev.readv = _comp_reduce_readv;
721 	comp_bdev->backing_dev.writev = _comp_reduce_writev;
722 	comp_bdev->backing_dev.compress = _comp_reduce_compress;
723 	comp_bdev->backing_dev.decompress = _comp_reduce_decompress;
724 
725 	comp_bdev->base_desc = bdev_desc;
726 	bdev = spdk_bdev_desc_get_bdev(bdev_desc);
727 	comp_bdev->base_bdev = bdev;
728 
729 	comp_bdev->backing_dev.blocklen = bdev->blocklen;
730 	comp_bdev->backing_dev.blockcnt = bdev->blockcnt;
731 
732 	comp_bdev->params.chunk_size = CHUNK_SIZE;
733 	if (lb_size == 0) {
734 		comp_bdev->params.logical_block_size = bdev->blocklen;
735 	} else {
736 		comp_bdev->params.logical_block_size = lb_size;
737 	}
738 
739 	comp_bdev->params.backing_io_unit_size = BACKING_IO_SZ;
740 	return comp_bdev;
741 }
742 
743 /* Call reducelib to initialize a new volume */
744 static int
745 vbdev_init_reduce(const char *bdev_name, const char *pm_path, uint32_t lb_size)
746 {
747 	struct spdk_bdev_desc *bdev_desc = NULL;
748 	struct vbdev_compress *comp_bdev;
749 	int rc;
750 
751 	rc = spdk_bdev_open_ext(bdev_name, true, vbdev_compress_base_bdev_event_cb,
752 				NULL, &bdev_desc);
753 	if (rc) {
754 		SPDK_ERRLOG("could not open bdev %s, error %s\n", bdev_name, spdk_strerror(-rc));
755 		return rc;
756 	}
757 
758 	comp_bdev = _prepare_for_load_init(bdev_desc, lb_size);
759 	if (comp_bdev == NULL) {
760 		spdk_bdev_close(bdev_desc);
761 		return -EINVAL;
762 	}
763 
764 	/* Save the thread where the base device is opened */
765 	comp_bdev->thread = spdk_get_thread();
766 
767 	comp_bdev->base_ch = spdk_bdev_get_io_channel(comp_bdev->base_desc);
768 
769 	spdk_reduce_vol_init(&comp_bdev->params, &comp_bdev->backing_dev,
770 			     pm_path,
771 			     vbdev_reduce_init_cb,
772 			     comp_bdev);
773 	return 0;
774 }
775 
776 /* We provide this callback for the SPDK channel code to create a channel using
777  * the channel struct we provided in our module get_io_channel() entry point. Here
778  * we get and save off an underlying base channel of the device below us so that
779  * we can communicate with the base bdev on a per channel basis.  If we needed
780  * our own poller for this vbdev, we'd register it here.
781  */
782 static int
783 comp_bdev_ch_create_cb(void *io_device, void *ctx_buf)
784 {
785 	struct vbdev_compress *comp_bdev = io_device;
786 
787 	/* Now set the reduce channel if it's not already set. */
788 	pthread_mutex_lock(&comp_bdev->reduce_lock);
789 	if (comp_bdev->ch_count == 0) {
790 		/* We use this queue to track outstanding IO in our layer. */
791 		TAILQ_INIT(&comp_bdev->pending_comp_ios);
792 
793 		/* We use this to queue up compression operations as needed. */
794 		TAILQ_INIT(&comp_bdev->queued_comp_ops);
795 
796 		comp_bdev->base_ch = spdk_bdev_get_io_channel(comp_bdev->base_desc);
797 		comp_bdev->reduce_thread = spdk_get_thread();
798 		comp_bdev->accel_channel = spdk_accel_get_io_channel();
799 	}
800 	comp_bdev->ch_count++;
801 	pthread_mutex_unlock(&comp_bdev->reduce_lock);
802 
803 	return 0;
804 }
805 
806 static void
807 _channel_cleanup(struct vbdev_compress *comp_bdev)
808 {
809 	spdk_put_io_channel(comp_bdev->base_ch);
810 	spdk_put_io_channel(comp_bdev->accel_channel);
811 	comp_bdev->reduce_thread = NULL;
812 }
813 
814 /* Used to reroute destroy_ch to the correct thread */
815 static void
816 _comp_bdev_ch_destroy_cb(void *arg)
817 {
818 	struct vbdev_compress *comp_bdev = arg;
819 
820 	pthread_mutex_lock(&comp_bdev->reduce_lock);
821 	_channel_cleanup(comp_bdev);
822 	pthread_mutex_unlock(&comp_bdev->reduce_lock);
823 }
824 
825 /* We provide this callback for the SPDK channel code to destroy a channel
826  * created with our create callback. We just need to undo anything we did
827  * when we created. If this bdev used its own poller, we'd unregister it here.
828  */
829 static void
830 comp_bdev_ch_destroy_cb(void *io_device, void *ctx_buf)
831 {
832 	struct vbdev_compress *comp_bdev = io_device;
833 
834 	pthread_mutex_lock(&comp_bdev->reduce_lock);
835 	comp_bdev->ch_count--;
836 	if (comp_bdev->ch_count == 0) {
837 		/* Send this request to the thread where the channel was created. */
838 		if (comp_bdev->reduce_thread != spdk_get_thread()) {
839 			spdk_thread_send_msg(comp_bdev->reduce_thread,
840 					     _comp_bdev_ch_destroy_cb, comp_bdev);
841 		} else {
842 			_channel_cleanup(comp_bdev);
843 		}
844 	}
845 	pthread_mutex_unlock(&comp_bdev->reduce_lock);
846 }
847 
848 /* RPC entry point for compression vbdev creation. */
849 int
850 create_compress_bdev(const char *bdev_name, const char *pm_path, uint32_t lb_size)
851 {
852 	struct vbdev_compress *comp_bdev = NULL;
853 	struct stat info;
854 
855 	if (stat(pm_path, &info) != 0) {
856 		SPDK_ERRLOG("PM path %s does not exist.\n", pm_path);
857 		return -EINVAL;
858 	} else if (!S_ISDIR(info.st_mode)) {
859 		SPDK_ERRLOG("PM path %s is not a directory.\n", pm_path);
860 		return -EINVAL;
861 	}
862 
863 	if ((lb_size != 0) && (lb_size != LB_SIZE_4K) && (lb_size != LB_SIZE_512B)) {
864 		SPDK_ERRLOG("Logical block size must be 512 or 4096\n");
865 		return -EINVAL;
866 	}
867 
868 	TAILQ_FOREACH(comp_bdev, &g_vbdev_comp, link) {
869 		if (strcmp(bdev_name, comp_bdev->base_bdev->name) == 0) {
870 			SPDK_ERRLOG("Bass bdev %s already being used for a compress bdev\n", bdev_name);
871 			return -EBUSY;
872 		}
873 	}
874 	return vbdev_init_reduce(bdev_name, pm_path, lb_size);
875 }
876 
877 static int
878 vbdev_compress_init(void)
879 {
880 	return 0;
881 }
882 
883 /* Called when the entire module is being torn down. */
884 static void
885 vbdev_compress_finish(void)
886 {
887 	/* TODO: unload vol in a future patch */
888 }
889 
890 /* During init we'll be asked how much memory we'd like passed to us
891  * in bev_io structures as context. Here's where we specify how
892  * much context we want per IO.
893  */
894 static int
895 vbdev_compress_get_ctx_size(void)
896 {
897 	return sizeof(struct comp_bdev_io);
898 }
899 
900 /* When we register our bdev this is how we specify our entry points. */
901 static const struct spdk_bdev_fn_table vbdev_compress_fn_table = {
902 	.destruct		= vbdev_compress_destruct,
903 	.submit_request		= vbdev_compress_submit_request,
904 	.io_type_supported	= vbdev_compress_io_type_supported,
905 	.get_io_channel		= vbdev_compress_get_io_channel,
906 	.dump_info_json		= vbdev_compress_dump_info_json,
907 	.write_config_json	= NULL,
908 };
909 
910 static struct spdk_bdev_module compress_if = {
911 	.name = "compress",
912 	.module_init = vbdev_compress_init,
913 	.get_ctx_size = vbdev_compress_get_ctx_size,
914 	.examine_disk = vbdev_compress_examine,
915 	.module_fini = vbdev_compress_finish,
916 	.config_json = vbdev_compress_config_json
917 };
918 
919 SPDK_BDEV_MODULE_REGISTER(compress, &compress_if)
920 
921 static int _set_compbdev_name(struct vbdev_compress *comp_bdev)
922 {
923 	struct spdk_bdev_alias *aliases;
924 
925 	if (!TAILQ_EMPTY(spdk_bdev_get_aliases(comp_bdev->base_bdev))) {
926 		aliases = TAILQ_FIRST(spdk_bdev_get_aliases(comp_bdev->base_bdev));
927 		comp_bdev->comp_bdev.name = spdk_sprintf_alloc("COMP_%s", aliases->alias.name);
928 		if (!comp_bdev->comp_bdev.name) {
929 			SPDK_ERRLOG("could not allocate comp_bdev name for alias\n");
930 			return -ENOMEM;
931 		}
932 	} else {
933 		comp_bdev->comp_bdev.name = spdk_sprintf_alloc("COMP_%s", comp_bdev->base_bdev->name);
934 		if (!comp_bdev->comp_bdev.name) {
935 			SPDK_ERRLOG("could not allocate comp_bdev name for unique name\n");
936 			return -ENOMEM;
937 		}
938 	}
939 	return 0;
940 }
941 
942 static int
943 vbdev_compress_claim(struct vbdev_compress *comp_bdev)
944 {
945 	struct spdk_uuid ns_uuid;
946 	int rc;
947 
948 	if (_set_compbdev_name(comp_bdev)) {
949 		return -EINVAL;
950 	}
951 
952 	/* Note: some of the fields below will change in the future - for example,
953 	 * blockcnt specifically will not match (the compressed volume size will
954 	 * be slightly less than the base bdev size)
955 	 */
956 	comp_bdev->comp_bdev.product_name = COMP_BDEV_NAME;
957 	comp_bdev->comp_bdev.write_cache = comp_bdev->base_bdev->write_cache;
958 
959 	comp_bdev->comp_bdev.optimal_io_boundary =
960 		comp_bdev->params.chunk_size / comp_bdev->params.logical_block_size;
961 
962 	comp_bdev->comp_bdev.split_on_optimal_io_boundary = true;
963 
964 	comp_bdev->comp_bdev.blocklen = comp_bdev->params.logical_block_size;
965 	comp_bdev->comp_bdev.blockcnt = comp_bdev->params.vol_size / comp_bdev->comp_bdev.blocklen;
966 	assert(comp_bdev->comp_bdev.blockcnt > 0);
967 
968 	/* This is the context that is passed to us when the bdev
969 	 * layer calls in so we'll save our comp_bdev node here.
970 	 */
971 	comp_bdev->comp_bdev.ctxt = comp_bdev;
972 	comp_bdev->comp_bdev.fn_table = &vbdev_compress_fn_table;
973 	comp_bdev->comp_bdev.module = &compress_if;
974 
975 	/* Generate UUID based on namespace UUID + base bdev UUID. */
976 	spdk_uuid_parse(&ns_uuid, BDEV_COMPRESS_NAMESPACE_UUID);
977 	rc = spdk_uuid_generate_sha1(&comp_bdev->comp_bdev.uuid, &ns_uuid,
978 				     (const char *)&comp_bdev->base_bdev->uuid, sizeof(struct spdk_uuid));
979 	if (rc) {
980 		SPDK_ERRLOG("Unable to generate new UUID for compress bdev, error %s\n", spdk_strerror(-rc));
981 		return -EINVAL;
982 	}
983 
984 	pthread_mutex_init(&comp_bdev->reduce_lock, NULL);
985 
986 	/* Save the thread where the base device is opened */
987 	comp_bdev->thread = spdk_get_thread();
988 
989 	spdk_io_device_register(comp_bdev, comp_bdev_ch_create_cb, comp_bdev_ch_destroy_cb,
990 				sizeof(struct comp_io_channel),
991 				comp_bdev->comp_bdev.name);
992 
993 	rc = spdk_bdev_module_claim_bdev(comp_bdev->base_bdev, comp_bdev->base_desc,
994 					 comp_bdev->comp_bdev.module);
995 	if (rc) {
996 		SPDK_ERRLOG("could not claim bdev %s, error %s\n", spdk_bdev_get_name(comp_bdev->base_bdev),
997 			    spdk_strerror(-rc));
998 		goto error_claim;
999 	}
1000 
1001 	rc = spdk_bdev_register(&comp_bdev->comp_bdev);
1002 	if (rc < 0) {
1003 		SPDK_ERRLOG("trying to register bdev, error %s\n", spdk_strerror(-rc));
1004 		goto error_bdev_register;
1005 	}
1006 
1007 	TAILQ_INSERT_TAIL(&g_vbdev_comp, comp_bdev, link);
1008 
1009 	SPDK_NOTICELOG("registered io_device and virtual bdev for: %s\n", comp_bdev->comp_bdev.name);
1010 
1011 	return 0;
1012 
1013 	/* Error cleanup paths. */
1014 error_bdev_register:
1015 	spdk_bdev_module_release_bdev(comp_bdev->base_bdev);
1016 error_claim:
1017 	spdk_io_device_unregister(comp_bdev, NULL);
1018 	free(comp_bdev->comp_bdev.name);
1019 	return rc;
1020 }
1021 
1022 static void
1023 _vbdev_compress_delete_done(void *_ctx)
1024 {
1025 	struct vbdev_comp_delete_ctx *ctx = _ctx;
1026 
1027 	ctx->cb_fn(ctx->cb_arg, ctx->cb_rc);
1028 
1029 	free(ctx);
1030 }
1031 
1032 static void
1033 vbdev_compress_delete_done(void *cb_arg, int bdeverrno)
1034 {
1035 	struct vbdev_comp_delete_ctx *ctx = cb_arg;
1036 
1037 	ctx->cb_rc = bdeverrno;
1038 
1039 	if (ctx->orig_thread != spdk_get_thread()) {
1040 		spdk_thread_send_msg(ctx->orig_thread, _vbdev_compress_delete_done, ctx);
1041 	} else {
1042 		_vbdev_compress_delete_done(ctx);
1043 	}
1044 }
1045 
1046 void
1047 bdev_compress_delete(const char *name, spdk_delete_compress_complete cb_fn, void *cb_arg)
1048 {
1049 	struct vbdev_compress *comp_bdev = NULL;
1050 	struct vbdev_comp_delete_ctx *ctx;
1051 
1052 	TAILQ_FOREACH(comp_bdev, &g_vbdev_comp, link) {
1053 		if (strcmp(name, comp_bdev->comp_bdev.name) == 0) {
1054 			break;
1055 		}
1056 	}
1057 
1058 	if (comp_bdev == NULL) {
1059 		cb_fn(cb_arg, -ENODEV);
1060 		return;
1061 	}
1062 
1063 	ctx = calloc(1, sizeof(*ctx));
1064 	if (ctx == NULL) {
1065 		SPDK_ERRLOG("Failed to allocate delete context\n");
1066 		cb_fn(cb_arg, -ENOMEM);
1067 		return;
1068 	}
1069 
1070 	/* Save these for after the vol is destroyed. */
1071 	ctx->cb_fn = cb_fn;
1072 	ctx->cb_arg = cb_arg;
1073 	ctx->orig_thread = spdk_get_thread();
1074 
1075 	comp_bdev->delete_ctx = ctx;
1076 
1077 	/* Tell reducelib that we're done with this volume. */
1078 	if (comp_bdev->orphaned == false) {
1079 		spdk_reduce_vol_unload(comp_bdev->vol, delete_vol_unload_cb, comp_bdev);
1080 	} else {
1081 		delete_vol_unload_cb(comp_bdev, 0);
1082 	}
1083 }
1084 
1085 static void
1086 _vbdev_reduce_load_cb(void *ctx)
1087 {
1088 	struct vbdev_compress *comp_bdev = ctx;
1089 	int rc;
1090 
1091 	assert(comp_bdev->base_desc != NULL);
1092 
1093 	/* Done with metadata operations */
1094 	spdk_put_io_channel(comp_bdev->base_ch);
1095 
1096 	if (comp_bdev->reduce_errno == 0) {
1097 		rc = vbdev_compress_claim(comp_bdev);
1098 		if (rc != 0) {
1099 			goto err;
1100 		}
1101 	} else if (comp_bdev->reduce_errno == -ENOENT) {
1102 		if (_set_compbdev_name(comp_bdev)) {
1103 			goto err;
1104 		}
1105 
1106 		/* Save the thread where the base device is opened */
1107 		comp_bdev->thread = spdk_get_thread();
1108 
1109 		comp_bdev->comp_bdev.module = &compress_if;
1110 		pthread_mutex_init(&comp_bdev->reduce_lock, NULL);
1111 		rc = spdk_bdev_module_claim_bdev(comp_bdev->base_bdev, comp_bdev->base_desc,
1112 						 comp_bdev->comp_bdev.module);
1113 		if (rc) {
1114 			SPDK_ERRLOG("could not claim bdev %s, error %s\n", spdk_bdev_get_name(comp_bdev->base_bdev),
1115 				    spdk_strerror(-rc));
1116 			free(comp_bdev->comp_bdev.name);
1117 			goto err;
1118 		}
1119 
1120 		comp_bdev->orphaned = true;
1121 		TAILQ_INSERT_TAIL(&g_vbdev_comp, comp_bdev, link);
1122 	} else {
1123 		if (comp_bdev->reduce_errno != -EILSEQ) {
1124 			SPDK_ERRLOG("for vol %s, error %s\n", spdk_bdev_get_name(comp_bdev->base_bdev),
1125 				    spdk_strerror(-comp_bdev->reduce_errno));
1126 		}
1127 		goto err;
1128 	}
1129 
1130 	spdk_bdev_module_examine_done(&compress_if);
1131 	return;
1132 
1133 err:
1134 	/* Close the underlying bdev on its same opened thread. */
1135 	spdk_bdev_close(comp_bdev->base_desc);
1136 	free(comp_bdev);
1137 	spdk_bdev_module_examine_done(&compress_if);
1138 }
1139 
1140 /* Callback from reduce for then load is complete. We'll pass the vbdev_comp struct
1141  * used for initial metadata operations to claim where it will be further filled out
1142  * and added to the global list.
1143  */
1144 static void
1145 vbdev_reduce_load_cb(void *cb_arg, struct spdk_reduce_vol *vol, int reduce_errno)
1146 {
1147 	struct vbdev_compress *comp_bdev = cb_arg;
1148 
1149 	if (reduce_errno == 0) {
1150 		/* Update information following volume load. */
1151 		comp_bdev->vol = vol;
1152 		memcpy(&comp_bdev->params, spdk_reduce_vol_get_params(vol),
1153 		       sizeof(struct spdk_reduce_vol_params));
1154 	}
1155 
1156 	comp_bdev->reduce_errno = reduce_errno;
1157 
1158 	if (comp_bdev->thread && comp_bdev->thread != spdk_get_thread()) {
1159 		spdk_thread_send_msg(comp_bdev->thread, _vbdev_reduce_load_cb, comp_bdev);
1160 	} else {
1161 		_vbdev_reduce_load_cb(comp_bdev);
1162 	}
1163 
1164 }
1165 
1166 /* Examine_disk entry point: will do a metadata load to see if this is ours,
1167  * and if so will go ahead and claim it.
1168  */
1169 static void
1170 vbdev_compress_examine(struct spdk_bdev *bdev)
1171 {
1172 	struct spdk_bdev_desc *bdev_desc = NULL;
1173 	struct vbdev_compress *comp_bdev;
1174 	int rc;
1175 
1176 	if (strcmp(bdev->product_name, COMP_BDEV_NAME) == 0) {
1177 		spdk_bdev_module_examine_done(&compress_if);
1178 		return;
1179 	}
1180 
1181 	rc = spdk_bdev_open_ext(spdk_bdev_get_name(bdev), false,
1182 				vbdev_compress_base_bdev_event_cb, NULL, &bdev_desc);
1183 	if (rc) {
1184 		SPDK_ERRLOG("could not open bdev %s, error %s\n", spdk_bdev_get_name(bdev),
1185 			    spdk_strerror(-rc));
1186 		spdk_bdev_module_examine_done(&compress_if);
1187 		return;
1188 	}
1189 
1190 	comp_bdev = _prepare_for_load_init(bdev_desc, 0);
1191 	if (comp_bdev == NULL) {
1192 		spdk_bdev_close(bdev_desc);
1193 		spdk_bdev_module_examine_done(&compress_if);
1194 		return;
1195 	}
1196 
1197 	/* Save the thread where the base device is opened */
1198 	comp_bdev->thread = spdk_get_thread();
1199 
1200 	comp_bdev->base_ch = spdk_bdev_get_io_channel(comp_bdev->base_desc);
1201 	spdk_reduce_vol_load(&comp_bdev->backing_dev, vbdev_reduce_load_cb, comp_bdev);
1202 }
1203 
1204 SPDK_LOG_REGISTER_COMPONENT(vbdev_compress)
1205