Lines Matching refs:zilog

134 zil_bp_tree_init(zilog_t *zilog)  in zil_bp_tree_init()  argument
136 avl_create(&zilog->zl_bp_tree, zil_bp_compare, in zil_bp_tree_init()
141 zil_bp_tree_fini(zilog_t *zilog) in zil_bp_tree_fini() argument
143 avl_tree_t *t = &zilog->zl_bp_tree; in zil_bp_tree_fini()
154 zil_bp_tree_add(zilog_t *zilog, const blkptr_t *bp) in zil_bp_tree_add() argument
156 avl_tree_t *t = &zilog->zl_bp_tree; in zil_bp_tree_add()
177 zil_header_in_syncing_context(zilog_t *zilog) in zil_header_in_syncing_context() argument
179 return ((zil_header_t *)zilog->zl_header); in zil_header_in_syncing_context()
183 zil_init_log_chain(zilog_t *zilog, blkptr_t *bp) in zil_init_log_chain() argument
189 zc->zc_word[ZIL_ZC_OBJSET] = dmu_objset_id(zilog->zl_os); in zil_init_log_chain()
197 zil_read_log_block(zilog_t *zilog, const blkptr_t *bp, blkptr_t *nbp, void *dst, in zil_read_log_block() argument
206 if (zilog->zl_header->zh_claim_txg == 0) in zil_read_log_block()
209 if (!(zilog->zl_header->zh_flags & ZIL_CLAIM_LR_SEQ_VALID)) in zil_read_log_block()
215 error = arc_read(NULL, zilog->zl_spa, bp, arc_getbuf_func, &abuf, in zil_read_log_block()
273 zil_read_log_data(zilog_t *zilog, const lr_write_t *lr, void *wbuf) in zil_read_log_data() argument
288 if (zilog->zl_header->zh_claim_txg == 0) in zil_read_log_data()
291 SET_BOOKMARK(&zb, dmu_objset_id(zilog->zl_os), lr->lr_foid, in zil_read_log_data()
294 error = arc_read(NULL, zilog->zl_spa, bp, arc_getbuf_func, &abuf, in zil_read_log_data()
310 zil_parse(zilog_t *zilog, zil_parse_blk_func_t *parse_blk_func, in zil_parse() argument
313 const zil_header_t *zh = zilog->zl_header; in zil_parse()
341 zil_bp_tree_init(zilog); in zil_parse()
350 if ((error = parse_blk_func(zilog, &blk, arg, txg)) != 0) in zil_parse()
359 error = zil_read_log_block(zilog, &blk, &next_blk, lrbuf, &end); in zil_parse()
369 if ((error = parse_lr_func(zilog, lr, arg, txg)) != 0) in zil_parse()
377 zilog->zl_parse_error = error; in zil_parse()
378 zilog->zl_parse_blk_seq = max_blk_seq; in zil_parse()
379 zilog->zl_parse_lr_seq = max_lr_seq; in zil_parse()
380 zilog->zl_parse_blk_count = blk_count; in zil_parse()
381 zilog->zl_parse_lr_count = lr_count; in zil_parse()
386 zil_bp_tree_fini(zilog); in zil_parse()
393 zil_claim_log_block(zilog_t *zilog, blkptr_t *bp, void *tx, uint64_t first_txg) in zil_claim_log_block() argument
400 zil_bp_tree_add(zilog, bp) != 0) in zil_claim_log_block()
403 return (zio_wait(zio_claim(NULL, zilog->zl_spa, in zil_claim_log_block()
409 zil_claim_log_record(zilog_t *zilog, lr_t *lrc, void *tx, uint64_t first_txg) in zil_claim_log_record() argument
426 (error = zil_read_log_data(zilog, lr, NULL)) != 0) in zil_claim_log_record()
428 return (zil_claim_log_block(zilog, &lr->lr_blkptr, tx, first_txg)); in zil_claim_log_record()
433 zil_free_log_block(zilog_t *zilog, blkptr_t *bp, void *tx, uint64_t claim_txg) in zil_free_log_block() argument
435 zio_free_zil(zilog->zl_spa, dmu_tx_get_txg(tx), bp); in zil_free_log_block()
441 zil_free_log_record(zilog_t *zilog, lr_t *lrc, void *tx, uint64_t claim_txg) in zil_free_log_record() argument
450 bp->blk_birth >= claim_txg && zil_bp_tree_add(zilog, bp) == 0 && in zil_free_log_record()
452 zio_free(zilog->zl_spa, dmu_tx_get_txg(tx), bp); in zil_free_log_record()
458 zil_alloc_lwb(zilog_t *zilog, blkptr_t *bp, boolean_t slog, uint64_t txg) in zil_alloc_lwb() argument
463 lwb->lwb_zilog = zilog; in zil_alloc_lwb()
478 mutex_enter(&zilog->zl_lock); in zil_alloc_lwb()
479 list_insert_tail(&zilog->zl_lwb_list, lwb); in zil_alloc_lwb()
480 mutex_exit(&zilog->zl_lock); in zil_alloc_lwb()
490 zilog_dirty(zilog_t *zilog, uint64_t txg) in zilog_dirty() argument
492 dsl_pool_t *dp = zilog->zl_dmu_pool; in zilog_dirty()
493 dsl_dataset_t *ds = dmu_objset_ds(zilog->zl_os); in zilog_dirty()
498 if (txg_list_add(&dp->dp_dirty_zilogs, zilog, txg)) { in zilog_dirty()
500 dmu_buf_add_ref(ds->ds_dbuf, zilog); in zilog_dirty()
512 zilog_is_dirty_in_txg(zilog_t *zilog, uint64_t txg) in zilog_is_dirty_in_txg() argument
514 dsl_pool_t *dp = zilog->zl_dmu_pool; in zilog_is_dirty_in_txg()
516 if (txg_list_member(&dp->dp_dirty_zilogs, zilog, txg & TXG_MASK)) in zilog_is_dirty_in_txg()
526 zilog_is_dirty(zilog_t *zilog) in zilog_is_dirty() argument
528 dsl_pool_t *dp = zilog->zl_dmu_pool; in zilog_is_dirty()
531 if (txg_list_member(&dp->dp_dirty_zilogs, zilog, t)) in zilog_is_dirty()
541 zil_create(zilog_t *zilog) in zil_create() argument
543 const zil_header_t *zh = zilog->zl_header; in zil_create()
554 txg_wait_synced(zilog->zl_dmu_pool, zilog->zl_destroy_txg); in zil_create()
567 tx = dmu_tx_create(zilog->zl_os); in zil_create()
569 dsl_dataset_dirty(dmu_objset_ds(zilog->zl_os), tx); in zil_create()
573 zio_free_zil(zilog->zl_spa, txg, &blk); in zil_create()
577 error = zio_alloc_zil(zilog->zl_spa, txg, &blk, NULL, in zil_create()
581 zil_init_log_chain(zilog, &blk); in zil_create()
588 lwb = zil_alloc_lwb(zilog, &blk, slog, txg); in zil_create()
597 txg_wait_synced(zilog->zl_dmu_pool, txg); in zil_create()
615 zil_destroy(zilog_t *zilog, boolean_t keep_first) in zil_destroy() argument
617 const zil_header_t *zh = zilog->zl_header; in zil_destroy()
625 txg_wait_synced(zilog->zl_dmu_pool, zilog->zl_destroy_txg); in zil_destroy()
627 zilog->zl_old_header = *zh; /* debugging aid */ in zil_destroy()
632 tx = dmu_tx_create(zilog->zl_os); in zil_destroy()
634 dsl_dataset_dirty(dmu_objset_ds(zilog->zl_os), tx); in zil_destroy()
637 mutex_enter(&zilog->zl_lock); in zil_destroy()
639 ASSERT3U(zilog->zl_destroy_txg, <, txg); in zil_destroy()
640 zilog->zl_destroy_txg = txg; in zil_destroy()
641 zilog->zl_keep_first = keep_first; in zil_destroy()
643 if (!list_is_empty(&zilog->zl_lwb_list)) { in zil_destroy()
646 while ((lwb = list_head(&zilog->zl_lwb_list)) != NULL) { in zil_destroy()
647 list_remove(&zilog->zl_lwb_list, lwb); in zil_destroy()
650 zio_free_zil(zilog->zl_spa, txg, &lwb->lwb_blk); in zil_destroy()
654 zil_destroy_sync(zilog, tx); in zil_destroy()
656 mutex_exit(&zilog->zl_lock); in zil_destroy()
662 zil_destroy_sync(zilog_t *zilog, dmu_tx_t *tx) in zil_destroy_sync() argument
664 ASSERT(list_is_empty(&zilog->zl_lwb_list)); in zil_destroy_sync()
665 (void) zil_parse(zilog, zil_free_log_block, in zil_destroy_sync()
666 zil_free_log_record, tx, zilog->zl_header->zh_claim_txg); in zil_destroy_sync()
674 zilog_t *zilog; in zil_claim() local
693 zilog = dmu_objset_zil(os); in zil_claim()
694 zh = zil_header_in_syncing_context(zilog); in zil_claim()
696 if (spa_get_log_state(zilog->zl_spa) == SPA_LOG_CLEAR) { in zil_claim()
698 zio_free_zil(zilog->zl_spa, first_txg, &zh->zh_log); in zil_claim()
714 (void) zil_parse(zilog, zil_claim_log_block, in zil_claim()
717 zh->zh_claim_blk_seq = zilog->zl_parse_blk_seq; in zil_claim()
718 zh->zh_claim_lr_seq = zilog->zl_parse_lr_seq; in zil_claim()
719 if (zilog->zl_parse_lr_count || zilog->zl_parse_blk_count > 1) in zil_claim()
725 ASSERT3U(first_txg, ==, (spa_last_synced_txg(zilog->zl_spa) + 1)); in zil_claim()
739 zilog_t *zilog; in zil_check_log_chain() local
753 zilog = dmu_objset_zil(os); in zil_check_log_chain()
754 bp = (blkptr_t *)&zilog->zl_header->zh_log; in zil_check_log_chain()
783 error = zil_parse(zilog, zil_claim_log_block, zil_claim_log_record, tx, in zil_check_log_chain()
784 zilog->zl_header->zh_claim_txg ? -1ULL : spa_first_txg(os->os_spa)); in zil_check_log_chain()
804 zil_add_block(zilog_t *zilog, const blkptr_t *bp) in zil_add_block() argument
806 avl_tree_t *t = &zilog->zl_vdev_tree; in zil_add_block()
815 ASSERT(zilog->zl_writer); in zil_add_block()
822 mutex_enter(&zilog->zl_vdev_lock); in zil_add_block()
831 mutex_exit(&zilog->zl_vdev_lock); in zil_add_block()
835 zil_flush_vdevs(zilog_t *zilog) in zil_flush_vdevs() argument
837 spa_t *spa = zilog->zl_spa; in zil_flush_vdevs()
838 avl_tree_t *t = &zilog->zl_vdev_tree; in zil_flush_vdevs()
843 ASSERT(zilog->zl_writer); in zil_flush_vdevs()
881 zilog_t *zilog = lwb->lwb_zilog; in zil_lwb_write_done() local
901 mutex_enter(&zilog->zl_lock); in zil_lwb_write_done()
904 mutex_exit(&zilog->zl_lock); in zil_lwb_write_done()
918 zil_lwb_write_init(zilog_t *zilog, lwb_t *lwb) in zil_lwb_write_init() argument
927 if (zilog->zl_root_zio == NULL) { in zil_lwb_write_init()
928 zilog->zl_root_zio = zio_root(zilog->zl_spa, NULL, NULL, in zil_lwb_write_init()
932 if (zilog->zl_cur_used <= zil_slog_limit || !lwb->lwb_slog) in zil_lwb_write_init()
936 lwb->lwb_zio = zio_rewrite(zilog->zl_root_zio, zilog->zl_spa, in zil_lwb_write_init()
962 zil_lwb_write_start(zilog_t *zilog, lwb_t *lwb, boolean_t last) in zil_lwb_write_start() argument
966 spa_t *spa = zilog->zl_spa; in zil_lwb_write_start()
993 tx = dmu_tx_create(zilog->zl_os); in zil_lwb_write_start()
995 dsl_dataset_dirty(dmu_objset_ds(zilog->zl_os), tx); in zil_lwb_write_start()
1016 zil_blksz = zilog->zl_cur_used + sizeof (zil_chain_t); in zil_lwb_write_start()
1022 zilog->zl_prev_blks[zilog->zl_prev_rotor] = zil_blksz; in zil_lwb_write_start()
1024 zil_blksz = MAX(zil_blksz, zilog->zl_prev_blks[i]); in zil_lwb_write_start()
1025 zilog->zl_prev_rotor = (zilog->zl_prev_rotor + 1) & (ZIL_PREV_BLKS - 1); in zil_lwb_write_start()
1038 nlwb = zil_alloc_lwb(zilog, bp, slog, txg); in zil_lwb_write_start()
1041 zil_add_block(zilog, &lwb->lwb_blk); in zil_lwb_write_start()
1075 zil_lwb_commit(zilog_t *zilog, itx_t *itx, lwb_t *lwb) in zil_lwb_commit() argument
1094 zilog->zl_cur_used += (reclen + dlen); in zil_lwb_commit()
1096 zil_lwb_write_init(zilog, lwb); in zil_lwb_commit()
1108 lwb = zil_lwb_write_start(zilog, lwb, B_FALSE); in zil_lwb_commit()
1111 zil_lwb_write_init(zilog, lwb); in zil_lwb_commit()
1127 if (txg > spa_freeze_txg(zilog->zl_spa)) in zil_lwb_commit()
1128 txg_wait_synced(zilog->zl_dmu_pool, txg); in zil_lwb_commit()
1144 error = zilog->zl_get_data( in zil_lwb_commit()
1147 txg_wait_synced(zilog->zl_dmu_pool, txg); in zil_lwb_commit()
1164 lrcb->lrc_seq = ++zilog->zl_lr_seq; /* we are single threaded */ in zil_lwb_commit()
1172 zilog->zl_cur_used += reclen; in zil_lwb_commit()
1256 zil_remove_async(zilog_t *zilog, uint64_t oid) in zil_remove_async() argument
1268 if (spa_freeze_txg(zilog->zl_spa) != UINT64_MAX) /* ziltest support */ in zil_remove_async()
1271 otxg = spa_last_synced_txg(zilog->zl_spa) + 1; in zil_remove_async()
1274 itxg_t *itxg = &zilog->zl_itxg[txg & TXG_MASK]; in zil_remove_async()
1300 zil_itx_assign(zilog_t *zilog, itx_t *itx, dmu_tx_t *tx) in zil_itx_assign() argument
1314 zil_remove_async(zilog, itx->itx_oid); in zil_itx_assign()
1320 zil_async_to_sync(zilog, itx->itx_oid); in zil_itx_assign()
1322 if (spa_freeze_txg(zilog->zl_spa) != UINT64_MAX) in zil_itx_assign()
1327 itxg = &zilog->zl_itxg[txg & TXG_MASK]; in zil_itx_assign()
1368 zilog_dirty(zilog, txg); in zil_itx_assign()
1384 zil_clean(zilog_t *zilog, uint64_t synced_txg) in zil_clean() argument
1386 itxg_t *itxg = &zilog->zl_itxg[synced_txg & TXG_MASK]; in zil_clean()
1396 ASSERT(zilog->zl_clean_taskq != NULL); in zil_clean()
1407 if (taskq_dispatch(zilog->zl_clean_taskq, in zil_clean()
1416 zil_get_commit_list(zilog_t *zilog) in zil_get_commit_list() argument
1419 list_t *commit_list = &zilog->zl_itx_commit_list; in zil_get_commit_list()
1421 if (spa_freeze_txg(zilog->zl_spa) != UINT64_MAX) /* ziltest support */ in zil_get_commit_list()
1424 otxg = spa_last_synced_txg(zilog->zl_spa) + 1; in zil_get_commit_list()
1432 itxg_t *itxg = &zilog->zl_itxg[txg & TXG_MASK]; in zil_get_commit_list()
1448 ASSERT(zilog_is_dirty_in_txg(zilog, txg) || in zil_get_commit_list()
1449 spa_freeze_txg(zilog->zl_spa) != UINT64_MAX); in zil_get_commit_list()
1460 zil_async_to_sync(zilog_t *zilog, uint64_t foid) in zil_async_to_sync() argument
1467 if (spa_freeze_txg(zilog->zl_spa) != UINT64_MAX) /* ziltest support */ in zil_async_to_sync()
1470 otxg = spa_last_synced_txg(zilog->zl_spa) + 1; in zil_async_to_sync()
1477 itxg_t *itxg = &zilog->zl_itxg[txg & TXG_MASK]; in zil_async_to_sync()
1513 zil_commit_writer(zilog_t *zilog) in zil_commit_writer() argument
1518 spa_t *spa = zilog->zl_spa; in zil_commit_writer()
1521 ASSERT(zilog->zl_root_zio == NULL); in zil_commit_writer()
1523 mutex_exit(&zilog->zl_lock); in zil_commit_writer()
1525 zil_get_commit_list(zilog); in zil_commit_writer()
1531 if (list_head(&zilog->zl_itx_commit_list) == NULL) { in zil_commit_writer()
1532 mutex_enter(&zilog->zl_lock); in zil_commit_writer()
1536 if (zilog->zl_suspend) { in zil_commit_writer()
1539 lwb = list_tail(&zilog->zl_lwb_list); in zil_commit_writer()
1541 lwb = zil_create(zilog); in zil_commit_writer()
1544 DTRACE_PROBE1(zil__cw1, zilog_t *, zilog); in zil_commit_writer()
1545 while (itx = list_head(&zilog->zl_itx_commit_list)) { in zil_commit_writer()
1556 lwb = zil_lwb_commit(zilog, itx, lwb); in zil_commit_writer()
1557 list_remove(&zilog->zl_itx_commit_list, itx); in zil_commit_writer()
1561 DTRACE_PROBE1(zil__cw2, zilog_t *, zilog); in zil_commit_writer()
1565 lwb = zil_lwb_write_start(zilog, lwb, B_TRUE); in zil_commit_writer()
1567 zilog->zl_cur_used = 0; in zil_commit_writer()
1572 if (zilog->zl_root_zio) { in zil_commit_writer()
1573 error = zio_wait(zilog->zl_root_zio); in zil_commit_writer()
1574 zilog->zl_root_zio = NULL; in zil_commit_writer()
1575 zil_flush_vdevs(zilog); in zil_commit_writer()
1579 txg_wait_synced(zilog->zl_dmu_pool, 0); in zil_commit_writer()
1581 mutex_enter(&zilog->zl_lock); in zil_commit_writer()
1589 zilog->zl_commit_lr_seq = zilog->zl_lr_seq; in zil_commit_writer()
1616 zil_commit(zilog_t *zilog, uint64_t foid) in zil_commit() argument
1620 if (zilog->zl_sync == ZFS_SYNC_DISABLED) in zil_commit()
1624 zil_async_to_sync(zilog, foid); in zil_commit()
1626 mutex_enter(&zilog->zl_lock); in zil_commit()
1627 mybatch = zilog->zl_next_batch; in zil_commit()
1628 while (zilog->zl_writer) { in zil_commit()
1629 cv_wait(&zilog->zl_cv_batch[mybatch & 1], &zilog->zl_lock); in zil_commit()
1630 if (mybatch <= zilog->zl_com_batch) { in zil_commit()
1631 mutex_exit(&zilog->zl_lock); in zil_commit()
1636 zilog->zl_next_batch++; in zil_commit()
1637 zilog->zl_writer = B_TRUE; in zil_commit()
1638 zil_commit_writer(zilog); in zil_commit()
1639 zilog->zl_com_batch = mybatch; in zil_commit()
1640 zilog->zl_writer = B_FALSE; in zil_commit()
1641 mutex_exit(&zilog->zl_lock); in zil_commit()
1644 cv_signal(&zilog->zl_cv_batch[(mybatch+1) & 1]); in zil_commit()
1647 cv_broadcast(&zilog->zl_cv_batch[mybatch & 1]); in zil_commit()
1654 zil_sync(zilog_t *zilog, dmu_tx_t *tx) in zil_sync() argument
1656 zil_header_t *zh = zil_header_in_syncing_context(zilog); in zil_sync()
1658 spa_t *spa = zilog->zl_spa; in zil_sync()
1659 uint64_t *replayed_seq = &zilog->zl_replayed_seq[txg & TXG_MASK]; in zil_sync()
1669 mutex_enter(&zilog->zl_lock); in zil_sync()
1671 ASSERT(zilog->zl_stop_sync == 0); in zil_sync()
1679 if (zilog->zl_destroy_txg == txg) { in zil_sync()
1682 ASSERT(list_head(&zilog->zl_lwb_list) == NULL); in zil_sync()
1685 bzero(zilog->zl_replayed_seq, sizeof (zilog->zl_replayed_seq)); in zil_sync()
1687 if (zilog->zl_keep_first) { in zil_sync()
1696 zil_init_log_chain(zilog, &blk); in zil_sync()
1701 while ((lwb = list_head(&zilog->zl_lwb_list)) != NULL) { in zil_sync()
1705 list_remove(&zilog->zl_lwb_list, lwb); in zil_sync()
1715 if (list_head(&zilog->zl_lwb_list) == NULL) in zil_sync()
1718 mutex_exit(&zilog->zl_lock); in zil_sync()
1735 zil_set_sync(zilog_t *zilog, uint64_t sync) in zil_set_sync() argument
1737 zilog->zl_sync = sync; in zil_set_sync()
1741 zil_set_logbias(zilog_t *zilog, uint64_t logbias) in zil_set_logbias() argument
1743 zilog->zl_logbias = logbias; in zil_set_logbias()
1749 zilog_t *zilog; in zil_alloc() local
1751 zilog = kmem_zalloc(sizeof (zilog_t), KM_SLEEP); in zil_alloc()
1753 zilog->zl_header = zh_phys; in zil_alloc()
1754 zilog->zl_os = os; in zil_alloc()
1755 zilog->zl_spa = dmu_objset_spa(os); in zil_alloc()
1756 zilog->zl_dmu_pool = dmu_objset_pool(os); in zil_alloc()
1757 zilog->zl_destroy_txg = TXG_INITIAL - 1; in zil_alloc()
1758 zilog->zl_logbias = dmu_objset_logbias(os); in zil_alloc()
1759 zilog->zl_sync = dmu_objset_syncprop(os); in zil_alloc()
1760 zilog->zl_next_batch = 1; in zil_alloc()
1762 mutex_init(&zilog->zl_lock, NULL, MUTEX_DEFAULT, NULL); in zil_alloc()
1765 mutex_init(&zilog->zl_itxg[i].itxg_lock, NULL, in zil_alloc()
1769 list_create(&zilog->zl_lwb_list, sizeof (lwb_t), in zil_alloc()
1772 list_create(&zilog->zl_itx_commit_list, sizeof (itx_t), in zil_alloc()
1775 mutex_init(&zilog->zl_vdev_lock, NULL, MUTEX_DEFAULT, NULL); in zil_alloc()
1777 avl_create(&zilog->zl_vdev_tree, zil_vdev_compare, in zil_alloc()
1780 cv_init(&zilog->zl_cv_writer, NULL, CV_DEFAULT, NULL); in zil_alloc()
1781 cv_init(&zilog->zl_cv_suspend, NULL, CV_DEFAULT, NULL); in zil_alloc()
1782 cv_init(&zilog->zl_cv_batch[0], NULL, CV_DEFAULT, NULL); in zil_alloc()
1783 cv_init(&zilog->zl_cv_batch[1], NULL, CV_DEFAULT, NULL); in zil_alloc()
1785 return (zilog); in zil_alloc()
1789 zil_free(zilog_t *zilog) in zil_free() argument
1791 zilog->zl_stop_sync = 1; in zil_free()
1793 ASSERT0(zilog->zl_suspend); in zil_free()
1794 ASSERT0(zilog->zl_suspending); in zil_free()
1796 ASSERT(list_is_empty(&zilog->zl_lwb_list)); in zil_free()
1797 list_destroy(&zilog->zl_lwb_list); in zil_free()
1799 avl_destroy(&zilog->zl_vdev_tree); in zil_free()
1800 mutex_destroy(&zilog->zl_vdev_lock); in zil_free()
1802 ASSERT(list_is_empty(&zilog->zl_itx_commit_list)); in zil_free()
1803 list_destroy(&zilog->zl_itx_commit_list); in zil_free()
1813 if (zilog->zl_itxg[i].itxg_itxs) in zil_free()
1814 zil_itxg_clean(zilog->zl_itxg[i].itxg_itxs); in zil_free()
1815 mutex_destroy(&zilog->zl_itxg[i].itxg_lock); in zil_free()
1818 mutex_destroy(&zilog->zl_lock); in zil_free()
1820 cv_destroy(&zilog->zl_cv_writer); in zil_free()
1821 cv_destroy(&zilog->zl_cv_suspend); in zil_free()
1822 cv_destroy(&zilog->zl_cv_batch[0]); in zil_free()
1823 cv_destroy(&zilog->zl_cv_batch[1]); in zil_free()
1825 kmem_free(zilog, sizeof (zilog_t)); in zil_free()
1834 zilog_t *zilog = dmu_objset_zil(os); in zil_open() local
1836 ASSERT(zilog->zl_clean_taskq == NULL); in zil_open()
1837 ASSERT(zilog->zl_get_data == NULL); in zil_open()
1838 ASSERT(list_is_empty(&zilog->zl_lwb_list)); in zil_open()
1840 zilog->zl_get_data = get_data; in zil_open()
1841 zilog->zl_clean_taskq = taskq_create("zil_clean", 1, minclsyspri, in zil_open()
1844 return (zilog); in zil_open()
1851 zil_close(zilog_t *zilog) in zil_close() argument
1856 zil_commit(zilog, 0); /* commit all itx */ in zil_close()
1864 mutex_enter(&zilog->zl_lock); in zil_close()
1865 lwb = list_tail(&zilog->zl_lwb_list); in zil_close()
1868 mutex_exit(&zilog->zl_lock); in zil_close()
1870 txg_wait_synced(zilog->zl_dmu_pool, txg); in zil_close()
1872 if (zilog_is_dirty(zilog)) in zil_close()
1873 zfs_dbgmsg("zil (%p) is dirty, txg %llu", zilog, txg); in zil_close()
1874 VERIFY(!zilog_is_dirty(zilog)); in zil_close()
1876 taskq_destroy(zilog->zl_clean_taskq); in zil_close()
1877 zilog->zl_clean_taskq = NULL; in zil_close()
1878 zilog->zl_get_data = NULL; in zil_close()
1883 mutex_enter(&zilog->zl_lock); in zil_close()
1884 lwb = list_head(&zilog->zl_lwb_list); in zil_close()
1886 ASSERT(lwb == list_tail(&zilog->zl_lwb_list)); in zil_close()
1887 list_remove(&zilog->zl_lwb_list, lwb); in zil_close()
1891 mutex_exit(&zilog->zl_lock); in zil_close()
1922 zilog_t *zilog; in zil_suspend() local
1929 zilog = dmu_objset_zil(os); in zil_suspend()
1931 mutex_enter(&zilog->zl_lock); in zil_suspend()
1932 zh = zilog->zl_header; in zil_suspend()
1935 mutex_exit(&zilog->zl_lock); in zil_suspend()
1946 if (cookiep == NULL && !zilog->zl_suspending && in zil_suspend()
1947 (zilog->zl_suspend > 0 || BP_IS_HOLE(&zh->zh_log))) { in zil_suspend()
1948 mutex_exit(&zilog->zl_lock); in zil_suspend()
1956 zilog->zl_suspend++; in zil_suspend()
1958 if (zilog->zl_suspend > 1) { in zil_suspend()
1964 while (zilog->zl_suspending) in zil_suspend()
1965 cv_wait(&zilog->zl_cv_suspend, &zilog->zl_lock); in zil_suspend()
1966 mutex_exit(&zilog->zl_lock); in zil_suspend()
1984 mutex_exit(&zilog->zl_lock); in zil_suspend()
1988 zilog->zl_suspending = B_TRUE; in zil_suspend()
1989 mutex_exit(&zilog->zl_lock); in zil_suspend()
1991 zil_commit(zilog, 0); in zil_suspend()
1993 zil_destroy(zilog, B_FALSE); in zil_suspend()
1995 mutex_enter(&zilog->zl_lock); in zil_suspend()
1996 zilog->zl_suspending = B_FALSE; in zil_suspend()
1997 cv_broadcast(&zilog->zl_cv_suspend); in zil_suspend()
1998 mutex_exit(&zilog->zl_lock); in zil_suspend()
2011 zilog_t *zilog = dmu_objset_zil(os); in zil_resume() local
2013 mutex_enter(&zilog->zl_lock); in zil_resume()
2014 ASSERT(zilog->zl_suspend != 0); in zil_resume()
2015 zilog->zl_suspend--; in zil_resume()
2016 mutex_exit(&zilog->zl_lock); in zil_resume()
2029 zil_replay_error(zilog_t *zilog, lr_t *lr, int error) in zil_replay_error() argument
2033 zilog->zl_replaying_seq--; /* didn't actually replay this one */ in zil_replay_error()
2035 dmu_objset_name(zilog->zl_os, name); in zil_replay_error()
2047 zil_replay_log_record(zilog_t *zilog, lr_t *lr, void *zra, uint64_t claim_txg) in zil_replay_log_record() argument
2050 const zil_header_t *zh = zilog->zl_header; in zil_replay_log_record()
2055 zilog->zl_replaying_seq = lr->lrc_seq; in zil_replay_log_record()
2067 return (zil_replay_error(zilog, lr, EINVAL)); in zil_replay_log_record()
2074 error = dmu_object_info(zilog->zl_os, in zil_replay_log_record()
2089 error = zil_read_log_data(zilog, (lr_write_t *)lr, in zil_replay_log_record()
2092 return (zil_replay_error(zilog, lr, error)); in zil_replay_log_record()
2120 txg_wait_synced(spa_get_dsl(zilog->zl_spa), 0); in zil_replay_log_record()
2123 return (zil_replay_error(zilog, lr, error)); in zil_replay_log_record()
2130 zil_incr_blks(zilog_t *zilog, blkptr_t *bp, void *arg, uint64_t claim_txg) in zil_incr_blks() argument
2132 zilog->zl_replay_blks++; in zil_incr_blks()
2143 zilog_t *zilog = dmu_objset_zil(os); in zil_replay() local
2144 const zil_header_t *zh = zilog->zl_header; in zil_replay()
2148 zil_destroy(zilog, B_TRUE); in zil_replay()
2160 txg_wait_synced(zilog->zl_dmu_pool, 0); in zil_replay()
2162 zilog->zl_replay = B_TRUE; in zil_replay()
2163 zilog->zl_replay_time = ddi_get_lbolt(); in zil_replay()
2164 ASSERT(zilog->zl_replay_blks == 0); in zil_replay()
2165 (void) zil_parse(zilog, zil_incr_blks, zil_replay_log_record, &zr, in zil_replay()
2169 zil_destroy(zilog, B_FALSE); in zil_replay()
2170 txg_wait_synced(zilog->zl_dmu_pool, zilog->zl_destroy_txg); in zil_replay()
2171 zilog->zl_replay = B_FALSE; in zil_replay()
2175 zil_replaying(zilog_t *zilog, dmu_tx_t *tx) in zil_replaying() argument
2177 if (zilog->zl_sync == ZFS_SYNC_DISABLED) in zil_replaying()
2180 if (zilog->zl_replay) { in zil_replaying()
2181 dsl_dataset_dirty(dmu_objset_ds(zilog->zl_os), tx); in zil_replaying()
2182 zilog->zl_replayed_seq[dmu_tx_get_txg(tx) & TXG_MASK] = in zil_replaying()
2183 zilog->zl_replaying_seq; in zil_replaying()