Lines Matching refs:lockroot
270 struct hammer_node_lock lockroot; in rebalance_node() local
292 hammer_node_lock_init(&lockroot, cursor->node); in rebalance_node()
296 error = hammer_btree_lock_children(cursor, 2, &lockroot, lcache); in rebalance_node()
305 hammer_btree_lock_copy(cursor, &lockroot); in rebalance_node()
310 if (TAILQ_FIRST(&lockroot.list) == NULL) in rebalance_node()
312 type1 = TAILQ_FIRST(&lockroot.list)->node->ondisk->type; in rebalance_node()
325 hdkprintf("lockroot %p count %d\n", &lockroot, lockroot.count); in rebalance_node()
327 TAILQ_FOREACH(item, &lockroot.list, entry) { in rebalance_node()
333 avg_elms = howmany(count, lockroot.count); in rebalance_node()
363 base_item = TAILQ_FIRST(&lockroot.list); in rebalance_node()
367 TAILQ_FOREACH(item, &lockroot.list, entry) { in rebalance_node()
378 elm = &lockroot.node->ondisk->elms[ in rebalance_node()
424 if (lockroot.copy->mirror_tid < tid) { in rebalance_node()
425 lockroot.copy->mirror_tid = tid; in rebalance_node()
426 lockroot.flags |= in rebalance_node()
429 if (lockroot.copy->elms[root_count]. in rebalance_node()
431 lockroot.copy->elms[root_count]. in rebalance_node()
433 lockroot.flags |= in rebalance_node()
475 elm = &lockroot.node->ondisk->elms[lockroot.node->ondisk->count]; in rebalance_node()
478 if (lockroot.copy->count != root_count) { in rebalance_node()
479 lockroot.copy->count = root_count; in rebalance_node()
480 lockroot.copy->elms[root_count] = *elm; in rebalance_node()
481 lockroot.flags |= HAMMER_NODE_LOCK_UPDATED; in rebalance_node()
495 hammer_cursor_removed_node(base_item->node, lockroot.node, in rebalance_node()
497 hammer_cursor_deleted_element(lockroot.node, base_count); in rebalance_node()
501 if (cursor->index > lockroot.copy->count) in rebalance_node()
510 rebal->stat_nrebal += hammer_btree_sync_copy(cursor, &lockroot); in rebalance_node()
512 hammer_btree_unlock_children(cursor->trans->hmp, &lockroot, lcache); in rebalance_node()