Home
last modified time | relevance | path

Searched refs:rb_root (Results 1 – 22 of 22) sorted by relevance

/netbsd-src/sys/external/bsd/drm2/include/linux/
H A Drbtree.h39 struct rb_root { struct
44 struct rb_root rb_root; /* Linux API name */ argument
68 RB_EMPTY_ROOT(const struct rb_root *root) in RB_EMPTY_ROOT()
75 rb_first(const struct rb_root *root) in rb_first()
85 rb_next2(const struct rb_root *root, const struct rb_node *rbnode) in rb_next2()
97 rb_last(const struct rb_root *root) in rb_last()
109 return rb_first(&root->rb_root); in rb_first_cached()
113 rb_erase(struct rb_node *rbnode, struct rb_root *root) in rb_erase()
124 rb_erase(rbnode, &root->rb_root); in rb_erase_cached()
128 rb_replace_node(struct rb_node *old, struct rb_node *new, struct rb_root *root) in rb_replace_node()
[all …]
H A Dinterval_tree.h91 rb_tree_init(&root->rb_root.rbr_tree, &interval_tree_ops); in interval_tree_init()
100 collision = rb_tree_insert_node(&root->rb_root.rbr_tree, node); in interval_tree_insert()
109 rb_tree_remove_node(&root->rb_root.rbr_tree, node); in interval_tree_remove()
118 node = rb_tree_find_node_geq(&root->rb_root.rbr_tree, &start); in interval_tree_iter_first()
141 next = rb_tree_iterate(&root->rb_root.rbr_tree, node, RB_DIR_RIGHT); in interval_tree_iter_next()
H A Dinterval_tree_generic.h84 rb_tree_init(&__root->rb_root.rbr_tree, &PREFIX##__rbtree_ops); \
92 __collision = rb_tree_insert_node(&__root->rb_root.rbr_tree, __node); \
99 rb_tree_remove_node(&__root->rb_root.rbr_tree, __node); \
107 __node = rb_tree_find_node_geq(&__root->rb_root.rbr_tree, &__start); \
124 __next = rb_tree_iterate(&__root->rb_root.rbr_tree, __node, \
/netbsd-src/sys/external/bsd/drm2/dist/drm/i915/
H A Di915_scheduler.c64 rb_first(&execlists->queue.rb_root)); in assert_priolists()
69 rb = rb_next2(&execlists->queue.rb_root, rb)) { in assert_priolists()
126 rb_tree_init(&execlists->queue.rb_root.rbr_tree, in i915_sched_init()
157 p = rb_tree_find_node(&execlists->queue.rb_root.rbr_tree, &prio); in i915_sched_lookup_priolist()
163 parent = &execlists->queue.rb_root.rb_node; in i915_sched_lookup_priolist()
204 collision = rb_tree_insert_node(&execlists->queue.rb_root.rbr_tree, in i915_sched_lookup_priolist()
H A Di915_active_types.h44 struct rb_root tree;
H A Di915_active.c178 struct rb_root root; in __active_retire()
H A Di915_drv.h979 struct rb_root uabi_engines;
/netbsd-src/sys/external/bsd/drm2/dist/include/drm/
H A Ddrm_prime.h50 struct rb_root dmabufs;
51 struct rb_root handles;
H A Ddrm_vma_manager.h57 struct rb_root vm_files;
H A Ddrm_mm.h216 struct rb_root holes_addr;
/netbsd-src/sys/external/bsd/drm2/dist/drm/i915/gem/
H A Di915_gem_object_types.h118 struct rb_root tree;
148 struct rb_root offsets;
H A Di915_gem_userptr.c109 if (RB_EMPTY_ROOT(&mn->objects.rb_root)) in userptr_mn_invalidate_range_start()
/netbsd-src/sys/external/bsd/drm2/dist/drm/
H A Ddrm_mm.c211 link = &mm->interval_tree.rb_root.rb_node; in drm_mm_interval_tree_add_node()
339 collision = rb_tree_insert_node(&root->rb_root.rbr_tree, node); in insert_hole_size()
342 struct rb_node **link = &root->rb_root.rb_node, *rb = NULL; in insert_hole_size()
413 best = rb_tree_find_node_leq(&mm->holes_size.rb_root.rbr_tree, &size); in best_hole()
418 struct rb_node *rb = mm->holes_size.rb_root.rb_node; in best_hole()
495 return RB_TREE_PREV(&mm->holes_size.rb_root.rbr_tree, node); in next_hole()
1066 rb_tree_init(&mm->holes_size.rb_root.rbr_tree, &holes_size_rb_ops); in drm_mm_init()
H A Ddrm_vma_manager.c153 iter = mgr->vm_addr_space_mm.interval_tree.rb_root.rb_node; in drm_vma_offset_lookup_locked()
/netbsd-src/sys/external/bsd/drm2/dist/drm/nouveau/include/nvkm/subdev/
H A Dmmu.h46 struct rb_root free;
47 struct rb_root root;
/netbsd-src/sys/external/bsd/drm2/dist/drm/nouveau/include/nvkm/core/
H A Dclient.h19 struct rb_root objroot;
/netbsd-src/sys/external/bsd/drm2/dist/drm/i915/gvt/
H A Dgvt.h213 struct rb_root gfn_cache;
214 struct rb_root dma_addr_cache;
/netbsd-src/sys/external/bsd/drm2/dist/drm/i915/gt/
H A Dintel_lrc.c1034 GEM_BUG_ON(RB_EMPTY_ROOT(&engine->execlists.queue.rb_root)); in __unwind_incomplete_requests()
1869 rb = rb_next2(&execlists->virtual.rb_root, rb); in execlists_dequeue()
1988 rb = rb_next2(&execlists->virtual.rb_root, in execlists_dequeue()
2868 GEM_BUG_ON(RB_EMPTY_ROOT(&engine->execlists.queue.rb_root)); in execlists_submit_request()
3753 rb_tree_init(&execlists->virtual.rb_root.rbr_tree, &ve_tree_ops); in execlists_reset_cancel()
3774 if (!RB_EMPTY_ROOT(&execlists->queue.rb_root)) in execlists_reset_finish()
4907 &sibling->execlists.virtual.rb_root.rbr_tree, in virtual_submission_tasklet()
4912 &sibling->execlists.virtual.rb_root.rbr_tree, in virtual_submission_tasklet()
4917 parent = &sibling->execlists.virtual.rb_root.rb_node; in virtual_submission_tasklet()
5280 rb = rb_next2(&execlists->queue.rb_root, rb)) { in intel_execlists_show_requests()
[all …]
H A Dintel_engine_cs.c1118 if (!RB_EMPTY_ROOT(&engine->execlists.queue.rb_root)) in intel_engine_is_idle()
/netbsd-src/sys/external/bsd/drm2/dist/drm/radeon/
H A Dradeon_vm.c1247 if (!RB_EMPTY_ROOT(&vm->va.rb_root)) in radeon_vm_fini()
1251 &vm->va.rb_root, it.rb) { in radeon_vm_fini()
/netbsd-src/sys/external/bsd/drm2/dist/drm/amd/amdgpu/
H A Damdgpu_vm.c3161 if (!RB_EMPTY_ROOT(&vm->va.rb_root)) { in amdgpu_vm_fini()
3165 &vm->va.rb_root, rb) { in amdgpu_vm_fini()
/netbsd-src/sys/external/bsd/drm2/dist/drm/vmwgfx/
H A Dvmwgfx_drv.h122 struct rb_root res_tree;