/netbsd-src/sys/external/bsd/drm2/include/linux/ |
H A D | rbtree.h | 39 struct rb_root { struct 44 struct rb_root rb_root; /* Linux API name */ argument 68 RB_EMPTY_ROOT(const struct rb_root *root) in RB_EMPTY_ROOT() 75 rb_first(const struct rb_root *root) in rb_first() 85 rb_next2(const struct rb_root *root, const struct rb_node *rbnode) in rb_next2() 97 rb_last(const struct rb_root *root) in rb_last() 109 return rb_first(&root->rb_root); in rb_first_cached() 113 rb_erase(struct rb_node *rbnode, struct rb_root *root) in rb_erase() 124 rb_erase(rbnode, &root->rb_root); in rb_erase_cached() 128 rb_replace_node(struct rb_node *old, struct rb_node *new, struct rb_root *root) in rb_replace_node() [all …]
|
H A D | interval_tree.h | 91 rb_tree_init(&root->rb_root.rbr_tree, &interval_tree_ops); in interval_tree_init() 100 collision = rb_tree_insert_node(&root->rb_root.rbr_tree, node); in interval_tree_insert() 109 rb_tree_remove_node(&root->rb_root.rbr_tree, node); in interval_tree_remove() 118 node = rb_tree_find_node_geq(&root->rb_root.rbr_tree, &start); in interval_tree_iter_first() 141 next = rb_tree_iterate(&root->rb_root.rbr_tree, node, RB_DIR_RIGHT); in interval_tree_iter_next()
|
H A D | interval_tree_generic.h | 84 rb_tree_init(&__root->rb_root.rbr_tree, &PREFIX##__rbtree_ops); \ 92 __collision = rb_tree_insert_node(&__root->rb_root.rbr_tree, __node); \ 99 rb_tree_remove_node(&__root->rb_root.rbr_tree, __node); \ 107 __node = rb_tree_find_node_geq(&__root->rb_root.rbr_tree, &__start); \ 124 __next = rb_tree_iterate(&__root->rb_root.rbr_tree, __node, \
|
/netbsd-src/sys/external/bsd/drm2/dist/drm/i915/ |
H A D | i915_scheduler.c | 64 rb_first(&execlists->queue.rb_root)); in assert_priolists() 69 rb = rb_next2(&execlists->queue.rb_root, rb)) { in assert_priolists() 126 rb_tree_init(&execlists->queue.rb_root.rbr_tree, in i915_sched_init() 157 p = rb_tree_find_node(&execlists->queue.rb_root.rbr_tree, &prio); in i915_sched_lookup_priolist() 163 parent = &execlists->queue.rb_root.rb_node; in i915_sched_lookup_priolist() 204 collision = rb_tree_insert_node(&execlists->queue.rb_root.rbr_tree, in i915_sched_lookup_priolist()
|
H A D | i915_active_types.h | 44 struct rb_root tree;
|
H A D | i915_active.c | 178 struct rb_root root; in __active_retire()
|
H A D | i915_drv.h | 979 struct rb_root uabi_engines;
|
/netbsd-src/sys/external/bsd/drm2/dist/include/drm/ |
H A D | drm_prime.h | 50 struct rb_root dmabufs; 51 struct rb_root handles;
|
H A D | drm_vma_manager.h | 57 struct rb_root vm_files;
|
H A D | drm_mm.h | 216 struct rb_root holes_addr;
|
/netbsd-src/sys/external/bsd/drm2/dist/drm/i915/gem/ |
H A D | i915_gem_object_types.h | 118 struct rb_root tree; 148 struct rb_root offsets;
|
H A D | i915_gem_userptr.c | 109 if (RB_EMPTY_ROOT(&mn->objects.rb_root)) in userptr_mn_invalidate_range_start()
|
/netbsd-src/sys/external/bsd/drm2/dist/drm/ |
H A D | drm_mm.c | 211 link = &mm->interval_tree.rb_root.rb_node; in drm_mm_interval_tree_add_node() 339 collision = rb_tree_insert_node(&root->rb_root.rbr_tree, node); in insert_hole_size() 342 struct rb_node **link = &root->rb_root.rb_node, *rb = NULL; in insert_hole_size() 413 best = rb_tree_find_node_leq(&mm->holes_size.rb_root.rbr_tree, &size); in best_hole() 418 struct rb_node *rb = mm->holes_size.rb_root.rb_node; in best_hole() 495 return RB_TREE_PREV(&mm->holes_size.rb_root.rbr_tree, node); in next_hole() 1066 rb_tree_init(&mm->holes_size.rb_root.rbr_tree, &holes_size_rb_ops); in drm_mm_init()
|
H A D | drm_vma_manager.c | 153 iter = mgr->vm_addr_space_mm.interval_tree.rb_root.rb_node; in drm_vma_offset_lookup_locked()
|
/netbsd-src/sys/external/bsd/drm2/dist/drm/nouveau/include/nvkm/subdev/ |
H A D | mmu.h | 46 struct rb_root free; 47 struct rb_root root;
|
/netbsd-src/sys/external/bsd/drm2/dist/drm/nouveau/include/nvkm/core/ |
H A D | client.h | 19 struct rb_root objroot;
|
/netbsd-src/sys/external/bsd/drm2/dist/drm/i915/gvt/ |
H A D | gvt.h | 213 struct rb_root gfn_cache; 214 struct rb_root dma_addr_cache;
|
/netbsd-src/sys/external/bsd/drm2/dist/drm/i915/gt/ |
H A D | intel_lrc.c | 1034 GEM_BUG_ON(RB_EMPTY_ROOT(&engine->execlists.queue.rb_root)); in __unwind_incomplete_requests() 1869 rb = rb_next2(&execlists->virtual.rb_root, rb); in execlists_dequeue() 1988 rb = rb_next2(&execlists->virtual.rb_root, in execlists_dequeue() 2868 GEM_BUG_ON(RB_EMPTY_ROOT(&engine->execlists.queue.rb_root)); in execlists_submit_request() 3753 rb_tree_init(&execlists->virtual.rb_root.rbr_tree, &ve_tree_ops); in execlists_reset_cancel() 3774 if (!RB_EMPTY_ROOT(&execlists->queue.rb_root)) in execlists_reset_finish() 4907 &sibling->execlists.virtual.rb_root.rbr_tree, in virtual_submission_tasklet() 4912 &sibling->execlists.virtual.rb_root.rbr_tree, in virtual_submission_tasklet() 4917 parent = &sibling->execlists.virtual.rb_root.rb_node; in virtual_submission_tasklet() 5280 rb = rb_next2(&execlists->queue.rb_root, rb)) { in intel_execlists_show_requests() [all …]
|
H A D | intel_engine_cs.c | 1118 if (!RB_EMPTY_ROOT(&engine->execlists.queue.rb_root)) in intel_engine_is_idle()
|
/netbsd-src/sys/external/bsd/drm2/dist/drm/radeon/ |
H A D | radeon_vm.c | 1247 if (!RB_EMPTY_ROOT(&vm->va.rb_root)) in radeon_vm_fini() 1251 &vm->va.rb_root, it.rb) { in radeon_vm_fini()
|
/netbsd-src/sys/external/bsd/drm2/dist/drm/amd/amdgpu/ |
H A D | amdgpu_vm.c | 3161 if (!RB_EMPTY_ROOT(&vm->va.rb_root)) { in amdgpu_vm_fini() 3165 &vm->va.rb_root, rb) { in amdgpu_vm_fini()
|
/netbsd-src/sys/external/bsd/drm2/dist/drm/vmwgfx/ |
H A D | vmwgfx_drv.h | 122 struct rb_root res_tree;
|