Home
last modified time | relevance | path

Searched refs:rb_entry (Results 1 – 22 of 22) sorted by relevance

/dragonfly/lib/libkvm/
H A Dkvm_vm_map.c72 while (copy->rb_entry.rbe_left) { in kvm_vm_map_entry_first()
73 ken = copy->rb_entry.rbe_left; in kvm_vm_map_entry_first()
88 if (copy->rb_entry.rbe_right) { in kvm_vm_map_entry_next()
89 ken = copy->rb_entry.rbe_right; in kvm_vm_map_entry_next()
92 while (copy->rb_entry.rbe_left) { in kvm_vm_map_entry_next()
93 ken = copy->rb_entry.rbe_left; in kvm_vm_map_entry_next()
98 if ((ken2 = copy->rb_entry.rbe_parent) == NULL) in kvm_vm_map_entry_next()
102 if (ken == copy->rb_entry.rbe_left) { in kvm_vm_map_entry_next()
105 while (ken == copy->rb_entry.rbe_right) { in kvm_vm_map_entry_next()
107 ken2 = copy->rb_entry.rbe_parent; in kvm_vm_map_entry_next()
/dragonfly/sys/dev/drm/
H A Ddrm_vma_manager.c155 node = rb_entry(iter, struct drm_vma_offset_node, vm_rb); in drm_vma_offset_lookup_locked()
188 iter_node = rb_entry(*iter, struct drm_vma_offset_node, vm_rb); in _drm_vma_offset_add_rb()
314 entry = rb_entry(*iter, struct drm_vma_offset_file, vm_rb); in drm_vma_node_allow()
367 entry = rb_entry(iter, struct drm_vma_offset_file, vm_rb); in drm_vma_node_revoke()
408 entry = rb_entry(iter, struct drm_vma_offset_file, vm_rb); in drm_vma_node_is_allowed()
H A Ddrm_prime.c111 pos = rb_entry(rb, struct drm_prime_member, dmabuf_rb); in drm_prime_add_buf_handle()
126 pos = rb_entry(rb, struct drm_prime_member, handle_rb); in drm_prime_add_buf_handle()
147 member = rb_entry(rb, struct drm_prime_member, handle_rb); in drm_prime_lookup_buf_by_handle()
169 member = rb_entry(rb, struct drm_prime_member, dmabuf_rb); in drm_prime_lookup_buf_handle()
240 member = rb_entry(rb, struct drm_prime_member, dmabuf_rb); in drm_prime_remove_buf_handle_locked()
/dragonfly/sbin/vquota/
H A Dvquota.c87 RB_ENTRY(hl_node) rb_entry;
94 RB_PROTOTYPE(hl_tree, hl_node, rb_entry, rb_hl_node_cmp);
99 RB_GENERATE(hl_tree, hl_node, rb_entry, rb_hl_node_cmp);
188 RB_PROTOTYPE(ac_utree, ac_unode, rb_entry, rb_ac_unode_cmp);
189 RB_PROTOTYPE(ac_gtree, ac_gnode, rb_entry, rb_ac_gnode_cmp);
190 RB_GENERATE(ac_utree, ac_unode, rb_entry, rb_ac_unode_cmp);
191 RB_GENERATE(ac_gtree, ac_gnode, rb_entry, rb_ac_gnode_cmp);
/dragonfly/sys/sys/
H A Dvkernel.h65 RB_PROTOTYPE(vmspace_rb_tree, vmspace_entry, rb_entry, rb_vmspace_compare);
94 RB_ENTRY(vmspace_entry) rb_entry;
H A Dmount.h163 RB_ENTRY(ac_unode) rb_entry;
169 RB_ENTRY(ac_gnode) rb_entry;
/dragonfly/sbin/hammer/
H A Dcmd_dedup.c47 RB_PROTOTYPE2(sim_dedup_entry_rb_tree, sim_dedup_entry, rb_entry,
54 RB_ENTRY(sim_dedup_entry) rb_entry;
67 RB_ENTRY(dedup_entry) rb_entry;
86 RB_PROTOTYPE2(dedup_entry_rb_tree, dedup_entry, rb_entry,
148 RB_GENERATE2(sim_dedup_entry_rb_tree, sim_dedup_entry, rb_entry,
150 RB_GENERATE2(dedup_entry_rb_tree, dedup_entry, rb_entry,
/dragonfly/sys/kern/
H A Dvfs_quota.c50 RB_PROTOTYPE(ac_utree, ac_unode, rb_entry, rb_ac_unode_cmp);
51 RB_PROTOTYPE(ac_gtree, ac_gnode, rb_entry, rb_ac_gnode_cmp);
58 RB_GENERATE(ac_utree, ac_unode, rb_entry, rb_ac_unode_cmp);
59 RB_GENERATE(ac_gtree, ac_gnode, rb_entry, rb_ac_gnode_cmp);
/dragonfly/sys/dev/drm/include/linux/
H A Drbtree.h83 #define rb_entry(ptr, type, member) container_of(ptr, type, member) macro
85 (ptr ? rb_entry(ptr, type, member) : NULL)
/dragonfly/sys/vm/
H A Dvm_map.h103 RB_PROTOTYPE(vm_map_rb_tree, vm_map_entry, rb_entry, rb_vm_map_compare);
225 RB_ENTRY(vm_map_entry) rb_entry;
239 #define MAPENT_FREELIST(ent) (ent)->rb_entry.rbe_left
H A Dvm_page.h177 RB_PROTOTYPE2(vm_page_rb_tree, vm_page, rb_entry,
183 RB_ENTRY(vm_page) rb_entry; /* Red-Black tree based at object */
H A Dvm_vmspace.c509 RB_GENERATE(vmspace_rb_tree, vmspace_entry, rb_entry, rb_vmspace_compare);
H A Dvm_map.c266 RB_GENERATE(vm_map_rb_tree, vm_map_entry, rb_entry, rb_vm_map_compare);
1213 tmp = RB_RIGHT(tmp, rb_entry); in vm_map_lookup_entry()
1215 tmp = RB_LEFT(tmp, rb_entry); in vm_map_lookup_entry()
H A Dvm_page.c167 RB_GENERATE2(vm_page_rb_tree, vm_page, rb_entry, rb_vm_page_compare,
/dragonfly/test/debug/
H A Dvmobjinfo.c236 dump_memq(kd, pg.rb_entry.rbe_left); in dump_memq()
261 dump_memq(kd, pg.rb_entry.rbe_right); in dump_memq()
/dragonfly/usr.bin/gcore/
H A Delfcore.c216 for (entry = map; entry; entry = entry->rb_entry.rbe_parent) in each_writable_segment()
362 vm_map_entry_t next = map->rb_entry.rbe_parent; in freemap()
534 linkp = &ent->rb_entry.rbe_parent; in readmap()
/dragonfly/sys/dev/drm/i915/
H A Dintel_breadcrumbs.c298 return rb_entry(node, struct intel_wait, node); in to_wait()
571 return rb_entry(rb, struct drm_i915_gem_request, signaling.node); in to_signaler()
H A Dintel_lrc.c308 p = rb_entry(rb, typeof(*p), node); in lookup_priolist()
571 rb_entry(rb, struct i915_priolist, node)->priority > in execlists_dequeue()
617 struct i915_priolist *p = rb_entry(rb, typeof(*p), node); in execlists_dequeue()
733 struct i915_priolist *p = rb_entry(rb, typeof(*p), node); in execlists_cancel_requests()
H A Di915_vma.c158 pos = rb_entry(rb, struct i915_vma, obj_node); in vma_create()
184 struct i915_vma *vma = rb_entry(rb, struct i915_vma, obj_node); in vma_lookup()
H A Dintel_engine_cs.c1773 rb_entry(rb, typeof(*p), node); in intel_engine_dump()
1782 struct intel_wait *w = rb_entry(rb, typeof(*w), node); in intel_engine_dump()
H A Di915_guc_submission.c578 struct i915_priolist *p = rb_entry(rb, typeof(*p), node); in i915_guc_dequeue()
/dragonfly/sys/dev/drm/amd/amdgpu/
H A Damdgpu_vm.c78 node = rb_entry(rb, typeof(*node), rb);
93 node = rb_entry(rb, typeof(*node), rb);
117 iter_node = rb_entry(*iter, struct amdgpu_bo_va_mapping, rb);