/linux/Documentation/translations/zh_CN/mm/ |
H A D | active_mm.rst | 59 计数器,即有多少 “真正的地址空间用户”,另一个是 “mm_count”计数器,即 “lazy” 63 一个lazy的用户仍在活动,所以你实际上得到的情况是,你有一个地址空间 **只** 64 被lazy的用户使用。这通常是一个短暂的生命周期状态,因为一旦这个线程被安排给一 68 “init_mm”应该被认为只是一个 “没有其他上下文时的lazy上下文”,事实上,它主
|
/linux/Documentation/mm/ |
H A D | active_mm.rst | 5 Note, the mm_count refcount may no longer include the "lazy" users 7 with CONFIG_MMU_LAZY_TLB_REFCOUNT=n. Taking and releasing these lazy 63 and a "mm_count" counter that is the number of "lazy" users (ie anonymous 67 user exited on another CPU while a lazy user was still active, so you do 69 lazy users. That is often a short-lived state, because once that thread 74 more. "init_mm" should be considered just a "lazy context when no other
|
/linux/kernel/rcu/ |
H A D | tree_nocb.h | 331 unsigned long j, bool lazy) in rcu_nocb_do_flush_bypass() argument 353 if (lazy && rhp) { in rcu_nocb_do_flush_bypass() 375 unsigned long j, bool lazy) in rcu_nocb_flush_bypass() argument 417 bool lazy) in rcu_nocb_try_bypass() argument 490 if (!rcu_nocb_flush_bypass(rdp, rhp, j, lazy)) { in rcu_nocb_try_bypass() 519 if (lazy) in rcu_nocb_try_bypass() 535 if (!ncbs || (bypass_is_lazy && !lazy)) { in rcu_nocb_try_bypass() 696 bool lazy = false; in nocb_gp_wait() local 760 lazy = true; in nocb_gp_wait() 813 if (lazy && !bypass) { in nocb_gp_wait() [all …]
|
H A D | tree.h | 488 unsigned long j, bool lazy); 490 rcu_callback_t func, unsigned long flags, bool lazy);
|
H A D | Kconfig | 314 bool "RCU callback lazy invocation functionality" 326 bool "Turn RCU lazy invocation off by default"
|
/linux/kernel/ |
H A D | irq_work.c | 186 struct llist_head *raised, *lazy; in irq_work_needs_cpu() local 189 lazy = this_cpu_ptr(&lazy_list); in irq_work_needs_cpu() 192 if (llist_empty(lazy)) in irq_work_needs_cpu()
|
/linux/drivers/opp/ |
H A D | of.c | 153 list_del(&opp_table->lazy); in _opp_table_free_required_tables() 167 bool lazy = false; in _opp_table_alloc_required_tables() local 200 lazy = true; in _opp_table_alloc_required_tables() 204 if (lazy) { in _opp_table_alloc_required_tables() 210 list_add(&opp_table->lazy, &lazy_opp_tables); in _opp_table_alloc_required_tables() 412 list_for_each_entry_safe(opp_table, temp, &lazy_opp_tables, lazy) { in lazy_link_required_opp_table() 413 bool lazy = false; in lazy_link_required_opp_table() local 437 lazy = true; in lazy_link_required_opp_table() 448 lazy = false; in lazy_link_required_opp_table() 456 if (!lazy) { in lazy_link_required_opp_table() [all …]
|
H A D | opp.h | 206 struct list_head node, lazy; member 271 return unlikely(!list_empty(&opp_table->lazy)); in lazy_linking_pending()
|
/linux/drivers/crypto/intel/qat/qat_common/ |
H A D | icp_qat_hw_20_comp.h | 69 __u16 lazy; member 102 QAT_FIELD_SET(val32, csr.lazy, in ICP_QAT_FW_COMP_20_BUILD_CONFIG_UPPER()
|
H A D | adf_gen4_dc.c | 42 hw_comp_upper_csr.lazy = ICP_QAT_HW_COMP_20_CONFIG_CSR_LAZY_PARAM_DEFAULT_VAL; in qat_comp_build_deflate()
|
/linux/drivers/gpu/drm/vmwgfx/ |
H A D | vmwgfx_irq.c | 163 bool lazy, in vmw_fallback_wait() argument 211 if (lazy) in vmw_fallback_wait()
|
H A D | vmwgfx_fence.h | 94 bool lazy,
|
H A D | vmwgfx_fence.c | 522 int vmw_fence_obj_wait(struct vmw_fence_obj *fence, bool lazy, in vmw_fence_obj_wait() argument 763 ret = vmw_fence_obj_wait(fence, arg->lazy, true, timeout); in vmw_fence_obj_wait_ioctl()
|
/linux/Documentation/arch/arm/ |
H A D | kernel_mode_neon.rst | 30 The NEON/VFP register file is managed using lazy preserve (on UP systems) and 31 lazy restore (on both SMP and UP systems). This means that the register file is 45 mode will hit the lazy restore trap upon next use. This is handled by the
|
/linux/drivers/gpu/drm/nouveau/ |
H A D | nouveau_fence.h | 26 int nouveau_fence_wait(struct nouveau_fence *, bool lazy, bool intr);
|
H A D | nouveau_fence.c | 340 nouveau_fence_wait(struct nouveau_fence *fence, bool lazy, bool intr) in nouveau_fence_wait() argument 344 if (!lazy) in nouveau_fence_wait()
|
/linux/mm/ |
H A D | vmalloc.c | 884 struct rb_list lazy; member 2253 spin_lock(&vn->lazy.lock); in __purge_vmap_area_lazy() 2256 spin_unlock(&vn->lazy.lock); in __purge_vmap_area_lazy() 2353 spin_lock(&vn->lazy.lock); in free_vmap_area_noflush() 2354 insert_vmap_area(va, &vn->lazy.root, &vn->lazy.head); in free_vmap_area_noflush() 2355 spin_unlock(&vn->lazy.lock); in free_vmap_area_noflush() 4858 spin_lock(&vn->lazy.lock); in show_purge_info() 4864 spin_unlock(&vn->lazy.lock); in show_purge_info() 5036 vn->lazy.root = RB_ROOT; in vmap_init_nodes() 5037 INIT_LIST_HEAD(&vn->lazy.head); in vmap_init_nodes() [all …]
|
/linux/tools/perf/Documentation/ |
H A D | perf-probe.txt | 165 3) Define event based on source file with lazy pattern 176 …ine, and '%return' means that it probes function return. And ';PTN' means lazy matching pattern (s… 177 …ber or lazy matching by using 'SRC:ALN' or 'SRC;PTN' syntax, where 'SRC' is the source file path, … 229 The lazy line matching is similar to glob matching but ignoring spaces in both of pattern and targe…
|
/linux/include/uapi/drm/ |
H A D | vmwgfx_drm.h | 647 __s32 lazy; member
|
/linux/Documentation/arch/parisc/ |
H A D | registers.rst | 18 CR10 (CCR) lazy FPU saving*
|
/linux/Documentation/arch/x86/ |
H A D | iommu.rst | 143 iommu: DMA domain TLB invalidation policy: lazy mode
|
/linux/include/asm-generic/ |
H A D | hyperv-tlfs.h | 355 u64 lazy:1; member
|
/linux/arch/ |
H A D | Kconfig | 484 # "lazy tlb" beyond its final refcount (i.e., by the time __mmdrop frees the mm 490 # the lazy tlb reference of a kthread's ->active_mm (non-arch code has been 497 # mm as a lazy tlb beyond its last reference count, by shooting down these 499 # be using the mm as a lazy tlb, so that they may switch themselves to using 501 # may be using mm as a lazy tlb mm. 505 # at least all possible CPUs in which the mm is lazy.
|
/linux/Documentation/filesystems/ |
H A D | fuse.rst | 27 umounted. Note that detaching (or lazy umounting) the filesystem 199 filesystem is still attached (it hasn't been lazy unmounted)
|
/linux/Documentation/driver-api/driver-model/ |
H A D | devres.rst | 30 that's probably because libata low level driver developers are lazy
|