/dports/multimedia/v4l_compat/linux-5.13-rc2/mm/ |
H A D | page_vma_mapped.c | 18 pvmw->pte = pte_offset_map(pvmw->pmd, pvmw->address); in map_pte() 50 pvmw->ptl = pte_lockptr(pvmw->vma->vm_mm, pvmw->pmd); in map_pte() 153 if (pvmw->pmd && !pvmw->pte) in page_vma_mapped_walk() 156 if (pvmw->pte) in page_vma_mapped_walk() 181 pvmw->pmd = pmd_offset(pud, pvmw->address); in page_vma_mapped_walk() 189 pvmw->ptl = pmd_lock(mm, pvmw->pmd); in page_vma_mapped_walk() 224 if (!PageTransHuge(pvmw->page) || PageHuge(pvmw->page)) in page_vma_mapped_walk() 228 if (pvmw->address >= pvmw->vma->vm_end || in page_vma_mapped_walk() 230 __vma_address(pvmw->page, pvmw->vma) + in page_vma_mapped_walk() 242 pvmw->pte++; in page_vma_mapped_walk() [all …]
|
H A D | rmap.c | 786 address = pvmw.address; in page_referenced_one() 794 if (pvmw.pte) { in page_referenced_one() 796 pvmw.pte)) { in page_referenced_one() 810 pvmw.pmd)) in page_referenced_one() 928 address = pvmw.address; in page_mkclean_one() 929 if (pvmw.pte) { in page_mkclean_one() 931 pte_t *pte = pvmw.pte; in page_mkclean_one() 944 pmd_t *pmd = pvmw.pmd; in page_mkclean_one() 1480 address = pvmw.address; in try_to_unmap_one() 1522 pteval = ptep_get_and_clear(mm, pvmw.address, pvmw.pte); in try_to_unmap_one() [all …]
|
H A D | page_idle.c | 51 struct page_vma_mapped_walk pvmw = { in page_idle_clear_pte_refs_one() local 58 while (page_vma_mapped_walk(&pvmw)) { in page_idle_clear_pte_refs_one() 59 addr = pvmw.address; in page_idle_clear_pte_refs_one() 60 if (pvmw.pte) { in page_idle_clear_pte_refs_one() 65 if (ptep_clear_young_notify(vma, addr, pvmw.pte)) in page_idle_clear_pte_refs_one() 68 if (pmdp_clear_young_notify(vma, addr, pvmw.pmd)) in page_idle_clear_pte_refs_one()
|
H A D | ksm.c | 1044 if (pvmw.address == -EFAULT) in write_protect_page() 1050 pvmw.address, in write_protect_page() 1051 pvmw.address + PAGE_SIZE); in write_protect_page() 1054 if (!page_vma_mapped_walk(&pvmw)) in write_protect_page() 1059 if (pte_write(*pvmw.pte) || pte_dirty(*pvmw.pte) || in write_protect_page() 1060 (pte_protnone(*pvmw.pte) && pte_savedwrite(*pvmw.pte)) || in write_protect_page() 1080 entry = ptep_clear_flush(vma, pvmw.address, pvmw.pte); in write_protect_page() 1086 set_pte_at(mm, pvmw.address, pvmw.pte, entry); in write_protect_page() 1096 set_pte_at_notify(mm, pvmw.address, pvmw.pte, entry); in write_protect_page() 1098 *orig_pte = *pvmw.pte; in write_protect_page() [all …]
|
H A D | migrate.c | 177 struct page_vma_mapped_walk pvmw = { in remove_migration_pte() local 188 while (page_vma_mapped_walk(&pvmw)) { in remove_migration_pte() 192 new = page - pvmw.page->index + in remove_migration_pte() 197 if (!pvmw.pte) { in remove_migration_pte() 199 remove_migration_pmd(&pvmw, new); in remove_migration_pte() 206 if (pte_swp_soft_dirty(*pvmw.pte)) in remove_migration_pte() 221 if (pte_swp_soft_dirty(*pvmw.pte)) in remove_migration_pte() 223 if (pte_swp_uffd_wp(*pvmw.pte)) in remove_migration_pte() 231 set_huge_pte_at(vma->vm_mm, pvmw.address, pvmw.pte, pte); in remove_migration_pte() 239 set_pte_at(vma->vm_mm, pvmw.address, pvmw.pte, pte); in remove_migration_pte() [all …]
|
H A D | huge_memory.c | 3201 struct vm_area_struct *vma = pvmw->vma; in set_pmd_migration_entry() 3203 unsigned long address = pvmw->address; in set_pmd_migration_entry() 3208 if (!(pvmw->pmd && !pvmw->pte)) in set_pmd_migration_entry() 3219 set_pmd_at(mm, address, pvmw->pmd, pmdswp); in set_pmd_migration_entry() 3226 struct vm_area_struct *vma = pvmw->vma; in remove_migration_pmd() 3228 unsigned long address = pvmw->address; in remove_migration_pmd() 3233 if (!(pvmw->pmd && !pvmw->pte)) in remove_migration_pmd() 3236 entry = pmd_to_swp_entry(*pvmw->pmd); in remove_migration_pmd() 3239 if (pmd_swp_soft_dirty(*pvmw->pmd)) in remove_migration_pmd() 3249 set_pmd_at(mm, mmun_start, pvmw->pmd, pmde); in remove_migration_pmd() [all …]
|
/dports/multimedia/libv4l/linux-5.13-rc2/mm/ |
H A D | page_vma_mapped.c | 18 pvmw->pte = pte_offset_map(pvmw->pmd, pvmw->address); in map_pte() 50 pvmw->ptl = pte_lockptr(pvmw->vma->vm_mm, pvmw->pmd); in map_pte() 153 if (pvmw->pmd && !pvmw->pte) in page_vma_mapped_walk() 156 if (pvmw->pte) in page_vma_mapped_walk() 181 pvmw->pmd = pmd_offset(pud, pvmw->address); in page_vma_mapped_walk() 189 pvmw->ptl = pmd_lock(mm, pvmw->pmd); in page_vma_mapped_walk() 224 if (!PageTransHuge(pvmw->page) || PageHuge(pvmw->page)) in page_vma_mapped_walk() 228 if (pvmw->address >= pvmw->vma->vm_end || in page_vma_mapped_walk() 230 __vma_address(pvmw->page, pvmw->vma) + in page_vma_mapped_walk() 242 pvmw->pte++; in page_vma_mapped_walk() [all …]
|
H A D | rmap.c | 786 address = pvmw.address; in page_referenced_one() 794 if (pvmw.pte) { in page_referenced_one() 796 pvmw.pte)) { in page_referenced_one() 810 pvmw.pmd)) in page_referenced_one() 928 address = pvmw.address; in page_mkclean_one() 929 if (pvmw.pte) { in page_mkclean_one() 931 pte_t *pte = pvmw.pte; in page_mkclean_one() 944 pmd_t *pmd = pvmw.pmd; in page_mkclean_one() 1480 address = pvmw.address; in try_to_unmap_one() 1522 pteval = ptep_get_and_clear(mm, pvmw.address, pvmw.pte); in try_to_unmap_one() [all …]
|
H A D | page_idle.c | 51 struct page_vma_mapped_walk pvmw = { in page_idle_clear_pte_refs_one() local 58 while (page_vma_mapped_walk(&pvmw)) { in page_idle_clear_pte_refs_one() 59 addr = pvmw.address; in page_idle_clear_pte_refs_one() 60 if (pvmw.pte) { in page_idle_clear_pte_refs_one() 65 if (ptep_clear_young_notify(vma, addr, pvmw.pte)) in page_idle_clear_pte_refs_one() 68 if (pmdp_clear_young_notify(vma, addr, pvmw.pmd)) in page_idle_clear_pte_refs_one()
|
H A D | ksm.c | 1044 if (pvmw.address == -EFAULT) in write_protect_page() 1050 pvmw.address, in write_protect_page() 1051 pvmw.address + PAGE_SIZE); in write_protect_page() 1054 if (!page_vma_mapped_walk(&pvmw)) in write_protect_page() 1059 if (pte_write(*pvmw.pte) || pte_dirty(*pvmw.pte) || in write_protect_page() 1060 (pte_protnone(*pvmw.pte) && pte_savedwrite(*pvmw.pte)) || in write_protect_page() 1080 entry = ptep_clear_flush(vma, pvmw.address, pvmw.pte); in write_protect_page() 1086 set_pte_at(mm, pvmw.address, pvmw.pte, entry); in write_protect_page() 1096 set_pte_at_notify(mm, pvmw.address, pvmw.pte, entry); in write_protect_page() 1098 *orig_pte = *pvmw.pte; in write_protect_page() [all …]
|
H A D | migrate.c | 177 struct page_vma_mapped_walk pvmw = { in remove_migration_pte() local 188 while (page_vma_mapped_walk(&pvmw)) { in remove_migration_pte() 192 new = page - pvmw.page->index + in remove_migration_pte() 197 if (!pvmw.pte) { in remove_migration_pte() 199 remove_migration_pmd(&pvmw, new); in remove_migration_pte() 206 if (pte_swp_soft_dirty(*pvmw.pte)) in remove_migration_pte() 221 if (pte_swp_soft_dirty(*pvmw.pte)) in remove_migration_pte() 223 if (pte_swp_uffd_wp(*pvmw.pte)) in remove_migration_pte() 231 set_huge_pte_at(vma->vm_mm, pvmw.address, pvmw.pte, pte); in remove_migration_pte() 239 set_pte_at(vma->vm_mm, pvmw.address, pvmw.pte, pte); in remove_migration_pte() [all …]
|
H A D | huge_memory.c | 3201 struct vm_area_struct *vma = pvmw->vma; in set_pmd_migration_entry() 3203 unsigned long address = pvmw->address; in set_pmd_migration_entry() 3208 if (!(pvmw->pmd && !pvmw->pte)) in set_pmd_migration_entry() 3219 set_pmd_at(mm, address, pvmw->pmd, pmdswp); in set_pmd_migration_entry() 3226 struct vm_area_struct *vma = pvmw->vma; in remove_migration_pmd() 3228 unsigned long address = pvmw->address; in remove_migration_pmd() 3233 if (!(pvmw->pmd && !pvmw->pte)) in remove_migration_pmd() 3236 entry = pmd_to_swp_entry(*pvmw->pmd); in remove_migration_pmd() 3239 if (pmd_swp_soft_dirty(*pvmw->pmd)) in remove_migration_pmd() 3249 set_pmd_at(mm, mmun_start, pvmw->pmd, pmde); in remove_migration_pmd() [all …]
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/mm/ |
H A D | page_vma_mapped.c | 18 pvmw->pte = pte_offset_map(pvmw->pmd, pvmw->address); in map_pte() 50 pvmw->ptl = pte_lockptr(pvmw->vma->vm_mm, pvmw->pmd); in map_pte() 153 if (pvmw->pmd && !pvmw->pte) in page_vma_mapped_walk() 156 if (pvmw->pte) in page_vma_mapped_walk() 181 pvmw->pmd = pmd_offset(pud, pvmw->address); in page_vma_mapped_walk() 189 pvmw->ptl = pmd_lock(mm, pvmw->pmd); in page_vma_mapped_walk() 224 if (!PageTransHuge(pvmw->page) || PageHuge(pvmw->page)) in page_vma_mapped_walk() 228 if (pvmw->address >= pvmw->vma->vm_end || in page_vma_mapped_walk() 230 __vma_address(pvmw->page, pvmw->vma) + in page_vma_mapped_walk() 242 pvmw->pte++; in page_vma_mapped_walk() [all …]
|
H A D | rmap.c | 786 address = pvmw.address; in page_referenced_one() 794 if (pvmw.pte) { in page_referenced_one() 796 pvmw.pte)) { in page_referenced_one() 810 pvmw.pmd)) in page_referenced_one() 928 address = pvmw.address; in page_mkclean_one() 929 if (pvmw.pte) { in page_mkclean_one() 931 pte_t *pte = pvmw.pte; in page_mkclean_one() 944 pmd_t *pmd = pvmw.pmd; in page_mkclean_one() 1480 address = pvmw.address; in try_to_unmap_one() 1522 pteval = ptep_get_and_clear(mm, pvmw.address, pvmw.pte); in try_to_unmap_one() [all …]
|
H A D | page_idle.c | 51 struct page_vma_mapped_walk pvmw = { in page_idle_clear_pte_refs_one() local 58 while (page_vma_mapped_walk(&pvmw)) { in page_idle_clear_pte_refs_one() 59 addr = pvmw.address; in page_idle_clear_pte_refs_one() 60 if (pvmw.pte) { in page_idle_clear_pte_refs_one() 65 if (ptep_clear_young_notify(vma, addr, pvmw.pte)) in page_idle_clear_pte_refs_one() 68 if (pmdp_clear_young_notify(vma, addr, pvmw.pmd)) in page_idle_clear_pte_refs_one()
|
H A D | ksm.c | 1044 if (pvmw.address == -EFAULT) in write_protect_page() 1050 pvmw.address, in write_protect_page() 1051 pvmw.address + PAGE_SIZE); in write_protect_page() 1054 if (!page_vma_mapped_walk(&pvmw)) in write_protect_page() 1059 if (pte_write(*pvmw.pte) || pte_dirty(*pvmw.pte) || in write_protect_page() 1060 (pte_protnone(*pvmw.pte) && pte_savedwrite(*pvmw.pte)) || in write_protect_page() 1080 entry = ptep_clear_flush(vma, pvmw.address, pvmw.pte); in write_protect_page() 1086 set_pte_at(mm, pvmw.address, pvmw.pte, entry); in write_protect_page() 1096 set_pte_at_notify(mm, pvmw.address, pvmw.pte, entry); in write_protect_page() 1098 *orig_pte = *pvmw.pte; in write_protect_page() [all …]
|
H A D | migrate.c | 177 struct page_vma_mapped_walk pvmw = { in remove_migration_pte() local 188 while (page_vma_mapped_walk(&pvmw)) { in remove_migration_pte() 192 new = page - pvmw.page->index + in remove_migration_pte() 197 if (!pvmw.pte) { in remove_migration_pte() 199 remove_migration_pmd(&pvmw, new); in remove_migration_pte() 206 if (pte_swp_soft_dirty(*pvmw.pte)) in remove_migration_pte() 221 if (pte_swp_soft_dirty(*pvmw.pte)) in remove_migration_pte() 223 if (pte_swp_uffd_wp(*pvmw.pte)) in remove_migration_pte() 231 set_huge_pte_at(vma->vm_mm, pvmw.address, pvmw.pte, pte); in remove_migration_pte() 239 set_pte_at(vma->vm_mm, pvmw.address, pvmw.pte, pte); in remove_migration_pte() [all …]
|
H A D | huge_memory.c | 3201 struct vm_area_struct *vma = pvmw->vma; in set_pmd_migration_entry() 3203 unsigned long address = pvmw->address; in set_pmd_migration_entry() 3208 if (!(pvmw->pmd && !pvmw->pte)) in set_pmd_migration_entry() 3219 set_pmd_at(mm, address, pvmw->pmd, pmdswp); in set_pmd_migration_entry() 3226 struct vm_area_struct *vma = pvmw->vma; in remove_migration_pmd() 3228 unsigned long address = pvmw->address; in remove_migration_pmd() 3233 if (!(pvmw->pmd && !pvmw->pte)) in remove_migration_pmd() 3236 entry = pmd_to_swp_entry(*pvmw->pmd); in remove_migration_pmd() 3239 if (pmd_swp_soft_dirty(*pvmw->pmd)) in remove_migration_pmd() 3249 set_pmd_at(mm, mmun_start, pvmw->pmd, pmde); in remove_migration_pmd() [all …]
|
/dports/multimedia/libv4l/linux-5.13-rc2/include/linux/ |
H A D | rmap.h | 214 static inline void page_vma_mapped_walk_done(struct page_vma_mapped_walk *pvmw) in page_vma_mapped_walk_done() argument 217 if (pvmw->pte && !PageHuge(pvmw->page)) in page_vma_mapped_walk_done() 218 pte_unmap(pvmw->pte); in page_vma_mapped_walk_done() 219 if (pvmw->ptl) in page_vma_mapped_walk_done() 220 spin_unlock(pvmw->ptl); in page_vma_mapped_walk_done() 223 bool page_vma_mapped_walk(struct page_vma_mapped_walk *pvmw);
|
H A D | swapops.h | 247 extern void set_pmd_migration_entry(struct page_vma_mapped_walk *pvmw, 250 extern void remove_migration_pmd(struct page_vma_mapped_walk *pvmw, 278 static inline void set_pmd_migration_entry(struct page_vma_mapped_walk *pvmw, in set_pmd_migration_entry() argument 284 static inline void remove_migration_pmd(struct page_vma_mapped_walk *pvmw, in remove_migration_pmd() argument
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/include/linux/ |
H A D | rmap.h | 214 static inline void page_vma_mapped_walk_done(struct page_vma_mapped_walk *pvmw) in page_vma_mapped_walk_done() argument 217 if (pvmw->pte && !PageHuge(pvmw->page)) in page_vma_mapped_walk_done() 218 pte_unmap(pvmw->pte); in page_vma_mapped_walk_done() 219 if (pvmw->ptl) in page_vma_mapped_walk_done() 220 spin_unlock(pvmw->ptl); in page_vma_mapped_walk_done() 223 bool page_vma_mapped_walk(struct page_vma_mapped_walk *pvmw);
|
H A D | swapops.h | 247 extern void set_pmd_migration_entry(struct page_vma_mapped_walk *pvmw, 250 extern void remove_migration_pmd(struct page_vma_mapped_walk *pvmw, 278 static inline void set_pmd_migration_entry(struct page_vma_mapped_walk *pvmw, in set_pmd_migration_entry() argument 284 static inline void remove_migration_pmd(struct page_vma_mapped_walk *pvmw, in remove_migration_pmd() argument
|
/dports/multimedia/v4l-utils/linux-5.13-rc2/include/linux/ |
H A D | rmap.h | 214 static inline void page_vma_mapped_walk_done(struct page_vma_mapped_walk *pvmw) in page_vma_mapped_walk_done() argument 217 if (pvmw->pte && !PageHuge(pvmw->page)) in page_vma_mapped_walk_done() 218 pte_unmap(pvmw->pte); in page_vma_mapped_walk_done() 219 if (pvmw->ptl) in page_vma_mapped_walk_done() 220 spin_unlock(pvmw->ptl); in page_vma_mapped_walk_done() 223 bool page_vma_mapped_walk(struct page_vma_mapped_walk *pvmw);
|
H A D | swapops.h | 247 extern void set_pmd_migration_entry(struct page_vma_mapped_walk *pvmw, 250 extern void remove_migration_pmd(struct page_vma_mapped_walk *pvmw, 278 static inline void set_pmd_migration_entry(struct page_vma_mapped_walk *pvmw, in set_pmd_migration_entry() argument 284 static inline void remove_migration_pmd(struct page_vma_mapped_walk *pvmw, in remove_migration_pmd() argument
|
/dports/multimedia/v4l_compat/linux-5.13-rc2/kernel/events/ |
H A D | uprobes.c | 158 struct page_vma_mapped_walk pvmw = { in __replace_page() local 180 if (!page_vma_mapped_walk(&pvmw)) in __replace_page() 182 VM_BUG_ON_PAGE(addr != pvmw.address, old_page); in __replace_page() 197 flush_cache_page(vma, addr, pte_pfn(*pvmw.pte)); in __replace_page() 198 ptep_clear_flush_notify(vma, addr, pvmw.pte); in __replace_page() 200 set_pte_at_notify(mm, addr, pvmw.pte, in __replace_page() 206 page_vma_mapped_walk_done(&pvmw); in __replace_page()
|