/linux/arch/s390/mm/ |
H A D | gmap.c | 2522 unsigned long end, struct mm_walk *walk) in thp_split_walk_pmd_entry() argument 2524 struct vm_area_struct *vma = walk->vma; in thp_split_walk_pmd_entry() 2575 unsigned long end, struct mm_walk *walk) in find_zeropage_pte_entry() argument 2577 unsigned long *found_addr = walk->private; in find_zeropage_pte_entry() 2587 if (!is_cow_mapping(walk->vma->vm_flags)) in find_zeropage_pte_entry() 2711 unsigned long next, struct mm_walk *walk) in __s390_enable_skey_pte() argument 2714 ptep_zap_key(walk->mm, addr, pte); in __s390_enable_skey_pte() 2724 unsigned long next, struct mm_walk *walk) in __s390_enable_skey_pmd() argument 2732 struct mm_walk *walk) in __s390_enable_skey_hugetlb() argument 2792 ptep_zap_unused(walk->mm, addr, pte, 1); in __s390_reset_cmma() [all …]
|
/linux/scripts/ |
H A D | leaking_addresses.pl | 185 walk(@DIRS); 536 sub walk subroutine
|
/linux/tools/testing/selftests/net/ |
H A D | psock_tpacket.c | 71 void (*walk)(int sock, struct ring *ring); member 641 ring->walk = walk_v1_v2; in __v1_v2_fill() 662 ring->walk = walk_v3; in __v3_fill() 748 ring->walk(sock, ring); in walk_ring()
|
/linux/drivers/crypto/ |
H A D | hifn_795x.c | 624 struct hifn_cipher_walk walk; member 1342 t = &rctx->walk.cache[0]; in hifn_setup_dma() 1345 if (t->length && rctx->walk.flags & ASYNC_FLAGS_MISALIGNED) { in hifn_setup_dma() 1537 rctx->walk.flags = 0; in hifn_setup_session() 1545 rctx->walk.flags |= ASYNC_FLAGS_MISALIGNED; in hifn_setup_session() 1551 if (rctx->walk.flags & ASYNC_FLAGS_MISALIGNED) { in hifn_setup_session() 1552 err = hifn_cipher_walk_init(&rctx->walk, idx, GFP_ATOMIC); in hifn_setup_session() 1557 sg_num = hifn_cipher_walk(req, &rctx->walk); in hifn_setup_session() 1670 if (rctx->walk.flags & ASYNC_FLAGS_MISALIGNED) { in hifn_process_ready() 1677 t = &rctx->walk.cache[idx]; in hifn_process_ready() [all …]
|
/linux/net/sched/ |
H A D | sch_ingress.c | 153 .walk = ingress_walk, 328 .walk = ingress_walk,
|
H A D | cls_cgroup.c | 208 .walk = cls_cgroup_walk,
|
H A D | sch_mq.c | 260 .walk = mq_walk,
|
H A D | sch_skbprio.c | 280 .walk = skbprio_walk,
|
/linux/drivers/vfio/pci/ |
H A D | vfio_pci_core.c | 864 struct vfio_pci_walk_info *walk = data; in vfio_pci_walk_wrapper() local 866 if (!walk->slot || vfio_pci_dev_below_slot(pdev, walk->pdev->slot)) in vfio_pci_walk_wrapper() 867 walk->ret = walk->fn(pdev, walk->data); in vfio_pci_walk_wrapper() 869 return walk->ret; in vfio_pci_walk_wrapper() 877 struct vfio_pci_walk_info walk = { in vfio_pci_for_each_slot_or_bus() local 881 pci_walk_bus(pdev->bus, vfio_pci_walk_wrapper, &walk); in vfio_pci_for_each_slot_or_bus() 883 return walk.ret; in vfio_pci_for_each_slot_or_bus()
|
/linux/Documentation/arch/arm64/ |
H A D | hugetlbpage.rst | 20 mappings reduce the depth of page table walk needed to translate hugepage
|
/linux/mm/ |
H A D | mlock.c | 364 unsigned long end, struct mm_walk *walk) in mlock_pte_range() argument 367 struct vm_area_struct *vma = walk->vma; in mlock_pte_range() 391 walk->action = ACTION_AGAIN; in mlock_pte_range()
|
H A D | memory-failure.c | 800 unsigned long end, struct mm_walk *walk) in hwpoison_pte_range() argument 802 struct hwpoison_walk *hwp = walk->private; in hwpoison_pte_range() 807 ptl = pmd_trans_huge_lock(pmdp, walk->vma); in hwpoison_pte_range() 814 mapped_pte = ptep = pte_offset_map_lock(walk->vma->vm_mm, pmdp, in hwpoison_pte_range() 834 struct mm_walk *walk) in hwpoison_hugetlb_range() argument 836 struct hwpoison_walk *hwp = walk->private; in hwpoison_hugetlb_range() 838 struct hstate *h = hstate_vma(walk->vma); in hwpoison_hugetlb_range()
|
/linux/scripts/dtc/ |
H A D | dt-extract-compatibles | 85 for path, dirs, files in os.walk(root):
|
/linux/include/net/ |
H A D | xfrm.h | 511 struct xfrm_policy_walk_entry walk; member 541 struct xfrm_policy_walk_entry walk; member 1565 void xfrm_state_walk_init(struct xfrm_state_walk *walk, u8 proto, 1567 int xfrm_state_walk(struct net *net, struct xfrm_state_walk *walk, 1569 void xfrm_state_walk_done(struct xfrm_state_walk *walk, struct net *net); 1738 void xfrm_policy_walk_init(struct xfrm_policy_walk *walk, u8 type); 1739 int xfrm_policy_walk(struct net *net, struct xfrm_policy_walk *walk, 1742 void xfrm_policy_walk_done(struct xfrm_policy_walk *walk, struct net *net);
|
/linux/tools/perf/util/ |
H A D | bpf-filter.l | 141 walk { return constant(PERF_MEM_TLB_WK); }
|
/linux/scripts/clang-tools/ |
H A D | gen_compile_commands.py | 88 for dirpath, dirnames, filenames in os.walk(directory, topdown=True):
|
/linux/Documentation/mm/ |
H A D | multigen_lru.rst | 30 accessed bit. A rmap walk targets a single page and does not try to 31 profit from discovering a young PTE. A page table walk can sweep all 181 Rmap/PT walk feedback 261 * Bloom filters for rmap/PT walk feedback
|
/linux/Documentation/filesystems/ |
H A D | porting.rst | 399 via rcu-walk path walk (basically, if the file can have had a path name in the 411 vfs now tries to do path walking in "rcu-walk mode", which avoids 415 filesystem callbacks, the vfs drops out of rcu-walk mode before the fs call, so 417 the benefits of rcu-walk mode. We will begin to add filesystem callbacks that 418 are rcu-walk aware, shown below. Filesystems should take advantage of this 426 the filesystem provides it), which requires dropping out of rcu-walk mode. This 427 may now be called in rcu-walk mode (nd->flags & LOOKUP_RCU). -ECHILD should be 428 returned if the filesystem cannot handle rcu-walk. See 432 directory inodes on the way down a path walk (to check for exec permission). It 433 must now be rcu-walk aware (mask & MAY_NOT_BLOCK). See
|
/linux/net/netfilter/ |
H A D | nft_set_hash.c | 752 .walk = nft_rhash_walk, 771 .walk = nft_hash_walk, 790 .walk = nft_hash_walk,
|
H A D | nft_lookup.c | 225 priv->set->ops->walk(ctx, priv->set, &iter); in nft_lookup_validate()
|
/linux/net/xfrm/ |
H A D | xfrm_user.c | 1340 xfrm_state_walk_done(walk, net); in xfrm_dump_sa_done() 1388 xfrm_state_walk_init(walk, proto, filter); in xfrm_dump_sa() 1392 (void) xfrm_state_walk(net, walk, dump_one_state, &info); in xfrm_dump_sa() 2066 xp->walk.dead = 1; in xfrm_policy_construct() 2243 struct xfrm_policy_walk *walk = (struct xfrm_policy_walk *)cb->args; in xfrm_dump_policy_done() local 2246 xfrm_policy_walk_done(walk, net); in xfrm_dump_policy_done() 2252 struct xfrm_policy_walk *walk = (struct xfrm_policy_walk *)cb->args; in xfrm_dump_policy_start() local 2254 BUILD_BUG_ON(sizeof(*walk) > sizeof(cb->args)); in xfrm_dump_policy_start() 2256 xfrm_policy_walk_init(walk, XFRM_POLICY_TYPE_ANY); in xfrm_dump_policy_start() 2271 (void) xfrm_policy_walk(net, walk, dump_one_policy, &info); in xfrm_dump_policy() [all …]
|
/linux/scripts/gdb/linux/ |
H A D | symbols.py | 78 for root, dirs, files in os.walk(path):
|
/linux/Documentation/networking/ |
H A D | fib_trie.rst | 80 it is run to optimize and reorganize. It will walk the trie upwards 107 slower than the corresponding fib_hash function, as we have to walk the
|
/linux/fs/jffs2/ |
H A D | README.Locking | 88 erase_completion_lock. So you can walk the list only while holding the 89 erase_completion_lock, and can drop the lock temporarily mid-walk as
|
/linux/fs/fat/ |
H A D | inode.c | 468 unsigned char exe_extensions[] = "EXECOMBAT", *walk; in is_exec() local 470 for (walk = exe_extensions; *walk; walk += 3) in is_exec() 471 if (!strncmp(extension, walk, 3)) in is_exec()
|