/linux/fs/nilfs2/ |
H A D | cpfile.c | 71 void *kaddr, in nilfs_cpfile_block_add_valid_checkpoints() argument 210 void *kaddr; in nilfs_cpfile_read_checkpoint() local 278 void *kaddr; in nilfs_cpfile_create_checkpoint() local 354 void *kaddr; in nilfs_cpfile_finalize_checkpoint() local 433 void *kaddr; in nilfs_cpfile_delete_checkpoints() local 545 void *kaddr; in nilfs_cpfile_do_get_cpinfo() local 598 void *kaddr; in nilfs_cpfile_do_get_ssinfo() local 756 void *kaddr; in nilfs_cpfile_set_snapshot() local 874 void *kaddr; in nilfs_cpfile_clear_snapshot() local 991 void *kaddr; in nilfs_cpfile_is_snapshot() local [all …]
|
H A D | dir.c | 186 void *kaddr; in nilfs_get_folio() local 198 return kaddr; in nilfs_get_folio() 273 char *kaddr, *limit; in nilfs_readdir() local 278 if (IS_ERR(kaddr)) { in nilfs_readdir() 344 if (!IS_ERR(kaddr)) { in nilfs_find_entry() 446 if (IS_ERR(kaddr)) in nilfs_add_link() 565 void *kaddr; in nilfs_make_empty() local 590 kunmap_local(kaddr); in nilfs_make_empty() 603 char *kaddr; in nilfs_empty_dir() local 610 if (IS_ERR(kaddr)) in nilfs_empty_dir() [all …]
|
H A D | sufile.c | 108 void *kaddr; in nilfs_sufile_mod_counter() local 309 void *kaddr; in nilfs_sufile_alloc() local 407 void *kaddr; in nilfs_sufile_do_cancel_free() local 432 void *kaddr; in nilfs_sufile_do_scrap() local 463 void *kaddr; in nilfs_sufile_do_free() local 503 void *kaddr; in nilfs_sufile_mark_dirty() local 557 void *kaddr; in nilfs_sufile_set_segment_usage() local 608 void *kaddr; in nilfs_sufile_get_stat() local 640 void *kaddr; in nilfs_sufile_do_set_error() local 689 void *kaddr; in nilfs_sufile_truncate_range() local [all …]
|
H A D | dat.c | 92 void *kaddr; in nilfs_dat_commit_alloc() local 100 kunmap_local(kaddr); in nilfs_dat_commit_alloc() 116 void *kaddr; in nilfs_dat_commit_free() local 124 kunmap_local(kaddr); in nilfs_dat_commit_free() 146 void *kaddr; in nilfs_dat_commit_start() local 163 void *kaddr; in nilfs_dat_prepare_end() local 203 void *kaddr; in nilfs_dat_commit_end() local 228 void *kaddr; in nilfs_dat_abort_end() local 342 void *kaddr; in nilfs_dat_move() local 413 void *kaddr; in nilfs_dat_translate() local [all …]
|
/linux/arch/arm64/kernel/ |
H A D | paravirt.c | 38 struct pvclock_vcpu_stolen_time __rcu *kaddr; member 67 kaddr = rcu_dereference(reg->kaddr); in para_steal_clock() 68 if (!kaddr) { in para_steal_clock() 84 if (!reg->kaddr) in stolen_time_cpu_down_prepare() 87 kaddr = rcu_replace_pointer(reg->kaddr, NULL, true); in stolen_time_cpu_down_prepare() 89 memunmap(kaddr); in stolen_time_cpu_down_prepare() 107 kaddr = memremap(res.a0, in stolen_time_cpu_online() 111 rcu_assign_pointer(reg->kaddr, kaddr); in stolen_time_cpu_online() 113 if (!reg->kaddr) { in stolen_time_cpu_online() 118 if (le32_to_cpu(kaddr->revision) != 0 || in stolen_time_cpu_online() [all …]
|
/linux/fs/btrfs/ |
H A D | accessors.c | 31 token->kaddr = folio_address(eb->folios[0]); in btrfs_init_map_token() 74 ASSERT(token->kaddr); \ 85 memcpy(lebytes, token->kaddr + oil, part); \ 106 return get_unaligned_le##bits(kaddr + oil); \ 108 memcpy(lebytes, kaddr + oil, part); \ 109 kaddr = folio_address(eb->folios[idx + 1]); \ 110 memcpy(lebytes + part, kaddr, size - part); \ 128 ASSERT(token->kaddr); \ 143 memcpy(token->kaddr + oil, lebytes, part); \ 164 put_unaligned_le##bits(val, kaddr + oil); \ [all …]
|
H A D | lzo.c | 141 char *kaddr; in copy_compressed_data_to_page() local 161 kaddr = kmap_local_folio(cur_folio, 0); in copy_compressed_data_to_page() 173 kunmap_local(kaddr); in copy_compressed_data_to_page() 186 kaddr = kmap_local_folio(cur_folio, 0); in copy_compressed_data_to_page() 188 memcpy(kaddr + offset_in_page(*cur_out), in copy_compressed_data_to_page() 208 kunmap_local(kaddr); in copy_compressed_data_to_page() 336 char *kaddr; in lzo_decompress_bio() local 346 len_in = read_compress_length(kaddr); in lzo_decompress_bio() 347 kunmap_local(kaddr); in lzo_decompress_bio() 381 kaddr = kmap_local_folio(cur_folio, 0); in lzo_decompress_bio() [all …]
|
/linux/fs/sysv/ |
H A D | dir.c | 88 char *kaddr, *limit; in sysv_readdir() local 93 if (IS_ERR(kaddr)) in sysv_readdir() 110 unmap_and_put_page(page, kaddr); in sysv_readdir() 159 if (!IS_ERR(kaddr)) { in sysv_find_entry() 193 char *kaddr; in sysv_add_link() local 200 if (IS_ERR(kaddr)) in sysv_add_link() 201 return PTR_ERR(kaddr); in sysv_add_link() 204 while ((char *)de <= kaddr) { in sysv_add_link() 231 unmap_and_put_page(page, kaddr); in sysv_add_link() 297 char *kaddr; in sysv_empty_dir() local [all …]
|
/linux/fs/ext2/ |
H A D | dir.c | 195 void *kaddr; in ext2_get_folio() local 205 return kaddr; in ext2_get_folio() 282 if (IS_ERR(kaddr)) { in ext2_readdir() 287 return PTR_ERR(kaddr); in ext2_readdir() 363 if (IS_ERR(kaddr)) in ext2_find_entry() 501 if (IS_ERR(kaddr)) in ext2_add_link() 576 char *kaddr; in ext2_delete_entry() local 624 void *kaddr; in ext2_make_empty() local 649 kunmap_local(kaddr); in ext2_make_empty() 663 char *kaddr; in ext2_empty_dir() local [all …]
|
/linux/fs/minix/ |
H A D | dir.c | 101 char *p, *kaddr, *limit; in minix_readdir() local 105 if (IS_ERR(kaddr)) in minix_readdir() 107 p = kaddr+offset; in minix_readdir() 169 char *kaddr, *limit; in minix_find_entry() local 172 if (IS_ERR(kaddr)) in minix_find_entry() 210 char *kaddr, *p; in minix_add_link() local 227 if (IS_ERR(kaddr)) in minix_add_link() 315 char *kaddr; in minix_make_empty() local 346 kunmap_local(kaddr); in minix_make_empty() 363 char *name, *kaddr; in minix_empty_dir() local [all …]
|
/linux/arch/loongarch/include/asm/ |
H A D | page.h | 78 struct page *dmw_virt_to_page(unsigned long kaddr); 79 struct page *tlb_virt_to_page(unsigned long kaddr); 90 #define virt_to_page(kaddr) phys_to_page(__pa(kaddr)) argument 102 #define virt_to_page(kaddr) \ argument 104 (likely((unsigned long)kaddr < vm_map_base)) ? \ 105 dmw_virt_to_page((unsigned long)kaddr) : tlb_virt_to_page((unsigned long)kaddr);\ 111 #define virt_to_pfn(kaddr) page_to_pfn(virt_to_page(kaddr)) argument 113 extern int __virt_addr_valid(volatile void *kaddr); 114 #define virt_addr_valid(kaddr) __virt_addr_valid((volatile void *)(kaddr)) argument
|
H A D | kfence.h | 19 char *kaddr, *vaddr; in arch_kfence_init_pool() local 39 kaddr = kfence_pool; in arch_kfence_init_pool() 41 while (kaddr < kfence_pool + KFENCE_POOL_SIZE) { in arch_kfence_init_pool() 42 set_page_address(virt_to_page(kaddr), vaddr); in arch_kfence_init_pool() 43 kaddr += PAGE_SIZE; in arch_kfence_init_pool()
|
/linux/arch/m68k/sun3/ |
H A D | dvma.c | 23 static unsigned long dvma_page(unsigned long kaddr, unsigned long vaddr) in dvma_page() argument 29 j = *(volatile unsigned long *)kaddr; in dvma_page() 30 *(volatile unsigned long *)kaddr = j; in dvma_page() 32 ptep = pfn_pte(virt_to_pfn((void *)kaddr), PAGE_KERNEL); in dvma_page() 40 return (vaddr + (kaddr & ~PAGE_MASK)); in dvma_page() 44 int dvma_map_iommu(unsigned long kaddr, unsigned long baddr, in dvma_map_iommu() argument 56 dvma_page(kaddr, vaddr); in dvma_map_iommu() 57 kaddr += PAGE_SIZE; in dvma_map_iommu()
|
H A D | sun3dvma.c | 266 unsigned long dvma_map_align(unsigned long kaddr, int len, int align) in dvma_map_align() argument 275 if(!kaddr || !len) { in dvma_map_align() 281 pr_debug("dvma_map request %08x bytes from %08lx\n", len, kaddr); in dvma_map_align() 282 off = kaddr & ~DVMA_PAGE_MASK; in dvma_map_align() 283 kaddr &= PAGE_MASK; in dvma_map_align() 295 if(!dvma_map_iommu(kaddr, baddr, len)) in dvma_map_align() 323 unsigned long kaddr; in dvma_malloc_align() local 337 free_pages(kaddr, get_order(len)); in dvma_malloc_align() 343 if(dvma_map_cpu(kaddr, vaddr, len) < 0) { in dvma_malloc_align() 345 free_pages(kaddr, get_order(len)); in dvma_malloc_align() [all …]
|
/linux/drivers/acpi/ |
H A D | nvs.c | 78 void *kaddr; member 138 if (entry->kaddr) { in suspend_nvs_free() 140 iounmap(entry->kaddr); in suspend_nvs_free() 143 acpi_os_unmap_iomem(entry->kaddr, in suspend_nvs_free() 146 entry->kaddr = NULL; in suspend_nvs_free() 182 entry->kaddr = acpi_os_get_iomem(phys, size); in suspend_nvs_save() 183 if (!entry->kaddr) { in suspend_nvs_save() 184 entry->kaddr = acpi_os_ioremap(phys, size); in suspend_nvs_save() 185 entry->unmap = !!entry->kaddr; in suspend_nvs_save() 187 if (!entry->kaddr) { in suspend_nvs_save() [all …]
|
/linux/fs/erofs/ |
H A D | inode.c | 19 void *kaddr; in erofs_read_inode() local 30 if (IS_ERR(kaddr)) { in erofs_read_inode() 32 vi->nid, PTR_ERR(kaddr)); in erofs_read_inode() 33 return kaddr; in erofs_read_inode() 36 dic = kaddr + *ofs; in erofs_read_inode() 71 if (IS_ERR(kaddr)) { in erofs_read_inode() 75 return kaddr; in erofs_read_inode() 168 return kaddr; in erofs_read_inode() 217 void *kaddr; in erofs_fill_inode() local 225 if (IS_ERR(kaddr)) in erofs_fill_inode() [all …]
|
H A D | xattr.c | 15 void *kaddr; member 85 if (IS_ERR(it.kaddr)) { in erofs_init_inode_xattrs() 86 ret = PTR_ERR(it.kaddr); in erofs_init_inode_xattrs() 107 if (IS_ERR(it.kaddr)) { in erofs_init_inode_xattrs() 110 ret = PTR_ERR(it.kaddr); in erofs_init_inode_xattrs() 190 if (IS_ERR(it->kaddr)) in erofs_xattr_copy_to_buffer() 191 return PTR_ERR(it->kaddr); in erofs_xattr_copy_to_buffer() 296 if (IS_ERR(it->kaddr)) in erofs_getxattr_foreach() 297 return PTR_ERR(it->kaddr); in erofs_getxattr_foreach() 341 if (IS_ERR(it->kaddr)) in erofs_xattr_iter_inline() [all …]
|
/linux/fs/ufs/ |
H A D | dir.c | 120 char *kaddr = page_address(page); in ufs_check_page() local 283 char *kaddr; in ufs_find_entry() local 286 kaddr = page_address(page); in ufs_find_entry() 324 char *kaddr; in ufs_add_link() local 343 kaddr = page_address(page); in ufs_add_link() 346 kaddr += PAGE_SIZE - reclen; in ufs_add_link() 347 while ((char *)de <= kaddr) { in ufs_add_link() 448 char *kaddr, *limit; in ufs_readdir() local 460 kaddr = page_address(page); in ufs_readdir() 607 char *kaddr; in ufs_empty_dir() local [all …]
|
/linux/arch/m68k/sun3x/ |
H A D | dvma.c | 78 inline int dvma_map_cpu(unsigned long kaddr, in dvma_map_cpu() argument 87 kaddr &= PAGE_MASK; in dvma_map_cpu() 92 pr_debug("dvma: mapping kern %08lx to virt %08lx\n", kaddr, vaddr); in dvma_map_cpu() 127 __pa(kaddr), vaddr); in dvma_map_cpu() 128 set_pte(pte, pfn_pte(virt_to_pfn((void *)kaddr), in dvma_map_cpu() 131 kaddr += PAGE_SIZE; in dvma_map_cpu() 146 int dvma_map_iommu(unsigned long kaddr, unsigned long baddr, int len) in dvma_map_iommu() argument 162 dvma_entry_set(index, __pa(kaddr)); in dvma_map_iommu() 167 kaddr += DVMA_PAGE_SIZE; in dvma_map_iommu()
|
/linux/arch/arm/include/asm/ |
H A D | tlbflush.h | 477 tlb_op(TLB_V4_U_PAGE, "c8, c7, 1", kaddr); in __local_flush_tlb_kernel_page() 478 tlb_op(TLB_V4_D_PAGE, "c8, c6, 1", kaddr); in __local_flush_tlb_kernel_page() 479 tlb_op(TLB_V4_I_PAGE, "c8, c5, 1", kaddr); in __local_flush_tlb_kernel_page() 483 tlb_op(TLB_V6_U_PAGE, "c8, c7, 1", kaddr); in __local_flush_tlb_kernel_page() 484 tlb_op(TLB_V6_D_PAGE, "c8, c6, 1", kaddr); in __local_flush_tlb_kernel_page() 485 tlb_op(TLB_V6_I_PAGE, "c8, c5, 1", kaddr); in __local_flush_tlb_kernel_page() 492 kaddr &= PAGE_MASK; in local_flush_tlb_kernel_page() 497 __local_flush_tlb_kernel_page(kaddr); in local_flush_tlb_kernel_page() 498 tlb_op(TLB_V7_UIS_PAGE, "c8, c7, 1", kaddr); in local_flush_tlb_kernel_page() 510 kaddr &= PAGE_MASK; in __flush_tlb_kernel_page() [all …]
|
/linux/include/linux/ |
H A D | highmem.h | 239 void *kaddr = kmap_local_page(page); in clear_highpage() local 240 clear_page(kaddr); in clear_highpage() 241 kunmap_local(kaddr); in clear_highpage() 248 clear_page(kasan_reset_tag(kaddr)); in clear_highpage_kasan_tagged() 249 kunmap_local(kaddr); in clear_highpage_kasan_tagged() 283 kunmap_local(kaddr); in zero_user_segments() 515 size_t offset, void *kaddr) in folio_zero_tail() argument 523 memset(kaddr, 0, max); in folio_zero_tail() 524 kunmap_local(kaddr); in folio_zero_tail() 532 memset(kaddr, 0, len); in folio_zero_tail() [all …]
|
/linux/arch/hexagon/include/asm/ |
H A D | page.h | 92 #define virt_to_page(kaddr) pfn_to_page(PFN_DOWN(__pa(kaddr))) argument 97 #define virt_addr_valid(kaddr) pfn_valid(__pa(kaddr) >> PAGE_SHIFT) argument 127 static inline unsigned long virt_to_pfn(const void *kaddr) in virt_to_pfn() argument 129 return __pa(kaddr) >> PAGE_SHIFT; in virt_to_pfn()
|
/linux/arch/arc/include/asm/ |
H A D | page.h | 123 static inline unsigned long virt_to_pfn(const void *kaddr) in virt_to_pfn() argument 125 return __pa(kaddr) >> PAGE_SHIFT; in virt_to_pfn() 128 #define virt_to_page(kaddr) pfn_to_page(virt_to_pfn(kaddr)) argument 129 #define virt_addr_valid(kaddr) pfn_valid(virt_to_pfn(kaddr)) argument
|
/linux/tools/testing/nvdimm/ |
H A D | pmem-dax.c | 12 long nr_pages, enum dax_access_mode mode, void **kaddr, in __pmem_direct_access() argument 28 if (kaddr) in __pmem_direct_access() 29 *kaddr = pmem->virt_addr + offset; in __pmem_direct_access() 39 if (kaddr) in __pmem_direct_access() 40 *kaddr = pmem->virt_addr + offset; in __pmem_direct_access()
|
/linux/arch/m68k/include/asm/ |
H A D | page_no.h | 22 static inline unsigned long virt_to_pfn(const void *kaddr) in virt_to_pfn() argument 24 return __pa(kaddr) >> PAGE_SHIFT; in virt_to_pfn() 35 #define virt_addr_valid(kaddr) (((unsigned long)(kaddr) >= PAGE_OFFSET) && \ argument 36 ((unsigned long)(kaddr) < memory_end))
|