1 /*- 2 * Copyright (c) 2010 Isilon Systems, Inc. 3 * Copyright (c) 2010 iX Systems, Inc. 4 * Copyright (c) 2010 Panasas, Inc. 5 * Copyright (c) 2013, 2014 Mellanox Technologies, Ltd. 6 * Copyright (c) 2015 Matthew Dillon <dillon@backplane.com> 7 * Copyright (c) 2015-2020 François Tigeot <ftigeot@wolfpond.org> 8 * All rights reserved. 9 * 10 * Redistribution and use in source and binary forms, with or without 11 * modification, are permitted provided that the following conditions 12 * are met: 13 * 1. Redistributions of source code must retain the above copyright 14 * notice unmodified, this list of conditions, and the following 15 * disclaimer. 16 * 2. Redistributions in binary form must reproduce the above copyright 17 * notice, this list of conditions and the following disclaimer in the 18 * documentation and/or other materials provided with the distribution. 19 * 20 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR 21 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES 22 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. 23 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT, 24 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT 25 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, 26 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY 27 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT 28 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF 29 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. 30 */ 31 #ifndef _LINUX_MM_H_ 32 #define _LINUX_MM_H_ 33 34 #include <linux/errno.h> 35 36 #include <linux/mmdebug.h> 37 #include <linux/gfp.h> 38 #include <linux/bug.h> 39 #include <linux/list.h> 40 #include <linux/mmzone.h> 41 #include <linux/rbtree.h> 42 #include <linux/atomic.h> 43 #include <linux/mm_types.h> 44 #include <linux/err.h> 45 #include <linux/shrinker.h> 46 47 #include <asm/page.h> 48 #include <asm/pgtable.h> 49 #include <asm/processor.h> 50 51 static inline struct page * 52 nth_page(struct page *page, int n) 53 { 54 return page + n; 55 } 56 57 #define PAGE_ALIGN(addr) round_page(addr) 58 59 #define VM_FAULT_RETRY 0x0400 60 61 #define FAULT_FLAG_ALLOW_RETRY 0x04 62 #define FAULT_FLAG_RETRY_NOWAIT 0x08 63 64 struct vm_fault { 65 unsigned int flags; 66 void __user *virtual_address; 67 }; 68 69 #define VM_FAULT_NOPAGE 0x0001 70 #define VM_FAULT_SIGBUS 0x0002 71 #define VM_FAULT_OOM 0x0004 72 73 #define VM_DONTDUMP 0x0001 74 #define VM_DONTEXPAND 0x0002 75 #define VM_IO 0x0004 76 #define VM_MIXEDMAP 0x0008 77 78 struct vm_operations_struct { 79 int (*fault)(struct vm_area_struct *vma, struct vm_fault *vmf); 80 void (*open)(struct vm_area_struct *vma); 81 void (*close)(struct vm_area_struct *vma); 82 }; 83 84 /* 85 * Compute log2 of the power of two rounded up count of pages 86 * needed for size bytes. 87 */ 88 static inline int 89 get_order(unsigned long size) 90 { 91 int order; 92 93 size = (size - 1) >> PAGE_SHIFT; 94 order = 0; 95 while (size) { 96 order++; 97 size >>= 1; 98 } 99 return (order); 100 } 101 102 /* 103 * This only works via mmap ops. 104 */ 105 static inline int 106 io_remap_pfn_range(struct vm_area_struct *vma, 107 unsigned long addr, unsigned long pfn, unsigned long size, 108 vm_memattr_t prot) 109 { 110 vma->vm_page_prot = prot; 111 vma->vm_pfn = pfn; 112 113 return (0); 114 } 115 116 static inline unsigned long 117 vma_pages(struct vm_area_struct *vma) 118 { 119 unsigned long size; 120 121 size = vma->vm_end - vma->vm_start; 122 123 return size >> PAGE_SHIFT; 124 } 125 126 #define offset_in_page(off) ((unsigned long)(off) & PAGE_MASK) 127 128 static inline void 129 set_page_dirty(struct page *page) 130 { 131 vm_page_dirty((struct vm_page *)page); 132 } 133 134 static inline void 135 get_page(struct vm_page *page) 136 { 137 vm_page_hold(page); 138 } 139 140 extern vm_paddr_t Realmem; 141 142 static inline unsigned long get_num_physpages(void) 143 { 144 return Realmem / PAGE_SIZE; 145 } 146 147 int is_vmalloc_addr(const void *x); 148 149 static inline void 150 unmap_mapping_range(struct address_space *mapping, 151 loff_t const holebegin, loff_t const holelen, int even_cows) 152 { 153 } 154 155 #define VM_SHARED 0x00000008 156 157 #define VM_PFNMAP 0x00000400 158 159 static inline struct page * 160 vmalloc_to_page(const void *addr) 161 { 162 vm_paddr_t paddr; 163 164 paddr = pmap_kextract((vm_offset_t)addr); 165 return (struct page *)(PHYS_TO_VM_PAGE(paddr)); 166 } 167 168 static inline void 169 put_page(struct page *page) 170 { 171 vm_page_busy_wait((struct vm_page *)page, FALSE, "i915gem"); 172 vm_page_unwire((struct vm_page *)page, 1); 173 vm_page_wakeup((struct vm_page *)page); 174 } 175 176 static inline void * 177 page_address(const struct page *page) 178 { 179 return (void *)VM_PAGE_TO_PHYS((const struct vm_page *)page); 180 } 181 182 #endif /* _LINUX_MM_H_ */ 183