xref: /linux/arch/x86/kernel/module.c (revision 44f57d78)
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*  Kernel module help for x86.
3     Copyright (C) 2001 Rusty Russell.
4 
5 */
6 
7 #define pr_fmt(fmt) KBUILD_MODNAME ": " fmt
8 
9 #include <linux/moduleloader.h>
10 #include <linux/elf.h>
11 #include <linux/vmalloc.h>
12 #include <linux/fs.h>
13 #include <linux/string.h>
14 #include <linux/kernel.h>
15 #include <linux/kasan.h>
16 #include <linux/bug.h>
17 #include <linux/mm.h>
18 #include <linux/gfp.h>
19 #include <linux/jump_label.h>
20 #include <linux/random.h>
21 
22 #include <asm/text-patching.h>
23 #include <asm/page.h>
24 #include <asm/pgtable.h>
25 #include <asm/setup.h>
26 #include <asm/unwind.h>
27 
28 #if 0
29 #define DEBUGP(fmt, ...)				\
30 	printk(KERN_DEBUG fmt, ##__VA_ARGS__)
31 #else
32 #define DEBUGP(fmt, ...)				\
33 do {							\
34 	if (0)						\
35 		printk(KERN_DEBUG fmt, ##__VA_ARGS__);	\
36 } while (0)
37 #endif
38 
39 #ifdef CONFIG_RANDOMIZE_BASE
40 static unsigned long module_load_offset;
41 
42 /* Mutex protects the module_load_offset. */
43 static DEFINE_MUTEX(module_kaslr_mutex);
44 
45 static unsigned long int get_module_load_offset(void)
46 {
47 	if (kaslr_enabled()) {
48 		mutex_lock(&module_kaslr_mutex);
49 		/*
50 		 * Calculate the module_load_offset the first time this
51 		 * code is called. Once calculated it stays the same until
52 		 * reboot.
53 		 */
54 		if (module_load_offset == 0)
55 			module_load_offset =
56 				(get_random_int() % 1024 + 1) * PAGE_SIZE;
57 		mutex_unlock(&module_kaslr_mutex);
58 	}
59 	return module_load_offset;
60 }
61 #else
62 static unsigned long int get_module_load_offset(void)
63 {
64 	return 0;
65 }
66 #endif
67 
68 void *module_alloc(unsigned long size)
69 {
70 	void *p;
71 
72 	if (PAGE_ALIGN(size) > MODULES_LEN)
73 		return NULL;
74 
75 	p = __vmalloc_node_range(size, MODULE_ALIGN,
76 				    MODULES_VADDR + get_module_load_offset(),
77 				    MODULES_END, GFP_KERNEL,
78 				    PAGE_KERNEL, 0, NUMA_NO_NODE,
79 				    __builtin_return_address(0));
80 	if (p && (kasan_module_alloc(p, size) < 0)) {
81 		vfree(p);
82 		return NULL;
83 	}
84 
85 	return p;
86 }
87 
88 #ifdef CONFIG_X86_32
89 int apply_relocate(Elf32_Shdr *sechdrs,
90 		   const char *strtab,
91 		   unsigned int symindex,
92 		   unsigned int relsec,
93 		   struct module *me)
94 {
95 	unsigned int i;
96 	Elf32_Rel *rel = (void *)sechdrs[relsec].sh_addr;
97 	Elf32_Sym *sym;
98 	uint32_t *location;
99 
100 	DEBUGP("Applying relocate section %u to %u\n",
101 	       relsec, sechdrs[relsec].sh_info);
102 	for (i = 0; i < sechdrs[relsec].sh_size / sizeof(*rel); i++) {
103 		/* This is where to make the change */
104 		location = (void *)sechdrs[sechdrs[relsec].sh_info].sh_addr
105 			+ rel[i].r_offset;
106 		/* This is the symbol it is referring to.  Note that all
107 		   undefined symbols have been resolved.  */
108 		sym = (Elf32_Sym *)sechdrs[symindex].sh_addr
109 			+ ELF32_R_SYM(rel[i].r_info);
110 
111 		switch (ELF32_R_TYPE(rel[i].r_info)) {
112 		case R_386_32:
113 			/* We add the value into the location given */
114 			*location += sym->st_value;
115 			break;
116 		case R_386_PC32:
117 			/* Add the value, subtract its position */
118 			*location += sym->st_value - (uint32_t)location;
119 			break;
120 		default:
121 			pr_err("%s: Unknown relocation: %u\n",
122 			       me->name, ELF32_R_TYPE(rel[i].r_info));
123 			return -ENOEXEC;
124 		}
125 	}
126 	return 0;
127 }
128 #else /*X86_64*/
129 int apply_relocate_add(Elf64_Shdr *sechdrs,
130 		   const char *strtab,
131 		   unsigned int symindex,
132 		   unsigned int relsec,
133 		   struct module *me)
134 {
135 	unsigned int i;
136 	Elf64_Rela *rel = (void *)sechdrs[relsec].sh_addr;
137 	Elf64_Sym *sym;
138 	void *loc;
139 	u64 val;
140 
141 	DEBUGP("Applying relocate section %u to %u\n",
142 	       relsec, sechdrs[relsec].sh_info);
143 	for (i = 0; i < sechdrs[relsec].sh_size / sizeof(*rel); i++) {
144 		/* This is where to make the change */
145 		loc = (void *)sechdrs[sechdrs[relsec].sh_info].sh_addr
146 			+ rel[i].r_offset;
147 
148 		/* This is the symbol it is referring to.  Note that all
149 		   undefined symbols have been resolved.  */
150 		sym = (Elf64_Sym *)sechdrs[symindex].sh_addr
151 			+ ELF64_R_SYM(rel[i].r_info);
152 
153 		DEBUGP("type %d st_value %Lx r_addend %Lx loc %Lx\n",
154 		       (int)ELF64_R_TYPE(rel[i].r_info),
155 		       sym->st_value, rel[i].r_addend, (u64)loc);
156 
157 		val = sym->st_value + rel[i].r_addend;
158 
159 		switch (ELF64_R_TYPE(rel[i].r_info)) {
160 		case R_X86_64_NONE:
161 			break;
162 		case R_X86_64_64:
163 			if (*(u64 *)loc != 0)
164 				goto invalid_relocation;
165 			*(u64 *)loc = val;
166 			break;
167 		case R_X86_64_32:
168 			if (*(u32 *)loc != 0)
169 				goto invalid_relocation;
170 			*(u32 *)loc = val;
171 			if (val != *(u32 *)loc)
172 				goto overflow;
173 			break;
174 		case R_X86_64_32S:
175 			if (*(s32 *)loc != 0)
176 				goto invalid_relocation;
177 			*(s32 *)loc = val;
178 			if ((s64)val != *(s32 *)loc)
179 				goto overflow;
180 			break;
181 		case R_X86_64_PC32:
182 		case R_X86_64_PLT32:
183 			if (*(u32 *)loc != 0)
184 				goto invalid_relocation;
185 			val -= (u64)loc;
186 			*(u32 *)loc = val;
187 #if 0
188 			if ((s64)val != *(s32 *)loc)
189 				goto overflow;
190 #endif
191 			break;
192 		case R_X86_64_PC64:
193 			if (*(u64 *)loc != 0)
194 				goto invalid_relocation;
195 			val -= (u64)loc;
196 			*(u64 *)loc = val;
197 			break;
198 		default:
199 			pr_err("%s: Unknown rela relocation: %llu\n",
200 			       me->name, ELF64_R_TYPE(rel[i].r_info));
201 			return -ENOEXEC;
202 		}
203 	}
204 	return 0;
205 
206 invalid_relocation:
207 	pr_err("x86/modules: Skipping invalid relocation target, existing value is nonzero for type %d, loc %p, val %Lx\n",
208 	       (int)ELF64_R_TYPE(rel[i].r_info), loc, val);
209 	return -ENOEXEC;
210 
211 overflow:
212 	pr_err("overflow in relocation type %d val %Lx\n",
213 	       (int)ELF64_R_TYPE(rel[i].r_info), val);
214 	pr_err("`%s' likely not compiled with -mcmodel=kernel\n",
215 	       me->name);
216 	return -ENOEXEC;
217 }
218 #endif
219 
220 int module_finalize(const Elf_Ehdr *hdr,
221 		    const Elf_Shdr *sechdrs,
222 		    struct module *me)
223 {
224 	const Elf_Shdr *s, *text = NULL, *alt = NULL, *locks = NULL,
225 		*para = NULL, *orc = NULL, *orc_ip = NULL;
226 	char *secstrings = (void *)hdr + sechdrs[hdr->e_shstrndx].sh_offset;
227 
228 	for (s = sechdrs; s < sechdrs + hdr->e_shnum; s++) {
229 		if (!strcmp(".text", secstrings + s->sh_name))
230 			text = s;
231 		if (!strcmp(".altinstructions", secstrings + s->sh_name))
232 			alt = s;
233 		if (!strcmp(".smp_locks", secstrings + s->sh_name))
234 			locks = s;
235 		if (!strcmp(".parainstructions", secstrings + s->sh_name))
236 			para = s;
237 		if (!strcmp(".orc_unwind", secstrings + s->sh_name))
238 			orc = s;
239 		if (!strcmp(".orc_unwind_ip", secstrings + s->sh_name))
240 			orc_ip = s;
241 	}
242 
243 	if (alt) {
244 		/* patch .altinstructions */
245 		void *aseg = (void *)alt->sh_addr;
246 		apply_alternatives(aseg, aseg + alt->sh_size);
247 	}
248 	if (locks && text) {
249 		void *lseg = (void *)locks->sh_addr;
250 		void *tseg = (void *)text->sh_addr;
251 		alternatives_smp_module_add(me, me->name,
252 					    lseg, lseg + locks->sh_size,
253 					    tseg, tseg + text->sh_size);
254 	}
255 
256 	if (para) {
257 		void *pseg = (void *)para->sh_addr;
258 		apply_paravirt(pseg, pseg + para->sh_size);
259 	}
260 
261 	/* make jump label nops */
262 	jump_label_apply_nops(me);
263 
264 	if (orc && orc_ip)
265 		unwind_module_init(me, (void *)orc_ip->sh_addr, orc_ip->sh_size,
266 				   (void *)orc->sh_addr, orc->sh_size);
267 
268 	return 0;
269 }
270 
271 void module_arch_cleanup(struct module *mod)
272 {
273 	alternatives_smp_module_del(mod);
274 }
275