xref: /linux/arch/x86/kernel/jump_label.c (revision 2da68a77)
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * jump label x86 support
4  *
5  * Copyright (C) 2009 Jason Baron <jbaron@redhat.com>
6  *
7  */
8 #include <linux/jump_label.h>
9 #include <linux/memory.h>
10 #include <linux/uaccess.h>
11 #include <linux/module.h>
12 #include <linux/list.h>
13 #include <linux/jhash.h>
14 #include <linux/cpu.h>
15 #include <asm/kprobes.h>
16 #include <asm/alternative.h>
17 #include <asm/text-patching.h>
18 #include <asm/insn.h>
19 
20 int arch_jump_entry_size(struct jump_entry *entry)
21 {
22 	struct insn insn = {};
23 
24 	insn_decode_kernel(&insn, (void *)jump_entry_code(entry));
25 	BUG_ON(insn.length != 2 && insn.length != 5);
26 
27 	return insn.length;
28 }
29 
30 struct jump_label_patch {
31 	const void *code;
32 	int size;
33 };
34 
35 static struct jump_label_patch
36 __jump_label_patch(struct jump_entry *entry, enum jump_label_type type)
37 {
38 	const void *expect, *code, *nop;
39 	const void *addr, *dest;
40 	int size;
41 
42 	addr = (void *)jump_entry_code(entry);
43 	dest = (void *)jump_entry_target(entry);
44 
45 	size = arch_jump_entry_size(entry);
46 	switch (size) {
47 	case JMP8_INSN_SIZE:
48 		code = text_gen_insn(JMP8_INSN_OPCODE, addr, dest);
49 		nop = x86_nops[size];
50 		break;
51 
52 	case JMP32_INSN_SIZE:
53 		code = text_gen_insn(JMP32_INSN_OPCODE, addr, dest);
54 		nop = x86_nops[size];
55 		break;
56 
57 	default: BUG();
58 	}
59 
60 	if (type == JUMP_LABEL_JMP)
61 		expect = nop;
62 	else
63 		expect = code;
64 
65 	if (memcmp(addr, expect, size)) {
66 		/*
67 		 * The location is not an op that we were expecting.
68 		 * Something went wrong. Crash the box, as something could be
69 		 * corrupting the kernel.
70 		 */
71 		pr_crit("jump_label: Fatal kernel bug, unexpected op at %pS [%p] (%5ph != %5ph)) size:%d type:%d\n",
72 				addr, addr, addr, expect, size, type);
73 		BUG();
74 	}
75 
76 	if (type == JUMP_LABEL_NOP)
77 		code = nop;
78 
79 	return (struct jump_label_patch){.code = code, .size = size};
80 }
81 
82 static __always_inline void
83 __jump_label_transform(struct jump_entry *entry,
84 		       enum jump_label_type type,
85 		       int init)
86 {
87 	const struct jump_label_patch jlp = __jump_label_patch(entry, type);
88 
89 	/*
90 	 * As long as only a single processor is running and the code is still
91 	 * not marked as RO, text_poke_early() can be used; Checking that
92 	 * system_state is SYSTEM_BOOTING guarantees it. It will be set to
93 	 * SYSTEM_SCHEDULING before other cores are awaken and before the
94 	 * code is write-protected.
95 	 *
96 	 * At the time the change is being done, just ignore whether we
97 	 * are doing nop -> jump or jump -> nop transition, and assume
98 	 * always nop being the 'currently valid' instruction
99 	 */
100 	if (init || system_state == SYSTEM_BOOTING) {
101 		text_poke_early((void *)jump_entry_code(entry), jlp.code, jlp.size);
102 		return;
103 	}
104 
105 	text_poke_bp((void *)jump_entry_code(entry), jlp.code, jlp.size, NULL);
106 }
107 
108 static void __ref jump_label_transform(struct jump_entry *entry,
109 				       enum jump_label_type type,
110 				       int init)
111 {
112 	mutex_lock(&text_mutex);
113 	__jump_label_transform(entry, type, init);
114 	mutex_unlock(&text_mutex);
115 }
116 
117 void arch_jump_label_transform(struct jump_entry *entry,
118 			       enum jump_label_type type)
119 {
120 	jump_label_transform(entry, type, 0);
121 }
122 
123 bool arch_jump_label_transform_queue(struct jump_entry *entry,
124 				     enum jump_label_type type)
125 {
126 	struct jump_label_patch jlp;
127 
128 	if (system_state == SYSTEM_BOOTING) {
129 		/*
130 		 * Fallback to the non-batching mode.
131 		 */
132 		arch_jump_label_transform(entry, type);
133 		return true;
134 	}
135 
136 	mutex_lock(&text_mutex);
137 	jlp = __jump_label_patch(entry, type);
138 	text_poke_queue((void *)jump_entry_code(entry), jlp.code, jlp.size, NULL);
139 	mutex_unlock(&text_mutex);
140 	return true;
141 }
142 
143 void arch_jump_label_transform_apply(void)
144 {
145 	mutex_lock(&text_mutex);
146 	text_poke_finish();
147 	mutex_unlock(&text_mutex);
148 }
149