xref: /freebsd/lib/libkvm/kvm_amd64.c (revision 8baaf913)
1 /*-
2  * Copyright (c) 1989, 1992, 1993
3  *	The Regents of the University of California.  All rights reserved.
4  *
5  * This code is derived from software developed by the Computer Systems
6  * Engineering group at Lawrence Berkeley Laboratory under DARPA contract
7  * BG 91-66 and contributed to Berkeley.
8  *
9  * Redistribution and use in source and binary forms, with or without
10  * modification, are permitted provided that the following conditions
11  * are met:
12  * 1. Redistributions of source code must retain the above copyright
13  *    notice, this list of conditions and the following disclaimer.
14  * 2. Redistributions in binary form must reproduce the above copyright
15  *    notice, this list of conditions and the following disclaimer in the
16  *    documentation and/or other materials provided with the distribution.
17  * 3. Neither the name of the University nor the names of its contributors
18  *    may be used to endorse or promote products derived from this software
19  *    without specific prior written permission.
20  *
21  * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
22  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
23  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
24  * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
25  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
26  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
27  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
28  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
29  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
30  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
31  * SUCH DAMAGE.
32  */
33 
34 #include <sys/cdefs.h>
35 __FBSDID("$FreeBSD$");
36 
37 #if defined(LIBC_SCCS) && !defined(lint)
38 #if 0
39 static char sccsid[] = "@(#)kvm_hp300.c	8.1 (Berkeley) 6/4/93";
40 #endif
41 #endif /* LIBC_SCCS and not lint */
42 
43 /*
44  * AMD64 machine dependent routines for kvm.  Hopefully, the forthcoming
45  * vm code will one day obsolete this module.
46  */
47 
48 #include <sys/param.h>
49 #include <sys/endian.h>
50 #include <stdint.h>
51 #include <stdlib.h>
52 #include <string.h>
53 #include <unistd.h>
54 #include <vm/vm.h>
55 #include <kvm.h>
56 
57 #include <limits.h>
58 
59 #include "kvm_private.h"
60 #include "kvm_amd64.h"
61 
62 struct vmstate {
63 	size_t		phnum;
64 	GElf_Phdr	*phdr;
65 	amd64_pml4e_t	*PML4;
66 };
67 
68 /*
69  * Translate a physical memory address to a file-offset in the crash-dump.
70  */
71 static size_t
72 _kvm_pa2off(kvm_t *kd, uint64_t pa, off_t *ofs)
73 {
74 	struct vmstate *vm = kd->vmst;
75 	GElf_Phdr *p;
76 	size_t n;
77 
78 	if (kd->rawdump) {
79 		*ofs = pa;
80 		return (AMD64_PAGE_SIZE - (pa & AMD64_PAGE_MASK));
81 	}
82 
83 	p = vm->phdr;
84 	n = vm->phnum;
85 	while (n && (pa < p->p_paddr || pa >= p->p_paddr + p->p_memsz))
86 		p++, n--;
87 	if (n == 0)
88 		return (0);
89 	*ofs = (pa - p->p_paddr) + p->p_offset;
90 	return (AMD64_PAGE_SIZE - (pa & AMD64_PAGE_MASK));
91 }
92 
93 static void
94 _amd64_freevtop(kvm_t *kd)
95 {
96 	struct vmstate *vm = kd->vmst;
97 
98 	if (vm->PML4)
99 		free(vm->PML4);
100 	free(vm->phdr);
101 	free(vm);
102 	kd->vmst = NULL;
103 }
104 
105 static int
106 _amd64_probe(kvm_t *kd)
107 {
108 
109 	return (_kvm_probe_elf_kernel(kd, ELFCLASS64, EM_X86_64) &&
110 	    !_kvm_is_minidump(kd));
111 }
112 
113 static int
114 _amd64_initvtop(kvm_t *kd)
115 {
116 	struct kvm_nlist nl[2];
117 	amd64_physaddr_t pa;
118 	kvaddr_t kernbase;
119 	amd64_pml4e_t *PML4;
120 
121 	kd->vmst = (struct vmstate *)_kvm_malloc(kd, sizeof(*kd->vmst));
122 	if (kd->vmst == NULL) {
123 		_kvm_err(kd, kd->program, "cannot allocate vm");
124 		return (-1);
125 	}
126 	kd->vmst->PML4 = 0;
127 
128 	if (kd->rawdump == 0) {
129 		if (_kvm_read_core_phdrs(kd, &kd->vmst->phnum,
130 		    &kd->vmst->phdr) == -1)
131 			return (-1);
132 	}
133 
134 	nl[0].n_name = "kernbase";
135 	nl[1].n_name = 0;
136 
137 	if (kvm_nlist2(kd, nl) != 0) {
138 		_kvm_err(kd, kd->program, "bad namelist - no kernbase");
139 		return (-1);
140 	}
141 	kernbase = nl[0].n_value;
142 
143 	nl[0].n_name = "KPML4phys";
144 	nl[1].n_name = 0;
145 
146 	if (kvm_nlist2(kd, nl) != 0) {
147 		_kvm_err(kd, kd->program, "bad namelist - no KPML4phys");
148 		return (-1);
149 	}
150 	if (kvm_read2(kd, (nl[0].n_value - kernbase), &pa, sizeof(pa)) !=
151 	    sizeof(pa)) {
152 		_kvm_err(kd, kd->program, "cannot read KPML4phys");
153 		return (-1);
154 	}
155 	pa = le64toh(pa);
156 	PML4 = _kvm_malloc(kd, AMD64_PAGE_SIZE);
157 	if (PML4 == NULL) {
158 		_kvm_err(kd, kd->program, "cannot allocate PML4");
159 		return (-1);
160 	}
161 	if (kvm_read2(kd, pa, PML4, AMD64_PAGE_SIZE) != AMD64_PAGE_SIZE) {
162 		_kvm_err(kd, kd->program, "cannot read KPML4phys");
163 		free(PML4);
164 		return (-1);
165 	}
166 	kd->vmst->PML4 = PML4;
167 	return (0);
168 }
169 
170 static int
171 _amd64_vatop(kvm_t *kd, kvaddr_t va, off_t *pa)
172 {
173 	struct vmstate *vm;
174 	amd64_physaddr_t offset;
175 	amd64_physaddr_t pdpe_pa;
176 	amd64_physaddr_t pde_pa;
177 	amd64_physaddr_t pte_pa;
178 	amd64_pml4e_t pml4e;
179 	amd64_pdpe_t pdpe;
180 	amd64_pde_t pde;
181 	amd64_pte_t pte;
182 	kvaddr_t pml4eindex;
183 	kvaddr_t pdpeindex;
184 	kvaddr_t pdeindex;
185 	kvaddr_t pteindex;
186 	amd64_physaddr_t a;
187 	off_t ofs;
188 	size_t s;
189 
190 	vm = kd->vmst;
191 	offset = va & AMD64_PAGE_MASK;
192 
193 	/*
194 	 * If we are initializing (kernel page table descriptor pointer
195 	 * not yet set) then return pa == va to avoid infinite recursion.
196 	 */
197 	if (vm->PML4 == NULL) {
198 		s = _kvm_pa2off(kd, va, pa);
199 		if (s == 0) {
200 			_kvm_err(kd, kd->program,
201 			    "_amd64_vatop: bootstrap data not in dump");
202 			goto invalid;
203 		} else
204 			return (AMD64_PAGE_SIZE - offset);
205 	}
206 
207 	pml4eindex = (va >> AMD64_PML4SHIFT) & (AMD64_NPML4EPG - 1);
208 	pml4e = le64toh(vm->PML4[pml4eindex]);
209 	if ((pml4e & AMD64_PG_V) == 0) {
210 		_kvm_err(kd, kd->program, "_amd64_vatop: pml4e not valid");
211 		goto invalid;
212 	}
213 
214 	pdpeindex = (va >> AMD64_PDPSHIFT) & (AMD64_NPDPEPG - 1);
215 	pdpe_pa = (pml4e & AMD64_PG_FRAME) + (pdpeindex * sizeof(amd64_pdpe_t));
216 
217 	s = _kvm_pa2off(kd, pdpe_pa, &ofs);
218 	if (s < sizeof(pdpe)) {
219 		_kvm_err(kd, kd->program, "_amd64_vatop: pdpe_pa not found");
220 		goto invalid;
221 	}
222 	if (pread(kd->pmfd, &pdpe, sizeof(pdpe), ofs) != sizeof(pdpe)) {
223 		_kvm_syserr(kd, kd->program, "_amd64_vatop: read pdpe");
224 		goto invalid;
225 	}
226 	pdpe = le64toh(pdpe);
227 	if ((pdpe & AMD64_PG_V) == 0) {
228 		_kvm_err(kd, kd->program, "_amd64_vatop: pdpe not valid");
229 		goto invalid;
230 	}
231 
232 	if (pdpe & AMD64_PG_PS) {
233 		/*
234 		 * No next-level page table; pdpe describes one 1GB page.
235 		 */
236 		a = (pdpe & AMD64_PG_1GB_FRAME) + (va & AMD64_PDPMASK);
237 		s = _kvm_pa2off(kd, a, pa);
238 		if (s == 0) {
239 			_kvm_err(kd, kd->program,
240 			    "_amd64_vatop: 1GB page address not in dump");
241 			goto invalid;
242 		} else
243 			return (AMD64_NBPDP - (va & AMD64_PDPMASK));
244 	}
245 
246 	pdeindex = (va >> AMD64_PDRSHIFT) & (AMD64_NPDEPG - 1);
247 	pde_pa = (pdpe & AMD64_PG_FRAME) + (pdeindex * sizeof(amd64_pde_t));
248 
249 	s = _kvm_pa2off(kd, pde_pa, &ofs);
250 	if (s < sizeof(pde)) {
251 		_kvm_syserr(kd, kd->program, "_amd64_vatop: pde_pa not found");
252 		goto invalid;
253 	}
254 	if (pread(kd->pmfd, &pde, sizeof(pde), ofs) != sizeof(pde)) {
255 		_kvm_syserr(kd, kd->program, "_amd64_vatop: read pde");
256 		goto invalid;
257 	}
258 	pde = le64toh(pde);
259 	if ((pde & AMD64_PG_V) == 0) {
260 		_kvm_err(kd, kd->program, "_amd64_vatop: pde not valid");
261 		goto invalid;
262 	}
263 
264 	if (pde & AMD64_PG_PS) {
265 		/*
266 		 * No final-level page table; pde describes one 2MB page.
267 		 */
268 		a = (pde & AMD64_PG_PS_FRAME) + (va & AMD64_PDRMASK);
269 		s = _kvm_pa2off(kd, a, pa);
270 		if (s == 0) {
271 			_kvm_err(kd, kd->program,
272 			    "_amd64_vatop: 2MB page address not in dump");
273 			goto invalid;
274 		} else
275 			return (AMD64_NBPDR - (va & AMD64_PDRMASK));
276 	}
277 
278 	pteindex = (va >> AMD64_PAGE_SHIFT) & (AMD64_NPTEPG - 1);
279 	pte_pa = (pde & AMD64_PG_FRAME) + (pteindex * sizeof(amd64_pte_t));
280 
281 	s = _kvm_pa2off(kd, pte_pa, &ofs);
282 	if (s < sizeof(pte)) {
283 		_kvm_err(kd, kd->program, "_amd64_vatop: pte_pa not found");
284 		goto invalid;
285 	}
286 	if (pread(kd->pmfd, &pte, sizeof(pte), ofs) != sizeof(pte)) {
287 		_kvm_syserr(kd, kd->program, "_amd64_vatop: read");
288 		goto invalid;
289 	}
290 	if ((pte & AMD64_PG_V) == 0) {
291 		_kvm_err(kd, kd->program, "_amd64_vatop: pte not valid");
292 		goto invalid;
293 	}
294 
295 	a = (pte & AMD64_PG_FRAME) + offset;
296 	s = _kvm_pa2off(kd, a, pa);
297 	if (s == 0) {
298 		_kvm_err(kd, kd->program, "_amd64_vatop: address not in dump");
299 		goto invalid;
300 	} else
301 		return (AMD64_PAGE_SIZE - offset);
302 
303 invalid:
304 	_kvm_err(kd, 0, "invalid address (0x%jx)", (uintmax_t)va);
305 	return (0);
306 }
307 
308 static int
309 _amd64_kvatop(kvm_t *kd, kvaddr_t va, off_t *pa)
310 {
311 
312 	if (ISALIVE(kd)) {
313 		_kvm_err(kd, 0, "kvm_kvatop called in live kernel!");
314 		return (0);
315 	}
316 	return (_amd64_vatop(kd, va, pa));
317 }
318 
319 int
320 _amd64_native(kvm_t *kd __unused)
321 {
322 
323 #ifdef __amd64__
324 	return (1);
325 #else
326 	return (0);
327 #endif
328 }
329 
330 static struct kvm_arch kvm_amd64 = {
331 	.ka_probe = _amd64_probe,
332 	.ka_initvtop = _amd64_initvtop,
333 	.ka_freevtop = _amd64_freevtop,
334 	.ka_kvatop = _amd64_kvatop,
335 	.ka_native = _amd64_native,
336 };
337 
338 KVM_ARCH(kvm_amd64);
339