kvm_amd64.c revision 330897
1/*-
2 * SPDX-License-Identifier: BSD-3-Clause
3 *
4 * Copyright (c) 1989, 1992, 1993
5 *	The Regents of the University of California.  All rights reserved.
6 *
7 * This code is derived from software developed by the Computer Systems
8 * Engineering group at Lawrence Berkeley Laboratory under DARPA contract
9 * BG 91-66 and contributed to Berkeley.
10 *
11 * Redistribution and use in source and binary forms, with or without
12 * modification, are permitted provided that the following conditions
13 * are met:
14 * 1. Redistributions of source code must retain the above copyright
15 *    notice, this list of conditions and the following disclaimer.
16 * 2. Redistributions in binary form must reproduce the above copyright
17 *    notice, this list of conditions and the following disclaimer in the
18 *    documentation and/or other materials provided with the distribution.
19 * 4. Neither the name of the University nor the names of its contributors
20 *    may be used to endorse or promote products derived from this software
21 *    without specific prior written permission.
22 *
23 * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
24 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
25 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
26 * ARE DISCLAIMED.  IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
27 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
28 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
29 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
30 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
31 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
32 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
33 * SUCH DAMAGE.
34 */
35
36#include <sys/cdefs.h>
37__FBSDID("$FreeBSD: stable/11/lib/libkvm/kvm_amd64.c 330897 2018-03-14 03:19:51Z eadler $");
38
39#if defined(LIBC_SCCS) && !defined(lint)
40#if 0
41static char sccsid[] = "@(#)kvm_hp300.c	8.1 (Berkeley) 6/4/93";
42#endif
43#endif /* LIBC_SCCS and not lint */
44
45/*
46 * AMD64 machine dependent routines for kvm.  Hopefully, the forthcoming
47 * vm code will one day obsolete this module.
48 */
49
50#include <sys/param.h>
51#include <sys/endian.h>
52#include <stdint.h>
53#include <stdlib.h>
54#include <string.h>
55#include <unistd.h>
56#include <kvm.h>
57
58#include <limits.h>
59
60#include "kvm_private.h"
61#include "kvm_amd64.h"
62
63struct vmstate {
64	size_t		phnum;
65	GElf_Phdr	*phdr;
66	amd64_pml4e_t	*PML4;
67};
68
69/*
70 * Translate a physical memory address to a file-offset in the crash-dump.
71 */
72static size_t
73_kvm_pa2off(kvm_t *kd, uint64_t pa, off_t *ofs)
74{
75	struct vmstate *vm = kd->vmst;
76	GElf_Phdr *p;
77	size_t n;
78
79	if (kd->rawdump) {
80		*ofs = pa;
81		return (AMD64_PAGE_SIZE - (pa & AMD64_PAGE_MASK));
82	}
83
84	p = vm->phdr;
85	n = vm->phnum;
86	while (n && (pa < p->p_paddr || pa >= p->p_paddr + p->p_memsz))
87		p++, n--;
88	if (n == 0)
89		return (0);
90	*ofs = (pa - p->p_paddr) + p->p_offset;
91	return (AMD64_PAGE_SIZE - (pa & AMD64_PAGE_MASK));
92}
93
94static void
95_amd64_freevtop(kvm_t *kd)
96{
97	struct vmstate *vm = kd->vmst;
98
99	if (vm->PML4)
100		free(vm->PML4);
101	free(vm->phdr);
102	free(vm);
103	kd->vmst = NULL;
104}
105
106static int
107_amd64_probe(kvm_t *kd)
108{
109
110	return (_kvm_probe_elf_kernel(kd, ELFCLASS64, EM_X86_64) &&
111	    !_kvm_is_minidump(kd));
112}
113
114static int
115_amd64_initvtop(kvm_t *kd)
116{
117	struct kvm_nlist nl[2];
118	amd64_physaddr_t pa;
119	kvaddr_t kernbase;
120	amd64_pml4e_t *PML4;
121
122	kd->vmst = (struct vmstate *)_kvm_malloc(kd, sizeof(*kd->vmst));
123	if (kd->vmst == NULL) {
124		_kvm_err(kd, kd->program, "cannot allocate vm");
125		return (-1);
126	}
127	kd->vmst->PML4 = 0;
128
129	if (kd->rawdump == 0) {
130		if (_kvm_read_core_phdrs(kd, &kd->vmst->phnum,
131		    &kd->vmst->phdr) == -1)
132			return (-1);
133	}
134
135	nl[0].n_name = "kernbase";
136	nl[1].n_name = 0;
137
138	if (kvm_nlist2(kd, nl) != 0) {
139		_kvm_err(kd, kd->program, "bad namelist - no kernbase");
140		return (-1);
141	}
142	kernbase = nl[0].n_value;
143
144	nl[0].n_name = "KPML4phys";
145	nl[1].n_name = 0;
146
147	if (kvm_nlist2(kd, nl) != 0) {
148		_kvm_err(kd, kd->program, "bad namelist - no KPML4phys");
149		return (-1);
150	}
151	if (kvm_read2(kd, (nl[0].n_value - kernbase), &pa, sizeof(pa)) !=
152	    sizeof(pa)) {
153		_kvm_err(kd, kd->program, "cannot read KPML4phys");
154		return (-1);
155	}
156	pa = le64toh(pa);
157	PML4 = _kvm_malloc(kd, AMD64_PAGE_SIZE);
158	if (PML4 == NULL) {
159		_kvm_err(kd, kd->program, "cannot allocate PML4");
160		return (-1);
161	}
162	if (kvm_read2(kd, pa, PML4, AMD64_PAGE_SIZE) != AMD64_PAGE_SIZE) {
163		_kvm_err(kd, kd->program, "cannot read KPML4phys");
164		free(PML4);
165		return (-1);
166	}
167	kd->vmst->PML4 = PML4;
168	return (0);
169}
170
171static int
172_amd64_vatop(kvm_t *kd, kvaddr_t va, off_t *pa)
173{
174	struct vmstate *vm;
175	amd64_physaddr_t offset;
176	amd64_physaddr_t pdpe_pa;
177	amd64_physaddr_t pde_pa;
178	amd64_physaddr_t pte_pa;
179	amd64_pml4e_t pml4e;
180	amd64_pdpe_t pdpe;
181	amd64_pde_t pde;
182	amd64_pte_t pte;
183	kvaddr_t pml4eindex;
184	kvaddr_t pdpeindex;
185	kvaddr_t pdeindex;
186	kvaddr_t pteindex;
187	amd64_physaddr_t a;
188	off_t ofs;
189	size_t s;
190
191	vm = kd->vmst;
192	offset = va & AMD64_PAGE_MASK;
193
194	/*
195	 * If we are initializing (kernel page table descriptor pointer
196	 * not yet set) then return pa == va to avoid infinite recursion.
197	 */
198	if (vm->PML4 == NULL) {
199		s = _kvm_pa2off(kd, va, pa);
200		if (s == 0) {
201			_kvm_err(kd, kd->program,
202			    "_amd64_vatop: bootstrap data not in dump");
203			goto invalid;
204		} else
205			return (AMD64_PAGE_SIZE - offset);
206	}
207
208	pml4eindex = (va >> AMD64_PML4SHIFT) & (AMD64_NPML4EPG - 1);
209	pml4e = le64toh(vm->PML4[pml4eindex]);
210	if ((pml4e & AMD64_PG_V) == 0) {
211		_kvm_err(kd, kd->program, "_amd64_vatop: pml4e not valid");
212		goto invalid;
213	}
214
215	pdpeindex = (va >> AMD64_PDPSHIFT) & (AMD64_NPDPEPG - 1);
216	pdpe_pa = (pml4e & AMD64_PG_FRAME) + (pdpeindex * sizeof(amd64_pdpe_t));
217
218	s = _kvm_pa2off(kd, pdpe_pa, &ofs);
219	if (s < sizeof(pdpe)) {
220		_kvm_err(kd, kd->program, "_amd64_vatop: pdpe_pa not found");
221		goto invalid;
222	}
223	if (pread(kd->pmfd, &pdpe, sizeof(pdpe), ofs) != sizeof(pdpe)) {
224		_kvm_syserr(kd, kd->program, "_amd64_vatop: read pdpe");
225		goto invalid;
226	}
227	pdpe = le64toh(pdpe);
228	if ((pdpe & AMD64_PG_V) == 0) {
229		_kvm_err(kd, kd->program, "_amd64_vatop: pdpe not valid");
230		goto invalid;
231	}
232
233	if (pdpe & AMD64_PG_PS) {
234		/*
235		 * No next-level page table; pdpe describes one 1GB page.
236		 */
237		a = (pdpe & AMD64_PG_1GB_FRAME) + (va & AMD64_PDPMASK);
238		s = _kvm_pa2off(kd, a, pa);
239		if (s == 0) {
240			_kvm_err(kd, kd->program,
241			    "_amd64_vatop: 1GB page address not in dump");
242			goto invalid;
243		} else
244			return (AMD64_NBPDP - (va & AMD64_PDPMASK));
245	}
246
247	pdeindex = (va >> AMD64_PDRSHIFT) & (AMD64_NPDEPG - 1);
248	pde_pa = (pdpe & AMD64_PG_FRAME) + (pdeindex * sizeof(amd64_pde_t));
249
250	s = _kvm_pa2off(kd, pde_pa, &ofs);
251	if (s < sizeof(pde)) {
252		_kvm_syserr(kd, kd->program, "_amd64_vatop: pde_pa not found");
253		goto invalid;
254	}
255	if (pread(kd->pmfd, &pde, sizeof(pde), ofs) != sizeof(pde)) {
256		_kvm_syserr(kd, kd->program, "_amd64_vatop: read pde");
257		goto invalid;
258	}
259	pde = le64toh(pde);
260	if ((pde & AMD64_PG_V) == 0) {
261		_kvm_err(kd, kd->program, "_amd64_vatop: pde not valid");
262		goto invalid;
263	}
264
265	if (pde & AMD64_PG_PS) {
266		/*
267		 * No final-level page table; pde describes one 2MB page.
268		 */
269		a = (pde & AMD64_PG_PS_FRAME) + (va & AMD64_PDRMASK);
270		s = _kvm_pa2off(kd, a, pa);
271		if (s == 0) {
272			_kvm_err(kd, kd->program,
273			    "_amd64_vatop: 2MB page address not in dump");
274			goto invalid;
275		} else
276			return (AMD64_NBPDR - (va & AMD64_PDRMASK));
277	}
278
279	pteindex = (va >> AMD64_PAGE_SHIFT) & (AMD64_NPTEPG - 1);
280	pte_pa = (pde & AMD64_PG_FRAME) + (pteindex * sizeof(amd64_pte_t));
281
282	s = _kvm_pa2off(kd, pte_pa, &ofs);
283	if (s < sizeof(pte)) {
284		_kvm_err(kd, kd->program, "_amd64_vatop: pte_pa not found");
285		goto invalid;
286	}
287	if (pread(kd->pmfd, &pte, sizeof(pte), ofs) != sizeof(pte)) {
288		_kvm_syserr(kd, kd->program, "_amd64_vatop: read");
289		goto invalid;
290	}
291	if ((pte & AMD64_PG_V) == 0) {
292		_kvm_err(kd, kd->program, "_amd64_vatop: pte not valid");
293		goto invalid;
294	}
295
296	a = (pte & AMD64_PG_FRAME) + offset;
297	s = _kvm_pa2off(kd, a, pa);
298	if (s == 0) {
299		_kvm_err(kd, kd->program, "_amd64_vatop: address not in dump");
300		goto invalid;
301	} else
302		return (AMD64_PAGE_SIZE - offset);
303
304invalid:
305	_kvm_err(kd, 0, "invalid address (0x%jx)", (uintmax_t)va);
306	return (0);
307}
308
309static int
310_amd64_kvatop(kvm_t *kd, kvaddr_t va, off_t *pa)
311{
312
313	if (ISALIVE(kd)) {
314		_kvm_err(kd, 0, "kvm_kvatop called in live kernel!");
315		return (0);
316	}
317	return (_amd64_vatop(kd, va, pa));
318}
319
320int
321_amd64_native(kvm_t *kd __unused)
322{
323
324#ifdef __amd64__
325	return (1);
326#else
327	return (0);
328#endif
329}
330
331static struct kvm_arch kvm_amd64 = {
332	.ka_probe = _amd64_probe,
333	.ka_initvtop = _amd64_initvtop,
334	.ka_freevtop = _amd64_freevtop,
335	.ka_kvatop = _amd64_kvatop,
336	.ka_native = _amd64_native,
337};
338
339KVM_ARCH(kvm_amd64);
340