2012-12-10 15:35:24 +00:00
|
|
|
/*
|
|
|
|
* Copyright (C) 2012,2013 - ARM Ltd
|
|
|
|
* Author: Marc Zyngier <marc.zyngier@arm.com>
|
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or modify
|
|
|
|
* it under the terms of the GNU General Public License version 2 as
|
|
|
|
* published by the Free Software Foundation.
|
|
|
|
*
|
|
|
|
* This program is distributed in the hope that it will be useful,
|
|
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
|
|
* GNU General Public License for more details.
|
|
|
|
*
|
|
|
|
* You should have received a copy of the GNU General Public License
|
|
|
|
* along with this program. If not, see <http://www.gnu.org/licenses/>.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#ifndef __ARM64_KVM_MMU_H__
|
|
|
|
#define __ARM64_KVM_MMU_H__
|
|
|
|
|
|
|
|
#include <asm/page.h>
|
|
|
|
#include <asm/memory.h>
|
2015-11-16 11:28:18 +00:00
|
|
|
#include <asm/cpufeature.h>
|
2012-12-10 15:35:24 +00:00
|
|
|
|
|
|
|
/*
|
2015-01-29 13:50:34 +00:00
|
|
|
* As ARMv8.0 only has the TTBR0_EL2 register, we cannot express
|
2012-12-10 15:35:24 +00:00
|
|
|
* "negative" addresses. This makes it impossible to directly share
|
|
|
|
* mappings with the kernel.
|
|
|
|
*
|
|
|
|
* Instead, give the HYP mode its own VA region at a fixed offset from
|
|
|
|
* the kernel by just masking the top bits (which are all ones for a
|
2016-06-30 17:40:34 +00:00
|
|
|
* kernel address). We need to find out how many bits to mask.
|
2015-01-29 13:50:34 +00:00
|
|
|
*
|
2016-06-30 17:40:34 +00:00
|
|
|
* We want to build a set of page tables that cover both parts of the
|
|
|
|
* idmap (the trampoline page used to initialize EL2), and our normal
|
|
|
|
* runtime VA space, at the same time.
|
|
|
|
*
|
|
|
|
* Given that the kernel uses VA_BITS for its entire address space,
|
|
|
|
* and that half of that space (VA_BITS - 1) is used for the linear
|
|
|
|
* mapping, we can also limit the EL2 space to (VA_BITS - 1).
|
|
|
|
*
|
|
|
|
* The main question is "Within the VA_BITS space, does EL2 use the
|
|
|
|
* top or the bottom half of that space to shadow the kernel's linear
|
|
|
|
* mapping?". As we need to idmap the trampoline page, this is
|
|
|
|
* determined by the range in which this page lives.
|
|
|
|
*
|
|
|
|
* If the page is in the bottom half, we have to use the top half. If
|
|
|
|
* the page is in the top half, we have to use the bottom half:
|
|
|
|
*
|
2017-01-10 21:35:49 +00:00
|
|
|
* T = __pa_symbol(__hyp_idmap_text_start)
|
2016-06-30 17:40:34 +00:00
|
|
|
* if (T & BIT(VA_BITS - 1))
|
|
|
|
* HYP_VA_MIN = 0 //idmap in upper half
|
|
|
|
* else
|
|
|
|
* HYP_VA_MIN = 1 << (VA_BITS - 1)
|
|
|
|
* HYP_VA_MAX = HYP_VA_MIN + (1 << (VA_BITS - 1)) - 1
|
|
|
|
*
|
|
|
|
* This of course assumes that the trampoline page exists within the
|
|
|
|
* VA_BITS range. If it doesn't, then it means we're in the odd case
|
|
|
|
* where the kernel idmap (as well as HYP) uses more levels than the
|
|
|
|
* kernel runtime page tables (as seen when the kernel is configured
|
|
|
|
* for 4k pages, 39bits VA, and yet memory lives just above that
|
|
|
|
* limit, forcing the idmap to use 4 levels of page tables while the
|
|
|
|
* kernel itself only uses 3). In this particular case, it doesn't
|
|
|
|
* matter which side of VA_BITS we use, as we're guaranteed not to
|
|
|
|
* conflict with anything.
|
|
|
|
*
|
|
|
|
* When using VHE, there are no separate hyp mappings and all KVM
|
|
|
|
* functionality is already mapped as part of the main kernel
|
|
|
|
* mappings, and none of this applies in that case.
|
2012-12-10 15:35:24 +00:00
|
|
|
*/
|
2016-06-30 17:40:39 +00:00
|
|
|
|
|
|
|
#define HYP_PAGE_OFFSET_HIGH_MASK ((UL(1) << VA_BITS) - 1)
|
|
|
|
#define HYP_PAGE_OFFSET_LOW_MASK ((UL(1) << (VA_BITS - 1)) - 1)
|
|
|
|
|
2012-12-10 15:35:24 +00:00
|
|
|
#ifdef __ASSEMBLY__
|
|
|
|
|
2015-01-29 13:50:34 +00:00
|
|
|
#include <asm/alternative.h>
|
|
|
|
#include <asm/cpufeature.h>
|
|
|
|
|
2012-12-10 15:35:24 +00:00
|
|
|
/*
|
|
|
|
* Convert a kernel VA into a HYP VA.
|
|
|
|
* reg: VA to be converted.
|
2016-06-30 17:40:40 +00:00
|
|
|
*
|
|
|
|
* This generates the following sequences:
|
|
|
|
* - High mask:
|
|
|
|
* and x0, x0, #HYP_PAGE_OFFSET_HIGH_MASK
|
|
|
|
* nop
|
|
|
|
* - Low mask:
|
|
|
|
* and x0, x0, #HYP_PAGE_OFFSET_HIGH_MASK
|
|
|
|
* and x0, x0, #HYP_PAGE_OFFSET_LOW_MASK
|
|
|
|
* - VHE:
|
|
|
|
* nop
|
|
|
|
* nop
|
|
|
|
*
|
|
|
|
* The "low mask" version works because the mask is a strict subset of
|
|
|
|
* the "high mask", hence performing the first mask for nothing.
|
|
|
|
* Should be completely invisible on any viable CPU.
|
2012-12-10 15:35:24 +00:00
|
|
|
*/
|
|
|
|
.macro kern_hyp_va reg
|
2016-06-30 17:40:40 +00:00
|
|
|
alternative_if_not ARM64_HAS_VIRT_HOST_EXTN
|
|
|
|
and \reg, \reg, #HYP_PAGE_OFFSET_HIGH_MASK
|
2016-09-07 10:07:10 +00:00
|
|
|
alternative_else_nop_endif
|
|
|
|
alternative_if ARM64_HYP_OFFSET_LOW
|
2016-06-30 17:40:40 +00:00
|
|
|
and \reg, \reg, #HYP_PAGE_OFFSET_LOW_MASK
|
2016-09-07 10:07:10 +00:00
|
|
|
alternative_else_nop_endif
|
2012-12-10 15:35:24 +00:00
|
|
|
.endm
|
|
|
|
|
|
|
|
#else
|
|
|
|
|
2014-10-10 10:14:28 +00:00
|
|
|
#include <asm/pgalloc.h>
|
2017-03-10 20:32:23 +00:00
|
|
|
#include <asm/cache.h>
|
2012-12-10 15:35:24 +00:00
|
|
|
#include <asm/cacheflush.h>
|
2015-03-19 16:42:28 +00:00
|
|
|
#include <asm/mmu_context.h>
|
|
|
|
#include <asm/pgtable.h>
|
2012-12-10 15:35:24 +00:00
|
|
|
|
2016-06-30 17:40:40 +00:00
|
|
|
static inline unsigned long __kern_hyp_va(unsigned long v)
|
|
|
|
{
|
|
|
|
asm volatile(ALTERNATIVE("and %0, %0, %1",
|
|
|
|
"nop",
|
|
|
|
ARM64_HAS_VIRT_HOST_EXTN)
|
|
|
|
: "+r" (v)
|
|
|
|
: "i" (HYP_PAGE_OFFSET_HIGH_MASK));
|
|
|
|
asm volatile(ALTERNATIVE("nop",
|
|
|
|
"and %0, %0, %1",
|
|
|
|
ARM64_HYP_OFFSET_LOW)
|
|
|
|
: "+r" (v)
|
|
|
|
: "i" (HYP_PAGE_OFFSET_LOW_MASK));
|
|
|
|
return v;
|
|
|
|
}
|
|
|
|
|
2016-10-18 17:37:49 +00:00
|
|
|
#define kern_hyp_va(v) ((typeof(v))(__kern_hyp_va((unsigned long)(v))))
|
2012-12-10 15:35:24 +00:00
|
|
|
|
|
|
|
/*
|
2014-07-09 16:17:04 +00:00
|
|
|
* We currently only support a 40bit IPA.
|
2012-12-10 15:35:24 +00:00
|
|
|
*/
|
2014-07-09 16:17:04 +00:00
|
|
|
#define KVM_PHYS_SHIFT (40)
|
2012-12-10 15:35:24 +00:00
|
|
|
#define KVM_PHYS_SIZE (1UL << KVM_PHYS_SHIFT)
|
|
|
|
#define KVM_PHYS_MASK (KVM_PHYS_SIZE - 1UL)
|
|
|
|
|
2016-03-22 14:16:52 +00:00
|
|
|
#include <asm/stage2_pgtable.h>
|
|
|
|
|
2016-06-13 14:00:45 +00:00
|
|
|
int create_hyp_mappings(void *from, void *to, pgprot_t prot);
|
2012-12-10 15:35:24 +00:00
|
|
|
int create_hyp_io_mappings(void *from, void *to, phys_addr_t);
|
|
|
|
void free_hyp_pgds(void);
|
|
|
|
|
2014-11-27 09:35:03 +00:00
|
|
|
void stage2_unmap_vm(struct kvm *kvm);
|
2012-12-10 15:35:24 +00:00
|
|
|
int kvm_alloc_stage2_pgd(struct kvm *kvm);
|
|
|
|
void kvm_free_stage2_pgd(struct kvm *kvm);
|
|
|
|
int kvm_phys_addr_ioremap(struct kvm *kvm, phys_addr_t guest_ipa,
|
2014-09-17 21:56:18 +00:00
|
|
|
phys_addr_t pa, unsigned long size, bool writable);
|
2012-12-10 15:35:24 +00:00
|
|
|
|
|
|
|
int kvm_handle_guest_abort(struct kvm_vcpu *vcpu, struct kvm_run *run);
|
|
|
|
|
|
|
|
void kvm_mmu_free_memory_caches(struct kvm_vcpu *vcpu);
|
|
|
|
|
|
|
|
phys_addr_t kvm_mmu_get_httbr(void);
|
|
|
|
phys_addr_t kvm_get_idmap_vector(void);
|
|
|
|
int kvm_mmu_init(void);
|
|
|
|
void kvm_clear_hyp_idmap(void);
|
|
|
|
|
|
|
|
#define kvm_set_pte(ptep, pte) set_pte(ptep, pte)
|
2012-11-01 16:14:45 +00:00
|
|
|
#define kvm_set_pmd(pmdp, pmd) set_pmd(pmdp, pmd)
|
2012-12-10 15:35:24 +00:00
|
|
|
|
2016-04-13 16:57:37 +00:00
|
|
|
static inline pte_t kvm_s2pte_mkwrite(pte_t pte)
|
2012-12-10 15:35:24 +00:00
|
|
|
{
|
2016-04-13 16:57:37 +00:00
|
|
|
pte_val(pte) |= PTE_S2_RDWR;
|
|
|
|
return pte;
|
2012-12-10 15:35:24 +00:00
|
|
|
}
|
|
|
|
|
2016-04-13 16:57:37 +00:00
|
|
|
static inline pmd_t kvm_s2pmd_mkwrite(pmd_t pmd)
|
2012-11-01 16:14:45 +00:00
|
|
|
{
|
2016-04-13 16:57:37 +00:00
|
|
|
pmd_val(pmd) |= PMD_S2_RDWR;
|
|
|
|
return pmd;
|
2012-11-01 16:14:45 +00:00
|
|
|
}
|
|
|
|
|
2017-10-23 16:11:19 +00:00
|
|
|
static inline pte_t kvm_s2pte_mkexec(pte_t pte)
|
|
|
|
{
|
|
|
|
pte_val(pte) &= ~PTE_S2_XN;
|
|
|
|
return pte;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline pmd_t kvm_s2pmd_mkexec(pmd_t pmd)
|
|
|
|
{
|
|
|
|
pmd_val(pmd) &= ~PMD_S2_XN;
|
|
|
|
return pmd;
|
|
|
|
}
|
|
|
|
|
2018-02-15 11:14:56 +00:00
|
|
|
static inline void kvm_set_s2pte_readonly(pte_t *ptep)
|
2015-01-15 23:58:59 +00:00
|
|
|
{
|
2017-07-06 10:46:39 +00:00
|
|
|
pteval_t old_pteval, pteval;
|
|
|
|
|
2018-02-15 11:14:56 +00:00
|
|
|
pteval = READ_ONCE(pte_val(*ptep));
|
2017-07-06 10:46:39 +00:00
|
|
|
do {
|
|
|
|
old_pteval = pteval;
|
|
|
|
pteval &= ~PTE_S2_RDWR;
|
|
|
|
pteval |= PTE_S2_RDONLY;
|
2018-02-15 11:14:56 +00:00
|
|
|
pteval = cmpxchg_relaxed(&pte_val(*ptep), old_pteval, pteval);
|
2017-07-06 10:46:39 +00:00
|
|
|
} while (pteval != old_pteval);
|
2015-01-15 23:58:59 +00:00
|
|
|
}
|
|
|
|
|
2018-02-15 11:14:56 +00:00
|
|
|
static inline bool kvm_s2pte_readonly(pte_t *ptep)
|
2015-01-15 23:58:59 +00:00
|
|
|
{
|
2018-02-15 11:14:56 +00:00
|
|
|
return (READ_ONCE(pte_val(*ptep)) & PTE_S2_RDWR) == PTE_S2_RDONLY;
|
2015-01-15 23:58:59 +00:00
|
|
|
}
|
|
|
|
|
2018-02-15 11:14:56 +00:00
|
|
|
static inline bool kvm_s2pte_exec(pte_t *ptep)
|
2017-10-23 16:11:21 +00:00
|
|
|
{
|
2018-02-15 11:14:56 +00:00
|
|
|
return !(READ_ONCE(pte_val(*ptep)) & PTE_S2_XN);
|
2017-10-23 16:11:21 +00:00
|
|
|
}
|
|
|
|
|
2018-02-15 11:14:56 +00:00
|
|
|
static inline void kvm_set_s2pmd_readonly(pmd_t *pmdp)
|
2015-01-15 23:58:59 +00:00
|
|
|
{
|
2018-02-15 11:14:56 +00:00
|
|
|
kvm_set_s2pte_readonly((pte_t *)pmdp);
|
2015-01-15 23:58:59 +00:00
|
|
|
}
|
|
|
|
|
2018-02-15 11:14:56 +00:00
|
|
|
static inline bool kvm_s2pmd_readonly(pmd_t *pmdp)
|
2015-01-15 23:58:59 +00:00
|
|
|
{
|
2018-02-15 11:14:56 +00:00
|
|
|
return kvm_s2pte_readonly((pte_t *)pmdp);
|
2014-10-10 10:14:28 +00:00
|
|
|
}
|
|
|
|
|
2018-02-15 11:14:56 +00:00
|
|
|
static inline bool kvm_s2pmd_exec(pmd_t *pmdp)
|
2017-10-23 16:11:21 +00:00
|
|
|
{
|
2018-02-15 11:14:56 +00:00
|
|
|
return !(READ_ONCE(pmd_val(*pmdp)) & PMD_S2_XN);
|
2017-10-23 16:11:21 +00:00
|
|
|
}
|
|
|
|
|
2014-05-09 21:31:31 +00:00
|
|
|
static inline bool kvm_page_empty(void *ptr)
|
|
|
|
{
|
|
|
|
struct page *ptr_page = virt_to_page(ptr);
|
|
|
|
return page_count(ptr_page) == 1;
|
|
|
|
}
|
|
|
|
|
2016-03-22 17:20:28 +00:00
|
|
|
#define hyp_pte_table_empty(ptep) kvm_page_empty(ptep)
|
2014-10-10 10:14:28 +00:00
|
|
|
|
|
|
|
#ifdef __PAGETABLE_PMD_FOLDED
|
2016-03-22 17:20:28 +00:00
|
|
|
#define hyp_pmd_table_empty(pmdp) (0)
|
2014-10-10 10:14:28 +00:00
|
|
|
#else
|
2016-03-22 17:20:28 +00:00
|
|
|
#define hyp_pmd_table_empty(pmdp) kvm_page_empty(pmdp)
|
2014-10-10 10:14:28 +00:00
|
|
|
#endif
|
|
|
|
|
|
|
|
#ifdef __PAGETABLE_PUD_FOLDED
|
2016-03-22 17:20:28 +00:00
|
|
|
#define hyp_pud_table_empty(pudp) (0)
|
2014-05-09 21:31:31 +00:00
|
|
|
#else
|
2016-03-22 17:20:28 +00:00
|
|
|
#define hyp_pud_table_empty(pudp) kvm_page_empty(pudp)
|
2014-05-09 21:31:31 +00:00
|
|
|
#endif
|
|
|
|
|
2012-12-10 15:35:24 +00:00
|
|
|
struct kvm;
|
|
|
|
|
2014-01-14 19:13:10 +00:00
|
|
|
#define kvm_flush_dcache_to_poc(a,l) __flush_dcache_area((a), (l))
|
|
|
|
|
|
|
|
static inline bool vcpu_has_cache_enabled(struct kvm_vcpu *vcpu)
|
2012-12-10 15:35:24 +00:00
|
|
|
{
|
2014-01-14 19:13:10 +00:00
|
|
|
return (vcpu_sys_reg(vcpu, SCTLR_EL1) & 0b101) == 0b101;
|
|
|
|
}
|
|
|
|
|
2017-10-23 16:11:22 +00:00
|
|
|
static inline void __clean_dcache_guest_page(kvm_pfn_t pfn, unsigned long size)
|
2014-01-14 19:13:10 +00:00
|
|
|
{
|
arm/arm64: KVM: Use kernel mapping to perform invalidation on page fault
When handling a fault in stage-2, we need to resync I$ and D$, just
to be sure we don't leave any old cache line behind.
That's very good, except that we do so using the *user* address.
Under heavy load (swapping like crazy), we may end up in a situation
where the page gets mapped in stage-2 while being unmapped from
userspace by another CPU.
At that point, the DC/IC instructions can generate a fault, which
we handle with kvm->mmu_lock held. The box quickly deadlocks, user
is unhappy.
Instead, perform this invalidation through the kernel mapping,
which is guaranteed to be present. The box is much happier, and so
am I.
Signed-off-by: Marc Zyngier <marc.zyngier@arm.com>
Signed-off-by: Christoffer Dall <christoffer.dall@linaro.org>
2015-01-05 21:13:24 +00:00
|
|
|
void *va = page_address(pfn_to_page(pfn));
|
|
|
|
|
2017-01-25 12:29:59 +00:00
|
|
|
kvm_flush_dcache_to_poc(va, size);
|
2017-10-23 16:11:15 +00:00
|
|
|
}
|
2014-01-14 19:13:10 +00:00
|
|
|
|
2017-10-23 16:11:22 +00:00
|
|
|
static inline void __invalidate_icache_guest_page(kvm_pfn_t pfn,
|
2017-10-23 16:11:15 +00:00
|
|
|
unsigned long size)
|
|
|
|
{
|
2017-03-10 20:32:25 +00:00
|
|
|
if (icache_is_aliasing()) {
|
2012-12-10 15:35:24 +00:00
|
|
|
/* any kind of VIPT cache */
|
|
|
|
__flush_icache_all();
|
2017-03-10 20:32:25 +00:00
|
|
|
} else if (is_kernel_in_hyp_mode() || !icache_is_vpipt()) {
|
|
|
|
/* PIPT or VPIPT at EL2 (see comment in __kvm_tlb_flush_vmid_ipa) */
|
2017-10-23 16:11:15 +00:00
|
|
|
void *va = page_address(pfn_to_page(pfn));
|
|
|
|
|
2017-10-23 16:11:16 +00:00
|
|
|
invalidate_icache_range((unsigned long)va,
|
|
|
|
(unsigned long)va + size);
|
2012-12-10 15:35:24 +00:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2014-12-19 16:48:06 +00:00
|
|
|
static inline void __kvm_flush_dcache_pte(pte_t pte)
|
|
|
|
{
|
|
|
|
struct page *page = pte_page(pte);
|
|
|
|
kvm_flush_dcache_to_poc(page_address(page), PAGE_SIZE);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void __kvm_flush_dcache_pmd(pmd_t pmd)
|
|
|
|
{
|
|
|
|
struct page *page = pmd_page(pmd);
|
|
|
|
kvm_flush_dcache_to_poc(page_address(page), PMD_SIZE);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void __kvm_flush_dcache_pud(pud_t pud)
|
|
|
|
{
|
|
|
|
struct page *page = pud_page(pud);
|
|
|
|
kvm_flush_dcache_to_poc(page_address(page), PUD_SIZE);
|
|
|
|
}
|
|
|
|
|
2017-01-10 21:35:49 +00:00
|
|
|
#define kvm_virt_to_phys(x) __pa_symbol(x)
|
2012-12-10 15:35:24 +00:00
|
|
|
|
2014-12-19 16:05:31 +00:00
|
|
|
void kvm_set_way_flush(struct kvm_vcpu *vcpu);
|
|
|
|
void kvm_toggle_cache(struct kvm_vcpu *vcpu, bool was_enabled);
|
2014-01-15 12:50:23 +00:00
|
|
|
|
2015-03-19 16:42:28 +00:00
|
|
|
static inline bool __kvm_cpu_uses_extended_idmap(void)
|
|
|
|
{
|
arm64: allow ID map to be extended to 52 bits
Currently, when using VA_BITS < 48, if the ID map text happens to be
placed in physical memory above VA_BITS, we increase the VA size (up to
48) and create a new table level, in order to map in the ID map text.
This is okay because the system always supports 48 bits of VA.
This patch extends the code such that if the system supports 52 bits of
VA, and the ID map text is placed that high up, then we increase the VA
size accordingly, up to 52.
One difference from the current implementation is that so far the
condition of VA_BITS < 48 has meant that the top level table is always
"full", with the maximum number of entries, and an extra table level is
always needed. Now, when VA_BITS = 48 (and using 64k pages), the top
level table is not full, and we simply need to increase the number of
entries in it, instead of creating a new table level.
Tested-by: Suzuki K Poulose <suzuki.poulose@arm.com>
Reviewed-by: Suzuki K Poulose <suzuki.poulose@arm.com>
Reviewed-by: Marc Zyngier <marc.zyngier@arm.com>
Tested-by: Bob Picco <bob.picco@oracle.com>
Reviewed-by: Bob Picco <bob.picco@oracle.com>
Signed-off-by: Kristina Martsenko <kristina.martsenko@arm.com>
[catalin.marinas@arm.com: reduce arguments to __create_hyp_mappings()]
[catalin.marinas@arm.com: reworked/renamed __cpu_uses_extended_idmap_level()]
Signed-off-by: Catalin Marinas <catalin.marinas@arm.com>
2017-12-13 17:07:24 +00:00
|
|
|
return __cpu_uses_extended_idmap_level();
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline unsigned long __kvm_idmap_ptrs_per_pgd(void)
|
|
|
|
{
|
|
|
|
return idmap_ptrs_per_pgd;
|
2015-03-19 16:42:28 +00:00
|
|
|
}
|
|
|
|
|
2017-12-13 17:07:20 +00:00
|
|
|
/*
|
|
|
|
* Can't use pgd_populate here, because the extended idmap adds an extra level
|
|
|
|
* above CONFIG_PGTABLE_LEVELS (which is 2 or 3 if we're using the extended
|
|
|
|
* idmap), and pgd_populate is only available if CONFIG_PGTABLE_LEVELS = 4.
|
|
|
|
*/
|
2015-03-19 16:42:28 +00:00
|
|
|
static inline void __kvm_extend_hypmap(pgd_t *boot_hyp_pgd,
|
|
|
|
pgd_t *hyp_pgd,
|
|
|
|
pgd_t *merged_hyp_pgd,
|
|
|
|
unsigned long hyp_idmap_start)
|
|
|
|
{
|
|
|
|
int idmap_idx;
|
2017-12-13 17:07:21 +00:00
|
|
|
u64 pgd_addr;
|
2015-03-19 16:42:28 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Use the first entry to access the HYP mappings. It is
|
|
|
|
* guaranteed to be free, otherwise we wouldn't use an
|
|
|
|
* extended idmap.
|
|
|
|
*/
|
|
|
|
VM_BUG_ON(pgd_val(merged_hyp_pgd[0]));
|
2017-12-13 17:07:21 +00:00
|
|
|
pgd_addr = __phys_to_pgd_val(__pa(hyp_pgd));
|
|
|
|
merged_hyp_pgd[0] = __pgd(pgd_addr | PMD_TYPE_TABLE);
|
2015-03-19 16:42:28 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Create another extended level entry that points to the boot HYP map,
|
|
|
|
* which contains an ID mapping of the HYP init code. We essentially
|
|
|
|
* merge the boot and runtime HYP maps by doing so, but they don't
|
|
|
|
* overlap anyway, so this is fine.
|
|
|
|
*/
|
|
|
|
idmap_idx = hyp_idmap_start >> VA_BITS;
|
|
|
|
VM_BUG_ON(pgd_val(merged_hyp_pgd[idmap_idx]));
|
2017-12-13 17:07:21 +00:00
|
|
|
pgd_addr = __phys_to_pgd_val(__pa(boot_hyp_pgd));
|
|
|
|
merged_hyp_pgd[idmap_idx] = __pgd(pgd_addr | PMD_TYPE_TABLE);
|
2015-03-19 16:42:28 +00:00
|
|
|
}
|
|
|
|
|
2015-11-16 11:28:18 +00:00
|
|
|
static inline unsigned int kvm_get_vmid_bits(void)
|
|
|
|
{
|
2017-03-23 15:14:39 +00:00
|
|
|
int reg = read_sanitised_ftr_reg(SYS_ID_AA64MMFR1_EL1);
|
2015-11-16 11:28:18 +00:00
|
|
|
|
2016-01-26 10:58:16 +00:00
|
|
|
return (cpuid_feature_extract_unsigned_field(reg, ID_AA64MMFR1_VMIDBITS_SHIFT) == 2) ? 16 : 8;
|
2015-11-16 11:28:18 +00:00
|
|
|
}
|
|
|
|
|
2018-01-03 16:38:35 +00:00
|
|
|
#ifdef CONFIG_HARDEN_BRANCH_PREDICTOR
|
|
|
|
#include <asm/mmu.h>
|
|
|
|
|
|
|
|
static inline void *kvm_get_hyp_vector(void)
|
|
|
|
{
|
|
|
|
struct bp_hardening_data *data = arm64_get_bp_hardening_data();
|
|
|
|
void *vect = kvm_ksym_ref(__kvm_hyp_vector);
|
|
|
|
|
|
|
|
if (data->fn) {
|
|
|
|
vect = __bp_harden_hyp_vecs_start +
|
|
|
|
data->hyp_vectors_slot * SZ_2K;
|
|
|
|
|
|
|
|
if (!has_vhe())
|
|
|
|
vect = lm_alias(vect);
|
|
|
|
}
|
|
|
|
|
|
|
|
return vect;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int kvm_map_vectors(void)
|
|
|
|
{
|
|
|
|
return create_hyp_mappings(kvm_ksym_ref(__bp_harden_hyp_vecs_start),
|
|
|
|
kvm_ksym_ref(__bp_harden_hyp_vecs_end),
|
|
|
|
PAGE_HYP_EXEC);
|
|
|
|
}
|
|
|
|
|
|
|
|
#else
|
|
|
|
static inline void *kvm_get_hyp_vector(void)
|
|
|
|
{
|
|
|
|
return kvm_ksym_ref(__kvm_hyp_vector);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int kvm_map_vectors(void)
|
|
|
|
{
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2017-12-13 17:07:18 +00:00
|
|
|
#define kvm_phys_to_vttbr(addr) phys_to_ttbr(addr)
|
|
|
|
|
2012-12-10 15:35:24 +00:00
|
|
|
#endif /* __ASSEMBLY__ */
|
|
|
|
#endif /* __ARM64_KVM_MMU_H__ */
|