2016-04-29 13:26:05 +00:00
|
|
|
/*
|
|
|
|
* TLB flush routines for radix kernels.
|
|
|
|
*
|
|
|
|
* Copyright 2015-2016, Aneesh Kumar K.V, IBM Corporation.
|
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or
|
|
|
|
* modify it under the terms of the GNU General Public License
|
|
|
|
* as published by the Free Software Foundation; either version
|
|
|
|
* 2 of the License, or (at your option) any later version.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/mm.h>
|
|
|
|
#include <linux/hugetlb.h>
|
|
|
|
#include <linux/memblock.h>
|
|
|
|
|
powerpc/mm/radix: Workaround prefetch issue with KVM
There's a somewhat architectural issue with Radix MMU and KVM.
When coming out of a guest with AIL (Alternate Interrupt Location, ie,
MMU enabled), we start executing hypervisor code with the PID register
still containing whatever the guest has been using.
The problem is that the CPU can (and will) then start prefetching or
speculatively load from whatever host context has that same PID (if
any), thus bringing translations for that context into the TLB, which
Linux doesn't know about.
This can cause stale translations and subsequent crashes.
Fixing this in a way that is neither racy nor a huge performance
impact is difficult. We could just make the host invalidations always
use broadcast forms but that would hurt single threaded programs for
example.
We chose to fix it instead by partitioning the PID space between guest
and host. This is possible because today Linux only use 19 out of the
20 bits of PID space, so existing guests will work if we make the host
use the top half of the 20 bits space.
We additionally add support for a property to indicate to Linux the
size of the PID register which will be useful if we eventually have
processors with a larger PID space available.
There is still an issue with malicious guests purposefully setting the
PID register to a value in the hosts PID range. Hopefully future HW
can prevent that, but in the meantime, we handle it with a pair of
kludges:
- On the way out of a guest, before we clear the current VCPU in the
PACA, we check the PID and if it's outside of the permitted range
we flush the TLB for that PID.
- When context switching, if the mm is "new" on that CPU (the
corresponding bit was set for the first time in the mm cpumask), we
check if any sibling thread is in KVM (has a non-NULL VCPU pointer
in the PACA). If that is the case, we also flush the PID for that
CPU (core).
This second part is needed to handle the case where a process is
migrated (or starts a new pthread) on a sibling thread of the CPU
coming out of KVM, as there's a window where stale translations can
exist before we detect it and flush them out.
A future optimization could be added by keeping track of whether the
PID has ever been used and avoid doing that for completely fresh PIDs.
We could similarily mark PIDs that have been the subject of a global
invalidation as "fresh". But for now this will do.
Signed-off-by: Benjamin Herrenschmidt <benh@kernel.crashing.org>
[mpe: Rework the asm to build with CONFIG_PPC_RADIX_MMU=n, drop
unneeded include of kvm_book3s_asm.h]
Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
2017-07-24 04:26:06 +00:00
|
|
|
#include <asm/ppc-opcode.h>
|
2016-04-29 13:26:05 +00:00
|
|
|
#include <asm/tlb.h>
|
|
|
|
#include <asm/tlbflush.h>
|
2017-04-11 05:23:25 +00:00
|
|
|
#include <asm/trace.h>
|
powerpc/mm/radix: Workaround prefetch issue with KVM
There's a somewhat architectural issue with Radix MMU and KVM.
When coming out of a guest with AIL (Alternate Interrupt Location, ie,
MMU enabled), we start executing hypervisor code with the PID register
still containing whatever the guest has been using.
The problem is that the CPU can (and will) then start prefetching or
speculatively load from whatever host context has that same PID (if
any), thus bringing translations for that context into the TLB, which
Linux doesn't know about.
This can cause stale translations and subsequent crashes.
Fixing this in a way that is neither racy nor a huge performance
impact is difficult. We could just make the host invalidations always
use broadcast forms but that would hurt single threaded programs for
example.
We chose to fix it instead by partitioning the PID space between guest
and host. This is possible because today Linux only use 19 out of the
20 bits of PID space, so existing guests will work if we make the host
use the top half of the 20 bits space.
We additionally add support for a property to indicate to Linux the
size of the PID register which will be useful if we eventually have
processors with a larger PID space available.
There is still an issue with malicious guests purposefully setting the
PID register to a value in the hosts PID range. Hopefully future HW
can prevent that, but in the meantime, we handle it with a pair of
kludges:
- On the way out of a guest, before we clear the current VCPU in the
PACA, we check the PID and if it's outside of the permitted range
we flush the TLB for that PID.
- When context switching, if the mm is "new" on that CPU (the
corresponding bit was set for the first time in the mm cpumask), we
check if any sibling thread is in KVM (has a non-NULL VCPU pointer
in the PACA). If that is the case, we also flush the PID for that
CPU (core).
This second part is needed to handle the case where a process is
migrated (or starts a new pthread) on a sibling thread of the CPU
coming out of KVM, as there's a window where stale translations can
exist before we detect it and flush them out.
A future optimization could be added by keeping track of whether the
PID has ever been used and avoid doing that for completely fresh PIDs.
We could similarily mark PIDs that have been the subject of a global
invalidation as "fresh". But for now this will do.
Signed-off-by: Benjamin Herrenschmidt <benh@kernel.crashing.org>
[mpe: Rework the asm to build with CONFIG_PPC_RADIX_MMU=n, drop
unneeded include of kvm_book3s_asm.h]
Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
2017-07-24 04:26:06 +00:00
|
|
|
#include <asm/cputhreads.h>
|
2016-04-29 13:26:05 +00:00
|
|
|
|
2016-06-08 14:25:50 +00:00
|
|
|
#define RIC_FLUSH_TLB 0
|
|
|
|
#define RIC_FLUSH_PWC 1
|
|
|
|
#define RIC_FLUSH_ALL 2
|
|
|
|
|
|
|
|
static inline void __tlbiel_pid(unsigned long pid, int set,
|
|
|
|
unsigned long ric)
|
2016-04-29 13:26:05 +00:00
|
|
|
{
|
2016-06-08 14:25:50 +00:00
|
|
|
unsigned long rb,rs,prs,r;
|
2016-04-29 13:26:05 +00:00
|
|
|
|
|
|
|
rb = PPC_BIT(53); /* IS = 1 */
|
|
|
|
rb |= set << PPC_BITLSHIFT(51);
|
|
|
|
rs = ((unsigned long)pid) << PPC_BITLSHIFT(31);
|
|
|
|
prs = 1; /* process scoped */
|
|
|
|
r = 1; /* raidx format */
|
|
|
|
|
2016-07-13 09:35:20 +00:00
|
|
|
asm volatile(PPC_TLBIEL(%0, %4, %3, %2, %1)
|
2016-04-29 13:26:05 +00:00
|
|
|
: : "r"(rb), "i"(r), "i"(prs), "i"(ric), "r"(rs) : "memory");
|
2017-04-11 05:23:25 +00:00
|
|
|
trace_tlbie(0, 1, rb, rs, ric, prs, r);
|
2016-04-29 13:26:05 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We use 128 set in radix mode and 256 set in hpt mode.
|
|
|
|
*/
|
2016-06-08 14:25:50 +00:00
|
|
|
static inline void _tlbiel_pid(unsigned long pid, unsigned long ric)
|
2016-04-29 13:26:05 +00:00
|
|
|
{
|
|
|
|
int set;
|
|
|
|
|
2017-04-01 14:41:48 +00:00
|
|
|
asm volatile("ptesync": : :"memory");
|
2017-04-26 11:38:17 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Flush the first set of the TLB, and if we're doing a RIC_FLUSH_ALL,
|
|
|
|
* also flush the entire Page Walk Cache.
|
|
|
|
*/
|
|
|
|
__tlbiel_pid(pid, 0, ric);
|
|
|
|
|
2017-07-19 04:49:04 +00:00
|
|
|
/* For PWC, only one flush is needed */
|
|
|
|
if (ric == RIC_FLUSH_PWC) {
|
|
|
|
asm volatile("ptesync": : :"memory");
|
|
|
|
return;
|
|
|
|
}
|
2017-04-26 11:38:17 +00:00
|
|
|
|
2017-07-19 04:49:04 +00:00
|
|
|
/* For the remaining sets, just flush the TLB */
|
2017-04-26 11:38:17 +00:00
|
|
|
for (set = 1; set < POWER9_TLB_SETS_RADIX ; set++)
|
2017-07-19 04:49:04 +00:00
|
|
|
__tlbiel_pid(pid, set, RIC_FLUSH_TLB);
|
2017-04-26 11:38:17 +00:00
|
|
|
|
2017-04-01 14:41:48 +00:00
|
|
|
asm volatile("ptesync": : :"memory");
|
2017-02-06 02:05:16 +00:00
|
|
|
asm volatile(PPC_INVALIDATE_ERAT "; isync" : : :"memory");
|
2016-04-29 13:26:05 +00:00
|
|
|
}
|
|
|
|
|
2016-06-08 14:25:50 +00:00
|
|
|
static inline void _tlbie_pid(unsigned long pid, unsigned long ric)
|
2016-04-29 13:26:05 +00:00
|
|
|
{
|
2016-06-08 14:25:50 +00:00
|
|
|
unsigned long rb,rs,prs,r;
|
2016-04-29 13:26:05 +00:00
|
|
|
|
|
|
|
rb = PPC_BIT(53); /* IS = 1 */
|
|
|
|
rs = pid << PPC_BITLSHIFT(31);
|
|
|
|
prs = 1; /* process scoped */
|
|
|
|
r = 1; /* raidx format */
|
|
|
|
|
|
|
|
asm volatile("ptesync": : :"memory");
|
2016-07-13 09:35:20 +00:00
|
|
|
asm volatile(PPC_TLBIE_5(%0, %4, %3, %2, %1)
|
2016-04-29 13:26:05 +00:00
|
|
|
: : "r"(rb), "i"(r), "i"(prs), "i"(ric), "r"(rs) : "memory");
|
|
|
|
asm volatile("eieio; tlbsync; ptesync": : :"memory");
|
2017-04-11 05:23:25 +00:00
|
|
|
trace_tlbie(0, 0, rb, rs, ric, prs, r);
|
2016-04-29 13:26:05 +00:00
|
|
|
}
|
|
|
|
|
2017-11-07 07:53:05 +00:00
|
|
|
static inline void __tlbiel_va(unsigned long va, unsigned long pid,
|
2017-11-07 07:53:06 +00:00
|
|
|
unsigned long ap, unsigned long ric)
|
2016-04-29 13:26:05 +00:00
|
|
|
{
|
2016-06-08 14:25:50 +00:00
|
|
|
unsigned long rb,rs,prs,r;
|
2016-04-29 13:26:05 +00:00
|
|
|
|
|
|
|
rb = va & ~(PPC_BITMASK(52, 63));
|
|
|
|
rb |= ap << PPC_BITLSHIFT(58);
|
|
|
|
rs = pid << PPC_BITLSHIFT(31);
|
|
|
|
prs = 1; /* process scoped */
|
|
|
|
r = 1; /* raidx format */
|
|
|
|
|
2016-07-13 09:35:20 +00:00
|
|
|
asm volatile(PPC_TLBIEL(%0, %4, %3, %2, %1)
|
2016-04-29 13:26:05 +00:00
|
|
|
: : "r"(rb), "i"(r), "i"(prs), "i"(ric), "r"(rs) : "memory");
|
2017-04-11 05:23:25 +00:00
|
|
|
trace_tlbie(0, 1, rb, rs, ric, prs, r);
|
2016-04-29 13:26:05 +00:00
|
|
|
}
|
|
|
|
|
2017-11-07 07:53:05 +00:00
|
|
|
static inline void _tlbiel_va(unsigned long va, unsigned long pid,
|
2017-11-07 07:53:06 +00:00
|
|
|
unsigned long psize, unsigned long ric)
|
2017-11-07 07:53:05 +00:00
|
|
|
{
|
2017-11-07 07:53:06 +00:00
|
|
|
unsigned long ap = mmu_get_ap(psize);
|
|
|
|
|
2017-11-07 07:53:05 +00:00
|
|
|
asm volatile("ptesync": : :"memory");
|
|
|
|
__tlbiel_va(va, pid, ap, ric);
|
|
|
|
asm volatile("ptesync": : :"memory");
|
|
|
|
}
|
|
|
|
|
2017-11-07 07:53:06 +00:00
|
|
|
static inline void _tlbiel_va_range(unsigned long start, unsigned long end,
|
|
|
|
unsigned long pid, unsigned long page_size,
|
|
|
|
unsigned long psize)
|
|
|
|
{
|
|
|
|
unsigned long addr;
|
|
|
|
unsigned long ap = mmu_get_ap(psize);
|
|
|
|
|
|
|
|
asm volatile("ptesync": : :"memory");
|
|
|
|
for (addr = start; addr < end; addr += page_size)
|
|
|
|
__tlbiel_va(addr, pid, ap, RIC_FLUSH_TLB);
|
|
|
|
asm volatile("ptesync": : :"memory");
|
|
|
|
}
|
|
|
|
|
2017-11-07 07:53:05 +00:00
|
|
|
static inline void __tlbie_va(unsigned long va, unsigned long pid,
|
2016-06-08 14:25:50 +00:00
|
|
|
unsigned long ap, unsigned long ric)
|
2016-04-29 13:26:05 +00:00
|
|
|
{
|
2016-06-08 14:25:50 +00:00
|
|
|
unsigned long rb,rs,prs,r;
|
2016-04-29 13:26:05 +00:00
|
|
|
|
|
|
|
rb = va & ~(PPC_BITMASK(52, 63));
|
|
|
|
rb |= ap << PPC_BITLSHIFT(58);
|
|
|
|
rs = pid << PPC_BITLSHIFT(31);
|
|
|
|
prs = 1; /* process scoped */
|
|
|
|
r = 1; /* raidx format */
|
|
|
|
|
2016-07-13 09:35:20 +00:00
|
|
|
asm volatile(PPC_TLBIE_5(%0, %4, %3, %2, %1)
|
2016-04-29 13:26:05 +00:00
|
|
|
: : "r"(rb), "i"(r), "i"(prs), "i"(ric), "r"(rs) : "memory");
|
2017-04-11 05:23:25 +00:00
|
|
|
trace_tlbie(0, 0, rb, rs, ric, prs, r);
|
2016-04-29 13:26:05 +00:00
|
|
|
}
|
|
|
|
|
2017-11-07 07:53:05 +00:00
|
|
|
static inline void _tlbie_va(unsigned long va, unsigned long pid,
|
2017-11-07 07:53:06 +00:00
|
|
|
unsigned long psize, unsigned long ric)
|
2017-11-07 07:53:05 +00:00
|
|
|
{
|
2017-11-07 07:53:06 +00:00
|
|
|
unsigned long ap = mmu_get_ap(psize);
|
|
|
|
|
2017-11-07 07:53:05 +00:00
|
|
|
asm volatile("ptesync": : :"memory");
|
|
|
|
__tlbie_va(va, pid, ap, ric);
|
|
|
|
asm volatile("eieio; tlbsync; ptesync": : :"memory");
|
|
|
|
}
|
|
|
|
|
2017-11-07 07:53:06 +00:00
|
|
|
static inline void _tlbie_va_range(unsigned long start, unsigned long end,
|
|
|
|
unsigned long pid, unsigned long page_size,
|
|
|
|
unsigned long psize)
|
|
|
|
{
|
|
|
|
unsigned long addr;
|
|
|
|
unsigned long ap = mmu_get_ap(psize);
|
|
|
|
|
|
|
|
asm volatile("ptesync": : :"memory");
|
|
|
|
for (addr = start; addr < end; addr += page_size)
|
|
|
|
__tlbie_va(addr, pid, ap, RIC_FLUSH_TLB);
|
|
|
|
asm volatile("eieio; tlbsync; ptesync": : :"memory");
|
|
|
|
}
|
2017-11-07 07:53:05 +00:00
|
|
|
|
2016-04-29 13:26:05 +00:00
|
|
|
/*
|
|
|
|
* Base TLB flushing operations:
|
|
|
|
*
|
|
|
|
* - flush_tlb_mm(mm) flushes the specified mm context TLB's
|
|
|
|
* - flush_tlb_page(vma, vmaddr) flushes one page
|
|
|
|
* - flush_tlb_range(vma, start, end) flushes a range of pages
|
|
|
|
* - flush_tlb_kernel_range(start, end) flushes kernel pages
|
|
|
|
*
|
|
|
|
* - local_* variants of page and mm only apply to the current
|
|
|
|
* processor
|
|
|
|
*/
|
|
|
|
void radix__local_flush_tlb_mm(struct mm_struct *mm)
|
|
|
|
{
|
2016-06-02 09:44:48 +00:00
|
|
|
unsigned long pid;
|
2016-04-29 13:26:05 +00:00
|
|
|
|
|
|
|
preempt_disable();
|
|
|
|
pid = mm->context.id;
|
|
|
|
if (pid != MMU_NO_CONTEXT)
|
2017-07-19 04:49:05 +00:00
|
|
|
_tlbiel_pid(pid, RIC_FLUSH_TLB);
|
2016-04-29 13:26:05 +00:00
|
|
|
preempt_enable();
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(radix__local_flush_tlb_mm);
|
|
|
|
|
2017-07-19 04:49:05 +00:00
|
|
|
#ifndef CONFIG_SMP
|
2017-09-03 18:15:12 +00:00
|
|
|
void radix__local_flush_all_mm(struct mm_struct *mm)
|
2016-06-08 14:25:51 +00:00
|
|
|
{
|
|
|
|
unsigned long pid;
|
|
|
|
|
|
|
|
preempt_disable();
|
|
|
|
pid = mm->context.id;
|
|
|
|
if (pid != MMU_NO_CONTEXT)
|
2017-07-19 04:49:05 +00:00
|
|
|
_tlbiel_pid(pid, RIC_FLUSH_ALL);
|
2016-06-08 14:25:51 +00:00
|
|
|
preempt_enable();
|
|
|
|
}
|
2017-09-03 18:15:12 +00:00
|
|
|
EXPORT_SYMBOL(radix__local_flush_all_mm);
|
2017-07-19 04:49:05 +00:00
|
|
|
#endif /* CONFIG_SMP */
|
2016-06-08 14:25:51 +00:00
|
|
|
|
2016-07-13 09:36:41 +00:00
|
|
|
void radix__local_flush_tlb_page_psize(struct mm_struct *mm, unsigned long vmaddr,
|
2016-07-13 09:36:42 +00:00
|
|
|
int psize)
|
2016-04-29 13:26:05 +00:00
|
|
|
{
|
2016-06-02 09:44:48 +00:00
|
|
|
unsigned long pid;
|
2016-04-29 13:26:05 +00:00
|
|
|
|
|
|
|
preempt_disable();
|
2017-10-16 07:11:00 +00:00
|
|
|
pid = mm->context.id;
|
2016-04-29 13:26:05 +00:00
|
|
|
if (pid != MMU_NO_CONTEXT)
|
2017-11-07 07:53:06 +00:00
|
|
|
_tlbiel_va(vmaddr, pid, psize, RIC_FLUSH_TLB);
|
2016-04-29 13:26:05 +00:00
|
|
|
preempt_enable();
|
|
|
|
}
|
|
|
|
|
|
|
|
void radix__local_flush_tlb_page(struct vm_area_struct *vma, unsigned long vmaddr)
|
|
|
|
{
|
2016-04-29 13:26:25 +00:00
|
|
|
#ifdef CONFIG_HUGETLB_PAGE
|
|
|
|
/* need the return fix for nohash.c */
|
2017-10-16 07:11:00 +00:00
|
|
|
if (is_vm_hugetlb_page(vma))
|
|
|
|
return radix__local_flush_hugetlb_page(vma, vmaddr);
|
2016-04-29 13:26:25 +00:00
|
|
|
#endif
|
2017-10-16 07:11:00 +00:00
|
|
|
radix__local_flush_tlb_page_psize(vma->vm_mm, vmaddr, mmu_virtual_psize);
|
2016-04-29 13:26:05 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(radix__local_flush_tlb_page);
|
|
|
|
|
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
void radix__flush_tlb_mm(struct mm_struct *mm)
|
|
|
|
{
|
2016-06-02 09:44:48 +00:00
|
|
|
unsigned long pid;
|
2016-04-29 13:26:05 +00:00
|
|
|
|
|
|
|
pid = mm->context.id;
|
|
|
|
if (unlikely(pid == MMU_NO_CONTEXT))
|
2017-10-24 13:06:53 +00:00
|
|
|
return;
|
2016-04-29 13:26:05 +00:00
|
|
|
|
2017-10-24 13:06:53 +00:00
|
|
|
preempt_disable();
|
2017-05-02 11:00:14 +00:00
|
|
|
if (!mm_is_thread_local(mm))
|
2017-07-19 04:49:05 +00:00
|
|
|
_tlbie_pid(pid, RIC_FLUSH_TLB);
|
2017-05-02 11:00:14 +00:00
|
|
|
else
|
2017-07-19 04:49:05 +00:00
|
|
|
_tlbiel_pid(pid, RIC_FLUSH_TLB);
|
2016-04-29 13:26:05 +00:00
|
|
|
preempt_enable();
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(radix__flush_tlb_mm);
|
|
|
|
|
2017-09-03 18:15:12 +00:00
|
|
|
void radix__flush_all_mm(struct mm_struct *mm)
|
2016-06-08 14:25:51 +00:00
|
|
|
{
|
|
|
|
unsigned long pid;
|
|
|
|
|
|
|
|
pid = mm->context.id;
|
|
|
|
if (unlikely(pid == MMU_NO_CONTEXT))
|
2017-10-24 13:06:53 +00:00
|
|
|
return;
|
2016-06-08 14:25:51 +00:00
|
|
|
|
2017-10-24 13:06:53 +00:00
|
|
|
preempt_disable();
|
2017-05-02 11:00:14 +00:00
|
|
|
if (!mm_is_thread_local(mm))
|
2017-07-19 04:49:05 +00:00
|
|
|
_tlbie_pid(pid, RIC_FLUSH_ALL);
|
2017-05-02 11:00:14 +00:00
|
|
|
else
|
2017-07-19 04:49:05 +00:00
|
|
|
_tlbiel_pid(pid, RIC_FLUSH_ALL);
|
2016-06-08 14:25:51 +00:00
|
|
|
preempt_enable();
|
|
|
|
}
|
2017-09-03 18:15:12 +00:00
|
|
|
EXPORT_SYMBOL(radix__flush_all_mm);
|
2017-07-19 04:49:05 +00:00
|
|
|
|
|
|
|
void radix__flush_tlb_pwc(struct mmu_gather *tlb, unsigned long addr)
|
|
|
|
{
|
|
|
|
tlb->need_flush_all = 1;
|
|
|
|
}
|
2016-06-08 14:25:51 +00:00
|
|
|
EXPORT_SYMBOL(radix__flush_tlb_pwc);
|
|
|
|
|
2016-07-13 09:36:41 +00:00
|
|
|
void radix__flush_tlb_page_psize(struct mm_struct *mm, unsigned long vmaddr,
|
2016-07-13 09:36:42 +00:00
|
|
|
int psize)
|
2016-04-29 13:26:05 +00:00
|
|
|
{
|
2016-06-02 09:44:48 +00:00
|
|
|
unsigned long pid;
|
2016-04-29 13:26:05 +00:00
|
|
|
|
2017-10-16 07:11:00 +00:00
|
|
|
pid = mm->context.id;
|
2016-04-29 13:26:05 +00:00
|
|
|
if (unlikely(pid == MMU_NO_CONTEXT))
|
2017-10-24 13:06:53 +00:00
|
|
|
return;
|
|
|
|
|
|
|
|
preempt_disable();
|
2017-05-02 11:00:14 +00:00
|
|
|
if (!mm_is_thread_local(mm))
|
2017-11-07 07:53:06 +00:00
|
|
|
_tlbie_va(vmaddr, pid, psize, RIC_FLUSH_TLB);
|
2017-05-02 11:00:14 +00:00
|
|
|
else
|
2017-11-07 07:53:06 +00:00
|
|
|
_tlbiel_va(vmaddr, pid, psize, RIC_FLUSH_TLB);
|
2016-04-29 13:26:05 +00:00
|
|
|
preempt_enable();
|
|
|
|
}
|
|
|
|
|
|
|
|
void radix__flush_tlb_page(struct vm_area_struct *vma, unsigned long vmaddr)
|
|
|
|
{
|
2016-04-29 13:26:25 +00:00
|
|
|
#ifdef CONFIG_HUGETLB_PAGE
|
2017-10-16 07:11:00 +00:00
|
|
|
if (is_vm_hugetlb_page(vma))
|
|
|
|
return radix__flush_hugetlb_page(vma, vmaddr);
|
2016-04-29 13:26:25 +00:00
|
|
|
#endif
|
2017-10-16 07:11:00 +00:00
|
|
|
radix__flush_tlb_page_psize(vma->vm_mm, vmaddr, mmu_virtual_psize);
|
2016-04-29 13:26:05 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(radix__flush_tlb_page);
|
|
|
|
|
2017-07-19 04:49:05 +00:00
|
|
|
#else /* CONFIG_SMP */
|
|
|
|
#define radix__flush_all_mm radix__local_flush_all_mm
|
2016-04-29 13:26:05 +00:00
|
|
|
#endif /* CONFIG_SMP */
|
|
|
|
|
|
|
|
void radix__flush_tlb_kernel_range(unsigned long start, unsigned long end)
|
|
|
|
{
|
2016-06-08 14:25:50 +00:00
|
|
|
_tlbie_pid(0, RIC_FLUSH_ALL);
|
2016-04-29 13:26:05 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(radix__flush_tlb_kernel_range);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Currently, for range flushing, we just do a full mm flush. Because
|
|
|
|
* we use this in code path where we don' track the page size.
|
|
|
|
*/
|
|
|
|
void radix__flush_tlb_range(struct vm_area_struct *vma, unsigned long start,
|
|
|
|
unsigned long end)
|
|
|
|
|
|
|
|
{
|
|
|
|
struct mm_struct *mm = vma->vm_mm;
|
2017-07-19 04:49:05 +00:00
|
|
|
|
2017-07-19 04:49:06 +00:00
|
|
|
radix__flush_tlb_mm(mm);
|
2016-04-29 13:26:05 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(radix__flush_tlb_range);
|
|
|
|
|
2016-07-13 09:35:29 +00:00
|
|
|
static int radix_get_mmu_psize(int page_size)
|
|
|
|
{
|
|
|
|
int psize;
|
|
|
|
|
|
|
|
if (page_size == (1UL << mmu_psize_defs[mmu_virtual_psize].shift))
|
|
|
|
psize = mmu_virtual_psize;
|
|
|
|
else if (page_size == (1UL << mmu_psize_defs[MMU_PAGE_2M].shift))
|
|
|
|
psize = MMU_PAGE_2M;
|
|
|
|
else if (page_size == (1UL << mmu_psize_defs[MMU_PAGE_1G].shift))
|
|
|
|
psize = MMU_PAGE_1G;
|
|
|
|
else
|
|
|
|
return -1;
|
|
|
|
return psize;
|
|
|
|
}
|
2016-04-29 13:26:05 +00:00
|
|
|
|
|
|
|
void radix__tlb_flush(struct mmu_gather *tlb)
|
|
|
|
{
|
2016-07-13 09:36:35 +00:00
|
|
|
int psize = 0;
|
2016-04-29 13:26:05 +00:00
|
|
|
struct mm_struct *mm = tlb->mm;
|
2016-07-13 09:36:35 +00:00
|
|
|
int page_size = tlb->page_size;
|
|
|
|
|
|
|
|
psize = radix_get_mmu_psize(page_size);
|
|
|
|
/*
|
|
|
|
* if page size is not something we understand, do a full mm flush
|
2017-10-24 13:06:54 +00:00
|
|
|
*
|
|
|
|
* A "fullmm" flush must always do a flush_all_mm (RIC=2) flush
|
|
|
|
* that flushes the process table entry cache upon process teardown.
|
|
|
|
* See the comment for radix in arch_exit_mmap().
|
2016-07-13 09:36:35 +00:00
|
|
|
*/
|
|
|
|
if (psize != -1 && !tlb->fullmm && !tlb->need_flush_all)
|
|
|
|
radix__flush_tlb_range_psize(mm, tlb->start, tlb->end, psize);
|
2017-10-24 13:06:54 +00:00
|
|
|
else if (tlb->fullmm || tlb->need_flush_all) {
|
2017-07-19 04:49:05 +00:00
|
|
|
tlb->need_flush_all = 0;
|
|
|
|
radix__flush_all_mm(mm);
|
|
|
|
} else
|
2016-07-13 09:36:35 +00:00
|
|
|
radix__flush_tlb_mm(mm);
|
|
|
|
}
|
|
|
|
|
|
|
|
#define TLB_FLUSH_ALL -1UL
|
|
|
|
/*
|
|
|
|
* Number of pages above which we will do a bcast tlbie. Just a
|
|
|
|
* number at this point copied from x86
|
|
|
|
*/
|
|
|
|
static unsigned long tlb_single_page_flush_ceiling __read_mostly = 33;
|
|
|
|
|
|
|
|
void radix__flush_tlb_range_psize(struct mm_struct *mm, unsigned long start,
|
|
|
|
unsigned long end, int psize)
|
|
|
|
{
|
|
|
|
unsigned long pid;
|
2017-10-24 13:06:53 +00:00
|
|
|
bool local;
|
2016-07-13 09:36:35 +00:00
|
|
|
unsigned long page_size = 1UL << mmu_psize_defs[psize].shift;
|
|
|
|
|
2017-10-16 07:11:00 +00:00
|
|
|
pid = mm->context.id;
|
2016-07-13 09:36:35 +00:00
|
|
|
if (unlikely(pid == MMU_NO_CONTEXT))
|
2017-10-24 13:06:53 +00:00
|
|
|
return;
|
2016-07-13 09:36:35 +00:00
|
|
|
|
2017-10-24 13:06:53 +00:00
|
|
|
preempt_disable();
|
|
|
|
local = mm_is_thread_local(mm);
|
2016-07-13 09:36:35 +00:00
|
|
|
if (end == TLB_FLUSH_ALL ||
|
|
|
|
(end - start) > tlb_single_page_flush_ceiling * page_size) {
|
|
|
|
if (local)
|
|
|
|
_tlbiel_pid(pid, RIC_FLUSH_TLB);
|
|
|
|
else
|
|
|
|
_tlbie_pid(pid, RIC_FLUSH_TLB);
|
2017-10-24 13:06:53 +00:00
|
|
|
} else {
|
2017-11-07 07:53:05 +00:00
|
|
|
if (local)
|
2017-11-07 07:53:06 +00:00
|
|
|
_tlbiel_va_range(start, end, pid, page_size, psize);
|
2017-11-07 07:53:05 +00:00
|
|
|
else
|
2017-11-07 07:53:06 +00:00
|
|
|
_tlbie_va_range(start, end, pid, page_size, psize);
|
2016-07-13 09:36:35 +00:00
|
|
|
}
|
|
|
|
preempt_enable();
|
2016-04-29 13:26:05 +00:00
|
|
|
}
|
2016-07-13 09:35:29 +00:00
|
|
|
|
2017-07-19 04:49:06 +00:00
|
|
|
#ifdef CONFIG_TRANSPARENT_HUGEPAGE
|
|
|
|
void radix__flush_tlb_collapsed_pmd(struct mm_struct *mm, unsigned long addr)
|
|
|
|
{
|
|
|
|
unsigned long pid, end;
|
2017-10-24 13:06:53 +00:00
|
|
|
bool local;
|
2017-07-19 04:49:06 +00:00
|
|
|
|
2017-10-16 07:11:00 +00:00
|
|
|
pid = mm->context.id;
|
2017-07-19 04:49:06 +00:00
|
|
|
if (unlikely(pid == MMU_NO_CONTEXT))
|
2017-10-24 13:06:53 +00:00
|
|
|
return;
|
2017-07-19 04:49:06 +00:00
|
|
|
|
|
|
|
/* 4k page size, just blow the world */
|
|
|
|
if (PAGE_SIZE == 0x1000) {
|
|
|
|
radix__flush_all_mm(mm);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2017-10-24 13:06:53 +00:00
|
|
|
preempt_disable();
|
|
|
|
local = mm_is_thread_local(mm);
|
2017-07-19 04:49:06 +00:00
|
|
|
/* Otherwise first do the PWC */
|
|
|
|
if (local)
|
|
|
|
_tlbiel_pid(pid, RIC_FLUSH_PWC);
|
|
|
|
else
|
|
|
|
_tlbie_pid(pid, RIC_FLUSH_PWC);
|
|
|
|
|
|
|
|
/* Then iterate the pages */
|
|
|
|
end = addr + HPAGE_PMD_SIZE;
|
2017-11-07 07:53:05 +00:00
|
|
|
if (local)
|
2017-11-07 07:53:06 +00:00
|
|
|
_tlbiel_va_range(addr, end, pid, PAGE_SIZE, mmu_virtual_psize);
|
2017-11-07 07:53:05 +00:00
|
|
|
else
|
2017-11-07 07:53:06 +00:00
|
|
|
_tlbie_va_range(addr, end, pid, PAGE_SIZE, mmu_virtual_psize);
|
2017-11-07 07:53:05 +00:00
|
|
|
|
2017-07-19 04:49:06 +00:00
|
|
|
preempt_enable();
|
|
|
|
}
|
|
|
|
#endif /* CONFIG_TRANSPARENT_HUGEPAGE */
|
|
|
|
|
2016-07-13 09:35:29 +00:00
|
|
|
void radix__flush_tlb_lpid_va(unsigned long lpid, unsigned long gpa,
|
|
|
|
unsigned long page_size)
|
|
|
|
{
|
|
|
|
unsigned long rb,rs,prs,r;
|
|
|
|
unsigned long ap;
|
|
|
|
unsigned long ric = RIC_FLUSH_TLB;
|
|
|
|
|
|
|
|
ap = mmu_get_ap(radix_get_mmu_psize(page_size));
|
|
|
|
rb = gpa & ~(PPC_BITMASK(52, 63));
|
|
|
|
rb |= ap << PPC_BITLSHIFT(58);
|
|
|
|
rs = lpid & ((1UL << 32) - 1);
|
|
|
|
prs = 0; /* process scoped */
|
|
|
|
r = 1; /* raidx format */
|
|
|
|
|
|
|
|
asm volatile("ptesync": : :"memory");
|
|
|
|
asm volatile(PPC_TLBIE_5(%0, %4, %3, %2, %1)
|
|
|
|
: : "r"(rb), "i"(r), "i"(prs), "i"(ric), "r"(rs) : "memory");
|
|
|
|
asm volatile("eieio; tlbsync; ptesync": : :"memory");
|
2017-04-11 05:23:25 +00:00
|
|
|
trace_tlbie(lpid, 0, rb, rs, ric, prs, r);
|
2016-07-13 09:35:29 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(radix__flush_tlb_lpid_va);
|
|
|
|
|
|
|
|
void radix__flush_tlb_lpid(unsigned long lpid)
|
|
|
|
{
|
|
|
|
unsigned long rb,rs,prs,r;
|
|
|
|
unsigned long ric = RIC_FLUSH_ALL;
|
|
|
|
|
|
|
|
rb = 0x2 << PPC_BITLSHIFT(53); /* IS = 2 */
|
|
|
|
rs = lpid & ((1UL << 32) - 1);
|
|
|
|
prs = 0; /* partition scoped */
|
|
|
|
r = 1; /* raidx format */
|
|
|
|
|
|
|
|
asm volatile("ptesync": : :"memory");
|
|
|
|
asm volatile(PPC_TLBIE_5(%0, %4, %3, %2, %1)
|
|
|
|
: : "r"(rb), "i"(r), "i"(prs), "i"(ric), "r"(rs) : "memory");
|
|
|
|
asm volatile("eieio; tlbsync; ptesync": : :"memory");
|
2017-04-11 05:23:25 +00:00
|
|
|
trace_tlbie(lpid, 0, rb, rs, ric, prs, r);
|
2016-07-13 09:35:29 +00:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(radix__flush_tlb_lpid);
|
2016-07-13 09:36:40 +00:00
|
|
|
|
|
|
|
void radix__flush_pmd_tlb_range(struct vm_area_struct *vma,
|
|
|
|
unsigned long start, unsigned long end)
|
|
|
|
{
|
|
|
|
radix__flush_tlb_range_psize(vma->vm_mm, start, end, MMU_PAGE_2M);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(radix__flush_pmd_tlb_range);
|
2016-08-23 10:57:48 +00:00
|
|
|
|
|
|
|
void radix__flush_tlb_all(void)
|
|
|
|
{
|
|
|
|
unsigned long rb,prs,r,rs;
|
|
|
|
unsigned long ric = RIC_FLUSH_ALL;
|
|
|
|
|
|
|
|
rb = 0x3 << PPC_BITLSHIFT(53); /* IS = 3 */
|
|
|
|
prs = 0; /* partition scoped */
|
|
|
|
r = 1; /* raidx format */
|
|
|
|
rs = 1 & ((1UL << 32) - 1); /* any LPID value to flush guest mappings */
|
|
|
|
|
|
|
|
asm volatile("ptesync": : :"memory");
|
|
|
|
/*
|
|
|
|
* now flush guest entries by passing PRS = 1 and LPID != 0
|
|
|
|
*/
|
|
|
|
asm volatile(PPC_TLBIE_5(%0, %4, %3, %2, %1)
|
|
|
|
: : "r"(rb), "i"(r), "i"(1), "i"(ric), "r"(rs) : "memory");
|
2017-04-11 05:23:25 +00:00
|
|
|
trace_tlbie(0, 0, rb, rs, ric, prs, r);
|
2016-08-23 10:57:48 +00:00
|
|
|
/*
|
|
|
|
* now flush host entires by passing PRS = 0 and LPID == 0
|
|
|
|
*/
|
|
|
|
asm volatile(PPC_TLBIE_5(%0, %4, %3, %2, %1)
|
|
|
|
: : "r"(rb), "i"(r), "i"(prs), "i"(ric), "r"(0) : "memory");
|
|
|
|
asm volatile("eieio; tlbsync; ptesync": : :"memory");
|
2017-04-11 05:23:25 +00:00
|
|
|
trace_tlbie(0, 0, rb, 0, ric, prs, r);
|
2016-08-23 10:57:48 +00:00
|
|
|
}
|
2016-11-28 06:17:01 +00:00
|
|
|
|
|
|
|
void radix__flush_tlb_pte_p9_dd1(unsigned long old_pte, struct mm_struct *mm,
|
|
|
|
unsigned long address)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
* We track page size in pte only for DD1, So we can
|
|
|
|
* call this only on DD1.
|
|
|
|
*/
|
|
|
|
if (!cpu_has_feature(CPU_FTR_POWER9_DD1)) {
|
|
|
|
VM_WARN_ON(1);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2017-03-21 17:29:54 +00:00
|
|
|
if (old_pte & R_PAGE_LARGE)
|
2016-11-28 06:17:01 +00:00
|
|
|
radix__flush_tlb_page_psize(mm, address, MMU_PAGE_2M);
|
|
|
|
else
|
|
|
|
radix__flush_tlb_page_psize(mm, address, mmu_virtual_psize);
|
|
|
|
}
|
powerpc/mm/radix: Workaround prefetch issue with KVM
There's a somewhat architectural issue with Radix MMU and KVM.
When coming out of a guest with AIL (Alternate Interrupt Location, ie,
MMU enabled), we start executing hypervisor code with the PID register
still containing whatever the guest has been using.
The problem is that the CPU can (and will) then start prefetching or
speculatively load from whatever host context has that same PID (if
any), thus bringing translations for that context into the TLB, which
Linux doesn't know about.
This can cause stale translations and subsequent crashes.
Fixing this in a way that is neither racy nor a huge performance
impact is difficult. We could just make the host invalidations always
use broadcast forms but that would hurt single threaded programs for
example.
We chose to fix it instead by partitioning the PID space between guest
and host. This is possible because today Linux only use 19 out of the
20 bits of PID space, so existing guests will work if we make the host
use the top half of the 20 bits space.
We additionally add support for a property to indicate to Linux the
size of the PID register which will be useful if we eventually have
processors with a larger PID space available.
There is still an issue with malicious guests purposefully setting the
PID register to a value in the hosts PID range. Hopefully future HW
can prevent that, but in the meantime, we handle it with a pair of
kludges:
- On the way out of a guest, before we clear the current VCPU in the
PACA, we check the PID and if it's outside of the permitted range
we flush the TLB for that PID.
- When context switching, if the mm is "new" on that CPU (the
corresponding bit was set for the first time in the mm cpumask), we
check if any sibling thread is in KVM (has a non-NULL VCPU pointer
in the PACA). If that is the case, we also flush the PID for that
CPU (core).
This second part is needed to handle the case where a process is
migrated (or starts a new pthread) on a sibling thread of the CPU
coming out of KVM, as there's a window where stale translations can
exist before we detect it and flush them out.
A future optimization could be added by keeping track of whether the
PID has ever been used and avoid doing that for completely fresh PIDs.
We could similarily mark PIDs that have been the subject of a global
invalidation as "fresh". But for now this will do.
Signed-off-by: Benjamin Herrenschmidt <benh@kernel.crashing.org>
[mpe: Rework the asm to build with CONFIG_PPC_RADIX_MMU=n, drop
unneeded include of kvm_book3s_asm.h]
Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
2017-07-24 04:26:06 +00:00
|
|
|
|
|
|
|
#ifdef CONFIG_KVM_BOOK3S_HV_POSSIBLE
|
|
|
|
extern void radix_kvm_prefetch_workaround(struct mm_struct *mm)
|
|
|
|
{
|
|
|
|
unsigned int pid = mm->context.id;
|
|
|
|
|
|
|
|
if (unlikely(pid == MMU_NO_CONTEXT))
|
|
|
|
return;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If this context hasn't run on that CPU before and KVM is
|
|
|
|
* around, there's a slim chance that the guest on another
|
|
|
|
* CPU just brought in obsolete translation into the TLB of
|
|
|
|
* this CPU due to a bad prefetch using the guest PID on
|
|
|
|
* the way into the hypervisor.
|
|
|
|
*
|
|
|
|
* We work around this here. If KVM is possible, we check if
|
|
|
|
* any sibling thread is in KVM. If it is, the window may exist
|
|
|
|
* and thus we flush that PID from the core.
|
|
|
|
*
|
|
|
|
* A potential future improvement would be to mark which PIDs
|
|
|
|
* have never been used on the system and avoid it if the PID
|
|
|
|
* is new and the process has no other cpumask bit set.
|
|
|
|
*/
|
|
|
|
if (cpu_has_feature(CPU_FTR_HVMODE) && radix_enabled()) {
|
|
|
|
int cpu = smp_processor_id();
|
|
|
|
int sib = cpu_first_thread_sibling(cpu);
|
|
|
|
bool flush = false;
|
|
|
|
|
|
|
|
for (; sib <= cpu_last_thread_sibling(cpu) && !flush; sib++) {
|
|
|
|
if (sib == cpu)
|
|
|
|
continue;
|
|
|
|
if (paca[sib].kvm_hstate.kvm_vcpu)
|
|
|
|
flush = true;
|
|
|
|
}
|
|
|
|
if (flush)
|
|
|
|
_tlbiel_pid(pid, RIC_FLUSH_ALL);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(radix_kvm_prefetch_workaround);
|
|
|
|
#endif /* CONFIG_KVM_BOOK3S_HV_POSSIBLE */
|