03f3e54abd
2nd DAWR can be set/unset using H_SET_MODE hcall with resource value 5. Enable powervm guest support with that. This has no effect on kvm guest because kvm will return error if guest does hcall with resource value 5. Signed-off-by: Ravi Bangoria <ravi.bangoria@linux.ibm.com> Signed-off-by: Michael Ellerman <mpe@ellerman.id.au> Link: https://lore.kernel.org/r/20200723090813.303838-9-ravi.bangoria@linux.ibm.com
409 lines
9.8 KiB
C
409 lines
9.8 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
#ifndef _ASM_POWERPC_PLPAR_WRAPPERS_H
|
|
#define _ASM_POWERPC_PLPAR_WRAPPERS_H
|
|
|
|
#ifdef CONFIG_PPC_PSERIES
|
|
|
|
#include <linux/string.h>
|
|
#include <linux/irqflags.h>
|
|
|
|
#include <asm/hvcall.h>
|
|
#include <asm/paca.h>
|
|
#include <asm/page.h>
|
|
|
|
static inline long poll_pending(void)
|
|
{
|
|
return plpar_hcall_norets(H_POLL_PENDING);
|
|
}
|
|
|
|
static inline u8 get_cede_latency_hint(void)
|
|
{
|
|
return get_lppaca()->cede_latency_hint;
|
|
}
|
|
|
|
static inline void set_cede_latency_hint(u8 latency_hint)
|
|
{
|
|
get_lppaca()->cede_latency_hint = latency_hint;
|
|
}
|
|
|
|
static inline long cede_processor(void)
|
|
{
|
|
return plpar_hcall_norets(H_CEDE);
|
|
}
|
|
|
|
static inline long extended_cede_processor(unsigned long latency_hint)
|
|
{
|
|
long rc;
|
|
u8 old_latency_hint = get_cede_latency_hint();
|
|
|
|
set_cede_latency_hint(latency_hint);
|
|
|
|
rc = cede_processor();
|
|
#ifdef CONFIG_PPC_IRQ_SOFT_MASK_DEBUG
|
|
/* Ensure that H_CEDE returns with IRQs on */
|
|
if (WARN_ON(!(mfmsr() & MSR_EE)))
|
|
__hard_irq_enable();
|
|
#endif
|
|
|
|
set_cede_latency_hint(old_latency_hint);
|
|
|
|
return rc;
|
|
}
|
|
|
|
static inline long vpa_call(unsigned long flags, unsigned long cpu,
|
|
unsigned long vpa)
|
|
{
|
|
flags = flags << H_VPA_FUNC_SHIFT;
|
|
|
|
return plpar_hcall_norets(H_REGISTER_VPA, flags, cpu, vpa);
|
|
}
|
|
|
|
static inline long unregister_vpa(unsigned long cpu)
|
|
{
|
|
return vpa_call(H_VPA_DEREG_VPA, cpu, 0);
|
|
}
|
|
|
|
static inline long register_vpa(unsigned long cpu, unsigned long vpa)
|
|
{
|
|
return vpa_call(H_VPA_REG_VPA, cpu, vpa);
|
|
}
|
|
|
|
static inline long unregister_slb_shadow(unsigned long cpu)
|
|
{
|
|
return vpa_call(H_VPA_DEREG_SLB, cpu, 0);
|
|
}
|
|
|
|
static inline long register_slb_shadow(unsigned long cpu, unsigned long vpa)
|
|
{
|
|
return vpa_call(H_VPA_REG_SLB, cpu, vpa);
|
|
}
|
|
|
|
static inline long unregister_dtl(unsigned long cpu)
|
|
{
|
|
return vpa_call(H_VPA_DEREG_DTL, cpu, 0);
|
|
}
|
|
|
|
static inline long register_dtl(unsigned long cpu, unsigned long vpa)
|
|
{
|
|
return vpa_call(H_VPA_REG_DTL, cpu, vpa);
|
|
}
|
|
|
|
extern void vpa_init(int cpu);
|
|
|
|
static inline long plpar_pte_enter(unsigned long flags,
|
|
unsigned long hpte_group, unsigned long hpte_v,
|
|
unsigned long hpte_r, unsigned long *slot)
|
|
{
|
|
long rc;
|
|
unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
|
|
|
|
rc = plpar_hcall(H_ENTER, retbuf, flags, hpte_group, hpte_v, hpte_r);
|
|
|
|
*slot = retbuf[0];
|
|
|
|
return rc;
|
|
}
|
|
|
|
static inline long plpar_pte_remove(unsigned long flags, unsigned long ptex,
|
|
unsigned long avpn, unsigned long *old_pteh_ret,
|
|
unsigned long *old_ptel_ret)
|
|
{
|
|
long rc;
|
|
unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
|
|
|
|
rc = plpar_hcall(H_REMOVE, retbuf, flags, ptex, avpn);
|
|
|
|
*old_pteh_ret = retbuf[0];
|
|
*old_ptel_ret = retbuf[1];
|
|
|
|
return rc;
|
|
}
|
|
|
|
/* plpar_pte_remove_raw can be called in real mode. It calls plpar_hcall_raw */
|
|
static inline long plpar_pte_remove_raw(unsigned long flags, unsigned long ptex,
|
|
unsigned long avpn, unsigned long *old_pteh_ret,
|
|
unsigned long *old_ptel_ret)
|
|
{
|
|
long rc;
|
|
unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
|
|
|
|
rc = plpar_hcall_raw(H_REMOVE, retbuf, flags, ptex, avpn);
|
|
|
|
*old_pteh_ret = retbuf[0];
|
|
*old_ptel_ret = retbuf[1];
|
|
|
|
return rc;
|
|
}
|
|
|
|
static inline long plpar_pte_read(unsigned long flags, unsigned long ptex,
|
|
unsigned long *old_pteh_ret, unsigned long *old_ptel_ret)
|
|
{
|
|
long rc;
|
|
unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
|
|
|
|
rc = plpar_hcall(H_READ, retbuf, flags, ptex);
|
|
|
|
*old_pteh_ret = retbuf[0];
|
|
*old_ptel_ret = retbuf[1];
|
|
|
|
return rc;
|
|
}
|
|
|
|
/* plpar_pte_read_raw can be called in real mode. It calls plpar_hcall_raw */
|
|
static inline long plpar_pte_read_raw(unsigned long flags, unsigned long ptex,
|
|
unsigned long *old_pteh_ret, unsigned long *old_ptel_ret)
|
|
{
|
|
long rc;
|
|
unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
|
|
|
|
rc = plpar_hcall_raw(H_READ, retbuf, flags, ptex);
|
|
|
|
*old_pteh_ret = retbuf[0];
|
|
*old_ptel_ret = retbuf[1];
|
|
|
|
return rc;
|
|
}
|
|
|
|
/*
|
|
* ptes must be 8*sizeof(unsigned long)
|
|
*/
|
|
static inline long plpar_pte_read_4(unsigned long flags, unsigned long ptex,
|
|
unsigned long *ptes)
|
|
|
|
{
|
|
long rc;
|
|
unsigned long retbuf[PLPAR_HCALL9_BUFSIZE];
|
|
|
|
rc = plpar_hcall9(H_READ, retbuf, flags | H_READ_4, ptex);
|
|
|
|
memcpy(ptes, retbuf, 8*sizeof(unsigned long));
|
|
|
|
return rc;
|
|
}
|
|
|
|
/*
|
|
* plpar_pte_read_4_raw can be called in real mode.
|
|
* ptes must be 8*sizeof(unsigned long)
|
|
*/
|
|
static inline long plpar_pte_read_4_raw(unsigned long flags, unsigned long ptex,
|
|
unsigned long *ptes)
|
|
|
|
{
|
|
long rc;
|
|
unsigned long retbuf[PLPAR_HCALL9_BUFSIZE];
|
|
|
|
rc = plpar_hcall9_raw(H_READ, retbuf, flags | H_READ_4, ptex);
|
|
|
|
memcpy(ptes, retbuf, 8*sizeof(unsigned long));
|
|
|
|
return rc;
|
|
}
|
|
|
|
static inline long plpar_pte_protect(unsigned long flags, unsigned long ptex,
|
|
unsigned long avpn)
|
|
{
|
|
return plpar_hcall_norets(H_PROTECT, flags, ptex, avpn);
|
|
}
|
|
|
|
static inline long plpar_resize_hpt_prepare(unsigned long flags,
|
|
unsigned long shift)
|
|
{
|
|
return plpar_hcall_norets(H_RESIZE_HPT_PREPARE, flags, shift);
|
|
}
|
|
|
|
static inline long plpar_resize_hpt_commit(unsigned long flags,
|
|
unsigned long shift)
|
|
{
|
|
return plpar_hcall_norets(H_RESIZE_HPT_COMMIT, flags, shift);
|
|
}
|
|
|
|
static inline long plpar_tce_get(unsigned long liobn, unsigned long ioba,
|
|
unsigned long *tce_ret)
|
|
{
|
|
long rc;
|
|
unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
|
|
|
|
rc = plpar_hcall(H_GET_TCE, retbuf, liobn, ioba);
|
|
|
|
*tce_ret = retbuf[0];
|
|
|
|
return rc;
|
|
}
|
|
|
|
static inline long plpar_tce_put(unsigned long liobn, unsigned long ioba,
|
|
unsigned long tceval)
|
|
{
|
|
return plpar_hcall_norets(H_PUT_TCE, liobn, ioba, tceval);
|
|
}
|
|
|
|
static inline long plpar_tce_put_indirect(unsigned long liobn,
|
|
unsigned long ioba, unsigned long page, unsigned long count)
|
|
{
|
|
return plpar_hcall_norets(H_PUT_TCE_INDIRECT, liobn, ioba, page, count);
|
|
}
|
|
|
|
static inline long plpar_tce_stuff(unsigned long liobn, unsigned long ioba,
|
|
unsigned long tceval, unsigned long count)
|
|
{
|
|
return plpar_hcall_norets(H_STUFF_TCE, liobn, ioba, tceval, count);
|
|
}
|
|
|
|
/* Set various resource mode parameters */
|
|
static inline long plpar_set_mode(unsigned long mflags, unsigned long resource,
|
|
unsigned long value1, unsigned long value2)
|
|
{
|
|
return plpar_hcall_norets(H_SET_MODE, mflags, resource, value1, value2);
|
|
}
|
|
|
|
/*
|
|
* Enable relocation on exceptions on this partition
|
|
*
|
|
* Note: this call has a partition wide scope and can take a while to complete.
|
|
* If it returns H_LONG_BUSY_* it should be retried periodically until it
|
|
* returns H_SUCCESS.
|
|
*/
|
|
static inline long enable_reloc_on_exceptions(void)
|
|
{
|
|
/* mflags = 3: Exceptions at 0xC000000000004000 */
|
|
return plpar_set_mode(3, H_SET_MODE_RESOURCE_ADDR_TRANS_MODE, 0, 0);
|
|
}
|
|
|
|
/*
|
|
* Disable relocation on exceptions on this partition
|
|
*
|
|
* Note: this call has a partition wide scope and can take a while to complete.
|
|
* If it returns H_LONG_BUSY_* it should be retried periodically until it
|
|
* returns H_SUCCESS.
|
|
*/
|
|
static inline long disable_reloc_on_exceptions(void) {
|
|
return plpar_set_mode(0, H_SET_MODE_RESOURCE_ADDR_TRANS_MODE, 0, 0);
|
|
}
|
|
|
|
/*
|
|
* Take exceptions in big endian mode on this partition
|
|
*
|
|
* Note: this call has a partition wide scope and can take a while to complete.
|
|
* If it returns H_LONG_BUSY_* it should be retried periodically until it
|
|
* returns H_SUCCESS.
|
|
*/
|
|
static inline long enable_big_endian_exceptions(void)
|
|
{
|
|
/* mflags = 0: big endian exceptions */
|
|
return plpar_set_mode(0, H_SET_MODE_RESOURCE_LE, 0, 0);
|
|
}
|
|
|
|
/*
|
|
* Take exceptions in little endian mode on this partition
|
|
*
|
|
* Note: this call has a partition wide scope and can take a while to complete.
|
|
* If it returns H_LONG_BUSY_* it should be retried periodically until it
|
|
* returns H_SUCCESS.
|
|
*/
|
|
static inline long enable_little_endian_exceptions(void)
|
|
{
|
|
/* mflags = 1: little endian exceptions */
|
|
return plpar_set_mode(1, H_SET_MODE_RESOURCE_LE, 0, 0);
|
|
}
|
|
|
|
static inline long plpar_set_ciabr(unsigned long ciabr)
|
|
{
|
|
return plpar_set_mode(0, H_SET_MODE_RESOURCE_SET_CIABR, ciabr, 0);
|
|
}
|
|
|
|
static inline long plpar_set_watchpoint0(unsigned long dawr0, unsigned long dawrx0)
|
|
{
|
|
return plpar_set_mode(0, H_SET_MODE_RESOURCE_SET_DAWR0, dawr0, dawrx0);
|
|
}
|
|
|
|
static inline long plpar_set_watchpoint1(unsigned long dawr1, unsigned long dawrx1)
|
|
{
|
|
return plpar_set_mode(0, H_SET_MODE_RESOURCE_SET_DAWR1, dawr1, dawrx1);
|
|
}
|
|
|
|
static inline long plpar_signal_sys_reset(long cpu)
|
|
{
|
|
return plpar_hcall_norets(H_SIGNAL_SYS_RESET, cpu);
|
|
}
|
|
|
|
static inline long plpar_get_cpu_characteristics(struct h_cpu_char_result *p)
|
|
{
|
|
unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
|
|
long rc;
|
|
|
|
rc = plpar_hcall(H_GET_CPU_CHARACTERISTICS, retbuf);
|
|
if (rc == H_SUCCESS) {
|
|
p->character = retbuf[0];
|
|
p->behaviour = retbuf[1];
|
|
}
|
|
|
|
return rc;
|
|
}
|
|
|
|
/*
|
|
* Wrapper to H_RPT_INVALIDATE hcall that handles return values appropriately
|
|
*
|
|
* - Returns H_SUCCESS on success
|
|
* - For H_BUSY return value, we retry the hcall.
|
|
* - For any other hcall failures, attempt a full flush once before
|
|
* resorting to BUG().
|
|
*
|
|
* Note: This hcall is expected to fail only very rarely. The correct
|
|
* error recovery of killing the process/guest will be eventually
|
|
* needed.
|
|
*/
|
|
static inline long pseries_rpt_invalidate(u32 pid, u64 target, u64 type,
|
|
u64 page_sizes, u64 start, u64 end)
|
|
{
|
|
long rc;
|
|
unsigned long all;
|
|
|
|
while (true) {
|
|
rc = plpar_hcall_norets(H_RPT_INVALIDATE, pid, target, type,
|
|
page_sizes, start, end);
|
|
if (rc == H_BUSY) {
|
|
cpu_relax();
|
|
continue;
|
|
} else if (rc == H_SUCCESS)
|
|
return rc;
|
|
|
|
/* Flush request failed, try with a full flush once */
|
|
if (type & H_RPTI_TYPE_NESTED)
|
|
all = H_RPTI_TYPE_NESTED | H_RPTI_TYPE_NESTED_ALL;
|
|
else
|
|
all = H_RPTI_TYPE_ALL;
|
|
retry:
|
|
rc = plpar_hcall_norets(H_RPT_INVALIDATE, pid, target,
|
|
all, page_sizes, 0, -1UL);
|
|
if (rc == H_BUSY) {
|
|
cpu_relax();
|
|
goto retry;
|
|
} else if (rc == H_SUCCESS)
|
|
return rc;
|
|
|
|
BUG();
|
|
}
|
|
}
|
|
|
|
#else /* !CONFIG_PPC_PSERIES */
|
|
|
|
static inline long plpar_set_ciabr(unsigned long ciabr)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline long plpar_pte_read_4(unsigned long flags, unsigned long ptex,
|
|
unsigned long *ptes)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
static inline long pseries_rpt_invalidate(u32 pid, u64 target, u64 type,
|
|
u64 page_sizes, u64 start, u64 end)
|
|
{
|
|
return 0;
|
|
}
|
|
|
|
#endif /* CONFIG_PPC_PSERIES */
|
|
|
|
#endif /* _ASM_POWERPC_PLPAR_WRAPPERS_H */
|