forked from Minki/linux
powerpc: Handle most loads and stores in instruction emulation code
This extends the instruction emulation infrastructure in sstep.c to handle all the load and store instructions defined in the Power ISA v3.0, except for the atomic memory operations, ldmx (which was never implemented), lfdp/stfdp, and the vector element load/stores. The instructions added are: Integer loads and stores: lbarx, lharx, lqarx, stbcx., sthcx., stqcx., lq, stq. VSX loads and stores: lxsiwzx, lxsiwax, stxsiwx, lxvx, lxvl, lxvll, lxvdsx, lxvwsx, stxvx, stxvl, stxvll, lxsspx, lxsdx, stxsspx, stxsdx, lxvw4x, lxsibzx, lxvh8x, lxsihzx, lxvb16x, stxvw4x, stxsibx, stxvh8x, stxsihx, stxvb16x, lxsd, lxssp, lxv, stxsd, stxssp, stxv. These instructions are handled both in the analyse_instr phase and in the emulate_step phase. The code for lxvd2ux and stxvd2ux has been taken out, as those instructions were never implemented in any processor and have been taken out of the architecture, and their opcodes have been reused for other instructions in POWER9 (lxvb16x and stxvb16x). The emulation for the VSX loads and stores uses helper functions which don't access registers or memory directly, which can hopefully be reused by KVM later. Signed-off-by: Paul Mackerras <paulus@ozlabs.org> Signed-off-by: Michael Ellerman <mpe@ellerman.id.au>
This commit is contained in:
parent
ee0a54d797
commit
350779a29f
@ -205,6 +205,8 @@
|
||||
#define PPC_INST_ISEL_MASK 0xfc00003e
|
||||
#define PPC_INST_LDARX 0x7c0000a8
|
||||
#define PPC_INST_STDCX 0x7c0001ad
|
||||
#define PPC_INST_LQARX 0x7c000228
|
||||
#define PPC_INST_STQCX 0x7c00016d
|
||||
#define PPC_INST_LSWI 0x7c0004aa
|
||||
#define PPC_INST_LSWX 0x7c00042a
|
||||
#define PPC_INST_LWARX 0x7c000028
|
||||
@ -403,12 +405,18 @@
|
||||
__PPC_RA(a) | __PPC_RB(b))
|
||||
#define PPC_DCBZL(a, b) stringify_in_c(.long PPC_INST_DCBZL | \
|
||||
__PPC_RA(a) | __PPC_RB(b))
|
||||
#define PPC_LQARX(t, a, b, eh) stringify_in_c(.long PPC_INST_LQARX | \
|
||||
___PPC_RT(t) | ___PPC_RA(a) | \
|
||||
___PPC_RB(b) | __PPC_EH(eh))
|
||||
#define PPC_LDARX(t, a, b, eh) stringify_in_c(.long PPC_INST_LDARX | \
|
||||
___PPC_RT(t) | ___PPC_RA(a) | \
|
||||
___PPC_RB(b) | __PPC_EH(eh))
|
||||
#define PPC_LWARX(t, a, b, eh) stringify_in_c(.long PPC_INST_LWARX | \
|
||||
___PPC_RT(t) | ___PPC_RA(a) | \
|
||||
___PPC_RB(b) | __PPC_EH(eh))
|
||||
#define PPC_STQCX(t, a, b) stringify_in_c(.long PPC_INST_STQCX | \
|
||||
___PPC_RT(t) | ___PPC_RA(a) | \
|
||||
___PPC_RB(b))
|
||||
#define PPC_MSGSND(b) stringify_in_c(.long PPC_INST_MSGSND | \
|
||||
___PPC_RB(b))
|
||||
#define PPC_MSGSYNC stringify_in_c(.long PPC_INST_MSGSYNC)
|
||||
|
@ -83,6 +83,12 @@ enum instruction_type {
|
||||
#define DCBT 0x300
|
||||
#define ICBI 0x400
|
||||
|
||||
/* VSX flags values */
|
||||
#define VSX_FPCONV 1 /* do floating point SP/DP conversion */
|
||||
#define VSX_SPLAT 2 /* store loaded value into all elements */
|
||||
#define VSX_LDLEFT 4 /* load VSX register from left */
|
||||
#define VSX_CHECK_VEC 8 /* check MSR_VEC not MSR_VSX for reg >= 32 */
|
||||
|
||||
/* Size field in type word */
|
||||
#define SIZE(n) ((n) << 8)
|
||||
#define GETSIZE(w) ((w) >> 8)
|
||||
@ -100,6 +106,17 @@ struct instruction_op {
|
||||
int spr;
|
||||
u32 ccval;
|
||||
u32 xerval;
|
||||
u8 element_size; /* for VSX/VMX loads/stores */
|
||||
u8 vsx_flags;
|
||||
};
|
||||
|
||||
union vsx_reg {
|
||||
u8 b[16];
|
||||
u16 h[8];
|
||||
u32 w[4];
|
||||
unsigned long d[2];
|
||||
float fp[4];
|
||||
double dp[2];
|
||||
};
|
||||
|
||||
/*
|
||||
@ -131,3 +148,7 @@ void emulate_update_regs(struct pt_regs *reg, struct instruction_op *op);
|
||||
*/
|
||||
extern int emulate_step(struct pt_regs *regs, unsigned int instr);
|
||||
|
||||
extern void emulate_vsx_load(struct instruction_op *op, union vsx_reg *reg,
|
||||
const void *mem);
|
||||
extern void emulate_vsx_store(struct instruction_op *op, const union vsx_reg *reg,
|
||||
void *mem);
|
||||
|
@ -32,6 +32,7 @@ obj64-$(CONFIG_KPROBES_SANITY_TEST) += test_emulate_step.o
|
||||
obj-y += checksum_$(BITS).o checksum_wrappers.o
|
||||
|
||||
obj-$(CONFIG_PPC_EMULATE_SSTEP) += sstep.o ldstfp.o
|
||||
obj64-$(CONFIG_PPC_EMULATE_SSTEP) += quad.o
|
||||
|
||||
obj-$(CONFIG_PPC_LIB_RHEAP) += rheap.o
|
||||
|
||||
|
@ -178,10 +178,10 @@ _GLOBAL(do_stfd)
|
||||
EX_TABLE(2b,3b)
|
||||
|
||||
#ifdef CONFIG_ALTIVEC
|
||||
/* Get the contents of vrN into v0; N is in r3. */
|
||||
/* Get the contents of vrN into v0; N is in r3. Doesn't touch r3 or r4. */
|
||||
_GLOBAL(get_vr)
|
||||
mflr r0
|
||||
rlwinm r3,r3,3,0xf8
|
||||
rlwinm r6,r3,3,0xf8
|
||||
bcl 20,31,1f
|
||||
blr /* v0 is already in v0 */
|
||||
nop
|
||||
@ -192,15 +192,15 @@ reg = 1
|
||||
reg = reg + 1
|
||||
.endr
|
||||
1: mflr r5
|
||||
add r5,r3,r5
|
||||
add r5,r6,r5
|
||||
mtctr r5
|
||||
mtlr r0
|
||||
bctr
|
||||
|
||||
/* Put the contents of v0 into vrN; N is in r3. */
|
||||
/* Put the contents of v0 into vrN; N is in r3. Doesn't touch r3 or r4. */
|
||||
_GLOBAL(put_vr)
|
||||
mflr r0
|
||||
rlwinm r3,r3,3,0xf8
|
||||
rlwinm r6,r3,3,0xf8
|
||||
bcl 20,31,1f
|
||||
blr /* v0 is already in v0 */
|
||||
nop
|
||||
@ -211,7 +211,7 @@ reg = 1
|
||||
reg = reg + 1
|
||||
.endr
|
||||
1: mflr r5
|
||||
add r5,r3,r5
|
||||
add r5,r6,r5
|
||||
mtctr r5
|
||||
mtlr r0
|
||||
bctr
|
||||
@ -313,7 +313,7 @@ reg = reg + 1
|
||||
bctr
|
||||
|
||||
/* Load VSX reg N from vector doubleword *p. N is in r3, p in r4. */
|
||||
_GLOBAL(do_lxvd2x)
|
||||
_GLOBAL(load_vsrn)
|
||||
PPC_STLU r1,-STKFRM(r1)
|
||||
mflr r0
|
||||
PPC_STL r0,STKFRM+PPC_LR_STKOFF(r1)
|
||||
@ -325,41 +325,38 @@ _GLOBAL(do_lxvd2x)
|
||||
isync
|
||||
beq cr7,1f
|
||||
STXVD2X(0,R1,R8)
|
||||
1: li r9,-EFAULT
|
||||
2: LXVD2X(0,R0,R4)
|
||||
li r9,0
|
||||
3: beq cr7,4f
|
||||
1: LXVD2X(0,R0,R4)
|
||||
#ifdef __LITTLE_ENDIAN__
|
||||
XXSWAPD(0,0)
|
||||
#endif
|
||||
beq cr7,4f
|
||||
bl put_vsr
|
||||
LXVD2X(0,R1,R8)
|
||||
4: PPC_LL r0,STKFRM+PPC_LR_STKOFF(r1)
|
||||
mtlr r0
|
||||
MTMSRD(r6)
|
||||
isync
|
||||
mr r3,r9
|
||||
addi r1,r1,STKFRM
|
||||
blr
|
||||
EX_TABLE(2b,3b)
|
||||
|
||||
/* Store VSX reg N to vector doubleword *p. N is in r3, p in r4. */
|
||||
_GLOBAL(do_stxvd2x)
|
||||
_GLOBAL(store_vsrn)
|
||||
PPC_STLU r1,-STKFRM(r1)
|
||||
mflr r0
|
||||
PPC_STL r0,STKFRM+PPC_LR_STKOFF(r1)
|
||||
mfmsr r6
|
||||
oris r7,r6,MSR_VSX@h
|
||||
cmpwi cr7,r3,0
|
||||
li r8,STKFRM-16
|
||||
MTMSRD(r7)
|
||||
isync
|
||||
beq cr7,1f
|
||||
STXVD2X(0,R1,R8)
|
||||
bl get_vsr
|
||||
1: li r9,-EFAULT
|
||||
2: STXVD2X(0,R0,R4)
|
||||
li r9,0
|
||||
3: beq cr7,4f
|
||||
#ifdef __LITTLE_ENDIAN__
|
||||
XXSWAPD(0,0)
|
||||
#endif
|
||||
STXVD2X(0,R0,R4)
|
||||
LXVD2X(0,R1,R8)
|
||||
4: PPC_LL r0,STKFRM+PPC_LR_STKOFF(r1)
|
||||
PPC_LL r0,STKFRM+PPC_LR_STKOFF(r1)
|
||||
mtlr r0
|
||||
MTMSRD(r6)
|
||||
isync
|
||||
@ -367,7 +364,36 @@ _GLOBAL(do_stxvd2x)
|
||||
addi r1,r1,STKFRM
|
||||
blr
|
||||
EX_TABLE(2b,3b)
|
||||
|
||||
#endif /* CONFIG_VSX */
|
||||
|
||||
/* Convert single-precision to double, without disturbing FPRs. */
|
||||
/* conv_sp_to_dp(float *sp, double *dp) */
|
||||
_GLOBAL(conv_sp_to_dp)
|
||||
mfmsr r6
|
||||
ori r7, r6, MSR_FP
|
||||
MTMSRD(r7)
|
||||
isync
|
||||
stfd fr0, -16(r1)
|
||||
lfs fr0, 0(r3)
|
||||
stfd fr0, 0(r4)
|
||||
lfd fr0, -16(r1)
|
||||
MTMSRD(r6)
|
||||
isync
|
||||
blr
|
||||
|
||||
/* Convert single-precision to double, without disturbing FPRs. */
|
||||
/* conv_sp_to_dp(double *dp, float *sp) */
|
||||
_GLOBAL(conv_dp_to_sp)
|
||||
mfmsr r6
|
||||
ori r7, r6, MSR_FP
|
||||
MTMSRD(r7)
|
||||
isync
|
||||
stfd fr0, -16(r1)
|
||||
lfd fr0, 0(r3)
|
||||
stfs fr0, 0(r4)
|
||||
lfd fr0, -16(r1)
|
||||
MTMSRD(r6)
|
||||
isync
|
||||
blr
|
||||
|
||||
#endif /* CONFIG_PPC_FPU */
|
||||
|
62
arch/powerpc/lib/quad.S
Normal file
62
arch/powerpc/lib/quad.S
Normal file
@ -0,0 +1,62 @@
|
||||
/*
|
||||
* Quadword loads and stores
|
||||
* for use in instruction emulation.
|
||||
*
|
||||
* Copyright 2017 Paul Mackerras, IBM Corp. <paulus@au1.ibm.com>
|
||||
*
|
||||
* This program is free software; you can redistribute it and/or
|
||||
* modify it under the terms of the GNU General Public License
|
||||
* as published by the Free Software Foundation; either version
|
||||
* 2 of the License, or (at your option) any later version.
|
||||
*/
|
||||
|
||||
#include <asm/processor.h>
|
||||
#include <asm/ppc_asm.h>
|
||||
#include <asm/ppc-opcode.h>
|
||||
#include <asm/reg.h>
|
||||
#include <asm/asm-offsets.h>
|
||||
#include <linux/errno.h>
|
||||
|
||||
/* do_lq(unsigned long ea, unsigned long *regs) */
|
||||
_GLOBAL(do_lq)
|
||||
1: lq r6, 0(r3)
|
||||
std r6, 0(r4)
|
||||
std r7, 8(r4)
|
||||
li r3, 0
|
||||
blr
|
||||
2: li r3, -EFAULT
|
||||
blr
|
||||
EX_TABLE(1b, 2b)
|
||||
|
||||
/* do_stq(unsigned long ea, unsigned long val0, unsigned long val1) */
|
||||
_GLOBAL(do_stq)
|
||||
1: stq r4, 0(r3)
|
||||
li r3, 0
|
||||
blr
|
||||
2: li r3, -EFAULT
|
||||
blr
|
||||
EX_TABLE(1b, 2b)
|
||||
|
||||
/* do_lqarx(unsigned long ea, unsigned long *regs) */
|
||||
_GLOBAL(do_lqarx)
|
||||
1: PPC_LQARX(6, 0, 3, 0)
|
||||
std r6, 0(r4)
|
||||
std r7, 8(r4)
|
||||
li r3, 0
|
||||
blr
|
||||
2: li r3, -EFAULT
|
||||
blr
|
||||
EX_TABLE(1b, 2b)
|
||||
|
||||
/* do_stqcx(unsigned long ea, unsigned long val0, unsigned long val1,
|
||||
unsigned int *crp) */
|
||||
|
||||
_GLOBAL(do_stqcx)
|
||||
1: PPC_STQCX(4, 0, 3)
|
||||
mfcr r5
|
||||
stw r5, 0(r6)
|
||||
li r3, 0
|
||||
blr
|
||||
2: li r3, -EFAULT
|
||||
blr
|
||||
EX_TABLE(1b, 2b)
|
@ -42,8 +42,29 @@ extern int do_stfs(int rn, unsigned long ea);
|
||||
extern int do_stfd(int rn, unsigned long ea);
|
||||
extern int do_lvx(int rn, unsigned long ea);
|
||||
extern int do_stvx(int rn, unsigned long ea);
|
||||
extern int do_lxvd2x(int rn, unsigned long ea);
|
||||
extern int do_stxvd2x(int rn, unsigned long ea);
|
||||
extern void load_vsrn(int vsr, const void *p);
|
||||
extern void store_vsrn(int vsr, void *p);
|
||||
extern void conv_sp_to_dp(const float *sp, double *dp);
|
||||
extern void conv_dp_to_sp(const double *dp, float *sp);
|
||||
#endif
|
||||
|
||||
#ifdef __powerpc64__
|
||||
/*
|
||||
* Functions in quad.S
|
||||
*/
|
||||
extern int do_lq(unsigned long ea, unsigned long *regs);
|
||||
extern int do_stq(unsigned long ea, unsigned long val0, unsigned long val1);
|
||||
extern int do_lqarx(unsigned long ea, unsigned long *regs);
|
||||
extern int do_stqcx(unsigned long ea, unsigned long val0, unsigned long val1,
|
||||
unsigned int *crp);
|
||||
#endif
|
||||
|
||||
#ifdef __LITTLE_ENDIAN__
|
||||
#define IS_LE 1
|
||||
#define IS_BE 0
|
||||
#else
|
||||
#define IS_LE 0
|
||||
#define IS_BE 1
|
||||
#endif
|
||||
|
||||
/*
|
||||
@ -125,6 +146,23 @@ static nokprobe_inline unsigned long dsform_ea(unsigned int instr,
|
||||
|
||||
return truncate_if_32bit(regs->msr, ea);
|
||||
}
|
||||
|
||||
/*
|
||||
* Calculate effective address for a DQ-form instruction
|
||||
*/
|
||||
static nokprobe_inline unsigned long dqform_ea(unsigned int instr,
|
||||
const struct pt_regs *regs)
|
||||
{
|
||||
int ra;
|
||||
unsigned long ea;
|
||||
|
||||
ra = (instr >> 16) & 0x1f;
|
||||
ea = (signed short) (instr & ~0xf); /* sign-extend */
|
||||
if (ra)
|
||||
ea += regs->gpr[ra];
|
||||
|
||||
return truncate_if_32bit(regs->msr, ea);
|
||||
}
|
||||
#endif /* __powerpc64 */
|
||||
|
||||
/*
|
||||
@ -454,43 +492,195 @@ static nokprobe_inline int do_vec_store(int rn, int (*func)(int, unsigned long),
|
||||
}
|
||||
#endif /* CONFIG_ALTIVEC */
|
||||
|
||||
#ifdef __powerpc64__
|
||||
static nokprobe_inline int emulate_lq(struct pt_regs *regs, unsigned long ea,
|
||||
int reg)
|
||||
{
|
||||
int err;
|
||||
|
||||
if (!address_ok(regs, ea, 16))
|
||||
return -EFAULT;
|
||||
/* if aligned, should be atomic */
|
||||
if ((ea & 0xf) == 0)
|
||||
return do_lq(ea, ®s->gpr[reg]);
|
||||
|
||||
err = read_mem(®s->gpr[reg + IS_LE], ea, 8, regs);
|
||||
if (!err)
|
||||
err = read_mem(®s->gpr[reg + IS_BE], ea + 8, 8, regs);
|
||||
return err;
|
||||
}
|
||||
|
||||
static nokprobe_inline int emulate_stq(struct pt_regs *regs, unsigned long ea,
|
||||
int reg)
|
||||
{
|
||||
int err;
|
||||
|
||||
if (!address_ok(regs, ea, 16))
|
||||
return -EFAULT;
|
||||
/* if aligned, should be atomic */
|
||||
if ((ea & 0xf) == 0)
|
||||
return do_stq(ea, regs->gpr[reg], regs->gpr[reg + 1]);
|
||||
|
||||
err = write_mem(regs->gpr[reg + IS_LE], ea, 8, regs);
|
||||
if (!err)
|
||||
err = write_mem(regs->gpr[reg + IS_BE], ea + 8, 8, regs);
|
||||
return err;
|
||||
}
|
||||
#endif /* __powerpc64 */
|
||||
|
||||
#ifdef CONFIG_VSX
|
||||
static nokprobe_inline int do_vsx_load(int rn, int (*func)(int, unsigned long),
|
||||
unsigned long ea, struct pt_regs *regs)
|
||||
void emulate_vsx_load(struct instruction_op *op, union vsx_reg *reg,
|
||||
const void *mem)
|
||||
{
|
||||
int err;
|
||||
unsigned long val[2];
|
||||
int size, read_size;
|
||||
int i, j;
|
||||
const unsigned int *wp;
|
||||
const unsigned short *hp;
|
||||
const unsigned char *bp;
|
||||
|
||||
if (!address_ok(regs, ea, 16))
|
||||
return -EFAULT;
|
||||
if ((ea & 3) == 0)
|
||||
return (*func)(rn, ea);
|
||||
err = read_mem_unaligned(&val[0], ea, 8, regs);
|
||||
if (!err)
|
||||
err = read_mem_unaligned(&val[1], ea + 8, 8, regs);
|
||||
if (!err)
|
||||
err = (*func)(rn, (unsigned long) &val[0]);
|
||||
return err;
|
||||
size = GETSIZE(op->type);
|
||||
reg->d[0] = reg->d[1] = 0;
|
||||
|
||||
switch (op->element_size) {
|
||||
case 16:
|
||||
/* whole vector; lxv[x] or lxvl[l] */
|
||||
if (size == 0)
|
||||
break;
|
||||
memcpy(reg, mem, size);
|
||||
if (IS_LE && (op->vsx_flags & VSX_LDLEFT)) {
|
||||
/* reverse 16 bytes */
|
||||
unsigned long tmp;
|
||||
tmp = byterev_8(reg->d[0]);
|
||||
reg->d[0] = byterev_8(reg->d[1]);
|
||||
reg->d[1] = tmp;
|
||||
}
|
||||
break;
|
||||
case 8:
|
||||
/* scalar loads, lxvd2x, lxvdsx */
|
||||
read_size = (size >= 8) ? 8 : size;
|
||||
i = IS_LE ? 8 : 8 - read_size;
|
||||
memcpy(®->b[i], mem, read_size);
|
||||
if (size < 8) {
|
||||
if (op->type & SIGNEXT) {
|
||||
/* size == 4 is the only case here */
|
||||
reg->d[IS_LE] = (signed int) reg->d[IS_LE];
|
||||
} else if (op->vsx_flags & VSX_FPCONV) {
|
||||
preempt_disable();
|
||||
conv_sp_to_dp(®->fp[1 + IS_LE],
|
||||
®->dp[IS_LE]);
|
||||
preempt_enable();
|
||||
}
|
||||
} else {
|
||||
if (size == 16)
|
||||
reg->d[IS_BE] = *(unsigned long *)(mem + 8);
|
||||
else if (op->vsx_flags & VSX_SPLAT)
|
||||
reg->d[IS_BE] = reg->d[IS_LE];
|
||||
}
|
||||
break;
|
||||
case 4:
|
||||
/* lxvw4x, lxvwsx */
|
||||
wp = mem;
|
||||
for (j = 0; j < size / 4; ++j) {
|
||||
i = IS_LE ? 3 - j : j;
|
||||
reg->w[i] = *wp++;
|
||||
}
|
||||
if (op->vsx_flags & VSX_SPLAT) {
|
||||
u32 val = reg->w[IS_LE ? 3 : 0];
|
||||
for (; j < 4; ++j) {
|
||||
i = IS_LE ? 3 - j : j;
|
||||
reg->w[i] = val;
|
||||
}
|
||||
}
|
||||
break;
|
||||
case 2:
|
||||
/* lxvh8x */
|
||||
hp = mem;
|
||||
for (j = 0; j < size / 2; ++j) {
|
||||
i = IS_LE ? 7 - j : j;
|
||||
reg->h[i] = *hp++;
|
||||
}
|
||||
break;
|
||||
case 1:
|
||||
/* lxvb16x */
|
||||
bp = mem;
|
||||
for (j = 0; j < size; ++j) {
|
||||
i = IS_LE ? 15 - j : j;
|
||||
reg->b[i] = *bp++;
|
||||
}
|
||||
break;
|
||||
}
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(emulate_vsx_load);
|
||||
NOKPROBE_SYMBOL(emulate_vsx_load);
|
||||
|
||||
static nokprobe_inline int do_vsx_store(int rn, int (*func)(int, unsigned long),
|
||||
unsigned long ea, struct pt_regs *regs)
|
||||
void emulate_vsx_store(struct instruction_op *op, const union vsx_reg *reg,
|
||||
void *mem)
|
||||
{
|
||||
int err;
|
||||
unsigned long val[2];
|
||||
int size, write_size;
|
||||
int i, j;
|
||||
union vsx_reg buf;
|
||||
unsigned int *wp;
|
||||
unsigned short *hp;
|
||||
unsigned char *bp;
|
||||
|
||||
if (!address_ok(regs, ea, 16))
|
||||
return -EFAULT;
|
||||
if ((ea & 3) == 0)
|
||||
return (*func)(rn, ea);
|
||||
err = (*func)(rn, (unsigned long) &val[0]);
|
||||
if (err)
|
||||
return err;
|
||||
err = write_mem_unaligned(val[0], ea, 8, regs);
|
||||
if (!err)
|
||||
err = write_mem_unaligned(val[1], ea + 8, 8, regs);
|
||||
return err;
|
||||
size = GETSIZE(op->type);
|
||||
|
||||
switch (op->element_size) {
|
||||
case 16:
|
||||
/* stxv, stxvx, stxvl, stxvll */
|
||||
if (size == 0)
|
||||
break;
|
||||
if (IS_LE && (op->vsx_flags & VSX_LDLEFT)) {
|
||||
/* reverse 16 bytes */
|
||||
buf.d[0] = byterev_8(reg->d[1]);
|
||||
buf.d[1] = byterev_8(reg->d[0]);
|
||||
reg = &buf;
|
||||
}
|
||||
memcpy(mem, reg, size);
|
||||
break;
|
||||
case 8:
|
||||
/* scalar stores, stxvd2x */
|
||||
write_size = (size >= 8) ? 8 : size;
|
||||
i = IS_LE ? 8 : 8 - write_size;
|
||||
if (size < 8 && op->vsx_flags & VSX_FPCONV) {
|
||||
buf.d[0] = buf.d[1] = 0;
|
||||
preempt_disable();
|
||||
conv_dp_to_sp(®->dp[IS_LE], &buf.fp[1 + IS_LE]);
|
||||
preempt_enable();
|
||||
reg = &buf;
|
||||
}
|
||||
memcpy(mem, ®->b[i], write_size);
|
||||
if (size == 16)
|
||||
memcpy(mem + 8, ®->d[IS_BE], 8);
|
||||
break;
|
||||
case 4:
|
||||
/* stxvw4x */
|
||||
wp = mem;
|
||||
for (j = 0; j < size / 4; ++j) {
|
||||
i = IS_LE ? 3 - j : j;
|
||||
*wp++ = reg->w[i];
|
||||
}
|
||||
break;
|
||||
case 2:
|
||||
/* stxvh8x */
|
||||
hp = mem;
|
||||
for (j = 0; j < size / 2; ++j) {
|
||||
i = IS_LE ? 7 - j : j;
|
||||
*hp++ = reg->h[i];
|
||||
}
|
||||
break;
|
||||
case 1:
|
||||
/* stvxb16x */
|
||||
bp = mem;
|
||||
for (j = 0; j < size; ++j) {
|
||||
i = IS_LE ? 15 - j : j;
|
||||
*bp++ = reg->b[i];
|
||||
}
|
||||
break;
|
||||
}
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(emulate_vsx_store);
|
||||
NOKPROBE_SYMBOL(emulate_vsx_store);
|
||||
#endif /* CONFIG_VSX */
|
||||
|
||||
#define __put_user_asmx(x, addr, err, op, cr) \
|
||||
@ -1455,14 +1645,15 @@ int analyse_instr(struct instruction_op *op, const struct pt_regs *regs,
|
||||
break;
|
||||
}
|
||||
|
||||
/*
|
||||
* Loads and stores.
|
||||
*/
|
||||
/*
|
||||
* Loads and stores.
|
||||
*/
|
||||
op->type = UNKNOWN;
|
||||
op->update_reg = ra;
|
||||
op->reg = rd;
|
||||
op->val = regs->gpr[rd];
|
||||
u = (instr >> 20) & UPDATE;
|
||||
op->vsx_flags = 0;
|
||||
|
||||
switch (opcode) {
|
||||
case 31:
|
||||
@ -1486,9 +1677,30 @@ int analyse_instr(struct instruction_op *op, const struct pt_regs *regs,
|
||||
op->type = MKOP(STCX, 0, 8);
|
||||
break;
|
||||
|
||||
case 21: /* ldx */
|
||||
case 53: /* ldux */
|
||||
op->type = MKOP(LOAD, u, 8);
|
||||
case 52: /* lbarx */
|
||||
op->type = MKOP(LARX, 0, 1);
|
||||
break;
|
||||
|
||||
case 694: /* stbcx. */
|
||||
op->type = MKOP(STCX, 0, 1);
|
||||
break;
|
||||
|
||||
case 116: /* lharx */
|
||||
op->type = MKOP(LARX, 0, 2);
|
||||
break;
|
||||
|
||||
case 726: /* sthcx. */
|
||||
op->type = MKOP(STCX, 0, 2);
|
||||
break;
|
||||
|
||||
case 276: /* lqarx */
|
||||
if (!((rd & 1) || rd == ra || rd == rb))
|
||||
op->type = MKOP(LARX, 0, 16);
|
||||
break;
|
||||
|
||||
case 182: /* stqcx. */
|
||||
if (!(rd & 1))
|
||||
op->type = MKOP(STCX, 0, 16);
|
||||
break;
|
||||
#endif
|
||||
|
||||
@ -1506,6 +1718,7 @@ int analyse_instr(struct instruction_op *op, const struct pt_regs *regs,
|
||||
case 103: /* lvx */
|
||||
case 359: /* lvxl */
|
||||
op->type = MKOP(LOAD_VMX, 0, 16);
|
||||
op->element_size = 16;
|
||||
break;
|
||||
|
||||
case 231: /* stvx */
|
||||
@ -1515,6 +1728,11 @@ int analyse_instr(struct instruction_op *op, const struct pt_regs *regs,
|
||||
#endif /* CONFIG_ALTIVEC */
|
||||
|
||||
#ifdef __powerpc64__
|
||||
case 21: /* ldx */
|
||||
case 53: /* ldux */
|
||||
op->type = MKOP(LOAD, u, 8);
|
||||
break;
|
||||
|
||||
case 149: /* stdx */
|
||||
case 181: /* stdux */
|
||||
op->type = MKOP(STORE, u, 8);
|
||||
@ -1635,16 +1853,184 @@ int analyse_instr(struct instruction_op *op, const struct pt_regs *regs,
|
||||
break;
|
||||
|
||||
#ifdef CONFIG_VSX
|
||||
case 844: /* lxvd2x */
|
||||
case 876: /* lxvd2ux */
|
||||
case 12: /* lxsiwzx */
|
||||
op->reg = rd | ((instr & 1) << 5);
|
||||
op->type = MKOP(LOAD_VSX, u, 16);
|
||||
op->type = MKOP(LOAD_VSX, 0, 4);
|
||||
op->element_size = 8;
|
||||
break;
|
||||
|
||||
case 76: /* lxsiwax */
|
||||
op->reg = rd | ((instr & 1) << 5);
|
||||
op->type = MKOP(LOAD_VSX, SIGNEXT, 4);
|
||||
op->element_size = 8;
|
||||
break;
|
||||
|
||||
case 140: /* stxsiwx */
|
||||
op->reg = rd | ((instr & 1) << 5);
|
||||
op->type = MKOP(STORE_VSX, 0, 4);
|
||||
op->element_size = 8;
|
||||
break;
|
||||
|
||||
case 268: /* lxvx */
|
||||
op->reg = rd | ((instr & 1) << 5);
|
||||
op->type = MKOP(LOAD_VSX, 0, 16);
|
||||
op->element_size = 16;
|
||||
op->vsx_flags = VSX_CHECK_VEC;
|
||||
break;
|
||||
|
||||
case 269: /* lxvl */
|
||||
case 301: { /* lxvll */
|
||||
int nb;
|
||||
op->reg = rd | ((instr & 1) << 5);
|
||||
op->ea = ra ? regs->gpr[ra] : 0;
|
||||
nb = regs->gpr[rb] & 0xff;
|
||||
if (nb > 16)
|
||||
nb = 16;
|
||||
op->type = MKOP(LOAD_VSX, 0, nb);
|
||||
op->element_size = 16;
|
||||
op->vsx_flags = ((instr & 0x20) ? VSX_LDLEFT : 0) |
|
||||
VSX_CHECK_VEC;
|
||||
break;
|
||||
}
|
||||
case 332: /* lxvdsx */
|
||||
op->reg = rd | ((instr & 1) << 5);
|
||||
op->type = MKOP(LOAD_VSX, 0, 8);
|
||||
op->element_size = 8;
|
||||
op->vsx_flags = VSX_SPLAT;
|
||||
break;
|
||||
|
||||
case 364: /* lxvwsx */
|
||||
op->reg = rd | ((instr & 1) << 5);
|
||||
op->type = MKOP(LOAD_VSX, 0, 4);
|
||||
op->element_size = 4;
|
||||
op->vsx_flags = VSX_SPLAT | VSX_CHECK_VEC;
|
||||
break;
|
||||
|
||||
case 396: /* stxvx */
|
||||
op->reg = rd | ((instr & 1) << 5);
|
||||
op->type = MKOP(STORE_VSX, 0, 16);
|
||||
op->element_size = 16;
|
||||
op->vsx_flags = VSX_CHECK_VEC;
|
||||
break;
|
||||
|
||||
case 397: /* stxvl */
|
||||
case 429: { /* stxvll */
|
||||
int nb;
|
||||
op->reg = rd | ((instr & 1) << 5);
|
||||
op->ea = ra ? regs->gpr[ra] : 0;
|
||||
nb = regs->gpr[rb] & 0xff;
|
||||
if (nb > 16)
|
||||
nb = 16;
|
||||
op->type = MKOP(STORE_VSX, 0, nb);
|
||||
op->element_size = 16;
|
||||
op->vsx_flags = ((instr & 0x20) ? VSX_LDLEFT : 0) |
|
||||
VSX_CHECK_VEC;
|
||||
break;
|
||||
}
|
||||
case 524: /* lxsspx */
|
||||
op->reg = rd | ((instr & 1) << 5);
|
||||
op->type = MKOP(LOAD_VSX, 0, 4);
|
||||
op->element_size = 8;
|
||||
op->vsx_flags = VSX_FPCONV;
|
||||
break;
|
||||
|
||||
case 588: /* lxsdx */
|
||||
op->reg = rd | ((instr & 1) << 5);
|
||||
op->type = MKOP(LOAD_VSX, 0, 8);
|
||||
op->element_size = 8;
|
||||
break;
|
||||
|
||||
case 652: /* stxsspx */
|
||||
op->reg = rd | ((instr & 1) << 5);
|
||||
op->type = MKOP(STORE_VSX, 0, 4);
|
||||
op->element_size = 8;
|
||||
op->vsx_flags = VSX_FPCONV;
|
||||
break;
|
||||
|
||||
case 716: /* stxsdx */
|
||||
op->reg = rd | ((instr & 1) << 5);
|
||||
op->type = MKOP(STORE_VSX, 0, 8);
|
||||
op->element_size = 8;
|
||||
break;
|
||||
|
||||
case 780: /* lxvw4x */
|
||||
op->reg = rd | ((instr & 1) << 5);
|
||||
op->type = MKOP(LOAD_VSX, 0, 16);
|
||||
op->element_size = 4;
|
||||
break;
|
||||
|
||||
case 781: /* lxsibzx */
|
||||
op->reg = rd | ((instr & 1) << 5);
|
||||
op->type = MKOP(LOAD_VSX, 0, 1);
|
||||
op->element_size = 8;
|
||||
op->vsx_flags = VSX_CHECK_VEC;
|
||||
break;
|
||||
|
||||
case 812: /* lxvh8x */
|
||||
op->reg = rd | ((instr & 1) << 5);
|
||||
op->type = MKOP(LOAD_VSX, 0, 16);
|
||||
op->element_size = 2;
|
||||
op->vsx_flags = VSX_CHECK_VEC;
|
||||
break;
|
||||
|
||||
case 813: /* lxsihzx */
|
||||
op->reg = rd | ((instr & 1) << 5);
|
||||
op->type = MKOP(LOAD_VSX, 0, 2);
|
||||
op->element_size = 8;
|
||||
op->vsx_flags = VSX_CHECK_VEC;
|
||||
break;
|
||||
|
||||
case 844: /* lxvd2x */
|
||||
op->reg = rd | ((instr & 1) << 5);
|
||||
op->type = MKOP(LOAD_VSX, 0, 16);
|
||||
op->element_size = 8;
|
||||
break;
|
||||
|
||||
case 876: /* lxvb16x */
|
||||
op->reg = rd | ((instr & 1) << 5);
|
||||
op->type = MKOP(LOAD_VSX, 0, 16);
|
||||
op->element_size = 1;
|
||||
op->vsx_flags = VSX_CHECK_VEC;
|
||||
break;
|
||||
|
||||
case 908: /* stxvw4x */
|
||||
op->reg = rd | ((instr & 1) << 5);
|
||||
op->type = MKOP(STORE_VSX, 0, 16);
|
||||
op->element_size = 4;
|
||||
break;
|
||||
|
||||
case 909: /* stxsibx */
|
||||
op->reg = rd | ((instr & 1) << 5);
|
||||
op->type = MKOP(STORE_VSX, 0, 1);
|
||||
op->element_size = 8;
|
||||
op->vsx_flags = VSX_CHECK_VEC;
|
||||
break;
|
||||
|
||||
case 940: /* stxvh8x */
|
||||
op->reg = rd | ((instr & 1) << 5);
|
||||
op->type = MKOP(STORE_VSX, 0, 16);
|
||||
op->element_size = 2;
|
||||
op->vsx_flags = VSX_CHECK_VEC;
|
||||
break;
|
||||
|
||||
case 941: /* stxsihx */
|
||||
op->reg = rd | ((instr & 1) << 5);
|
||||
op->type = MKOP(STORE_VSX, 0, 2);
|
||||
op->element_size = 8;
|
||||
op->vsx_flags = VSX_CHECK_VEC;
|
||||
break;
|
||||
|
||||
case 972: /* stxvd2x */
|
||||
case 1004: /* stxvd2ux */
|
||||
op->reg = rd | ((instr & 1) << 5);
|
||||
op->type = MKOP(STORE_VSX, u, 16);
|
||||
op->type = MKOP(STORE_VSX, 0, 16);
|
||||
op->element_size = 8;
|
||||
break;
|
||||
|
||||
case 1004: /* stxvb16x */
|
||||
op->reg = rd | ((instr & 1) << 5);
|
||||
op->type = MKOP(STORE_VSX, 0, 16);
|
||||
op->element_size = 1;
|
||||
op->vsx_flags = VSX_CHECK_VEC;
|
||||
break;
|
||||
|
||||
#endif /* CONFIG_VSX */
|
||||
@ -1731,6 +2117,34 @@ int analyse_instr(struct instruction_op *op, const struct pt_regs *regs,
|
||||
break;
|
||||
#endif
|
||||
|
||||
#ifdef __powerpc64__
|
||||
case 56: /* lq */
|
||||
if (!((rd & 1) || (rd == ra)))
|
||||
op->type = MKOP(LOAD, 0, 16);
|
||||
op->ea = dqform_ea(instr, regs);
|
||||
break;
|
||||
#endif
|
||||
|
||||
#ifdef CONFIG_VSX
|
||||
case 57: /* lxsd, lxssp */
|
||||
op->ea = dsform_ea(instr, regs);
|
||||
switch (instr & 3) {
|
||||
case 2: /* lxsd */
|
||||
op->reg = rd + 32;
|
||||
op->type = MKOP(LOAD_VSX, 0, 8);
|
||||
op->element_size = 8;
|
||||
op->vsx_flags = VSX_CHECK_VEC;
|
||||
break;
|
||||
case 3: /* lxssp */
|
||||
op->reg = rd + 32;
|
||||
op->type = MKOP(LOAD_VSX, 0, 4);
|
||||
op->element_size = 8;
|
||||
op->vsx_flags = VSX_FPCONV | VSX_CHECK_VEC;
|
||||
break;
|
||||
}
|
||||
break;
|
||||
#endif /* CONFIG_VSX */
|
||||
|
||||
#ifdef __powerpc64__
|
||||
case 58: /* ld[u], lwa */
|
||||
op->ea = dsform_ea(instr, regs);
|
||||
@ -1746,7 +2160,51 @@ int analyse_instr(struct instruction_op *op, const struct pt_regs *regs,
|
||||
break;
|
||||
}
|
||||
break;
|
||||
#endif
|
||||
|
||||
#ifdef CONFIG_VSX
|
||||
case 61: /* lxv, stxsd, stxssp, stxv */
|
||||
switch (instr & 7) {
|
||||
case 1: /* lxv */
|
||||
op->ea = dqform_ea(instr, regs);
|
||||
if (instr & 8)
|
||||
op->reg = rd + 32;
|
||||
op->type = MKOP(LOAD_VSX, 0, 16);
|
||||
op->element_size = 16;
|
||||
op->vsx_flags = VSX_CHECK_VEC;
|
||||
break;
|
||||
|
||||
case 2: /* stxsd with LSB of DS field = 0 */
|
||||
case 6: /* stxsd with LSB of DS field = 1 */
|
||||
op->ea = dsform_ea(instr, regs);
|
||||
op->reg = rd + 32;
|
||||
op->type = MKOP(STORE_VSX, 0, 8);
|
||||
op->element_size = 8;
|
||||
op->vsx_flags = VSX_CHECK_VEC;
|
||||
break;
|
||||
|
||||
case 3: /* stxssp with LSB of DS field = 0 */
|
||||
case 7: /* stxssp with LSB of DS field = 1 */
|
||||
op->ea = dsform_ea(instr, regs);
|
||||
op->reg = rd + 32;
|
||||
op->type = MKOP(STORE_VSX, 0, 4);
|
||||
op->element_size = 8;
|
||||
op->vsx_flags = VSX_FPCONV | VSX_CHECK_VEC;
|
||||
break;
|
||||
|
||||
case 5: /* stxv */
|
||||
op->ea = dqform_ea(instr, regs);
|
||||
if (instr & 8)
|
||||
op->reg = rd + 32;
|
||||
op->type = MKOP(STORE_VSX, 0, 16);
|
||||
op->element_size = 16;
|
||||
op->vsx_flags = VSX_CHECK_VEC;
|
||||
break;
|
||||
}
|
||||
break;
|
||||
#endif /* CONFIG_VSX */
|
||||
|
||||
#ifdef __powerpc64__
|
||||
case 62: /* std[u] */
|
||||
op->ea = dsform_ea(instr, regs);
|
||||
switch (instr & 3) {
|
||||
@ -1756,6 +2214,10 @@ int analyse_instr(struct instruction_op *op, const struct pt_regs *regs,
|
||||
case 1: /* stdu */
|
||||
op->type = MKOP(STORE, UPDATE, 8);
|
||||
break;
|
||||
case 2: /* stq */
|
||||
if (!(rd & 1))
|
||||
op->type = MKOP(STORE, 0, 16);
|
||||
break;
|
||||
}
|
||||
break;
|
||||
#endif /* __powerpc64__ */
|
||||
@ -1994,6 +2456,14 @@ int emulate_step(struct pt_regs *regs, unsigned int instr)
|
||||
return 0;
|
||||
err = 0;
|
||||
switch (size) {
|
||||
#ifdef __powerpc64__
|
||||
case 1:
|
||||
__get_user_asmx(val, op.ea, err, "lbarx");
|
||||
break;
|
||||
case 2:
|
||||
__get_user_asmx(val, op.ea, err, "lharx");
|
||||
break;
|
||||
#endif
|
||||
case 4:
|
||||
__get_user_asmx(val, op.ea, err, "lwarx");
|
||||
break;
|
||||
@ -2001,6 +2471,9 @@ int emulate_step(struct pt_regs *regs, unsigned int instr)
|
||||
case 8:
|
||||
__get_user_asmx(val, op.ea, err, "ldarx");
|
||||
break;
|
||||
case 16:
|
||||
err = do_lqarx(op.ea, ®s->gpr[op.reg]);
|
||||
goto ldst_done;
|
||||
#endif
|
||||
default:
|
||||
return 0;
|
||||
@ -2016,6 +2489,14 @@ int emulate_step(struct pt_regs *regs, unsigned int instr)
|
||||
return 0;
|
||||
err = 0;
|
||||
switch (size) {
|
||||
#ifdef __powerpc64__
|
||||
case 1:
|
||||
__put_user_asmx(op.val, op.ea, err, "stbcx.", cr);
|
||||
break;
|
||||
case 2:
|
||||
__put_user_asmx(op.val, op.ea, err, "stbcx.", cr);
|
||||
break;
|
||||
#endif
|
||||
case 4:
|
||||
__put_user_asmx(op.val, op.ea, err, "stwcx.", cr);
|
||||
break;
|
||||
@ -2023,6 +2504,10 @@ int emulate_step(struct pt_regs *regs, unsigned int instr)
|
||||
case 8:
|
||||
__put_user_asmx(op.val, op.ea, err, "stdcx.", cr);
|
||||
break;
|
||||
case 16:
|
||||
err = do_stqcx(op.ea, regs->gpr[op.reg],
|
||||
regs->gpr[op.reg + 1], &cr);
|
||||
break;
|
||||
#endif
|
||||
default:
|
||||
return 0;
|
||||
@ -2034,6 +2519,12 @@ int emulate_step(struct pt_regs *regs, unsigned int instr)
|
||||
goto ldst_done;
|
||||
|
||||
case LOAD:
|
||||
#ifdef __powerpc64__
|
||||
if (size == 16) {
|
||||
err = emulate_lq(regs, op.ea, op.reg);
|
||||
goto ldst_done;
|
||||
}
|
||||
#endif
|
||||
err = read_mem(®s->gpr[op.reg], op.ea, size, regs);
|
||||
if (!err) {
|
||||
if (op.type & SIGNEXT)
|
||||
@ -2057,15 +2548,31 @@ int emulate_step(struct pt_regs *regs, unsigned int instr)
|
||||
case LOAD_VMX:
|
||||
if (!(regs->msr & MSR_VEC))
|
||||
return 0;
|
||||
err = do_vec_load(op.reg, do_lvx, op.ea & ~0xfUL, regs);
|
||||
err = do_vec_load(op.reg, do_lvx, op.ea, regs);
|
||||
goto ldst_done;
|
||||
#endif
|
||||
#ifdef CONFIG_VSX
|
||||
case LOAD_VSX:
|
||||
if (!(regs->msr & MSR_VSX))
|
||||
case LOAD_VSX: {
|
||||
char mem[16];
|
||||
union vsx_reg buf;
|
||||
unsigned long msrbit = MSR_VSX;
|
||||
|
||||
/*
|
||||
* Some VSX instructions check the MSR_VEC bit rather than MSR_VSX
|
||||
* when the target of the instruction is a vector register.
|
||||
*/
|
||||
if (op.reg >= 32 && (op.vsx_flags & VSX_CHECK_VEC))
|
||||
msrbit = MSR_VEC;
|
||||
if (!(regs->msr & msrbit))
|
||||
return 0;
|
||||
err = do_vsx_load(op.reg, do_lxvd2x, op.ea, regs);
|
||||
if (!address_ok(regs, op.ea, size) ||
|
||||
__copy_from_user(mem, (void __user *)op.ea, size))
|
||||
return 0;
|
||||
|
||||
emulate_vsx_load(&op, &buf, mem);
|
||||
load_vsrn(op.reg, &buf);
|
||||
goto ldst_done;
|
||||
}
|
||||
#endif
|
||||
case LOAD_MULTI:
|
||||
if (regs->msr & MSR_LE)
|
||||
@ -2086,6 +2593,12 @@ int emulate_step(struct pt_regs *regs, unsigned int instr)
|
||||
goto instr_done;
|
||||
|
||||
case STORE:
|
||||
#ifdef __powerpc64__
|
||||
if (size == 16) {
|
||||
err = emulate_stq(regs, op.ea, op.reg);
|
||||
goto ldst_done;
|
||||
}
|
||||
#endif
|
||||
if ((op.type & UPDATE) && size == sizeof(long) &&
|
||||
op.reg == 1 && op.update_reg == 1 &&
|
||||
!(regs->msr & MSR_PR) &&
|
||||
@ -2110,15 +2623,32 @@ int emulate_step(struct pt_regs *regs, unsigned int instr)
|
||||
case STORE_VMX:
|
||||
if (!(regs->msr & MSR_VEC))
|
||||
return 0;
|
||||
err = do_vec_store(op.reg, do_stvx, op.ea & ~0xfUL, regs);
|
||||
err = do_vec_store(op.reg, do_stvx, op.ea, regs);
|
||||
goto ldst_done;
|
||||
#endif
|
||||
#ifdef CONFIG_VSX
|
||||
case STORE_VSX:
|
||||
if (!(regs->msr & MSR_VSX))
|
||||
case STORE_VSX: {
|
||||
char mem[16];
|
||||
union vsx_reg buf;
|
||||
unsigned long msrbit = MSR_VSX;
|
||||
|
||||
/*
|
||||
* Some VSX instructions check the MSR_VEC bit rather than MSR_VSX
|
||||
* when the target of the instruction is a vector register.
|
||||
*/
|
||||
if (op.reg >= 32 && (op.vsx_flags & VSX_CHECK_VEC))
|
||||
msrbit = MSR_VEC;
|
||||
if (!(regs->msr & msrbit))
|
||||
return 0;
|
||||
if (!address_ok(regs, op.ea, size))
|
||||
return 0;
|
||||
|
||||
store_vsrn(op.reg, &buf);
|
||||
emulate_vsx_store(&op, &buf, mem);
|
||||
if (__copy_to_user((void __user *)op.ea, mem, size))
|
||||
return 0;
|
||||
err = do_vsx_store(op.reg, do_stxvd2x, op.ea, regs);
|
||||
goto ldst_done;
|
||||
}
|
||||
#endif
|
||||
case STORE_MULTI:
|
||||
if (regs->msr & MSR_LE)
|
||||
|
Loading…
Reference in New Issue
Block a user