mirror of
https://github.com/torvalds/linux.git
synced 2024-12-11 13:41:55 +00:00
ec6347bb43
In reaction to a proposal to introduce a memcpy_mcsafe_fast() implementation Linus points out that memcpy_mcsafe() is poorly named relative to communicating the scope of the interface. Specifically what addresses are valid to pass as source, destination, and what faults / exceptions are handled. Of particular concern is that even though x86 might be able to handle the semantics of copy_mc_to_user() with its common copy_user_generic() implementation other archs likely need / want an explicit path for this case: On Fri, May 1, 2020 at 11:28 AM Linus Torvalds <torvalds@linux-foundation.org> wrote: > > On Thu, Apr 30, 2020 at 6:21 PM Dan Williams <dan.j.williams@intel.com> wrote: > > > > However now I see that copy_user_generic() works for the wrong reason. > > It works because the exception on the source address due to poison > > looks no different than a write fault on the user address to the > > caller, it's still just a short copy. So it makes copy_to_user() work > > for the wrong reason relative to the name. > > Right. > > And it won't work that way on other architectures. On x86, we have a > generic function that can take faults on either side, and we use it > for both cases (and for the "in_user" case too), but that's an > artifact of the architecture oddity. > > In fact, it's probably wrong even on x86 - because it can hide bugs - > but writing those things is painful enough that everybody prefers > having just one function. Replace a single top-level memcpy_mcsafe() with either copy_mc_to_user(), or copy_mc_to_kernel(). Introduce an x86 copy_mc_fragile() name as the rename for the low-level x86 implementation formerly named memcpy_mcsafe(). It is used as the slow / careful backend that is supplanted by a fast copy_mc_generic() in a follow-on patch. One side-effect of this reorganization is that separating copy_mc_64.S to its own file means that perf no longer needs to track dependencies for its memcpy_64.S benchmarks. [ bp: Massage a bit. ] Signed-off-by: Dan Williams <dan.j.williams@intel.com> Signed-off-by: Borislav Petkov <bp@suse.de> Reviewed-by: Tony Luck <tony.luck@intel.com> Acked-by: Michael Ellerman <mpe@ellerman.id.au> Cc: <stable@vger.kernel.org> Link: http://lore.kernel.org/r/CAHk-=wjSqtXAqfUJxFtWNwmguFASTgB0dz1dT3V-78Quiezqbg@mail.gmail.com Link: https://lkml.kernel.org/r/160195561680.2163339.11574962055305783722.stgit@dwillia2-desk3.amr.corp.intel.com
670 lines
18 KiB
C
670 lines
18 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* This file contains work-arounds for x86 and x86_64 platform bugs.
|
|
*/
|
|
#include <linux/dmi.h>
|
|
#include <linux/pci.h>
|
|
#include <linux/irq.h>
|
|
|
|
#include <asm/hpet.h>
|
|
#include <asm/setup.h>
|
|
#include <asm/mce.h>
|
|
|
|
#if defined(CONFIG_X86_IO_APIC) && defined(CONFIG_SMP) && defined(CONFIG_PCI)
|
|
|
|
static void quirk_intel_irqbalance(struct pci_dev *dev)
|
|
{
|
|
u8 config;
|
|
u16 word;
|
|
|
|
/* BIOS may enable hardware IRQ balancing for
|
|
* E7520/E7320/E7525(revision ID 0x9 and below)
|
|
* based platforms.
|
|
* Disable SW irqbalance/affinity on those platforms.
|
|
*/
|
|
if (dev->revision > 0x9)
|
|
return;
|
|
|
|
/* enable access to config space*/
|
|
pci_read_config_byte(dev, 0xf4, &config);
|
|
pci_write_config_byte(dev, 0xf4, config|0x2);
|
|
|
|
/*
|
|
* read xTPR register. We may not have a pci_dev for device 8
|
|
* because it might be hidden until the above write.
|
|
*/
|
|
pci_bus_read_config_word(dev->bus, PCI_DEVFN(8, 0), 0x4c, &word);
|
|
|
|
if (!(word & (1 << 13))) {
|
|
dev_info(&dev->dev, "Intel E7520/7320/7525 detected; "
|
|
"disabling irq balancing and affinity\n");
|
|
noirqdebug_setup("");
|
|
#ifdef CONFIG_PROC_FS
|
|
no_irq_affinity = 1;
|
|
#endif
|
|
}
|
|
|
|
/* put back the original value for config space*/
|
|
if (!(config & 0x2))
|
|
pci_write_config_byte(dev, 0xf4, config);
|
|
}
|
|
DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_E7320_MCH,
|
|
quirk_intel_irqbalance);
|
|
DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_E7525_MCH,
|
|
quirk_intel_irqbalance);
|
|
DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_E7520_MCH,
|
|
quirk_intel_irqbalance);
|
|
#endif
|
|
|
|
#if defined(CONFIG_HPET_TIMER)
|
|
unsigned long force_hpet_address;
|
|
|
|
static enum {
|
|
NONE_FORCE_HPET_RESUME,
|
|
OLD_ICH_FORCE_HPET_RESUME,
|
|
ICH_FORCE_HPET_RESUME,
|
|
VT8237_FORCE_HPET_RESUME,
|
|
NVIDIA_FORCE_HPET_RESUME,
|
|
ATI_FORCE_HPET_RESUME,
|
|
} force_hpet_resume_type;
|
|
|
|
static void __iomem *rcba_base;
|
|
|
|
static void ich_force_hpet_resume(void)
|
|
{
|
|
u32 val;
|
|
|
|
if (!force_hpet_address)
|
|
return;
|
|
|
|
BUG_ON(rcba_base == NULL);
|
|
|
|
/* read the Function Disable register, dword mode only */
|
|
val = readl(rcba_base + 0x3404);
|
|
if (!(val & 0x80)) {
|
|
/* HPET disabled in HPTC. Trying to enable */
|
|
writel(val | 0x80, rcba_base + 0x3404);
|
|
}
|
|
|
|
val = readl(rcba_base + 0x3404);
|
|
if (!(val & 0x80))
|
|
BUG();
|
|
else
|
|
printk(KERN_DEBUG "Force enabled HPET at resume\n");
|
|
}
|
|
|
|
static void ich_force_enable_hpet(struct pci_dev *dev)
|
|
{
|
|
u32 val;
|
|
u32 rcba;
|
|
int err = 0;
|
|
|
|
if (hpet_address || force_hpet_address)
|
|
return;
|
|
|
|
pci_read_config_dword(dev, 0xF0, &rcba);
|
|
rcba &= 0xFFFFC000;
|
|
if (rcba == 0) {
|
|
dev_printk(KERN_DEBUG, &dev->dev, "RCBA disabled; "
|
|
"cannot force enable HPET\n");
|
|
return;
|
|
}
|
|
|
|
/* use bits 31:14, 16 kB aligned */
|
|
rcba_base = ioremap(rcba, 0x4000);
|
|
if (rcba_base == NULL) {
|
|
dev_printk(KERN_DEBUG, &dev->dev, "ioremap failed; "
|
|
"cannot force enable HPET\n");
|
|
return;
|
|
}
|
|
|
|
/* read the Function Disable register, dword mode only */
|
|
val = readl(rcba_base + 0x3404);
|
|
|
|
if (val & 0x80) {
|
|
/* HPET is enabled in HPTC. Just not reported by BIOS */
|
|
val = val & 0x3;
|
|
force_hpet_address = 0xFED00000 | (val << 12);
|
|
dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
|
|
"0x%lx\n", force_hpet_address);
|
|
iounmap(rcba_base);
|
|
return;
|
|
}
|
|
|
|
/* HPET disabled in HPTC. Trying to enable */
|
|
writel(val | 0x80, rcba_base + 0x3404);
|
|
|
|
val = readl(rcba_base + 0x3404);
|
|
if (!(val & 0x80)) {
|
|
err = 1;
|
|
} else {
|
|
val = val & 0x3;
|
|
force_hpet_address = 0xFED00000 | (val << 12);
|
|
}
|
|
|
|
if (err) {
|
|
force_hpet_address = 0;
|
|
iounmap(rcba_base);
|
|
dev_printk(KERN_DEBUG, &dev->dev,
|
|
"Failed to force enable HPET\n");
|
|
} else {
|
|
force_hpet_resume_type = ICH_FORCE_HPET_RESUME;
|
|
dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
|
|
"0x%lx\n", force_hpet_address);
|
|
}
|
|
}
|
|
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ESB2_0,
|
|
ich_force_enable_hpet);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH6_0,
|
|
ich_force_enable_hpet);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH6_1,
|
|
ich_force_enable_hpet);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH7_0,
|
|
ich_force_enable_hpet);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH7_1,
|
|
ich_force_enable_hpet);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH7_31,
|
|
ich_force_enable_hpet);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH8_1,
|
|
ich_force_enable_hpet);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH8_4,
|
|
ich_force_enable_hpet);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ICH9_7,
|
|
ich_force_enable_hpet);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, 0x3a16, /* ICH10 */
|
|
ich_force_enable_hpet);
|
|
|
|
static struct pci_dev *cached_dev;
|
|
|
|
static void hpet_print_force_info(void)
|
|
{
|
|
printk(KERN_INFO "HPET not enabled in BIOS. "
|
|
"You might try hpet=force boot option\n");
|
|
}
|
|
|
|
static void old_ich_force_hpet_resume(void)
|
|
{
|
|
u32 val;
|
|
u32 gen_cntl;
|
|
|
|
if (!force_hpet_address || !cached_dev)
|
|
return;
|
|
|
|
pci_read_config_dword(cached_dev, 0xD0, &gen_cntl);
|
|
gen_cntl &= (~(0x7 << 15));
|
|
gen_cntl |= (0x4 << 15);
|
|
|
|
pci_write_config_dword(cached_dev, 0xD0, gen_cntl);
|
|
pci_read_config_dword(cached_dev, 0xD0, &gen_cntl);
|
|
val = gen_cntl >> 15;
|
|
val &= 0x7;
|
|
if (val == 0x4)
|
|
printk(KERN_DEBUG "Force enabled HPET at resume\n");
|
|
else
|
|
BUG();
|
|
}
|
|
|
|
static void old_ich_force_enable_hpet(struct pci_dev *dev)
|
|
{
|
|
u32 val;
|
|
u32 gen_cntl;
|
|
|
|
if (hpet_address || force_hpet_address)
|
|
return;
|
|
|
|
pci_read_config_dword(dev, 0xD0, &gen_cntl);
|
|
/*
|
|
* Bit 17 is HPET enable bit.
|
|
* Bit 16:15 control the HPET base address.
|
|
*/
|
|
val = gen_cntl >> 15;
|
|
val &= 0x7;
|
|
if (val & 0x4) {
|
|
val &= 0x3;
|
|
force_hpet_address = 0xFED00000 | (val << 12);
|
|
dev_printk(KERN_DEBUG, &dev->dev, "HPET at 0x%lx\n",
|
|
force_hpet_address);
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* HPET is disabled. Trying enabling at FED00000 and check
|
|
* whether it sticks
|
|
*/
|
|
gen_cntl &= (~(0x7 << 15));
|
|
gen_cntl |= (0x4 << 15);
|
|
pci_write_config_dword(dev, 0xD0, gen_cntl);
|
|
|
|
pci_read_config_dword(dev, 0xD0, &gen_cntl);
|
|
|
|
val = gen_cntl >> 15;
|
|
val &= 0x7;
|
|
if (val & 0x4) {
|
|
/* HPET is enabled in HPTC. Just not reported by BIOS */
|
|
val &= 0x3;
|
|
force_hpet_address = 0xFED00000 | (val << 12);
|
|
dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
|
|
"0x%lx\n", force_hpet_address);
|
|
cached_dev = dev;
|
|
force_hpet_resume_type = OLD_ICH_FORCE_HPET_RESUME;
|
|
return;
|
|
}
|
|
|
|
dev_printk(KERN_DEBUG, &dev->dev, "Failed to force enable HPET\n");
|
|
}
|
|
|
|
/*
|
|
* Undocumented chipset features. Make sure that the user enforced
|
|
* this.
|
|
*/
|
|
static void old_ich_force_enable_hpet_user(struct pci_dev *dev)
|
|
{
|
|
if (hpet_force_user)
|
|
old_ich_force_enable_hpet(dev);
|
|
}
|
|
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_ESB_1,
|
|
old_ich_force_enable_hpet_user);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801CA_0,
|
|
old_ich_force_enable_hpet_user);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801CA_12,
|
|
old_ich_force_enable_hpet_user);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801DB_0,
|
|
old_ich_force_enable_hpet_user);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801DB_12,
|
|
old_ich_force_enable_hpet_user);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801EB_0,
|
|
old_ich_force_enable_hpet);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_82801EB_12,
|
|
old_ich_force_enable_hpet);
|
|
|
|
|
|
static void vt8237_force_hpet_resume(void)
|
|
{
|
|
u32 val;
|
|
|
|
if (!force_hpet_address || !cached_dev)
|
|
return;
|
|
|
|
val = 0xfed00000 | 0x80;
|
|
pci_write_config_dword(cached_dev, 0x68, val);
|
|
|
|
pci_read_config_dword(cached_dev, 0x68, &val);
|
|
if (val & 0x80)
|
|
printk(KERN_DEBUG "Force enabled HPET at resume\n");
|
|
else
|
|
BUG();
|
|
}
|
|
|
|
static void vt8237_force_enable_hpet(struct pci_dev *dev)
|
|
{
|
|
u32 val;
|
|
|
|
if (hpet_address || force_hpet_address)
|
|
return;
|
|
|
|
if (!hpet_force_user) {
|
|
hpet_print_force_info();
|
|
return;
|
|
}
|
|
|
|
pci_read_config_dword(dev, 0x68, &val);
|
|
/*
|
|
* Bit 7 is HPET enable bit.
|
|
* Bit 31:10 is HPET base address (contrary to what datasheet claims)
|
|
*/
|
|
if (val & 0x80) {
|
|
force_hpet_address = (val & ~0x3ff);
|
|
dev_printk(KERN_DEBUG, &dev->dev, "HPET at 0x%lx\n",
|
|
force_hpet_address);
|
|
return;
|
|
}
|
|
|
|
/*
|
|
* HPET is disabled. Trying enabling at FED00000 and check
|
|
* whether it sticks
|
|
*/
|
|
val = 0xfed00000 | 0x80;
|
|
pci_write_config_dword(dev, 0x68, val);
|
|
|
|
pci_read_config_dword(dev, 0x68, &val);
|
|
if (val & 0x80) {
|
|
force_hpet_address = (val & ~0x3ff);
|
|
dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
|
|
"0x%lx\n", force_hpet_address);
|
|
cached_dev = dev;
|
|
force_hpet_resume_type = VT8237_FORCE_HPET_RESUME;
|
|
return;
|
|
}
|
|
|
|
dev_printk(KERN_DEBUG, &dev->dev, "Failed to force enable HPET\n");
|
|
}
|
|
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_VIA, PCI_DEVICE_ID_VIA_8235,
|
|
vt8237_force_enable_hpet);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_VIA, PCI_DEVICE_ID_VIA_8237,
|
|
vt8237_force_enable_hpet);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_VIA, PCI_DEVICE_ID_VIA_CX700,
|
|
vt8237_force_enable_hpet);
|
|
|
|
static void ati_force_hpet_resume(void)
|
|
{
|
|
pci_write_config_dword(cached_dev, 0x14, 0xfed00000);
|
|
printk(KERN_DEBUG "Force enabled HPET at resume\n");
|
|
}
|
|
|
|
static u32 ati_ixp4x0_rev(struct pci_dev *dev)
|
|
{
|
|
int err = 0;
|
|
u32 d = 0;
|
|
u8 b = 0;
|
|
|
|
err = pci_read_config_byte(dev, 0xac, &b);
|
|
b &= ~(1<<5);
|
|
err |= pci_write_config_byte(dev, 0xac, b);
|
|
err |= pci_read_config_dword(dev, 0x70, &d);
|
|
d |= 1<<8;
|
|
err |= pci_write_config_dword(dev, 0x70, d);
|
|
err |= pci_read_config_dword(dev, 0x8, &d);
|
|
d &= 0xff;
|
|
dev_printk(KERN_DEBUG, &dev->dev, "SB4X0 revision 0x%x\n", d);
|
|
|
|
WARN_ON_ONCE(err);
|
|
|
|
return d;
|
|
}
|
|
|
|
static void ati_force_enable_hpet(struct pci_dev *dev)
|
|
{
|
|
u32 d, val;
|
|
u8 b;
|
|
|
|
if (hpet_address || force_hpet_address)
|
|
return;
|
|
|
|
if (!hpet_force_user) {
|
|
hpet_print_force_info();
|
|
return;
|
|
}
|
|
|
|
d = ati_ixp4x0_rev(dev);
|
|
if (d < 0x82)
|
|
return;
|
|
|
|
/* base address */
|
|
pci_write_config_dword(dev, 0x14, 0xfed00000);
|
|
pci_read_config_dword(dev, 0x14, &val);
|
|
|
|
/* enable interrupt */
|
|
outb(0x72, 0xcd6); b = inb(0xcd7);
|
|
b |= 0x1;
|
|
outb(0x72, 0xcd6); outb(b, 0xcd7);
|
|
outb(0x72, 0xcd6); b = inb(0xcd7);
|
|
if (!(b & 0x1))
|
|
return;
|
|
pci_read_config_dword(dev, 0x64, &d);
|
|
d |= (1<<10);
|
|
pci_write_config_dword(dev, 0x64, d);
|
|
pci_read_config_dword(dev, 0x64, &d);
|
|
if (!(d & (1<<10)))
|
|
return;
|
|
|
|
force_hpet_address = val;
|
|
force_hpet_resume_type = ATI_FORCE_HPET_RESUME;
|
|
dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at 0x%lx\n",
|
|
force_hpet_address);
|
|
cached_dev = dev;
|
|
}
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_ATI, PCI_DEVICE_ID_ATI_IXP400_SMBUS,
|
|
ati_force_enable_hpet);
|
|
|
|
/*
|
|
* Undocumented chipset feature taken from LinuxBIOS.
|
|
*/
|
|
static void nvidia_force_hpet_resume(void)
|
|
{
|
|
pci_write_config_dword(cached_dev, 0x44, 0xfed00001);
|
|
printk(KERN_DEBUG "Force enabled HPET at resume\n");
|
|
}
|
|
|
|
static void nvidia_force_enable_hpet(struct pci_dev *dev)
|
|
{
|
|
u32 val;
|
|
|
|
if (hpet_address || force_hpet_address)
|
|
return;
|
|
|
|
if (!hpet_force_user) {
|
|
hpet_print_force_info();
|
|
return;
|
|
}
|
|
|
|
pci_write_config_dword(dev, 0x44, 0xfed00001);
|
|
pci_read_config_dword(dev, 0x44, &val);
|
|
force_hpet_address = val & 0xfffffffe;
|
|
force_hpet_resume_type = NVIDIA_FORCE_HPET_RESUME;
|
|
dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at 0x%lx\n",
|
|
force_hpet_address);
|
|
cached_dev = dev;
|
|
}
|
|
|
|
/* ISA Bridges */
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0050,
|
|
nvidia_force_enable_hpet);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0051,
|
|
nvidia_force_enable_hpet);
|
|
|
|
/* LPC bridges */
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0260,
|
|
nvidia_force_enable_hpet);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0360,
|
|
nvidia_force_enable_hpet);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0361,
|
|
nvidia_force_enable_hpet);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0362,
|
|
nvidia_force_enable_hpet);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0363,
|
|
nvidia_force_enable_hpet);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0364,
|
|
nvidia_force_enable_hpet);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0365,
|
|
nvidia_force_enable_hpet);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0366,
|
|
nvidia_force_enable_hpet);
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_NVIDIA, 0x0367,
|
|
nvidia_force_enable_hpet);
|
|
|
|
void force_hpet_resume(void)
|
|
{
|
|
switch (force_hpet_resume_type) {
|
|
case ICH_FORCE_HPET_RESUME:
|
|
ich_force_hpet_resume();
|
|
return;
|
|
case OLD_ICH_FORCE_HPET_RESUME:
|
|
old_ich_force_hpet_resume();
|
|
return;
|
|
case VT8237_FORCE_HPET_RESUME:
|
|
vt8237_force_hpet_resume();
|
|
return;
|
|
case NVIDIA_FORCE_HPET_RESUME:
|
|
nvidia_force_hpet_resume();
|
|
return;
|
|
case ATI_FORCE_HPET_RESUME:
|
|
ati_force_hpet_resume();
|
|
return;
|
|
default:
|
|
break;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* According to the datasheet e6xx systems have the HPET hardwired to
|
|
* 0xfed00000
|
|
*/
|
|
static void e6xx_force_enable_hpet(struct pci_dev *dev)
|
|
{
|
|
if (hpet_address || force_hpet_address)
|
|
return;
|
|
|
|
force_hpet_address = 0xFED00000;
|
|
force_hpet_resume_type = NONE_FORCE_HPET_RESUME;
|
|
dev_printk(KERN_DEBUG, &dev->dev, "Force enabled HPET at "
|
|
"0x%lx\n", force_hpet_address);
|
|
}
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_INTEL, PCI_DEVICE_ID_INTEL_E6XX_CU,
|
|
e6xx_force_enable_hpet);
|
|
|
|
/*
|
|
* HPET MSI on some boards (ATI SB700/SB800) has side effect on
|
|
* floppy DMA. Disable HPET MSI on such platforms.
|
|
* See erratum #27 (Misinterpreted MSI Requests May Result in
|
|
* Corrupted LPC DMA Data) in AMD Publication #46837,
|
|
* "SB700 Family Product Errata", Rev. 1.0, March 2010.
|
|
*/
|
|
static void force_disable_hpet_msi(struct pci_dev *unused)
|
|
{
|
|
hpet_msi_disable = true;
|
|
}
|
|
|
|
DECLARE_PCI_FIXUP_HEADER(PCI_VENDOR_ID_ATI, PCI_DEVICE_ID_ATI_SBX00_SMBUS,
|
|
force_disable_hpet_msi);
|
|
|
|
#endif
|
|
|
|
#if defined(CONFIG_PCI) && defined(CONFIG_NUMA)
|
|
/* Set correct numa_node information for AMD NB functions */
|
|
static void quirk_amd_nb_node(struct pci_dev *dev)
|
|
{
|
|
struct pci_dev *nb_ht;
|
|
unsigned int devfn;
|
|
u32 node;
|
|
u32 val;
|
|
|
|
devfn = PCI_DEVFN(PCI_SLOT(dev->devfn), 0);
|
|
nb_ht = pci_get_slot(dev->bus, devfn);
|
|
if (!nb_ht)
|
|
return;
|
|
|
|
pci_read_config_dword(nb_ht, 0x60, &val);
|
|
node = pcibus_to_node(dev->bus) | (val & 7);
|
|
/*
|
|
* Some hardware may return an invalid node ID,
|
|
* so check it first:
|
|
*/
|
|
if (node_online(node))
|
|
set_dev_node(&dev->dev, node);
|
|
pci_dev_put(nb_ht);
|
|
}
|
|
|
|
DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_K8_NB,
|
|
quirk_amd_nb_node);
|
|
DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_K8_NB_ADDRMAP,
|
|
quirk_amd_nb_node);
|
|
DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_K8_NB_MEMCTL,
|
|
quirk_amd_nb_node);
|
|
DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_K8_NB_MISC,
|
|
quirk_amd_nb_node);
|
|
DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_HT,
|
|
quirk_amd_nb_node);
|
|
DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_MAP,
|
|
quirk_amd_nb_node);
|
|
DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_DRAM,
|
|
quirk_amd_nb_node);
|
|
DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_MISC,
|
|
quirk_amd_nb_node);
|
|
DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_10H_NB_LINK,
|
|
quirk_amd_nb_node);
|
|
DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_F0,
|
|
quirk_amd_nb_node);
|
|
DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_F1,
|
|
quirk_amd_nb_node);
|
|
DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_F2,
|
|
quirk_amd_nb_node);
|
|
DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_F3,
|
|
quirk_amd_nb_node);
|
|
DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_F4,
|
|
quirk_amd_nb_node);
|
|
DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_15H_NB_F5,
|
|
quirk_amd_nb_node);
|
|
|
|
#endif
|
|
|
|
#ifdef CONFIG_PCI
|
|
/*
|
|
* Processor does not ensure DRAM scrub read/write sequence
|
|
* is atomic wrt accesses to CC6 save state area. Therefore
|
|
* if a concurrent scrub read/write access is to same address
|
|
* the entry may appear as if it is not written. This quirk
|
|
* applies to Fam16h models 00h-0Fh
|
|
*
|
|
* See "Revision Guide" for AMD F16h models 00h-0fh,
|
|
* document 51810 rev. 3.04, Nov 2013
|
|
*/
|
|
static void amd_disable_seq_and_redirect_scrub(struct pci_dev *dev)
|
|
{
|
|
u32 val;
|
|
|
|
/*
|
|
* Suggested workaround:
|
|
* set D18F3x58[4:0] = 00h and set D18F3x5C[0] = 0b
|
|
*/
|
|
pci_read_config_dword(dev, 0x58, &val);
|
|
if (val & 0x1F) {
|
|
val &= ~(0x1F);
|
|
pci_write_config_dword(dev, 0x58, val);
|
|
}
|
|
|
|
pci_read_config_dword(dev, 0x5C, &val);
|
|
if (val & BIT(0)) {
|
|
val &= ~BIT(0);
|
|
pci_write_config_dword(dev, 0x5c, val);
|
|
}
|
|
}
|
|
|
|
DECLARE_PCI_FIXUP_FINAL(PCI_VENDOR_ID_AMD, PCI_DEVICE_ID_AMD_16H_NB_F3,
|
|
amd_disable_seq_and_redirect_scrub);
|
|
|
|
/* Ivy Bridge, Haswell, Broadwell */
|
|
static void quirk_intel_brickland_xeon_ras_cap(struct pci_dev *pdev)
|
|
{
|
|
u32 capid0;
|
|
|
|
pci_read_config_dword(pdev, 0x84, &capid0);
|
|
|
|
if (capid0 & 0x10)
|
|
enable_copy_mc_fragile();
|
|
}
|
|
|
|
/* Skylake */
|
|
static void quirk_intel_purley_xeon_ras_cap(struct pci_dev *pdev)
|
|
{
|
|
u32 capid0, capid5;
|
|
|
|
pci_read_config_dword(pdev, 0x84, &capid0);
|
|
pci_read_config_dword(pdev, 0x98, &capid5);
|
|
|
|
/*
|
|
* CAPID0{7:6} indicate whether this is an advanced RAS SKU
|
|
* CAPID5{8:5} indicate that various NVDIMM usage modes are
|
|
* enabled, so memory machine check recovery is also enabled.
|
|
*/
|
|
if ((capid0 & 0xc0) == 0xc0 || (capid5 & 0x1e0))
|
|
enable_copy_mc_fragile();
|
|
|
|
}
|
|
DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_INTEL, 0x0ec3, quirk_intel_brickland_xeon_ras_cap);
|
|
DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_INTEL, 0x2fc0, quirk_intel_brickland_xeon_ras_cap);
|
|
DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_INTEL, 0x6fc0, quirk_intel_brickland_xeon_ras_cap);
|
|
DECLARE_PCI_FIXUP_EARLY(PCI_VENDOR_ID_INTEL, 0x2083, quirk_intel_purley_xeon_ras_cap);
|
|
#endif
|
|
|
|
bool x86_apple_machine;
|
|
EXPORT_SYMBOL(x86_apple_machine);
|
|
|
|
void __init early_platform_quirks(void)
|
|
{
|
|
x86_apple_machine = dmi_match(DMI_SYS_VENDOR, "Apple Inc.") ||
|
|
dmi_match(DMI_SYS_VENDOR, "Apple Computer, Inc.");
|
|
}
|