forked from Minki/linux
d43f20bae5
Error reporting interrupt is very useful to notify users that some errors are detected by the hardware. Once users are notified, they could query hardware logged error states, no need to continuously poll on these states. This patch adds interrupt support for fme global error reporting sub feature. It follows the common DFL interrupt notification and handling mechanism. And it implements two ioctls below for user to query number of irqs supported, and set/unset interrupt triggers. Ioctls: * DFL_FPGA_FME_ERR_GET_IRQ_NUM get the number of irqs, which is used to determine whether/how many interrupts fme error reporting feature supports. * DFL_FPGA_FME_ERR_SET_IRQ set/unset given eventfds as fme error reporting interrupt triggers. Signed-off-by: Luwei Kang <luwei.kang@intel.com> Signed-off-by: Wu Hao <hao.wu@intel.com> Signed-off-by: Xu Yilun <yilun.xu@intel.com> Reviewed-by: Marcelo Tosatti <mtosatti@redhat.com> Acked-by: Wu Hao <hao.wu@intel.com> Signed-off-by: Moritz Fischer <mdf@kernel.org>
378 lines
9.6 KiB
C
378 lines
9.6 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Driver for FPGA Management Engine Error Management
|
|
*
|
|
* Copyright 2019 Intel Corporation, Inc.
|
|
*
|
|
* Authors:
|
|
* Kang Luwei <luwei.kang@intel.com>
|
|
* Xiao Guangrong <guangrong.xiao@linux.intel.com>
|
|
* Wu Hao <hao.wu@intel.com>
|
|
* Joseph Grecco <joe.grecco@intel.com>
|
|
* Enno Luebbers <enno.luebbers@intel.com>
|
|
* Tim Whisonant <tim.whisonant@intel.com>
|
|
* Ananda Ravuri <ananda.ravuri@intel.com>
|
|
* Mitchel, Henry <henry.mitchel@intel.com>
|
|
*/
|
|
|
|
#include <linux/fpga-dfl.h>
|
|
#include <linux/uaccess.h>
|
|
|
|
#include "dfl.h"
|
|
#include "dfl-fme.h"
|
|
|
|
#define FME_ERROR_MASK 0x8
|
|
#define FME_ERROR 0x10
|
|
#define MBP_ERROR BIT_ULL(6)
|
|
#define PCIE0_ERROR_MASK 0x18
|
|
#define PCIE0_ERROR 0x20
|
|
#define PCIE1_ERROR_MASK 0x28
|
|
#define PCIE1_ERROR 0x30
|
|
#define FME_FIRST_ERROR 0x38
|
|
#define FME_NEXT_ERROR 0x40
|
|
#define RAS_NONFAT_ERROR_MASK 0x48
|
|
#define RAS_NONFAT_ERROR 0x50
|
|
#define RAS_CATFAT_ERROR_MASK 0x58
|
|
#define RAS_CATFAT_ERROR 0x60
|
|
#define RAS_ERROR_INJECT 0x68
|
|
#define INJECT_ERROR_MASK GENMASK_ULL(2, 0)
|
|
|
|
#define ERROR_MASK GENMASK_ULL(63, 0)
|
|
|
|
static ssize_t pcie0_errors_show(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
struct dfl_feature_platform_data *pdata = dev_get_platdata(dev);
|
|
void __iomem *base;
|
|
u64 value;
|
|
|
|
base = dfl_get_feature_ioaddr_by_id(dev, FME_FEATURE_ID_GLOBAL_ERR);
|
|
|
|
mutex_lock(&pdata->lock);
|
|
value = readq(base + PCIE0_ERROR);
|
|
mutex_unlock(&pdata->lock);
|
|
|
|
return sprintf(buf, "0x%llx\n", (unsigned long long)value);
|
|
}
|
|
|
|
static ssize_t pcie0_errors_store(struct device *dev,
|
|
struct device_attribute *attr,
|
|
const char *buf, size_t count)
|
|
{
|
|
struct dfl_feature_platform_data *pdata = dev_get_platdata(dev);
|
|
void __iomem *base;
|
|
int ret = 0;
|
|
u64 v, val;
|
|
|
|
if (kstrtou64(buf, 0, &val))
|
|
return -EINVAL;
|
|
|
|
base = dfl_get_feature_ioaddr_by_id(dev, FME_FEATURE_ID_GLOBAL_ERR);
|
|
|
|
mutex_lock(&pdata->lock);
|
|
writeq(GENMASK_ULL(63, 0), base + PCIE0_ERROR_MASK);
|
|
|
|
v = readq(base + PCIE0_ERROR);
|
|
if (val == v)
|
|
writeq(v, base + PCIE0_ERROR);
|
|
else
|
|
ret = -EINVAL;
|
|
|
|
writeq(0ULL, base + PCIE0_ERROR_MASK);
|
|
mutex_unlock(&pdata->lock);
|
|
return ret ? ret : count;
|
|
}
|
|
static DEVICE_ATTR_RW(pcie0_errors);
|
|
|
|
static ssize_t pcie1_errors_show(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
struct dfl_feature_platform_data *pdata = dev_get_platdata(dev);
|
|
void __iomem *base;
|
|
u64 value;
|
|
|
|
base = dfl_get_feature_ioaddr_by_id(dev, FME_FEATURE_ID_GLOBAL_ERR);
|
|
|
|
mutex_lock(&pdata->lock);
|
|
value = readq(base + PCIE1_ERROR);
|
|
mutex_unlock(&pdata->lock);
|
|
|
|
return sprintf(buf, "0x%llx\n", (unsigned long long)value);
|
|
}
|
|
|
|
static ssize_t pcie1_errors_store(struct device *dev,
|
|
struct device_attribute *attr,
|
|
const char *buf, size_t count)
|
|
{
|
|
struct dfl_feature_platform_data *pdata = dev_get_platdata(dev);
|
|
void __iomem *base;
|
|
int ret = 0;
|
|
u64 v, val;
|
|
|
|
if (kstrtou64(buf, 0, &val))
|
|
return -EINVAL;
|
|
|
|
base = dfl_get_feature_ioaddr_by_id(dev, FME_FEATURE_ID_GLOBAL_ERR);
|
|
|
|
mutex_lock(&pdata->lock);
|
|
writeq(GENMASK_ULL(63, 0), base + PCIE1_ERROR_MASK);
|
|
|
|
v = readq(base + PCIE1_ERROR);
|
|
if (val == v)
|
|
writeq(v, base + PCIE1_ERROR);
|
|
else
|
|
ret = -EINVAL;
|
|
|
|
writeq(0ULL, base + PCIE1_ERROR_MASK);
|
|
mutex_unlock(&pdata->lock);
|
|
return ret ? ret : count;
|
|
}
|
|
static DEVICE_ATTR_RW(pcie1_errors);
|
|
|
|
static ssize_t nonfatal_errors_show(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
void __iomem *base;
|
|
|
|
base = dfl_get_feature_ioaddr_by_id(dev, FME_FEATURE_ID_GLOBAL_ERR);
|
|
|
|
return sprintf(buf, "0x%llx\n",
|
|
(unsigned long long)readq(base + RAS_NONFAT_ERROR));
|
|
}
|
|
static DEVICE_ATTR_RO(nonfatal_errors);
|
|
|
|
static ssize_t catfatal_errors_show(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
void __iomem *base;
|
|
|
|
base = dfl_get_feature_ioaddr_by_id(dev, FME_FEATURE_ID_GLOBAL_ERR);
|
|
|
|
return sprintf(buf, "0x%llx\n",
|
|
(unsigned long long)readq(base + RAS_CATFAT_ERROR));
|
|
}
|
|
static DEVICE_ATTR_RO(catfatal_errors);
|
|
|
|
static ssize_t inject_errors_show(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
struct dfl_feature_platform_data *pdata = dev_get_platdata(dev);
|
|
void __iomem *base;
|
|
u64 v;
|
|
|
|
base = dfl_get_feature_ioaddr_by_id(dev, FME_FEATURE_ID_GLOBAL_ERR);
|
|
|
|
mutex_lock(&pdata->lock);
|
|
v = readq(base + RAS_ERROR_INJECT);
|
|
mutex_unlock(&pdata->lock);
|
|
|
|
return sprintf(buf, "0x%llx\n",
|
|
(unsigned long long)FIELD_GET(INJECT_ERROR_MASK, v));
|
|
}
|
|
|
|
static ssize_t inject_errors_store(struct device *dev,
|
|
struct device_attribute *attr,
|
|
const char *buf, size_t count)
|
|
{
|
|
struct dfl_feature_platform_data *pdata = dev_get_platdata(dev);
|
|
void __iomem *base;
|
|
u8 inject_error;
|
|
u64 v;
|
|
|
|
if (kstrtou8(buf, 0, &inject_error))
|
|
return -EINVAL;
|
|
|
|
if (inject_error & ~INJECT_ERROR_MASK)
|
|
return -EINVAL;
|
|
|
|
base = dfl_get_feature_ioaddr_by_id(dev, FME_FEATURE_ID_GLOBAL_ERR);
|
|
|
|
mutex_lock(&pdata->lock);
|
|
v = readq(base + RAS_ERROR_INJECT);
|
|
v &= ~INJECT_ERROR_MASK;
|
|
v |= FIELD_PREP(INJECT_ERROR_MASK, inject_error);
|
|
writeq(v, base + RAS_ERROR_INJECT);
|
|
mutex_unlock(&pdata->lock);
|
|
|
|
return count;
|
|
}
|
|
static DEVICE_ATTR_RW(inject_errors);
|
|
|
|
static ssize_t fme_errors_show(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
struct dfl_feature_platform_data *pdata = dev_get_platdata(dev);
|
|
void __iomem *base;
|
|
u64 value;
|
|
|
|
base = dfl_get_feature_ioaddr_by_id(dev, FME_FEATURE_ID_GLOBAL_ERR);
|
|
|
|
mutex_lock(&pdata->lock);
|
|
value = readq(base + FME_ERROR);
|
|
mutex_unlock(&pdata->lock);
|
|
|
|
return sprintf(buf, "0x%llx\n", (unsigned long long)value);
|
|
}
|
|
|
|
static ssize_t fme_errors_store(struct device *dev,
|
|
struct device_attribute *attr,
|
|
const char *buf, size_t count)
|
|
{
|
|
struct dfl_feature_platform_data *pdata = dev_get_platdata(dev);
|
|
void __iomem *base;
|
|
u64 v, val;
|
|
int ret = 0;
|
|
|
|
if (kstrtou64(buf, 0, &val))
|
|
return -EINVAL;
|
|
|
|
base = dfl_get_feature_ioaddr_by_id(dev, FME_FEATURE_ID_GLOBAL_ERR);
|
|
|
|
mutex_lock(&pdata->lock);
|
|
writeq(GENMASK_ULL(63, 0), base + FME_ERROR_MASK);
|
|
|
|
v = readq(base + FME_ERROR);
|
|
if (val == v)
|
|
writeq(v, base + FME_ERROR);
|
|
else
|
|
ret = -EINVAL;
|
|
|
|
/* Workaround: disable MBP_ERROR if feature revision is 0 */
|
|
writeq(dfl_feature_revision(base) ? 0ULL : MBP_ERROR,
|
|
base + FME_ERROR_MASK);
|
|
mutex_unlock(&pdata->lock);
|
|
return ret ? ret : count;
|
|
}
|
|
static DEVICE_ATTR_RW(fme_errors);
|
|
|
|
static ssize_t first_error_show(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
struct dfl_feature_platform_data *pdata = dev_get_platdata(dev);
|
|
void __iomem *base;
|
|
u64 value;
|
|
|
|
base = dfl_get_feature_ioaddr_by_id(dev, FME_FEATURE_ID_GLOBAL_ERR);
|
|
|
|
mutex_lock(&pdata->lock);
|
|
value = readq(base + FME_FIRST_ERROR);
|
|
mutex_unlock(&pdata->lock);
|
|
|
|
return sprintf(buf, "0x%llx\n", (unsigned long long)value);
|
|
}
|
|
static DEVICE_ATTR_RO(first_error);
|
|
|
|
static ssize_t next_error_show(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
struct dfl_feature_platform_data *pdata = dev_get_platdata(dev);
|
|
void __iomem *base;
|
|
u64 value;
|
|
|
|
base = dfl_get_feature_ioaddr_by_id(dev, FME_FEATURE_ID_GLOBAL_ERR);
|
|
|
|
mutex_lock(&pdata->lock);
|
|
value = readq(base + FME_NEXT_ERROR);
|
|
mutex_unlock(&pdata->lock);
|
|
|
|
return sprintf(buf, "0x%llx\n", (unsigned long long)value);
|
|
}
|
|
static DEVICE_ATTR_RO(next_error);
|
|
|
|
static struct attribute *fme_global_err_attrs[] = {
|
|
&dev_attr_pcie0_errors.attr,
|
|
&dev_attr_pcie1_errors.attr,
|
|
&dev_attr_nonfatal_errors.attr,
|
|
&dev_attr_catfatal_errors.attr,
|
|
&dev_attr_inject_errors.attr,
|
|
&dev_attr_fme_errors.attr,
|
|
&dev_attr_first_error.attr,
|
|
&dev_attr_next_error.attr,
|
|
NULL,
|
|
};
|
|
|
|
static umode_t fme_global_err_attrs_visible(struct kobject *kobj,
|
|
struct attribute *attr, int n)
|
|
{
|
|
struct device *dev = kobj_to_dev(kobj);
|
|
|
|
/*
|
|
* sysfs entries are visible only if related private feature is
|
|
* enumerated.
|
|
*/
|
|
if (!dfl_get_feature_by_id(dev, FME_FEATURE_ID_GLOBAL_ERR))
|
|
return 0;
|
|
|
|
return attr->mode;
|
|
}
|
|
|
|
const struct attribute_group fme_global_err_group = {
|
|
.name = "errors",
|
|
.attrs = fme_global_err_attrs,
|
|
.is_visible = fme_global_err_attrs_visible,
|
|
};
|
|
|
|
static void fme_err_mask(struct device *dev, bool mask)
|
|
{
|
|
struct dfl_feature_platform_data *pdata = dev_get_platdata(dev);
|
|
void __iomem *base;
|
|
|
|
base = dfl_get_feature_ioaddr_by_id(dev, FME_FEATURE_ID_GLOBAL_ERR);
|
|
|
|
mutex_lock(&pdata->lock);
|
|
|
|
/* Workaround: keep MBP_ERROR always masked if revision is 0 */
|
|
if (dfl_feature_revision(base))
|
|
writeq(mask ? ERROR_MASK : 0, base + FME_ERROR_MASK);
|
|
else
|
|
writeq(mask ? ERROR_MASK : MBP_ERROR, base + FME_ERROR_MASK);
|
|
|
|
writeq(mask ? ERROR_MASK : 0, base + PCIE0_ERROR_MASK);
|
|
writeq(mask ? ERROR_MASK : 0, base + PCIE1_ERROR_MASK);
|
|
writeq(mask ? ERROR_MASK : 0, base + RAS_NONFAT_ERROR_MASK);
|
|
writeq(mask ? ERROR_MASK : 0, base + RAS_CATFAT_ERROR_MASK);
|
|
|
|
mutex_unlock(&pdata->lock);
|
|
}
|
|
|
|
static int fme_global_err_init(struct platform_device *pdev,
|
|
struct dfl_feature *feature)
|
|
{
|
|
fme_err_mask(&pdev->dev, false);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void fme_global_err_uinit(struct platform_device *pdev,
|
|
struct dfl_feature *feature)
|
|
{
|
|
fme_err_mask(&pdev->dev, true);
|
|
}
|
|
|
|
static long
|
|
fme_global_error_ioctl(struct platform_device *pdev,
|
|
struct dfl_feature *feature,
|
|
unsigned int cmd, unsigned long arg)
|
|
{
|
|
switch (cmd) {
|
|
case DFL_FPGA_FME_ERR_GET_IRQ_NUM:
|
|
return dfl_feature_ioctl_get_num_irqs(pdev, feature, arg);
|
|
case DFL_FPGA_FME_ERR_SET_IRQ:
|
|
return dfl_feature_ioctl_set_irq(pdev, feature, arg);
|
|
default:
|
|
dev_dbg(&pdev->dev, "%x cmd not handled", cmd);
|
|
return -ENODEV;
|
|
}
|
|
}
|
|
|
|
const struct dfl_feature_id fme_global_err_id_table[] = {
|
|
{.id = FME_FEATURE_ID_GLOBAL_ERR,},
|
|
{0,}
|
|
};
|
|
|
|
const struct dfl_feature_ops fme_global_err_ops = {
|
|
.init = fme_global_err_init,
|
|
.uinit = fme_global_err_uinit,
|
|
.ioctl = fme_global_error_ioctl,
|
|
};
|