mirror of
https://github.com/torvalds/linux.git
synced 2024-11-25 13:41:51 +00:00
26e9baa849
There are currently 2 ways to create a set of sysfs files for a
kobj_type, through the default_attrs field, and the default_groups
field. Move the ioatdma sysfs code to use default_groups field which has
been the preferred way since aa30f47cf6
("kobject: Add support for
default attribute groups to kobj_type") so that we can soon get rid of
the obsolete default_attrs field.
Cc: Vinod Koul <vkoul@kernel.org>
Cc: dmaengine@vger.kernel.org
Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Acked-by: Dave Jiang <dave.jiang@intel.com>
Link: https://lore.kernel.org/r/20220104163330.1338824-1-gregkh@linuxfoundation.org
Signed-off-by: Vinod Koul <vkoul@kernel.org>
167 lines
4.4 KiB
C
167 lines
4.4 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* Intel I/OAT DMA Linux driver
|
|
* Copyright(c) 2004 - 2015 Intel Corporation.
|
|
*/
|
|
|
|
#include <linux/init.h>
|
|
#include <linux/module.h>
|
|
#include <linux/dmaengine.h>
|
|
#include <linux/pci.h>
|
|
#include "dma.h"
|
|
#include "registers.h"
|
|
#include "hw.h"
|
|
|
|
#include "../dmaengine.h"
|
|
|
|
static ssize_t cap_show(struct dma_chan *c, char *page)
|
|
{
|
|
struct dma_device *dma = c->device;
|
|
|
|
return sprintf(page, "copy%s%s%s%s%s\n",
|
|
dma_has_cap(DMA_PQ, dma->cap_mask) ? " pq" : "",
|
|
dma_has_cap(DMA_PQ_VAL, dma->cap_mask) ? " pq_val" : "",
|
|
dma_has_cap(DMA_XOR, dma->cap_mask) ? " xor" : "",
|
|
dma_has_cap(DMA_XOR_VAL, dma->cap_mask) ? " xor_val" : "",
|
|
dma_has_cap(DMA_INTERRUPT, dma->cap_mask) ? " intr" : "");
|
|
|
|
}
|
|
struct ioat_sysfs_entry ioat_cap_attr = __ATTR_RO(cap);
|
|
|
|
static ssize_t version_show(struct dma_chan *c, char *page)
|
|
{
|
|
struct dma_device *dma = c->device;
|
|
struct ioatdma_device *ioat_dma = to_ioatdma_device(dma);
|
|
|
|
return sprintf(page, "%d.%d\n",
|
|
ioat_dma->version >> 4, ioat_dma->version & 0xf);
|
|
}
|
|
struct ioat_sysfs_entry ioat_version_attr = __ATTR_RO(version);
|
|
|
|
static ssize_t
|
|
ioat_attr_show(struct kobject *kobj, struct attribute *attr, char *page)
|
|
{
|
|
struct ioat_sysfs_entry *entry;
|
|
struct ioatdma_chan *ioat_chan;
|
|
|
|
entry = container_of(attr, struct ioat_sysfs_entry, attr);
|
|
ioat_chan = container_of(kobj, struct ioatdma_chan, kobj);
|
|
|
|
if (!entry->show)
|
|
return -EIO;
|
|
return entry->show(&ioat_chan->dma_chan, page);
|
|
}
|
|
|
|
static ssize_t
|
|
ioat_attr_store(struct kobject *kobj, struct attribute *attr,
|
|
const char *page, size_t count)
|
|
{
|
|
struct ioat_sysfs_entry *entry;
|
|
struct ioatdma_chan *ioat_chan;
|
|
|
|
entry = container_of(attr, struct ioat_sysfs_entry, attr);
|
|
ioat_chan = container_of(kobj, struct ioatdma_chan, kobj);
|
|
|
|
if (!entry->store)
|
|
return -EIO;
|
|
return entry->store(&ioat_chan->dma_chan, page, count);
|
|
}
|
|
|
|
const struct sysfs_ops ioat_sysfs_ops = {
|
|
.show = ioat_attr_show,
|
|
.store = ioat_attr_store,
|
|
};
|
|
|
|
void ioat_kobject_add(struct ioatdma_device *ioat_dma, struct kobj_type *type)
|
|
{
|
|
struct dma_device *dma = &ioat_dma->dma_dev;
|
|
struct dma_chan *c;
|
|
|
|
list_for_each_entry(c, &dma->channels, device_node) {
|
|
struct ioatdma_chan *ioat_chan = to_ioat_chan(c);
|
|
struct kobject *parent = &c->dev->device.kobj;
|
|
int err;
|
|
|
|
err = kobject_init_and_add(&ioat_chan->kobj, type,
|
|
parent, "quickdata");
|
|
if (err) {
|
|
dev_warn(to_dev(ioat_chan),
|
|
"sysfs init error (%d), continuing...\n", err);
|
|
kobject_put(&ioat_chan->kobj);
|
|
set_bit(IOAT_KOBJ_INIT_FAIL, &ioat_chan->state);
|
|
}
|
|
}
|
|
}
|
|
|
|
void ioat_kobject_del(struct ioatdma_device *ioat_dma)
|
|
{
|
|
struct dma_device *dma = &ioat_dma->dma_dev;
|
|
struct dma_chan *c;
|
|
|
|
list_for_each_entry(c, &dma->channels, device_node) {
|
|
struct ioatdma_chan *ioat_chan = to_ioat_chan(c);
|
|
|
|
if (!test_bit(IOAT_KOBJ_INIT_FAIL, &ioat_chan->state)) {
|
|
kobject_del(&ioat_chan->kobj);
|
|
kobject_put(&ioat_chan->kobj);
|
|
}
|
|
}
|
|
}
|
|
|
|
static ssize_t ring_size_show(struct dma_chan *c, char *page)
|
|
{
|
|
struct ioatdma_chan *ioat_chan = to_ioat_chan(c);
|
|
|
|
return sprintf(page, "%d\n", (1 << ioat_chan->alloc_order) & ~1);
|
|
}
|
|
static struct ioat_sysfs_entry ring_size_attr = __ATTR_RO(ring_size);
|
|
|
|
static ssize_t ring_active_show(struct dma_chan *c, char *page)
|
|
{
|
|
struct ioatdma_chan *ioat_chan = to_ioat_chan(c);
|
|
|
|
/* ...taken outside the lock, no need to be precise */
|
|
return sprintf(page, "%d\n", ioat_ring_active(ioat_chan));
|
|
}
|
|
static struct ioat_sysfs_entry ring_active_attr = __ATTR_RO(ring_active);
|
|
|
|
static ssize_t intr_coalesce_show(struct dma_chan *c, char *page)
|
|
{
|
|
struct ioatdma_chan *ioat_chan = to_ioat_chan(c);
|
|
|
|
return sprintf(page, "%d\n", ioat_chan->intr_coalesce);
|
|
}
|
|
|
|
static ssize_t intr_coalesce_store(struct dma_chan *c, const char *page,
|
|
size_t count)
|
|
{
|
|
int intr_coalesce = 0;
|
|
struct ioatdma_chan *ioat_chan = to_ioat_chan(c);
|
|
|
|
if (sscanf(page, "%du", &intr_coalesce) != -1) {
|
|
if ((intr_coalesce < 0) ||
|
|
(intr_coalesce > IOAT_INTRDELAY_MASK))
|
|
return -EINVAL;
|
|
ioat_chan->intr_coalesce = intr_coalesce;
|
|
}
|
|
|
|
return count;
|
|
}
|
|
|
|
static struct ioat_sysfs_entry intr_coalesce_attr = __ATTR_RW(intr_coalesce);
|
|
|
|
static struct attribute *ioat_attrs[] = {
|
|
&ring_size_attr.attr,
|
|
&ring_active_attr.attr,
|
|
&ioat_cap_attr.attr,
|
|
&ioat_version_attr.attr,
|
|
&intr_coalesce_attr.attr,
|
|
NULL,
|
|
};
|
|
ATTRIBUTE_GROUPS(ioat);
|
|
|
|
struct kobj_type ioat_ktype = {
|
|
.sysfs_ops = &ioat_sysfs_ops,
|
|
.default_groups = ioat_groups,
|
|
};
|