mirror of
https://github.com/torvalds/linux.git
synced 2024-11-25 13:41:51 +00:00
825d8bbd2f
Commit4f563a6473
("block: add a max_user_discard_sectors queue limit") changed block core to set max_discard_sectors to: min(lim->max_hw_discard_sectors, lim->max_user_discard_sectors) Since commit1c0e720228
("dm: use queue_limits_set") it was reported dm-thinp was failing in a few fstests (generic/347 and generic/405) with the first WARN_ON_ONCE in dm_cell_key_has_valid_range() being reported, e.g.: WARNING: CPU: 1 PID: 30 at drivers/md/dm-bio-prison-v1.c:128 dm_cell_key_has_valid_range+0x3d/0x50 blk_set_stacking_limits() sets max_user_discard_sectors to UINT_MAX, so given how block core now sets max_discard_sectors (detailed above) it follows that blk_stack_limits() stacks up the underlying device's max_hw_discard_sectors and max_discard_sectors is set to match it. If max_hw_discard_sectors exceeds dm's BIO_PRISON_MAX_RANGE, then dm_cell_key_has_valid_range() will trigger the warning with: WARN_ON_ONCE(key->block_end - key->block_begin > BIO_PRISON_MAX_RANGE) Aside from this warning, the discard will fail. Fix this and other DM issues by governing discard support in terms of max_hw_discard_sectors instead of max_discard_sectors. Reported-by: Theodore Ts'o <tytso@mit.edu> Fixes:1c0e720228
("dm: use queue_limits_set") Signed-off-by: Mike Snitzer <snitzer@kernel.org>
287 lines
5.6 KiB
C
287 lines
5.6 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* Copyright (C) 2001 Sistina Software (UK) Limited
|
|
*
|
|
* This file is released under the GPL.
|
|
*/
|
|
|
|
#include "dm-core.h"
|
|
|
|
#include <linux/module.h>
|
|
#include <linux/init.h>
|
|
#include <linux/kmod.h>
|
|
#include <linux/bio.h>
|
|
#include <linux/dax.h>
|
|
|
|
#define DM_MSG_PREFIX "target"
|
|
|
|
static LIST_HEAD(_targets);
|
|
static DECLARE_RWSEM(_lock);
|
|
|
|
static inline struct target_type *__find_target_type(const char *name)
|
|
{
|
|
struct target_type *tt;
|
|
|
|
list_for_each_entry(tt, &_targets, list)
|
|
if (!strcmp(name, tt->name))
|
|
return tt;
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static struct target_type *get_target_type(const char *name)
|
|
{
|
|
struct target_type *tt;
|
|
|
|
down_read(&_lock);
|
|
|
|
tt = __find_target_type(name);
|
|
if (tt && !try_module_get(tt->module))
|
|
tt = NULL;
|
|
|
|
up_read(&_lock);
|
|
return tt;
|
|
}
|
|
|
|
static void load_module(const char *name)
|
|
{
|
|
request_module("dm-%s", name);
|
|
}
|
|
|
|
struct target_type *dm_get_target_type(const char *name)
|
|
{
|
|
struct target_type *tt = get_target_type(name);
|
|
|
|
if (!tt) {
|
|
load_module(name);
|
|
tt = get_target_type(name);
|
|
}
|
|
|
|
return tt;
|
|
}
|
|
|
|
void dm_put_target_type(struct target_type *tt)
|
|
{
|
|
down_read(&_lock);
|
|
module_put(tt->module);
|
|
up_read(&_lock);
|
|
}
|
|
|
|
int dm_target_iterate(void (*iter_func)(struct target_type *tt,
|
|
void *param), void *param)
|
|
{
|
|
struct target_type *tt;
|
|
|
|
down_read(&_lock);
|
|
list_for_each_entry(tt, &_targets, list)
|
|
iter_func(tt, param);
|
|
up_read(&_lock);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int dm_register_target(struct target_type *tt)
|
|
{
|
|
int rv = 0;
|
|
|
|
down_write(&_lock);
|
|
if (__find_target_type(tt->name)) {
|
|
DMERR("%s: '%s' target already registered",
|
|
__func__, tt->name);
|
|
rv = -EEXIST;
|
|
} else {
|
|
list_add(&tt->list, &_targets);
|
|
}
|
|
up_write(&_lock);
|
|
|
|
return rv;
|
|
}
|
|
EXPORT_SYMBOL(dm_register_target);
|
|
|
|
void dm_unregister_target(struct target_type *tt)
|
|
{
|
|
down_write(&_lock);
|
|
if (!__find_target_type(tt->name)) {
|
|
DMCRIT("Unregistering unrecognised target: %s", tt->name);
|
|
BUG();
|
|
}
|
|
|
|
list_del(&tt->list);
|
|
|
|
up_write(&_lock);
|
|
}
|
|
EXPORT_SYMBOL(dm_unregister_target);
|
|
|
|
/*
|
|
* io-err: always fails an io, useful for bringing
|
|
* up LVs that have holes in them.
|
|
*/
|
|
struct io_err_c {
|
|
struct dm_dev *dev;
|
|
sector_t start;
|
|
};
|
|
|
|
static int io_err_get_args(struct dm_target *tt, unsigned int argc, char **args)
|
|
{
|
|
unsigned long long start;
|
|
struct io_err_c *ioec;
|
|
char dummy;
|
|
int ret;
|
|
|
|
ioec = kmalloc(sizeof(*ioec), GFP_KERNEL);
|
|
if (!ioec) {
|
|
tt->error = "Cannot allocate io_err context";
|
|
return -ENOMEM;
|
|
}
|
|
|
|
ret = -EINVAL;
|
|
if (sscanf(args[1], "%llu%c", &start, &dummy) != 1 ||
|
|
start != (sector_t)start) {
|
|
tt->error = "Invalid device sector";
|
|
goto bad;
|
|
}
|
|
ioec->start = start;
|
|
|
|
ret = dm_get_device(tt, args[0], dm_table_get_mode(tt->table), &ioec->dev);
|
|
if (ret) {
|
|
tt->error = "Device lookup failed";
|
|
goto bad;
|
|
}
|
|
|
|
tt->private = ioec;
|
|
|
|
return 0;
|
|
|
|
bad:
|
|
kfree(ioec);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int io_err_ctr(struct dm_target *tt, unsigned int argc, char **args)
|
|
{
|
|
/*
|
|
* If we have arguments, assume it is the path to the backing
|
|
* block device and its mapping start sector (same as dm-linear).
|
|
* In this case, get the device so that we can get its limits.
|
|
*/
|
|
if (argc == 2) {
|
|
int ret = io_err_get_args(tt, argc, args);
|
|
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
/*
|
|
* Return error for discards instead of -EOPNOTSUPP
|
|
*/
|
|
tt->num_discard_bios = 1;
|
|
tt->discards_supported = true;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void io_err_dtr(struct dm_target *tt)
|
|
{
|
|
struct io_err_c *ioec = tt->private;
|
|
|
|
if (ioec) {
|
|
dm_put_device(tt, ioec->dev);
|
|
kfree(ioec);
|
|
}
|
|
}
|
|
|
|
static int io_err_map(struct dm_target *tt, struct bio *bio)
|
|
{
|
|
return DM_MAPIO_KILL;
|
|
}
|
|
|
|
static int io_err_clone_and_map_rq(struct dm_target *ti, struct request *rq,
|
|
union map_info *map_context,
|
|
struct request **clone)
|
|
{
|
|
return DM_MAPIO_KILL;
|
|
}
|
|
|
|
static void io_err_release_clone_rq(struct request *clone,
|
|
union map_info *map_context)
|
|
{
|
|
}
|
|
|
|
#ifdef CONFIG_BLK_DEV_ZONED
|
|
static sector_t io_err_map_sector(struct dm_target *ti, sector_t bi_sector)
|
|
{
|
|
struct io_err_c *ioec = ti->private;
|
|
|
|
return ioec->start + dm_target_offset(ti, bi_sector);
|
|
}
|
|
|
|
static int io_err_report_zones(struct dm_target *ti,
|
|
struct dm_report_zones_args *args, unsigned int nr_zones)
|
|
{
|
|
struct io_err_c *ioec = ti->private;
|
|
|
|
/*
|
|
* This should never be called when we do not have a backing device
|
|
* as that mean the target is not a zoned one.
|
|
*/
|
|
if (WARN_ON_ONCE(!ioec))
|
|
return -EIO;
|
|
|
|
return dm_report_zones(ioec->dev->bdev, ioec->start,
|
|
io_err_map_sector(ti, args->next_sector),
|
|
args, nr_zones);
|
|
}
|
|
#else
|
|
#define io_err_report_zones NULL
|
|
#endif
|
|
|
|
static int io_err_iterate_devices(struct dm_target *ti,
|
|
iterate_devices_callout_fn fn, void *data)
|
|
{
|
|
struct io_err_c *ioec = ti->private;
|
|
|
|
if (!ioec)
|
|
return 0;
|
|
|
|
return fn(ti, ioec->dev, ioec->start, ti->len, data);
|
|
}
|
|
|
|
static void io_err_io_hints(struct dm_target *ti, struct queue_limits *limits)
|
|
{
|
|
limits->max_hw_discard_sectors = UINT_MAX;
|
|
limits->discard_granularity = 512;
|
|
}
|
|
|
|
static long io_err_dax_direct_access(struct dm_target *ti, pgoff_t pgoff,
|
|
long nr_pages, enum dax_access_mode mode, void **kaddr,
|
|
pfn_t *pfn)
|
|
{
|
|
return -EIO;
|
|
}
|
|
|
|
static struct target_type error_target = {
|
|
.name = "error",
|
|
.version = {1, 7, 0},
|
|
.features = DM_TARGET_WILDCARD | DM_TARGET_ZONED_HM,
|
|
.ctr = io_err_ctr,
|
|
.dtr = io_err_dtr,
|
|
.map = io_err_map,
|
|
.clone_and_map_rq = io_err_clone_and_map_rq,
|
|
.release_clone_rq = io_err_release_clone_rq,
|
|
.iterate_devices = io_err_iterate_devices,
|
|
.io_hints = io_err_io_hints,
|
|
.direct_access = io_err_dax_direct_access,
|
|
.report_zones = io_err_report_zones,
|
|
};
|
|
|
|
int __init dm_target_init(void)
|
|
{
|
|
return dm_register_target(&error_target);
|
|
}
|
|
|
|
void dm_target_exit(void)
|
|
{
|
|
dm_unregister_target(&error_target);
|
|
}
|