mirror of
https://github.com/torvalds/linux.git
synced 2024-11-27 14:41:39 +00:00
2d594dfb53
The trigger flags really belong with individual btree_insert_entries, not the transaction commit flags - this splits out those flags and unifies them with the BCH_BUCKET_MARK flags. Todo - split out btree_trigger.c from buckets.c Signed-off-by: Kent Overstreet <kent.overstreet@gmail.com> Signed-off-by: Kent Overstreet <kent.overstreet@linux.dev>
783 lines
18 KiB
C
783 lines
18 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
#include "bcachefs.h"
|
|
#include "btree_update.h"
|
|
#include "inode.h"
|
|
#include "quota.h"
|
|
#include "super-io.h"
|
|
|
|
static const char *bch2_sb_validate_quota(struct bch_sb *sb,
|
|
struct bch_sb_field *f)
|
|
{
|
|
struct bch_sb_field_quota *q = field_to_type(f, quota);
|
|
|
|
if (vstruct_bytes(&q->field) != sizeof(*q))
|
|
return "invalid field quota: wrong size";
|
|
|
|
return NULL;
|
|
}
|
|
|
|
const struct bch_sb_field_ops bch_sb_field_ops_quota = {
|
|
.validate = bch2_sb_validate_quota,
|
|
};
|
|
|
|
const char *bch2_quota_invalid(const struct bch_fs *c, struct bkey_s_c k)
|
|
{
|
|
if (k.k->p.inode >= QTYP_NR)
|
|
return "invalid quota type";
|
|
|
|
if (bkey_val_bytes(k.k) != sizeof(struct bch_quota))
|
|
return "incorrect value size";
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static const char * const bch2_quota_counters[] = {
|
|
"space",
|
|
"inodes",
|
|
};
|
|
|
|
void bch2_quota_to_text(struct printbuf *out, struct bch_fs *c,
|
|
struct bkey_s_c k)
|
|
{
|
|
struct bkey_s_c_quota dq = bkey_s_c_to_quota(k);
|
|
unsigned i;
|
|
|
|
for (i = 0; i < Q_COUNTERS; i++)
|
|
pr_buf(out, "%s hardlimit %llu softlimit %llu",
|
|
bch2_quota_counters[i],
|
|
le64_to_cpu(dq.v->c[i].hardlimit),
|
|
le64_to_cpu(dq.v->c[i].softlimit));
|
|
}
|
|
|
|
#ifdef CONFIG_BCACHEFS_QUOTA
|
|
|
|
#include <linux/cred.h>
|
|
#include <linux/fs.h>
|
|
#include <linux/quota.h>
|
|
|
|
static inline unsigned __next_qtype(unsigned i, unsigned qtypes)
|
|
{
|
|
qtypes >>= i;
|
|
return qtypes ? i + __ffs(qtypes) : QTYP_NR;
|
|
}
|
|
|
|
#define for_each_set_qtype(_c, _i, _q, _qtypes) \
|
|
for (_i = 0; \
|
|
(_i = __next_qtype(_i, _qtypes), \
|
|
_q = &(_c)->quotas[_i], \
|
|
_i < QTYP_NR); \
|
|
_i++)
|
|
|
|
static bool ignore_hardlimit(struct bch_memquota_type *q)
|
|
{
|
|
if (capable(CAP_SYS_RESOURCE))
|
|
return true;
|
|
#if 0
|
|
struct mem_dqinfo *info = &sb_dqopt(dquot->dq_sb)->info[dquot->dq_id.type];
|
|
|
|
return capable(CAP_SYS_RESOURCE) &&
|
|
(info->dqi_format->qf_fmt_id != QFMT_VFS_OLD ||
|
|
!(info->dqi_flags & DQF_ROOT_SQUASH));
|
|
#endif
|
|
return false;
|
|
}
|
|
|
|
enum quota_msg {
|
|
SOFTWARN, /* Softlimit reached */
|
|
SOFTLONGWARN, /* Grace time expired */
|
|
HARDWARN, /* Hardlimit reached */
|
|
|
|
HARDBELOW, /* Usage got below inode hardlimit */
|
|
SOFTBELOW, /* Usage got below inode softlimit */
|
|
};
|
|
|
|
static int quota_nl[][Q_COUNTERS] = {
|
|
[HARDWARN][Q_SPC] = QUOTA_NL_BHARDWARN,
|
|
[SOFTLONGWARN][Q_SPC] = QUOTA_NL_BSOFTLONGWARN,
|
|
[SOFTWARN][Q_SPC] = QUOTA_NL_BSOFTWARN,
|
|
[HARDBELOW][Q_SPC] = QUOTA_NL_BHARDBELOW,
|
|
[SOFTBELOW][Q_SPC] = QUOTA_NL_BSOFTBELOW,
|
|
|
|
[HARDWARN][Q_INO] = QUOTA_NL_IHARDWARN,
|
|
[SOFTLONGWARN][Q_INO] = QUOTA_NL_ISOFTLONGWARN,
|
|
[SOFTWARN][Q_INO] = QUOTA_NL_ISOFTWARN,
|
|
[HARDBELOW][Q_INO] = QUOTA_NL_IHARDBELOW,
|
|
[SOFTBELOW][Q_INO] = QUOTA_NL_ISOFTBELOW,
|
|
};
|
|
|
|
struct quota_msgs {
|
|
u8 nr;
|
|
struct {
|
|
u8 qtype;
|
|
u8 msg;
|
|
} m[QTYP_NR * Q_COUNTERS];
|
|
};
|
|
|
|
static void prepare_msg(unsigned qtype,
|
|
enum quota_counters counter,
|
|
struct quota_msgs *msgs,
|
|
enum quota_msg msg_type)
|
|
{
|
|
BUG_ON(msgs->nr >= ARRAY_SIZE(msgs->m));
|
|
|
|
msgs->m[msgs->nr].qtype = qtype;
|
|
msgs->m[msgs->nr].msg = quota_nl[msg_type][counter];
|
|
msgs->nr++;
|
|
}
|
|
|
|
static void prepare_warning(struct memquota_counter *qc,
|
|
unsigned qtype,
|
|
enum quota_counters counter,
|
|
struct quota_msgs *msgs,
|
|
enum quota_msg msg_type)
|
|
{
|
|
if (qc->warning_issued & (1 << msg_type))
|
|
return;
|
|
|
|
prepare_msg(qtype, counter, msgs, msg_type);
|
|
}
|
|
|
|
static void flush_warnings(struct bch_qid qid,
|
|
struct super_block *sb,
|
|
struct quota_msgs *msgs)
|
|
{
|
|
unsigned i;
|
|
|
|
for (i = 0; i < msgs->nr; i++)
|
|
quota_send_warning(make_kqid(&init_user_ns, msgs->m[i].qtype, qid.q[i]),
|
|
sb->s_dev, msgs->m[i].msg);
|
|
}
|
|
|
|
static int bch2_quota_check_limit(struct bch_fs *c,
|
|
unsigned qtype,
|
|
struct bch_memquota *mq,
|
|
struct quota_msgs *msgs,
|
|
enum quota_counters counter,
|
|
s64 v,
|
|
enum quota_acct_mode mode)
|
|
{
|
|
struct bch_memquota_type *q = &c->quotas[qtype];
|
|
struct memquota_counter *qc = &mq->c[counter];
|
|
u64 n = qc->v + v;
|
|
|
|
BUG_ON((s64) n < 0);
|
|
|
|
if (mode == KEY_TYPE_QUOTA_NOCHECK)
|
|
return 0;
|
|
|
|
if (v <= 0) {
|
|
if (n < qc->hardlimit &&
|
|
(qc->warning_issued & (1 << HARDWARN))) {
|
|
qc->warning_issued &= ~(1 << HARDWARN);
|
|
prepare_msg(qtype, counter, msgs, HARDBELOW);
|
|
}
|
|
|
|
if (n < qc->softlimit &&
|
|
(qc->warning_issued & (1 << SOFTWARN))) {
|
|
qc->warning_issued &= ~(1 << SOFTWARN);
|
|
prepare_msg(qtype, counter, msgs, SOFTBELOW);
|
|
}
|
|
|
|
qc->warning_issued = 0;
|
|
return 0;
|
|
}
|
|
|
|
if (qc->hardlimit &&
|
|
qc->hardlimit < n &&
|
|
!ignore_hardlimit(q)) {
|
|
if (mode == KEY_TYPE_QUOTA_PREALLOC)
|
|
return -EDQUOT;
|
|
|
|
prepare_warning(qc, qtype, counter, msgs, HARDWARN);
|
|
}
|
|
|
|
if (qc->softlimit &&
|
|
qc->softlimit < n &&
|
|
qc->timer &&
|
|
ktime_get_real_seconds() >= qc->timer &&
|
|
!ignore_hardlimit(q)) {
|
|
if (mode == KEY_TYPE_QUOTA_PREALLOC)
|
|
return -EDQUOT;
|
|
|
|
prepare_warning(qc, qtype, counter, msgs, SOFTLONGWARN);
|
|
}
|
|
|
|
if (qc->softlimit &&
|
|
qc->softlimit < n &&
|
|
qc->timer == 0) {
|
|
if (mode == KEY_TYPE_QUOTA_PREALLOC)
|
|
return -EDQUOT;
|
|
|
|
prepare_warning(qc, qtype, counter, msgs, SOFTWARN);
|
|
|
|
/* XXX is this the right one? */
|
|
qc->timer = ktime_get_real_seconds() +
|
|
q->limits[counter].warnlimit;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int bch2_quota_acct(struct bch_fs *c, struct bch_qid qid,
|
|
enum quota_counters counter, s64 v,
|
|
enum quota_acct_mode mode)
|
|
{
|
|
unsigned qtypes = enabled_qtypes(c);
|
|
struct bch_memquota_type *q;
|
|
struct bch_memquota *mq[QTYP_NR];
|
|
struct quota_msgs msgs;
|
|
unsigned i;
|
|
int ret = 0;
|
|
|
|
memset(&msgs, 0, sizeof(msgs));
|
|
|
|
for_each_set_qtype(c, i, q, qtypes)
|
|
mutex_lock_nested(&q->lock, i);
|
|
|
|
for_each_set_qtype(c, i, q, qtypes) {
|
|
mq[i] = genradix_ptr_alloc(&q->table, qid.q[i], GFP_NOFS);
|
|
if (!mq[i]) {
|
|
ret = -ENOMEM;
|
|
goto err;
|
|
}
|
|
|
|
ret = bch2_quota_check_limit(c, i, mq[i], &msgs, counter, v, mode);
|
|
if (ret)
|
|
goto err;
|
|
}
|
|
|
|
for_each_set_qtype(c, i, q, qtypes)
|
|
mq[i]->c[counter].v += v;
|
|
err:
|
|
for_each_set_qtype(c, i, q, qtypes)
|
|
mutex_unlock(&q->lock);
|
|
|
|
flush_warnings(qid, c->vfs_sb, &msgs);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void __bch2_quota_transfer(struct bch_memquota *src_q,
|
|
struct bch_memquota *dst_q,
|
|
enum quota_counters counter, s64 v)
|
|
{
|
|
BUG_ON(v > src_q->c[counter].v);
|
|
BUG_ON(v + dst_q->c[counter].v < v);
|
|
|
|
src_q->c[counter].v -= v;
|
|
dst_q->c[counter].v += v;
|
|
}
|
|
|
|
int bch2_quota_transfer(struct bch_fs *c, unsigned qtypes,
|
|
struct bch_qid dst,
|
|
struct bch_qid src, u64 space,
|
|
enum quota_acct_mode mode)
|
|
{
|
|
struct bch_memquota_type *q;
|
|
struct bch_memquota *src_q[3], *dst_q[3];
|
|
struct quota_msgs msgs;
|
|
unsigned i;
|
|
int ret = 0;
|
|
|
|
qtypes &= enabled_qtypes(c);
|
|
|
|
memset(&msgs, 0, sizeof(msgs));
|
|
|
|
for_each_set_qtype(c, i, q, qtypes)
|
|
mutex_lock_nested(&q->lock, i);
|
|
|
|
for_each_set_qtype(c, i, q, qtypes) {
|
|
src_q[i] = genradix_ptr_alloc(&q->table, src.q[i], GFP_NOFS);
|
|
dst_q[i] = genradix_ptr_alloc(&q->table, dst.q[i], GFP_NOFS);
|
|
|
|
if (!src_q[i] || !dst_q[i]) {
|
|
ret = -ENOMEM;
|
|
goto err;
|
|
}
|
|
|
|
ret = bch2_quota_check_limit(c, i, dst_q[i], &msgs, Q_SPC,
|
|
dst_q[i]->c[Q_SPC].v + space,
|
|
mode);
|
|
if (ret)
|
|
goto err;
|
|
|
|
ret = bch2_quota_check_limit(c, i, dst_q[i], &msgs, Q_INO,
|
|
dst_q[i]->c[Q_INO].v + 1,
|
|
mode);
|
|
if (ret)
|
|
goto err;
|
|
}
|
|
|
|
for_each_set_qtype(c, i, q, qtypes) {
|
|
__bch2_quota_transfer(src_q[i], dst_q[i], Q_SPC, space);
|
|
__bch2_quota_transfer(src_q[i], dst_q[i], Q_INO, 1);
|
|
}
|
|
|
|
err:
|
|
for_each_set_qtype(c, i, q, qtypes)
|
|
mutex_unlock(&q->lock);
|
|
|
|
flush_warnings(dst, c->vfs_sb, &msgs);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int __bch2_quota_set(struct bch_fs *c, struct bkey_s_c k)
|
|
{
|
|
struct bkey_s_c_quota dq;
|
|
struct bch_memquota_type *q;
|
|
struct bch_memquota *mq;
|
|
unsigned i;
|
|
|
|
BUG_ON(k.k->p.inode >= QTYP_NR);
|
|
|
|
switch (k.k->type) {
|
|
case KEY_TYPE_quota:
|
|
dq = bkey_s_c_to_quota(k);
|
|
q = &c->quotas[k.k->p.inode];
|
|
|
|
mutex_lock(&q->lock);
|
|
mq = genradix_ptr_alloc(&q->table, k.k->p.offset, GFP_KERNEL);
|
|
if (!mq) {
|
|
mutex_unlock(&q->lock);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
for (i = 0; i < Q_COUNTERS; i++) {
|
|
mq->c[i].hardlimit = le64_to_cpu(dq.v->c[i].hardlimit);
|
|
mq->c[i].softlimit = le64_to_cpu(dq.v->c[i].softlimit);
|
|
}
|
|
|
|
mutex_unlock(&q->lock);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int bch2_quota_init_type(struct bch_fs *c, enum quota_types type)
|
|
{
|
|
struct btree_trans trans;
|
|
struct btree_iter *iter;
|
|
struct bkey_s_c k;
|
|
int ret = 0;
|
|
|
|
bch2_trans_init(&trans, c, 0, 0);
|
|
|
|
for_each_btree_key(&trans, iter, BTREE_ID_QUOTAS, POS(type, 0),
|
|
BTREE_ITER_PREFETCH, k, ret) {
|
|
if (k.k->p.inode != type)
|
|
break;
|
|
|
|
ret = __bch2_quota_set(c, k);
|
|
if (ret)
|
|
break;
|
|
}
|
|
|
|
return bch2_trans_exit(&trans) ?: ret;
|
|
}
|
|
|
|
void bch2_fs_quota_exit(struct bch_fs *c)
|
|
{
|
|
unsigned i;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(c->quotas); i++)
|
|
genradix_free(&c->quotas[i].table);
|
|
}
|
|
|
|
void bch2_fs_quota_init(struct bch_fs *c)
|
|
{
|
|
unsigned i;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(c->quotas); i++)
|
|
mutex_init(&c->quotas[i].lock);
|
|
}
|
|
|
|
static void bch2_sb_quota_read(struct bch_fs *c)
|
|
{
|
|
struct bch_sb_field_quota *sb_quota;
|
|
unsigned i, j;
|
|
|
|
sb_quota = bch2_sb_get_quota(c->disk_sb.sb);
|
|
if (!sb_quota)
|
|
return;
|
|
|
|
for (i = 0; i < QTYP_NR; i++) {
|
|
struct bch_memquota_type *q = &c->quotas[i];
|
|
|
|
for (j = 0; j < Q_COUNTERS; j++) {
|
|
q->limits[j].timelimit =
|
|
le32_to_cpu(sb_quota->q[i].c[j].timelimit);
|
|
q->limits[j].warnlimit =
|
|
le32_to_cpu(sb_quota->q[i].c[j].warnlimit);
|
|
}
|
|
}
|
|
}
|
|
|
|
int bch2_fs_quota_read(struct bch_fs *c)
|
|
{
|
|
unsigned i, qtypes = enabled_qtypes(c);
|
|
struct bch_memquota_type *q;
|
|
struct btree_trans trans;
|
|
struct btree_iter *iter;
|
|
struct bch_inode_unpacked u;
|
|
struct bkey_s_c k;
|
|
int ret;
|
|
|
|
mutex_lock(&c->sb_lock);
|
|
bch2_sb_quota_read(c);
|
|
mutex_unlock(&c->sb_lock);
|
|
|
|
for_each_set_qtype(c, i, q, qtypes) {
|
|
ret = bch2_quota_init_type(c, i);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
bch2_trans_init(&trans, c, 0, 0);
|
|
|
|
for_each_btree_key(&trans, iter, BTREE_ID_INODES, POS_MIN,
|
|
BTREE_ITER_PREFETCH, k, ret) {
|
|
switch (k.k->type) {
|
|
case KEY_TYPE_inode:
|
|
ret = bch2_inode_unpack(bkey_s_c_to_inode(k), &u);
|
|
if (ret)
|
|
return ret;
|
|
|
|
bch2_quota_acct(c, bch_qid(&u), Q_SPC, u.bi_sectors,
|
|
KEY_TYPE_QUOTA_NOCHECK);
|
|
bch2_quota_acct(c, bch_qid(&u), Q_INO, 1,
|
|
KEY_TYPE_QUOTA_NOCHECK);
|
|
}
|
|
}
|
|
return bch2_trans_exit(&trans) ?: ret;
|
|
}
|
|
|
|
/* Enable/disable/delete quotas for an entire filesystem: */
|
|
|
|
static int bch2_quota_enable(struct super_block *sb, unsigned uflags)
|
|
{
|
|
struct bch_fs *c = sb->s_fs_info;
|
|
|
|
if (sb->s_flags & SB_RDONLY)
|
|
return -EROFS;
|
|
|
|
/* Accounting must be enabled at mount time: */
|
|
if (uflags & (FS_QUOTA_UDQ_ACCT|FS_QUOTA_GDQ_ACCT|FS_QUOTA_PDQ_ACCT))
|
|
return -EINVAL;
|
|
|
|
/* Can't enable enforcement without accounting: */
|
|
if ((uflags & FS_QUOTA_UDQ_ENFD) && !c->opts.usrquota)
|
|
return -EINVAL;
|
|
|
|
if ((uflags & FS_QUOTA_GDQ_ENFD) && !c->opts.grpquota)
|
|
return -EINVAL;
|
|
|
|
if (uflags & FS_QUOTA_PDQ_ENFD && !c->opts.prjquota)
|
|
return -EINVAL;
|
|
|
|
mutex_lock(&c->sb_lock);
|
|
if (uflags & FS_QUOTA_UDQ_ENFD)
|
|
SET_BCH_SB_USRQUOTA(c->disk_sb.sb, true);
|
|
|
|
if (uflags & FS_QUOTA_GDQ_ENFD)
|
|
SET_BCH_SB_GRPQUOTA(c->disk_sb.sb, true);
|
|
|
|
if (uflags & FS_QUOTA_PDQ_ENFD)
|
|
SET_BCH_SB_PRJQUOTA(c->disk_sb.sb, true);
|
|
|
|
bch2_write_super(c);
|
|
mutex_unlock(&c->sb_lock);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int bch2_quota_disable(struct super_block *sb, unsigned uflags)
|
|
{
|
|
struct bch_fs *c = sb->s_fs_info;
|
|
|
|
if (sb->s_flags & SB_RDONLY)
|
|
return -EROFS;
|
|
|
|
mutex_lock(&c->sb_lock);
|
|
if (uflags & FS_QUOTA_UDQ_ENFD)
|
|
SET_BCH_SB_USRQUOTA(c->disk_sb.sb, false);
|
|
|
|
if (uflags & FS_QUOTA_GDQ_ENFD)
|
|
SET_BCH_SB_GRPQUOTA(c->disk_sb.sb, false);
|
|
|
|
if (uflags & FS_QUOTA_PDQ_ENFD)
|
|
SET_BCH_SB_PRJQUOTA(c->disk_sb.sb, false);
|
|
|
|
bch2_write_super(c);
|
|
mutex_unlock(&c->sb_lock);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int bch2_quota_remove(struct super_block *sb, unsigned uflags)
|
|
{
|
|
struct bch_fs *c = sb->s_fs_info;
|
|
int ret;
|
|
|
|
if (sb->s_flags & SB_RDONLY)
|
|
return -EROFS;
|
|
|
|
if (uflags & FS_USER_QUOTA) {
|
|
if (c->opts.usrquota)
|
|
return -EINVAL;
|
|
|
|
ret = bch2_btree_delete_range(c, BTREE_ID_QUOTAS,
|
|
POS(QTYP_USR, 0),
|
|
POS(QTYP_USR + 1, 0),
|
|
NULL);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
if (uflags & FS_GROUP_QUOTA) {
|
|
if (c->opts.grpquota)
|
|
return -EINVAL;
|
|
|
|
ret = bch2_btree_delete_range(c, BTREE_ID_QUOTAS,
|
|
POS(QTYP_GRP, 0),
|
|
POS(QTYP_GRP + 1, 0),
|
|
NULL);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
if (uflags & FS_PROJ_QUOTA) {
|
|
if (c->opts.prjquota)
|
|
return -EINVAL;
|
|
|
|
ret = bch2_btree_delete_range(c, BTREE_ID_QUOTAS,
|
|
POS(QTYP_PRJ, 0),
|
|
POS(QTYP_PRJ + 1, 0),
|
|
NULL);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Return quota status information, such as enforcements, quota file inode
|
|
* numbers etc.
|
|
*/
|
|
static int bch2_quota_get_state(struct super_block *sb, struct qc_state *state)
|
|
{
|
|
struct bch_fs *c = sb->s_fs_info;
|
|
unsigned qtypes = enabled_qtypes(c);
|
|
unsigned i;
|
|
|
|
memset(state, 0, sizeof(*state));
|
|
|
|
for (i = 0; i < QTYP_NR; i++) {
|
|
state->s_state[i].flags |= QCI_SYSFILE;
|
|
|
|
if (!(qtypes & (1 << i)))
|
|
continue;
|
|
|
|
state->s_state[i].flags |= QCI_ACCT_ENABLED;
|
|
|
|
state->s_state[i].spc_timelimit = c->quotas[i].limits[Q_SPC].timelimit;
|
|
state->s_state[i].spc_warnlimit = c->quotas[i].limits[Q_SPC].warnlimit;
|
|
|
|
state->s_state[i].ino_timelimit = c->quotas[i].limits[Q_INO].timelimit;
|
|
state->s_state[i].ino_warnlimit = c->quotas[i].limits[Q_INO].warnlimit;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* Adjust quota timers & warnings
|
|
*/
|
|
static int bch2_quota_set_info(struct super_block *sb, int type,
|
|
struct qc_info *info)
|
|
{
|
|
struct bch_fs *c = sb->s_fs_info;
|
|
struct bch_sb_field_quota *sb_quota;
|
|
struct bch_memquota_type *q;
|
|
|
|
if (sb->s_flags & SB_RDONLY)
|
|
return -EROFS;
|
|
|
|
if (type >= QTYP_NR)
|
|
return -EINVAL;
|
|
|
|
if (!((1 << type) & enabled_qtypes(c)))
|
|
return -ESRCH;
|
|
|
|
if (info->i_fieldmask &
|
|
~(QC_SPC_TIMER|QC_INO_TIMER|QC_SPC_WARNS|QC_INO_WARNS))
|
|
return -EINVAL;
|
|
|
|
q = &c->quotas[type];
|
|
|
|
mutex_lock(&c->sb_lock);
|
|
sb_quota = bch2_sb_get_quota(c->disk_sb.sb);
|
|
if (!sb_quota) {
|
|
sb_quota = bch2_sb_resize_quota(&c->disk_sb,
|
|
sizeof(*sb_quota) / sizeof(u64));
|
|
if (!sb_quota)
|
|
return -ENOSPC;
|
|
}
|
|
|
|
if (info->i_fieldmask & QC_SPC_TIMER)
|
|
sb_quota->q[type].c[Q_SPC].timelimit =
|
|
cpu_to_le32(info->i_spc_timelimit);
|
|
|
|
if (info->i_fieldmask & QC_SPC_WARNS)
|
|
sb_quota->q[type].c[Q_SPC].warnlimit =
|
|
cpu_to_le32(info->i_spc_warnlimit);
|
|
|
|
if (info->i_fieldmask & QC_INO_TIMER)
|
|
sb_quota->q[type].c[Q_INO].timelimit =
|
|
cpu_to_le32(info->i_ino_timelimit);
|
|
|
|
if (info->i_fieldmask & QC_INO_WARNS)
|
|
sb_quota->q[type].c[Q_INO].warnlimit =
|
|
cpu_to_le32(info->i_ino_warnlimit);
|
|
|
|
bch2_sb_quota_read(c);
|
|
|
|
bch2_write_super(c);
|
|
mutex_unlock(&c->sb_lock);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* Get/set individual quotas: */
|
|
|
|
static void __bch2_quota_get(struct qc_dqblk *dst, struct bch_memquota *src)
|
|
{
|
|
dst->d_space = src->c[Q_SPC].v << 9;
|
|
dst->d_spc_hardlimit = src->c[Q_SPC].hardlimit << 9;
|
|
dst->d_spc_softlimit = src->c[Q_SPC].softlimit << 9;
|
|
dst->d_spc_timer = src->c[Q_SPC].timer;
|
|
dst->d_spc_warns = src->c[Q_SPC].warns;
|
|
|
|
dst->d_ino_count = src->c[Q_INO].v;
|
|
dst->d_ino_hardlimit = src->c[Q_INO].hardlimit;
|
|
dst->d_ino_softlimit = src->c[Q_INO].softlimit;
|
|
dst->d_ino_timer = src->c[Q_INO].timer;
|
|
dst->d_ino_warns = src->c[Q_INO].warns;
|
|
}
|
|
|
|
static int bch2_get_quota(struct super_block *sb, struct kqid kqid,
|
|
struct qc_dqblk *qdq)
|
|
{
|
|
struct bch_fs *c = sb->s_fs_info;
|
|
struct bch_memquota_type *q = &c->quotas[kqid.type];
|
|
qid_t qid = from_kqid(&init_user_ns, kqid);
|
|
struct bch_memquota *mq;
|
|
|
|
memset(qdq, 0, sizeof(*qdq));
|
|
|
|
mutex_lock(&q->lock);
|
|
mq = genradix_ptr(&q->table, qid);
|
|
if (mq)
|
|
__bch2_quota_get(qdq, mq);
|
|
mutex_unlock(&q->lock);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int bch2_get_next_quota(struct super_block *sb, struct kqid *kqid,
|
|
struct qc_dqblk *qdq)
|
|
{
|
|
struct bch_fs *c = sb->s_fs_info;
|
|
struct bch_memquota_type *q = &c->quotas[kqid->type];
|
|
qid_t qid = from_kqid(&init_user_ns, *kqid);
|
|
struct genradix_iter iter;
|
|
struct bch_memquota *mq;
|
|
int ret = 0;
|
|
|
|
mutex_lock(&q->lock);
|
|
|
|
genradix_for_each_from(&q->table, iter, mq, qid)
|
|
if (memcmp(mq, page_address(ZERO_PAGE(0)), sizeof(*mq))) {
|
|
__bch2_quota_get(qdq, mq);
|
|
*kqid = make_kqid(current_user_ns(), kqid->type, iter.pos);
|
|
goto found;
|
|
}
|
|
|
|
ret = -ENOENT;
|
|
found:
|
|
mutex_unlock(&q->lock);
|
|
return ret;
|
|
}
|
|
|
|
static int bch2_set_quota(struct super_block *sb, struct kqid qid,
|
|
struct qc_dqblk *qdq)
|
|
{
|
|
struct bch_fs *c = sb->s_fs_info;
|
|
struct btree_trans trans;
|
|
struct btree_iter *iter;
|
|
struct bkey_s_c k;
|
|
struct bkey_i_quota new_quota;
|
|
int ret;
|
|
|
|
if (sb->s_flags & SB_RDONLY)
|
|
return -EROFS;
|
|
|
|
bkey_quota_init(&new_quota.k_i);
|
|
new_quota.k.p = POS(qid.type, from_kqid(&init_user_ns, qid));
|
|
|
|
bch2_trans_init(&trans, c, 0, 0);
|
|
|
|
iter = bch2_trans_get_iter(&trans, BTREE_ID_QUOTAS, new_quota.k.p,
|
|
BTREE_ITER_SLOTS|BTREE_ITER_INTENT);
|
|
k = bch2_btree_iter_peek_slot(iter);
|
|
|
|
ret = bkey_err(k);
|
|
if (unlikely(ret))
|
|
return ret;
|
|
|
|
switch (k.k->type) {
|
|
case KEY_TYPE_quota:
|
|
new_quota.v = *bkey_s_c_to_quota(k).v;
|
|
break;
|
|
}
|
|
|
|
if (qdq->d_fieldmask & QC_SPC_SOFT)
|
|
new_quota.v.c[Q_SPC].softlimit = cpu_to_le64(qdq->d_spc_softlimit >> 9);
|
|
if (qdq->d_fieldmask & QC_SPC_HARD)
|
|
new_quota.v.c[Q_SPC].hardlimit = cpu_to_le64(qdq->d_spc_hardlimit >> 9);
|
|
|
|
if (qdq->d_fieldmask & QC_INO_SOFT)
|
|
new_quota.v.c[Q_INO].softlimit = cpu_to_le64(qdq->d_ino_softlimit);
|
|
if (qdq->d_fieldmask & QC_INO_HARD)
|
|
new_quota.v.c[Q_INO].hardlimit = cpu_to_le64(qdq->d_ino_hardlimit);
|
|
|
|
bch2_trans_update(&trans, iter, &new_quota.k_i, 0);
|
|
|
|
ret = bch2_trans_commit(&trans, NULL, NULL, 0);
|
|
|
|
bch2_trans_exit(&trans);
|
|
|
|
if (ret)
|
|
return ret;
|
|
|
|
ret = __bch2_quota_set(c, bkey_i_to_s_c(&new_quota.k_i));
|
|
|
|
return ret;
|
|
}
|
|
|
|
const struct quotactl_ops bch2_quotactl_operations = {
|
|
.quota_enable = bch2_quota_enable,
|
|
.quota_disable = bch2_quota_disable,
|
|
.rm_xquota = bch2_quota_remove,
|
|
|
|
.get_state = bch2_quota_get_state,
|
|
.set_info = bch2_quota_set_info,
|
|
|
|
.get_dqblk = bch2_get_quota,
|
|
.get_nextdqblk = bch2_get_next_quota,
|
|
.set_dqblk = bch2_set_quota,
|
|
};
|
|
|
|
#endif /* CONFIG_BCACHEFS_QUOTA */
|