mirror of
https://github.com/torvalds/linux.git
synced 2024-11-29 07:31:29 +00:00
c915d8f591
When inotify_freeing_mark() races with inotify_handle_inode_event() it
can happen that inotify_handle_inode_event() sees that i_mark->wd got
already reset to -1 and reports this value to userspace which can
confuse the inotify listener. Avoid the problem by validating that wd is
sensible (and pretend the mark got removed before the event got
generated otherwise).
CC: stable@vger.kernel.org
Fixes: 7e790dd5fc
("inotify: fix error paths in inotify_update_watch")
Message-Id: <20230424163219.9250-1-jack@suse.cz>
Reported-by: syzbot+4a06d4373fd52f0b2f9c@syzkaller.appspotmail.com
Reviewed-by: Amir Goldstein <amir73il@gmail.com>
Signed-off-by: Jan Kara <jack@suse.cz>
210 lines
5.8 KiB
C
210 lines
5.8 KiB
C
// SPDX-License-Identifier: GPL-2.0-or-later
|
|
/*
|
|
* fs/inotify_user.c - inotify support for userspace
|
|
*
|
|
* Authors:
|
|
* John McCutchan <ttb@tentacle.dhs.org>
|
|
* Robert Love <rml@novell.com>
|
|
*
|
|
* Copyright (C) 2005 John McCutchan
|
|
* Copyright 2006 Hewlett-Packard Development Company, L.P.
|
|
*
|
|
* Copyright (C) 2009 Eric Paris <Red Hat Inc>
|
|
* inotify was largely rewriten to make use of the fsnotify infrastructure
|
|
*/
|
|
|
|
#include <linux/dcache.h> /* d_unlinked */
|
|
#include <linux/fs.h> /* struct inode */
|
|
#include <linux/fsnotify_backend.h>
|
|
#include <linux/inotify.h>
|
|
#include <linux/path.h> /* struct path */
|
|
#include <linux/slab.h> /* kmem_* */
|
|
#include <linux/types.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/sched/user.h>
|
|
#include <linux/sched/mm.h>
|
|
|
|
#include "inotify.h"
|
|
|
|
/*
|
|
* Check if 2 events contain the same information.
|
|
*/
|
|
static bool event_compare(struct fsnotify_event *old_fsn,
|
|
struct fsnotify_event *new_fsn)
|
|
{
|
|
struct inotify_event_info *old, *new;
|
|
|
|
old = INOTIFY_E(old_fsn);
|
|
new = INOTIFY_E(new_fsn);
|
|
if (old->mask & FS_IN_IGNORED)
|
|
return false;
|
|
if ((old->mask == new->mask) &&
|
|
(old->wd == new->wd) &&
|
|
(old->name_len == new->name_len) &&
|
|
(!old->name_len || !strcmp(old->name, new->name)))
|
|
return true;
|
|
return false;
|
|
}
|
|
|
|
static int inotify_merge(struct fsnotify_group *group,
|
|
struct fsnotify_event *event)
|
|
{
|
|
struct list_head *list = &group->notification_list;
|
|
struct fsnotify_event *last_event;
|
|
|
|
last_event = list_entry(list->prev, struct fsnotify_event, list);
|
|
return event_compare(last_event, event);
|
|
}
|
|
|
|
int inotify_handle_inode_event(struct fsnotify_mark *inode_mark, u32 mask,
|
|
struct inode *inode, struct inode *dir,
|
|
const struct qstr *name, u32 cookie)
|
|
{
|
|
struct inotify_inode_mark *i_mark;
|
|
struct inotify_event_info *event;
|
|
struct fsnotify_event *fsn_event;
|
|
struct fsnotify_group *group = inode_mark->group;
|
|
int ret;
|
|
int len = 0, wd;
|
|
int alloc_len = sizeof(struct inotify_event_info);
|
|
struct mem_cgroup *old_memcg;
|
|
|
|
if (name) {
|
|
len = name->len;
|
|
alloc_len += len + 1;
|
|
}
|
|
|
|
pr_debug("%s: group=%p mark=%p mask=%x\n", __func__, group, inode_mark,
|
|
mask);
|
|
|
|
i_mark = container_of(inode_mark, struct inotify_inode_mark,
|
|
fsn_mark);
|
|
|
|
/*
|
|
* We can be racing with mark being detached. Don't report event with
|
|
* invalid wd.
|
|
*/
|
|
wd = READ_ONCE(i_mark->wd);
|
|
if (wd == -1)
|
|
return 0;
|
|
/*
|
|
* Whoever is interested in the event, pays for the allocation. Do not
|
|
* trigger OOM killer in the target monitoring memcg as it may have
|
|
* security repercussion.
|
|
*/
|
|
old_memcg = set_active_memcg(group->memcg);
|
|
event = kmalloc(alloc_len, GFP_KERNEL_ACCOUNT | __GFP_RETRY_MAYFAIL);
|
|
set_active_memcg(old_memcg);
|
|
|
|
if (unlikely(!event)) {
|
|
/*
|
|
* Treat lost event due to ENOMEM the same way as queue
|
|
* overflow to let userspace know event was lost.
|
|
*/
|
|
fsnotify_queue_overflow(group);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
/*
|
|
* We now report FS_ISDIR flag with MOVE_SELF and DELETE_SELF events
|
|
* for fanotify. inotify never reported IN_ISDIR with those events.
|
|
* It looks like an oversight, but to avoid the risk of breaking
|
|
* existing inotify programs, mask the flag out from those events.
|
|
*/
|
|
if (mask & (IN_MOVE_SELF | IN_DELETE_SELF))
|
|
mask &= ~IN_ISDIR;
|
|
|
|
fsn_event = &event->fse;
|
|
fsnotify_init_event(fsn_event);
|
|
event->mask = mask;
|
|
event->wd = wd;
|
|
event->sync_cookie = cookie;
|
|
event->name_len = len;
|
|
if (len)
|
|
strcpy(event->name, name->name);
|
|
|
|
ret = fsnotify_add_event(group, fsn_event, inotify_merge);
|
|
if (ret) {
|
|
/* Our event wasn't used in the end. Free it. */
|
|
fsnotify_destroy_event(group, fsn_event);
|
|
}
|
|
|
|
if (inode_mark->flags & FSNOTIFY_MARK_FLAG_IN_ONESHOT)
|
|
fsnotify_destroy_mark(inode_mark, group);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void inotify_freeing_mark(struct fsnotify_mark *fsn_mark, struct fsnotify_group *group)
|
|
{
|
|
inotify_ignored_and_remove_idr(fsn_mark, group);
|
|
}
|
|
|
|
/*
|
|
* This is NEVER supposed to be called. Inotify marks should either have been
|
|
* removed from the idr when the watch was removed or in the
|
|
* fsnotify_destroy_mark_by_group() call when the inotify instance was being
|
|
* torn down. This is only called if the idr is about to be freed but there
|
|
* are still marks in it.
|
|
*/
|
|
static int idr_callback(int id, void *p, void *data)
|
|
{
|
|
struct fsnotify_mark *fsn_mark;
|
|
struct inotify_inode_mark *i_mark;
|
|
static bool warned = false;
|
|
|
|
if (warned)
|
|
return 0;
|
|
|
|
warned = true;
|
|
fsn_mark = p;
|
|
i_mark = container_of(fsn_mark, struct inotify_inode_mark, fsn_mark);
|
|
|
|
WARN(1, "inotify closing but id=%d for fsn_mark=%p in group=%p still in "
|
|
"idr. Probably leaking memory\n", id, p, data);
|
|
|
|
/*
|
|
* I'm taking the liberty of assuming that the mark in question is a
|
|
* valid address and I'm dereferencing it. This might help to figure
|
|
* out why we got here and the panic is no worse than the original
|
|
* BUG() that was here.
|
|
*/
|
|
if (fsn_mark)
|
|
printk(KERN_WARNING "fsn_mark->group=%p wd=%d\n",
|
|
fsn_mark->group, i_mark->wd);
|
|
return 0;
|
|
}
|
|
|
|
static void inotify_free_group_priv(struct fsnotify_group *group)
|
|
{
|
|
/* ideally the idr is empty and we won't hit the BUG in the callback */
|
|
idr_for_each(&group->inotify_data.idr, idr_callback, group);
|
|
idr_destroy(&group->inotify_data.idr);
|
|
if (group->inotify_data.ucounts)
|
|
dec_inotify_instances(group->inotify_data.ucounts);
|
|
}
|
|
|
|
static void inotify_free_event(struct fsnotify_group *group,
|
|
struct fsnotify_event *fsn_event)
|
|
{
|
|
kfree(INOTIFY_E(fsn_event));
|
|
}
|
|
|
|
/* ding dong the mark is dead */
|
|
static void inotify_free_mark(struct fsnotify_mark *fsn_mark)
|
|
{
|
|
struct inotify_inode_mark *i_mark;
|
|
|
|
i_mark = container_of(fsn_mark, struct inotify_inode_mark, fsn_mark);
|
|
|
|
kmem_cache_free(inotify_inode_mark_cachep, i_mark);
|
|
}
|
|
|
|
const struct fsnotify_ops inotify_fsnotify_ops = {
|
|
.handle_inode_event = inotify_handle_inode_event,
|
|
.free_group_priv = inotify_free_group_priv,
|
|
.free_event = inotify_free_event,
|
|
.freeing_mark = inotify_freeing_mark,
|
|
.free_mark = inotify_free_mark,
|
|
};
|