mirror of
https://github.com/torvalds/linux.git
synced 2024-11-22 12:11:40 +00:00
Remove SLAB_CTOR_CONSTRUCTOR
SLAB_CTOR_CONSTRUCTOR is always specified. No point in checking it. Signed-off-by: Christoph Lameter <clameter@sgi.com> Cc: David Howells <dhowells@redhat.com> Cc: Jens Axboe <jens.axboe@oracle.com> Cc: Steven French <sfrench@us.ibm.com> Cc: Michael Halcrow <mhalcrow@us.ibm.com> Cc: OGAWA Hirofumi <hirofumi@mail.parknet.co.jp> Cc: Miklos Szeredi <miklos@szeredi.hu> Cc: Steven Whitehouse <swhiteho@redhat.com> Cc: Roman Zippel <zippel@linux-m68k.org> Cc: David Woodhouse <dwmw2@infradead.org> Cc: Dave Kleikamp <shaggy@austin.ibm.com> Cc: Trond Myklebust <trond.myklebust@fys.uio.no> Cc: "J. Bruce Fields" <bfields@fieldses.org> Cc: Anton Altaparmakov <aia21@cantab.net> Cc: Mark Fasheh <mark.fasheh@oracle.com> Cc: Paul Mackerras <paulus@samba.org> Cc: Christoph Hellwig <hch@lst.de> Cc: Jan Kara <jack@ucw.cz> Cc: David Chinner <dgc@sgi.com> Cc: "David S. Miller" <davem@davemloft.net> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
parent
5577bd8a85
commit
a35afb830f
@ -71,10 +71,8 @@ spufs_init_once(void *p, struct kmem_cache * cachep, unsigned long flags)
|
||||
{
|
||||
struct spufs_inode_info *ei = p;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
}
|
||||
}
|
||||
|
||||
static struct inode *
|
||||
spufs_new_inode(struct super_block *sb, int mode)
|
||||
|
@ -940,9 +940,6 @@ static void ltree_entry_ctor(void *obj, struct kmem_cache *cache,
|
||||
{
|
||||
struct ltree_entry *le = obj;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR)
|
||||
return;
|
||||
|
||||
le->users = 0;
|
||||
init_rwsem(&le->mutex);
|
||||
}
|
||||
|
@ -232,7 +232,6 @@ static void init_once(void * foo, struct kmem_cache * cachep, unsigned long flag
|
||||
{
|
||||
struct adfs_inode_info *ei = (struct adfs_inode_info *) foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR)
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
}
|
||||
|
||||
|
@ -87,12 +87,10 @@ static void init_once(void * foo, struct kmem_cache * cachep, unsigned long flag
|
||||
{
|
||||
struct affs_inode_info *ei = (struct affs_inode_info *) foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
init_MUTEX(&ei->i_link_lock);
|
||||
init_MUTEX(&ei->i_ext_lock);
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
}
|
||||
}
|
||||
|
||||
static int init_inodecache(void)
|
||||
{
|
||||
|
@ -451,7 +451,6 @@ static void afs_i_init_once(void *_vnode, struct kmem_cache *cachep,
|
||||
{
|
||||
struct afs_vnode *vnode = _vnode;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
memset(vnode, 0, sizeof(*vnode));
|
||||
inode_init_once(&vnode->vfs_inode);
|
||||
init_waitqueue_head(&vnode->update_waitq);
|
||||
@ -462,7 +461,6 @@ static void afs_i_init_once(void *_vnode, struct kmem_cache *cachep,
|
||||
INIT_LIST_HEAD(&vnode->writebacks);
|
||||
INIT_WORK(&vnode->cb_broken_work, afs_broken_callback_work);
|
||||
}
|
||||
}
|
||||
|
||||
/*
|
||||
* allocate an AFS inode struct from our slab cache
|
||||
|
@ -293,10 +293,8 @@ static void init_once(void * foo, struct kmem_cache * cachep, unsigned long flag
|
||||
{
|
||||
struct befs_inode_info *bi = (struct befs_inode_info *) foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
inode_init_once(&bi->vfs_inode);
|
||||
}
|
||||
}
|
||||
|
||||
static void
|
||||
befs_read_inode(struct inode *inode)
|
||||
|
@ -248,7 +248,6 @@ static void init_once(void * foo, struct kmem_cache * cachep, unsigned long flag
|
||||
{
|
||||
struct bfs_inode_info *bi = foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR)
|
||||
inode_init_once(&bi->vfs_inode);
|
||||
}
|
||||
|
||||
|
@ -458,7 +458,6 @@ static void init_once(void * foo, struct kmem_cache * cachep, unsigned long flag
|
||||
struct bdev_inode *ei = (struct bdev_inode *) foo;
|
||||
struct block_device *bdev = &ei->bdev;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
memset(bdev, 0, sizeof(*bdev));
|
||||
mutex_init(&bdev->bd_mutex);
|
||||
sema_init(&bdev->bd_mount_sem, 1);
|
||||
@ -469,7 +468,6 @@ static void init_once(void * foo, struct kmem_cache * cachep, unsigned long flag
|
||||
#endif
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
}
|
||||
}
|
||||
|
||||
static inline void __bd_forget(struct inode *inode)
|
||||
{
|
||||
|
22
fs/buffer.c
22
fs/buffer.c
@ -2898,8 +2898,9 @@ static void recalc_bh_state(void)
|
||||
|
||||
struct buffer_head *alloc_buffer_head(gfp_t gfp_flags)
|
||||
{
|
||||
struct buffer_head *ret = kmem_cache_alloc(bh_cachep, gfp_flags);
|
||||
struct buffer_head *ret = kmem_cache_zalloc(bh_cachep, gfp_flags);
|
||||
if (ret) {
|
||||
INIT_LIST_HEAD(&ret->b_assoc_buffers);
|
||||
get_cpu_var(bh_accounting).nr++;
|
||||
recalc_bh_state();
|
||||
put_cpu_var(bh_accounting);
|
||||
@ -2918,17 +2919,6 @@ void free_buffer_head(struct buffer_head *bh)
|
||||
}
|
||||
EXPORT_SYMBOL(free_buffer_head);
|
||||
|
||||
static void
|
||||
init_buffer_head(void *data, struct kmem_cache *cachep, unsigned long flags)
|
||||
{
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
struct buffer_head * bh = (struct buffer_head *)data;
|
||||
|
||||
memset(bh, 0, sizeof(*bh));
|
||||
INIT_LIST_HEAD(&bh->b_assoc_buffers);
|
||||
}
|
||||
}
|
||||
|
||||
static void buffer_exit_cpu(int cpu)
|
||||
{
|
||||
int i;
|
||||
@ -2955,12 +2945,8 @@ void __init buffer_init(void)
|
||||
{
|
||||
int nrpages;
|
||||
|
||||
bh_cachep = kmem_cache_create("buffer_head",
|
||||
sizeof(struct buffer_head), 0,
|
||||
(SLAB_RECLAIM_ACCOUNT|SLAB_PANIC|
|
||||
SLAB_MEM_SPREAD),
|
||||
init_buffer_head,
|
||||
NULL);
|
||||
bh_cachep = KMEM_CACHE(buffer_head,
|
||||
SLAB_RECLAIM_ACCOUNT|SLAB_PANIC|SLAB_MEM_SPREAD);
|
||||
|
||||
/*
|
||||
* Limit the bh occupancy to 10% of ZONE_NORMAL
|
||||
|
@ -701,11 +701,9 @@ cifs_init_once(void *inode, struct kmem_cache * cachep, unsigned long flags)
|
||||
{
|
||||
struct cifsInodeInfo *cifsi = inode;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
inode_init_once(&cifsi->vfs_inode);
|
||||
INIT_LIST_HEAD(&cifsi->lockList);
|
||||
}
|
||||
}
|
||||
|
||||
static int
|
||||
cifs_init_inodecache(void)
|
||||
|
@ -62,7 +62,6 @@ static void init_once(void * foo, struct kmem_cache * cachep, unsigned long flag
|
||||
{
|
||||
struct coda_inode_info *ei = (struct coda_inode_info *) foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR)
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
}
|
||||
|
||||
|
@ -583,7 +583,6 @@ inode_info_init_once(void *vptr, struct kmem_cache *cachep, unsigned long flags)
|
||||
{
|
||||
struct ecryptfs_inode_info *ei = (struct ecryptfs_inode_info *)vptr;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR)
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
}
|
||||
|
||||
|
@ -72,7 +72,6 @@ static void init_once(void * foo, struct kmem_cache * cachep, unsigned long flag
|
||||
{
|
||||
struct efs_inode_info *ei = (struct efs_inode_info *) foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR)
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
}
|
||||
|
||||
|
@ -160,14 +160,12 @@ static void init_once(void * foo, struct kmem_cache * cachep, unsigned long flag
|
||||
{
|
||||
struct ext2_inode_info *ei = (struct ext2_inode_info *) foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
rwlock_init(&ei->i_meta_lock);
|
||||
#ifdef CONFIG_EXT2_FS_XATTR
|
||||
init_rwsem(&ei->xattr_sem);
|
||||
#endif
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
}
|
||||
}
|
||||
|
||||
static int init_inodecache(void)
|
||||
{
|
||||
|
@ -466,7 +466,6 @@ static void init_once(void * foo, struct kmem_cache * cachep, unsigned long flag
|
||||
{
|
||||
struct ext3_inode_info *ei = (struct ext3_inode_info *) foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
INIT_LIST_HEAD(&ei->i_orphan);
|
||||
#ifdef CONFIG_EXT3_FS_XATTR
|
||||
init_rwsem(&ei->xattr_sem);
|
||||
@ -474,7 +473,6 @@ static void init_once(void * foo, struct kmem_cache * cachep, unsigned long flag
|
||||
mutex_init(&ei->truncate_mutex);
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
}
|
||||
}
|
||||
|
||||
static int init_inodecache(void)
|
||||
{
|
||||
|
@ -517,7 +517,6 @@ static void init_once(void * foo, struct kmem_cache * cachep, unsigned long flag
|
||||
{
|
||||
struct ext4_inode_info *ei = (struct ext4_inode_info *) foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
INIT_LIST_HEAD(&ei->i_orphan);
|
||||
#ifdef CONFIG_EXT4DEV_FS_XATTR
|
||||
init_rwsem(&ei->xattr_sem);
|
||||
@ -525,7 +524,6 @@ static void init_once(void * foo, struct kmem_cache * cachep, unsigned long flag
|
||||
mutex_init(&ei->truncate_mutex);
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
}
|
||||
}
|
||||
|
||||
static int init_inodecache(void)
|
||||
{
|
||||
|
@ -40,7 +40,6 @@ static void init_once(void *foo, struct kmem_cache *cachep, unsigned long flags)
|
||||
{
|
||||
struct fat_cache *cache = (struct fat_cache *)foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR)
|
||||
INIT_LIST_HEAD(&cache->cache_list);
|
||||
}
|
||||
|
||||
|
@ -500,7 +500,6 @@ static void init_once(void * foo, struct kmem_cache * cachep, unsigned long flag
|
||||
{
|
||||
struct msdos_inode_info *ei = (struct msdos_inode_info *)foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
spin_lock_init(&ei->cache_lru_lock);
|
||||
ei->nr_caches = 0;
|
||||
ei->cache_valid_id = FAT_CACHE_VALID + 1;
|
||||
@ -508,7 +507,6 @@ static void init_once(void * foo, struct kmem_cache * cachep, unsigned long flag
|
||||
INIT_HLIST_NODE(&ei->i_fat_hash);
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
}
|
||||
}
|
||||
|
||||
static int __init fat_init_inodecache(void)
|
||||
{
|
||||
|
@ -687,7 +687,6 @@ static void fuse_inode_init_once(void *foo, struct kmem_cache *cachep,
|
||||
{
|
||||
struct inode * inode = foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR)
|
||||
inode_init_once(inode);
|
||||
}
|
||||
|
||||
|
@ -27,18 +27,17 @@
|
||||
static void gfs2_init_inode_once(void *foo, struct kmem_cache *cachep, unsigned long flags)
|
||||
{
|
||||
struct gfs2_inode *ip = foo;
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
|
||||
inode_init_once(&ip->i_inode);
|
||||
spin_lock_init(&ip->i_spin);
|
||||
init_rwsem(&ip->i_rw_mutex);
|
||||
memset(ip->i_cache, 0, sizeof(ip->i_cache));
|
||||
}
|
||||
}
|
||||
|
||||
static void gfs2_init_glock_once(void *foo, struct kmem_cache *cachep, unsigned long flags)
|
||||
{
|
||||
struct gfs2_glock *gl = foo;
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
|
||||
INIT_HLIST_NODE(&gl->gl_list);
|
||||
spin_lock_init(&gl->gl_spin);
|
||||
INIT_LIST_HEAD(&gl->gl_holders);
|
||||
@ -50,7 +49,6 @@ static void gfs2_init_glock_once(void *foo, struct kmem_cache *cachep, unsigned
|
||||
INIT_LIST_HEAD(&gl->gl_ail_list);
|
||||
atomic_set(&gl->gl_ail_count, 0);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* init_gfs2_fs - Register GFS2 as a filesystem
|
||||
|
@ -434,7 +434,6 @@ static void hfs_init_once(void *p, struct kmem_cache *cachep, unsigned long flag
|
||||
{
|
||||
struct hfs_inode_info *i = p;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR)
|
||||
inode_init_once(&i->vfs_inode);
|
||||
}
|
||||
|
||||
|
@ -470,7 +470,6 @@ static void hfsplus_init_once(void *p, struct kmem_cache *cachep, unsigned long
|
||||
{
|
||||
struct hfsplus_inode_info *i = p;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR)
|
||||
inode_init_once(&i->vfs_inode);
|
||||
}
|
||||
|
||||
|
@ -176,12 +176,10 @@ static void init_once(void * foo, struct kmem_cache * cachep, unsigned long flag
|
||||
{
|
||||
struct hpfs_inode_info *ei = (struct hpfs_inode_info *) foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
mutex_init(&ei->i_mutex);
|
||||
mutex_init(&ei->i_parent_mutex);
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
}
|
||||
}
|
||||
|
||||
static int init_inodecache(void)
|
||||
{
|
||||
|
@ -556,7 +556,6 @@ static void init_once(void *foo, struct kmem_cache *cachep, unsigned long flags)
|
||||
{
|
||||
struct hugetlbfs_inode_info *ei = (struct hugetlbfs_inode_info *)foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR)
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
}
|
||||
|
||||
|
@ -213,7 +213,6 @@ static void init_once(void * foo, struct kmem_cache * cachep, unsigned long flag
|
||||
{
|
||||
struct inode * inode = (struct inode *) foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR)
|
||||
inode_init_once(inode);
|
||||
}
|
||||
|
||||
|
@ -77,7 +77,6 @@ static void init_once(void *foo, struct kmem_cache * cachep, unsigned long flags
|
||||
{
|
||||
struct iso_inode_info *ei = foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR)
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
}
|
||||
|
||||
|
@ -47,11 +47,9 @@ static void jffs2_i_init_once(void * foo, struct kmem_cache * cachep, unsigned l
|
||||
{
|
||||
struct jffs2_inode_info *ei = (struct jffs2_inode_info *) foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
init_MUTEX(&ei->sem);
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
}
|
||||
}
|
||||
|
||||
static int jffs2_sync_fs(struct super_block *sb, int wait)
|
||||
{
|
||||
|
@ -184,7 +184,6 @@ static void init_once(void *foo, struct kmem_cache *cachep, unsigned long flags)
|
||||
{
|
||||
struct metapage *mp = (struct metapage *)foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
mp->lid = 0;
|
||||
mp->lsn = 0;
|
||||
mp->flag = 0;
|
||||
@ -194,7 +193,6 @@ static void init_once(void *foo, struct kmem_cache *cachep, unsigned long flags)
|
||||
set_bit(META_free, &mp->flag);
|
||||
init_waitqueue_head(&mp->wait);
|
||||
}
|
||||
}
|
||||
|
||||
static inline struct metapage *alloc_metapage(gfp_t gfp_mask)
|
||||
{
|
||||
|
@ -752,7 +752,6 @@ static void init_once(void *foo, struct kmem_cache * cachep, unsigned long flags
|
||||
{
|
||||
struct jfs_inode_info *jfs_ip = (struct jfs_inode_info *) foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
memset(jfs_ip, 0, sizeof(struct jfs_inode_info));
|
||||
INIT_LIST_HEAD(&jfs_ip->anon_inode_list);
|
||||
init_rwsem(&jfs_ip->rdwrlock);
|
||||
@ -766,7 +765,6 @@ static void init_once(void *foo, struct kmem_cache * cachep, unsigned long flags
|
||||
#endif
|
||||
inode_init_once(&jfs_ip->vfs_inode);
|
||||
}
|
||||
}
|
||||
|
||||
static int __init init_jfs_fs(void)
|
||||
{
|
||||
|
@ -203,9 +203,6 @@ static void init_once(void *foo, struct kmem_cache *cache, unsigned long flags)
|
||||
{
|
||||
struct file_lock *lock = (struct file_lock *) foo;
|
||||
|
||||
if (!(flags & SLAB_CTOR_CONSTRUCTOR))
|
||||
return;
|
||||
|
||||
locks_init_lock(lock);
|
||||
}
|
||||
|
||||
|
@ -73,7 +73,6 @@ static void init_once(void * foo, struct kmem_cache * cachep, unsigned long flag
|
||||
{
|
||||
struct minix_inode_info *ei = (struct minix_inode_info *) foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR)
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
}
|
||||
|
||||
|
@ -60,11 +60,9 @@ static void init_once(void * foo, struct kmem_cache * cachep, unsigned long flag
|
||||
{
|
||||
struct ncp_inode_info *ei = (struct ncp_inode_info *) foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
mutex_init(&ei->open_mutex);
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
}
|
||||
}
|
||||
|
||||
static int init_inodecache(void)
|
||||
{
|
||||
|
@ -1164,7 +1164,6 @@ static void init_once(void * foo, struct kmem_cache * cachep, unsigned long flag
|
||||
{
|
||||
struct nfs_inode *nfsi = (struct nfs_inode *) foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
inode_init_once(&nfsi->vfs_inode);
|
||||
spin_lock_init(&nfsi->req_lock);
|
||||
INIT_LIST_HEAD(&nfsi->dirty);
|
||||
@ -1179,7 +1178,6 @@ static void init_once(void * foo, struct kmem_cache * cachep, unsigned long flag
|
||||
nfsi->npages = 0;
|
||||
nfs4_init_once(nfsi);
|
||||
}
|
||||
}
|
||||
|
||||
static int __init nfs_init_inodecache(void)
|
||||
{
|
||||
|
@ -3085,7 +3085,6 @@ static void ntfs_big_inode_init_once(void *foo, struct kmem_cache *cachep,
|
||||
{
|
||||
ntfs_inode *ni = (ntfs_inode *)foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR)
|
||||
inode_init_once(VFS_I(ni));
|
||||
}
|
||||
|
||||
|
@ -262,13 +262,11 @@ static void dlmfs_init_once(void *foo,
|
||||
struct dlmfs_inode_private *ip =
|
||||
(struct dlmfs_inode_private *) foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
ip->ip_dlm = NULL;
|
||||
ip->ip_parent = NULL;
|
||||
|
||||
inode_init_once(&ip->ip_vfs_inode);
|
||||
}
|
||||
}
|
||||
|
||||
static struct inode *dlmfs_alloc_inode(struct super_block *sb)
|
||||
{
|
||||
|
@ -937,7 +937,6 @@ static void ocfs2_inode_init_once(void *data,
|
||||
{
|
||||
struct ocfs2_inode_info *oi = data;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
oi->ip_flags = 0;
|
||||
oi->ip_open_count = 0;
|
||||
spin_lock_init(&oi->ip_lock);
|
||||
@ -962,7 +961,6 @@ static void ocfs2_inode_init_once(void *data,
|
||||
|
||||
inode_init_once(&oi->vfs_inode);
|
||||
}
|
||||
}
|
||||
|
||||
static int ocfs2_initialize_mem_caches(void)
|
||||
{
|
||||
|
@ -419,7 +419,6 @@ static void op_inode_init_once(void *data, struct kmem_cache * cachep, unsigned
|
||||
{
|
||||
struct op_inode_info *oi = (struct op_inode_info *) data;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR)
|
||||
inode_init_once(&oi->vfs_inode);
|
||||
}
|
||||
|
||||
|
@ -109,7 +109,6 @@ static void init_once(void * foo, struct kmem_cache * cachep, unsigned long flag
|
||||
{
|
||||
struct proc_inode *ei = (struct proc_inode *) foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR)
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
}
|
||||
|
||||
|
@ -536,7 +536,6 @@ static void init_once(void *foo, struct kmem_cache * cachep,
|
||||
{
|
||||
struct qnx4_inode_info *ei = (struct qnx4_inode_info *) foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR)
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
}
|
||||
|
||||
|
@ -511,7 +511,6 @@ static void init_once(void *foo, struct kmem_cache * cachep, unsigned long flags
|
||||
{
|
||||
struct reiserfs_inode_info *ei = (struct reiserfs_inode_info *)foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
INIT_LIST_HEAD(&ei->i_prealloc_list);
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
#ifdef CONFIG_REISERFS_FS_POSIX_ACL
|
||||
@ -519,7 +518,6 @@ static void init_once(void *foo, struct kmem_cache * cachep, unsigned long flags
|
||||
ei->i_acl_default = NULL;
|
||||
#endif
|
||||
}
|
||||
}
|
||||
|
||||
static int init_inodecache(void)
|
||||
{
|
||||
|
@ -568,9 +568,8 @@ static void romfs_destroy_inode(struct inode *inode)
|
||||
|
||||
static void init_once(void *foo, struct kmem_cache *cachep, unsigned long flags)
|
||||
{
|
||||
struct romfs_inode_info *ei = (struct romfs_inode_info *) foo;
|
||||
struct romfs_inode_info *ei = foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR)
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
}
|
||||
|
||||
|
@ -70,7 +70,6 @@ static void init_once(void * foo, struct kmem_cache * cachep, unsigned long flag
|
||||
{
|
||||
struct smb_inode_info *ei = (struct smb_inode_info *) foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR)
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
}
|
||||
|
||||
|
@ -322,7 +322,6 @@ static void init_once(void *p, struct kmem_cache *cachep, unsigned long flags)
|
||||
{
|
||||
struct sysv_inode_info *si = (struct sysv_inode_info *)p;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR)
|
||||
inode_init_once(&si->vfs_inode);
|
||||
}
|
||||
|
||||
|
@ -134,11 +134,9 @@ static void init_once(void * foo, struct kmem_cache * cachep, unsigned long flag
|
||||
{
|
||||
struct udf_inode_info *ei = (struct udf_inode_info *) foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
ei->i_ext.i_data = NULL;
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
}
|
||||
}
|
||||
|
||||
static int init_inodecache(void)
|
||||
{
|
||||
|
@ -1237,7 +1237,6 @@ static void init_once(void * foo, struct kmem_cache * cachep, unsigned long flag
|
||||
{
|
||||
struct ufs_inode_info *ei = (struct ufs_inode_info *) foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR)
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
}
|
||||
|
||||
|
@ -360,7 +360,6 @@ xfs_fs_inode_init_once(
|
||||
kmem_zone_t *zonep,
|
||||
unsigned long flags)
|
||||
{
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR)
|
||||
inode_init_once(vn_to_inode((bhv_vnode_t *)vnode));
|
||||
}
|
||||
|
||||
|
@ -32,9 +32,6 @@ typedef struct kmem_cache kmem_cache_t __deprecated;
|
||||
#define SLAB_MEM_SPREAD 0x00100000UL /* Spread some memory over cpuset */
|
||||
#define SLAB_TRACE 0x00200000UL /* Trace allocations and frees */
|
||||
|
||||
/* Flags passed to a constructor functions */
|
||||
#define SLAB_CTOR_CONSTRUCTOR 0x001UL /* If not set, then deconstructor */
|
||||
|
||||
/*
|
||||
* struct kmem_cache related prototypes
|
||||
*/
|
||||
|
@ -215,7 +215,6 @@ static void init_once(void *foo, struct kmem_cache * cachep, unsigned long flags
|
||||
{
|
||||
struct mqueue_inode_info *p = (struct mqueue_inode_info *) foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR)
|
||||
inode_init_once(&p->vfs_inode);
|
||||
}
|
||||
|
||||
|
@ -1427,11 +1427,9 @@ static void sighand_ctor(void *data, struct kmem_cache *cachep,
|
||||
{
|
||||
struct sighand_struct *sighand = data;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
spin_lock_init(&sighand->siglock);
|
||||
INIT_LIST_HEAD(&sighand->signalfd_list);
|
||||
}
|
||||
}
|
||||
|
||||
void __init proc_caches_init(void)
|
||||
{
|
||||
|
@ -162,13 +162,11 @@ void anon_vma_unlink(struct vm_area_struct *vma)
|
||||
static void anon_vma_ctor(void *data, struct kmem_cache *cachep,
|
||||
unsigned long flags)
|
||||
{
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
struct anon_vma *anon_vma = data;
|
||||
|
||||
spin_lock_init(&anon_vma->lock);
|
||||
INIT_LIST_HEAD(&anon_vma->head);
|
||||
}
|
||||
}
|
||||
|
||||
void __init anon_vma_init(void)
|
||||
{
|
||||
|
@ -2358,14 +2358,12 @@ static void init_once(void *foo, struct kmem_cache *cachep,
|
||||
{
|
||||
struct shmem_inode_info *p = (struct shmem_inode_info *) foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
inode_init_once(&p->vfs_inode);
|
||||
#ifdef CONFIG_TMPFS_POSIX_ACL
|
||||
p->i_acl = NULL;
|
||||
p->i_default_acl = NULL;
|
||||
#endif
|
||||
}
|
||||
}
|
||||
|
||||
static int init_inodecache(void)
|
||||
{
|
||||
|
12
mm/slab.c
12
mm/slab.c
@ -2610,7 +2610,7 @@ static inline kmem_bufctl_t *slab_bufctl(struct slab *slabp)
|
||||
}
|
||||
|
||||
static void cache_init_objs(struct kmem_cache *cachep,
|
||||
struct slab *slabp, unsigned long ctor_flags)
|
||||
struct slab *slabp)
|
||||
{
|
||||
int i;
|
||||
|
||||
@ -2634,7 +2634,7 @@ static void cache_init_objs(struct kmem_cache *cachep,
|
||||
*/
|
||||
if (cachep->ctor && !(cachep->flags & SLAB_POISON))
|
||||
cachep->ctor(objp + obj_offset(cachep), cachep,
|
||||
ctor_flags);
|
||||
0);
|
||||
|
||||
if (cachep->flags & SLAB_RED_ZONE) {
|
||||
if (*dbg_redzone2(cachep, objp) != RED_INACTIVE)
|
||||
@ -2650,7 +2650,7 @@ static void cache_init_objs(struct kmem_cache *cachep,
|
||||
cachep->buffer_size / PAGE_SIZE, 0);
|
||||
#else
|
||||
if (cachep->ctor)
|
||||
cachep->ctor(objp, cachep, ctor_flags);
|
||||
cachep->ctor(objp, cachep, 0);
|
||||
#endif
|
||||
slab_bufctl(slabp)[i] = i + 1;
|
||||
}
|
||||
@ -2739,7 +2739,6 @@ static int cache_grow(struct kmem_cache *cachep,
|
||||
struct slab *slabp;
|
||||
size_t offset;
|
||||
gfp_t local_flags;
|
||||
unsigned long ctor_flags;
|
||||
struct kmem_list3 *l3;
|
||||
|
||||
/*
|
||||
@ -2748,7 +2747,6 @@ static int cache_grow(struct kmem_cache *cachep,
|
||||
*/
|
||||
BUG_ON(flags & ~(GFP_DMA | GFP_LEVEL_MASK));
|
||||
|
||||
ctor_flags = SLAB_CTOR_CONSTRUCTOR;
|
||||
local_flags = (flags & GFP_LEVEL_MASK);
|
||||
/* Take the l3 list lock to change the colour_next on this node */
|
||||
check_irq_off();
|
||||
@ -2793,7 +2791,7 @@ static int cache_grow(struct kmem_cache *cachep,
|
||||
slabp->nodeid = nodeid;
|
||||
slab_map_pages(cachep, slabp, objp);
|
||||
|
||||
cache_init_objs(cachep, slabp, ctor_flags);
|
||||
cache_init_objs(cachep, slabp);
|
||||
|
||||
if (local_flags & __GFP_WAIT)
|
||||
local_irq_disable();
|
||||
@ -3077,7 +3075,7 @@ static void *cache_alloc_debugcheck_after(struct kmem_cache *cachep,
|
||||
#endif
|
||||
objp += obj_offset(cachep);
|
||||
if (cachep->ctor && cachep->flags & SLAB_POISON)
|
||||
cachep->ctor(objp, cachep, SLAB_CTOR_CONSTRUCTOR);
|
||||
cachep->ctor(objp, cachep, 0);
|
||||
#if ARCH_SLAB_MINALIGN
|
||||
if ((u32)objp & (ARCH_SLAB_MINALIGN-1)) {
|
||||
printk(KERN_ERR "0x%p: not aligned to ARCH_SLAB_MINALIGN=%d\n",
|
||||
|
@ -327,7 +327,7 @@ void *kmem_cache_alloc(struct kmem_cache *c, gfp_t flags)
|
||||
b = (void *)__get_free_pages(flags, get_order(c->size));
|
||||
|
||||
if (c->ctor)
|
||||
c->ctor(b, c, SLAB_CTOR_CONSTRUCTOR);
|
||||
c->ctor(b, c, 0);
|
||||
|
||||
return b;
|
||||
}
|
||||
|
@ -994,7 +994,7 @@ static void setup_object(struct kmem_cache *s, struct page *page,
|
||||
}
|
||||
|
||||
if (unlikely(s->ctor))
|
||||
s->ctor(object, s, SLAB_CTOR_CONSTRUCTOR);
|
||||
s->ctor(object, s, 0);
|
||||
}
|
||||
|
||||
static struct page *new_slab(struct kmem_cache *s, gfp_t flags, int node)
|
||||
|
@ -261,7 +261,6 @@ static void init_once(void *foo, struct kmem_cache *cachep, unsigned long flags)
|
||||
{
|
||||
struct socket_alloc *ei = (struct socket_alloc *)foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR)
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
}
|
||||
|
||||
|
@ -828,7 +828,6 @@ init_once(void * foo, struct kmem_cache * cachep, unsigned long flags)
|
||||
{
|
||||
struct rpc_inode *rpci = (struct rpc_inode *) foo;
|
||||
|
||||
if (flags & SLAB_CTOR_CONSTRUCTOR) {
|
||||
inode_init_once(&rpci->vfs_inode);
|
||||
rpci->private = NULL;
|
||||
rpci->nreaders = 0;
|
||||
@ -841,7 +840,6 @@ init_once(void * foo, struct kmem_cache * cachep, unsigned long flags)
|
||||
rpc_timeout_upcall_queue);
|
||||
rpci->ops = NULL;
|
||||
}
|
||||
}
|
||||
|
||||
int register_rpc_pipefs(void)
|
||||
{
|
||||
|
Loading…
Reference in New Issue
Block a user