mirror of
https://github.com/torvalds/linux.git
synced 2024-12-27 21:33:00 +00:00
4e57b68178
I recently picked up my older work to remove unnecessary #includes of sched.h, starting from a patch by Dave Jones to not include sched.h from module.h. This reduces the number of indirect includes of sched.h by ~300. Another ~400 pointless direct includes can be removed after this disentangling (patch to follow later). However, quite a few indirect includes need to be fixed up for this. In order to feed the patches through -mm with as little disturbance as possible, I've split out the fixes I accumulated up to now (complete for i386 and x86_64, more archs to follow later) and post them before the real patch. This way this large part of the patch is kept simple with only adding #includes, and all hunks are independent of each other. So if any hunk rejects or gets in the way of other patches, just drop it. My scripts will pick it up again in the next round. Signed-off-by: Tim Schmielau <tim@physik3.uni-rostock.de> Signed-off-by: Andrew Morton <akpm@osdl.org> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
674 lines
14 KiB
C
674 lines
14 KiB
C
/*
|
|
* $Id: mtdchar.c,v 1.73 2005/07/04 17:36:41 gleixner Exp $
|
|
*
|
|
* Character-device access to raw MTD devices.
|
|
*
|
|
*/
|
|
|
|
#include <linux/config.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/module.h>
|
|
#include <linux/mtd/mtd.h>
|
|
#include <linux/mtd/compatmac.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/init.h>
|
|
#include <linux/fs.h>
|
|
#include <linux/sched.h> /* TASK_* */
|
|
#include <asm/uaccess.h>
|
|
|
|
#include <linux/device.h>
|
|
|
|
static struct class *mtd_class;
|
|
|
|
static void mtd_notify_add(struct mtd_info* mtd)
|
|
{
|
|
if (!mtd)
|
|
return;
|
|
|
|
class_device_create(mtd_class, NULL, MKDEV(MTD_CHAR_MAJOR, mtd->index*2),
|
|
NULL, "mtd%d", mtd->index);
|
|
|
|
class_device_create(mtd_class, NULL,
|
|
MKDEV(MTD_CHAR_MAJOR, mtd->index*2+1),
|
|
NULL, "mtd%dro", mtd->index);
|
|
}
|
|
|
|
static void mtd_notify_remove(struct mtd_info* mtd)
|
|
{
|
|
if (!mtd)
|
|
return;
|
|
|
|
class_device_destroy(mtd_class, MKDEV(MTD_CHAR_MAJOR, mtd->index*2));
|
|
class_device_destroy(mtd_class, MKDEV(MTD_CHAR_MAJOR, mtd->index*2+1));
|
|
}
|
|
|
|
static struct mtd_notifier notifier = {
|
|
.add = mtd_notify_add,
|
|
.remove = mtd_notify_remove,
|
|
};
|
|
|
|
/*
|
|
* We use file->private_data to store a pointer to the MTDdevice.
|
|
* Since alighment is at least 32 bits, we have 2 bits free for OTP
|
|
* modes as well.
|
|
*/
|
|
|
|
#define TO_MTD(file) (struct mtd_info *)((long)((file)->private_data) & ~3L)
|
|
|
|
#define MTD_MODE_OTP_FACT 1
|
|
#define MTD_MODE_OTP_USER 2
|
|
#define MTD_MODE(file) ((long)((file)->private_data) & 3)
|
|
|
|
#define SET_MTD_MODE(file, mode) \
|
|
do { long __p = (long)((file)->private_data); \
|
|
(file)->private_data = (void *)((__p & ~3L) | mode); } while (0)
|
|
|
|
static loff_t mtd_lseek (struct file *file, loff_t offset, int orig)
|
|
{
|
|
struct mtd_info *mtd = TO_MTD(file);
|
|
|
|
switch (orig) {
|
|
case 0:
|
|
/* SEEK_SET */
|
|
file->f_pos = offset;
|
|
break;
|
|
case 1:
|
|
/* SEEK_CUR */
|
|
file->f_pos += offset;
|
|
break;
|
|
case 2:
|
|
/* SEEK_END */
|
|
file->f_pos =mtd->size + offset;
|
|
break;
|
|
default:
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (file->f_pos < 0)
|
|
file->f_pos = 0;
|
|
else if (file->f_pos >= mtd->size)
|
|
file->f_pos = mtd->size - 1;
|
|
|
|
return file->f_pos;
|
|
}
|
|
|
|
|
|
|
|
static int mtd_open(struct inode *inode, struct file *file)
|
|
{
|
|
int minor = iminor(inode);
|
|
int devnum = minor >> 1;
|
|
struct mtd_info *mtd;
|
|
|
|
DEBUG(MTD_DEBUG_LEVEL0, "MTD_open\n");
|
|
|
|
if (devnum >= MAX_MTD_DEVICES)
|
|
return -ENODEV;
|
|
|
|
/* You can't open the RO devices RW */
|
|
if ((file->f_mode & 2) && (minor & 1))
|
|
return -EACCES;
|
|
|
|
mtd = get_mtd_device(NULL, devnum);
|
|
|
|
if (!mtd)
|
|
return -ENODEV;
|
|
|
|
if (MTD_ABSENT == mtd->type) {
|
|
put_mtd_device(mtd);
|
|
return -ENODEV;
|
|
}
|
|
|
|
file->private_data = mtd;
|
|
|
|
/* You can't open it RW if it's not a writeable device */
|
|
if ((file->f_mode & 2) && !(mtd->flags & MTD_WRITEABLE)) {
|
|
put_mtd_device(mtd);
|
|
return -EACCES;
|
|
}
|
|
|
|
return 0;
|
|
} /* mtd_open */
|
|
|
|
/*====================================================================*/
|
|
|
|
static int mtd_close(struct inode *inode, struct file *file)
|
|
{
|
|
struct mtd_info *mtd;
|
|
|
|
DEBUG(MTD_DEBUG_LEVEL0, "MTD_close\n");
|
|
|
|
mtd = TO_MTD(file);
|
|
|
|
if (mtd->sync)
|
|
mtd->sync(mtd);
|
|
|
|
put_mtd_device(mtd);
|
|
|
|
return 0;
|
|
} /* mtd_close */
|
|
|
|
/* FIXME: This _really_ needs to die. In 2.5, we should lock the
|
|
userspace buffer down and use it directly with readv/writev.
|
|
*/
|
|
#define MAX_KMALLOC_SIZE 0x20000
|
|
|
|
static ssize_t mtd_read(struct file *file, char __user *buf, size_t count,loff_t *ppos)
|
|
{
|
|
struct mtd_info *mtd = TO_MTD(file);
|
|
size_t retlen=0;
|
|
size_t total_retlen=0;
|
|
int ret=0;
|
|
int len;
|
|
char *kbuf;
|
|
|
|
DEBUG(MTD_DEBUG_LEVEL0,"MTD_read\n");
|
|
|
|
if (*ppos + count > mtd->size)
|
|
count = mtd->size - *ppos;
|
|
|
|
if (!count)
|
|
return 0;
|
|
|
|
/* FIXME: Use kiovec in 2.5 to lock down the user's buffers
|
|
and pass them directly to the MTD functions */
|
|
while (count) {
|
|
if (count > MAX_KMALLOC_SIZE)
|
|
len = MAX_KMALLOC_SIZE;
|
|
else
|
|
len = count;
|
|
|
|
kbuf=kmalloc(len,GFP_KERNEL);
|
|
if (!kbuf)
|
|
return -ENOMEM;
|
|
|
|
switch (MTD_MODE(file)) {
|
|
case MTD_MODE_OTP_FACT:
|
|
ret = mtd->read_fact_prot_reg(mtd, *ppos, len, &retlen, kbuf);
|
|
break;
|
|
case MTD_MODE_OTP_USER:
|
|
ret = mtd->read_user_prot_reg(mtd, *ppos, len, &retlen, kbuf);
|
|
break;
|
|
default:
|
|
ret = MTD_READ(mtd, *ppos, len, &retlen, kbuf);
|
|
}
|
|
/* Nand returns -EBADMSG on ecc errors, but it returns
|
|
* the data. For our userspace tools it is important
|
|
* to dump areas with ecc errors !
|
|
* Userspace software which accesses NAND this way
|
|
* must be aware of the fact that it deals with NAND
|
|
*/
|
|
if (!ret || (ret == -EBADMSG)) {
|
|
*ppos += retlen;
|
|
if (copy_to_user(buf, kbuf, retlen)) {
|
|
kfree(kbuf);
|
|
return -EFAULT;
|
|
}
|
|
else
|
|
total_retlen += retlen;
|
|
|
|
count -= retlen;
|
|
buf += retlen;
|
|
if (retlen == 0)
|
|
count = 0;
|
|
}
|
|
else {
|
|
kfree(kbuf);
|
|
return ret;
|
|
}
|
|
|
|
kfree(kbuf);
|
|
}
|
|
|
|
return total_retlen;
|
|
} /* mtd_read */
|
|
|
|
static ssize_t mtd_write(struct file *file, const char __user *buf, size_t count,loff_t *ppos)
|
|
{
|
|
struct mtd_info *mtd = TO_MTD(file);
|
|
char *kbuf;
|
|
size_t retlen;
|
|
size_t total_retlen=0;
|
|
int ret=0;
|
|
int len;
|
|
|
|
DEBUG(MTD_DEBUG_LEVEL0,"MTD_write\n");
|
|
|
|
if (*ppos == mtd->size)
|
|
return -ENOSPC;
|
|
|
|
if (*ppos + count > mtd->size)
|
|
count = mtd->size - *ppos;
|
|
|
|
if (!count)
|
|
return 0;
|
|
|
|
while (count) {
|
|
if (count > MAX_KMALLOC_SIZE)
|
|
len = MAX_KMALLOC_SIZE;
|
|
else
|
|
len = count;
|
|
|
|
kbuf=kmalloc(len,GFP_KERNEL);
|
|
if (!kbuf) {
|
|
printk("kmalloc is null\n");
|
|
return -ENOMEM;
|
|
}
|
|
|
|
if (copy_from_user(kbuf, buf, len)) {
|
|
kfree(kbuf);
|
|
return -EFAULT;
|
|
}
|
|
|
|
switch (MTD_MODE(file)) {
|
|
case MTD_MODE_OTP_FACT:
|
|
ret = -EROFS;
|
|
break;
|
|
case MTD_MODE_OTP_USER:
|
|
if (!mtd->write_user_prot_reg) {
|
|
ret = -EOPNOTSUPP;
|
|
break;
|
|
}
|
|
ret = mtd->write_user_prot_reg(mtd, *ppos, len, &retlen, kbuf);
|
|
break;
|
|
default:
|
|
ret = (*(mtd->write))(mtd, *ppos, len, &retlen, kbuf);
|
|
}
|
|
if (!ret) {
|
|
*ppos += retlen;
|
|
total_retlen += retlen;
|
|
count -= retlen;
|
|
buf += retlen;
|
|
}
|
|
else {
|
|
kfree(kbuf);
|
|
return ret;
|
|
}
|
|
|
|
kfree(kbuf);
|
|
}
|
|
|
|
return total_retlen;
|
|
} /* mtd_write */
|
|
|
|
/*======================================================================
|
|
|
|
IOCTL calls for getting device parameters.
|
|
|
|
======================================================================*/
|
|
static void mtdchar_erase_callback (struct erase_info *instr)
|
|
{
|
|
wake_up((wait_queue_head_t *)instr->priv);
|
|
}
|
|
|
|
static int mtd_ioctl(struct inode *inode, struct file *file,
|
|
u_int cmd, u_long arg)
|
|
{
|
|
struct mtd_info *mtd = TO_MTD(file);
|
|
void __user *argp = (void __user *)arg;
|
|
int ret = 0;
|
|
u_long size;
|
|
|
|
DEBUG(MTD_DEBUG_LEVEL0, "MTD_ioctl\n");
|
|
|
|
size = (cmd & IOCSIZE_MASK) >> IOCSIZE_SHIFT;
|
|
if (cmd & IOC_IN) {
|
|
if (!access_ok(VERIFY_READ, argp, size))
|
|
return -EFAULT;
|
|
}
|
|
if (cmd & IOC_OUT) {
|
|
if (!access_ok(VERIFY_WRITE, argp, size))
|
|
return -EFAULT;
|
|
}
|
|
|
|
switch (cmd) {
|
|
case MEMGETREGIONCOUNT:
|
|
if (copy_to_user(argp, &(mtd->numeraseregions), sizeof(int)))
|
|
return -EFAULT;
|
|
break;
|
|
|
|
case MEMGETREGIONINFO:
|
|
{
|
|
struct region_info_user ur;
|
|
|
|
if (copy_from_user(&ur, argp, sizeof(struct region_info_user)))
|
|
return -EFAULT;
|
|
|
|
if (ur.regionindex >= mtd->numeraseregions)
|
|
return -EINVAL;
|
|
if (copy_to_user(argp, &(mtd->eraseregions[ur.regionindex]),
|
|
sizeof(struct mtd_erase_region_info)))
|
|
return -EFAULT;
|
|
break;
|
|
}
|
|
|
|
case MEMGETINFO:
|
|
if (copy_to_user(argp, mtd, sizeof(struct mtd_info_user)))
|
|
return -EFAULT;
|
|
break;
|
|
|
|
case MEMERASE:
|
|
{
|
|
struct erase_info *erase;
|
|
|
|
if(!(file->f_mode & 2))
|
|
return -EPERM;
|
|
|
|
erase=kmalloc(sizeof(struct erase_info),GFP_KERNEL);
|
|
if (!erase)
|
|
ret = -ENOMEM;
|
|
else {
|
|
wait_queue_head_t waitq;
|
|
DECLARE_WAITQUEUE(wait, current);
|
|
|
|
init_waitqueue_head(&waitq);
|
|
|
|
memset (erase,0,sizeof(struct erase_info));
|
|
if (copy_from_user(&erase->addr, argp,
|
|
sizeof(struct erase_info_user))) {
|
|
kfree(erase);
|
|
return -EFAULT;
|
|
}
|
|
erase->mtd = mtd;
|
|
erase->callback = mtdchar_erase_callback;
|
|
erase->priv = (unsigned long)&waitq;
|
|
|
|
/*
|
|
FIXME: Allow INTERRUPTIBLE. Which means
|
|
not having the wait_queue head on the stack.
|
|
|
|
If the wq_head is on the stack, and we
|
|
leave because we got interrupted, then the
|
|
wq_head is no longer there when the
|
|
callback routine tries to wake us up.
|
|
*/
|
|
ret = mtd->erase(mtd, erase);
|
|
if (!ret) {
|
|
set_current_state(TASK_UNINTERRUPTIBLE);
|
|
add_wait_queue(&waitq, &wait);
|
|
if (erase->state != MTD_ERASE_DONE &&
|
|
erase->state != MTD_ERASE_FAILED)
|
|
schedule();
|
|
remove_wait_queue(&waitq, &wait);
|
|
set_current_state(TASK_RUNNING);
|
|
|
|
ret = (erase->state == MTD_ERASE_FAILED)?-EIO:0;
|
|
}
|
|
kfree(erase);
|
|
}
|
|
break;
|
|
}
|
|
|
|
case MEMWRITEOOB:
|
|
{
|
|
struct mtd_oob_buf buf;
|
|
void *databuf;
|
|
ssize_t retlen;
|
|
|
|
if(!(file->f_mode & 2))
|
|
return -EPERM;
|
|
|
|
if (copy_from_user(&buf, argp, sizeof(struct mtd_oob_buf)))
|
|
return -EFAULT;
|
|
|
|
if (buf.length > 0x4096)
|
|
return -EINVAL;
|
|
|
|
if (!mtd->write_oob)
|
|
ret = -EOPNOTSUPP;
|
|
else
|
|
ret = access_ok(VERIFY_READ, buf.ptr,
|
|
buf.length) ? 0 : EFAULT;
|
|
|
|
if (ret)
|
|
return ret;
|
|
|
|
databuf = kmalloc(buf.length, GFP_KERNEL);
|
|
if (!databuf)
|
|
return -ENOMEM;
|
|
|
|
if (copy_from_user(databuf, buf.ptr, buf.length)) {
|
|
kfree(databuf);
|
|
return -EFAULT;
|
|
}
|
|
|
|
ret = (mtd->write_oob)(mtd, buf.start, buf.length, &retlen, databuf);
|
|
|
|
if (copy_to_user(argp + sizeof(uint32_t), &retlen, sizeof(uint32_t)))
|
|
ret = -EFAULT;
|
|
|
|
kfree(databuf);
|
|
break;
|
|
|
|
}
|
|
|
|
case MEMREADOOB:
|
|
{
|
|
struct mtd_oob_buf buf;
|
|
void *databuf;
|
|
ssize_t retlen;
|
|
|
|
if (copy_from_user(&buf, argp, sizeof(struct mtd_oob_buf)))
|
|
return -EFAULT;
|
|
|
|
if (buf.length > 0x4096)
|
|
return -EINVAL;
|
|
|
|
if (!mtd->read_oob)
|
|
ret = -EOPNOTSUPP;
|
|
else
|
|
ret = access_ok(VERIFY_WRITE, buf.ptr,
|
|
buf.length) ? 0 : -EFAULT;
|
|
|
|
if (ret)
|
|
return ret;
|
|
|
|
databuf = kmalloc(buf.length, GFP_KERNEL);
|
|
if (!databuf)
|
|
return -ENOMEM;
|
|
|
|
ret = (mtd->read_oob)(mtd, buf.start, buf.length, &retlen, databuf);
|
|
|
|
if (put_user(retlen, (uint32_t __user *)argp))
|
|
ret = -EFAULT;
|
|
else if (retlen && copy_to_user(buf.ptr, databuf, retlen))
|
|
ret = -EFAULT;
|
|
|
|
kfree(databuf);
|
|
break;
|
|
}
|
|
|
|
case MEMLOCK:
|
|
{
|
|
struct erase_info_user info;
|
|
|
|
if (copy_from_user(&info, argp, sizeof(info)))
|
|
return -EFAULT;
|
|
|
|
if (!mtd->lock)
|
|
ret = -EOPNOTSUPP;
|
|
else
|
|
ret = mtd->lock(mtd, info.start, info.length);
|
|
break;
|
|
}
|
|
|
|
case MEMUNLOCK:
|
|
{
|
|
struct erase_info_user info;
|
|
|
|
if (copy_from_user(&info, argp, sizeof(info)))
|
|
return -EFAULT;
|
|
|
|
if (!mtd->unlock)
|
|
ret = -EOPNOTSUPP;
|
|
else
|
|
ret = mtd->unlock(mtd, info.start, info.length);
|
|
break;
|
|
}
|
|
|
|
case MEMSETOOBSEL:
|
|
{
|
|
if (copy_from_user(&mtd->oobinfo, argp, sizeof(struct nand_oobinfo)))
|
|
return -EFAULT;
|
|
break;
|
|
}
|
|
|
|
case MEMGETOOBSEL:
|
|
{
|
|
if (copy_to_user(argp, &(mtd->oobinfo), sizeof(struct nand_oobinfo)))
|
|
return -EFAULT;
|
|
break;
|
|
}
|
|
|
|
case MEMGETBADBLOCK:
|
|
{
|
|
loff_t offs;
|
|
|
|
if (copy_from_user(&offs, argp, sizeof(loff_t)))
|
|
return -EFAULT;
|
|
if (!mtd->block_isbad)
|
|
ret = -EOPNOTSUPP;
|
|
else
|
|
return mtd->block_isbad(mtd, offs);
|
|
break;
|
|
}
|
|
|
|
case MEMSETBADBLOCK:
|
|
{
|
|
loff_t offs;
|
|
|
|
if (copy_from_user(&offs, argp, sizeof(loff_t)))
|
|
return -EFAULT;
|
|
if (!mtd->block_markbad)
|
|
ret = -EOPNOTSUPP;
|
|
else
|
|
return mtd->block_markbad(mtd, offs);
|
|
break;
|
|
}
|
|
|
|
#ifdef CONFIG_MTD_OTP
|
|
case OTPSELECT:
|
|
{
|
|
int mode;
|
|
if (copy_from_user(&mode, argp, sizeof(int)))
|
|
return -EFAULT;
|
|
SET_MTD_MODE(file, 0);
|
|
switch (mode) {
|
|
case MTD_OTP_FACTORY:
|
|
if (!mtd->read_fact_prot_reg)
|
|
ret = -EOPNOTSUPP;
|
|
else
|
|
SET_MTD_MODE(file, MTD_MODE_OTP_FACT);
|
|
break;
|
|
case MTD_OTP_USER:
|
|
if (!mtd->read_fact_prot_reg)
|
|
ret = -EOPNOTSUPP;
|
|
else
|
|
SET_MTD_MODE(file, MTD_MODE_OTP_USER);
|
|
break;
|
|
default:
|
|
ret = -EINVAL;
|
|
case MTD_OTP_OFF:
|
|
break;
|
|
}
|
|
file->f_pos = 0;
|
|
break;
|
|
}
|
|
|
|
case OTPGETREGIONCOUNT:
|
|
case OTPGETREGIONINFO:
|
|
{
|
|
struct otp_info *buf = kmalloc(4096, GFP_KERNEL);
|
|
if (!buf)
|
|
return -ENOMEM;
|
|
ret = -EOPNOTSUPP;
|
|
switch (MTD_MODE(file)) {
|
|
case MTD_MODE_OTP_FACT:
|
|
if (mtd->get_fact_prot_info)
|
|
ret = mtd->get_fact_prot_info(mtd, buf, 4096);
|
|
break;
|
|
case MTD_MODE_OTP_USER:
|
|
if (mtd->get_user_prot_info)
|
|
ret = mtd->get_user_prot_info(mtd, buf, 4096);
|
|
break;
|
|
}
|
|
if (ret >= 0) {
|
|
if (cmd == OTPGETREGIONCOUNT) {
|
|
int nbr = ret / sizeof(struct otp_info);
|
|
ret = copy_to_user(argp, &nbr, sizeof(int));
|
|
} else
|
|
ret = copy_to_user(argp, buf, ret);
|
|
if (ret)
|
|
ret = -EFAULT;
|
|
}
|
|
kfree(buf);
|
|
break;
|
|
}
|
|
|
|
case OTPLOCK:
|
|
{
|
|
struct otp_info info;
|
|
|
|
if (MTD_MODE(file) != MTD_MODE_OTP_USER)
|
|
return -EINVAL;
|
|
if (copy_from_user(&info, argp, sizeof(info)))
|
|
return -EFAULT;
|
|
if (!mtd->lock_user_prot_reg)
|
|
return -EOPNOTSUPP;
|
|
ret = mtd->lock_user_prot_reg(mtd, info.start, info.length);
|
|
break;
|
|
}
|
|
#endif
|
|
|
|
default:
|
|
ret = -ENOTTY;
|
|
}
|
|
|
|
return ret;
|
|
} /* memory_ioctl */
|
|
|
|
static struct file_operations mtd_fops = {
|
|
.owner = THIS_MODULE,
|
|
.llseek = mtd_lseek,
|
|
.read = mtd_read,
|
|
.write = mtd_write,
|
|
.ioctl = mtd_ioctl,
|
|
.open = mtd_open,
|
|
.release = mtd_close,
|
|
};
|
|
|
|
static int __init init_mtdchar(void)
|
|
{
|
|
if (register_chrdev(MTD_CHAR_MAJOR, "mtd", &mtd_fops)) {
|
|
printk(KERN_NOTICE "Can't allocate major number %d for Memory Technology Devices.\n",
|
|
MTD_CHAR_MAJOR);
|
|
return -EAGAIN;
|
|
}
|
|
|
|
mtd_class = class_create(THIS_MODULE, "mtd");
|
|
|
|
if (IS_ERR(mtd_class)) {
|
|
printk(KERN_ERR "Error creating mtd class.\n");
|
|
unregister_chrdev(MTD_CHAR_MAJOR, "mtd");
|
|
return PTR_ERR(mtd_class);
|
|
}
|
|
|
|
register_mtd_user(¬ifier);
|
|
return 0;
|
|
}
|
|
|
|
static void __exit cleanup_mtdchar(void)
|
|
{
|
|
unregister_mtd_user(¬ifier);
|
|
class_destroy(mtd_class);
|
|
unregister_chrdev(MTD_CHAR_MAJOR, "mtd");
|
|
}
|
|
|
|
module_init(init_mtdchar);
|
|
module_exit(cleanup_mtdchar);
|
|
|
|
|
|
MODULE_LICENSE("GPL");
|
|
MODULE_AUTHOR("David Woodhouse <dwmw2@infradead.org>");
|
|
MODULE_DESCRIPTION("Direct character-device access to MTD devices");
|