mirror of
https://github.com/torvalds/linux.git
synced 2024-11-27 22:51:35 +00:00
12a26d0879
Received from Mark Salyzyn from Adaptec: If the Adapter is quiet and does not produce an AIF event packets to be picked up by the management applications for longer than the timeout interval of two minutes, the cleanup code that deals with aging out registrants could erroneously drop the registration. The timeout is there to clean up should the management application die and fail to poll for updated AIF event packets. Moving the timer update from the ioctl code that delivers an AIF to the polling registrant to the bottom of the ioctl means the timeout is reset with any management application polling activity regardless if an AIF is delivered or not removing the erroneous timeout cleanups. Signed-off-by: Mark Haverkamp <markh@osdl.org> Signed-off-by: James Bottomley <James.Bottomley@SteelEye.com>
752 lines
20 KiB
C
752 lines
20 KiB
C
/*
|
|
* Adaptec AAC series RAID controller driver
|
|
* (c) Copyright 2001 Red Hat Inc. <alan@redhat.com>
|
|
*
|
|
* based on the old aacraid driver that is..
|
|
* Adaptec aacraid device driver for Linux.
|
|
*
|
|
* Copyright (c) 2000 Adaptec, Inc. (aacraid@adaptec.com)
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation; either version 2, or (at your option)
|
|
* any later version.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program; see the file COPYING. If not, write to
|
|
* the Free Software Foundation, 675 Mass Ave, Cambridge, MA 02139, USA.
|
|
*
|
|
* Module Name:
|
|
* commctrl.c
|
|
*
|
|
* Abstract: Contains all routines for control of the AFA comm layer
|
|
*
|
|
*/
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/init.h>
|
|
#include <linux/types.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/pci.h>
|
|
#include <linux/spinlock.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/completion.h>
|
|
#include <linux/dma-mapping.h>
|
|
#include <linux/blkdev.h>
|
|
#include <asm/semaphore.h>
|
|
#include <asm/uaccess.h>
|
|
|
|
#include "aacraid.h"
|
|
|
|
/**
|
|
* ioctl_send_fib - send a FIB from userspace
|
|
* @dev: adapter is being processed
|
|
* @arg: arguments to the ioctl call
|
|
*
|
|
* This routine sends a fib to the adapter on behalf of a user level
|
|
* program.
|
|
*/
|
|
# define AAC_DEBUG_PREAMBLE KERN_INFO
|
|
# define AAC_DEBUG_POSTAMBLE
|
|
|
|
static int ioctl_send_fib(struct aac_dev * dev, void __user *arg)
|
|
{
|
|
struct hw_fib * kfib;
|
|
struct fib *fibptr;
|
|
struct hw_fib * hw_fib = (struct hw_fib *)0;
|
|
dma_addr_t hw_fib_pa = (dma_addr_t)0LL;
|
|
unsigned size;
|
|
int retval;
|
|
|
|
fibptr = fib_alloc(dev);
|
|
if(fibptr == NULL) {
|
|
return -ENOMEM;
|
|
}
|
|
|
|
kfib = fibptr->hw_fib;
|
|
/*
|
|
* First copy in the header so that we can check the size field.
|
|
*/
|
|
if (copy_from_user((void *)kfib, arg, sizeof(struct aac_fibhdr))) {
|
|
fib_free(fibptr);
|
|
return -EFAULT;
|
|
}
|
|
/*
|
|
* Since we copy based on the fib header size, make sure that we
|
|
* will not overrun the buffer when we copy the memory. Return
|
|
* an error if we would.
|
|
*/
|
|
size = le16_to_cpu(kfib->header.Size) + sizeof(struct aac_fibhdr);
|
|
if (size < le16_to_cpu(kfib->header.SenderSize))
|
|
size = le16_to_cpu(kfib->header.SenderSize);
|
|
if (size > dev->max_fib_size) {
|
|
/* Highjack the hw_fib */
|
|
hw_fib = fibptr->hw_fib;
|
|
hw_fib_pa = fibptr->hw_fib_pa;
|
|
fibptr->hw_fib = kfib = pci_alloc_consistent(dev->pdev, size, &fibptr->hw_fib_pa);
|
|
memset(((char *)kfib) + dev->max_fib_size, 0, size - dev->max_fib_size);
|
|
memcpy(kfib, hw_fib, dev->max_fib_size);
|
|
}
|
|
|
|
if (copy_from_user(kfib, arg, size)) {
|
|
retval = -EFAULT;
|
|
goto cleanup;
|
|
}
|
|
|
|
if (kfib->header.Command == cpu_to_le16(TakeABreakPt)) {
|
|
aac_adapter_interrupt(dev);
|
|
/*
|
|
* Since we didn't really send a fib, zero out the state to allow
|
|
* cleanup code not to assert.
|
|
*/
|
|
kfib->header.XferState = 0;
|
|
} else {
|
|
retval = fib_send(le16_to_cpu(kfib->header.Command), fibptr,
|
|
le16_to_cpu(kfib->header.Size) , FsaNormal,
|
|
1, 1, NULL, NULL);
|
|
if (retval) {
|
|
goto cleanup;
|
|
}
|
|
if (fib_complete(fibptr) != 0) {
|
|
retval = -EINVAL;
|
|
goto cleanup;
|
|
}
|
|
}
|
|
/*
|
|
* Make sure that the size returned by the adapter (which includes
|
|
* the header) is less than or equal to the size of a fib, so we
|
|
* don't corrupt application data. Then copy that size to the user
|
|
* buffer. (Don't try to add the header information again, since it
|
|
* was already included by the adapter.)
|
|
*/
|
|
|
|
retval = 0;
|
|
if (copy_to_user(arg, (void *)kfib, size))
|
|
retval = -EFAULT;
|
|
cleanup:
|
|
if (hw_fib) {
|
|
pci_free_consistent(dev->pdev, size, kfib, fibptr->hw_fib_pa);
|
|
fibptr->hw_fib_pa = hw_fib_pa;
|
|
fibptr->hw_fib = hw_fib;
|
|
}
|
|
fib_free(fibptr);
|
|
return retval;
|
|
}
|
|
|
|
/**
|
|
* open_getadapter_fib - Get the next fib
|
|
*
|
|
* This routine will get the next Fib, if available, from the AdapterFibContext
|
|
* passed in from the user.
|
|
*/
|
|
|
|
static int open_getadapter_fib(struct aac_dev * dev, void __user *arg)
|
|
{
|
|
struct aac_fib_context * fibctx;
|
|
int status;
|
|
|
|
fibctx = kmalloc(sizeof(struct aac_fib_context), GFP_KERNEL);
|
|
if (fibctx == NULL) {
|
|
status = -ENOMEM;
|
|
} else {
|
|
unsigned long flags;
|
|
struct list_head * entry;
|
|
struct aac_fib_context * context;
|
|
|
|
fibctx->type = FSAFS_NTC_GET_ADAPTER_FIB_CONTEXT;
|
|
fibctx->size = sizeof(struct aac_fib_context);
|
|
/*
|
|
* Yes yes, I know this could be an index, but we have a
|
|
* better guarantee of uniqueness for the locked loop below.
|
|
* Without the aid of a persistent history, this also helps
|
|
* reduce the chance that the opaque context would be reused.
|
|
*/
|
|
fibctx->unique = (u32)((ulong)fibctx & 0xFFFFFFFF);
|
|
/*
|
|
* Initialize the mutex used to wait for the next AIF.
|
|
*/
|
|
init_MUTEX_LOCKED(&fibctx->wait_sem);
|
|
fibctx->wait = 0;
|
|
/*
|
|
* Initialize the fibs and set the count of fibs on
|
|
* the list to 0.
|
|
*/
|
|
fibctx->count = 0;
|
|
INIT_LIST_HEAD(&fibctx->fib_list);
|
|
fibctx->jiffies = jiffies/HZ;
|
|
/*
|
|
* Now add this context onto the adapter's
|
|
* AdapterFibContext list.
|
|
*/
|
|
spin_lock_irqsave(&dev->fib_lock, flags);
|
|
/* Ensure that we have a unique identifier */
|
|
entry = dev->fib_list.next;
|
|
while (entry != &dev->fib_list) {
|
|
context = list_entry(entry, struct aac_fib_context, next);
|
|
if (context->unique == fibctx->unique) {
|
|
/* Not unique (32 bits) */
|
|
fibctx->unique++;
|
|
entry = dev->fib_list.next;
|
|
} else {
|
|
entry = entry->next;
|
|
}
|
|
}
|
|
list_add_tail(&fibctx->next, &dev->fib_list);
|
|
spin_unlock_irqrestore(&dev->fib_lock, flags);
|
|
if (copy_to_user(arg, &fibctx->unique,
|
|
sizeof(fibctx->unique))) {
|
|
status = -EFAULT;
|
|
} else {
|
|
status = 0;
|
|
}
|
|
}
|
|
return status;
|
|
}
|
|
|
|
/**
|
|
* next_getadapter_fib - get the next fib
|
|
* @dev: adapter to use
|
|
* @arg: ioctl argument
|
|
*
|
|
* This routine will get the next Fib, if available, from the AdapterFibContext
|
|
* passed in from the user.
|
|
*/
|
|
|
|
static int next_getadapter_fib(struct aac_dev * dev, void __user *arg)
|
|
{
|
|
struct fib_ioctl f;
|
|
struct fib *fib;
|
|
struct aac_fib_context *fibctx;
|
|
int status;
|
|
struct list_head * entry;
|
|
unsigned long flags;
|
|
|
|
if(copy_from_user((void *)&f, arg, sizeof(struct fib_ioctl)))
|
|
return -EFAULT;
|
|
/*
|
|
* Verify that the HANDLE passed in was a valid AdapterFibContext
|
|
*
|
|
* Search the list of AdapterFibContext addresses on the adapter
|
|
* to be sure this is a valid address
|
|
*/
|
|
entry = dev->fib_list.next;
|
|
fibctx = NULL;
|
|
|
|
while (entry != &dev->fib_list) {
|
|
fibctx = list_entry(entry, struct aac_fib_context, next);
|
|
/*
|
|
* Extract the AdapterFibContext from the Input parameters.
|
|
*/
|
|
if (fibctx->unique == f.fibctx) { /* We found a winner */
|
|
break;
|
|
}
|
|
entry = entry->next;
|
|
fibctx = NULL;
|
|
}
|
|
if (!fibctx) {
|
|
dprintk ((KERN_INFO "Fib Context not found\n"));
|
|
return -EINVAL;
|
|
}
|
|
|
|
if((fibctx->type != FSAFS_NTC_GET_ADAPTER_FIB_CONTEXT) ||
|
|
(fibctx->size != sizeof(struct aac_fib_context))) {
|
|
dprintk ((KERN_INFO "Fib Context corrupt?\n"));
|
|
return -EINVAL;
|
|
}
|
|
status = 0;
|
|
spin_lock_irqsave(&dev->fib_lock, flags);
|
|
/*
|
|
* If there are no fibs to send back, then either wait or return
|
|
* -EAGAIN
|
|
*/
|
|
return_fib:
|
|
if (!list_empty(&fibctx->fib_list)) {
|
|
struct list_head * entry;
|
|
/*
|
|
* Pull the next fib from the fibs
|
|
*/
|
|
entry = fibctx->fib_list.next;
|
|
list_del(entry);
|
|
|
|
fib = list_entry(entry, struct fib, fiblink);
|
|
fibctx->count--;
|
|
spin_unlock_irqrestore(&dev->fib_lock, flags);
|
|
if (copy_to_user(f.fib, fib->hw_fib, sizeof(struct hw_fib))) {
|
|
kfree(fib->hw_fib);
|
|
kfree(fib);
|
|
return -EFAULT;
|
|
}
|
|
/*
|
|
* Free the space occupied by this copy of the fib.
|
|
*/
|
|
kfree(fib->hw_fib);
|
|
kfree(fib);
|
|
status = 0;
|
|
} else {
|
|
spin_unlock_irqrestore(&dev->fib_lock, flags);
|
|
if (f.wait) {
|
|
if(down_interruptible(&fibctx->wait_sem) < 0) {
|
|
status = -EINTR;
|
|
} else {
|
|
/* Lock again and retry */
|
|
spin_lock_irqsave(&dev->fib_lock, flags);
|
|
goto return_fib;
|
|
}
|
|
} else {
|
|
status = -EAGAIN;
|
|
}
|
|
}
|
|
fibctx->jiffies = jiffies/HZ;
|
|
return status;
|
|
}
|
|
|
|
int aac_close_fib_context(struct aac_dev * dev, struct aac_fib_context * fibctx)
|
|
{
|
|
struct fib *fib;
|
|
|
|
/*
|
|
* First free any FIBs that have not been consumed.
|
|
*/
|
|
while (!list_empty(&fibctx->fib_list)) {
|
|
struct list_head * entry;
|
|
/*
|
|
* Pull the next fib from the fibs
|
|
*/
|
|
entry = fibctx->fib_list.next;
|
|
list_del(entry);
|
|
fib = list_entry(entry, struct fib, fiblink);
|
|
fibctx->count--;
|
|
/*
|
|
* Free the space occupied by this copy of the fib.
|
|
*/
|
|
kfree(fib->hw_fib);
|
|
kfree(fib);
|
|
}
|
|
/*
|
|
* Remove the Context from the AdapterFibContext List
|
|
*/
|
|
list_del(&fibctx->next);
|
|
/*
|
|
* Invalidate context
|
|
*/
|
|
fibctx->type = 0;
|
|
/*
|
|
* Free the space occupied by the Context
|
|
*/
|
|
kfree(fibctx);
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* close_getadapter_fib - close down user fib context
|
|
* @dev: adapter
|
|
* @arg: ioctl arguments
|
|
*
|
|
* This routine will close down the fibctx passed in from the user.
|
|
*/
|
|
|
|
static int close_getadapter_fib(struct aac_dev * dev, void __user *arg)
|
|
{
|
|
struct aac_fib_context *fibctx;
|
|
int status;
|
|
unsigned long flags;
|
|
struct list_head * entry;
|
|
|
|
/*
|
|
* Verify that the HANDLE passed in was a valid AdapterFibContext
|
|
*
|
|
* Search the list of AdapterFibContext addresses on the adapter
|
|
* to be sure this is a valid address
|
|
*/
|
|
|
|
entry = dev->fib_list.next;
|
|
fibctx = NULL;
|
|
|
|
while(entry != &dev->fib_list) {
|
|
fibctx = list_entry(entry, struct aac_fib_context, next);
|
|
/*
|
|
* Extract the fibctx from the input parameters
|
|
*/
|
|
if (fibctx->unique == (u32)(unsigned long)arg) {
|
|
/* We found a winner */
|
|
break;
|
|
}
|
|
entry = entry->next;
|
|
fibctx = NULL;
|
|
}
|
|
|
|
if (!fibctx)
|
|
return 0; /* Already gone */
|
|
|
|
if((fibctx->type != FSAFS_NTC_GET_ADAPTER_FIB_CONTEXT) ||
|
|
(fibctx->size != sizeof(struct aac_fib_context)))
|
|
return -EINVAL;
|
|
spin_lock_irqsave(&dev->fib_lock, flags);
|
|
status = aac_close_fib_context(dev, fibctx);
|
|
spin_unlock_irqrestore(&dev->fib_lock, flags);
|
|
return status;
|
|
}
|
|
|
|
/**
|
|
* check_revision - close down user fib context
|
|
* @dev: adapter
|
|
* @arg: ioctl arguments
|
|
*
|
|
* This routine returns the driver version.
|
|
* Under Linux, there have been no version incompatibilities, so this is
|
|
* simple!
|
|
*/
|
|
|
|
static int check_revision(struct aac_dev *dev, void __user *arg)
|
|
{
|
|
struct revision response;
|
|
char *driver_version = aac_driver_version;
|
|
u32 version;
|
|
|
|
response.compat = cpu_to_le32(1);
|
|
version = (simple_strtol(driver_version,
|
|
&driver_version, 10) << 24) | 0x00000400;
|
|
version += simple_strtol(driver_version + 1, &driver_version, 10) << 16;
|
|
version += simple_strtol(driver_version + 1, NULL, 10);
|
|
response.version = cpu_to_le32(version);
|
|
# if (defined(AAC_DRIVER_BUILD))
|
|
response.build = cpu_to_le32(AAC_DRIVER_BUILD);
|
|
# else
|
|
response.build = cpu_to_le32(9999);
|
|
# endif
|
|
|
|
if (copy_to_user(arg, &response, sizeof(response)))
|
|
return -EFAULT;
|
|
return 0;
|
|
}
|
|
|
|
|
|
/**
|
|
*
|
|
* aac_send_raw_scb
|
|
*
|
|
*/
|
|
|
|
static int aac_send_raw_srb(struct aac_dev* dev, void __user * arg)
|
|
{
|
|
struct fib* srbfib;
|
|
int status;
|
|
struct aac_srb *srbcmd = NULL;
|
|
struct user_aac_srb *user_srbcmd = NULL;
|
|
struct user_aac_srb __user *user_srb = arg;
|
|
struct aac_srb_reply __user *user_reply;
|
|
struct aac_srb_reply* reply;
|
|
u32 fibsize = 0;
|
|
u32 flags = 0;
|
|
s32 rcode = 0;
|
|
u32 data_dir;
|
|
void __user *sg_user[32];
|
|
void *sg_list[32];
|
|
u32 sg_indx = 0;
|
|
u32 byte_count = 0;
|
|
u32 actual_fibsize = 0;
|
|
int i;
|
|
|
|
|
|
if (!capable(CAP_SYS_ADMIN)){
|
|
dprintk((KERN_DEBUG"aacraid: No permission to send raw srb\n"));
|
|
return -EPERM;
|
|
}
|
|
/*
|
|
* Allocate and initialize a Fib then setup a BlockWrite command
|
|
*/
|
|
if (!(srbfib = fib_alloc(dev))) {
|
|
return -ENOMEM;
|
|
}
|
|
fib_init(srbfib);
|
|
|
|
srbcmd = (struct aac_srb*) fib_data(srbfib);
|
|
|
|
memset(sg_list, 0, sizeof(sg_list)); /* cleanup may take issue */
|
|
if(copy_from_user(&fibsize, &user_srb->count,sizeof(u32))){
|
|
dprintk((KERN_DEBUG"aacraid: Could not copy data size from user\n"));
|
|
rcode = -EFAULT;
|
|
goto cleanup;
|
|
}
|
|
|
|
if (fibsize > (dev->max_fib_size - sizeof(struct aac_fibhdr))) {
|
|
rcode = -EINVAL;
|
|
goto cleanup;
|
|
}
|
|
|
|
user_srbcmd = kmalloc(fibsize, GFP_KERNEL);
|
|
if (!user_srbcmd) {
|
|
dprintk((KERN_DEBUG"aacraid: Could not make a copy of the srb\n"));
|
|
rcode = -ENOMEM;
|
|
goto cleanup;
|
|
}
|
|
if(copy_from_user(user_srbcmd, user_srb,fibsize)){
|
|
dprintk((KERN_DEBUG"aacraid: Could not copy srb from user\n"));
|
|
rcode = -EFAULT;
|
|
goto cleanup;
|
|
}
|
|
|
|
user_reply = arg+fibsize;
|
|
|
|
flags = user_srbcmd->flags; /* from user in cpu order */
|
|
// Fix up srb for endian and force some values
|
|
|
|
srbcmd->function = cpu_to_le32(SRBF_ExecuteScsi); // Force this
|
|
srbcmd->channel = cpu_to_le32(user_srbcmd->channel);
|
|
srbcmd->id = cpu_to_le32(user_srbcmd->id);
|
|
srbcmd->lun = cpu_to_le32(user_srbcmd->lun);
|
|
srbcmd->timeout = cpu_to_le32(user_srbcmd->timeout);
|
|
srbcmd->flags = cpu_to_le32(flags);
|
|
srbcmd->retry_limit = 0; // Obsolete parameter
|
|
srbcmd->cdb_size = cpu_to_le32(user_srbcmd->cdb_size);
|
|
memcpy(srbcmd->cdb, user_srbcmd->cdb, sizeof(srbcmd->cdb));
|
|
|
|
switch (flags & (SRB_DataIn | SRB_DataOut)) {
|
|
case SRB_DataOut:
|
|
data_dir = DMA_TO_DEVICE;
|
|
break;
|
|
case (SRB_DataIn | SRB_DataOut):
|
|
data_dir = DMA_BIDIRECTIONAL;
|
|
break;
|
|
case SRB_DataIn:
|
|
data_dir = DMA_FROM_DEVICE;
|
|
break;
|
|
default:
|
|
data_dir = DMA_NONE;
|
|
}
|
|
if (user_srbcmd->sg.count > (sizeof(sg_list)/sizeof(sg_list[0]))) {
|
|
dprintk((KERN_DEBUG"aacraid: too many sg entries %d\n",
|
|
le32_to_cpu(srbcmd->sg.count)));
|
|
rcode = -EINVAL;
|
|
goto cleanup;
|
|
}
|
|
if (dev->dac_support == 1) {
|
|
struct user_sgmap64* upsg = (struct user_sgmap64*)&user_srbcmd->sg;
|
|
struct sgmap64* psg = (struct sgmap64*)&srbcmd->sg;
|
|
struct user_sgmap* usg;
|
|
byte_count = 0;
|
|
|
|
/*
|
|
* This should also catch if user used the 32 bit sgmap
|
|
*/
|
|
actual_fibsize = sizeof(struct aac_srb) -
|
|
sizeof(struct sgentry) +
|
|
((upsg->count & 0xff) *
|
|
sizeof(struct sgentry));
|
|
if(actual_fibsize != fibsize){ // User made a mistake - should not continue
|
|
dprintk((KERN_DEBUG"aacraid: Bad Size specified in Raw SRB command\n"));
|
|
rcode = -EINVAL;
|
|
goto cleanup;
|
|
}
|
|
usg = kmalloc(actual_fibsize - sizeof(struct aac_srb)
|
|
+ sizeof(struct sgmap), GFP_KERNEL);
|
|
if (!usg) {
|
|
dprintk((KERN_DEBUG"aacraid: Allocation error in Raw SRB command\n"));
|
|
rcode = -ENOMEM;
|
|
goto cleanup;
|
|
}
|
|
memcpy (usg, upsg, actual_fibsize - sizeof(struct aac_srb)
|
|
+ sizeof(struct sgmap));
|
|
actual_fibsize = sizeof(struct aac_srb) -
|
|
sizeof(struct sgentry) + ((usg->count & 0xff) *
|
|
sizeof(struct sgentry64));
|
|
if ((data_dir == DMA_NONE) && upsg->count) {
|
|
kfree (usg);
|
|
dprintk((KERN_DEBUG"aacraid: SG with no direction specified in Raw SRB command\n"));
|
|
rcode = -EINVAL;
|
|
goto cleanup;
|
|
}
|
|
|
|
for (i = 0; i < usg->count; i++) {
|
|
u64 addr;
|
|
void* p;
|
|
/* Does this really need to be GFP_DMA? */
|
|
p = kmalloc(usg->sg[i].count,GFP_KERNEL|__GFP_DMA);
|
|
if(p == 0) {
|
|
kfree (usg);
|
|
dprintk((KERN_DEBUG"aacraid: Could not allocate SG buffer - size = %d buffer number %d of %d\n",
|
|
usg->sg[i].count,i,usg->count));
|
|
rcode = -ENOMEM;
|
|
goto cleanup;
|
|
}
|
|
sg_user[i] = (void __user *)usg->sg[i].addr;
|
|
sg_list[i] = p; // save so we can clean up later
|
|
sg_indx = i;
|
|
|
|
if( flags & SRB_DataOut ){
|
|
if(copy_from_user(p,sg_user[i],upsg->sg[i].count)){
|
|
kfree (usg);
|
|
dprintk((KERN_DEBUG"aacraid: Could not copy sg data from user\n"));
|
|
rcode = -EFAULT;
|
|
goto cleanup;
|
|
}
|
|
}
|
|
addr = pci_map_single(dev->pdev, p, usg->sg[i].count, data_dir);
|
|
|
|
psg->sg[i].addr[0] = cpu_to_le32(addr & 0xffffffff);
|
|
psg->sg[i].addr[1] = cpu_to_le32(addr>>32);
|
|
psg->sg[i].count = cpu_to_le32(usg->sg[i].count);
|
|
byte_count += usg->sg[i].count;
|
|
}
|
|
kfree (usg);
|
|
|
|
srbcmd->count = cpu_to_le32(byte_count);
|
|
psg->count = cpu_to_le32(sg_indx+1);
|
|
status = fib_send(ScsiPortCommand64, srbfib, actual_fibsize, FsaNormal, 1, 1,NULL,NULL);
|
|
} else {
|
|
struct user_sgmap* upsg = &user_srbcmd->sg;
|
|
struct sgmap* psg = &srbcmd->sg;
|
|
byte_count = 0;
|
|
|
|
actual_fibsize = sizeof (struct aac_srb) + (((user_srbcmd->sg.count & 0xff) - 1) * sizeof (struct sgentry));
|
|
if(actual_fibsize != fibsize){ // User made a mistake - should not continue
|
|
dprintk((KERN_DEBUG"aacraid: Bad Size specified in Raw SRB command\n"));
|
|
rcode = -EINVAL;
|
|
goto cleanup;
|
|
}
|
|
if ((data_dir == DMA_NONE) && upsg->count) {
|
|
dprintk((KERN_DEBUG"aacraid: SG with no direction specified in Raw SRB command\n"));
|
|
rcode = -EINVAL;
|
|
goto cleanup;
|
|
}
|
|
for (i = 0; i < upsg->count; i++) {
|
|
dma_addr_t addr;
|
|
void* p;
|
|
p = kmalloc(upsg->sg[i].count, GFP_KERNEL);
|
|
if(p == 0) {
|
|
dprintk((KERN_DEBUG"aacraid: Could not allocate SG buffer - size = %d buffer number %d of %d\n",
|
|
upsg->sg[i].count, i, upsg->count));
|
|
rcode = -ENOMEM;
|
|
goto cleanup;
|
|
}
|
|
sg_user[i] = (void __user *)upsg->sg[i].addr;
|
|
sg_list[i] = p; // save so we can clean up later
|
|
sg_indx = i;
|
|
|
|
if( flags & SRB_DataOut ){
|
|
if(copy_from_user(p, sg_user[i],
|
|
upsg->sg[i].count)) {
|
|
dprintk((KERN_DEBUG"aacraid: Could not copy sg data from user\n"));
|
|
rcode = -EFAULT;
|
|
goto cleanup;
|
|
}
|
|
}
|
|
addr = pci_map_single(dev->pdev, p,
|
|
upsg->sg[i].count, data_dir);
|
|
|
|
psg->sg[i].addr = cpu_to_le32(addr);
|
|
psg->sg[i].count = cpu_to_le32(upsg->sg[i].count);
|
|
byte_count += upsg->sg[i].count;
|
|
}
|
|
srbcmd->count = cpu_to_le32(byte_count);
|
|
psg->count = cpu_to_le32(sg_indx+1);
|
|
status = fib_send(ScsiPortCommand, srbfib, actual_fibsize, FsaNormal, 1, 1, NULL, NULL);
|
|
}
|
|
|
|
if (status != 0){
|
|
dprintk((KERN_DEBUG"aacraid: Could not send raw srb fib to hba\n"));
|
|
rcode = -ENXIO;
|
|
goto cleanup;
|
|
}
|
|
|
|
if( flags & SRB_DataIn ) {
|
|
for(i = 0 ; i <= sg_indx; i++){
|
|
byte_count = le32_to_cpu((dev->dac_support == 1)
|
|
? ((struct sgmap64*)&srbcmd->sg)->sg[i].count
|
|
: srbcmd->sg.sg[i].count);
|
|
if(copy_to_user(sg_user[i], sg_list[i], byte_count)){
|
|
dprintk((KERN_DEBUG"aacraid: Could not copy sg data to user\n"));
|
|
rcode = -EFAULT;
|
|
goto cleanup;
|
|
|
|
}
|
|
}
|
|
}
|
|
|
|
reply = (struct aac_srb_reply *) fib_data(srbfib);
|
|
if(copy_to_user(user_reply,reply,sizeof(struct aac_srb_reply))){
|
|
dprintk((KERN_DEBUG"aacraid: Could not copy reply to user\n"));
|
|
rcode = -EFAULT;
|
|
goto cleanup;
|
|
}
|
|
|
|
cleanup:
|
|
kfree(user_srbcmd);
|
|
for(i=0; i <= sg_indx; i++){
|
|
kfree(sg_list[i]);
|
|
}
|
|
fib_complete(srbfib);
|
|
fib_free(srbfib);
|
|
|
|
return rcode;
|
|
}
|
|
|
|
struct aac_pci_info {
|
|
u32 bus;
|
|
u32 slot;
|
|
};
|
|
|
|
|
|
static int aac_get_pci_info(struct aac_dev* dev, void __user *arg)
|
|
{
|
|
struct aac_pci_info pci_info;
|
|
|
|
pci_info.bus = dev->pdev->bus->number;
|
|
pci_info.slot = PCI_SLOT(dev->pdev->devfn);
|
|
|
|
if (copy_to_user(arg, &pci_info, sizeof(struct aac_pci_info))) {
|
|
dprintk((KERN_DEBUG "aacraid: Could not copy pci info\n"));
|
|
return -EFAULT;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
|
|
int aac_do_ioctl(struct aac_dev * dev, int cmd, void __user *arg)
|
|
{
|
|
int status;
|
|
|
|
/*
|
|
* HBA gets first crack
|
|
*/
|
|
|
|
status = aac_dev_ioctl(dev, cmd, arg);
|
|
if(status != -ENOTTY)
|
|
return status;
|
|
|
|
switch (cmd) {
|
|
case FSACTL_MINIPORT_REV_CHECK:
|
|
status = check_revision(dev, arg);
|
|
break;
|
|
case FSACTL_SEND_LARGE_FIB:
|
|
case FSACTL_SENDFIB:
|
|
status = ioctl_send_fib(dev, arg);
|
|
break;
|
|
case FSACTL_OPEN_GET_ADAPTER_FIB:
|
|
status = open_getadapter_fib(dev, arg);
|
|
break;
|
|
case FSACTL_GET_NEXT_ADAPTER_FIB:
|
|
status = next_getadapter_fib(dev, arg);
|
|
break;
|
|
case FSACTL_CLOSE_GET_ADAPTER_FIB:
|
|
status = close_getadapter_fib(dev, arg);
|
|
break;
|
|
case FSACTL_SEND_RAW_SRB:
|
|
status = aac_send_raw_srb(dev,arg);
|
|
break;
|
|
case FSACTL_GET_PCI_INFO:
|
|
status = aac_get_pci_info(dev,arg);
|
|
break;
|
|
default:
|
|
status = -ENOTTY;
|
|
break;
|
|
}
|
|
return status;
|
|
}
|
|
|