mirror of
https://github.com/torvalds/linux.git
synced 2024-11-27 14:41:39 +00:00
f486893288
mptlan_probe() calls mpt_register_lan_device() which initializes the &priv->post_buckets_task workqueue. A call to mpt_lan_wake_post_buckets_task() will subsequently start the work. During driver unload in mptlan_remove() the following race may occur: CPU0 CPU1 |mpt_lan_post_receive_buckets_work() mptlan_remove() | free_netdev() | kfree(dev); | | | dev->mtu | //use Fix this by finishing the work prior to cleaning up in mptlan_remove(). [mkp: we really should remove mptlan instead of attempting to fix it] Signed-off-by: Zheng Wang <zyytlz.wz@163.com> Link: https://lore.kernel.org/r/20230318081635.796479-1-zyytlz.wz@163.com Signed-off-by: Martin K. Petersen <martin.petersen@oracle.com>
1544 lines
43 KiB
C
1544 lines
43 KiB
C
/*
|
|
* linux/drivers/message/fusion/mptlan.c
|
|
* IP Over Fibre Channel device driver.
|
|
* For use with LSI Fibre Channel PCI chip/adapters
|
|
* running LSI Fusion MPT (Message Passing Technology) firmware.
|
|
*
|
|
* Copyright (c) 2000-2008 LSI Corporation
|
|
* (mailto:DL-MPTFusionLinux@lsi.com)
|
|
*
|
|
*/
|
|
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|
|
/*
|
|
This program is free software; you can redistribute it and/or modify
|
|
it under the terms of the GNU General Public License as published by
|
|
the Free Software Foundation; version 2 of the License.
|
|
|
|
This program is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
GNU General Public License for more details.
|
|
|
|
NO WARRANTY
|
|
THE PROGRAM IS PROVIDED ON AN "AS IS" BASIS, WITHOUT WARRANTIES OR
|
|
CONDITIONS OF ANY KIND, EITHER EXPRESS OR IMPLIED INCLUDING, WITHOUT
|
|
LIMITATION, ANY WARRANTIES OR CONDITIONS OF TITLE, NON-INFRINGEMENT,
|
|
MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Each Recipient is
|
|
solely responsible for determining the appropriateness of using and
|
|
distributing the Program and assumes all risks associated with its
|
|
exercise of rights under this Agreement, including but not limited to
|
|
the risks and costs of program errors, damage to or loss of data,
|
|
programs or equipment, and unavailability or interruption of operations.
|
|
|
|
DISCLAIMER OF LIABILITY
|
|
NEITHER RECIPIENT NOR ANY CONTRIBUTORS SHALL HAVE ANY LIABILITY FOR ANY
|
|
DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
DAMAGES (INCLUDING WITHOUT LIMITATION LOST PROFITS), HOWEVER CAUSED AND
|
|
ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR
|
|
TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE
|
|
USE OR DISTRIBUTION OF THE PROGRAM OR THE EXERCISE OF ANY RIGHTS GRANTED
|
|
HEREUNDER, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGES
|
|
|
|
You should have received a copy of the GNU General Public License
|
|
along with this program; if not, write to the Free Software
|
|
Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
|
*/
|
|
|
|
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|
|
/*
|
|
* Define statements used for debugging
|
|
*/
|
|
//#define MPT_LAN_IO_DEBUG
|
|
|
|
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|
|
|
|
#include "mptlan.h"
|
|
#include <linux/init.h>
|
|
#include <linux/module.h>
|
|
#include <linux/fs.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/slab.h>
|
|
|
|
#define my_VERSION MPT_LINUX_VERSION_COMMON
|
|
#define MYNAM "mptlan"
|
|
|
|
MODULE_LICENSE("GPL");
|
|
MODULE_VERSION(my_VERSION);
|
|
|
|
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|
|
/*
|
|
* MPT LAN message sizes without variable part.
|
|
*/
|
|
#define MPT_LAN_RECEIVE_POST_REQUEST_SIZE \
|
|
(sizeof(LANReceivePostRequest_t) - sizeof(SGE_MPI_UNION))
|
|
|
|
/*
|
|
* Fusion MPT LAN private structures
|
|
*/
|
|
|
|
struct BufferControl {
|
|
struct sk_buff *skb;
|
|
dma_addr_t dma;
|
|
unsigned int len;
|
|
};
|
|
|
|
struct mpt_lan_priv {
|
|
MPT_ADAPTER *mpt_dev;
|
|
u8 pnum; /* Port number in the IOC. This is not a Unix network port! */
|
|
|
|
atomic_t buckets_out; /* number of unused buckets on IOC */
|
|
int bucketthresh; /* Send more when this many left */
|
|
|
|
int *mpt_txfidx; /* Free Tx Context list */
|
|
int mpt_txfidx_tail;
|
|
spinlock_t txfidx_lock;
|
|
|
|
int *mpt_rxfidx; /* Free Rx Context list */
|
|
int mpt_rxfidx_tail;
|
|
spinlock_t rxfidx_lock;
|
|
|
|
struct BufferControl *RcvCtl; /* Receive BufferControl structs */
|
|
struct BufferControl *SendCtl; /* Send BufferControl structs */
|
|
|
|
int max_buckets_out; /* Max buckets to send to IOC */
|
|
int tx_max_out; /* IOC's Tx queue len */
|
|
|
|
u32 total_posted;
|
|
u32 total_received;
|
|
|
|
struct delayed_work post_buckets_task;
|
|
struct net_device *dev;
|
|
unsigned long post_buckets_active;
|
|
};
|
|
|
|
struct mpt_lan_ohdr {
|
|
u16 dtype;
|
|
u8 daddr[FC_ALEN];
|
|
u16 stype;
|
|
u8 saddr[FC_ALEN];
|
|
};
|
|
|
|
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|
|
|
|
/*
|
|
* Forward protos...
|
|
*/
|
|
static int lan_reply (MPT_ADAPTER *ioc, MPT_FRAME_HDR *mf,
|
|
MPT_FRAME_HDR *reply);
|
|
static int mpt_lan_open(struct net_device *dev);
|
|
static int mpt_lan_reset(struct net_device *dev);
|
|
static int mpt_lan_close(struct net_device *dev);
|
|
static void mpt_lan_post_receive_buckets(struct mpt_lan_priv *priv);
|
|
static void mpt_lan_wake_post_buckets_task(struct net_device *dev,
|
|
int priority);
|
|
static int mpt_lan_receive_post_turbo(struct net_device *dev, u32 tmsg);
|
|
static int mpt_lan_receive_post_reply(struct net_device *dev,
|
|
LANReceivePostReply_t *pRecvRep);
|
|
static int mpt_lan_send_turbo(struct net_device *dev, u32 tmsg);
|
|
static int mpt_lan_send_reply(struct net_device *dev,
|
|
LANSendReply_t *pSendRep);
|
|
static int mpt_lan_ioc_reset(MPT_ADAPTER *ioc, int reset_phase);
|
|
static int mpt_lan_event_process(MPT_ADAPTER *ioc, EventNotificationReply_t *pEvReply);
|
|
static unsigned short mpt_lan_type_trans(struct sk_buff *skb,
|
|
struct net_device *dev);
|
|
|
|
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|
|
/*
|
|
* Fusion MPT LAN private data
|
|
*/
|
|
static u8 LanCtx = MPT_MAX_PROTOCOL_DRIVERS;
|
|
|
|
static u32 max_buckets_out = 127;
|
|
static u32 tx_max_out_p = 127 - 16;
|
|
|
|
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|
|
/**
|
|
* lan_reply - Handle all data sent from the hardware.
|
|
* @ioc: Pointer to MPT_ADAPTER structure
|
|
* @mf: Pointer to original MPT request frame (NULL if TurboReply)
|
|
* @reply: Pointer to MPT reply frame
|
|
*
|
|
* Returns 1 indicating original alloc'd request frame ptr
|
|
* should be freed, or 0 if it shouldn't.
|
|
*/
|
|
static int
|
|
lan_reply (MPT_ADAPTER *ioc, MPT_FRAME_HDR *mf, MPT_FRAME_HDR *reply)
|
|
{
|
|
struct net_device *dev = ioc->netdev;
|
|
int FreeReqFrame = 0;
|
|
|
|
dioprintk((KERN_INFO MYNAM ": %s/%s: Got reply.\n",
|
|
IOC_AND_NETDEV_NAMES_s_s(dev)));
|
|
|
|
// dioprintk((KERN_INFO MYNAM "@lan_reply: mf = %p, reply = %p\n",
|
|
// mf, reply));
|
|
|
|
if (mf == NULL) {
|
|
u32 tmsg = CAST_PTR_TO_U32(reply);
|
|
|
|
dioprintk((KERN_INFO MYNAM ": %s/%s: @lan_reply, tmsg %08x\n",
|
|
IOC_AND_NETDEV_NAMES_s_s(dev),
|
|
tmsg));
|
|
|
|
switch (GET_LAN_FORM(tmsg)) {
|
|
|
|
// NOTE! (Optimization) First case here is now caught in
|
|
// mptbase.c::mpt_interrupt() routine and callcack here
|
|
// is now skipped for this case!
|
|
#if 0
|
|
case LAN_REPLY_FORM_MESSAGE_CONTEXT:
|
|
// dioprintk((KERN_INFO MYNAM "/lan_reply: "
|
|
// "MessageContext turbo reply received\n"));
|
|
FreeReqFrame = 1;
|
|
break;
|
|
#endif
|
|
|
|
case LAN_REPLY_FORM_SEND_SINGLE:
|
|
// dioprintk((MYNAM "/lan_reply: "
|
|
// "calling mpt_lan_send_reply (turbo)\n"));
|
|
|
|
// Potential BUG here?
|
|
// FreeReqFrame = mpt_lan_send_turbo(dev, tmsg);
|
|
// If/when mpt_lan_send_turbo would return 1 here,
|
|
// calling routine (mptbase.c|mpt_interrupt)
|
|
// would Oops because mf has already been set
|
|
// to NULL. So after return from this func,
|
|
// mpt_interrupt() will attempt to put (NULL) mf ptr
|
|
// item back onto its adapter FreeQ - Oops!:-(
|
|
// It's Ok, since mpt_lan_send_turbo() *currently*
|
|
// always returns 0, but..., just in case:
|
|
|
|
(void) mpt_lan_send_turbo(dev, tmsg);
|
|
FreeReqFrame = 0;
|
|
|
|
break;
|
|
|
|
case LAN_REPLY_FORM_RECEIVE_SINGLE:
|
|
// dioprintk((KERN_INFO MYNAM "@lan_reply: "
|
|
// "rcv-Turbo = %08x\n", tmsg));
|
|
mpt_lan_receive_post_turbo(dev, tmsg);
|
|
break;
|
|
|
|
default:
|
|
printk (KERN_ERR MYNAM "/lan_reply: Got a turbo reply "
|
|
"that I don't know what to do with\n");
|
|
|
|
/* CHECKME! Hmmm... FreeReqFrame is 0 here; is that right? */
|
|
|
|
break;
|
|
}
|
|
|
|
return FreeReqFrame;
|
|
}
|
|
|
|
// msg = (u32 *) reply;
|
|
// dioprintk((KERN_INFO MYNAM "@lan_reply: msg = %08x %08x %08x %08x\n",
|
|
// le32_to_cpu(msg[0]), le32_to_cpu(msg[1]),
|
|
// le32_to_cpu(msg[2]), le32_to_cpu(msg[3])));
|
|
// dioprintk((KERN_INFO MYNAM "@lan_reply: Function = %02xh\n",
|
|
// reply->u.hdr.Function));
|
|
|
|
switch (reply->u.hdr.Function) {
|
|
|
|
case MPI_FUNCTION_LAN_SEND:
|
|
{
|
|
LANSendReply_t *pSendRep;
|
|
|
|
pSendRep = (LANSendReply_t *) reply;
|
|
FreeReqFrame = mpt_lan_send_reply(dev, pSendRep);
|
|
break;
|
|
}
|
|
|
|
case MPI_FUNCTION_LAN_RECEIVE:
|
|
{
|
|
LANReceivePostReply_t *pRecvRep;
|
|
|
|
pRecvRep = (LANReceivePostReply_t *) reply;
|
|
if (pRecvRep->NumberOfContexts) {
|
|
mpt_lan_receive_post_reply(dev, pRecvRep);
|
|
if (!(pRecvRep->MsgFlags & MPI_MSGFLAGS_CONTINUATION_REPLY))
|
|
FreeReqFrame = 1;
|
|
} else
|
|
dioprintk((KERN_INFO MYNAM "@lan_reply: zero context "
|
|
"ReceivePostReply received.\n"));
|
|
break;
|
|
}
|
|
|
|
case MPI_FUNCTION_LAN_RESET:
|
|
/* Just a default reply. Might want to check it to
|
|
* make sure that everything went ok.
|
|
*/
|
|
FreeReqFrame = 1;
|
|
break;
|
|
|
|
case MPI_FUNCTION_EVENT_NOTIFICATION:
|
|
case MPI_FUNCTION_EVENT_ACK:
|
|
/* _EVENT_NOTIFICATION should NOT come down this path any more.
|
|
* Should be routed to mpt_lan_event_process(), but just in case...
|
|
*/
|
|
FreeReqFrame = 1;
|
|
break;
|
|
|
|
default:
|
|
printk (KERN_ERR MYNAM "/lan_reply: Got a non-turbo "
|
|
"reply that I don't know what to do with\n");
|
|
|
|
/* CHECKME! Hmmm... FreeReqFrame is 0 here; is that right? */
|
|
FreeReqFrame = 1;
|
|
|
|
break;
|
|
}
|
|
|
|
return FreeReqFrame;
|
|
}
|
|
|
|
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|
|
static int
|
|
mpt_lan_ioc_reset(MPT_ADAPTER *ioc, int reset_phase)
|
|
{
|
|
struct net_device *dev = ioc->netdev;
|
|
struct mpt_lan_priv *priv;
|
|
|
|
if (dev == NULL)
|
|
return(1);
|
|
else
|
|
priv = netdev_priv(dev);
|
|
|
|
dlprintk((KERN_INFO MYNAM ": IOC %s_reset routed to LAN driver!\n",
|
|
reset_phase==MPT_IOC_SETUP_RESET ? "setup" : (
|
|
reset_phase==MPT_IOC_PRE_RESET ? "pre" : "post")));
|
|
|
|
if (priv->mpt_rxfidx == NULL)
|
|
return (1);
|
|
|
|
if (reset_phase == MPT_IOC_SETUP_RESET) {
|
|
;
|
|
} else if (reset_phase == MPT_IOC_PRE_RESET) {
|
|
int i;
|
|
unsigned long flags;
|
|
|
|
netif_stop_queue(dev);
|
|
|
|
dlprintk ((KERN_INFO "mptlan/ioc_reset: called netif_stop_queue for %s.\n", dev->name));
|
|
|
|
atomic_set(&priv->buckets_out, 0);
|
|
|
|
/* Reset Rx Free Tail index and re-populate the queue. */
|
|
spin_lock_irqsave(&priv->rxfidx_lock, flags);
|
|
priv->mpt_rxfidx_tail = -1;
|
|
for (i = 0; i < priv->max_buckets_out; i++)
|
|
priv->mpt_rxfidx[++priv->mpt_rxfidx_tail] = i;
|
|
spin_unlock_irqrestore(&priv->rxfidx_lock, flags);
|
|
} else {
|
|
mpt_lan_post_receive_buckets(priv);
|
|
netif_wake_queue(dev);
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|
|
static int
|
|
mpt_lan_event_process(MPT_ADAPTER *ioc, EventNotificationReply_t *pEvReply)
|
|
{
|
|
dlprintk((KERN_INFO MYNAM ": MPT event routed to LAN driver!\n"));
|
|
|
|
switch (le32_to_cpu(pEvReply->Event)) {
|
|
case MPI_EVENT_NONE: /* 00 */
|
|
case MPI_EVENT_LOG_DATA: /* 01 */
|
|
case MPI_EVENT_STATE_CHANGE: /* 02 */
|
|
case MPI_EVENT_UNIT_ATTENTION: /* 03 */
|
|
case MPI_EVENT_IOC_BUS_RESET: /* 04 */
|
|
case MPI_EVENT_EXT_BUS_RESET: /* 05 */
|
|
case MPI_EVENT_RESCAN: /* 06 */
|
|
/* Ok, do we need to do anything here? As far as
|
|
I can tell, this is when a new device gets added
|
|
to the loop. */
|
|
case MPI_EVENT_LINK_STATUS_CHANGE: /* 07 */
|
|
case MPI_EVENT_LOOP_STATE_CHANGE: /* 08 */
|
|
case MPI_EVENT_LOGOUT: /* 09 */
|
|
case MPI_EVENT_EVENT_CHANGE: /* 0A */
|
|
default:
|
|
break;
|
|
}
|
|
|
|
/*
|
|
* NOTE: pEvent->AckRequired handling now done in mptbase.c;
|
|
* Do NOT do it here now!
|
|
*/
|
|
|
|
return 1;
|
|
}
|
|
|
|
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|
|
static int
|
|
mpt_lan_open(struct net_device *dev)
|
|
{
|
|
struct mpt_lan_priv *priv = netdev_priv(dev);
|
|
int i;
|
|
|
|
if (mpt_lan_reset(dev) != 0) {
|
|
MPT_ADAPTER *mpt_dev = priv->mpt_dev;
|
|
|
|
printk (KERN_WARNING MYNAM "/lan_open: lan_reset failed.");
|
|
|
|
if (mpt_dev->active)
|
|
printk ("The ioc is active. Perhaps it needs to be"
|
|
" reset?\n");
|
|
else
|
|
printk ("The ioc in inactive, most likely in the "
|
|
"process of being reset. Please try again in "
|
|
"a moment.\n");
|
|
}
|
|
|
|
priv->mpt_txfidx = kmalloc_array(priv->tx_max_out, sizeof(int),
|
|
GFP_KERNEL);
|
|
if (priv->mpt_txfidx == NULL)
|
|
goto out;
|
|
priv->mpt_txfidx_tail = -1;
|
|
|
|
priv->SendCtl = kcalloc(priv->tx_max_out, sizeof(struct BufferControl),
|
|
GFP_KERNEL);
|
|
if (priv->SendCtl == NULL)
|
|
goto out_mpt_txfidx;
|
|
for (i = 0; i < priv->tx_max_out; i++)
|
|
priv->mpt_txfidx[++priv->mpt_txfidx_tail] = i;
|
|
|
|
dlprintk((KERN_INFO MYNAM "@lo: Finished initializing SendCtl\n"));
|
|
|
|
priv->mpt_rxfidx = kmalloc_array(priv->max_buckets_out, sizeof(int),
|
|
GFP_KERNEL);
|
|
if (priv->mpt_rxfidx == NULL)
|
|
goto out_SendCtl;
|
|
priv->mpt_rxfidx_tail = -1;
|
|
|
|
priv->RcvCtl = kcalloc(priv->max_buckets_out,
|
|
sizeof(struct BufferControl),
|
|
GFP_KERNEL);
|
|
if (priv->RcvCtl == NULL)
|
|
goto out_mpt_rxfidx;
|
|
for (i = 0; i < priv->max_buckets_out; i++)
|
|
priv->mpt_rxfidx[++priv->mpt_rxfidx_tail] = i;
|
|
|
|
/**/ dlprintk((KERN_INFO MYNAM "/lo: txfidx contains - "));
|
|
/**/ for (i = 0; i < priv->tx_max_out; i++)
|
|
/**/ dlprintk((" %xh", priv->mpt_txfidx[i]));
|
|
/**/ dlprintk(("\n"));
|
|
|
|
dlprintk((KERN_INFO MYNAM "/lo: Finished initializing RcvCtl\n"));
|
|
|
|
mpt_lan_post_receive_buckets(priv);
|
|
printk(KERN_INFO MYNAM ": %s/%s: interface up & active\n",
|
|
IOC_AND_NETDEV_NAMES_s_s(dev));
|
|
|
|
if (mpt_event_register(LanCtx, mpt_lan_event_process) != 0) {
|
|
printk (KERN_WARNING MYNAM "/lo: Unable to register for Event"
|
|
" Notifications. This is a bad thing! We're not going "
|
|
"to go ahead, but I'd be leery of system stability at "
|
|
"this point.\n");
|
|
}
|
|
|
|
netif_start_queue(dev);
|
|
dlprintk((KERN_INFO MYNAM "/lo: Done.\n"));
|
|
|
|
return 0;
|
|
out_mpt_rxfidx:
|
|
kfree(priv->mpt_rxfidx);
|
|
priv->mpt_rxfidx = NULL;
|
|
out_SendCtl:
|
|
kfree(priv->SendCtl);
|
|
priv->SendCtl = NULL;
|
|
out_mpt_txfidx:
|
|
kfree(priv->mpt_txfidx);
|
|
priv->mpt_txfidx = NULL;
|
|
out: return -ENOMEM;
|
|
}
|
|
|
|
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|
|
/* Send a LanReset message to the FW. This should result in the FW returning
|
|
any buckets it still has. */
|
|
static int
|
|
mpt_lan_reset(struct net_device *dev)
|
|
{
|
|
MPT_FRAME_HDR *mf;
|
|
LANResetRequest_t *pResetReq;
|
|
struct mpt_lan_priv *priv = netdev_priv(dev);
|
|
|
|
mf = mpt_get_msg_frame(LanCtx, priv->mpt_dev);
|
|
|
|
if (mf == NULL) {
|
|
/* dlprintk((KERN_ERR MYNAM "/reset: Evil funkiness abounds! "
|
|
"Unable to allocate a request frame.\n"));
|
|
*/
|
|
return -1;
|
|
}
|
|
|
|
pResetReq = (LANResetRequest_t *) mf;
|
|
|
|
pResetReq->Function = MPI_FUNCTION_LAN_RESET;
|
|
pResetReq->ChainOffset = 0;
|
|
pResetReq->Reserved = 0;
|
|
pResetReq->PortNumber = priv->pnum;
|
|
pResetReq->MsgFlags = 0;
|
|
pResetReq->Reserved2 = 0;
|
|
|
|
mpt_put_msg_frame(LanCtx, priv->mpt_dev, mf);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|
|
static int
|
|
mpt_lan_close(struct net_device *dev)
|
|
{
|
|
struct mpt_lan_priv *priv = netdev_priv(dev);
|
|
MPT_ADAPTER *mpt_dev = priv->mpt_dev;
|
|
unsigned long timeout;
|
|
int i;
|
|
|
|
dlprintk((KERN_INFO MYNAM ": mpt_lan_close called\n"));
|
|
|
|
mpt_event_deregister(LanCtx);
|
|
|
|
dlprintk((KERN_INFO MYNAM ":lan_close: Posted %d buckets "
|
|
"since driver was loaded, %d still out\n",
|
|
priv->total_posted,atomic_read(&priv->buckets_out)));
|
|
|
|
netif_stop_queue(dev);
|
|
|
|
mpt_lan_reset(dev);
|
|
|
|
timeout = jiffies + 2 * HZ;
|
|
while (atomic_read(&priv->buckets_out) && time_before(jiffies, timeout))
|
|
schedule_timeout_interruptible(1);
|
|
|
|
for (i = 0; i < priv->max_buckets_out; i++) {
|
|
if (priv->RcvCtl[i].skb != NULL) {
|
|
/**/ dlprintk((KERN_INFO MYNAM "/lan_close: bucket %05x "
|
|
/**/ "is still out\n", i));
|
|
dma_unmap_single(&mpt_dev->pcidev->dev,
|
|
priv->RcvCtl[i].dma,
|
|
priv->RcvCtl[i].len, DMA_FROM_DEVICE);
|
|
dev_kfree_skb(priv->RcvCtl[i].skb);
|
|
}
|
|
}
|
|
|
|
kfree(priv->RcvCtl);
|
|
kfree(priv->mpt_rxfidx);
|
|
|
|
for (i = 0; i < priv->tx_max_out; i++) {
|
|
if (priv->SendCtl[i].skb != NULL) {
|
|
dma_unmap_single(&mpt_dev->pcidev->dev,
|
|
priv->SendCtl[i].dma,
|
|
priv->SendCtl[i].len, DMA_TO_DEVICE);
|
|
dev_kfree_skb(priv->SendCtl[i].skb);
|
|
}
|
|
}
|
|
|
|
kfree(priv->SendCtl);
|
|
kfree(priv->mpt_txfidx);
|
|
|
|
atomic_set(&priv->buckets_out, 0);
|
|
|
|
printk(KERN_INFO MYNAM ": %s/%s: interface down & inactive\n",
|
|
IOC_AND_NETDEV_NAMES_s_s(dev));
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|
|
/* Tx timeout handler. */
|
|
static void
|
|
mpt_lan_tx_timeout(struct net_device *dev, unsigned int txqueue)
|
|
{
|
|
struct mpt_lan_priv *priv = netdev_priv(dev);
|
|
MPT_ADAPTER *mpt_dev = priv->mpt_dev;
|
|
|
|
if (mpt_dev->active) {
|
|
dlprintk (("mptlan/tx_timeout: calling netif_wake_queue for %s.\n", dev->name));
|
|
netif_wake_queue(dev);
|
|
}
|
|
}
|
|
|
|
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|
|
//static inline int
|
|
static int
|
|
mpt_lan_send_turbo(struct net_device *dev, u32 tmsg)
|
|
{
|
|
struct mpt_lan_priv *priv = netdev_priv(dev);
|
|
MPT_ADAPTER *mpt_dev = priv->mpt_dev;
|
|
struct sk_buff *sent;
|
|
unsigned long flags;
|
|
u32 ctx;
|
|
|
|
ctx = GET_LAN_BUFFER_CONTEXT(tmsg);
|
|
sent = priv->SendCtl[ctx].skb;
|
|
|
|
dev->stats.tx_packets++;
|
|
dev->stats.tx_bytes += sent->len;
|
|
|
|
dioprintk((KERN_INFO MYNAM ": %s/%s: @%s, skb %p sent.\n",
|
|
IOC_AND_NETDEV_NAMES_s_s(dev),
|
|
__func__, sent));
|
|
|
|
priv->SendCtl[ctx].skb = NULL;
|
|
dma_unmap_single(&mpt_dev->pcidev->dev, priv->SendCtl[ctx].dma,
|
|
priv->SendCtl[ctx].len, DMA_TO_DEVICE);
|
|
dev_kfree_skb_irq(sent);
|
|
|
|
spin_lock_irqsave(&priv->txfidx_lock, flags);
|
|
priv->mpt_txfidx[++priv->mpt_txfidx_tail] = ctx;
|
|
spin_unlock_irqrestore(&priv->txfidx_lock, flags);
|
|
|
|
netif_wake_queue(dev);
|
|
return 0;
|
|
}
|
|
|
|
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|
|
static int
|
|
mpt_lan_send_reply(struct net_device *dev, LANSendReply_t *pSendRep)
|
|
{
|
|
struct mpt_lan_priv *priv = netdev_priv(dev);
|
|
MPT_ADAPTER *mpt_dev = priv->mpt_dev;
|
|
struct sk_buff *sent;
|
|
unsigned long flags;
|
|
int FreeReqFrame = 0;
|
|
u32 *pContext;
|
|
u32 ctx;
|
|
u8 count;
|
|
|
|
count = pSendRep->NumberOfContexts;
|
|
|
|
dioprintk((KERN_INFO MYNAM ": send_reply: IOCStatus: %04x\n",
|
|
le16_to_cpu(pSendRep->IOCStatus)));
|
|
|
|
/* Add check for Loginfo Flag in IOCStatus */
|
|
|
|
switch (le16_to_cpu(pSendRep->IOCStatus) & MPI_IOCSTATUS_MASK) {
|
|
case MPI_IOCSTATUS_SUCCESS:
|
|
dev->stats.tx_packets += count;
|
|
break;
|
|
|
|
case MPI_IOCSTATUS_LAN_CANCELED:
|
|
case MPI_IOCSTATUS_LAN_TRANSMIT_ABORTED:
|
|
break;
|
|
|
|
case MPI_IOCSTATUS_INVALID_SGL:
|
|
dev->stats.tx_errors += count;
|
|
printk (KERN_ERR MYNAM ": %s/%s: ERROR - Invalid SGL sent to IOC!\n",
|
|
IOC_AND_NETDEV_NAMES_s_s(dev));
|
|
goto out;
|
|
|
|
default:
|
|
dev->stats.tx_errors += count;
|
|
break;
|
|
}
|
|
|
|
pContext = &pSendRep->BufferContext;
|
|
|
|
spin_lock_irqsave(&priv->txfidx_lock, flags);
|
|
while (count > 0) {
|
|
ctx = GET_LAN_BUFFER_CONTEXT(le32_to_cpu(*pContext));
|
|
|
|
sent = priv->SendCtl[ctx].skb;
|
|
dev->stats.tx_bytes += sent->len;
|
|
|
|
dioprintk((KERN_INFO MYNAM ": %s/%s: @%s, skb %p sent.\n",
|
|
IOC_AND_NETDEV_NAMES_s_s(dev),
|
|
__func__, sent));
|
|
|
|
priv->SendCtl[ctx].skb = NULL;
|
|
dma_unmap_single(&mpt_dev->pcidev->dev,
|
|
priv->SendCtl[ctx].dma,
|
|
priv->SendCtl[ctx].len, DMA_TO_DEVICE);
|
|
dev_kfree_skb_irq(sent);
|
|
|
|
priv->mpt_txfidx[++priv->mpt_txfidx_tail] = ctx;
|
|
|
|
pContext++;
|
|
count--;
|
|
}
|
|
spin_unlock_irqrestore(&priv->txfidx_lock, flags);
|
|
|
|
out:
|
|
if (!(pSendRep->MsgFlags & MPI_MSGFLAGS_CONTINUATION_REPLY))
|
|
FreeReqFrame = 1;
|
|
|
|
netif_wake_queue(dev);
|
|
return FreeReqFrame;
|
|
}
|
|
|
|
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|
|
static netdev_tx_t
|
|
mpt_lan_sdu_send (struct sk_buff *skb, struct net_device *dev)
|
|
{
|
|
struct mpt_lan_priv *priv = netdev_priv(dev);
|
|
MPT_ADAPTER *mpt_dev = priv->mpt_dev;
|
|
MPT_FRAME_HDR *mf;
|
|
LANSendRequest_t *pSendReq;
|
|
SGETransaction32_t *pTrans;
|
|
SGESimple64_t *pSimple;
|
|
const unsigned char *mac;
|
|
dma_addr_t dma;
|
|
unsigned long flags;
|
|
int ctx;
|
|
u16 cur_naa = 0x1000;
|
|
|
|
dioprintk((KERN_INFO MYNAM ": %s called, skb_addr = %p\n",
|
|
__func__, skb));
|
|
|
|
spin_lock_irqsave(&priv->txfidx_lock, flags);
|
|
if (priv->mpt_txfidx_tail < 0) {
|
|
netif_stop_queue(dev);
|
|
spin_unlock_irqrestore(&priv->txfidx_lock, flags);
|
|
|
|
printk (KERN_ERR "%s: no tx context available: %u\n",
|
|
__func__, priv->mpt_txfidx_tail);
|
|
return NETDEV_TX_BUSY;
|
|
}
|
|
|
|
mf = mpt_get_msg_frame(LanCtx, mpt_dev);
|
|
if (mf == NULL) {
|
|
netif_stop_queue(dev);
|
|
spin_unlock_irqrestore(&priv->txfidx_lock, flags);
|
|
|
|
printk (KERN_ERR "%s: Unable to alloc request frame\n",
|
|
__func__);
|
|
return NETDEV_TX_BUSY;
|
|
}
|
|
|
|
ctx = priv->mpt_txfidx[priv->mpt_txfidx_tail--];
|
|
spin_unlock_irqrestore(&priv->txfidx_lock, flags);
|
|
|
|
// dioprintk((KERN_INFO MYNAM ": %s/%s: Creating new msg frame (send).\n",
|
|
// IOC_AND_NETDEV_NAMES_s_s(dev)));
|
|
|
|
pSendReq = (LANSendRequest_t *) mf;
|
|
|
|
/* Set the mac.raw pointer, since this apparently isn't getting
|
|
* done before we get the skb. Pull the data pointer past the mac data.
|
|
*/
|
|
skb_reset_mac_header(skb);
|
|
skb_pull(skb, 12);
|
|
|
|
dma = dma_map_single(&mpt_dev->pcidev->dev, skb->data, skb->len,
|
|
DMA_TO_DEVICE);
|
|
|
|
priv->SendCtl[ctx].skb = skb;
|
|
priv->SendCtl[ctx].dma = dma;
|
|
priv->SendCtl[ctx].len = skb->len;
|
|
|
|
/* Message Header */
|
|
pSendReq->Reserved = 0;
|
|
pSendReq->Function = MPI_FUNCTION_LAN_SEND;
|
|
pSendReq->ChainOffset = 0;
|
|
pSendReq->Reserved2 = 0;
|
|
pSendReq->MsgFlags = 0;
|
|
pSendReq->PortNumber = priv->pnum;
|
|
|
|
/* Transaction Context Element */
|
|
pTrans = (SGETransaction32_t *) pSendReq->SG_List;
|
|
|
|
/* No Flags, 8 bytes of Details, 32bit Context (bloody turbo replies) */
|
|
pTrans->ContextSize = sizeof(u32);
|
|
pTrans->DetailsLength = 2 * sizeof(u32);
|
|
pTrans->Flags = 0;
|
|
pTrans->TransactionContext = cpu_to_le32(ctx);
|
|
|
|
// dioprintk((KERN_INFO MYNAM ": %s/%s: BC = %08x, skb = %p, buff = %p\n",
|
|
// IOC_AND_NETDEV_NAMES_s_s(dev),
|
|
// ctx, skb, skb->data));
|
|
|
|
mac = skb_mac_header(skb);
|
|
|
|
pTrans->TransactionDetails[0] = cpu_to_le32((cur_naa << 16) |
|
|
(mac[0] << 8) |
|
|
(mac[1] << 0));
|
|
pTrans->TransactionDetails[1] = cpu_to_le32((mac[2] << 24) |
|
|
(mac[3] << 16) |
|
|
(mac[4] << 8) |
|
|
(mac[5] << 0));
|
|
|
|
pSimple = (SGESimple64_t *) &pTrans->TransactionDetails[2];
|
|
|
|
/* If we ever decide to send more than one Simple SGE per LANSend, then
|
|
we will need to make sure that LAST_ELEMENT only gets set on the
|
|
last one. Otherwise, bad voodoo and evil funkiness will commence. */
|
|
pSimple->FlagsLength = cpu_to_le32(
|
|
((MPI_SGE_FLAGS_LAST_ELEMENT |
|
|
MPI_SGE_FLAGS_END_OF_BUFFER |
|
|
MPI_SGE_FLAGS_SIMPLE_ELEMENT |
|
|
MPI_SGE_FLAGS_SYSTEM_ADDRESS |
|
|
MPI_SGE_FLAGS_HOST_TO_IOC |
|
|
MPI_SGE_FLAGS_64_BIT_ADDRESSING |
|
|
MPI_SGE_FLAGS_END_OF_LIST) << MPI_SGE_FLAGS_SHIFT) |
|
|
skb->len);
|
|
pSimple->Address.Low = cpu_to_le32((u32) dma);
|
|
if (sizeof(dma_addr_t) > sizeof(u32))
|
|
pSimple->Address.High = cpu_to_le32((u32) ((u64) dma >> 32));
|
|
else
|
|
pSimple->Address.High = 0;
|
|
|
|
mpt_put_msg_frame (LanCtx, mpt_dev, mf);
|
|
netif_trans_update(dev);
|
|
|
|
dioprintk((KERN_INFO MYNAM ": %s/%s: Sending packet. FlagsLength = %08x.\n",
|
|
IOC_AND_NETDEV_NAMES_s_s(dev),
|
|
le32_to_cpu(pSimple->FlagsLength)));
|
|
|
|
return NETDEV_TX_OK;
|
|
}
|
|
|
|
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|
|
static void
|
|
mpt_lan_wake_post_buckets_task(struct net_device *dev, int priority)
|
|
/*
|
|
* @priority: 0 = put it on the timer queue, 1 = put it on the immediate queue
|
|
*/
|
|
{
|
|
struct mpt_lan_priv *priv = netdev_priv(dev);
|
|
|
|
if (test_and_set_bit(0, &priv->post_buckets_active) == 0) {
|
|
if (priority) {
|
|
schedule_delayed_work(&priv->post_buckets_task, 0);
|
|
} else {
|
|
schedule_delayed_work(&priv->post_buckets_task, 1);
|
|
dioprintk((KERN_INFO MYNAM ": post_buckets queued on "
|
|
"timer.\n"));
|
|
}
|
|
dioprintk((KERN_INFO MYNAM ": %s/%s: Queued post_buckets task.\n",
|
|
IOC_AND_NETDEV_NAMES_s_s(dev) ));
|
|
}
|
|
}
|
|
|
|
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|
|
static int
|
|
mpt_lan_receive_skb(struct net_device *dev, struct sk_buff *skb)
|
|
{
|
|
struct mpt_lan_priv *priv = netdev_priv(dev);
|
|
|
|
skb->protocol = mpt_lan_type_trans(skb, dev);
|
|
|
|
dioprintk((KERN_INFO MYNAM ": %s/%s: Incoming packet (%d bytes) "
|
|
"delivered to upper level.\n",
|
|
IOC_AND_NETDEV_NAMES_s_s(dev), skb->len));
|
|
|
|
dev->stats.rx_bytes += skb->len;
|
|
dev->stats.rx_packets++;
|
|
|
|
skb->dev = dev;
|
|
netif_rx(skb);
|
|
|
|
dioprintk((MYNAM "/receive_skb: %d buckets remaining\n",
|
|
atomic_read(&priv->buckets_out)));
|
|
|
|
if (atomic_read(&priv->buckets_out) < priv->bucketthresh)
|
|
mpt_lan_wake_post_buckets_task(dev, 1);
|
|
|
|
dioprintk((KERN_INFO MYNAM "/receive_post_reply: %d buckets "
|
|
"remaining, %d received back since sod\n",
|
|
atomic_read(&priv->buckets_out), priv->total_received));
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|
|
//static inline int
|
|
static int
|
|
mpt_lan_receive_post_turbo(struct net_device *dev, u32 tmsg)
|
|
{
|
|
struct mpt_lan_priv *priv = netdev_priv(dev);
|
|
MPT_ADAPTER *mpt_dev = priv->mpt_dev;
|
|
struct sk_buff *skb, *old_skb;
|
|
unsigned long flags;
|
|
u32 ctx, len;
|
|
|
|
ctx = GET_LAN_BUCKET_CONTEXT(tmsg);
|
|
skb = priv->RcvCtl[ctx].skb;
|
|
|
|
len = GET_LAN_PACKET_LENGTH(tmsg);
|
|
|
|
if (len < MPT_LAN_RX_COPYBREAK) {
|
|
old_skb = skb;
|
|
|
|
skb = (struct sk_buff *)dev_alloc_skb(len);
|
|
if (!skb) {
|
|
printk (KERN_ERR MYNAM ": %s/%s: ERROR - Can't allocate skb! (%s@%d)\n",
|
|
IOC_AND_NETDEV_NAMES_s_s(dev),
|
|
__FILE__, __LINE__);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
dma_sync_single_for_cpu(&mpt_dev->pcidev->dev,
|
|
priv->RcvCtl[ctx].dma,
|
|
priv->RcvCtl[ctx].len,
|
|
DMA_FROM_DEVICE);
|
|
|
|
skb_copy_from_linear_data(old_skb, skb_put(skb, len), len);
|
|
|
|
dma_sync_single_for_device(&mpt_dev->pcidev->dev,
|
|
priv->RcvCtl[ctx].dma,
|
|
priv->RcvCtl[ctx].len,
|
|
DMA_FROM_DEVICE);
|
|
goto out;
|
|
}
|
|
|
|
skb_put(skb, len);
|
|
|
|
priv->RcvCtl[ctx].skb = NULL;
|
|
|
|
dma_unmap_single(&mpt_dev->pcidev->dev, priv->RcvCtl[ctx].dma,
|
|
priv->RcvCtl[ctx].len, DMA_FROM_DEVICE);
|
|
|
|
out:
|
|
spin_lock_irqsave(&priv->rxfidx_lock, flags);
|
|
priv->mpt_rxfidx[++priv->mpt_rxfidx_tail] = ctx;
|
|
spin_unlock_irqrestore(&priv->rxfidx_lock, flags);
|
|
|
|
atomic_dec(&priv->buckets_out);
|
|
priv->total_received++;
|
|
|
|
return mpt_lan_receive_skb(dev, skb);
|
|
}
|
|
|
|
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|
|
static int
|
|
mpt_lan_receive_post_free(struct net_device *dev,
|
|
LANReceivePostReply_t *pRecvRep)
|
|
{
|
|
struct mpt_lan_priv *priv = netdev_priv(dev);
|
|
MPT_ADAPTER *mpt_dev = priv->mpt_dev;
|
|
unsigned long flags;
|
|
struct sk_buff *skb;
|
|
u32 ctx;
|
|
int count;
|
|
int i;
|
|
|
|
count = pRecvRep->NumberOfContexts;
|
|
|
|
/**/ dlprintk((KERN_INFO MYNAM "/receive_post_reply: "
|
|
"IOC returned %d buckets, freeing them...\n", count));
|
|
|
|
spin_lock_irqsave(&priv->rxfidx_lock, flags);
|
|
for (i = 0; i < count; i++) {
|
|
ctx = le32_to_cpu(pRecvRep->BucketContext[i]);
|
|
|
|
skb = priv->RcvCtl[ctx].skb;
|
|
|
|
// dlprintk((KERN_INFO MYNAM ": %s: dev_name = %s\n",
|
|
// IOC_AND_NETDEV_NAMES_s_s(dev)));
|
|
// dlprintk((KERN_INFO MYNAM "@rpr[2], priv = %p, buckets_out addr = %p",
|
|
// priv, &(priv->buckets_out)));
|
|
// dlprintk((KERN_INFO MYNAM "@rpr[2] TC + 3\n"));
|
|
|
|
priv->RcvCtl[ctx].skb = NULL;
|
|
dma_unmap_single(&mpt_dev->pcidev->dev, priv->RcvCtl[ctx].dma,
|
|
priv->RcvCtl[ctx].len, DMA_FROM_DEVICE);
|
|
dev_kfree_skb_any(skb);
|
|
|
|
priv->mpt_rxfidx[++priv->mpt_rxfidx_tail] = ctx;
|
|
}
|
|
spin_unlock_irqrestore(&priv->rxfidx_lock, flags);
|
|
|
|
atomic_sub(count, &priv->buckets_out);
|
|
|
|
// for (i = 0; i < priv->max_buckets_out; i++)
|
|
// if (priv->RcvCtl[i].skb != NULL)
|
|
// dlprintk((KERN_INFO MYNAM "@rpr: bucket %03x "
|
|
// "is still out\n", i));
|
|
|
|
/* dlprintk((KERN_INFO MYNAM "/receive_post_reply: freed %d buckets\n",
|
|
count));
|
|
*/
|
|
/**/ dlprintk((KERN_INFO MYNAM "@receive_post_reply: %d buckets "
|
|
/**/ "remaining, %d received back since sod.\n",
|
|
/**/ atomic_read(&priv->buckets_out), priv->total_received));
|
|
return 0;
|
|
}
|
|
|
|
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|
|
static int
|
|
mpt_lan_receive_post_reply(struct net_device *dev,
|
|
LANReceivePostReply_t *pRecvRep)
|
|
{
|
|
struct mpt_lan_priv *priv = netdev_priv(dev);
|
|
MPT_ADAPTER *mpt_dev = priv->mpt_dev;
|
|
struct sk_buff *skb, *old_skb;
|
|
unsigned long flags;
|
|
u32 len, ctx, offset;
|
|
u32 remaining = le32_to_cpu(pRecvRep->BucketsRemaining);
|
|
int count;
|
|
int i, l;
|
|
|
|
dioprintk((KERN_INFO MYNAM ": mpt_lan_receive_post_reply called\n"));
|
|
dioprintk((KERN_INFO MYNAM ": receive_post_reply: IOCStatus: %04x\n",
|
|
le16_to_cpu(pRecvRep->IOCStatus)));
|
|
|
|
if ((le16_to_cpu(pRecvRep->IOCStatus) & MPI_IOCSTATUS_MASK) ==
|
|
MPI_IOCSTATUS_LAN_CANCELED)
|
|
return mpt_lan_receive_post_free(dev, pRecvRep);
|
|
|
|
len = le32_to_cpu(pRecvRep->PacketLength);
|
|
if (len == 0) {
|
|
printk (KERN_ERR MYNAM ": %s/%s: ERROR - Got a non-TURBO "
|
|
"ReceivePostReply w/ PacketLength zero!\n",
|
|
IOC_AND_NETDEV_NAMES_s_s(dev));
|
|
printk (KERN_ERR MYNAM ": MsgFlags = %02x, IOCStatus = %04x\n",
|
|
pRecvRep->MsgFlags, le16_to_cpu(pRecvRep->IOCStatus));
|
|
return -1;
|
|
}
|
|
|
|
ctx = le32_to_cpu(pRecvRep->BucketContext[0]);
|
|
count = pRecvRep->NumberOfContexts;
|
|
skb = priv->RcvCtl[ctx].skb;
|
|
|
|
offset = le32_to_cpu(pRecvRep->PacketOffset);
|
|
// if (offset != 0) {
|
|
// printk (KERN_INFO MYNAM ": %s/%s: Got a ReceivePostReply "
|
|
// "w/ PacketOffset %u\n",
|
|
// IOC_AND_NETDEV_NAMES_s_s(dev),
|
|
// offset);
|
|
// }
|
|
|
|
dioprintk((KERN_INFO MYNAM ": %s/%s: @rpr, offset = %d, len = %d\n",
|
|
IOC_AND_NETDEV_NAMES_s_s(dev),
|
|
offset, len));
|
|
|
|
if (count > 1) {
|
|
int szrem = len;
|
|
|
|
// dioprintk((KERN_INFO MYNAM ": %s/%s: Multiple buckets returned "
|
|
// "for single packet, concatenating...\n",
|
|
// IOC_AND_NETDEV_NAMES_s_s(dev)));
|
|
|
|
skb = (struct sk_buff *)dev_alloc_skb(len);
|
|
if (!skb) {
|
|
printk (KERN_ERR MYNAM ": %s/%s: ERROR - Can't allocate skb! (%s@%d)\n",
|
|
IOC_AND_NETDEV_NAMES_s_s(dev),
|
|
__FILE__, __LINE__);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
spin_lock_irqsave(&priv->rxfidx_lock, flags);
|
|
for (i = 0; i < count; i++) {
|
|
|
|
ctx = le32_to_cpu(pRecvRep->BucketContext[i]);
|
|
old_skb = priv->RcvCtl[ctx].skb;
|
|
|
|
l = priv->RcvCtl[ctx].len;
|
|
if (szrem < l)
|
|
l = szrem;
|
|
|
|
// dioprintk((KERN_INFO MYNAM ": %s/%s: Buckets = %d, len = %u\n",
|
|
// IOC_AND_NETDEV_NAMES_s_s(dev),
|
|
// i, l));
|
|
|
|
dma_sync_single_for_cpu(&mpt_dev->pcidev->dev,
|
|
priv->RcvCtl[ctx].dma,
|
|
priv->RcvCtl[ctx].len,
|
|
DMA_FROM_DEVICE);
|
|
skb_copy_from_linear_data(old_skb, skb_put(skb, l), l);
|
|
|
|
dma_sync_single_for_device(&mpt_dev->pcidev->dev,
|
|
priv->RcvCtl[ctx].dma,
|
|
priv->RcvCtl[ctx].len,
|
|
DMA_FROM_DEVICE);
|
|
|
|
priv->mpt_rxfidx[++priv->mpt_rxfidx_tail] = ctx;
|
|
szrem -= l;
|
|
}
|
|
spin_unlock_irqrestore(&priv->rxfidx_lock, flags);
|
|
|
|
} else if (len < MPT_LAN_RX_COPYBREAK) {
|
|
|
|
old_skb = skb;
|
|
|
|
skb = (struct sk_buff *)dev_alloc_skb(len);
|
|
if (!skb) {
|
|
printk (KERN_ERR MYNAM ": %s/%s: ERROR - Can't allocate skb! (%s@%d)\n",
|
|
IOC_AND_NETDEV_NAMES_s_s(dev),
|
|
__FILE__, __LINE__);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
dma_sync_single_for_cpu(&mpt_dev->pcidev->dev,
|
|
priv->RcvCtl[ctx].dma,
|
|
priv->RcvCtl[ctx].len,
|
|
DMA_FROM_DEVICE);
|
|
|
|
skb_copy_from_linear_data(old_skb, skb_put(skb, len), len);
|
|
|
|
dma_sync_single_for_device(&mpt_dev->pcidev->dev,
|
|
priv->RcvCtl[ctx].dma,
|
|
priv->RcvCtl[ctx].len,
|
|
DMA_FROM_DEVICE);
|
|
|
|
spin_lock_irqsave(&priv->rxfidx_lock, flags);
|
|
priv->mpt_rxfidx[++priv->mpt_rxfidx_tail] = ctx;
|
|
spin_unlock_irqrestore(&priv->rxfidx_lock, flags);
|
|
|
|
} else {
|
|
spin_lock_irqsave(&priv->rxfidx_lock, flags);
|
|
|
|
priv->RcvCtl[ctx].skb = NULL;
|
|
|
|
dma_unmap_single(&mpt_dev->pcidev->dev, priv->RcvCtl[ctx].dma,
|
|
priv->RcvCtl[ctx].len, DMA_FROM_DEVICE);
|
|
priv->RcvCtl[ctx].dma = 0;
|
|
|
|
priv->mpt_rxfidx[++priv->mpt_rxfidx_tail] = ctx;
|
|
spin_unlock_irqrestore(&priv->rxfidx_lock, flags);
|
|
|
|
skb_put(skb,len);
|
|
}
|
|
|
|
atomic_sub(count, &priv->buckets_out);
|
|
priv->total_received += count;
|
|
|
|
if (priv->mpt_rxfidx_tail >= MPT_LAN_MAX_BUCKETS_OUT) {
|
|
printk (KERN_ERR MYNAM ": %s/%s: Yoohoo! mpt_rxfidx_tail = %d, "
|
|
"MPT_LAN_MAX_BUCKETS_OUT = %d\n",
|
|
IOC_AND_NETDEV_NAMES_s_s(dev),
|
|
priv->mpt_rxfidx_tail,
|
|
MPT_LAN_MAX_BUCKETS_OUT);
|
|
|
|
return -1;
|
|
}
|
|
|
|
if (remaining == 0)
|
|
printk (KERN_WARNING MYNAM ": %s/%s: WARNING - IOC out of buckets! "
|
|
"(priv->buckets_out = %d)\n",
|
|
IOC_AND_NETDEV_NAMES_s_s(dev),
|
|
atomic_read(&priv->buckets_out));
|
|
else if (remaining < 10)
|
|
printk (KERN_INFO MYNAM ": %s/%s: IOC says %d buckets left. "
|
|
"(priv->buckets_out = %d)\n",
|
|
IOC_AND_NETDEV_NAMES_s_s(dev),
|
|
remaining, atomic_read(&priv->buckets_out));
|
|
|
|
if ((remaining < priv->bucketthresh) &&
|
|
((atomic_read(&priv->buckets_out) - remaining) >
|
|
MPT_LAN_BUCKETS_REMAIN_MISMATCH_THRESH)) {
|
|
|
|
printk (KERN_WARNING MYNAM " Mismatch between driver's "
|
|
"buckets_out count and fw's BucketsRemaining "
|
|
"count has crossed the threshold, issuing a "
|
|
"LanReset to clear the fw's hashtable. You may "
|
|
"want to check your /var/log/messages for \"CRC "
|
|
"error\" event notifications.\n");
|
|
|
|
mpt_lan_reset(dev);
|
|
mpt_lan_wake_post_buckets_task(dev, 0);
|
|
}
|
|
|
|
return mpt_lan_receive_skb(dev, skb);
|
|
}
|
|
|
|
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|
|
/* Simple SGE's only at the moment */
|
|
|
|
static void
|
|
mpt_lan_post_receive_buckets(struct mpt_lan_priv *priv)
|
|
{
|
|
struct net_device *dev = priv->dev;
|
|
MPT_ADAPTER *mpt_dev = priv->mpt_dev;
|
|
MPT_FRAME_HDR *mf;
|
|
LANReceivePostRequest_t *pRecvReq;
|
|
SGETransaction32_t *pTrans;
|
|
SGESimple64_t *pSimple;
|
|
struct sk_buff *skb;
|
|
dma_addr_t dma;
|
|
u32 curr, buckets, count, max;
|
|
u32 len = (dev->mtu + dev->hard_header_len + 4);
|
|
unsigned long flags;
|
|
int i;
|
|
|
|
curr = atomic_read(&priv->buckets_out);
|
|
buckets = (priv->max_buckets_out - curr);
|
|
|
|
dioprintk((KERN_INFO MYNAM ": %s/%s: @%s, Start_buckets = %u, buckets_out = %u\n",
|
|
IOC_AND_NETDEV_NAMES_s_s(dev),
|
|
__func__, buckets, curr));
|
|
|
|
max = (mpt_dev->req_sz - MPT_LAN_RECEIVE_POST_REQUEST_SIZE) /
|
|
(sizeof(SGETransaction32_t) + sizeof(SGESimple64_t));
|
|
|
|
while (buckets) {
|
|
mf = mpt_get_msg_frame(LanCtx, mpt_dev);
|
|
if (mf == NULL) {
|
|
printk (KERN_ERR "%s: Unable to alloc request frame\n",
|
|
__func__);
|
|
dioprintk((KERN_ERR "%s: %u buckets remaining\n",
|
|
__func__, buckets));
|
|
goto out;
|
|
}
|
|
pRecvReq = (LANReceivePostRequest_t *) mf;
|
|
|
|
i = le16_to_cpu(mf->u.frame.hwhdr.msgctxu.fld.req_idx);
|
|
mpt_dev->RequestNB[i] = 0;
|
|
count = buckets;
|
|
if (count > max)
|
|
count = max;
|
|
|
|
pRecvReq->Function = MPI_FUNCTION_LAN_RECEIVE;
|
|
pRecvReq->ChainOffset = 0;
|
|
pRecvReq->MsgFlags = 0;
|
|
pRecvReq->PortNumber = priv->pnum;
|
|
|
|
pTrans = (SGETransaction32_t *) pRecvReq->SG_List;
|
|
pSimple = NULL;
|
|
|
|
for (i = 0; i < count; i++) {
|
|
int ctx;
|
|
|
|
spin_lock_irqsave(&priv->rxfidx_lock, flags);
|
|
if (priv->mpt_rxfidx_tail < 0) {
|
|
printk (KERN_ERR "%s: Can't alloc context\n",
|
|
__func__);
|
|
spin_unlock_irqrestore(&priv->rxfidx_lock,
|
|
flags);
|
|
break;
|
|
}
|
|
|
|
ctx = priv->mpt_rxfidx[priv->mpt_rxfidx_tail--];
|
|
|
|
skb = priv->RcvCtl[ctx].skb;
|
|
if (skb && (priv->RcvCtl[ctx].len != len)) {
|
|
dma_unmap_single(&mpt_dev->pcidev->dev,
|
|
priv->RcvCtl[ctx].dma,
|
|
priv->RcvCtl[ctx].len,
|
|
DMA_FROM_DEVICE);
|
|
dev_kfree_skb(priv->RcvCtl[ctx].skb);
|
|
skb = priv->RcvCtl[ctx].skb = NULL;
|
|
}
|
|
|
|
if (skb == NULL) {
|
|
skb = dev_alloc_skb(len);
|
|
if (skb == NULL) {
|
|
printk (KERN_WARNING
|
|
MYNAM "/%s: Can't alloc skb\n",
|
|
__func__);
|
|
priv->mpt_rxfidx[++priv->mpt_rxfidx_tail] = ctx;
|
|
spin_unlock_irqrestore(&priv->rxfidx_lock, flags);
|
|
break;
|
|
}
|
|
|
|
dma = dma_map_single(&mpt_dev->pcidev->dev,
|
|
skb->data, len,
|
|
DMA_FROM_DEVICE);
|
|
|
|
priv->RcvCtl[ctx].skb = skb;
|
|
priv->RcvCtl[ctx].dma = dma;
|
|
priv->RcvCtl[ctx].len = len;
|
|
}
|
|
|
|
spin_unlock_irqrestore(&priv->rxfidx_lock, flags);
|
|
|
|
pTrans->ContextSize = sizeof(u32);
|
|
pTrans->DetailsLength = 0;
|
|
pTrans->Flags = 0;
|
|
pTrans->TransactionContext = cpu_to_le32(ctx);
|
|
|
|
pSimple = (SGESimple64_t *) pTrans->TransactionDetails;
|
|
|
|
pSimple->FlagsLength = cpu_to_le32(
|
|
((MPI_SGE_FLAGS_END_OF_BUFFER |
|
|
MPI_SGE_FLAGS_SIMPLE_ELEMENT |
|
|
MPI_SGE_FLAGS_64_BIT_ADDRESSING) << MPI_SGE_FLAGS_SHIFT) | len);
|
|
pSimple->Address.Low = cpu_to_le32((u32) priv->RcvCtl[ctx].dma);
|
|
if (sizeof(dma_addr_t) > sizeof(u32))
|
|
pSimple->Address.High = cpu_to_le32((u32) ((u64) priv->RcvCtl[ctx].dma >> 32));
|
|
else
|
|
pSimple->Address.High = 0;
|
|
|
|
pTrans = (SGETransaction32_t *) (pSimple + 1);
|
|
}
|
|
|
|
if (pSimple == NULL) {
|
|
/**/ printk (KERN_WARNING MYNAM "/%s: No buckets posted\n",
|
|
/**/ __func__);
|
|
mpt_free_msg_frame(mpt_dev, mf);
|
|
goto out;
|
|
}
|
|
|
|
pSimple->FlagsLength |= cpu_to_le32(MPI_SGE_FLAGS_END_OF_LIST << MPI_SGE_FLAGS_SHIFT);
|
|
|
|
pRecvReq->BucketCount = cpu_to_le32(i);
|
|
|
|
/* printk(KERN_INFO MYNAM ": posting buckets\n ");
|
|
* for (i = 0; i < j + 2; i ++)
|
|
* printk (" %08x", le32_to_cpu(msg[i]));
|
|
* printk ("\n");
|
|
*/
|
|
|
|
mpt_put_msg_frame(LanCtx, mpt_dev, mf);
|
|
|
|
priv->total_posted += i;
|
|
buckets -= i;
|
|
atomic_add(i, &priv->buckets_out);
|
|
}
|
|
|
|
out:
|
|
dioprintk((KERN_INFO MYNAM "/%s: End_buckets = %u, priv->buckets_out = %u\n",
|
|
__func__, buckets, atomic_read(&priv->buckets_out)));
|
|
dioprintk((KERN_INFO MYNAM "/%s: Posted %u buckets and received %u back\n",
|
|
__func__, priv->total_posted, priv->total_received));
|
|
|
|
clear_bit(0, &priv->post_buckets_active);
|
|
}
|
|
|
|
static void
|
|
mpt_lan_post_receive_buckets_work(struct work_struct *work)
|
|
{
|
|
mpt_lan_post_receive_buckets(container_of(work, struct mpt_lan_priv,
|
|
post_buckets_task.work));
|
|
}
|
|
|
|
static const struct net_device_ops mpt_netdev_ops = {
|
|
.ndo_open = mpt_lan_open,
|
|
.ndo_stop = mpt_lan_close,
|
|
.ndo_start_xmit = mpt_lan_sdu_send,
|
|
.ndo_tx_timeout = mpt_lan_tx_timeout,
|
|
};
|
|
|
|
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|
|
static struct net_device *
|
|
mpt_register_lan_device (MPT_ADAPTER *mpt_dev, int pnum)
|
|
{
|
|
struct net_device *dev;
|
|
struct mpt_lan_priv *priv;
|
|
u8 HWaddr[FC_ALEN], *a;
|
|
|
|
dev = alloc_fcdev(sizeof(struct mpt_lan_priv));
|
|
if (!dev)
|
|
return NULL;
|
|
|
|
dev->mtu = MPT_LAN_MTU;
|
|
|
|
priv = netdev_priv(dev);
|
|
|
|
priv->dev = dev;
|
|
priv->mpt_dev = mpt_dev;
|
|
priv->pnum = pnum;
|
|
|
|
INIT_DELAYED_WORK(&priv->post_buckets_task,
|
|
mpt_lan_post_receive_buckets_work);
|
|
priv->post_buckets_active = 0;
|
|
|
|
dlprintk((KERN_INFO MYNAM "@%d: bucketlen = %d\n",
|
|
__LINE__, dev->mtu + dev->hard_header_len + 4));
|
|
|
|
atomic_set(&priv->buckets_out, 0);
|
|
priv->total_posted = 0;
|
|
priv->total_received = 0;
|
|
priv->max_buckets_out = max_buckets_out;
|
|
if (mpt_dev->pfacts[0].MaxLanBuckets < max_buckets_out)
|
|
priv->max_buckets_out = mpt_dev->pfacts[0].MaxLanBuckets;
|
|
|
|
dlprintk((KERN_INFO MYNAM "@%d: MaxLanBuckets=%d, max_buckets_out/priv=%d/%d\n",
|
|
__LINE__,
|
|
mpt_dev->pfacts[0].MaxLanBuckets,
|
|
max_buckets_out,
|
|
priv->max_buckets_out));
|
|
|
|
priv->bucketthresh = priv->max_buckets_out * 2 / 3;
|
|
spin_lock_init(&priv->txfidx_lock);
|
|
spin_lock_init(&priv->rxfidx_lock);
|
|
|
|
/* Grab pre-fetched LANPage1 stuff. :-) */
|
|
a = (u8 *) &mpt_dev->lan_cnfg_page1.HardwareAddressLow;
|
|
|
|
HWaddr[0] = a[5];
|
|
HWaddr[1] = a[4];
|
|
HWaddr[2] = a[3];
|
|
HWaddr[3] = a[2];
|
|
HWaddr[4] = a[1];
|
|
HWaddr[5] = a[0];
|
|
|
|
dev->addr_len = FC_ALEN;
|
|
dev_addr_set(dev, HWaddr);
|
|
memset(dev->broadcast, 0xff, FC_ALEN);
|
|
|
|
/* The Tx queue is 127 deep on the 909.
|
|
* Give ourselves some breathing room.
|
|
*/
|
|
priv->tx_max_out = (tx_max_out_p <= MPT_TX_MAX_OUT_LIM) ?
|
|
tx_max_out_p : MPT_TX_MAX_OUT_LIM;
|
|
|
|
dev->netdev_ops = &mpt_netdev_ops;
|
|
dev->watchdog_timeo = MPT_LAN_TX_TIMEOUT;
|
|
|
|
/* MTU range: 96 - 65280 */
|
|
dev->min_mtu = MPT_LAN_MIN_MTU;
|
|
dev->max_mtu = MPT_LAN_MAX_MTU;
|
|
|
|
dlprintk((KERN_INFO MYNAM ": Finished registering dev "
|
|
"and setting initial values\n"));
|
|
|
|
if (register_netdev(dev) != 0) {
|
|
free_netdev(dev);
|
|
dev = NULL;
|
|
}
|
|
return dev;
|
|
}
|
|
|
|
static int
|
|
mptlan_probe(struct pci_dev *pdev)
|
|
{
|
|
MPT_ADAPTER *ioc = pci_get_drvdata(pdev);
|
|
struct net_device *dev;
|
|
int i;
|
|
|
|
for (i = 0; i < ioc->facts.NumberOfPorts; i++) {
|
|
printk(KERN_INFO MYNAM ": %s: PortNum=%x, "
|
|
"ProtocolFlags=%02Xh (%c%c%c%c)\n",
|
|
ioc->name, ioc->pfacts[i].PortNumber,
|
|
ioc->pfacts[i].ProtocolFlags,
|
|
MPT_PROTOCOL_FLAGS_c_c_c_c(
|
|
ioc->pfacts[i].ProtocolFlags));
|
|
|
|
if (!(ioc->pfacts[i].ProtocolFlags &
|
|
MPI_PORTFACTS_PROTOCOL_LAN)) {
|
|
printk(KERN_INFO MYNAM ": %s: Hmmm... LAN protocol "
|
|
"seems to be disabled on this adapter port!\n",
|
|
ioc->name);
|
|
continue;
|
|
}
|
|
|
|
dev = mpt_register_lan_device(ioc, i);
|
|
if (!dev) {
|
|
printk(KERN_ERR MYNAM ": %s: Unable to register "
|
|
"port%d as a LAN device\n", ioc->name,
|
|
ioc->pfacts[i].PortNumber);
|
|
continue;
|
|
}
|
|
|
|
printk(KERN_INFO MYNAM ": %s: Fusion MPT LAN device "
|
|
"registered as '%s'\n", ioc->name, dev->name);
|
|
printk(KERN_INFO MYNAM ": %s/%s: "
|
|
"LanAddr = %pM\n",
|
|
IOC_AND_NETDEV_NAMES_s_s(dev),
|
|
dev->dev_addr);
|
|
|
|
ioc->netdev = dev;
|
|
|
|
return 0;
|
|
}
|
|
|
|
return -ENODEV;
|
|
}
|
|
|
|
static void
|
|
mptlan_remove(struct pci_dev *pdev)
|
|
{
|
|
MPT_ADAPTER *ioc = pci_get_drvdata(pdev);
|
|
struct net_device *dev = ioc->netdev;
|
|
struct mpt_lan_priv *priv = netdev_priv(dev);
|
|
|
|
cancel_delayed_work_sync(&priv->post_buckets_task);
|
|
if(dev != NULL) {
|
|
unregister_netdev(dev);
|
|
free_netdev(dev);
|
|
}
|
|
}
|
|
|
|
static struct mpt_pci_driver mptlan_driver = {
|
|
.probe = mptlan_probe,
|
|
.remove = mptlan_remove,
|
|
};
|
|
|
|
static int __init mpt_lan_init (void)
|
|
{
|
|
show_mptmod_ver(LANAME, LANVER);
|
|
|
|
LanCtx = mpt_register(lan_reply, MPTLAN_DRIVER,
|
|
"lan_reply");
|
|
if (LanCtx <= 0) {
|
|
printk (KERN_ERR MYNAM ": Failed to register with MPT base driver\n");
|
|
return -EBUSY;
|
|
}
|
|
|
|
dlprintk((KERN_INFO MYNAM ": assigned context of %d\n", LanCtx));
|
|
|
|
if (mpt_reset_register(LanCtx, mpt_lan_ioc_reset)) {
|
|
printk(KERN_ERR MYNAM ": Eieee! unable to register a reset "
|
|
"handler with mptbase! The world is at an end! "
|
|
"Everything is fading to black! Goodbye.\n");
|
|
return -EBUSY;
|
|
}
|
|
|
|
dlprintk((KERN_INFO MYNAM ": Registered for IOC reset notifications\n"));
|
|
|
|
mpt_device_driver_register(&mptlan_driver, MPTLAN_DRIVER);
|
|
return 0;
|
|
}
|
|
|
|
static void __exit mpt_lan_exit(void)
|
|
{
|
|
mpt_device_driver_deregister(MPTLAN_DRIVER);
|
|
mpt_reset_deregister(LanCtx);
|
|
|
|
if (LanCtx) {
|
|
mpt_deregister(LanCtx);
|
|
LanCtx = MPT_MAX_PROTOCOL_DRIVERS;
|
|
}
|
|
}
|
|
|
|
module_init(mpt_lan_init);
|
|
module_exit(mpt_lan_exit);
|
|
|
|
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|
|
static unsigned short
|
|
mpt_lan_type_trans(struct sk_buff *skb, struct net_device *dev)
|
|
{
|
|
struct mpt_lan_ohdr *fch = (struct mpt_lan_ohdr *)skb->data;
|
|
struct fcllc *fcllc;
|
|
|
|
skb_reset_mac_header(skb);
|
|
skb_pull(skb, sizeof(struct mpt_lan_ohdr));
|
|
|
|
if (fch->dtype == htons(0xffff)) {
|
|
u32 *p = (u32 *) fch;
|
|
|
|
swab32s(p + 0);
|
|
swab32s(p + 1);
|
|
swab32s(p + 2);
|
|
swab32s(p + 3);
|
|
|
|
printk (KERN_WARNING MYNAM ": %s: WARNING - Broadcast swap F/W bug detected!\n",
|
|
NETDEV_PTR_TO_IOC_NAME_s(dev));
|
|
printk (KERN_WARNING MYNAM ": Please update sender @ MAC_addr = %pM\n",
|
|
fch->saddr);
|
|
}
|
|
|
|
if (*fch->daddr & 1) {
|
|
if (!memcmp(fch->daddr, dev->broadcast, FC_ALEN)) {
|
|
skb->pkt_type = PACKET_BROADCAST;
|
|
} else {
|
|
skb->pkt_type = PACKET_MULTICAST;
|
|
}
|
|
} else {
|
|
if (memcmp(fch->daddr, dev->dev_addr, FC_ALEN)) {
|
|
skb->pkt_type = PACKET_OTHERHOST;
|
|
} else {
|
|
skb->pkt_type = PACKET_HOST;
|
|
}
|
|
}
|
|
|
|
fcllc = (struct fcllc *)skb->data;
|
|
|
|
/* Strip the SNAP header from ARP packets since we don't
|
|
* pass them through to the 802.2/SNAP layers.
|
|
*/
|
|
if (fcllc->dsap == EXTENDED_SAP &&
|
|
(fcllc->ethertype == htons(ETH_P_IP) ||
|
|
fcllc->ethertype == htons(ETH_P_ARP))) {
|
|
skb_pull(skb, sizeof(struct fcllc));
|
|
return fcllc->ethertype;
|
|
}
|
|
|
|
return htons(ETH_P_802_2);
|
|
}
|
|
|
|
/*=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=-=*/
|