mirror of
https://github.com/torvalds/linux.git
synced 2024-11-27 06:31:52 +00:00
386a966f5c
The driver core ignores the return value of struct bus_type::remove() because there is only little that can be done. To simplify the quest to make this function return void, let struct vio_driver::remove() return void, too. All users already unconditionally return 0, this commit makes it obvious that returning an error code is a bad idea. Note there are two nominally different implementations for a vio bus: one in arch/sparc/kernel/vio.c and the other in arch/powerpc/platforms/pseries/vio.c. This patch only adapts the powerpc one. Before this patch for a device that was bound to a driver without a remove callback vio_cmo_bus_remove(viodev) wasn't called. As the device core still considers the device unbound after vio_bus_remove() returns calling this unconditionally is the consistent behaviour which is implemented here. Signed-off-by: Uwe Kleine-König <uwe@kleine-koenig.org> Reviewed-by: Tyrel Datwyler <tyreld@linux.ibm.com> Acked-by: Lijun Pan <ljp@linux.ibm.com> Acked-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org> [mpe: Drop unneeded hvcs_remove() forward declaration, squash in change from sfr to drop ibmvnic_remove() forward declaration] Signed-off-by: Michael Ellerman <mpe@ellerman.id.au> Link: https://lore.kernel.org/r/20210225221834.160083-1-uwe@kleine-koenig.org
2420 lines
61 KiB
C
2420 lines
61 KiB
C
// SPDX-License-Identifier: GPL-2.0+
|
|
/*
|
|
* IBM Power Systems Virtual Management Channel Support.
|
|
*
|
|
* Copyright (c) 2004, 2018 IBM Corp.
|
|
* Dave Engebretsen engebret@us.ibm.com
|
|
* Steven Royer seroyer@linux.vnet.ibm.com
|
|
* Adam Reznechek adreznec@linux.vnet.ibm.com
|
|
* Bryant G. Ly <bryantly@linux.vnet.ibm.com>
|
|
*/
|
|
|
|
#include <linux/module.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/kthread.h>
|
|
#include <linux/major.h>
|
|
#include <linux/string.h>
|
|
#include <linux/fcntl.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/poll.h>
|
|
#include <linux/init.h>
|
|
#include <linux/fs.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/spinlock.h>
|
|
#include <linux/percpu.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/uaccess.h>
|
|
#include <linux/io.h>
|
|
#include <linux/miscdevice.h>
|
|
#include <linux/sched/signal.h>
|
|
|
|
#include <asm/byteorder.h>
|
|
#include <asm/irq.h>
|
|
#include <asm/vio.h>
|
|
|
|
#include "ibmvmc.h"
|
|
|
|
#define IBMVMC_DRIVER_VERSION "1.0"
|
|
|
|
/*
|
|
* Static global variables
|
|
*/
|
|
static DECLARE_WAIT_QUEUE_HEAD(ibmvmc_read_wait);
|
|
|
|
static const char ibmvmc_driver_name[] = "ibmvmc";
|
|
|
|
static struct ibmvmc_struct ibmvmc;
|
|
static struct ibmvmc_hmc hmcs[MAX_HMCS];
|
|
static struct crq_server_adapter ibmvmc_adapter;
|
|
|
|
static int ibmvmc_max_buf_pool_size = DEFAULT_BUF_POOL_SIZE;
|
|
static int ibmvmc_max_hmcs = DEFAULT_HMCS;
|
|
static int ibmvmc_max_mtu = DEFAULT_MTU;
|
|
|
|
static inline long h_copy_rdma(s64 length, u64 sliobn, u64 slioba,
|
|
u64 dliobn, u64 dlioba)
|
|
{
|
|
long rc = 0;
|
|
|
|
/* Ensure all writes to source memory are visible before hcall */
|
|
dma_wmb();
|
|
pr_debug("ibmvmc: h_copy_rdma(0x%llx, 0x%llx, 0x%llx, 0x%llx, 0x%llx\n",
|
|
length, sliobn, slioba, dliobn, dlioba);
|
|
rc = plpar_hcall_norets(H_COPY_RDMA, length, sliobn, slioba,
|
|
dliobn, dlioba);
|
|
pr_debug("ibmvmc: h_copy_rdma rc = 0x%lx\n", rc);
|
|
|
|
return rc;
|
|
}
|
|
|
|
static inline void h_free_crq(uint32_t unit_address)
|
|
{
|
|
long rc = 0;
|
|
|
|
do {
|
|
if (H_IS_LONG_BUSY(rc))
|
|
msleep(get_longbusy_msecs(rc));
|
|
|
|
rc = plpar_hcall_norets(H_FREE_CRQ, unit_address);
|
|
} while ((rc == H_BUSY) || (H_IS_LONG_BUSY(rc)));
|
|
}
|
|
|
|
/**
|
|
* h_request_vmc: - request a hypervisor virtual management channel device
|
|
* @vmc_index: drc index of the vmc device created
|
|
*
|
|
* Requests the hypervisor create a new virtual management channel device,
|
|
* allowing this partition to send hypervisor virtualization control
|
|
* commands.
|
|
*
|
|
* Return:
|
|
* 0 - Success
|
|
* Non-zero - Failure
|
|
*/
|
|
static inline long h_request_vmc(u32 *vmc_index)
|
|
{
|
|
long rc = 0;
|
|
unsigned long retbuf[PLPAR_HCALL_BUFSIZE];
|
|
|
|
do {
|
|
if (H_IS_LONG_BUSY(rc))
|
|
msleep(get_longbusy_msecs(rc));
|
|
|
|
/* Call to request the VMC device from phyp */
|
|
rc = plpar_hcall(H_REQUEST_VMC, retbuf);
|
|
pr_debug("ibmvmc: %s rc = 0x%lx\n", __func__, rc);
|
|
*vmc_index = retbuf[0];
|
|
} while ((rc == H_BUSY) || (H_IS_LONG_BUSY(rc)));
|
|
|
|
return rc;
|
|
}
|
|
|
|
/* routines for managing a command/response queue */
|
|
/**
|
|
* ibmvmc_handle_event: - Interrupt handler for crq events
|
|
* @irq: number of irq to handle, not used
|
|
* @dev_instance: crq_server_adapter that received interrupt
|
|
*
|
|
* Disables interrupts and schedules ibmvmc_task
|
|
*
|
|
* Always returns IRQ_HANDLED
|
|
*/
|
|
static irqreturn_t ibmvmc_handle_event(int irq, void *dev_instance)
|
|
{
|
|
struct crq_server_adapter *adapter =
|
|
(struct crq_server_adapter *)dev_instance;
|
|
|
|
vio_disable_interrupts(to_vio_dev(adapter->dev));
|
|
tasklet_schedule(&adapter->work_task);
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_release_crq_queue - Release CRQ Queue
|
|
*
|
|
* @adapter: crq_server_adapter struct
|
|
*
|
|
* Return:
|
|
* 0 - Success
|
|
* Non-Zero - Failure
|
|
*/
|
|
static void ibmvmc_release_crq_queue(struct crq_server_adapter *adapter)
|
|
{
|
|
struct vio_dev *vdev = to_vio_dev(adapter->dev);
|
|
struct crq_queue *queue = &adapter->queue;
|
|
|
|
free_irq(vdev->irq, (void *)adapter);
|
|
tasklet_kill(&adapter->work_task);
|
|
|
|
if (adapter->reset_task)
|
|
kthread_stop(adapter->reset_task);
|
|
|
|
h_free_crq(vdev->unit_address);
|
|
dma_unmap_single(adapter->dev,
|
|
queue->msg_token,
|
|
queue->size * sizeof(*queue->msgs), DMA_BIDIRECTIONAL);
|
|
free_page((unsigned long)queue->msgs);
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_reset_crq_queue - Reset CRQ Queue
|
|
*
|
|
* @adapter: crq_server_adapter struct
|
|
*
|
|
* This function calls h_free_crq and then calls H_REG_CRQ and does all the
|
|
* bookkeeping to get us back to where we can communicate.
|
|
*
|
|
* Return:
|
|
* 0 - Success
|
|
* Non-Zero - Failure
|
|
*/
|
|
static int ibmvmc_reset_crq_queue(struct crq_server_adapter *adapter)
|
|
{
|
|
struct vio_dev *vdev = to_vio_dev(adapter->dev);
|
|
struct crq_queue *queue = &adapter->queue;
|
|
int rc = 0;
|
|
|
|
/* Close the CRQ */
|
|
h_free_crq(vdev->unit_address);
|
|
|
|
/* Clean out the queue */
|
|
memset(queue->msgs, 0x00, PAGE_SIZE);
|
|
queue->cur = 0;
|
|
|
|
/* And re-open it again */
|
|
rc = plpar_hcall_norets(H_REG_CRQ,
|
|
vdev->unit_address,
|
|
queue->msg_token, PAGE_SIZE);
|
|
if (rc == 2)
|
|
/* Adapter is good, but other end is not ready */
|
|
dev_warn(adapter->dev, "Partner adapter not ready\n");
|
|
else if (rc != 0)
|
|
dev_err(adapter->dev, "couldn't register crq--rc 0x%x\n", rc);
|
|
|
|
return rc;
|
|
}
|
|
|
|
/**
|
|
* crq_queue_next_crq: - Returns the next entry in message queue
|
|
* @queue: crq_queue to use
|
|
*
|
|
* Returns pointer to next entry in queue, or NULL if there are no new
|
|
* entried in the CRQ.
|
|
*/
|
|
static struct ibmvmc_crq_msg *crq_queue_next_crq(struct crq_queue *queue)
|
|
{
|
|
struct ibmvmc_crq_msg *crq;
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&queue->lock, flags);
|
|
crq = &queue->msgs[queue->cur];
|
|
if (crq->valid & 0x80) {
|
|
if (++queue->cur == queue->size)
|
|
queue->cur = 0;
|
|
|
|
/* Ensure the read of the valid bit occurs before reading any
|
|
* other bits of the CRQ entry
|
|
*/
|
|
dma_rmb();
|
|
} else {
|
|
crq = NULL;
|
|
}
|
|
|
|
spin_unlock_irqrestore(&queue->lock, flags);
|
|
|
|
return crq;
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_send_crq - Send CRQ
|
|
*
|
|
* @adapter: crq_server_adapter struct
|
|
* @word1: Word1 Data field
|
|
* @word2: Word2 Data field
|
|
*
|
|
* Return:
|
|
* 0 - Success
|
|
* Non-Zero - Failure
|
|
*/
|
|
static long ibmvmc_send_crq(struct crq_server_adapter *adapter,
|
|
u64 word1, u64 word2)
|
|
{
|
|
struct vio_dev *vdev = to_vio_dev(adapter->dev);
|
|
long rc = 0;
|
|
|
|
dev_dbg(adapter->dev, "(0x%x, 0x%016llx, 0x%016llx)\n",
|
|
vdev->unit_address, word1, word2);
|
|
|
|
/*
|
|
* Ensure the command buffer is flushed to memory before handing it
|
|
* over to the other side to prevent it from fetching any stale data.
|
|
*/
|
|
dma_wmb();
|
|
rc = plpar_hcall_norets(H_SEND_CRQ, vdev->unit_address, word1, word2);
|
|
dev_dbg(adapter->dev, "rc = 0x%lx\n", rc);
|
|
|
|
return rc;
|
|
}
|
|
|
|
/**
|
|
* alloc_dma_buffer - Create DMA Buffer
|
|
*
|
|
* @vdev: vio_dev struct
|
|
* @size: Size field
|
|
* @dma_handle: DMA address field
|
|
*
|
|
* Allocates memory for the command queue and maps remote memory into an
|
|
* ioba.
|
|
*
|
|
* Returns a pointer to the buffer
|
|
*/
|
|
static void *alloc_dma_buffer(struct vio_dev *vdev, size_t size,
|
|
dma_addr_t *dma_handle)
|
|
{
|
|
/* allocate memory */
|
|
void *buffer = kzalloc(size, GFP_ATOMIC);
|
|
|
|
if (!buffer) {
|
|
*dma_handle = 0;
|
|
return NULL;
|
|
}
|
|
|
|
/* DMA map */
|
|
*dma_handle = dma_map_single(&vdev->dev, buffer, size,
|
|
DMA_BIDIRECTIONAL);
|
|
|
|
if (dma_mapping_error(&vdev->dev, *dma_handle)) {
|
|
*dma_handle = 0;
|
|
kfree_sensitive(buffer);
|
|
return NULL;
|
|
}
|
|
|
|
return buffer;
|
|
}
|
|
|
|
/**
|
|
* free_dma_buffer - Free DMA Buffer
|
|
*
|
|
* @vdev: vio_dev struct
|
|
* @size: Size field
|
|
* @vaddr: Address field
|
|
* @dma_handle: DMA address field
|
|
*
|
|
* Releases memory for a command queue and unmaps mapped remote memory.
|
|
*/
|
|
static void free_dma_buffer(struct vio_dev *vdev, size_t size, void *vaddr,
|
|
dma_addr_t dma_handle)
|
|
{
|
|
/* DMA unmap */
|
|
dma_unmap_single(&vdev->dev, dma_handle, size, DMA_BIDIRECTIONAL);
|
|
|
|
/* deallocate memory */
|
|
kfree_sensitive(vaddr);
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_get_valid_hmc_buffer - Retrieve Valid HMC Buffer
|
|
*
|
|
* @hmc_index: HMC Index Field
|
|
*
|
|
* Return:
|
|
* Pointer to ibmvmc_buffer
|
|
*/
|
|
static struct ibmvmc_buffer *ibmvmc_get_valid_hmc_buffer(u8 hmc_index)
|
|
{
|
|
struct ibmvmc_buffer *buffer;
|
|
struct ibmvmc_buffer *ret_buf = NULL;
|
|
unsigned long i;
|
|
|
|
if (hmc_index > ibmvmc.max_hmc_index)
|
|
return NULL;
|
|
|
|
buffer = hmcs[hmc_index].buffer;
|
|
|
|
for (i = 0; i < ibmvmc_max_buf_pool_size; i++) {
|
|
if (buffer[i].valid && buffer[i].free &&
|
|
buffer[i].owner == VMC_BUF_OWNER_ALPHA) {
|
|
buffer[i].free = 0;
|
|
ret_buf = &buffer[i];
|
|
break;
|
|
}
|
|
}
|
|
|
|
return ret_buf;
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_get_free_hmc_buffer - Get Free HMC Buffer
|
|
*
|
|
* @adapter: crq_server_adapter struct
|
|
* @hmc_index: Hmc Index field
|
|
*
|
|
* Return:
|
|
* Pointer to ibmvmc_buffer
|
|
*/
|
|
static struct ibmvmc_buffer *ibmvmc_get_free_hmc_buffer(struct crq_server_adapter *adapter,
|
|
u8 hmc_index)
|
|
{
|
|
struct ibmvmc_buffer *buffer;
|
|
struct ibmvmc_buffer *ret_buf = NULL;
|
|
unsigned long i;
|
|
|
|
if (hmc_index > ibmvmc.max_hmc_index) {
|
|
dev_info(adapter->dev, "get_free_hmc_buffer: invalid hmc_index=0x%x\n",
|
|
hmc_index);
|
|
return NULL;
|
|
}
|
|
|
|
buffer = hmcs[hmc_index].buffer;
|
|
|
|
for (i = 0; i < ibmvmc_max_buf_pool_size; i++) {
|
|
if (buffer[i].free &&
|
|
buffer[i].owner == VMC_BUF_OWNER_ALPHA) {
|
|
buffer[i].free = 0;
|
|
ret_buf = &buffer[i];
|
|
break;
|
|
}
|
|
}
|
|
|
|
return ret_buf;
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_free_hmc_buffer - Free an HMC Buffer
|
|
*
|
|
* @hmc: ibmvmc_hmc struct
|
|
* @buffer: ibmvmc_buffer struct
|
|
*
|
|
*/
|
|
static void ibmvmc_free_hmc_buffer(struct ibmvmc_hmc *hmc,
|
|
struct ibmvmc_buffer *buffer)
|
|
{
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&hmc->lock, flags);
|
|
buffer->free = 1;
|
|
spin_unlock_irqrestore(&hmc->lock, flags);
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_count_hmc_buffers - Count HMC Buffers
|
|
*
|
|
* @hmc_index: HMC Index field
|
|
* @valid: Valid number of buffers field
|
|
* @free: Free number of buffers field
|
|
*
|
|
*/
|
|
static void ibmvmc_count_hmc_buffers(u8 hmc_index, unsigned int *valid,
|
|
unsigned int *free)
|
|
{
|
|
struct ibmvmc_buffer *buffer;
|
|
unsigned long i;
|
|
unsigned long flags;
|
|
|
|
if (hmc_index > ibmvmc.max_hmc_index)
|
|
return;
|
|
|
|
if (!valid || !free)
|
|
return;
|
|
|
|
*valid = 0; *free = 0;
|
|
|
|
buffer = hmcs[hmc_index].buffer;
|
|
spin_lock_irqsave(&hmcs[hmc_index].lock, flags);
|
|
|
|
for (i = 0; i < ibmvmc_max_buf_pool_size; i++) {
|
|
if (buffer[i].valid) {
|
|
*valid = *valid + 1;
|
|
if (buffer[i].free)
|
|
*free = *free + 1;
|
|
}
|
|
}
|
|
|
|
spin_unlock_irqrestore(&hmcs[hmc_index].lock, flags);
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_get_free_hmc - Get Free HMC
|
|
*
|
|
* Return:
|
|
* Pointer to an available HMC Connection
|
|
* Null otherwise
|
|
*/
|
|
static struct ibmvmc_hmc *ibmvmc_get_free_hmc(void)
|
|
{
|
|
unsigned long i;
|
|
unsigned long flags;
|
|
|
|
/*
|
|
* Find an available HMC connection.
|
|
*/
|
|
for (i = 0; i <= ibmvmc.max_hmc_index; i++) {
|
|
spin_lock_irqsave(&hmcs[i].lock, flags);
|
|
if (hmcs[i].state == ibmhmc_state_free) {
|
|
hmcs[i].index = i;
|
|
hmcs[i].state = ibmhmc_state_initial;
|
|
spin_unlock_irqrestore(&hmcs[i].lock, flags);
|
|
return &hmcs[i];
|
|
}
|
|
spin_unlock_irqrestore(&hmcs[i].lock, flags);
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_return_hmc - Return an HMC Connection
|
|
*
|
|
* @hmc: ibmvmc_hmc struct
|
|
* @release_readers: Number of readers connected to session
|
|
*
|
|
* This function releases the HMC connections back into the pool.
|
|
*
|
|
* Return:
|
|
* 0 - Success
|
|
* Non-zero - Failure
|
|
*/
|
|
static int ibmvmc_return_hmc(struct ibmvmc_hmc *hmc, bool release_readers)
|
|
{
|
|
struct ibmvmc_buffer *buffer;
|
|
struct crq_server_adapter *adapter;
|
|
struct vio_dev *vdev;
|
|
unsigned long i;
|
|
unsigned long flags;
|
|
|
|
if (!hmc || !hmc->adapter)
|
|
return -EIO;
|
|
|
|
if (release_readers) {
|
|
if (hmc->file_session) {
|
|
struct ibmvmc_file_session *session = hmc->file_session;
|
|
|
|
session->valid = 0;
|
|
wake_up_interruptible(&ibmvmc_read_wait);
|
|
}
|
|
}
|
|
|
|
adapter = hmc->adapter;
|
|
vdev = to_vio_dev(adapter->dev);
|
|
|
|
spin_lock_irqsave(&hmc->lock, flags);
|
|
hmc->index = 0;
|
|
hmc->state = ibmhmc_state_free;
|
|
hmc->queue_head = 0;
|
|
hmc->queue_tail = 0;
|
|
buffer = hmc->buffer;
|
|
for (i = 0; i < ibmvmc_max_buf_pool_size; i++) {
|
|
if (buffer[i].valid) {
|
|
free_dma_buffer(vdev,
|
|
ibmvmc.max_mtu,
|
|
buffer[i].real_addr_local,
|
|
buffer[i].dma_addr_local);
|
|
dev_dbg(adapter->dev, "Forgot buffer id 0x%lx\n", i);
|
|
}
|
|
memset(&buffer[i], 0, sizeof(struct ibmvmc_buffer));
|
|
|
|
hmc->queue_outbound_msgs[i] = VMC_INVALID_BUFFER_ID;
|
|
}
|
|
|
|
spin_unlock_irqrestore(&hmc->lock, flags);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_send_open - Interface Open
|
|
* @buffer: Pointer to ibmvmc_buffer struct
|
|
* @hmc: Pointer to ibmvmc_hmc struct
|
|
*
|
|
* This command is sent by the management partition as the result of a
|
|
* management partition device request. It causes the hypervisor to
|
|
* prepare a set of data buffers for the management application connection
|
|
* indicated HMC idx. A unique HMC Idx would be used if multiple management
|
|
* applications running concurrently were desired. Before responding to this
|
|
* command, the hypervisor must provide the management partition with at
|
|
* least one of these new buffers via the Add Buffer. This indicates whether
|
|
* the messages are inbound or outbound from the hypervisor.
|
|
*
|
|
* Return:
|
|
* 0 - Success
|
|
* Non-zero - Failure
|
|
*/
|
|
static int ibmvmc_send_open(struct ibmvmc_buffer *buffer,
|
|
struct ibmvmc_hmc *hmc)
|
|
{
|
|
struct ibmvmc_crq_msg crq_msg;
|
|
struct crq_server_adapter *adapter;
|
|
__be64 *crq_as_u64 = (__be64 *)&crq_msg;
|
|
int rc = 0;
|
|
|
|
if (!hmc || !hmc->adapter)
|
|
return -EIO;
|
|
|
|
adapter = hmc->adapter;
|
|
|
|
dev_dbg(adapter->dev, "send_open: 0x%lx 0x%lx 0x%lx 0x%lx 0x%lx\n",
|
|
(unsigned long)buffer->size, (unsigned long)adapter->liobn,
|
|
(unsigned long)buffer->dma_addr_local,
|
|
(unsigned long)adapter->riobn,
|
|
(unsigned long)buffer->dma_addr_remote);
|
|
|
|
rc = h_copy_rdma(buffer->size,
|
|
adapter->liobn,
|
|
buffer->dma_addr_local,
|
|
adapter->riobn,
|
|
buffer->dma_addr_remote);
|
|
if (rc) {
|
|
dev_err(adapter->dev, "Error: In send_open, h_copy_rdma rc 0x%x\n",
|
|
rc);
|
|
return -EIO;
|
|
}
|
|
|
|
hmc->state = ibmhmc_state_opening;
|
|
|
|
crq_msg.valid = 0x80;
|
|
crq_msg.type = VMC_MSG_OPEN;
|
|
crq_msg.status = 0;
|
|
crq_msg.var1.rsvd = 0;
|
|
crq_msg.hmc_session = hmc->session;
|
|
crq_msg.hmc_index = hmc->index;
|
|
crq_msg.var2.buffer_id = cpu_to_be16(buffer->id);
|
|
crq_msg.rsvd = 0;
|
|
crq_msg.var3.rsvd = 0;
|
|
|
|
ibmvmc_send_crq(adapter, be64_to_cpu(crq_as_u64[0]),
|
|
be64_to_cpu(crq_as_u64[1]));
|
|
|
|
return rc;
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_send_close - Interface Close
|
|
* @hmc: Pointer to ibmvmc_hmc struct
|
|
*
|
|
* This command is sent by the management partition to terminate a
|
|
* management application to hypervisor connection. When this command is
|
|
* sent, the management partition has quiesced all I/O operations to all
|
|
* buffers associated with this management application connection, and
|
|
* has freed any storage for these buffers.
|
|
*
|
|
* Return:
|
|
* 0 - Success
|
|
* Non-zero - Failure
|
|
*/
|
|
static int ibmvmc_send_close(struct ibmvmc_hmc *hmc)
|
|
{
|
|
struct ibmvmc_crq_msg crq_msg;
|
|
struct crq_server_adapter *adapter;
|
|
__be64 *crq_as_u64 = (__be64 *)&crq_msg;
|
|
int rc = 0;
|
|
|
|
if (!hmc || !hmc->adapter)
|
|
return -EIO;
|
|
|
|
adapter = hmc->adapter;
|
|
|
|
dev_info(adapter->dev, "CRQ send: close\n");
|
|
|
|
crq_msg.valid = 0x80;
|
|
crq_msg.type = VMC_MSG_CLOSE;
|
|
crq_msg.status = 0;
|
|
crq_msg.var1.rsvd = 0;
|
|
crq_msg.hmc_session = hmc->session;
|
|
crq_msg.hmc_index = hmc->index;
|
|
crq_msg.var2.rsvd = 0;
|
|
crq_msg.rsvd = 0;
|
|
crq_msg.var3.rsvd = 0;
|
|
|
|
ibmvmc_send_crq(adapter, be64_to_cpu(crq_as_u64[0]),
|
|
be64_to_cpu(crq_as_u64[1]));
|
|
|
|
return rc;
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_send_capabilities - Send VMC Capabilities
|
|
*
|
|
* @adapter: crq_server_adapter struct
|
|
*
|
|
* The capabilities message is an administrative message sent after the CRQ
|
|
* initialization sequence of messages and is used to exchange VMC capabilities
|
|
* between the management partition and the hypervisor. The management
|
|
* partition must send this message and the hypervisor must respond with VMC
|
|
* capabilities Response message before HMC interface message can begin. Any
|
|
* HMC interface messages received before the exchange of capabilities has
|
|
* complete are dropped.
|
|
*
|
|
* Return:
|
|
* 0 - Success
|
|
*/
|
|
static int ibmvmc_send_capabilities(struct crq_server_adapter *adapter)
|
|
{
|
|
struct ibmvmc_admin_crq_msg crq_msg;
|
|
__be64 *crq_as_u64 = (__be64 *)&crq_msg;
|
|
|
|
dev_dbg(adapter->dev, "ibmvmc: CRQ send: capabilities\n");
|
|
crq_msg.valid = 0x80;
|
|
crq_msg.type = VMC_MSG_CAP;
|
|
crq_msg.status = 0;
|
|
crq_msg.rsvd[0] = 0;
|
|
crq_msg.rsvd[1] = 0;
|
|
crq_msg.max_hmc = ibmvmc_max_hmcs;
|
|
crq_msg.max_mtu = cpu_to_be32(ibmvmc_max_mtu);
|
|
crq_msg.pool_size = cpu_to_be16(ibmvmc_max_buf_pool_size);
|
|
crq_msg.crq_size = cpu_to_be16(adapter->queue.size);
|
|
crq_msg.version = cpu_to_be16(IBMVMC_PROTOCOL_VERSION);
|
|
|
|
ibmvmc_send_crq(adapter, be64_to_cpu(crq_as_u64[0]),
|
|
be64_to_cpu(crq_as_u64[1]));
|
|
|
|
ibmvmc.state = ibmvmc_state_capabilities;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_send_add_buffer_resp - Add Buffer Response
|
|
*
|
|
* @adapter: crq_server_adapter struct
|
|
* @status: Status field
|
|
* @hmc_session: HMC Session field
|
|
* @hmc_index: HMC Index field
|
|
* @buffer_id: Buffer Id field
|
|
*
|
|
* This command is sent by the management partition to the hypervisor in
|
|
* response to the Add Buffer message. The Status field indicates the result of
|
|
* the command.
|
|
*
|
|
* Return:
|
|
* 0 - Success
|
|
*/
|
|
static int ibmvmc_send_add_buffer_resp(struct crq_server_adapter *adapter,
|
|
u8 status, u8 hmc_session,
|
|
u8 hmc_index, u16 buffer_id)
|
|
{
|
|
struct ibmvmc_crq_msg crq_msg;
|
|
__be64 *crq_as_u64 = (__be64 *)&crq_msg;
|
|
|
|
dev_dbg(adapter->dev, "CRQ send: add_buffer_resp\n");
|
|
crq_msg.valid = 0x80;
|
|
crq_msg.type = VMC_MSG_ADD_BUF_RESP;
|
|
crq_msg.status = status;
|
|
crq_msg.var1.rsvd = 0;
|
|
crq_msg.hmc_session = hmc_session;
|
|
crq_msg.hmc_index = hmc_index;
|
|
crq_msg.var2.buffer_id = cpu_to_be16(buffer_id);
|
|
crq_msg.rsvd = 0;
|
|
crq_msg.var3.rsvd = 0;
|
|
|
|
ibmvmc_send_crq(adapter, be64_to_cpu(crq_as_u64[0]),
|
|
be64_to_cpu(crq_as_u64[1]));
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_send_rem_buffer_resp - Remove Buffer Response
|
|
*
|
|
* @adapter: crq_server_adapter struct
|
|
* @status: Status field
|
|
* @hmc_session: HMC Session field
|
|
* @hmc_index: HMC Index field
|
|
* @buffer_id: Buffer Id field
|
|
*
|
|
* This command is sent by the management partition to the hypervisor in
|
|
* response to the Remove Buffer message. The Buffer ID field indicates
|
|
* which buffer the management partition selected to remove. The Status
|
|
* field indicates the result of the command.
|
|
*
|
|
* Return:
|
|
* 0 - Success
|
|
*/
|
|
static int ibmvmc_send_rem_buffer_resp(struct crq_server_adapter *adapter,
|
|
u8 status, u8 hmc_session,
|
|
u8 hmc_index, u16 buffer_id)
|
|
{
|
|
struct ibmvmc_crq_msg crq_msg;
|
|
__be64 *crq_as_u64 = (__be64 *)&crq_msg;
|
|
|
|
dev_dbg(adapter->dev, "CRQ send: rem_buffer_resp\n");
|
|
crq_msg.valid = 0x80;
|
|
crq_msg.type = VMC_MSG_REM_BUF_RESP;
|
|
crq_msg.status = status;
|
|
crq_msg.var1.rsvd = 0;
|
|
crq_msg.hmc_session = hmc_session;
|
|
crq_msg.hmc_index = hmc_index;
|
|
crq_msg.var2.buffer_id = cpu_to_be16(buffer_id);
|
|
crq_msg.rsvd = 0;
|
|
crq_msg.var3.rsvd = 0;
|
|
|
|
ibmvmc_send_crq(adapter, be64_to_cpu(crq_as_u64[0]),
|
|
be64_to_cpu(crq_as_u64[1]));
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_send_msg - Signal Message
|
|
*
|
|
* @adapter: crq_server_adapter struct
|
|
* @buffer: ibmvmc_buffer struct
|
|
* @hmc: ibmvmc_hmc struct
|
|
* @msg_len: message length field
|
|
*
|
|
* This command is sent between the management partition and the hypervisor
|
|
* in order to signal the arrival of an HMC protocol message. The command
|
|
* can be sent by both the management partition and the hypervisor. It is
|
|
* used for all traffic between the management application and the hypervisor,
|
|
* regardless of who initiated the communication.
|
|
*
|
|
* There is no response to this message.
|
|
*
|
|
* Return:
|
|
* 0 - Success
|
|
* Non-zero - Failure
|
|
*/
|
|
static int ibmvmc_send_msg(struct crq_server_adapter *adapter,
|
|
struct ibmvmc_buffer *buffer,
|
|
struct ibmvmc_hmc *hmc, int msg_len)
|
|
{
|
|
struct ibmvmc_crq_msg crq_msg;
|
|
__be64 *crq_as_u64 = (__be64 *)&crq_msg;
|
|
int rc = 0;
|
|
|
|
dev_dbg(adapter->dev, "CRQ send: rdma to HV\n");
|
|
rc = h_copy_rdma(msg_len,
|
|
adapter->liobn,
|
|
buffer->dma_addr_local,
|
|
adapter->riobn,
|
|
buffer->dma_addr_remote);
|
|
if (rc) {
|
|
dev_err(adapter->dev, "Error in send_msg, h_copy_rdma rc 0x%x\n",
|
|
rc);
|
|
return rc;
|
|
}
|
|
|
|
crq_msg.valid = 0x80;
|
|
crq_msg.type = VMC_MSG_SIGNAL;
|
|
crq_msg.status = 0;
|
|
crq_msg.var1.rsvd = 0;
|
|
crq_msg.hmc_session = hmc->session;
|
|
crq_msg.hmc_index = hmc->index;
|
|
crq_msg.var2.buffer_id = cpu_to_be16(buffer->id);
|
|
crq_msg.var3.msg_len = cpu_to_be32(msg_len);
|
|
dev_dbg(adapter->dev, "CRQ send: msg to HV 0x%llx 0x%llx\n",
|
|
be64_to_cpu(crq_as_u64[0]), be64_to_cpu(crq_as_u64[1]));
|
|
|
|
buffer->owner = VMC_BUF_OWNER_HV;
|
|
ibmvmc_send_crq(adapter, be64_to_cpu(crq_as_u64[0]),
|
|
be64_to_cpu(crq_as_u64[1]));
|
|
|
|
return rc;
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_open - Open Session
|
|
*
|
|
* @inode: inode struct
|
|
* @file: file struct
|
|
*
|
|
* Return:
|
|
* 0 - Success
|
|
* Non-zero - Failure
|
|
*/
|
|
static int ibmvmc_open(struct inode *inode, struct file *file)
|
|
{
|
|
struct ibmvmc_file_session *session;
|
|
|
|
pr_debug("%s: inode = 0x%lx, file = 0x%lx, state = 0x%x\n", __func__,
|
|
(unsigned long)inode, (unsigned long)file,
|
|
ibmvmc.state);
|
|
|
|
session = kzalloc(sizeof(*session), GFP_KERNEL);
|
|
if (!session)
|
|
return -ENOMEM;
|
|
|
|
session->file = file;
|
|
file->private_data = session;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_close - Close Session
|
|
*
|
|
* @inode: inode struct
|
|
* @file: file struct
|
|
*
|
|
* Return:
|
|
* 0 - Success
|
|
* Non-zero - Failure
|
|
*/
|
|
static int ibmvmc_close(struct inode *inode, struct file *file)
|
|
{
|
|
struct ibmvmc_file_session *session;
|
|
struct ibmvmc_hmc *hmc;
|
|
int rc = 0;
|
|
unsigned long flags;
|
|
|
|
pr_debug("%s: file = 0x%lx, state = 0x%x\n", __func__,
|
|
(unsigned long)file, ibmvmc.state);
|
|
|
|
session = file->private_data;
|
|
if (!session)
|
|
return -EIO;
|
|
|
|
hmc = session->hmc;
|
|
if (hmc) {
|
|
if (!hmc->adapter)
|
|
return -EIO;
|
|
|
|
if (ibmvmc.state == ibmvmc_state_failed) {
|
|
dev_warn(hmc->adapter->dev, "close: state_failed\n");
|
|
return -EIO;
|
|
}
|
|
|
|
spin_lock_irqsave(&hmc->lock, flags);
|
|
if (hmc->state >= ibmhmc_state_opening) {
|
|
rc = ibmvmc_send_close(hmc);
|
|
if (rc)
|
|
dev_warn(hmc->adapter->dev, "close: send_close failed.\n");
|
|
}
|
|
spin_unlock_irqrestore(&hmc->lock, flags);
|
|
}
|
|
|
|
kfree_sensitive(session);
|
|
|
|
return rc;
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_read - Read
|
|
*
|
|
* @file: file struct
|
|
* @buf: Character buffer
|
|
* @nbytes: Size in bytes
|
|
* @ppos: Offset
|
|
*
|
|
* Return:
|
|
* 0 - Success
|
|
* Non-zero - Failure
|
|
*/
|
|
static ssize_t ibmvmc_read(struct file *file, char *buf, size_t nbytes,
|
|
loff_t *ppos)
|
|
{
|
|
struct ibmvmc_file_session *session;
|
|
struct ibmvmc_hmc *hmc;
|
|
struct crq_server_adapter *adapter;
|
|
struct ibmvmc_buffer *buffer;
|
|
ssize_t n;
|
|
ssize_t retval = 0;
|
|
unsigned long flags;
|
|
DEFINE_WAIT(wait);
|
|
|
|
pr_debug("ibmvmc: read: file = 0x%lx, buf = 0x%lx, nbytes = 0x%lx\n",
|
|
(unsigned long)file, (unsigned long)buf,
|
|
(unsigned long)nbytes);
|
|
|
|
if (nbytes == 0)
|
|
return 0;
|
|
|
|
if (nbytes > ibmvmc.max_mtu) {
|
|
pr_warn("ibmvmc: read: nbytes invalid 0x%x\n",
|
|
(unsigned int)nbytes);
|
|
return -EINVAL;
|
|
}
|
|
|
|
session = file->private_data;
|
|
if (!session) {
|
|
pr_warn("ibmvmc: read: no session\n");
|
|
return -EIO;
|
|
}
|
|
|
|
hmc = session->hmc;
|
|
if (!hmc) {
|
|
pr_warn("ibmvmc: read: no hmc\n");
|
|
return -EIO;
|
|
}
|
|
|
|
adapter = hmc->adapter;
|
|
if (!adapter) {
|
|
pr_warn("ibmvmc: read: no adapter\n");
|
|
return -EIO;
|
|
}
|
|
|
|
do {
|
|
prepare_to_wait(&ibmvmc_read_wait, &wait, TASK_INTERRUPTIBLE);
|
|
|
|
spin_lock_irqsave(&hmc->lock, flags);
|
|
if (hmc->queue_tail != hmc->queue_head)
|
|
/* Data is available */
|
|
break;
|
|
|
|
spin_unlock_irqrestore(&hmc->lock, flags);
|
|
|
|
if (!session->valid) {
|
|
retval = -EBADFD;
|
|
goto out;
|
|
}
|
|
if (file->f_flags & O_NONBLOCK) {
|
|
retval = -EAGAIN;
|
|
goto out;
|
|
}
|
|
|
|
schedule();
|
|
|
|
if (signal_pending(current)) {
|
|
retval = -ERESTARTSYS;
|
|
goto out;
|
|
}
|
|
} while (1);
|
|
|
|
buffer = &(hmc->buffer[hmc->queue_outbound_msgs[hmc->queue_tail]]);
|
|
hmc->queue_tail++;
|
|
if (hmc->queue_tail == ibmvmc_max_buf_pool_size)
|
|
hmc->queue_tail = 0;
|
|
spin_unlock_irqrestore(&hmc->lock, flags);
|
|
|
|
nbytes = min_t(size_t, nbytes, buffer->msg_len);
|
|
n = copy_to_user((void *)buf, buffer->real_addr_local, nbytes);
|
|
dev_dbg(adapter->dev, "read: copy to user nbytes = 0x%lx.\n", nbytes);
|
|
ibmvmc_free_hmc_buffer(hmc, buffer);
|
|
retval = nbytes;
|
|
|
|
if (n) {
|
|
dev_warn(adapter->dev, "read: copy to user failed.\n");
|
|
retval = -EFAULT;
|
|
}
|
|
|
|
out:
|
|
finish_wait(&ibmvmc_read_wait, &wait);
|
|
dev_dbg(adapter->dev, "read: out %ld\n", retval);
|
|
return retval;
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_poll - Poll
|
|
*
|
|
* @file: file struct
|
|
* @wait: Poll Table
|
|
*
|
|
* Return:
|
|
* poll.h return values
|
|
*/
|
|
static unsigned int ibmvmc_poll(struct file *file, poll_table *wait)
|
|
{
|
|
struct ibmvmc_file_session *session;
|
|
struct ibmvmc_hmc *hmc;
|
|
unsigned int mask = 0;
|
|
|
|
session = file->private_data;
|
|
if (!session)
|
|
return 0;
|
|
|
|
hmc = session->hmc;
|
|
if (!hmc)
|
|
return 0;
|
|
|
|
poll_wait(file, &ibmvmc_read_wait, wait);
|
|
|
|
if (hmc->queue_head != hmc->queue_tail)
|
|
mask |= POLLIN | POLLRDNORM;
|
|
|
|
return mask;
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_write - Write
|
|
*
|
|
* @file: file struct
|
|
* @buffer: Character buffer
|
|
* @count: Count field
|
|
* @ppos: Offset
|
|
*
|
|
* Return:
|
|
* 0 - Success
|
|
* Non-zero - Failure
|
|
*/
|
|
static ssize_t ibmvmc_write(struct file *file, const char *buffer,
|
|
size_t count, loff_t *ppos)
|
|
{
|
|
struct ibmvmc_buffer *vmc_buffer;
|
|
struct ibmvmc_file_session *session;
|
|
struct crq_server_adapter *adapter;
|
|
struct ibmvmc_hmc *hmc;
|
|
unsigned char *buf;
|
|
unsigned long flags;
|
|
size_t bytes;
|
|
const char *p = buffer;
|
|
size_t c = count;
|
|
int ret = 0;
|
|
|
|
session = file->private_data;
|
|
if (!session)
|
|
return -EIO;
|
|
|
|
hmc = session->hmc;
|
|
if (!hmc)
|
|
return -EIO;
|
|
|
|
spin_lock_irqsave(&hmc->lock, flags);
|
|
if (hmc->state == ibmhmc_state_free) {
|
|
/* HMC connection is not valid (possibly was reset under us). */
|
|
ret = -EIO;
|
|
goto out;
|
|
}
|
|
|
|
adapter = hmc->adapter;
|
|
if (!adapter) {
|
|
ret = -EIO;
|
|
goto out;
|
|
}
|
|
|
|
if (count > ibmvmc.max_mtu) {
|
|
dev_warn(adapter->dev, "invalid buffer size 0x%lx\n",
|
|
(unsigned long)count);
|
|
ret = -EIO;
|
|
goto out;
|
|
}
|
|
|
|
/* Waiting for the open resp message to the ioctl(1) - retry */
|
|
if (hmc->state == ibmhmc_state_opening) {
|
|
ret = -EBUSY;
|
|
goto out;
|
|
}
|
|
|
|
/* Make sure the ioctl() was called & the open msg sent, and that
|
|
* the HMC connection has not failed.
|
|
*/
|
|
if (hmc->state != ibmhmc_state_ready) {
|
|
ret = -EIO;
|
|
goto out;
|
|
}
|
|
|
|
vmc_buffer = ibmvmc_get_valid_hmc_buffer(hmc->index);
|
|
if (!vmc_buffer) {
|
|
/* No buffer available for the msg send, or we have not yet
|
|
* completed the open/open_resp sequence. Retry until this is
|
|
* complete.
|
|
*/
|
|
ret = -EBUSY;
|
|
goto out;
|
|
}
|
|
if (!vmc_buffer->real_addr_local) {
|
|
dev_err(adapter->dev, "no buffer storage assigned\n");
|
|
ret = -EIO;
|
|
goto out;
|
|
}
|
|
buf = vmc_buffer->real_addr_local;
|
|
|
|
while (c > 0) {
|
|
bytes = min_t(size_t, c, vmc_buffer->size);
|
|
|
|
bytes -= copy_from_user(buf, p, bytes);
|
|
if (!bytes) {
|
|
ret = -EFAULT;
|
|
goto out;
|
|
}
|
|
c -= bytes;
|
|
p += bytes;
|
|
}
|
|
if (p == buffer)
|
|
goto out;
|
|
|
|
file->f_path.dentry->d_inode->i_mtime = current_time(file_inode(file));
|
|
mark_inode_dirty(file->f_path.dentry->d_inode);
|
|
|
|
dev_dbg(adapter->dev, "write: file = 0x%lx, count = 0x%lx\n",
|
|
(unsigned long)file, (unsigned long)count);
|
|
|
|
ibmvmc_send_msg(adapter, vmc_buffer, hmc, count);
|
|
ret = p - buffer;
|
|
out:
|
|
spin_unlock_irqrestore(&hmc->lock, flags);
|
|
return (ssize_t)(ret);
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_setup_hmc - Setup the HMC
|
|
*
|
|
* @session: ibmvmc_file_session struct
|
|
*
|
|
* Return:
|
|
* 0 - Success
|
|
* Non-zero - Failure
|
|
*/
|
|
static long ibmvmc_setup_hmc(struct ibmvmc_file_session *session)
|
|
{
|
|
struct ibmvmc_hmc *hmc;
|
|
unsigned int valid, free, index;
|
|
|
|
if (ibmvmc.state == ibmvmc_state_failed) {
|
|
pr_warn("ibmvmc: Reserve HMC: state_failed\n");
|
|
return -EIO;
|
|
}
|
|
|
|
if (ibmvmc.state < ibmvmc_state_ready) {
|
|
pr_warn("ibmvmc: Reserve HMC: not state_ready\n");
|
|
return -EAGAIN;
|
|
}
|
|
|
|
/* Device is busy until capabilities have been exchanged and we
|
|
* have a generic buffer for each possible HMC connection.
|
|
*/
|
|
for (index = 0; index <= ibmvmc.max_hmc_index; index++) {
|
|
valid = 0;
|
|
ibmvmc_count_hmc_buffers(index, &valid, &free);
|
|
if (valid == 0) {
|
|
pr_warn("ibmvmc: buffers not ready for index %d\n",
|
|
index);
|
|
return -ENOBUFS;
|
|
}
|
|
}
|
|
|
|
/* Get an hmc object, and transition to ibmhmc_state_initial */
|
|
hmc = ibmvmc_get_free_hmc();
|
|
if (!hmc) {
|
|
pr_warn("%s: free hmc not found\n", __func__);
|
|
return -EBUSY;
|
|
}
|
|
|
|
hmc->session = hmc->session + 1;
|
|
if (hmc->session == 0xff)
|
|
hmc->session = 1;
|
|
|
|
session->hmc = hmc;
|
|
hmc->adapter = &ibmvmc_adapter;
|
|
hmc->file_session = session;
|
|
session->valid = 1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_ioctl_sethmcid - IOCTL Set HMC ID
|
|
*
|
|
* @session: ibmvmc_file_session struct
|
|
* @new_hmc_id: HMC id field
|
|
*
|
|
* IOCTL command to setup the hmc id
|
|
*
|
|
* Return:
|
|
* 0 - Success
|
|
* Non-zero - Failure
|
|
*/
|
|
static long ibmvmc_ioctl_sethmcid(struct ibmvmc_file_session *session,
|
|
unsigned char __user *new_hmc_id)
|
|
{
|
|
struct ibmvmc_hmc *hmc;
|
|
struct ibmvmc_buffer *buffer;
|
|
size_t bytes;
|
|
char print_buffer[HMC_ID_LEN + 1];
|
|
unsigned long flags;
|
|
long rc = 0;
|
|
|
|
/* Reserve HMC session */
|
|
hmc = session->hmc;
|
|
if (!hmc) {
|
|
rc = ibmvmc_setup_hmc(session);
|
|
if (rc)
|
|
return rc;
|
|
|
|
hmc = session->hmc;
|
|
if (!hmc) {
|
|
pr_err("ibmvmc: setup_hmc success but no hmc\n");
|
|
return -EIO;
|
|
}
|
|
}
|
|
|
|
if (hmc->state != ibmhmc_state_initial) {
|
|
pr_warn("ibmvmc: sethmcid: invalid state to send open 0x%x\n",
|
|
hmc->state);
|
|
return -EIO;
|
|
}
|
|
|
|
bytes = copy_from_user(hmc->hmc_id, new_hmc_id, HMC_ID_LEN);
|
|
if (bytes)
|
|
return -EFAULT;
|
|
|
|
/* Send Open Session command */
|
|
spin_lock_irqsave(&hmc->lock, flags);
|
|
buffer = ibmvmc_get_valid_hmc_buffer(hmc->index);
|
|
spin_unlock_irqrestore(&hmc->lock, flags);
|
|
|
|
if (!buffer || !buffer->real_addr_local) {
|
|
pr_warn("ibmvmc: sethmcid: no buffer available\n");
|
|
return -EIO;
|
|
}
|
|
|
|
/* Make sure buffer is NULL terminated before trying to print it */
|
|
memset(print_buffer, 0, HMC_ID_LEN + 1);
|
|
strncpy(print_buffer, hmc->hmc_id, HMC_ID_LEN);
|
|
pr_info("ibmvmc: sethmcid: Set HMC ID: \"%s\"\n", print_buffer);
|
|
|
|
memcpy(buffer->real_addr_local, hmc->hmc_id, HMC_ID_LEN);
|
|
/* RDMA over ID, send open msg, change state to ibmhmc_state_opening */
|
|
rc = ibmvmc_send_open(buffer, hmc);
|
|
|
|
return rc;
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_ioctl_query - IOCTL Query
|
|
*
|
|
* @session: ibmvmc_file_session struct
|
|
* @ret_struct: ibmvmc_query_struct
|
|
*
|
|
* Return:
|
|
* 0 - Success
|
|
* Non-zero - Failure
|
|
*/
|
|
static long ibmvmc_ioctl_query(struct ibmvmc_file_session *session,
|
|
struct ibmvmc_query_struct __user *ret_struct)
|
|
{
|
|
struct ibmvmc_query_struct query_struct;
|
|
size_t bytes;
|
|
|
|
memset(&query_struct, 0, sizeof(query_struct));
|
|
query_struct.have_vmc = (ibmvmc.state > ibmvmc_state_initial);
|
|
query_struct.state = ibmvmc.state;
|
|
query_struct.vmc_drc_index = ibmvmc.vmc_drc_index;
|
|
|
|
bytes = copy_to_user(ret_struct, &query_struct,
|
|
sizeof(query_struct));
|
|
if (bytes)
|
|
return -EFAULT;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_ioctl_requestvmc - IOCTL Request VMC
|
|
*
|
|
* @session: ibmvmc_file_session struct
|
|
* @ret_vmc_index: VMC Index
|
|
*
|
|
* Return:
|
|
* 0 - Success
|
|
* Non-zero - Failure
|
|
*/
|
|
static long ibmvmc_ioctl_requestvmc(struct ibmvmc_file_session *session,
|
|
u32 __user *ret_vmc_index)
|
|
{
|
|
/* TODO: (adreznec) Add locking to control multiple process access */
|
|
size_t bytes;
|
|
long rc;
|
|
u32 vmc_drc_index;
|
|
|
|
/* Call to request the VMC device from phyp*/
|
|
rc = h_request_vmc(&vmc_drc_index);
|
|
pr_debug("ibmvmc: requestvmc: H_REQUEST_VMC rc = 0x%lx\n", rc);
|
|
|
|
if (rc == H_SUCCESS) {
|
|
rc = 0;
|
|
} else if (rc == H_FUNCTION) {
|
|
pr_err("ibmvmc: requestvmc: h_request_vmc not supported\n");
|
|
return -EPERM;
|
|
} else if (rc == H_AUTHORITY) {
|
|
pr_err("ibmvmc: requestvmc: hypervisor denied vmc request\n");
|
|
return -EPERM;
|
|
} else if (rc == H_HARDWARE) {
|
|
pr_err("ibmvmc: requestvmc: hypervisor hardware fault\n");
|
|
return -EIO;
|
|
} else if (rc == H_RESOURCE) {
|
|
pr_err("ibmvmc: requestvmc: vmc resource unavailable\n");
|
|
return -ENODEV;
|
|
} else if (rc == H_NOT_AVAILABLE) {
|
|
pr_err("ibmvmc: requestvmc: system cannot be vmc managed\n");
|
|
return -EPERM;
|
|
} else if (rc == H_PARAMETER) {
|
|
pr_err("ibmvmc: requestvmc: invalid parameter\n");
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* Success, set the vmc index in global struct */
|
|
ibmvmc.vmc_drc_index = vmc_drc_index;
|
|
|
|
bytes = copy_to_user(ret_vmc_index, &vmc_drc_index,
|
|
sizeof(*ret_vmc_index));
|
|
if (bytes) {
|
|
pr_warn("ibmvmc: requestvmc: copy to user failed.\n");
|
|
return -EFAULT;
|
|
}
|
|
return rc;
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_ioctl - IOCTL
|
|
*
|
|
* @file: file information
|
|
* @cmd: cmd field
|
|
* @arg: Argument field
|
|
*
|
|
* Return:
|
|
* 0 - Success
|
|
* Non-zero - Failure
|
|
*/
|
|
static long ibmvmc_ioctl(struct file *file,
|
|
unsigned int cmd, unsigned long arg)
|
|
{
|
|
struct ibmvmc_file_session *session = file->private_data;
|
|
|
|
pr_debug("ibmvmc: ioctl file=0x%lx, cmd=0x%x, arg=0x%lx, ses=0x%lx\n",
|
|
(unsigned long)file, cmd, arg,
|
|
(unsigned long)session);
|
|
|
|
if (!session) {
|
|
pr_warn("ibmvmc: ioctl: no session\n");
|
|
return -EIO;
|
|
}
|
|
|
|
switch (cmd) {
|
|
case VMC_IOCTL_SETHMCID:
|
|
return ibmvmc_ioctl_sethmcid(session,
|
|
(unsigned char __user *)arg);
|
|
case VMC_IOCTL_QUERY:
|
|
return ibmvmc_ioctl_query(session,
|
|
(struct ibmvmc_query_struct __user *)arg);
|
|
case VMC_IOCTL_REQUESTVMC:
|
|
return ibmvmc_ioctl_requestvmc(session,
|
|
(unsigned int __user *)arg);
|
|
default:
|
|
pr_warn("ibmvmc: unknown ioctl 0x%x\n", cmd);
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
|
|
static const struct file_operations ibmvmc_fops = {
|
|
.owner = THIS_MODULE,
|
|
.read = ibmvmc_read,
|
|
.write = ibmvmc_write,
|
|
.poll = ibmvmc_poll,
|
|
.unlocked_ioctl = ibmvmc_ioctl,
|
|
.open = ibmvmc_open,
|
|
.release = ibmvmc_close,
|
|
};
|
|
|
|
/**
|
|
* ibmvmc_add_buffer - Add Buffer
|
|
*
|
|
* @adapter: crq_server_adapter struct
|
|
* @crq: ibmvmc_crq_msg struct
|
|
*
|
|
* This message transfers a buffer from hypervisor ownership to management
|
|
* partition ownership. The LIOBA is obtained from the virtual TCE table
|
|
* associated with the hypervisor side of the VMC device, and points to a
|
|
* buffer of size MTU (as established in the capabilities exchange).
|
|
*
|
|
* Typical flow for ading buffers:
|
|
* 1. A new management application connection is opened by the management
|
|
* partition.
|
|
* 2. The hypervisor assigns new buffers for the traffic associated with
|
|
* that connection.
|
|
* 3. The hypervisor sends VMC Add Buffer messages to the management
|
|
* partition, informing it of the new buffers.
|
|
* 4. The hypervisor sends an HMC protocol message (to the management
|
|
* application) notifying it of the new buffers. This informs the
|
|
* application that it has buffers available for sending HMC
|
|
* commands.
|
|
*
|
|
* Return:
|
|
* 0 - Success
|
|
* Non-zero - Failure
|
|
*/
|
|
static int ibmvmc_add_buffer(struct crq_server_adapter *adapter,
|
|
struct ibmvmc_crq_msg *crq)
|
|
{
|
|
struct ibmvmc_buffer *buffer;
|
|
u8 hmc_index;
|
|
u8 hmc_session;
|
|
u16 buffer_id;
|
|
unsigned long flags;
|
|
int rc = 0;
|
|
|
|
if (!crq)
|
|
return -1;
|
|
|
|
hmc_session = crq->hmc_session;
|
|
hmc_index = crq->hmc_index;
|
|
buffer_id = be16_to_cpu(crq->var2.buffer_id);
|
|
|
|
if (hmc_index > ibmvmc.max_hmc_index) {
|
|
dev_err(adapter->dev, "add_buffer: invalid hmc_index = 0x%x\n",
|
|
hmc_index);
|
|
ibmvmc_send_add_buffer_resp(adapter, VMC_MSG_INVALID_HMC_INDEX,
|
|
hmc_session, hmc_index, buffer_id);
|
|
return -1;
|
|
}
|
|
|
|
if (buffer_id >= ibmvmc.max_buffer_pool_size) {
|
|
dev_err(adapter->dev, "add_buffer: invalid buffer_id = 0x%x\n",
|
|
buffer_id);
|
|
ibmvmc_send_add_buffer_resp(adapter, VMC_MSG_INVALID_BUFFER_ID,
|
|
hmc_session, hmc_index, buffer_id);
|
|
return -1;
|
|
}
|
|
|
|
spin_lock_irqsave(&hmcs[hmc_index].lock, flags);
|
|
buffer = &hmcs[hmc_index].buffer[buffer_id];
|
|
|
|
if (buffer->real_addr_local || buffer->dma_addr_local) {
|
|
dev_warn(adapter->dev, "add_buffer: already allocated id = 0x%lx\n",
|
|
(unsigned long)buffer_id);
|
|
spin_unlock_irqrestore(&hmcs[hmc_index].lock, flags);
|
|
ibmvmc_send_add_buffer_resp(adapter, VMC_MSG_INVALID_BUFFER_ID,
|
|
hmc_session, hmc_index, buffer_id);
|
|
return -1;
|
|
}
|
|
|
|
buffer->real_addr_local = alloc_dma_buffer(to_vio_dev(adapter->dev),
|
|
ibmvmc.max_mtu,
|
|
&buffer->dma_addr_local);
|
|
|
|
if (!buffer->real_addr_local) {
|
|
dev_err(adapter->dev, "add_buffer: alloc_dma_buffer failed.\n");
|
|
spin_unlock_irqrestore(&hmcs[hmc_index].lock, flags);
|
|
ibmvmc_send_add_buffer_resp(adapter, VMC_MSG_INTERFACE_FAILURE,
|
|
hmc_session, hmc_index, buffer_id);
|
|
return -1;
|
|
}
|
|
|
|
buffer->dma_addr_remote = be32_to_cpu(crq->var3.lioba);
|
|
buffer->size = ibmvmc.max_mtu;
|
|
buffer->owner = crq->var1.owner;
|
|
buffer->free = 1;
|
|
/* Must ensure valid==1 is observable only after all other fields are */
|
|
dma_wmb();
|
|
buffer->valid = 1;
|
|
buffer->id = buffer_id;
|
|
|
|
dev_dbg(adapter->dev, "add_buffer: successfully added a buffer:\n");
|
|
dev_dbg(adapter->dev, " index: %d, session: %d, buffer: 0x%x, owner: %d\n",
|
|
hmc_index, hmc_session, buffer_id, buffer->owner);
|
|
dev_dbg(adapter->dev, " local: 0x%x, remote: 0x%x\n",
|
|
(u32)buffer->dma_addr_local,
|
|
(u32)buffer->dma_addr_remote);
|
|
spin_unlock_irqrestore(&hmcs[hmc_index].lock, flags);
|
|
|
|
ibmvmc_send_add_buffer_resp(adapter, VMC_MSG_SUCCESS, hmc_session,
|
|
hmc_index, buffer_id);
|
|
|
|
return rc;
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_rem_buffer - Remove Buffer
|
|
*
|
|
* @adapter: crq_server_adapter struct
|
|
* @crq: ibmvmc_crq_msg struct
|
|
*
|
|
* This message requests an HMC buffer to be transferred from management
|
|
* partition ownership to hypervisor ownership. The management partition may
|
|
* not be able to satisfy the request at a particular point in time if all its
|
|
* buffers are in use. The management partition requires a depth of at least
|
|
* one inbound buffer to allow management application commands to flow to the
|
|
* hypervisor. It is, therefore, an interface error for the hypervisor to
|
|
* attempt to remove the management partition's last buffer.
|
|
*
|
|
* The hypervisor is expected to manage buffer usage with the management
|
|
* application directly and inform the management partition when buffers may be
|
|
* removed. The typical flow for removing buffers:
|
|
*
|
|
* 1. The management application no longer needs a communication path to a
|
|
* particular hypervisor function. That function is closed.
|
|
* 2. The hypervisor and the management application quiesce all traffic to that
|
|
* function. The hypervisor requests a reduction in buffer pool size.
|
|
* 3. The management application acknowledges the reduction in buffer pool size.
|
|
* 4. The hypervisor sends a Remove Buffer message to the management partition,
|
|
* informing it of the reduction in buffers.
|
|
* 5. The management partition verifies it can remove the buffer. This is
|
|
* possible if buffers have been quiesced.
|
|
*
|
|
* Return:
|
|
* 0 - Success
|
|
* Non-zero - Failure
|
|
*/
|
|
/*
|
|
* The hypervisor requested that we pick an unused buffer, and return it.
|
|
* Before sending the buffer back, we free any storage associated with the
|
|
* buffer.
|
|
*/
|
|
static int ibmvmc_rem_buffer(struct crq_server_adapter *adapter,
|
|
struct ibmvmc_crq_msg *crq)
|
|
{
|
|
struct ibmvmc_buffer *buffer;
|
|
u8 hmc_index;
|
|
u8 hmc_session;
|
|
u16 buffer_id = 0;
|
|
unsigned long flags;
|
|
int rc = 0;
|
|
|
|
if (!crq)
|
|
return -1;
|
|
|
|
hmc_session = crq->hmc_session;
|
|
hmc_index = crq->hmc_index;
|
|
|
|
if (hmc_index > ibmvmc.max_hmc_index) {
|
|
dev_warn(adapter->dev, "rem_buffer: invalid hmc_index = 0x%x\n",
|
|
hmc_index);
|
|
ibmvmc_send_rem_buffer_resp(adapter, VMC_MSG_INVALID_HMC_INDEX,
|
|
hmc_session, hmc_index, buffer_id);
|
|
return -1;
|
|
}
|
|
|
|
spin_lock_irqsave(&hmcs[hmc_index].lock, flags);
|
|
buffer = ibmvmc_get_free_hmc_buffer(adapter, hmc_index);
|
|
if (!buffer) {
|
|
dev_info(adapter->dev, "rem_buffer: no buffer to remove\n");
|
|
spin_unlock_irqrestore(&hmcs[hmc_index].lock, flags);
|
|
ibmvmc_send_rem_buffer_resp(adapter, VMC_MSG_NO_BUFFER,
|
|
hmc_session, hmc_index,
|
|
VMC_INVALID_BUFFER_ID);
|
|
return -1;
|
|
}
|
|
|
|
buffer_id = buffer->id;
|
|
|
|
if (buffer->valid)
|
|
free_dma_buffer(to_vio_dev(adapter->dev),
|
|
ibmvmc.max_mtu,
|
|
buffer->real_addr_local,
|
|
buffer->dma_addr_local);
|
|
|
|
memset(buffer, 0, sizeof(struct ibmvmc_buffer));
|
|
spin_unlock_irqrestore(&hmcs[hmc_index].lock, flags);
|
|
|
|
dev_dbg(adapter->dev, "rem_buffer: removed buffer 0x%x.\n", buffer_id);
|
|
ibmvmc_send_rem_buffer_resp(adapter, VMC_MSG_SUCCESS, hmc_session,
|
|
hmc_index, buffer_id);
|
|
|
|
return rc;
|
|
}
|
|
|
|
static int ibmvmc_recv_msg(struct crq_server_adapter *adapter,
|
|
struct ibmvmc_crq_msg *crq)
|
|
{
|
|
struct ibmvmc_buffer *buffer;
|
|
struct ibmvmc_hmc *hmc;
|
|
unsigned long msg_len;
|
|
u8 hmc_index;
|
|
u8 hmc_session;
|
|
u16 buffer_id;
|
|
unsigned long flags;
|
|
int rc = 0;
|
|
|
|
if (!crq)
|
|
return -1;
|
|
|
|
/* Hypervisor writes CRQs directly into our memory in big endian */
|
|
dev_dbg(adapter->dev, "Recv_msg: msg from HV 0x%016llx 0x%016llx\n",
|
|
be64_to_cpu(*((unsigned long *)crq)),
|
|
be64_to_cpu(*(((unsigned long *)crq) + 1)));
|
|
|
|
hmc_session = crq->hmc_session;
|
|
hmc_index = crq->hmc_index;
|
|
buffer_id = be16_to_cpu(crq->var2.buffer_id);
|
|
msg_len = be32_to_cpu(crq->var3.msg_len);
|
|
|
|
if (hmc_index > ibmvmc.max_hmc_index) {
|
|
dev_err(adapter->dev, "Recv_msg: invalid hmc_index = 0x%x\n",
|
|
hmc_index);
|
|
ibmvmc_send_add_buffer_resp(adapter, VMC_MSG_INVALID_HMC_INDEX,
|
|
hmc_session, hmc_index, buffer_id);
|
|
return -1;
|
|
}
|
|
|
|
if (buffer_id >= ibmvmc.max_buffer_pool_size) {
|
|
dev_err(adapter->dev, "Recv_msg: invalid buffer_id = 0x%x\n",
|
|
buffer_id);
|
|
ibmvmc_send_add_buffer_resp(adapter, VMC_MSG_INVALID_BUFFER_ID,
|
|
hmc_session, hmc_index, buffer_id);
|
|
return -1;
|
|
}
|
|
|
|
hmc = &hmcs[hmc_index];
|
|
spin_lock_irqsave(&hmc->lock, flags);
|
|
|
|
if (hmc->state == ibmhmc_state_free) {
|
|
dev_err(adapter->dev, "Recv_msg: invalid hmc state = 0x%x\n",
|
|
hmc->state);
|
|
/* HMC connection is not valid (possibly was reset under us). */
|
|
spin_unlock_irqrestore(&hmc->lock, flags);
|
|
return -1;
|
|
}
|
|
|
|
buffer = &hmc->buffer[buffer_id];
|
|
|
|
if (buffer->valid == 0 || buffer->owner == VMC_BUF_OWNER_ALPHA) {
|
|
dev_err(adapter->dev, "Recv_msg: not valid, or not HV. 0x%x 0x%x\n",
|
|
buffer->valid, buffer->owner);
|
|
spin_unlock_irqrestore(&hmc->lock, flags);
|
|
return -1;
|
|
}
|
|
|
|
/* RDMA the data into the partition. */
|
|
rc = h_copy_rdma(msg_len,
|
|
adapter->riobn,
|
|
buffer->dma_addr_remote,
|
|
adapter->liobn,
|
|
buffer->dma_addr_local);
|
|
|
|
dev_dbg(adapter->dev, "Recv_msg: msg_len = 0x%x, buffer_id = 0x%x, queue_head = 0x%x, hmc_idx = 0x%x\n",
|
|
(unsigned int)msg_len, (unsigned int)buffer_id,
|
|
(unsigned int)hmc->queue_head, (unsigned int)hmc_index);
|
|
buffer->msg_len = msg_len;
|
|
buffer->free = 0;
|
|
buffer->owner = VMC_BUF_OWNER_ALPHA;
|
|
|
|
if (rc) {
|
|
dev_err(adapter->dev, "Failure in recv_msg: h_copy_rdma = 0x%x\n",
|
|
rc);
|
|
spin_unlock_irqrestore(&hmc->lock, flags);
|
|
return -1;
|
|
}
|
|
|
|
/* Must be locked because read operates on the same data */
|
|
hmc->queue_outbound_msgs[hmc->queue_head] = buffer_id;
|
|
hmc->queue_head++;
|
|
if (hmc->queue_head == ibmvmc_max_buf_pool_size)
|
|
hmc->queue_head = 0;
|
|
|
|
if (hmc->queue_head == hmc->queue_tail)
|
|
dev_err(adapter->dev, "outbound buffer queue wrapped.\n");
|
|
|
|
spin_unlock_irqrestore(&hmc->lock, flags);
|
|
|
|
wake_up_interruptible(&ibmvmc_read_wait);
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_process_capabilities - Process Capabilities
|
|
*
|
|
* @adapter: crq_server_adapter struct
|
|
* @crqp: ibmvmc_crq_msg struct
|
|
*
|
|
*/
|
|
static void ibmvmc_process_capabilities(struct crq_server_adapter *adapter,
|
|
struct ibmvmc_crq_msg *crqp)
|
|
{
|
|
struct ibmvmc_admin_crq_msg *crq = (struct ibmvmc_admin_crq_msg *)crqp;
|
|
|
|
if ((be16_to_cpu(crq->version) >> 8) !=
|
|
(IBMVMC_PROTOCOL_VERSION >> 8)) {
|
|
dev_err(adapter->dev, "init failed, incompatible versions 0x%x 0x%x\n",
|
|
be16_to_cpu(crq->version),
|
|
IBMVMC_PROTOCOL_VERSION);
|
|
ibmvmc.state = ibmvmc_state_failed;
|
|
return;
|
|
}
|
|
|
|
ibmvmc.max_mtu = min_t(u32, ibmvmc_max_mtu, be32_to_cpu(crq->max_mtu));
|
|
ibmvmc.max_buffer_pool_size = min_t(u16, ibmvmc_max_buf_pool_size,
|
|
be16_to_cpu(crq->pool_size));
|
|
ibmvmc.max_hmc_index = min_t(u8, ibmvmc_max_hmcs, crq->max_hmc) - 1;
|
|
ibmvmc.state = ibmvmc_state_ready;
|
|
|
|
dev_info(adapter->dev, "Capabilities: mtu=0x%x, pool_size=0x%x, max_hmc=0x%x\n",
|
|
ibmvmc.max_mtu, ibmvmc.max_buffer_pool_size,
|
|
ibmvmc.max_hmc_index);
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_validate_hmc_session - Validate HMC Session
|
|
*
|
|
* @adapter: crq_server_adapter struct
|
|
* @crq: ibmvmc_crq_msg struct
|
|
*
|
|
* Return:
|
|
* 0 - Success
|
|
* Non-zero - Failure
|
|
*/
|
|
static int ibmvmc_validate_hmc_session(struct crq_server_adapter *adapter,
|
|
struct ibmvmc_crq_msg *crq)
|
|
{
|
|
unsigned char hmc_index;
|
|
|
|
hmc_index = crq->hmc_index;
|
|
|
|
if (crq->hmc_session == 0)
|
|
return 0;
|
|
|
|
if (hmc_index > ibmvmc.max_hmc_index)
|
|
return -1;
|
|
|
|
if (hmcs[hmc_index].session != crq->hmc_session) {
|
|
dev_warn(adapter->dev, "Drop, bad session: expected 0x%x, recv 0x%x\n",
|
|
hmcs[hmc_index].session, crq->hmc_session);
|
|
return -1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_reset - Reset
|
|
*
|
|
* @adapter: crq_server_adapter struct
|
|
* @xport_event: export_event field
|
|
*
|
|
* Closes all HMC sessions and conditionally schedules a CRQ reset.
|
|
* @xport_event: If true, the partner closed their CRQ; we don't need to reset.
|
|
* If false, we need to schedule a CRQ reset.
|
|
*/
|
|
static void ibmvmc_reset(struct crq_server_adapter *adapter, bool xport_event)
|
|
{
|
|
int i;
|
|
|
|
if (ibmvmc.state != ibmvmc_state_sched_reset) {
|
|
dev_info(adapter->dev, "*** Reset to initial state.\n");
|
|
for (i = 0; i < ibmvmc_max_hmcs; i++)
|
|
ibmvmc_return_hmc(&hmcs[i], xport_event);
|
|
|
|
if (xport_event) {
|
|
/* CRQ was closed by the partner. We don't need to do
|
|
* anything except set ourself to the correct state to
|
|
* handle init msgs.
|
|
*/
|
|
ibmvmc.state = ibmvmc_state_crqinit;
|
|
} else {
|
|
/* The partner did not close their CRQ - instead, we're
|
|
* closing the CRQ on our end. Need to schedule this
|
|
* for process context, because CRQ reset may require a
|
|
* sleep.
|
|
*
|
|
* Setting ibmvmc.state here immediately prevents
|
|
* ibmvmc_open from completing until the reset
|
|
* completes in process context.
|
|
*/
|
|
ibmvmc.state = ibmvmc_state_sched_reset;
|
|
dev_dbg(adapter->dev, "Device reset scheduled");
|
|
wake_up_interruptible(&adapter->reset_wait_queue);
|
|
}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_reset_task - Reset Task
|
|
*
|
|
* @data: Data field
|
|
*
|
|
* Performs a CRQ reset of the VMC device in process context.
|
|
* NOTE: This function should not be called directly, use ibmvmc_reset.
|
|
*/
|
|
static int ibmvmc_reset_task(void *data)
|
|
{
|
|
struct crq_server_adapter *adapter = data;
|
|
int rc;
|
|
|
|
set_user_nice(current, -20);
|
|
|
|
while (!kthread_should_stop()) {
|
|
wait_event_interruptible(adapter->reset_wait_queue,
|
|
(ibmvmc.state == ibmvmc_state_sched_reset) ||
|
|
kthread_should_stop());
|
|
|
|
if (kthread_should_stop())
|
|
break;
|
|
|
|
dev_dbg(adapter->dev, "CRQ resetting in process context");
|
|
tasklet_disable(&adapter->work_task);
|
|
|
|
rc = ibmvmc_reset_crq_queue(adapter);
|
|
|
|
if (rc != H_SUCCESS && rc != H_RESOURCE) {
|
|
dev_err(adapter->dev, "Error initializing CRQ. rc = 0x%x\n",
|
|
rc);
|
|
ibmvmc.state = ibmvmc_state_failed;
|
|
} else {
|
|
ibmvmc.state = ibmvmc_state_crqinit;
|
|
|
|
if (ibmvmc_send_crq(adapter, 0xC001000000000000LL, 0)
|
|
!= 0 && rc != H_RESOURCE)
|
|
dev_warn(adapter->dev, "Failed to send initialize CRQ message\n");
|
|
}
|
|
|
|
vio_enable_interrupts(to_vio_dev(adapter->dev));
|
|
tasklet_enable(&adapter->work_task);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_process_open_resp - Process Open Response
|
|
*
|
|
* @crq: ibmvmc_crq_msg struct
|
|
* @adapter: crq_server_adapter struct
|
|
*
|
|
* This command is sent by the hypervisor in response to the Interface
|
|
* Open message. When this message is received, the indicated buffer is
|
|
* again available for management partition use.
|
|
*/
|
|
static void ibmvmc_process_open_resp(struct ibmvmc_crq_msg *crq,
|
|
struct crq_server_adapter *adapter)
|
|
{
|
|
unsigned char hmc_index;
|
|
unsigned short buffer_id;
|
|
|
|
hmc_index = crq->hmc_index;
|
|
if (hmc_index > ibmvmc.max_hmc_index) {
|
|
/* Why would PHYP give an index > max negotiated? */
|
|
ibmvmc_reset(adapter, false);
|
|
return;
|
|
}
|
|
|
|
if (crq->status) {
|
|
dev_warn(adapter->dev, "open_resp: failed - status 0x%x\n",
|
|
crq->status);
|
|
ibmvmc_return_hmc(&hmcs[hmc_index], false);
|
|
return;
|
|
}
|
|
|
|
if (hmcs[hmc_index].state == ibmhmc_state_opening) {
|
|
buffer_id = be16_to_cpu(crq->var2.buffer_id);
|
|
if (buffer_id >= ibmvmc.max_buffer_pool_size) {
|
|
dev_err(adapter->dev, "open_resp: invalid buffer_id = 0x%x\n",
|
|
buffer_id);
|
|
hmcs[hmc_index].state = ibmhmc_state_failed;
|
|
} else {
|
|
ibmvmc_free_hmc_buffer(&hmcs[hmc_index],
|
|
&hmcs[hmc_index].buffer[buffer_id]);
|
|
hmcs[hmc_index].state = ibmhmc_state_ready;
|
|
dev_dbg(adapter->dev, "open_resp: set hmc state = ready\n");
|
|
}
|
|
} else {
|
|
dev_warn(adapter->dev, "open_resp: invalid hmc state (0x%x)\n",
|
|
hmcs[hmc_index].state);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_process_close_resp - Process Close Response
|
|
*
|
|
* @crq: ibmvmc_crq_msg struct
|
|
* @adapter: crq_server_adapter struct
|
|
*
|
|
* This command is sent by the hypervisor in response to the managemant
|
|
* application Interface Close message.
|
|
*
|
|
* If the close fails, simply reset the entire driver as the state of the VMC
|
|
* must be in tough shape.
|
|
*/
|
|
static void ibmvmc_process_close_resp(struct ibmvmc_crq_msg *crq,
|
|
struct crq_server_adapter *adapter)
|
|
{
|
|
unsigned char hmc_index;
|
|
|
|
hmc_index = crq->hmc_index;
|
|
if (hmc_index > ibmvmc.max_hmc_index) {
|
|
ibmvmc_reset(adapter, false);
|
|
return;
|
|
}
|
|
|
|
if (crq->status) {
|
|
dev_warn(adapter->dev, "close_resp: failed - status 0x%x\n",
|
|
crq->status);
|
|
ibmvmc_reset(adapter, false);
|
|
return;
|
|
}
|
|
|
|
ibmvmc_return_hmc(&hmcs[hmc_index], false);
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_crq_process - Process CRQ
|
|
*
|
|
* @adapter: crq_server_adapter struct
|
|
* @crq: ibmvmc_crq_msg struct
|
|
*
|
|
* Process the CRQ message based upon the type of message received.
|
|
*
|
|
*/
|
|
static void ibmvmc_crq_process(struct crq_server_adapter *adapter,
|
|
struct ibmvmc_crq_msg *crq)
|
|
{
|
|
switch (crq->type) {
|
|
case VMC_MSG_CAP_RESP:
|
|
dev_dbg(adapter->dev, "CRQ recv: capabilities resp (0x%x)\n",
|
|
crq->type);
|
|
if (ibmvmc.state == ibmvmc_state_capabilities)
|
|
ibmvmc_process_capabilities(adapter, crq);
|
|
else
|
|
dev_warn(adapter->dev, "caps msg invalid in state 0x%x\n",
|
|
ibmvmc.state);
|
|
break;
|
|
case VMC_MSG_OPEN_RESP:
|
|
dev_dbg(adapter->dev, "CRQ recv: open resp (0x%x)\n",
|
|
crq->type);
|
|
if (ibmvmc_validate_hmc_session(adapter, crq) == 0)
|
|
ibmvmc_process_open_resp(crq, adapter);
|
|
break;
|
|
case VMC_MSG_ADD_BUF:
|
|
dev_dbg(adapter->dev, "CRQ recv: add buf (0x%x)\n",
|
|
crq->type);
|
|
if (ibmvmc_validate_hmc_session(adapter, crq) == 0)
|
|
ibmvmc_add_buffer(adapter, crq);
|
|
break;
|
|
case VMC_MSG_REM_BUF:
|
|
dev_dbg(adapter->dev, "CRQ recv: rem buf (0x%x)\n",
|
|
crq->type);
|
|
if (ibmvmc_validate_hmc_session(adapter, crq) == 0)
|
|
ibmvmc_rem_buffer(adapter, crq);
|
|
break;
|
|
case VMC_MSG_SIGNAL:
|
|
dev_dbg(adapter->dev, "CRQ recv: signal msg (0x%x)\n",
|
|
crq->type);
|
|
if (ibmvmc_validate_hmc_session(adapter, crq) == 0)
|
|
ibmvmc_recv_msg(adapter, crq);
|
|
break;
|
|
case VMC_MSG_CLOSE_RESP:
|
|
dev_dbg(adapter->dev, "CRQ recv: close resp (0x%x)\n",
|
|
crq->type);
|
|
if (ibmvmc_validate_hmc_session(adapter, crq) == 0)
|
|
ibmvmc_process_close_resp(crq, adapter);
|
|
break;
|
|
case VMC_MSG_CAP:
|
|
case VMC_MSG_OPEN:
|
|
case VMC_MSG_CLOSE:
|
|
case VMC_MSG_ADD_BUF_RESP:
|
|
case VMC_MSG_REM_BUF_RESP:
|
|
dev_warn(adapter->dev, "CRQ recv: unexpected msg (0x%x)\n",
|
|
crq->type);
|
|
break;
|
|
default:
|
|
dev_warn(adapter->dev, "CRQ recv: unknown msg (0x%x)\n",
|
|
crq->type);
|
|
break;
|
|
}
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_handle_crq_init - Handle CRQ Init
|
|
*
|
|
* @crq: ibmvmc_crq_msg struct
|
|
* @adapter: crq_server_adapter struct
|
|
*
|
|
* Handle the type of crq initialization based on whether
|
|
* it is a message or a response.
|
|
*
|
|
*/
|
|
static void ibmvmc_handle_crq_init(struct ibmvmc_crq_msg *crq,
|
|
struct crq_server_adapter *adapter)
|
|
{
|
|
switch (crq->type) {
|
|
case 0x01: /* Initialization message */
|
|
dev_dbg(adapter->dev, "CRQ recv: CRQ init msg - state 0x%x\n",
|
|
ibmvmc.state);
|
|
if (ibmvmc.state == ibmvmc_state_crqinit) {
|
|
/* Send back a response */
|
|
if (ibmvmc_send_crq(adapter, 0xC002000000000000,
|
|
0) == 0)
|
|
ibmvmc_send_capabilities(adapter);
|
|
else
|
|
dev_err(adapter->dev, " Unable to send init rsp\n");
|
|
} else {
|
|
dev_err(adapter->dev, "Invalid state 0x%x mtu = 0x%x\n",
|
|
ibmvmc.state, ibmvmc.max_mtu);
|
|
}
|
|
|
|
break;
|
|
case 0x02: /* Initialization response */
|
|
dev_dbg(adapter->dev, "CRQ recv: initialization resp msg - state 0x%x\n",
|
|
ibmvmc.state);
|
|
if (ibmvmc.state == ibmvmc_state_crqinit)
|
|
ibmvmc_send_capabilities(adapter);
|
|
break;
|
|
default:
|
|
dev_warn(adapter->dev, "Unknown crq message type 0x%lx\n",
|
|
(unsigned long)crq->type);
|
|
}
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_handle_crq - Handle CRQ
|
|
*
|
|
* @crq: ibmvmc_crq_msg struct
|
|
* @adapter: crq_server_adapter struct
|
|
*
|
|
* Read the command elements from the command queue and execute the
|
|
* requests based upon the type of crq message.
|
|
*
|
|
*/
|
|
static void ibmvmc_handle_crq(struct ibmvmc_crq_msg *crq,
|
|
struct crq_server_adapter *adapter)
|
|
{
|
|
switch (crq->valid) {
|
|
case 0xC0: /* initialization */
|
|
ibmvmc_handle_crq_init(crq, adapter);
|
|
break;
|
|
case 0xFF: /* Hypervisor telling us the connection is closed */
|
|
dev_warn(adapter->dev, "CRQ recv: virtual adapter failed - resetting.\n");
|
|
ibmvmc_reset(adapter, true);
|
|
break;
|
|
case 0x80: /* real payload */
|
|
ibmvmc_crq_process(adapter, crq);
|
|
break;
|
|
default:
|
|
dev_warn(adapter->dev, "CRQ recv: unknown msg 0x%02x.\n",
|
|
crq->valid);
|
|
break;
|
|
}
|
|
}
|
|
|
|
static void ibmvmc_task(unsigned long data)
|
|
{
|
|
struct crq_server_adapter *adapter =
|
|
(struct crq_server_adapter *)data;
|
|
struct vio_dev *vdev = to_vio_dev(adapter->dev);
|
|
struct ibmvmc_crq_msg *crq;
|
|
int done = 0;
|
|
|
|
while (!done) {
|
|
/* Pull all the valid messages off the CRQ */
|
|
while ((crq = crq_queue_next_crq(&adapter->queue)) != NULL) {
|
|
ibmvmc_handle_crq(crq, adapter);
|
|
crq->valid = 0x00;
|
|
/* CRQ reset was requested, stop processing CRQs.
|
|
* Interrupts will be re-enabled by the reset task.
|
|
*/
|
|
if (ibmvmc.state == ibmvmc_state_sched_reset)
|
|
return;
|
|
}
|
|
|
|
vio_enable_interrupts(vdev);
|
|
crq = crq_queue_next_crq(&adapter->queue);
|
|
if (crq) {
|
|
vio_disable_interrupts(vdev);
|
|
ibmvmc_handle_crq(crq, adapter);
|
|
crq->valid = 0x00;
|
|
/* CRQ reset was requested, stop processing CRQs.
|
|
* Interrupts will be re-enabled by the reset task.
|
|
*/
|
|
if (ibmvmc.state == ibmvmc_state_sched_reset)
|
|
return;
|
|
} else {
|
|
done = 1;
|
|
}
|
|
}
|
|
}
|
|
|
|
/**
|
|
* ibmvmc_init_crq_queue - Init CRQ Queue
|
|
*
|
|
* @adapter: crq_server_adapter struct
|
|
*
|
|
* Return:
|
|
* 0 - Success
|
|
* Non-zero - Failure
|
|
*/
|
|
static int ibmvmc_init_crq_queue(struct crq_server_adapter *adapter)
|
|
{
|
|
struct vio_dev *vdev = to_vio_dev(adapter->dev);
|
|
struct crq_queue *queue = &adapter->queue;
|
|
int rc = 0;
|
|
int retrc = 0;
|
|
|
|
queue->msgs = (struct ibmvmc_crq_msg *)get_zeroed_page(GFP_KERNEL);
|
|
|
|
if (!queue->msgs)
|
|
goto malloc_failed;
|
|
|
|
queue->size = PAGE_SIZE / sizeof(*queue->msgs);
|
|
|
|
queue->msg_token = dma_map_single(adapter->dev, queue->msgs,
|
|
queue->size * sizeof(*queue->msgs),
|
|
DMA_BIDIRECTIONAL);
|
|
|
|
if (dma_mapping_error(adapter->dev, queue->msg_token))
|
|
goto map_failed;
|
|
|
|
retrc = plpar_hcall_norets(H_REG_CRQ,
|
|
vdev->unit_address,
|
|
queue->msg_token, PAGE_SIZE);
|
|
rc = retrc;
|
|
|
|
if (rc == H_RESOURCE)
|
|
rc = ibmvmc_reset_crq_queue(adapter);
|
|
|
|
if (rc == 2) {
|
|
dev_warn(adapter->dev, "Partner adapter not ready\n");
|
|
retrc = 0;
|
|
} else if (rc != 0) {
|
|
dev_err(adapter->dev, "Error %d opening adapter\n", rc);
|
|
goto reg_crq_failed;
|
|
}
|
|
|
|
queue->cur = 0;
|
|
spin_lock_init(&queue->lock);
|
|
|
|
tasklet_init(&adapter->work_task, ibmvmc_task, (unsigned long)adapter);
|
|
|
|
if (request_irq(vdev->irq,
|
|
ibmvmc_handle_event,
|
|
0, "ibmvmc", (void *)adapter) != 0) {
|
|
dev_err(adapter->dev, "couldn't register irq 0x%x\n",
|
|
vdev->irq);
|
|
goto req_irq_failed;
|
|
}
|
|
|
|
rc = vio_enable_interrupts(vdev);
|
|
if (rc != 0) {
|
|
dev_err(adapter->dev, "Error %d enabling interrupts!!!\n", rc);
|
|
goto req_irq_failed;
|
|
}
|
|
|
|
return retrc;
|
|
|
|
req_irq_failed:
|
|
/* Cannot have any work since we either never got our IRQ registered,
|
|
* or never got interrupts enabled
|
|
*/
|
|
tasklet_kill(&adapter->work_task);
|
|
h_free_crq(vdev->unit_address);
|
|
reg_crq_failed:
|
|
dma_unmap_single(adapter->dev,
|
|
queue->msg_token,
|
|
queue->size * sizeof(*queue->msgs), DMA_BIDIRECTIONAL);
|
|
map_failed:
|
|
free_page((unsigned long)queue->msgs);
|
|
malloc_failed:
|
|
return -ENOMEM;
|
|
}
|
|
|
|
/* Fill in the liobn and riobn fields on the adapter */
|
|
static int read_dma_window(struct vio_dev *vdev,
|
|
struct crq_server_adapter *adapter)
|
|
{
|
|
const __be32 *dma_window;
|
|
const __be32 *prop;
|
|
|
|
/* TODO Using of_parse_dma_window would be better, but it doesn't give
|
|
* a way to read multiple windows without already knowing the size of
|
|
* a window or the number of windows
|
|
*/
|
|
dma_window =
|
|
(const __be32 *)vio_get_attribute(vdev, "ibm,my-dma-window",
|
|
NULL);
|
|
if (!dma_window) {
|
|
dev_warn(adapter->dev, "Couldn't find ibm,my-dma-window property\n");
|
|
return -1;
|
|
}
|
|
|
|
adapter->liobn = be32_to_cpu(*dma_window);
|
|
dma_window++;
|
|
|
|
prop = (const __be32 *)vio_get_attribute(vdev, "ibm,#dma-address-cells",
|
|
NULL);
|
|
if (!prop) {
|
|
dev_warn(adapter->dev, "Couldn't find ibm,#dma-address-cells property\n");
|
|
dma_window++;
|
|
} else {
|
|
dma_window += be32_to_cpu(*prop);
|
|
}
|
|
|
|
prop = (const __be32 *)vio_get_attribute(vdev, "ibm,#dma-size-cells",
|
|
NULL);
|
|
if (!prop) {
|
|
dev_warn(adapter->dev, "Couldn't find ibm,#dma-size-cells property\n");
|
|
dma_window++;
|
|
} else {
|
|
dma_window += be32_to_cpu(*prop);
|
|
}
|
|
|
|
/* dma_window should point to the second window now */
|
|
adapter->riobn = be32_to_cpu(*dma_window);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ibmvmc_probe(struct vio_dev *vdev, const struct vio_device_id *id)
|
|
{
|
|
struct crq_server_adapter *adapter = &ibmvmc_adapter;
|
|
int rc;
|
|
|
|
dev_set_drvdata(&vdev->dev, NULL);
|
|
memset(adapter, 0, sizeof(*adapter));
|
|
adapter->dev = &vdev->dev;
|
|
|
|
dev_info(adapter->dev, "Probe for UA 0x%x\n", vdev->unit_address);
|
|
|
|
rc = read_dma_window(vdev, adapter);
|
|
if (rc != 0) {
|
|
ibmvmc.state = ibmvmc_state_failed;
|
|
return -1;
|
|
}
|
|
|
|
dev_dbg(adapter->dev, "Probe: liobn 0x%x, riobn 0x%x\n",
|
|
adapter->liobn, adapter->riobn);
|
|
|
|
init_waitqueue_head(&adapter->reset_wait_queue);
|
|
adapter->reset_task = kthread_run(ibmvmc_reset_task, adapter, "ibmvmc");
|
|
if (IS_ERR(adapter->reset_task)) {
|
|
dev_err(adapter->dev, "Failed to start reset thread\n");
|
|
ibmvmc.state = ibmvmc_state_failed;
|
|
rc = PTR_ERR(adapter->reset_task);
|
|
adapter->reset_task = NULL;
|
|
return rc;
|
|
}
|
|
|
|
rc = ibmvmc_init_crq_queue(adapter);
|
|
if (rc != 0 && rc != H_RESOURCE) {
|
|
dev_err(adapter->dev, "Error initializing CRQ. rc = 0x%x\n",
|
|
rc);
|
|
ibmvmc.state = ibmvmc_state_failed;
|
|
goto crq_failed;
|
|
}
|
|
|
|
ibmvmc.state = ibmvmc_state_crqinit;
|
|
|
|
/* Try to send an initialization message. Note that this is allowed
|
|
* to fail if the other end is not acive. In that case we just wait
|
|
* for the other side to initialize.
|
|
*/
|
|
if (ibmvmc_send_crq(adapter, 0xC001000000000000LL, 0) != 0 &&
|
|
rc != H_RESOURCE)
|
|
dev_warn(adapter->dev, "Failed to send initialize CRQ message\n");
|
|
|
|
dev_set_drvdata(&vdev->dev, adapter);
|
|
|
|
return 0;
|
|
|
|
crq_failed:
|
|
kthread_stop(adapter->reset_task);
|
|
adapter->reset_task = NULL;
|
|
return -EPERM;
|
|
}
|
|
|
|
static void ibmvmc_remove(struct vio_dev *vdev)
|
|
{
|
|
struct crq_server_adapter *adapter = dev_get_drvdata(&vdev->dev);
|
|
|
|
dev_info(adapter->dev, "Entering remove for UA 0x%x\n",
|
|
vdev->unit_address);
|
|
ibmvmc_release_crq_queue(adapter);
|
|
}
|
|
|
|
static struct vio_device_id ibmvmc_device_table[] = {
|
|
{ "ibm,vmc", "IBM,vmc" },
|
|
{ "", "" }
|
|
};
|
|
MODULE_DEVICE_TABLE(vio, ibmvmc_device_table);
|
|
|
|
static struct vio_driver ibmvmc_driver = {
|
|
.name = ibmvmc_driver_name,
|
|
.id_table = ibmvmc_device_table,
|
|
.probe = ibmvmc_probe,
|
|
.remove = ibmvmc_remove,
|
|
};
|
|
|
|
static void __init ibmvmc_scrub_module_parms(void)
|
|
{
|
|
if (ibmvmc_max_mtu > MAX_MTU) {
|
|
pr_warn("ibmvmc: Max MTU reduced to %d\n", MAX_MTU);
|
|
ibmvmc_max_mtu = MAX_MTU;
|
|
} else if (ibmvmc_max_mtu < MIN_MTU) {
|
|
pr_warn("ibmvmc: Max MTU increased to %d\n", MIN_MTU);
|
|
ibmvmc_max_mtu = MIN_MTU;
|
|
}
|
|
|
|
if (ibmvmc_max_buf_pool_size > MAX_BUF_POOL_SIZE) {
|
|
pr_warn("ibmvmc: Max buffer pool size reduced to %d\n",
|
|
MAX_BUF_POOL_SIZE);
|
|
ibmvmc_max_buf_pool_size = MAX_BUF_POOL_SIZE;
|
|
} else if (ibmvmc_max_buf_pool_size < MIN_BUF_POOL_SIZE) {
|
|
pr_warn("ibmvmc: Max buffer pool size increased to %d\n",
|
|
MIN_BUF_POOL_SIZE);
|
|
ibmvmc_max_buf_pool_size = MIN_BUF_POOL_SIZE;
|
|
}
|
|
|
|
if (ibmvmc_max_hmcs > MAX_HMCS) {
|
|
pr_warn("ibmvmc: Max HMCs reduced to %d\n", MAX_HMCS);
|
|
ibmvmc_max_hmcs = MAX_HMCS;
|
|
} else if (ibmvmc_max_hmcs < MIN_HMCS) {
|
|
pr_warn("ibmvmc: Max HMCs increased to %d\n", MIN_HMCS);
|
|
ibmvmc_max_hmcs = MIN_HMCS;
|
|
}
|
|
}
|
|
|
|
static struct miscdevice ibmvmc_miscdev = {
|
|
.name = ibmvmc_driver_name,
|
|
.minor = MISC_DYNAMIC_MINOR,
|
|
.fops = &ibmvmc_fops,
|
|
};
|
|
|
|
static int __init ibmvmc_module_init(void)
|
|
{
|
|
int rc, i, j;
|
|
|
|
ibmvmc.state = ibmvmc_state_initial;
|
|
pr_info("ibmvmc: version %s\n", IBMVMC_DRIVER_VERSION);
|
|
|
|
rc = misc_register(&ibmvmc_miscdev);
|
|
if (rc) {
|
|
pr_err("ibmvmc: misc registration failed\n");
|
|
goto misc_register_failed;
|
|
}
|
|
pr_info("ibmvmc: node %d:%d\n", MISC_MAJOR,
|
|
ibmvmc_miscdev.minor);
|
|
|
|
/* Initialize data structures */
|
|
memset(hmcs, 0, sizeof(struct ibmvmc_hmc) * MAX_HMCS);
|
|
for (i = 0; i < MAX_HMCS; i++) {
|
|
spin_lock_init(&hmcs[i].lock);
|
|
hmcs[i].state = ibmhmc_state_free;
|
|
for (j = 0; j < MAX_BUF_POOL_SIZE; j++)
|
|
hmcs[i].queue_outbound_msgs[j] = VMC_INVALID_BUFFER_ID;
|
|
}
|
|
|
|
/* Sanity check module parms */
|
|
ibmvmc_scrub_module_parms();
|
|
|
|
/*
|
|
* Initialize some reasonable values. Might be negotiated smaller
|
|
* values during the capabilities exchange.
|
|
*/
|
|
ibmvmc.max_mtu = ibmvmc_max_mtu;
|
|
ibmvmc.max_buffer_pool_size = ibmvmc_max_buf_pool_size;
|
|
ibmvmc.max_hmc_index = ibmvmc_max_hmcs - 1;
|
|
|
|
rc = vio_register_driver(&ibmvmc_driver);
|
|
|
|
if (rc) {
|
|
pr_err("ibmvmc: rc %d from vio_register_driver\n", rc);
|
|
goto vio_reg_failed;
|
|
}
|
|
|
|
return 0;
|
|
|
|
vio_reg_failed:
|
|
misc_deregister(&ibmvmc_miscdev);
|
|
misc_register_failed:
|
|
return rc;
|
|
}
|
|
|
|
static void __exit ibmvmc_module_exit(void)
|
|
{
|
|
pr_info("ibmvmc: module exit\n");
|
|
vio_unregister_driver(&ibmvmc_driver);
|
|
misc_deregister(&ibmvmc_miscdev);
|
|
}
|
|
|
|
module_init(ibmvmc_module_init);
|
|
module_exit(ibmvmc_module_exit);
|
|
|
|
module_param_named(buf_pool_size, ibmvmc_max_buf_pool_size,
|
|
int, 0644);
|
|
MODULE_PARM_DESC(buf_pool_size, "Buffer pool size");
|
|
module_param_named(max_hmcs, ibmvmc_max_hmcs, int, 0644);
|
|
MODULE_PARM_DESC(max_hmcs, "Max HMCs");
|
|
module_param_named(max_mtu, ibmvmc_max_mtu, int, 0644);
|
|
MODULE_PARM_DESC(max_mtu, "Max MTU");
|
|
|
|
MODULE_AUTHOR("Steven Royer <seroyer@linux.vnet.ibm.com>");
|
|
MODULE_DESCRIPTION("IBM VMC");
|
|
MODULE_VERSION(IBMVMC_DRIVER_VERSION);
|
|
MODULE_LICENSE("GPL v2");
|