usb: wusbcore: fix short transfers
If a URB is broken up into multiple transfer segments and a short transfer occurs in any segment other than the last, the URB will currently get stuck in the driver forever. This patch adds a check for a short transfer and cleans up any pending segments so the URB can complete properly. Signed-off-by: Thomas Pugliese <thomas.pugliese@gmail.com> Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
This commit is contained in:
committed by
Greg Kroah-Hartman
parent
e0e6a356a0
commit
7005234c18
@@ -1993,7 +1993,7 @@ static int wa_xfer_status_to_errno(u8 status)
|
|||||||
* the xfer will complete cleanly.
|
* the xfer will complete cleanly.
|
||||||
*/
|
*/
|
||||||
static void wa_complete_remaining_xfer_segs(struct wa_xfer *xfer,
|
static void wa_complete_remaining_xfer_segs(struct wa_xfer *xfer,
|
||||||
struct wa_seg *incoming_seg)
|
struct wa_seg *incoming_seg, enum wa_seg_status status)
|
||||||
{
|
{
|
||||||
int index;
|
int index;
|
||||||
struct wa_rpipe *rpipe = xfer->ep->hcpriv;
|
struct wa_rpipe *rpipe = xfer->ep->hcpriv;
|
||||||
@@ -2015,7 +2015,7 @@ static void wa_complete_remaining_xfer_segs(struct wa_xfer *xfer,
|
|||||||
*/
|
*/
|
||||||
case WA_SEG_DELAYED:
|
case WA_SEG_DELAYED:
|
||||||
xfer->segs_done++;
|
xfer->segs_done++;
|
||||||
current_seg->status = incoming_seg->status;
|
current_seg->status = status;
|
||||||
break;
|
break;
|
||||||
case WA_SEG_ABORTED:
|
case WA_SEG_ABORTED:
|
||||||
break;
|
break;
|
||||||
@@ -2028,6 +2028,58 @@ static void wa_complete_remaining_xfer_segs(struct wa_xfer *xfer,
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* Populate the wa->buf_in_urb based on the current transfer state. */
|
||||||
|
static int wa_populate_buf_in_urb(struct wahc *wa, struct wa_xfer *xfer,
|
||||||
|
unsigned int seg_idx, unsigned int bytes_transferred)
|
||||||
|
{
|
||||||
|
int result = 0;
|
||||||
|
struct wa_seg *seg = xfer->seg[seg_idx];
|
||||||
|
|
||||||
|
BUG_ON(wa->buf_in_urb->status == -EINPROGRESS);
|
||||||
|
/* this should always be 0 before a resubmit. */
|
||||||
|
wa->buf_in_urb->num_mapped_sgs = 0;
|
||||||
|
|
||||||
|
if (xfer->is_dma) {
|
||||||
|
wa->buf_in_urb->transfer_dma = xfer->urb->transfer_dma
|
||||||
|
+ (seg_idx * xfer->seg_size);
|
||||||
|
wa->buf_in_urb->transfer_flags |= URB_NO_TRANSFER_DMA_MAP;
|
||||||
|
wa->buf_in_urb->transfer_buffer = NULL;
|
||||||
|
wa->buf_in_urb->sg = NULL;
|
||||||
|
wa->buf_in_urb->num_sgs = 0;
|
||||||
|
} else {
|
||||||
|
/* do buffer or SG processing. */
|
||||||
|
wa->buf_in_urb->transfer_flags &= ~URB_NO_TRANSFER_DMA_MAP;
|
||||||
|
|
||||||
|
if (xfer->urb->transfer_buffer) {
|
||||||
|
wa->buf_in_urb->transfer_buffer =
|
||||||
|
xfer->urb->transfer_buffer
|
||||||
|
+ (seg_idx * xfer->seg_size);
|
||||||
|
wa->buf_in_urb->sg = NULL;
|
||||||
|
wa->buf_in_urb->num_sgs = 0;
|
||||||
|
} else {
|
||||||
|
/* allocate an SG list to store seg_size bytes
|
||||||
|
and copy the subset of the xfer->urb->sg
|
||||||
|
that matches the buffer subset we are
|
||||||
|
about to read. */
|
||||||
|
wa->buf_in_urb->sg = wa_xfer_create_subset_sg(
|
||||||
|
xfer->urb->sg,
|
||||||
|
seg_idx * xfer->seg_size,
|
||||||
|
bytes_transferred,
|
||||||
|
&(wa->buf_in_urb->num_sgs));
|
||||||
|
|
||||||
|
if (!(wa->buf_in_urb->sg)) {
|
||||||
|
wa->buf_in_urb->num_sgs = 0;
|
||||||
|
result = -ENOMEM;
|
||||||
|
}
|
||||||
|
wa->buf_in_urb->transfer_buffer = NULL;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
wa->buf_in_urb->transfer_buffer_length = bytes_transferred;
|
||||||
|
wa->buf_in_urb->context = seg;
|
||||||
|
|
||||||
|
return result;
|
||||||
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Process a xfer result completion message
|
* Process a xfer result completion message
|
||||||
*
|
*
|
||||||
@@ -2041,12 +2093,13 @@ static void wa_xfer_result_chew(struct wahc *wa, struct wa_xfer *xfer,
|
|||||||
int result;
|
int result;
|
||||||
struct device *dev = &wa->usb_iface->dev;
|
struct device *dev = &wa->usb_iface->dev;
|
||||||
unsigned long flags;
|
unsigned long flags;
|
||||||
u8 seg_idx;
|
unsigned int seg_idx;
|
||||||
struct wa_seg *seg;
|
struct wa_seg *seg;
|
||||||
struct wa_rpipe *rpipe;
|
struct wa_rpipe *rpipe;
|
||||||
unsigned done = 0;
|
unsigned done = 0;
|
||||||
u8 usb_status;
|
u8 usb_status;
|
||||||
unsigned rpipe_ready = 0;
|
unsigned rpipe_ready = 0;
|
||||||
|
unsigned bytes_transferred = le32_to_cpu(xfer_result->dwTransferLength);
|
||||||
|
|
||||||
spin_lock_irqsave(&xfer->lock, flags);
|
spin_lock_irqsave(&xfer->lock, flags);
|
||||||
seg_idx = xfer_result->bTransferSegment & 0x7f;
|
seg_idx = xfer_result->bTransferSegment & 0x7f;
|
||||||
@@ -2079,66 +2132,33 @@ static void wa_xfer_result_chew(struct wahc *wa, struct wa_xfer *xfer,
|
|||||||
/* FIXME: we ignore warnings, tally them for stats */
|
/* FIXME: we ignore warnings, tally them for stats */
|
||||||
if (usb_status & 0x40) /* Warning?... */
|
if (usb_status & 0x40) /* Warning?... */
|
||||||
usb_status = 0; /* ... pass */
|
usb_status = 0; /* ... pass */
|
||||||
|
/*
|
||||||
|
* If the last segment bit is set, complete the remaining segments.
|
||||||
|
* When the current segment is completed, either in wa_buf_in_cb for
|
||||||
|
* transfers with data or below for no data, the xfer will complete.
|
||||||
|
*/
|
||||||
|
if (xfer_result->bTransferSegment & 0x80)
|
||||||
|
wa_complete_remaining_xfer_segs(xfer, seg, WA_SEG_DONE);
|
||||||
if (usb_pipeisoc(xfer->urb->pipe)) {
|
if (usb_pipeisoc(xfer->urb->pipe)) {
|
||||||
/* set up WA state to read the isoc packet status next. */
|
/* set up WA state to read the isoc packet status next. */
|
||||||
wa->dti_isoc_xfer_in_progress = wa_xfer_id(xfer);
|
wa->dti_isoc_xfer_in_progress = wa_xfer_id(xfer);
|
||||||
wa->dti_isoc_xfer_seg = seg_idx;
|
wa->dti_isoc_xfer_seg = seg_idx;
|
||||||
wa->dti_state = WA_DTI_ISOC_PACKET_STATUS_PENDING;
|
wa->dti_state = WA_DTI_ISOC_PACKET_STATUS_PENDING;
|
||||||
} else if (xfer->is_inbound) { /* IN data phase: read to buffer */
|
} else if ((xfer->is_inbound)
|
||||||
|
&& (bytes_transferred > 0)) {
|
||||||
|
/* IN data phase: read to buffer */
|
||||||
seg->status = WA_SEG_DTI_PENDING;
|
seg->status = WA_SEG_DTI_PENDING;
|
||||||
BUG_ON(wa->buf_in_urb->status == -EINPROGRESS);
|
result = wa_populate_buf_in_urb(wa, xfer, seg_idx,
|
||||||
/* this should always be 0 before a resubmit. */
|
bytes_transferred);
|
||||||
wa->buf_in_urb->num_mapped_sgs = 0;
|
if (result < 0)
|
||||||
|
goto error_buf_in_populate;
|
||||||
if (xfer->is_dma) {
|
|
||||||
wa->buf_in_urb->transfer_dma =
|
|
||||||
xfer->urb->transfer_dma
|
|
||||||
+ (seg_idx * xfer->seg_size);
|
|
||||||
wa->buf_in_urb->transfer_flags
|
|
||||||
|= URB_NO_TRANSFER_DMA_MAP;
|
|
||||||
wa->buf_in_urb->transfer_buffer = NULL;
|
|
||||||
wa->buf_in_urb->sg = NULL;
|
|
||||||
wa->buf_in_urb->num_sgs = 0;
|
|
||||||
} else {
|
|
||||||
/* do buffer or SG processing. */
|
|
||||||
wa->buf_in_urb->transfer_flags
|
|
||||||
&= ~URB_NO_TRANSFER_DMA_MAP;
|
|
||||||
|
|
||||||
if (xfer->urb->transfer_buffer) {
|
|
||||||
wa->buf_in_urb->transfer_buffer =
|
|
||||||
xfer->urb->transfer_buffer
|
|
||||||
+ (seg_idx * xfer->seg_size);
|
|
||||||
wa->buf_in_urb->sg = NULL;
|
|
||||||
wa->buf_in_urb->num_sgs = 0;
|
|
||||||
} else {
|
|
||||||
/* allocate an SG list to store seg_size bytes
|
|
||||||
and copy the subset of the xfer->urb->sg
|
|
||||||
that matches the buffer subset we are
|
|
||||||
about to read. */
|
|
||||||
wa->buf_in_urb->sg = wa_xfer_create_subset_sg(
|
|
||||||
xfer->urb->sg,
|
|
||||||
seg_idx * xfer->seg_size,
|
|
||||||
le32_to_cpu(
|
|
||||||
xfer_result->dwTransferLength),
|
|
||||||
&(wa->buf_in_urb->num_sgs));
|
|
||||||
|
|
||||||
if (!(wa->buf_in_urb->sg)) {
|
|
||||||
wa->buf_in_urb->num_sgs = 0;
|
|
||||||
goto error_sg_alloc;
|
|
||||||
}
|
|
||||||
wa->buf_in_urb->transfer_buffer = NULL;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
wa->buf_in_urb->transfer_buffer_length =
|
|
||||||
le32_to_cpu(xfer_result->dwTransferLength);
|
|
||||||
wa->buf_in_urb->context = seg;
|
|
||||||
result = usb_submit_urb(wa->buf_in_urb, GFP_ATOMIC);
|
result = usb_submit_urb(wa->buf_in_urb, GFP_ATOMIC);
|
||||||
if (result < 0)
|
if (result < 0)
|
||||||
goto error_submit_buf_in;
|
goto error_submit_buf_in;
|
||||||
} else {
|
} else {
|
||||||
/* OUT data phase, complete it -- */
|
/* OUT data phase or no data, complete it -- */
|
||||||
seg->status = WA_SEG_DONE;
|
seg->status = WA_SEG_DONE;
|
||||||
seg->result = le32_to_cpu(xfer_result->dwTransferLength);
|
seg->result = bytes_transferred;
|
||||||
xfer->segs_done++;
|
xfer->segs_done++;
|
||||||
rpipe_ready = rpipe_avail_inc(rpipe);
|
rpipe_ready = rpipe_avail_inc(rpipe);
|
||||||
done = __wa_xfer_is_done(xfer);
|
done = __wa_xfer_is_done(xfer);
|
||||||
@@ -2162,13 +2182,13 @@ error_submit_buf_in:
|
|||||||
seg->result = result;
|
seg->result = result;
|
||||||
kfree(wa->buf_in_urb->sg);
|
kfree(wa->buf_in_urb->sg);
|
||||||
wa->buf_in_urb->sg = NULL;
|
wa->buf_in_urb->sg = NULL;
|
||||||
error_sg_alloc:
|
error_buf_in_populate:
|
||||||
__wa_xfer_abort(xfer);
|
__wa_xfer_abort(xfer);
|
||||||
seg->status = WA_SEG_ERROR;
|
seg->status = WA_SEG_ERROR;
|
||||||
error_complete:
|
error_complete:
|
||||||
xfer->segs_done++;
|
xfer->segs_done++;
|
||||||
rpipe_ready = rpipe_avail_inc(rpipe);
|
rpipe_ready = rpipe_avail_inc(rpipe);
|
||||||
wa_complete_remaining_xfer_segs(xfer, seg);
|
wa_complete_remaining_xfer_segs(xfer, seg, seg->status);
|
||||||
done = __wa_xfer_is_done(xfer);
|
done = __wa_xfer_is_done(xfer);
|
||||||
/*
|
/*
|
||||||
* queue work item to clear STALL for control endpoints.
|
* queue work item to clear STALL for control endpoints.
|
||||||
|
|||||||
Reference in New Issue
Block a user