[CCID2]: Deadlock and spurious timeouts when Ack Ratio > cwnd

This patch removes a bug in the current code. I agree with Andrea's comment
that there is a problem here but the way it is treated does not fix it.

The problem is that whenever Ack Ratio > cwnd, starvation/deadlock occurs:
 * the receiver will not send an Ack until (Ack Ratio - cwnd) data packets
   have arrived;
 * the sender will not send any data packet before the receipt of an Ack
   advances the send window.
The only way that the connection then progresses was via RTO timeout. In one
extreme case (bulk transfer), it was observed that this happened for every single
packet; i.e. hundreds of packets, each a RTO timeout of 1..3 seconds apart:
a transfer which normally would take a fraction of a second thus grew to
several minutes.

The solution taken by this approach is to observe the relation

                   "Ack Ratio <= cwnd"

by using the constraint (1) from RFC 4341, 6.1.2; i.e. set

                 Ack Ratio = ceil(cwnd / 2)

and update it whenever either Ack Ratio or cwnd change. This ensures that
the deadlock problem can not arise.

Signed-off-by: Gerrit Renker <gerrit@erg.abdn.ac.uk>
Acked-by: Ian McDonald <ian.mcdonald@jandi.co.nz>
Signed-off-by: Arnaldo Carvalho de Melo <acme@redhat.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
Gerrit Renker 2007-11-24 21:40:24 -02:00 committed by David S. Miller
parent df054e1d00
commit d50ad163e6

View File

@ -143,32 +143,30 @@ static int ccid2_hc_tx_send_packet(struct sock *sk, struct sk_buff *skb)
static void ccid2_change_l_ack_ratio(struct sock *sk, u32 val) static void ccid2_change_l_ack_ratio(struct sock *sk, u32 val)
{ {
struct dccp_sock *dp = dccp_sk(sk); struct dccp_sock *dp = dccp_sk(sk);
u32 max_ratio = DIV_ROUND_UP(ccid2_hc_tx_sk(sk)->ccid2hctx_cwnd, 2);
/* /*
* XXX I don't really agree with val != 2. If cwnd is 1, ack ratio * Ensure that Ack Ratio does not exceed ceil(cwnd/2), which is (2) from
* should be 1... it shouldn't be allowed to become 2. * RFC 4341, 6.1.2. We ignore the statement that Ack Ratio 2 is always
* -sorbo. * acceptable since this causes starvation/deadlock whenever cwnd < 2.
* The same problem arises when Ack Ratio is 0 (ie. Ack Ratio disabled).
*/ */
if (val != 2) { if (val == 0 || val > max_ratio) {
const struct ccid2_hc_tx_sock *hctx = ccid2_hc_tx_sk(sk); DCCP_WARN("Limiting Ack Ratio (%u) to %u\n", val, max_ratio);
int max = hctx->ccid2hctx_cwnd / 2; val = max_ratio;
/* round up */
if (hctx->ccid2hctx_cwnd & 1)
max++;
if (val > max)
val = max;
} }
if (val > 0xFFFF) /* RFC 4340, 11.3 */ if (val > 0xFFFF) /* RFC 4340, 11.3 */
val = 0xFFFF; val = 0xFFFF;
if (val == dp->dccps_l_ack_ratio)
return;
ccid2_pr_debug("changing local ack ratio to %u\n", val); ccid2_pr_debug("changing local ack ratio to %u\n", val);
dp->dccps_l_ack_ratio = val; dp->dccps_l_ack_ratio = val;
} }
static void ccid2_change_cwnd(struct ccid2_hc_tx_sock *hctx, u32 val) static void ccid2_change_cwnd(struct ccid2_hc_tx_sock *hctx, u32 val)
{ {
/* XXX do we need to change ack ratio? */
hctx->ccid2hctx_cwnd = val? : 1; hctx->ccid2hctx_cwnd = val? : 1;
ccid2_pr_debug("changed cwnd to %u\n", hctx->ccid2hctx_cwnd); ccid2_pr_debug("changed cwnd to %u\n", hctx->ccid2hctx_cwnd);
} }
@ -519,9 +517,10 @@ static void ccid2_hc_tx_dec_pipe(struct sock *sk)
ccid2_hc_tx_kill_rto_timer(sk); ccid2_hc_tx_kill_rto_timer(sk);
} }
static void ccid2_congestion_event(struct ccid2_hc_tx_sock *hctx, static void ccid2_congestion_event(struct sock *sk, struct ccid2_seq *seqp)
struct ccid2_seq *seqp)
{ {
struct ccid2_hc_tx_sock *hctx = ccid2_hc_tx_sk(sk);
if (time_before(seqp->ccid2s_sent, hctx->ccid2hctx_last_cong)) { if (time_before(seqp->ccid2s_sent, hctx->ccid2hctx_last_cong)) {
ccid2_pr_debug("Multiple losses in an RTT---treating as one\n"); ccid2_pr_debug("Multiple losses in an RTT---treating as one\n");
return; return;
@ -533,6 +532,10 @@ static void ccid2_congestion_event(struct ccid2_hc_tx_sock *hctx,
hctx->ccid2hctx_ssthresh = hctx->ccid2hctx_cwnd; hctx->ccid2hctx_ssthresh = hctx->ccid2hctx_cwnd;
if (hctx->ccid2hctx_ssthresh < 2) if (hctx->ccid2hctx_ssthresh < 2)
hctx->ccid2hctx_ssthresh = 2; hctx->ccid2hctx_ssthresh = 2;
/* Avoid spurious timeouts resulting from Ack Ratio > cwnd */
if (dccp_sk(sk)->dccps_l_ack_ratio > hctx->ccid2hctx_cwnd)
ccid2_change_l_ack_ratio(sk, hctx->ccid2hctx_cwnd);
} }
static void ccid2_hc_tx_packet_recv(struct sock *sk, struct sk_buff *skb) static void ccid2_hc_tx_packet_recv(struct sock *sk, struct sk_buff *skb)
@ -648,7 +651,7 @@ static void ccid2_hc_tx_packet_recv(struct sock *sk, struct sk_buff *skb)
!seqp->ccid2s_acked) { !seqp->ccid2s_acked) {
if (state == if (state ==
DCCP_ACKVEC_STATE_ECN_MARKED) { DCCP_ACKVEC_STATE_ECN_MARKED) {
ccid2_congestion_event(hctx, ccid2_congestion_event(sk,
seqp); seqp);
} else } else
ccid2_new_ack(sk, seqp, ccid2_new_ack(sk, seqp,
@ -713,7 +716,7 @@ static void ccid2_hc_tx_packet_recv(struct sock *sk, struct sk_buff *skb)
* order to detect multiple congestion events in * order to detect multiple congestion events in
* one ack vector. * one ack vector.
*/ */
ccid2_congestion_event(hctx, seqp); ccid2_congestion_event(sk, seqp);
ccid2_hc_tx_dec_pipe(sk); ccid2_hc_tx_dec_pipe(sk);
} }
if (seqp == hctx->ccid2hctx_seqt) if (seqp == hctx->ccid2hctx_seqt)