2006-01-02 18:04:38 +00:00
|
|
|
/*
|
|
|
|
* net/tipc/msg.h: Include file for TIPC message header routines
|
2007-02-09 14:25:21 +00:00
|
|
|
*
|
2015-03-13 20:08:06 +00:00
|
|
|
* Copyright (c) 2000-2007, 2014-2015 Ericsson AB
|
2011-01-25 18:33:31 +00:00
|
|
|
* Copyright (c) 2005-2008, 2010-2011, Wind River Systems
|
2006-01-02 18:04:38 +00:00
|
|
|
* All rights reserved.
|
|
|
|
*
|
2006-01-11 12:30:43 +00:00
|
|
|
* Redistribution and use in source and binary forms, with or without
|
2006-01-02 18:04:38 +00:00
|
|
|
* modification, are permitted provided that the following conditions are met:
|
|
|
|
*
|
2006-01-11 12:30:43 +00:00
|
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer.
|
|
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
|
|
* documentation and/or other materials provided with the distribution.
|
|
|
|
* 3. Neither the names of the copyright holders nor the names of its
|
|
|
|
* contributors may be used to endorse or promote products derived from
|
|
|
|
* this software without specific prior written permission.
|
2006-01-02 18:04:38 +00:00
|
|
|
*
|
2006-01-11 12:30:43 +00:00
|
|
|
* Alternatively, this software may be distributed under the terms of the
|
|
|
|
* GNU General Public License ("GPL") version 2 as published by the Free
|
|
|
|
* Software Foundation.
|
|
|
|
*
|
|
|
|
* THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
|
|
|
|
* AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
|
|
|
|
* LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
|
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
|
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
|
|
|
|
* INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
|
|
|
|
* CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
|
|
|
|
* ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
|
2006-01-02 18:04:38 +00:00
|
|
|
* POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#ifndef _TIPC_MSG_H
|
|
|
|
#define _TIPC_MSG_H
|
|
|
|
|
2015-01-09 07:27:07 +00:00
|
|
|
#include <linux/tipc.h>
|
tipc: reduce locking scope during packet reception
We convert packet/message reception according to the same principle
we have been using for message sending and timeout handling:
We move the function tipc_rcv() to node.c, hence handling the initial
packet reception at the link aggregation level. The function grabs
the node lock, selects the receiving link, and accesses it via a new
call tipc_link_rcv(). This function appends buffers to the input
queue for delivery upwards, but it may also append outgoing packets
to the xmit queue, just as we do during regular message sending. The
latter will happen when buffers are forwarded from the link backlog,
or when retransmission is requested.
Upon return of this function, and after having released the node lock,
tipc_rcv() delivers/tranmsits the contents of those queues, but it may
also perform actions such as link activation or reset, as indicated by
the return flags from the link.
This reduces the number of cpu cycles spent inside the node spinlock,
and reduces contention on that lock.
Reviewed-by: Ying Xue <ying.xue@windriver.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2015-07-16 20:54:31 +00:00
|
|
|
#include "core.h"
|
2006-01-02 18:04:38 +00:00
|
|
|
|
2011-04-08 14:50:52 +00:00
|
|
|
/*
|
|
|
|
* Constants and routines used to read and write TIPC payload message headers
|
|
|
|
*
|
|
|
|
* Note: Some items are also used with TIPC internal message headers
|
|
|
|
*/
|
2006-01-02 18:04:38 +00:00
|
|
|
#define TIPC_VERSION 2
|
tipc: resolve race problem at unicast message reception
TIPC handles message cardinality and sequencing at the link layer,
before passing messages upwards to the destination sockets. During the
upcall from link to socket no locks are held. It is therefore possible,
and we see it happen occasionally, that messages arriving in different
threads and delivered in sequence still bypass each other before they
reach the destination socket. This must not happen, since it violates
the sequentiality guarantee.
We solve this by adding a new input buffer queue to the link structure.
Arriving messages are added safely to the tail of that queue by the
link, while the head of the queue is consumed, also safely, by the
receiving socket. Sequentiality is secured per socket by only allowing
buffers to be dequeued inside the socket lock. Since there may be multiple
simultaneous readers of the queue, we use a 'filter' parameter to reduce
the risk that they peek the same buffer from the queue, hence also
reducing the risk of contention on the receiving socket locks.
This solves the sequentiality problem, and seems to cause no measurable
performance degradation.
A nice side effect of this change is that lock handling in the functions
tipc_rcv() and tipc_bcast_rcv() now becomes uniform, something that
will enable future simplifications of those functions.
Reviewed-by: Ying Xue <ying.xue@windriver.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2015-02-05 13:36:41 +00:00
|
|
|
struct plist;
|
2008-03-06 23:06:55 +00:00
|
|
|
|
2010-11-30 12:00:53 +00:00
|
|
|
/*
|
2011-04-08 14:50:52 +00:00
|
|
|
* Payload message users are defined in TIPC's public API:
|
|
|
|
* - TIPC_LOW_IMPORTANCE
|
|
|
|
* - TIPC_MEDIUM_IMPORTANCE
|
|
|
|
* - TIPC_HIGH_IMPORTANCE
|
|
|
|
* - TIPC_CRITICAL_IMPORTANCE
|
|
|
|
*/
|
tipc: clean up handling of message priorities
Messages transferred by TIPC are assigned an "importance priority", -an
integer value indicating how to treat the message when there is link or
destination socket congestion.
There is no separate header field for this value. Instead, the message
user values have been chosen in ascending order according to perceived
importance, so that the message user field can be used for this.
This is not a good solution. First, we have many more users than the
needed priority levels, so we end up with treating more priority
levels than necessary. Second, the user field cannot always
accurately reflect the priority of the message. E.g., a message
fragment packet should really have the priority of the enveloped
user data message, and not the priority of the MSG_FRAGMENTER user.
Until now, we have been working around this problem in different ways,
but it is now time to implement a consistent way of handling such
priorities, although still within the constraint that we cannot
allocate any more bits in the regular data message header for this.
In this commit, we define a new priority level, TIPC_SYSTEM_IMPORTANCE,
that will be the only one used apart from the four (lower) user data
levels. All non-data messages map down to this priority. Furthermore,
we take some free bits from the MSG_FRAGMENTER header and allocate
them to store the priority of the enveloped message. We then adjust
the functions msg_importance()/msg_set_importance() so that they
read/set the correct header fields depending on user type.
This small protocol change is fully compatible, because the code at
the receiving end of a link currently reads the importance level
only from user data messages, where there is no change.
Reviewed-by: Erik Hugne <erik.hugne@ericsson.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2015-03-13 20:08:11 +00:00
|
|
|
#define TIPC_SYSTEM_IMPORTANCE 4
|
|
|
|
|
2011-04-08 14:50:52 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Payload message types
|
2010-11-30 12:00:53 +00:00
|
|
|
*/
|
|
|
|
#define TIPC_CONN_MSG 0
|
|
|
|
#define TIPC_MCAST_MSG 1
|
|
|
|
#define TIPC_NAMED_MSG 2
|
|
|
|
#define TIPC_DIRECT_MSG 3
|
|
|
|
|
tipc: clean up handling of message priorities
Messages transferred by TIPC are assigned an "importance priority", -an
integer value indicating how to treat the message when there is link or
destination socket congestion.
There is no separate header field for this value. Instead, the message
user values have been chosen in ascending order according to perceived
importance, so that the message user field can be used for this.
This is not a good solution. First, we have many more users than the
needed priority levels, so we end up with treating more priority
levels than necessary. Second, the user field cannot always
accurately reflect the priority of the message. E.g., a message
fragment packet should really have the priority of the enveloped
user data message, and not the priority of the MSG_FRAGMENTER user.
Until now, we have been working around this problem in different ways,
but it is now time to implement a consistent way of handling such
priorities, although still within the constraint that we cannot
allocate any more bits in the regular data message header for this.
In this commit, we define a new priority level, TIPC_SYSTEM_IMPORTANCE,
that will be the only one used apart from the four (lower) user data
levels. All non-data messages map down to this priority. Furthermore,
we take some free bits from the MSG_FRAGMENTER header and allocate
them to store the priority of the enveloped message. We then adjust
the functions msg_importance()/msg_set_importance() so that they
read/set the correct header fields depending on user type.
This small protocol change is fully compatible, because the code at
the receiving end of a link currently reads the importance level
only from user data messages, where there is no change.
Reviewed-by: Erik Hugne <erik.hugne@ericsson.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2015-03-13 20:08:11 +00:00
|
|
|
/*
|
|
|
|
* Internal message users
|
|
|
|
*/
|
|
|
|
#define BCAST_PROTOCOL 5
|
|
|
|
#define MSG_BUNDLER 6
|
|
|
|
#define LINK_PROTOCOL 7
|
|
|
|
#define CONN_MANAGER 8
|
tipc: eliminate delayed link deletion at link failover
When a bearer is disabled manually, all its links have to be reset
and deleted. However, if there is a remaining, parallel link ready
to take over a deleted link's traffic, we currently delay the delete
of the removed link until the failover procedure is finished. This
is because the remaining link needs to access state from the reset
link, such as the last received packet number, and any partially
reassembled buffer, in order to perform a successful failover.
In this commit, we do instead move the state data over to the new
link, so that it can fulfill the procedure autonomously, without
accessing any data on the old link. This means that we can now
proceed and delete all pertaining links immediately when a bearer
is disabled. This saves us from some unnecessary complexity in such
situations.
We also choose to change the confusing definitions CHANGEOVER_PROTOCOL,
ORIGINAL_MSG and DUPLICATE_MSG to the more descriptive TUNNEL_PROTOCOL,
FAILOVER_MSG and SYNCH_MSG respectively.
Reviewed-by: Ying Xue <ying.xue@windriver.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2015-04-02 13:33:01 +00:00
|
|
|
#define TUNNEL_PROTOCOL 10
|
tipc: clean up handling of message priorities
Messages transferred by TIPC are assigned an "importance priority", -an
integer value indicating how to treat the message when there is link or
destination socket congestion.
There is no separate header field for this value. Instead, the message
user values have been chosen in ascending order according to perceived
importance, so that the message user field can be used for this.
This is not a good solution. First, we have many more users than the
needed priority levels, so we end up with treating more priority
levels than necessary. Second, the user field cannot always
accurately reflect the priority of the message. E.g., a message
fragment packet should really have the priority of the enveloped
user data message, and not the priority of the MSG_FRAGMENTER user.
Until now, we have been working around this problem in different ways,
but it is now time to implement a consistent way of handling such
priorities, although still within the constraint that we cannot
allocate any more bits in the regular data message header for this.
In this commit, we define a new priority level, TIPC_SYSTEM_IMPORTANCE,
that will be the only one used apart from the four (lower) user data
levels. All non-data messages map down to this priority. Furthermore,
we take some free bits from the MSG_FRAGMENTER header and allocate
them to store the priority of the enveloped message. We then adjust
the functions msg_importance()/msg_set_importance() so that they
read/set the correct header fields depending on user type.
This small protocol change is fully compatible, because the code at
the receiving end of a link currently reads the importance level
only from user data messages, where there is no change.
Reviewed-by: Erik Hugne <erik.hugne@ericsson.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2015-03-13 20:08:11 +00:00
|
|
|
#define NAME_DISTRIBUTOR 11
|
|
|
|
#define MSG_FRAGMENTER 12
|
|
|
|
#define LINK_CONFIG 13
|
|
|
|
#define SOCK_WAKEUP 14 /* pseudo user */
|
|
|
|
|
2011-04-08 14:50:52 +00:00
|
|
|
/*
|
|
|
|
* Message header sizes
|
|
|
|
*/
|
2011-05-31 19:03:18 +00:00
|
|
|
#define SHORT_H_SIZE 24 /* In-cluster basic payload message */
|
|
|
|
#define BASIC_H_SIZE 32 /* Basic payload message */
|
|
|
|
#define NAMED_H_SIZE 40 /* Named payload message */
|
|
|
|
#define MCAST_H_SIZE 44 /* Multicast payload message */
|
2008-03-06 23:06:55 +00:00
|
|
|
#define INT_H_SIZE 40 /* Internal messages */
|
|
|
|
#define MIN_H_SIZE 24 /* Smallest legal TIPC header size */
|
|
|
|
#define MAX_H_SIZE 60 /* Largest possible TIPC header size */
|
|
|
|
|
2006-01-02 18:04:38 +00:00
|
|
|
#define MAX_MSG_SIZE (MAX_H_SIZE + TIPC_MAX_USER_MSG_SIZE)
|
|
|
|
|
2015-02-27 07:56:57 +00:00
|
|
|
#define TIPC_MEDIA_INFO_OFFSET 5
|
2011-10-07 19:19:11 +00:00
|
|
|
|
2015-01-09 07:27:01 +00:00
|
|
|
/**
|
|
|
|
* TIPC message buffer code
|
|
|
|
*
|
|
|
|
* TIPC message buffer headroom reserves space for the worst-case
|
|
|
|
* link-level device header (in case the message is sent off-node).
|
|
|
|
*
|
|
|
|
* Note: Headroom should be a multiple of 4 to ensure the TIPC header fields
|
|
|
|
* are word aligned for quicker access
|
|
|
|
*/
|
2015-03-05 09:23:49 +00:00
|
|
|
#define BUF_HEADROOM (LL_MAX_HEADER + 48)
|
2015-01-09 07:27:01 +00:00
|
|
|
|
|
|
|
struct tipc_skb_cb {
|
|
|
|
void *handle;
|
|
|
|
struct sk_buff *tail;
|
2015-03-13 20:08:06 +00:00
|
|
|
bool validated;
|
2015-01-09 07:27:01 +00:00
|
|
|
bool wakeup_pending;
|
|
|
|
u16 chain_sz;
|
|
|
|
u16 chain_imp;
|
tipc: make struct tipc_link generic to support broadcast
Realizing that unicast is just a special case of broadcast, we also see
that we can go in the other direction, i.e., that modest changes to the
current unicast link can make it generic enough to support broadcast.
The following changes are introduced here:
- A new counter ("ackers") in struct tipc_link, to indicate how many
peers need to ack a packet before it can be released.
- A corresponding counter in the skb user area, to keep track of how
many peers a are left to ack before a buffer can be released.
- A new counter ("acked"), to keep persistent track of how far a peer
has acked at the moment, i.e., where in the transmission queue to
start updating buffers when the next ack arrives. This is to avoid
double acknowledgements from a peer, with inadvertent relase of
packets as a result.
- A more generic tipc_link_retrans() function, where retransmit starts
from a given sequence number, instead of the first packet in the
transmision queue. This is to minimize the number of retransmitted
packets on the broadcast media.
When the new functionality is taken into use in the next commits,
we expect it to have minimal effect on unicast mode performance.
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Reviewed-by: Ying Xue <ying.xue@windriver.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2015-10-22 12:51:38 +00:00
|
|
|
u16 ackers;
|
2015-01-09 07:27:01 +00:00
|
|
|
};
|
|
|
|
|
|
|
|
#define TIPC_SKB_CB(__skb) ((struct tipc_skb_cb *)&((__skb)->cb[0]))
|
|
|
|
|
2010-11-30 12:00:53 +00:00
|
|
|
struct tipc_msg {
|
|
|
|
__be32 hdr[15];
|
|
|
|
};
|
2007-02-09 14:25:21 +00:00
|
|
|
|
2015-01-09 07:27:01 +00:00
|
|
|
static inline struct tipc_msg *buf_msg(struct sk_buff *skb)
|
|
|
|
{
|
|
|
|
return (struct tipc_msg *)skb->data;
|
|
|
|
}
|
|
|
|
|
2010-11-30 12:00:53 +00:00
|
|
|
static inline u32 msg_word(struct tipc_msg *m, u32 pos)
|
|
|
|
{
|
|
|
|
return ntohl(m->hdr[pos]);
|
|
|
|
}
|
2006-01-02 18:04:38 +00:00
|
|
|
|
|
|
|
static inline void msg_set_word(struct tipc_msg *m, u32 w, u32 val)
|
|
|
|
{
|
|
|
|
m->hdr[w] = htonl(val);
|
|
|
|
}
|
|
|
|
|
2010-11-30 12:00:53 +00:00
|
|
|
static inline u32 msg_bits(struct tipc_msg *m, u32 w, u32 pos, u32 mask)
|
|
|
|
{
|
|
|
|
return (msg_word(m, w) >> pos) & mask;
|
|
|
|
}
|
|
|
|
|
2006-01-02 18:04:38 +00:00
|
|
|
static inline void msg_set_bits(struct tipc_msg *m, u32 w,
|
|
|
|
u32 pos, u32 mask, u32 val)
|
|
|
|
{
|
2007-04-24 21:51:55 +00:00
|
|
|
val = (val & mask) << pos;
|
2008-04-27 05:42:14 +00:00
|
|
|
mask = mask << pos;
|
|
|
|
m->hdr[w] &= ~htonl(mask);
|
|
|
|
m->hdr[w] |= htonl(val);
|
2006-01-02 18:04:38 +00:00
|
|
|
}
|
|
|
|
|
2008-06-05 00:54:48 +00:00
|
|
|
static inline void msg_swap_words(struct tipc_msg *msg, u32 a, u32 b)
|
|
|
|
{
|
|
|
|
u32 temp = msg->hdr[a];
|
|
|
|
|
|
|
|
msg->hdr[a] = msg->hdr[b];
|
|
|
|
msg->hdr[b] = temp;
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
/*
|
2006-01-02 18:04:38 +00:00
|
|
|
* Word 0
|
|
|
|
*/
|
|
|
|
static inline u32 msg_version(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_bits(m, 0, 29, 7);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline void msg_set_version(struct tipc_msg *m)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
2008-03-06 23:07:42 +00:00
|
|
|
msg_set_bits(m, 0, 29, 7, TIPC_VERSION);
|
2006-01-02 18:04:38 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 msg_user(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_bits(m, 0, 25, 0xf);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 msg_isdata(struct tipc_msg *m)
|
|
|
|
{
|
2010-09-22 20:43:57 +00:00
|
|
|
return msg_user(m) <= TIPC_CRITICAL_IMPORTANCE;
|
2006-01-02 18:04:38 +00:00
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline void msg_set_user(struct tipc_msg *m, u32 n)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_bits(m, 0, 25, 0xf, n);
|
|
|
|
}
|
|
|
|
|
2010-11-30 12:00:53 +00:00
|
|
|
static inline u32 msg_hdr_sz(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_bits(m, 0, 21, 0xf) << 2;
|
|
|
|
}
|
|
|
|
|
2010-12-31 18:59:32 +00:00
|
|
|
static inline void msg_set_hdr_sz(struct tipc_msg *m, u32 n)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_bits(m, 0, 21, 0xf, n>>2);
|
|
|
|
}
|
|
|
|
|
2010-11-30 12:00:53 +00:00
|
|
|
static inline u32 msg_size(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_bits(m, 0, 0, 0x1ffff);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 msg_data_sz(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_size(m) - msg_hdr_sz(m);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline int msg_non_seq(struct tipc_msg *m)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
return msg_bits(m, 0, 20, 1);
|
|
|
|
}
|
|
|
|
|
2008-06-05 00:54:48 +00:00
|
|
|
static inline void msg_set_non_seq(struct tipc_msg *m, u32 n)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
2008-06-05 00:54:48 +00:00
|
|
|
msg_set_bits(m, 0, 20, 1, n);
|
2006-01-02 18:04:38 +00:00
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline int msg_dest_droppable(struct tipc_msg *m)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
return msg_bits(m, 0, 19, 1);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline void msg_set_dest_droppable(struct tipc_msg *m, u32 d)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_bits(m, 0, 19, 1, d);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline int msg_src_droppable(struct tipc_msg *m)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
return msg_bits(m, 0, 18, 1);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline void msg_set_src_droppable(struct tipc_msg *m, u32 d)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_bits(m, 0, 18, 1, d);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void msg_set_size(struct tipc_msg *m, u32 sz)
|
|
|
|
{
|
|
|
|
m->hdr[0] = htonl((msg_word(m, 0) & ~0x1ffff) | sz);
|
|
|
|
}
|
|
|
|
|
2015-03-25 16:07:25 +00:00
|
|
|
static inline unchar *msg_data(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return ((unchar *)m) + msg_hdr_sz(m);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline struct tipc_msg *msg_get_wrapped(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return (struct tipc_msg *)msg_data(m);
|
|
|
|
}
|
2006-01-02 18:04:38 +00:00
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
/*
|
2006-01-02 18:04:38 +00:00
|
|
|
* Word 1
|
|
|
|
*/
|
2010-11-30 12:00:53 +00:00
|
|
|
static inline u32 msg_type(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_bits(m, 1, 29, 0x7);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline void msg_set_type(struct tipc_msg *m, u32 n)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_bits(m, 1, 29, 0x7, n);
|
|
|
|
}
|
|
|
|
|
2010-11-30 12:00:53 +00:00
|
|
|
static inline u32 msg_named(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_type(m) == TIPC_NAMED_MSG;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 msg_mcast(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_type(m) == TIPC_MCAST_MSG;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 msg_connected(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_type(m) == TIPC_CONN_MSG;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 msg_errcode(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_bits(m, 1, 25, 0xf);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline void msg_set_errcode(struct tipc_msg *m, u32 err)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_bits(m, 1, 25, 0xf, err);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline u32 msg_reroute_cnt(struct tipc_msg *m)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
return msg_bits(m, 1, 21, 0xf);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline void msg_incr_reroute_cnt(struct tipc_msg *m)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_bits(m, 1, 21, 0xf, msg_reroute_cnt(m) + 1);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline void msg_reset_reroute_cnt(struct tipc_msg *m)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_bits(m, 1, 21, 0xf, 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 msg_lookup_scope(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_bits(m, 1, 19, 0x3);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline void msg_set_lookup_scope(struct tipc_msg *m, u32 n)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_bits(m, 1, 19, 0x3, n);
|
|
|
|
}
|
|
|
|
|
2015-05-14 14:46:14 +00:00
|
|
|
static inline u16 msg_bcast_ack(struct tipc_msg *m)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
return msg_bits(m, 1, 0, 0xffff);
|
|
|
|
}
|
|
|
|
|
2015-05-14 14:46:14 +00:00
|
|
|
static inline void msg_set_bcast_ack(struct tipc_msg *m, u16 n)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_bits(m, 1, 0, 0xffff, n);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
/*
|
2006-01-02 18:04:38 +00:00
|
|
|
* Word 2
|
|
|
|
*/
|
2015-05-14 14:46:14 +00:00
|
|
|
static inline u16 msg_ack(struct tipc_msg *m)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
return msg_bits(m, 2, 16, 0xffff);
|
|
|
|
}
|
|
|
|
|
2015-05-14 14:46:14 +00:00
|
|
|
static inline void msg_set_ack(struct tipc_msg *m, u16 n)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_bits(m, 2, 16, 0xffff, n);
|
|
|
|
}
|
|
|
|
|
2015-05-14 14:46:14 +00:00
|
|
|
static inline u16 msg_seqno(struct tipc_msg *m)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
return msg_bits(m, 2, 0, 0xffff);
|
|
|
|
}
|
|
|
|
|
2015-05-14 14:46:14 +00:00
|
|
|
static inline void msg_set_seqno(struct tipc_msg *m, u16 n)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_bits(m, 2, 0, 0xffff, n);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
/*
|
2006-01-02 18:04:38 +00:00
|
|
|
* Words 3-10
|
|
|
|
*/
|
tipc: clean up handling of message priorities
Messages transferred by TIPC are assigned an "importance priority", -an
integer value indicating how to treat the message when there is link or
destination socket congestion.
There is no separate header field for this value. Instead, the message
user values have been chosen in ascending order according to perceived
importance, so that the message user field can be used for this.
This is not a good solution. First, we have many more users than the
needed priority levels, so we end up with treating more priority
levels than necessary. Second, the user field cannot always
accurately reflect the priority of the message. E.g., a message
fragment packet should really have the priority of the enveloped
user data message, and not the priority of the MSG_FRAGMENTER user.
Until now, we have been working around this problem in different ways,
but it is now time to implement a consistent way of handling such
priorities, although still within the constraint that we cannot
allocate any more bits in the regular data message header for this.
In this commit, we define a new priority level, TIPC_SYSTEM_IMPORTANCE,
that will be the only one used apart from the four (lower) user data
levels. All non-data messages map down to this priority. Furthermore,
we take some free bits from the MSG_FRAGMENTER header and allocate
them to store the priority of the enveloped message. We then adjust
the functions msg_importance()/msg_set_importance() so that they
read/set the correct header fields depending on user type.
This small protocol change is fully compatible, because the code at
the receiving end of a link currently reads the importance level
only from user data messages, where there is no change.
Reviewed-by: Erik Hugne <erik.hugne@ericsson.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2015-03-13 20:08:11 +00:00
|
|
|
static inline u32 msg_importance(struct tipc_msg *m)
|
|
|
|
{
|
tipc: improve link congestion algorithm
The link congestion algorithm used until now implies two problems.
- It is too generous towards lower-level messages in situations of high
load by giving "absolute" bandwidth guarantees to the different
priority levels. LOW traffic is guaranteed 10%, MEDIUM is guaranted
20%, HIGH is guaranteed 30%, and CRITICAL is guaranteed 40% of the
available bandwidth. But, in the absence of higher level traffic, the
ratio between two distinct levels becomes unreasonable. E.g. if there
is only LOW and MEDIUM traffic on a system, the former is guaranteed
1/3 of the bandwidth, and the latter 2/3. This again means that if
there is e.g. one LOW user and 10 MEDIUM users, the former will have
33.3% of the bandwidth, and the others will have to compete for the
remainder, i.e. each will end up with 6.7% of the capacity.
- Packets of type MSG_BUNDLER are created at SYSTEM importance level,
but only after the packets bundled into it have passed the congestion
test for their own respective levels. Since bundled packets don't
result in incrementing the level counter for their own importance,
only occasionally for the SYSTEM level counter, they do in practice
obtain SYSTEM level importance. Hence, the current implementation
provides a gap in the congestion algorithm that in the worst case
may lead to a link reset.
We now refine the congestion algorithm as follows:
- A message is accepted to the link backlog only if its own level
counter, and all superior level counters, permit it.
- The importance of a created bundle packet is set according to its
contents. A bundle packet created from messges at levels LOW to
CRITICAL is given importance level CRITICAL, while a bundle created
from a SYSTEM level message is given importance SYSTEM. In the latter
case only subsequent SYSTEM level messages are allowed to be bundled
into it.
This solves the first problem described above, by making the bandwidth
guarantee relative to the total number of users at all levels; only
the upper limit for each level remains absolute. In the example
described above, the single LOW user would use 1/11th of the bandwidth,
the same as each of the ten MEDIUM users, but he still has the same
guarantee against starvation as the latter ones.
The fix also solves the second problem. If the CRITICAL level is filled
up by bundle packets of that level, no lower level packets will be
accepted any more.
Suggested-by: Gergely Kiss <gergely.kiss@ericsson.com>
Reviewed-by: Ying Xue <ying.xue@windriver.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2015-05-14 14:46:17 +00:00
|
|
|
int usr = msg_user(m);
|
|
|
|
|
|
|
|
if (likely((usr <= TIPC_CRITICAL_IMPORTANCE) && !msg_errcode(m)))
|
|
|
|
return usr;
|
|
|
|
if ((usr == MSG_FRAGMENTER) || (usr == MSG_BUNDLER))
|
2015-10-14 13:23:18 +00:00
|
|
|
return msg_bits(m, 9, 0, 0x7);
|
tipc: clean up handling of message priorities
Messages transferred by TIPC are assigned an "importance priority", -an
integer value indicating how to treat the message when there is link or
destination socket congestion.
There is no separate header field for this value. Instead, the message
user values have been chosen in ascending order according to perceived
importance, so that the message user field can be used for this.
This is not a good solution. First, we have many more users than the
needed priority levels, so we end up with treating more priority
levels than necessary. Second, the user field cannot always
accurately reflect the priority of the message. E.g., a message
fragment packet should really have the priority of the enveloped
user data message, and not the priority of the MSG_FRAGMENTER user.
Until now, we have been working around this problem in different ways,
but it is now time to implement a consistent way of handling such
priorities, although still within the constraint that we cannot
allocate any more bits in the regular data message header for this.
In this commit, we define a new priority level, TIPC_SYSTEM_IMPORTANCE,
that will be the only one used apart from the four (lower) user data
levels. All non-data messages map down to this priority. Furthermore,
we take some free bits from the MSG_FRAGMENTER header and allocate
them to store the priority of the enveloped message. We then adjust
the functions msg_importance()/msg_set_importance() so that they
read/set the correct header fields depending on user type.
This small protocol change is fully compatible, because the code at
the receiving end of a link currently reads the importance level
only from user data messages, where there is no change.
Reviewed-by: Erik Hugne <erik.hugne@ericsson.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2015-03-13 20:08:11 +00:00
|
|
|
return TIPC_SYSTEM_IMPORTANCE;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void msg_set_importance(struct tipc_msg *m, u32 i)
|
|
|
|
{
|
tipc: improve link congestion algorithm
The link congestion algorithm used until now implies two problems.
- It is too generous towards lower-level messages in situations of high
load by giving "absolute" bandwidth guarantees to the different
priority levels. LOW traffic is guaranteed 10%, MEDIUM is guaranted
20%, HIGH is guaranteed 30%, and CRITICAL is guaranteed 40% of the
available bandwidth. But, in the absence of higher level traffic, the
ratio between two distinct levels becomes unreasonable. E.g. if there
is only LOW and MEDIUM traffic on a system, the former is guaranteed
1/3 of the bandwidth, and the latter 2/3. This again means that if
there is e.g. one LOW user and 10 MEDIUM users, the former will have
33.3% of the bandwidth, and the others will have to compete for the
remainder, i.e. each will end up with 6.7% of the capacity.
- Packets of type MSG_BUNDLER are created at SYSTEM importance level,
but only after the packets bundled into it have passed the congestion
test for their own respective levels. Since bundled packets don't
result in incrementing the level counter for their own importance,
only occasionally for the SYSTEM level counter, they do in practice
obtain SYSTEM level importance. Hence, the current implementation
provides a gap in the congestion algorithm that in the worst case
may lead to a link reset.
We now refine the congestion algorithm as follows:
- A message is accepted to the link backlog only if its own level
counter, and all superior level counters, permit it.
- The importance of a created bundle packet is set according to its
contents. A bundle packet created from messges at levels LOW to
CRITICAL is given importance level CRITICAL, while a bundle created
from a SYSTEM level message is given importance SYSTEM. In the latter
case only subsequent SYSTEM level messages are allowed to be bundled
into it.
This solves the first problem described above, by making the bandwidth
guarantee relative to the total number of users at all levels; only
the upper limit for each level remains absolute. In the example
described above, the single LOW user would use 1/11th of the bandwidth,
the same as each of the ten MEDIUM users, but he still has the same
guarantee against starvation as the latter ones.
The fix also solves the second problem. If the CRITICAL level is filled
up by bundle packets of that level, no lower level packets will be
accepted any more.
Suggested-by: Gergely Kiss <gergely.kiss@ericsson.com>
Reviewed-by: Ying Xue <ying.xue@windriver.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2015-05-14 14:46:17 +00:00
|
|
|
int usr = msg_user(m);
|
|
|
|
|
|
|
|
if (likely((usr == MSG_FRAGMENTER) || (usr == MSG_BUNDLER)))
|
2015-10-14 13:23:18 +00:00
|
|
|
msg_set_bits(m, 9, 0, 0x7, i);
|
tipc: improve link congestion algorithm
The link congestion algorithm used until now implies two problems.
- It is too generous towards lower-level messages in situations of high
load by giving "absolute" bandwidth guarantees to the different
priority levels. LOW traffic is guaranteed 10%, MEDIUM is guaranted
20%, HIGH is guaranteed 30%, and CRITICAL is guaranteed 40% of the
available bandwidth. But, in the absence of higher level traffic, the
ratio between two distinct levels becomes unreasonable. E.g. if there
is only LOW and MEDIUM traffic on a system, the former is guaranteed
1/3 of the bandwidth, and the latter 2/3. This again means that if
there is e.g. one LOW user and 10 MEDIUM users, the former will have
33.3% of the bandwidth, and the others will have to compete for the
remainder, i.e. each will end up with 6.7% of the capacity.
- Packets of type MSG_BUNDLER are created at SYSTEM importance level,
but only after the packets bundled into it have passed the congestion
test for their own respective levels. Since bundled packets don't
result in incrementing the level counter for their own importance,
only occasionally for the SYSTEM level counter, they do in practice
obtain SYSTEM level importance. Hence, the current implementation
provides a gap in the congestion algorithm that in the worst case
may lead to a link reset.
We now refine the congestion algorithm as follows:
- A message is accepted to the link backlog only if its own level
counter, and all superior level counters, permit it.
- The importance of a created bundle packet is set according to its
contents. A bundle packet created from messges at levels LOW to
CRITICAL is given importance level CRITICAL, while a bundle created
from a SYSTEM level message is given importance SYSTEM. In the latter
case only subsequent SYSTEM level messages are allowed to be bundled
into it.
This solves the first problem described above, by making the bandwidth
guarantee relative to the total number of users at all levels; only
the upper limit for each level remains absolute. In the example
described above, the single LOW user would use 1/11th of the bandwidth,
the same as each of the ten MEDIUM users, but he still has the same
guarantee against starvation as the latter ones.
The fix also solves the second problem. If the CRITICAL level is filled
up by bundle packets of that level, no lower level packets will be
accepted any more.
Suggested-by: Gergely Kiss <gergely.kiss@ericsson.com>
Reviewed-by: Ying Xue <ying.xue@windriver.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2015-05-14 14:46:17 +00:00
|
|
|
else if (i < TIPC_SYSTEM_IMPORTANCE)
|
tipc: clean up handling of message priorities
Messages transferred by TIPC are assigned an "importance priority", -an
integer value indicating how to treat the message when there is link or
destination socket congestion.
There is no separate header field for this value. Instead, the message
user values have been chosen in ascending order according to perceived
importance, so that the message user field can be used for this.
This is not a good solution. First, we have many more users than the
needed priority levels, so we end up with treating more priority
levels than necessary. Second, the user field cannot always
accurately reflect the priority of the message. E.g., a message
fragment packet should really have the priority of the enveloped
user data message, and not the priority of the MSG_FRAGMENTER user.
Until now, we have been working around this problem in different ways,
but it is now time to implement a consistent way of handling such
priorities, although still within the constraint that we cannot
allocate any more bits in the regular data message header for this.
In this commit, we define a new priority level, TIPC_SYSTEM_IMPORTANCE,
that will be the only one used apart from the four (lower) user data
levels. All non-data messages map down to this priority. Furthermore,
we take some free bits from the MSG_FRAGMENTER header and allocate
them to store the priority of the enveloped message. We then adjust
the functions msg_importance()/msg_set_importance() so that they
read/set the correct header fields depending on user type.
This small protocol change is fully compatible, because the code at
the receiving end of a link currently reads the importance level
only from user data messages, where there is no change.
Reviewed-by: Erik Hugne <erik.hugne@ericsson.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2015-03-13 20:08:11 +00:00
|
|
|
msg_set_user(m, i);
|
|
|
|
else
|
|
|
|
pr_warn("Trying to set illegal importance in message\n");
|
|
|
|
}
|
|
|
|
|
2010-11-30 12:00:53 +00:00
|
|
|
static inline u32 msg_prevnode(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_word(m, 3);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline void msg_set_prevnode(struct tipc_msg *m, u32 a)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_word(m, 3, a);
|
|
|
|
}
|
|
|
|
|
2010-11-30 12:00:53 +00:00
|
|
|
static inline u32 msg_origport(struct tipc_msg *m)
|
|
|
|
{
|
2015-03-25 16:07:25 +00:00
|
|
|
if (msg_user(m) == MSG_FRAGMENTER)
|
|
|
|
m = msg_get_wrapped(m);
|
2010-11-30 12:00:53 +00:00
|
|
|
return msg_word(m, 4);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline void msg_set_origport(struct tipc_msg *m, u32 p)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_word(m, 4, p);
|
|
|
|
}
|
|
|
|
|
2010-11-30 12:00:53 +00:00
|
|
|
static inline u32 msg_destport(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_word(m, 5);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline void msg_set_destport(struct tipc_msg *m, u32 p)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_word(m, 5, p);
|
|
|
|
}
|
|
|
|
|
2010-11-30 12:00:53 +00:00
|
|
|
static inline u32 msg_mc_netid(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_word(m, 5);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline void msg_set_mc_netid(struct tipc_msg *m, u32 p)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_word(m, 5, p);
|
|
|
|
}
|
|
|
|
|
2010-11-30 12:00:53 +00:00
|
|
|
static inline int msg_short(struct tipc_msg *m)
|
|
|
|
{
|
2011-05-31 19:03:18 +00:00
|
|
|
return msg_hdr_sz(m) == SHORT_H_SIZE;
|
2010-11-30 12:00:53 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 msg_orignode(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
if (likely(msg_short(m)))
|
|
|
|
return msg_prevnode(m);
|
|
|
|
return msg_word(m, 6);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline void msg_set_orignode(struct tipc_msg *m, u32 a)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_word(m, 6, a);
|
|
|
|
}
|
|
|
|
|
2010-11-30 12:00:53 +00:00
|
|
|
static inline u32 msg_destnode(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_word(m, 7);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline void msg_set_destnode(struct tipc_msg *m, u32 a)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_word(m, 7, a);
|
|
|
|
}
|
|
|
|
|
2010-11-30 12:00:53 +00:00
|
|
|
static inline u32 msg_nametype(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_word(m, 8);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline void msg_set_nametype(struct tipc_msg *m, u32 n)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_word(m, 8, n);
|
|
|
|
}
|
|
|
|
|
2010-11-30 12:00:53 +00:00
|
|
|
static inline u32 msg_nameinst(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_word(m, 9);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 msg_namelower(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_nameinst(m);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline void msg_set_namelower(struct tipc_msg *m, u32 n)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_word(m, 9, n);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline void msg_set_nameinst(struct tipc_msg *m, u32 n)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_namelower(m, n);
|
|
|
|
}
|
|
|
|
|
2010-11-30 12:00:53 +00:00
|
|
|
static inline u32 msg_nameupper(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_word(m, 10);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline void msg_set_nameupper(struct tipc_msg *m, u32 n)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_word(m, 10, n);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2011-04-08 14:50:52 +00:00
|
|
|
* Constants and routines used to read and write TIPC internal message headers
|
|
|
|
*/
|
2006-01-02 18:04:38 +00:00
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
/*
|
2011-04-08 14:50:52 +00:00
|
|
|
* Connection management protocol message types
|
2006-01-02 18:04:38 +00:00
|
|
|
*/
|
|
|
|
#define CONN_PROBE 0
|
|
|
|
#define CONN_PROBE_REPLY 1
|
|
|
|
#define CONN_ACK 2
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
/*
|
2011-04-08 14:50:52 +00:00
|
|
|
* Name distributor message types
|
2006-01-02 18:04:38 +00:00
|
|
|
*/
|
|
|
|
#define PUBLICATION 0
|
|
|
|
#define WITHDRAWAL 1
|
|
|
|
|
2011-04-08 15:04:15 +00:00
|
|
|
/*
|
|
|
|
* Segmentation message types
|
|
|
|
*/
|
|
|
|
#define FIRST_FRAGMENT 0
|
|
|
|
#define FRAGMENT 1
|
|
|
|
#define LAST_FRAGMENT 2
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Link management protocol message types
|
|
|
|
*/
|
|
|
|
#define STATE_MSG 0
|
|
|
|
#define RESET_MSG 1
|
|
|
|
#define ACTIVATE_MSG 2
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Changeover tunnel message types
|
|
|
|
*/
|
tipc: eliminate delayed link deletion at link failover
When a bearer is disabled manually, all its links have to be reset
and deleted. However, if there is a remaining, parallel link ready
to take over a deleted link's traffic, we currently delay the delete
of the removed link until the failover procedure is finished. This
is because the remaining link needs to access state from the reset
link, such as the last received packet number, and any partially
reassembled buffer, in order to perform a successful failover.
In this commit, we do instead move the state data over to the new
link, so that it can fulfill the procedure autonomously, without
accessing any data on the old link. This means that we can now
proceed and delete all pertaining links immediately when a bearer
is disabled. This saves us from some unnecessary complexity in such
situations.
We also choose to change the confusing definitions CHANGEOVER_PROTOCOL,
ORIGINAL_MSG and DUPLICATE_MSG to the more descriptive TUNNEL_PROTOCOL,
FAILOVER_MSG and SYNCH_MSG respectively.
Reviewed-by: Ying Xue <ying.xue@windriver.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2015-04-02 13:33:01 +00:00
|
|
|
#define SYNCH_MSG 0
|
|
|
|
#define FAILOVER_MSG 1
|
2011-04-08 15:04:15 +00:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Config protocol message types
|
|
|
|
*/
|
|
|
|
#define DSC_REQ_MSG 0
|
|
|
|
#define DSC_RESP_MSG 1
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
/*
|
2006-01-02 18:04:38 +00:00
|
|
|
* Word 1
|
|
|
|
*/
|
|
|
|
static inline u32 msg_seq_gap(struct tipc_msg *m)
|
|
|
|
{
|
2008-06-05 00:47:30 +00:00
|
|
|
return msg_bits(m, 1, 16, 0x1fff);
|
2006-01-02 18:04:38 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static inline void msg_set_seq_gap(struct tipc_msg *m, u32 n)
|
|
|
|
{
|
2008-06-05 00:47:30 +00:00
|
|
|
msg_set_bits(m, 1, 16, 0x1fff, n);
|
2006-01-02 18:04:38 +00:00
|
|
|
}
|
|
|
|
|
2011-10-28 20:26:41 +00:00
|
|
|
static inline u32 msg_node_sig(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_bits(m, 1, 0, 0xffff);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void msg_set_node_sig(struct tipc_msg *m, u32 n)
|
|
|
|
{
|
|
|
|
msg_set_bits(m, 1, 0, 0xffff, n);
|
|
|
|
}
|
|
|
|
|
2015-03-13 20:08:05 +00:00
|
|
|
static inline u32 msg_node_capabilities(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_bits(m, 1, 15, 0x1fff);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void msg_set_node_capabilities(struct tipc_msg *m, u32 n)
|
|
|
|
{
|
|
|
|
msg_set_bits(m, 1, 15, 0x1fff, n);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
/*
|
2006-01-02 18:04:38 +00:00
|
|
|
* Word 2
|
|
|
|
*/
|
|
|
|
static inline u32 msg_dest_domain(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_word(m, 2);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline void msg_set_dest_domain(struct tipc_msg *m, u32 n)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_word(m, 2, n);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 msg_bcgap_after(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_bits(m, 2, 16, 0xffff);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void msg_set_bcgap_after(struct tipc_msg *m, u32 n)
|
|
|
|
{
|
|
|
|
msg_set_bits(m, 2, 16, 0xffff, n);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 msg_bcgap_to(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_bits(m, 2, 0, 0xffff);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline void msg_set_bcgap_to(struct tipc_msg *m, u32 n)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_bits(m, 2, 0, 0xffff, n);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
/*
|
2006-01-02 18:04:38 +00:00
|
|
|
* Word 4
|
|
|
|
*/
|
|
|
|
static inline u32 msg_last_bcast(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_bits(m, 4, 16, 0xffff);
|
|
|
|
}
|
|
|
|
|
2015-10-22 12:51:41 +00:00
|
|
|
static inline u32 msg_bc_snd_nxt(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_last_bcast(m) + 1;
|
|
|
|
}
|
|
|
|
|
2006-01-02 18:04:38 +00:00
|
|
|
static inline void msg_set_last_bcast(struct tipc_msg *m, u32 n)
|
|
|
|
{
|
|
|
|
msg_set_bits(m, 4, 16, 0xffff, n);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void msg_set_fragm_no(struct tipc_msg *m, u32 n)
|
|
|
|
{
|
|
|
|
msg_set_bits(m, 4, 16, 0xffff, n);
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2015-07-30 22:24:19 +00:00
|
|
|
static inline u16 msg_next_sent(struct tipc_msg *m)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
return msg_bits(m, 4, 0, 0xffff);
|
|
|
|
}
|
|
|
|
|
2015-07-30 22:24:19 +00:00
|
|
|
static inline void msg_set_next_sent(struct tipc_msg *m, u16 n)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_bits(m, 4, 0, 0xffff, n);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void msg_set_long_msgno(struct tipc_msg *m, u32 n)
|
|
|
|
{
|
|
|
|
msg_set_bits(m, 4, 0, 0xffff, n);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 msg_bc_netid(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_word(m, 4);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void msg_set_bc_netid(struct tipc_msg *m, u32 id)
|
|
|
|
{
|
|
|
|
msg_set_word(m, 4, id);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 msg_link_selector(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_bits(m, 4, 0, 1);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void msg_set_link_selector(struct tipc_msg *m, u32 n)
|
|
|
|
{
|
2011-05-25 17:28:27 +00:00
|
|
|
msg_set_bits(m, 4, 0, 1, n);
|
2006-01-02 18:04:38 +00:00
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
/*
|
2006-01-02 18:04:38 +00:00
|
|
|
* Word 5
|
|
|
|
*/
|
tipc: reduce locking scope during packet reception
We convert packet/message reception according to the same principle
we have been using for message sending and timeout handling:
We move the function tipc_rcv() to node.c, hence handling the initial
packet reception at the link aggregation level. The function grabs
the node lock, selects the receiving link, and accesses it via a new
call tipc_link_rcv(). This function appends buffers to the input
queue for delivery upwards, but it may also append outgoing packets
to the xmit queue, just as we do during regular message sending. The
latter will happen when buffers are forwarded from the link backlog,
or when retransmission is requested.
Upon return of this function, and after having released the node lock,
tipc_rcv() delivers/tranmsits the contents of those queues, but it may
also perform actions such as link activation or reset, as indicated by
the return flags from the link.
This reduces the number of cpu cycles spent inside the node spinlock,
and reduces contention on that lock.
Reviewed-by: Ying Xue <ying.xue@windriver.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2015-07-16 20:54:31 +00:00
|
|
|
static inline u16 msg_session(struct tipc_msg *m)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
return msg_bits(m, 5, 16, 0xffff);
|
|
|
|
}
|
|
|
|
|
tipc: reduce locking scope during packet reception
We convert packet/message reception according to the same principle
we have been using for message sending and timeout handling:
We move the function tipc_rcv() to node.c, hence handling the initial
packet reception at the link aggregation level. The function grabs
the node lock, selects the receiving link, and accesses it via a new
call tipc_link_rcv(). This function appends buffers to the input
queue for delivery upwards, but it may also append outgoing packets
to the xmit queue, just as we do during regular message sending. The
latter will happen when buffers are forwarded from the link backlog,
or when retransmission is requested.
Upon return of this function, and after having released the node lock,
tipc_rcv() delivers/tranmsits the contents of those queues, but it may
also perform actions such as link activation or reset, as indicated by
the return flags from the link.
This reduces the number of cpu cycles spent inside the node spinlock,
and reduces contention on that lock.
Reviewed-by: Ying Xue <ying.xue@windriver.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2015-07-16 20:54:31 +00:00
|
|
|
static inline void msg_set_session(struct tipc_msg *m, u16 n)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_bits(m, 5, 16, 0xffff, n);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 msg_probe(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_bits(m, 5, 0, 1);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void msg_set_probe(struct tipc_msg *m, u32 val)
|
|
|
|
{
|
2011-05-25 17:28:27 +00:00
|
|
|
msg_set_bits(m, 5, 0, 1, val);
|
2006-01-02 18:04:38 +00:00
|
|
|
}
|
|
|
|
|
|
|
|
static inline char msg_net_plane(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_bits(m, 5, 1, 7) + 'A';
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void msg_set_net_plane(struct tipc_msg *m, char n)
|
|
|
|
{
|
|
|
|
msg_set_bits(m, 5, 1, 7, (n - 'A'));
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 msg_linkprio(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_bits(m, 5, 4, 0x1f);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void msg_set_linkprio(struct tipc_msg *m, u32 n)
|
|
|
|
{
|
|
|
|
msg_set_bits(m, 5, 4, 0x1f, n);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 msg_bearer_id(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_bits(m, 5, 9, 0x7);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void msg_set_bearer_id(struct tipc_msg *m, u32 n)
|
|
|
|
{
|
|
|
|
msg_set_bits(m, 5, 9, 0x7, n);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 msg_redundant_link(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_bits(m, 5, 12, 0x1);
|
|
|
|
}
|
|
|
|
|
2011-02-28 20:30:20 +00:00
|
|
|
static inline void msg_set_redundant_link(struct tipc_msg *m, u32 r)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
2011-02-28 20:30:20 +00:00
|
|
|
msg_set_bits(m, 5, 12, 0x1, r);
|
2006-01-02 18:04:38 +00:00
|
|
|
}
|
|
|
|
|
tipc: guarantee peer bearer id exchange after reboot
When a link endpoint is going down locally, e.g., because its interface
is being stopped, it will spontaneously send out a RESET message to
its peer, informing it about this fact. This saves the peer from
detecting the failure via probing, and hence gives both speedier and
less resource consuming failure detection on the peer side.
According to the link FSM, a receiver of a RESET message, ignoring the
reason for it, must now consider the sender ready to come back up, and
starts periodically sending out ACTIVATE messages to the peer in order
to re-establish the link. Also, according to the FSM, the receiver of
an ACTIVATE message can now go directly to state ESTABLISHED and start
sending regular traffic packets. This is a well-proven and robust FSM.
However, in the case of a reboot, there is a small possibilty that link
endpoint on the rebooted node may have been re-created with a new bearer
identity between the moment it sent its (pre-boot) RESET and the moment
it receives the ACTIVATE from the peer. The new bearer identity cannot
be known by the peer according to this scenario, since traffic headers
don't convey such information. This is a problem, because both endpoints
need to know the correct value of the peer's bearer id at any moment in
time in order to be able to produce correct link events for their users.
The only way to guarantee this is to enforce a full setup message
exchange (RESET + ACTIVATE) even after the reboot, since those messages
carry the bearer idientity in their header.
In this commit we do this by introducing and setting a "stopping" bit in
the header of the spontaneously generated RESET messages, informing the
peer that the sender will not be immediately ready to re-establish the
link. A receiver seeing this bit must act as if this were a locally
detected connectivity failure, and hence has to go through a full two-
way setup message exchange before any link can be re-established.
Although never reported, this problem seems to have always been around.
This protocol addition is fully backwards compatible.
Acked-by: Ying Xue <ying.xue@windriver.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2016-04-15 17:33:03 +00:00
|
|
|
static inline u32 msg_peer_stopping(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_bits(m, 5, 13, 0x1);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void msg_set_peer_stopping(struct tipc_msg *m, u32 s)
|
|
|
|
{
|
|
|
|
msg_set_bits(m, 5, 13, 0x1, s);
|
|
|
|
}
|
|
|
|
|
2011-10-07 19:19:11 +00:00
|
|
|
static inline char *msg_media_addr(struct tipc_msg *m)
|
|
|
|
{
|
2015-02-27 07:56:57 +00:00
|
|
|
return (char *)&m->hdr[TIPC_MEDIA_INFO_OFFSET];
|
2011-10-07 19:19:11 +00:00
|
|
|
}
|
2006-01-02 18:04:38 +00:00
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
/*
|
2006-01-02 18:04:38 +00:00
|
|
|
* Word 9
|
|
|
|
*/
|
2015-07-30 22:24:19 +00:00
|
|
|
static inline u16 msg_msgcnt(struct tipc_msg *m)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
return msg_bits(m, 9, 16, 0xffff);
|
|
|
|
}
|
|
|
|
|
2015-07-30 22:24:19 +00:00
|
|
|
static inline void msg_set_msgcnt(struct tipc_msg *m, u16 n)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_bits(m, 9, 16, 0xffff, n);
|
|
|
|
}
|
|
|
|
|
tipc: redesign connection-level flow control
There are two flow control mechanisms in TIPC; one at link level that
handles network congestion, burst control, and retransmission, and one
at connection level which' only remaining task is to prevent overflow
in the receiving socket buffer. In TIPC, the latter task has to be
solved end-to-end because messages can not be thrown away once they
have been accepted and delivered upwards from the link layer, i.e, we
can never permit the receive buffer to overflow.
Currently, this algorithm is message based. A counter in the receiving
socket keeps track of number of consumed messages, and sends a dedicated
acknowledge message back to the sender for each 256 consumed message.
A counter at the sending end keeps track of the sent, not yet
acknowledged messages, and blocks the sender if this number ever reaches
512 unacknowledged messages. When the missing acknowledge arrives, the
socket is then woken up for renewed transmission. This works well for
keeping the message flow running, as it almost never happens that a
sender socket is blocked this way.
A problem with the current mechanism is that it potentially is very
memory consuming. Since we don't distinguish between small and large
messages, we have to dimension the socket receive buffer according
to a worst-case of both. I.e., the window size must be chosen large
enough to sustain a reasonable throughput even for the smallest
messages, while we must still consider a scenario where all messages
are of maximum size. Hence, the current fix window size of 512 messages
and a maximum message size of 66k results in a receive buffer of 66 MB
when truesize(66k) = 131k is taken into account. It is possible to do
much better.
This commit introduces an algorithm where we instead use 1024-byte
blocks as base unit. This unit, always rounded upwards from the
actual message size, is used when we advertise windows as well as when
we count and acknowledge transmitted data. The advertised window is
based on the configured receive buffer size in such a way that even
the worst-case truesize/msgsize ratio always is covered. Since the
smallest possible message size (from a flow control viewpoint) now is
1024 bytes, we can safely assume this ratio to be less than four, which
is the value we are now using.
This way, we have been able to reduce the default receive buffer size
from 66 MB to 2 MB with maintained performance.
In order to keep this solution backwards compatible, we introduce a
new capability bit in the discovery protocol, and use this throughout
the message sending/reception path to always select the right unit.
Acked-by: Ying Xue <ying.xue@windriver.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2016-05-02 15:58:47 +00:00
|
|
|
static inline u32 msg_conn_ack(struct tipc_msg *m)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
return msg_bits(m, 9, 16, 0xffff);
|
|
|
|
}
|
|
|
|
|
tipc: redesign connection-level flow control
There are two flow control mechanisms in TIPC; one at link level that
handles network congestion, burst control, and retransmission, and one
at connection level which' only remaining task is to prevent overflow
in the receiving socket buffer. In TIPC, the latter task has to be
solved end-to-end because messages can not be thrown away once they
have been accepted and delivered upwards from the link layer, i.e, we
can never permit the receive buffer to overflow.
Currently, this algorithm is message based. A counter in the receiving
socket keeps track of number of consumed messages, and sends a dedicated
acknowledge message back to the sender for each 256 consumed message.
A counter at the sending end keeps track of the sent, not yet
acknowledged messages, and blocks the sender if this number ever reaches
512 unacknowledged messages. When the missing acknowledge arrives, the
socket is then woken up for renewed transmission. This works well for
keeping the message flow running, as it almost never happens that a
sender socket is blocked this way.
A problem with the current mechanism is that it potentially is very
memory consuming. Since we don't distinguish between small and large
messages, we have to dimension the socket receive buffer according
to a worst-case of both. I.e., the window size must be chosen large
enough to sustain a reasonable throughput even for the smallest
messages, while we must still consider a scenario where all messages
are of maximum size. Hence, the current fix window size of 512 messages
and a maximum message size of 66k results in a receive buffer of 66 MB
when truesize(66k) = 131k is taken into account. It is possible to do
much better.
This commit introduces an algorithm where we instead use 1024-byte
blocks as base unit. This unit, always rounded upwards from the
actual message size, is used when we advertise windows as well as when
we count and acknowledge transmitted data. The advertised window is
based on the configured receive buffer size in such a way that even
the worst-case truesize/msgsize ratio always is covered. Since the
smallest possible message size (from a flow control viewpoint) now is
1024 bytes, we can safely assume this ratio to be less than four, which
is the value we are now using.
This way, we have been able to reduce the default receive buffer size
from 66 MB to 2 MB with maintained performance.
In order to keep this solution backwards compatible, we introduce a
new capability bit in the discovery protocol, and use this throughout
the message sending/reception path to always select the right unit.
Acked-by: Ying Xue <ying.xue@windriver.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2016-05-02 15:58:47 +00:00
|
|
|
static inline void msg_set_conn_ack(struct tipc_msg *m, u32 n)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_bits(m, 9, 16, 0xffff, n);
|
|
|
|
}
|
|
|
|
|
tipc: redesign connection-level flow control
There are two flow control mechanisms in TIPC; one at link level that
handles network congestion, burst control, and retransmission, and one
at connection level which' only remaining task is to prevent overflow
in the receiving socket buffer. In TIPC, the latter task has to be
solved end-to-end because messages can not be thrown away once they
have been accepted and delivered upwards from the link layer, i.e, we
can never permit the receive buffer to overflow.
Currently, this algorithm is message based. A counter in the receiving
socket keeps track of number of consumed messages, and sends a dedicated
acknowledge message back to the sender for each 256 consumed message.
A counter at the sending end keeps track of the sent, not yet
acknowledged messages, and blocks the sender if this number ever reaches
512 unacknowledged messages. When the missing acknowledge arrives, the
socket is then woken up for renewed transmission. This works well for
keeping the message flow running, as it almost never happens that a
sender socket is blocked this way.
A problem with the current mechanism is that it potentially is very
memory consuming. Since we don't distinguish between small and large
messages, we have to dimension the socket receive buffer according
to a worst-case of both. I.e., the window size must be chosen large
enough to sustain a reasonable throughput even for the smallest
messages, while we must still consider a scenario where all messages
are of maximum size. Hence, the current fix window size of 512 messages
and a maximum message size of 66k results in a receive buffer of 66 MB
when truesize(66k) = 131k is taken into account. It is possible to do
much better.
This commit introduces an algorithm where we instead use 1024-byte
blocks as base unit. This unit, always rounded upwards from the
actual message size, is used when we advertise windows as well as when
we count and acknowledge transmitted data. The advertised window is
based on the configured receive buffer size in such a way that even
the worst-case truesize/msgsize ratio always is covered. Since the
smallest possible message size (from a flow control viewpoint) now is
1024 bytes, we can safely assume this ratio to be less than four, which
is the value we are now using.
This way, we have been able to reduce the default receive buffer size
from 66 MB to 2 MB with maintained performance.
In order to keep this solution backwards compatible, we introduce a
new capability bit in the discovery protocol, and use this throughout
the message sending/reception path to always select the right unit.
Acked-by: Ying Xue <ying.xue@windriver.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2016-05-02 15:58:47 +00:00
|
|
|
static inline u32 msg_adv_win(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_bits(m, 9, 0, 0xffff);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void msg_set_adv_win(struct tipc_msg *m, u32 n)
|
|
|
|
{
|
|
|
|
msg_set_bits(m, 9, 0, 0xffff, n);
|
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline u32 msg_max_pkt(struct tipc_msg *m)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
2010-09-22 20:43:57 +00:00
|
|
|
return msg_bits(m, 9, 16, 0xffff) * 4;
|
2006-01-02 18:04:38 +00:00
|
|
|
}
|
|
|
|
|
2007-02-09 14:25:21 +00:00
|
|
|
static inline void msg_set_max_pkt(struct tipc_msg *m, u32 n)
|
2006-01-02 18:04:38 +00:00
|
|
|
{
|
|
|
|
msg_set_bits(m, 9, 16, 0xffff, (n / 4));
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 msg_link_tolerance(struct tipc_msg *m)
|
|
|
|
{
|
|
|
|
return msg_bits(m, 9, 0, 0xffff);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void msg_set_link_tolerance(struct tipc_msg *m, u32 n)
|
|
|
|
{
|
|
|
|
msg_set_bits(m, 9, 0, 0xffff, n);
|
|
|
|
}
|
|
|
|
|
2015-07-30 22:24:19 +00:00
|
|
|
static inline bool msg_peer_link_is_up(struct tipc_msg *m)
|
2015-07-16 20:54:30 +00:00
|
|
|
{
|
tipc: reduce locking scope during packet reception
We convert packet/message reception according to the same principle
we have been using for message sending and timeout handling:
We move the function tipc_rcv() to node.c, hence handling the initial
packet reception at the link aggregation level. The function grabs
the node lock, selects the receiving link, and accesses it via a new
call tipc_link_rcv(). This function appends buffers to the input
queue for delivery upwards, but it may also append outgoing packets
to the xmit queue, just as we do during regular message sending. The
latter will happen when buffers are forwarded from the link backlog,
or when retransmission is requested.
Upon return of this function, and after having released the node lock,
tipc_rcv() delivers/tranmsits the contents of those queues, but it may
also perform actions such as link activation or reset, as indicated by
the return flags from the link.
This reduces the number of cpu cycles spent inside the node spinlock,
and reduces contention on that lock.
Reviewed-by: Ying Xue <ying.xue@windriver.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2015-07-16 20:54:31 +00:00
|
|
|
if (likely(msg_user(m) != LINK_PROTOCOL))
|
2015-07-16 20:54:30 +00:00
|
|
|
return true;
|
2015-07-30 22:24:19 +00:00
|
|
|
if (msg_type(m) == STATE_MSG)
|
|
|
|
return true;
|
|
|
|
return false;
|
tipc: reduce locking scope during packet reception
We convert packet/message reception according to the same principle
we have been using for message sending and timeout handling:
We move the function tipc_rcv() to node.c, hence handling the initial
packet reception at the link aggregation level. The function grabs
the node lock, selects the receiving link, and accesses it via a new
call tipc_link_rcv(). This function appends buffers to the input
queue for delivery upwards, but it may also append outgoing packets
to the xmit queue, just as we do during regular message sending. The
latter will happen when buffers are forwarded from the link backlog,
or when retransmission is requested.
Upon return of this function, and after having released the node lock,
tipc_rcv() delivers/tranmsits the contents of those queues, but it may
also perform actions such as link activation or reset, as indicated by
the return flags from the link.
This reduces the number of cpu cycles spent inside the node spinlock,
and reduces contention on that lock.
Reviewed-by: Ying Xue <ying.xue@windriver.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2015-07-16 20:54:31 +00:00
|
|
|
}
|
|
|
|
|
2015-07-30 22:24:19 +00:00
|
|
|
static inline bool msg_peer_node_is_up(struct tipc_msg *m)
|
tipc: reduce locking scope during packet reception
We convert packet/message reception according to the same principle
we have been using for message sending and timeout handling:
We move the function tipc_rcv() to node.c, hence handling the initial
packet reception at the link aggregation level. The function grabs
the node lock, selects the receiving link, and accesses it via a new
call tipc_link_rcv(). This function appends buffers to the input
queue for delivery upwards, but it may also append outgoing packets
to the xmit queue, just as we do during regular message sending. The
latter will happen when buffers are forwarded from the link backlog,
or when retransmission is requested.
Upon return of this function, and after having released the node lock,
tipc_rcv() delivers/tranmsits the contents of those queues, but it may
also perform actions such as link activation or reset, as indicated by
the return flags from the link.
This reduces the number of cpu cycles spent inside the node spinlock,
and reduces contention on that lock.
Reviewed-by: Ying Xue <ying.xue@windriver.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2015-07-16 20:54:31 +00:00
|
|
|
{
|
2015-07-30 22:24:19 +00:00
|
|
|
if (msg_peer_link_is_up(m))
|
|
|
|
return true;
|
2015-07-16 20:54:30 +00:00
|
|
|
return msg_redundant_link(m);
|
|
|
|
}
|
|
|
|
|
2016-04-07 14:09:14 +00:00
|
|
|
static inline bool msg_is_reset(struct tipc_msg *hdr)
|
|
|
|
{
|
|
|
|
return (msg_user(hdr) == LINK_PROTOCOL) && (msg_type(hdr) == RESET_MSG);
|
|
|
|
}
|
|
|
|
|
2015-01-09 07:27:01 +00:00
|
|
|
struct sk_buff *tipc_buf_acquire(u32 size);
|
2015-03-13 20:08:06 +00:00
|
|
|
bool tipc_msg_validate(struct sk_buff *skb);
|
tipc: introduce new tipc_sk_respond() function
Currently, we use the code sequence
if (msg_reverse())
tipc_link_xmit_skb()
at numerous locations in socket.c. The preparation of arguments
for these calls, as well as the sequence itself, makes the code
unecessarily complex.
In this commit, we introduce a new function, tipc_sk_respond(),
that performs this call combination. We also replace some, but not
yet all, of these explicit call sequences with calls to the new
function. Notably, we let the function tipc_sk_proto_rcv() use
the new function to directly send out PROBE_REPLY messages,
instead of deferring this to the calling tipc_sk_rcv() function,
as we do now.
Reviewed-by: Ying Xue <ying.xue@windriver.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2015-07-22 14:11:19 +00:00
|
|
|
bool tipc_msg_reverse(u32 own_addr, struct sk_buff **skb, int err);
|
2015-02-05 13:36:36 +00:00
|
|
|
void tipc_msg_init(u32 own_addr, struct tipc_msg *m, u32 user, u32 type,
|
2015-01-09 07:27:10 +00:00
|
|
|
u32 hsize, u32 destnode);
|
tipc: split up function tipc_msg_eval()
The function tipc_msg_eval() is in reality doing two related, but
different tasks. First it tries to find a new destination for named
messages, in case there was no first lookup, or if the first lookup
failed. Second, it does what its name suggests, evaluating the validity
of the message and its destination, and returning an appropriate error
code depending on the result.
This is confusing, and in this commit we choose to break it up into two
functions. A new function, tipc_msg_lookup_dest(), first attempts to find
a new destination, if the message is of the right type. If this lookup
fails, or if the message should not be subject to a second lookup, the
already existing tipc_msg_reverse() is called. This function performs
prepares the message for rejection, if applicable.
Reviewed-by: Ying Xue <ying.xue@windriver.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2015-02-05 13:36:39 +00:00
|
|
|
struct sk_buff *tipc_msg_create(uint user, uint type, uint hdr_sz,
|
|
|
|
uint data_sz, u32 dnode, u32 onode,
|
|
|
|
u32 dport, u32 oport, int errcode);
|
2014-05-14 09:39:12 +00:00
|
|
|
int tipc_buf_append(struct sk_buff **headbuf, struct sk_buff **buf);
|
2015-05-14 14:46:18 +00:00
|
|
|
bool tipc_msg_bundle(struct sk_buff *skb, struct tipc_msg *msg, u32 mtu);
|
|
|
|
bool tipc_msg_make_bundle(struct sk_buff **skb, struct tipc_msg *msg,
|
|
|
|
u32 mtu, u32 dnode);
|
tipc: resolve race problem at unicast message reception
TIPC handles message cardinality and sequencing at the link layer,
before passing messages upwards to the destination sockets. During the
upcall from link to socket no locks are held. It is therefore possible,
and we see it happen occasionally, that messages arriving in different
threads and delivered in sequence still bypass each other before they
reach the destination socket. This must not happen, since it violates
the sequentiality guarantee.
We solve this by adding a new input buffer queue to the link structure.
Arriving messages are added safely to the tail of that queue by the
link, while the head of the queue is consumed, also safely, by the
receiving socket. Sequentiality is secured per socket by only allowing
buffers to be dequeued inside the socket lock. Since there may be multiple
simultaneous readers of the queue, we use a 'filter' parameter to reduce
the risk that they peek the same buffer from the queue, hence also
reducing the risk of contention on the receiving socket locks.
This solves the sequentiality problem, and seems to cause no measurable
performance degradation.
A nice side effect of this change is that lock handling in the functions
tipc_rcv() and tipc_bcast_rcv() now becomes uniform, something that
will enable future simplifications of those functions.
Reviewed-by: Ying Xue <ying.xue@windriver.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2015-02-05 13:36:41 +00:00
|
|
|
bool tipc_msg_extract(struct sk_buff *skb, struct sk_buff **iskb, int *pos);
|
2015-02-05 13:36:36 +00:00
|
|
|
int tipc_msg_build(struct tipc_msg *mhdr, struct msghdr *m,
|
2015-01-09 07:27:10 +00:00
|
|
|
int offset, int dsz, int mtu, struct sk_buff_head *list);
|
2015-07-22 14:11:20 +00:00
|
|
|
bool tipc_msg_lookup_dest(struct net *net, struct sk_buff *skb, int *err);
|
2015-10-22 12:51:39 +00:00
|
|
|
bool tipc_msg_reassemble(struct sk_buff_head *list, struct sk_buff_head *rcvq);
|
2015-10-15 18:52:43 +00:00
|
|
|
void __tipc_skb_queue_sorted(struct sk_buff_head *list, u16 seqno,
|
|
|
|
struct sk_buff *skb);
|
2014-07-17 00:41:00 +00:00
|
|
|
|
2015-05-14 14:46:14 +00:00
|
|
|
static inline u16 buf_seqno(struct sk_buff *skb)
|
|
|
|
{
|
|
|
|
return msg_seqno(buf_msg(skb));
|
|
|
|
}
|
|
|
|
|
2015-02-05 13:36:44 +00:00
|
|
|
/* tipc_skb_peek(): peek and reserve first buffer in list
|
|
|
|
* @list: list to be peeked in
|
|
|
|
* Returns pointer to first buffer in list, if any
|
|
|
|
*/
|
|
|
|
static inline struct sk_buff *tipc_skb_peek(struct sk_buff_head *list,
|
|
|
|
spinlock_t *lock)
|
|
|
|
{
|
|
|
|
struct sk_buff *skb;
|
|
|
|
|
|
|
|
spin_lock_bh(lock);
|
|
|
|
skb = skb_peek(list);
|
|
|
|
if (skb)
|
|
|
|
skb_get(skb);
|
|
|
|
spin_unlock_bh(lock);
|
|
|
|
return skb;
|
|
|
|
}
|
|
|
|
|
tipc: resolve race problem at unicast message reception
TIPC handles message cardinality and sequencing at the link layer,
before passing messages upwards to the destination sockets. During the
upcall from link to socket no locks are held. It is therefore possible,
and we see it happen occasionally, that messages arriving in different
threads and delivered in sequence still bypass each other before they
reach the destination socket. This must not happen, since it violates
the sequentiality guarantee.
We solve this by adding a new input buffer queue to the link structure.
Arriving messages are added safely to the tail of that queue by the
link, while the head of the queue is consumed, also safely, by the
receiving socket. Sequentiality is secured per socket by only allowing
buffers to be dequeued inside the socket lock. Since there may be multiple
simultaneous readers of the queue, we use a 'filter' parameter to reduce
the risk that they peek the same buffer from the queue, hence also
reducing the risk of contention on the receiving socket locks.
This solves the sequentiality problem, and seems to cause no measurable
performance degradation.
A nice side effect of this change is that lock handling in the functions
tipc_rcv() and tipc_bcast_rcv() now becomes uniform, something that
will enable future simplifications of those functions.
Reviewed-by: Ying Xue <ying.xue@windriver.com>
Signed-off-by: Jon Maloy <jon.maloy@ericsson.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2015-02-05 13:36:41 +00:00
|
|
|
/* tipc_skb_peek_port(): find a destination port, ignoring all destinations
|
|
|
|
* up to and including 'filter'.
|
|
|
|
* Note: ignoring previously tried destinations minimizes the risk of
|
|
|
|
* contention on the socket lock
|
|
|
|
* @list: list to be peeked in
|
|
|
|
* @filter: last destination to be ignored from search
|
|
|
|
* Returns a destination port number, of applicable.
|
|
|
|
*/
|
|
|
|
static inline u32 tipc_skb_peek_port(struct sk_buff_head *list, u32 filter)
|
|
|
|
{
|
|
|
|
struct sk_buff *skb;
|
|
|
|
u32 dport = 0;
|
|
|
|
bool ignore = true;
|
|
|
|
|
|
|
|
spin_lock_bh(&list->lock);
|
|
|
|
skb_queue_walk(list, skb) {
|
|
|
|
dport = msg_destport(buf_msg(skb));
|
|
|
|
if (!filter || skb_queue_is_last(list, skb))
|
|
|
|
break;
|
|
|
|
if (dport == filter)
|
|
|
|
ignore = false;
|
|
|
|
else if (!ignore)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
spin_unlock_bh(&list->lock);
|
|
|
|
return dport;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* tipc_skb_dequeue(): unlink first buffer with dest 'dport' from list
|
|
|
|
* @list: list to be unlinked from
|
|
|
|
* @dport: selection criteria for buffer to unlink
|
|
|
|
*/
|
|
|
|
static inline struct sk_buff *tipc_skb_dequeue(struct sk_buff_head *list,
|
|
|
|
u32 dport)
|
|
|
|
{
|
|
|
|
struct sk_buff *_skb, *tmp, *skb = NULL;
|
|
|
|
|
|
|
|
spin_lock_bh(&list->lock);
|
|
|
|
skb_queue_walk_safe(list, _skb, tmp) {
|
|
|
|
if (msg_destport(buf_msg(_skb)) == dport) {
|
|
|
|
__skb_unlink(_skb, list);
|
|
|
|
skb = _skb;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
spin_unlock_bh(&list->lock);
|
|
|
|
return skb;
|
|
|
|
}
|
|
|
|
|
2015-07-30 22:24:23 +00:00
|
|
|
/* tipc_skb_queue_splice_tail - append an skb list to lock protected list
|
|
|
|
* @list: the new list to append. Not lock protected
|
|
|
|
* @head: target list. Lock protected.
|
|
|
|
*/
|
|
|
|
static inline void tipc_skb_queue_splice_tail(struct sk_buff_head *list,
|
|
|
|
struct sk_buff_head *head)
|
|
|
|
{
|
|
|
|
spin_lock_bh(&head->lock);
|
|
|
|
skb_queue_splice_tail(list, head);
|
|
|
|
spin_unlock_bh(&head->lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* tipc_skb_queue_splice_tail_init - merge two lock protected skb lists
|
|
|
|
* @list: the new list to add. Lock protected. Will be reinitialized
|
|
|
|
* @head: target list. Lock protected.
|
|
|
|
*/
|
|
|
|
static inline void tipc_skb_queue_splice_tail_init(struct sk_buff_head *list,
|
|
|
|
struct sk_buff_head *head)
|
|
|
|
{
|
|
|
|
struct sk_buff_head tmp;
|
|
|
|
|
|
|
|
__skb_queue_head_init(&tmp);
|
|
|
|
|
|
|
|
spin_lock_bh(&list->lock);
|
|
|
|
skb_queue_splice_tail_init(list, &tmp);
|
|
|
|
spin_unlock_bh(&list->lock);
|
|
|
|
tipc_skb_queue_splice_tail(&tmp, head);
|
|
|
|
}
|
|
|
|
|
2006-01-02 18:04:38 +00:00
|
|
|
#endif
|