2005-08-12 12:26:18 +00:00
|
|
|
/*
|
|
|
|
* INET An implementation of the TCP/IP protocol suite for the LINUX
|
|
|
|
* operating system. INET is implemented using the BSD Socket
|
|
|
|
* interface as the means of communication with the user level.
|
|
|
|
*
|
|
|
|
* Generic INET6 transport hashtables
|
|
|
|
*
|
2005-12-14 07:25:44 +00:00
|
|
|
* Authors: Lotsa people, from code originally in tcp, generalised here
|
|
|
|
* by Arnaldo Carvalho de Melo <acme@mandriva.com>
|
2005-08-12 12:26:18 +00:00
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or
|
|
|
|
* modify it under the terms of the GNU General Public License
|
|
|
|
* as published by the Free Software Foundation; either version
|
|
|
|
* 2 of the License, or (at your option) any later version.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/module.h>
|
2005-12-14 07:25:44 +00:00
|
|
|
#include <linux/random.h>
|
2005-08-12 12:26:18 +00:00
|
|
|
|
|
|
|
#include <net/inet_connection_sock.h>
|
|
|
|
#include <net/inet_hashtables.h>
|
|
|
|
#include <net/inet6_hashtables.h>
|
2005-12-14 07:25:44 +00:00
|
|
|
#include <net/ip.h>
|
2005-08-12 12:26:18 +00:00
|
|
|
|
2006-04-10 05:48:59 +00:00
|
|
|
void __inet6_hash(struct inet_hashinfo *hashinfo,
|
|
|
|
struct sock *sk)
|
|
|
|
{
|
|
|
|
struct hlist_head *list;
|
|
|
|
rwlock_t *lock;
|
|
|
|
|
|
|
|
BUG_TRAP(sk_unhashed(sk));
|
|
|
|
|
|
|
|
if (sk->sk_state == TCP_LISTEN) {
|
|
|
|
list = &hashinfo->listening_hash[inet_sk_listen_hashfn(sk)];
|
|
|
|
lock = &hashinfo->lhash_lock;
|
|
|
|
inet_listen_wlock(hashinfo);
|
|
|
|
} else {
|
|
|
|
unsigned int hash;
|
|
|
|
sk->sk_hash = hash = inet6_sk_ehashfn(sk);
|
|
|
|
hash &= (hashinfo->ehash_size - 1);
|
|
|
|
list = &hashinfo->ehash[hash].chain;
|
|
|
|
lock = &hashinfo->ehash[hash].lock;
|
|
|
|
write_lock(lock);
|
|
|
|
}
|
|
|
|
|
|
|
|
__sk_add_node(sk, list);
|
|
|
|
sock_prot_inc_use(sk->sk_prot);
|
|
|
|
write_unlock(lock);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(__inet6_hash);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Sockets in TCP_CLOSE state are _always_ taken out of the hash, so
|
|
|
|
* we need not check it for TCP lookups anymore, thanks Alexey. -DaveM
|
|
|
|
*
|
|
|
|
* The sockhash lock must be held as a reader here.
|
|
|
|
*/
|
|
|
|
struct sock *__inet6_lookup_established(struct inet_hashinfo *hashinfo,
|
|
|
|
const struct in6_addr *saddr,
|
2006-11-08 08:20:00 +00:00
|
|
|
const __be16 sport,
|
2006-04-10 05:48:59 +00:00
|
|
|
const struct in6_addr *daddr,
|
|
|
|
const u16 hnum,
|
|
|
|
const int dif)
|
|
|
|
{
|
|
|
|
struct sock *sk;
|
|
|
|
const struct hlist_node *node;
|
2006-09-28 01:43:07 +00:00
|
|
|
const __portpair ports = INET_COMBINED_PORTS(sport, hnum);
|
2006-04-10 05:48:59 +00:00
|
|
|
/* Optimize here for direct hit, only listening connections can
|
|
|
|
* have wildcards anyways.
|
|
|
|
*/
|
|
|
|
unsigned int hash = inet6_ehashfn(daddr, hnum, saddr, sport);
|
|
|
|
struct inet_ehash_bucket *head = inet_ehash_bucket(hashinfo, hash);
|
|
|
|
|
|
|
|
prefetch(head->chain.first);
|
|
|
|
read_lock(&head->lock);
|
|
|
|
sk_for_each(sk, node, &head->chain) {
|
|
|
|
/* For IPV6 do the cheaper port and family tests first. */
|
|
|
|
if (INET6_MATCH(sk, hash, saddr, daddr, ports, dif))
|
|
|
|
goto hit; /* You sunk my battleship! */
|
|
|
|
}
|
|
|
|
/* Must check for a TIME_WAIT'er before going to listener hash. */
|
[NET]: change layout of ehash table
ehash table layout is currently this one :
First half of this table is used by sockets not in TIME_WAIT state
Second half of it is used by sockets in TIME_WAIT state.
This is non optimal because of for a given hash or socket, the two chain heads
are located in separate cache lines.
Moreover the locks of the second half are never used.
If instead of this halving, we use two list heads in inet_ehash_bucket instead
of only one, we probably can avoid one cache miss, and reduce ram usage,
particularly if sizeof(rwlock_t) is big (various CONFIG_DEBUG_SPINLOCK,
CONFIG_DEBUG_LOCK_ALLOC settings). So we still halves the table but we keep
together related chains to speedup lookups and socket state change.
In this patch I did not try to align struct inet_ehash_bucket, but a future
patch could try to make this structure have a convenient size (a power of two
or a multiple of L1_CACHE_SIZE).
I guess rwlock will just vanish as soon as RCU is plugged into ehash :) , so
maybe we dont need to scratch our heads to align the bucket...
Note : In case struct inet_ehash_bucket is not a power of two, we could
probably change alloc_large_system_hash() (in case it use __get_free_pages())
to free the unused space. It currently allocates a big zone, but the last
quarter of it could be freed. Again, this should be a temporary 'problem'.
Patch tested on ipv4 tcp only, but should be OK for IPV6 and DCCP.
Signed-off-by: Eric Dumazet <dada1@cosmosbay.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2007-02-08 22:16:46 +00:00
|
|
|
sk_for_each(sk, node, &head->twchain) {
|
2006-04-10 05:48:59 +00:00
|
|
|
const struct inet_timewait_sock *tw = inet_twsk(sk);
|
|
|
|
|
2006-09-28 01:43:07 +00:00
|
|
|
if(*((__portpair *)&(tw->tw_dport)) == ports &&
|
2006-04-10 05:48:59 +00:00
|
|
|
sk->sk_family == PF_INET6) {
|
|
|
|
const struct inet6_timewait_sock *tw6 = inet6_twsk(sk);
|
|
|
|
|
|
|
|
if (ipv6_addr_equal(&tw6->tw_v6_daddr, saddr) &&
|
|
|
|
ipv6_addr_equal(&tw6->tw_v6_rcv_saddr, daddr) &&
|
|
|
|
(!sk->sk_bound_dev_if || sk->sk_bound_dev_if == dif))
|
|
|
|
goto hit;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
read_unlock(&head->lock);
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
hit:
|
|
|
|
sock_hold(sk);
|
|
|
|
read_unlock(&head->lock);
|
|
|
|
return sk;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(__inet6_lookup_established);
|
|
|
|
|
2005-08-12 12:26:18 +00:00
|
|
|
struct sock *inet6_lookup_listener(struct inet_hashinfo *hashinfo,
|
|
|
|
const struct in6_addr *daddr,
|
|
|
|
const unsigned short hnum, const int dif)
|
|
|
|
{
|
|
|
|
struct sock *sk;
|
|
|
|
const struct hlist_node *node;
|
|
|
|
struct sock *result = NULL;
|
|
|
|
int score, hiscore = 0;
|
|
|
|
|
|
|
|
read_lock(&hashinfo->lhash_lock);
|
|
|
|
sk_for_each(sk, node, &hashinfo->listening_hash[inet_lhashfn(hnum)]) {
|
|
|
|
if (inet_sk(sk)->num == hnum && sk->sk_family == PF_INET6) {
|
|
|
|
const struct ipv6_pinfo *np = inet6_sk(sk);
|
2007-02-09 14:24:49 +00:00
|
|
|
|
2005-08-12 12:26:18 +00:00
|
|
|
score = 1;
|
|
|
|
if (!ipv6_addr_any(&np->rcv_saddr)) {
|
|
|
|
if (!ipv6_addr_equal(&np->rcv_saddr, daddr))
|
|
|
|
continue;
|
|
|
|
score++;
|
|
|
|
}
|
|
|
|
if (sk->sk_bound_dev_if) {
|
|
|
|
if (sk->sk_bound_dev_if != dif)
|
|
|
|
continue;
|
|
|
|
score++;
|
|
|
|
}
|
|
|
|
if (score == 3) {
|
|
|
|
result = sk;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (score > hiscore) {
|
|
|
|
hiscore = score;
|
|
|
|
result = sk;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
if (result)
|
|
|
|
sock_hold(result);
|
|
|
|
read_unlock(&hashinfo->lhash_lock);
|
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL_GPL(inet6_lookup_listener);
|
|
|
|
|
|
|
|
struct sock *inet6_lookup(struct inet_hashinfo *hashinfo,
|
2006-11-08 08:20:00 +00:00
|
|
|
const struct in6_addr *saddr, const __be16 sport,
|
|
|
|
const struct in6_addr *daddr, const __be16 dport,
|
2005-08-12 12:26:18 +00:00
|
|
|
const int dif)
|
|
|
|
{
|
|
|
|
struct sock *sk;
|
|
|
|
|
|
|
|
local_bh_disable();
|
|
|
|
sk = __inet6_lookup(hashinfo, saddr, sport, daddr, ntohs(dport), dif);
|
|
|
|
local_bh_enable();
|
|
|
|
|
|
|
|
return sk;
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL_GPL(inet6_lookup);
|
2005-12-14 07:25:44 +00:00
|
|
|
|
|
|
|
static int __inet6_check_established(struct inet_timewait_death_row *death_row,
|
|
|
|
struct sock *sk, const __u16 lport,
|
|
|
|
struct inet_timewait_sock **twp)
|
|
|
|
{
|
|
|
|
struct inet_hashinfo *hinfo = death_row->hashinfo;
|
2006-03-13 22:26:12 +00:00
|
|
|
struct inet_sock *inet = inet_sk(sk);
|
2005-12-14 07:25:44 +00:00
|
|
|
const struct ipv6_pinfo *np = inet6_sk(sk);
|
|
|
|
const struct in6_addr *daddr = &np->rcv_saddr;
|
|
|
|
const struct in6_addr *saddr = &np->daddr;
|
|
|
|
const int dif = sk->sk_bound_dev_if;
|
2006-09-28 01:43:07 +00:00
|
|
|
const __portpair ports = INET_COMBINED_PORTS(inet->dport, lport);
|
2007-02-13 04:26:39 +00:00
|
|
|
const unsigned int hash = inet6_ehashfn(daddr, lport, saddr,
|
2005-12-14 07:25:44 +00:00
|
|
|
inet->dport);
|
|
|
|
struct inet_ehash_bucket *head = inet_ehash_bucket(hinfo, hash);
|
|
|
|
struct sock *sk2;
|
|
|
|
const struct hlist_node *node;
|
|
|
|
struct inet_timewait_sock *tw;
|
|
|
|
|
|
|
|
prefetch(head->chain.first);
|
|
|
|
write_lock(&head->lock);
|
|
|
|
|
|
|
|
/* Check TIME-WAIT sockets first. */
|
[NET]: change layout of ehash table
ehash table layout is currently this one :
First half of this table is used by sockets not in TIME_WAIT state
Second half of it is used by sockets in TIME_WAIT state.
This is non optimal because of for a given hash or socket, the two chain heads
are located in separate cache lines.
Moreover the locks of the second half are never used.
If instead of this halving, we use two list heads in inet_ehash_bucket instead
of only one, we probably can avoid one cache miss, and reduce ram usage,
particularly if sizeof(rwlock_t) is big (various CONFIG_DEBUG_SPINLOCK,
CONFIG_DEBUG_LOCK_ALLOC settings). So we still halves the table but we keep
together related chains to speedup lookups and socket state change.
In this patch I did not try to align struct inet_ehash_bucket, but a future
patch could try to make this structure have a convenient size (a power of two
or a multiple of L1_CACHE_SIZE).
I guess rwlock will just vanish as soon as RCU is plugged into ehash :) , so
maybe we dont need to scratch our heads to align the bucket...
Note : In case struct inet_ehash_bucket is not a power of two, we could
probably change alloc_large_system_hash() (in case it use __get_free_pages())
to free the unused space. It currently allocates a big zone, but the last
quarter of it could be freed. Again, this should be a temporary 'problem'.
Patch tested on ipv4 tcp only, but should be OK for IPV6 and DCCP.
Signed-off-by: Eric Dumazet <dada1@cosmosbay.com>
Signed-off-by: David S. Miller <davem@davemloft.net>
2007-02-08 22:16:46 +00:00
|
|
|
sk_for_each(sk2, node, &head->twchain) {
|
2005-12-14 07:25:44 +00:00
|
|
|
const struct inet6_timewait_sock *tw6 = inet6_twsk(sk2);
|
|
|
|
|
|
|
|
tw = inet_twsk(sk2);
|
|
|
|
|
2006-09-28 01:43:07 +00:00
|
|
|
if(*((__portpair *)&(tw->tw_dport)) == ports &&
|
2005-12-14 07:25:44 +00:00
|
|
|
sk2->sk_family == PF_INET6 &&
|
|
|
|
ipv6_addr_equal(&tw6->tw_v6_daddr, saddr) &&
|
|
|
|
ipv6_addr_equal(&tw6->tw_v6_rcv_saddr, daddr) &&
|
|
|
|
sk2->sk_bound_dev_if == sk->sk_bound_dev_if) {
|
|
|
|
if (twsk_unique(sk, sk2, twp))
|
|
|
|
goto unique;
|
|
|
|
else
|
|
|
|
goto not_unique;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
tw = NULL;
|
|
|
|
|
|
|
|
/* And established part... */
|
|
|
|
sk_for_each(sk2, node, &head->chain) {
|
|
|
|
if (INET6_MATCH(sk2, hash, saddr, daddr, ports, dif))
|
|
|
|
goto not_unique;
|
|
|
|
}
|
|
|
|
|
|
|
|
unique:
|
2006-03-13 22:26:12 +00:00
|
|
|
/* Must record num and sport now. Otherwise we will see
|
|
|
|
* in hash table socket with a funny identity. */
|
|
|
|
inet->num = lport;
|
|
|
|
inet->sport = htons(lport);
|
2005-12-14 07:25:44 +00:00
|
|
|
BUG_TRAP(sk_unhashed(sk));
|
|
|
|
__sk_add_node(sk, &head->chain);
|
|
|
|
sk->sk_hash = hash;
|
|
|
|
sock_prot_inc_use(sk->sk_prot);
|
|
|
|
write_unlock(&head->lock);
|
|
|
|
|
|
|
|
if (twp != NULL) {
|
|
|
|
*twp = tw;
|
|
|
|
NET_INC_STATS_BH(LINUX_MIB_TIMEWAITRECYCLED);
|
|
|
|
} else if (tw != NULL) {
|
|
|
|
/* Silly. Should hash-dance instead... */
|
|
|
|
inet_twsk_deschedule(tw, death_row);
|
|
|
|
NET_INC_STATS_BH(LINUX_MIB_TIMEWAITRECYCLED);
|
|
|
|
|
|
|
|
inet_twsk_put(tw);
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
not_unique:
|
|
|
|
write_unlock(&head->lock);
|
|
|
|
return -EADDRNOTAVAIL;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32 inet6_sk_port_offset(const struct sock *sk)
|
|
|
|
{
|
|
|
|
const struct inet_sock *inet = inet_sk(sk);
|
|
|
|
const struct ipv6_pinfo *np = inet6_sk(sk);
|
|
|
|
return secure_ipv6_port_ephemeral(np->rcv_saddr.s6_addr32,
|
|
|
|
np->daddr.s6_addr32,
|
|
|
|
inet->dport);
|
|
|
|
}
|
|
|
|
|
|
|
|
int inet6_hash_connect(struct inet_timewait_death_row *death_row,
|
|
|
|
struct sock *sk)
|
|
|
|
{
|
|
|
|
struct inet_hashinfo *hinfo = death_row->hashinfo;
|
|
|
|
const unsigned short snum = inet_sk(sk)->num;
|
2007-02-09 14:24:49 +00:00
|
|
|
struct inet_bind_hashbucket *head;
|
|
|
|
struct inet_bind_bucket *tb;
|
2005-12-14 07:25:44 +00:00
|
|
|
int ret;
|
|
|
|
|
2007-02-09 14:24:49 +00:00
|
|
|
if (snum == 0) {
|
|
|
|
const int low = sysctl_local_port_range[0];
|
|
|
|
const int high = sysctl_local_port_range[1];
|
2005-12-14 07:25:44 +00:00
|
|
|
const int range = high - low;
|
2007-02-09 14:24:49 +00:00
|
|
|
int i, port;
|
2005-12-14 07:25:44 +00:00
|
|
|
static u32 hint;
|
|
|
|
const u32 offset = hint + inet6_sk_port_offset(sk);
|
|
|
|
struct hlist_node *node;
|
2007-02-09 14:24:49 +00:00
|
|
|
struct inet_timewait_sock *tw = NULL;
|
2005-12-14 07:25:44 +00:00
|
|
|
|
2007-02-09 14:24:49 +00:00
|
|
|
local_bh_disable();
|
2005-12-14 07:25:44 +00:00
|
|
|
for (i = 1; i <= range; i++) {
|
|
|
|
port = low + (i + offset) % range;
|
2007-02-09 14:24:49 +00:00
|
|
|
head = &hinfo->bhash[inet_bhashfn(port, hinfo->bhash_size)];
|
|
|
|
spin_lock(&head->lock);
|
2005-12-14 07:25:44 +00:00
|
|
|
|
2007-02-09 14:24:49 +00:00
|
|
|
/* Does not bother with rcv_saddr checks,
|
|
|
|
* because the established check is already
|
|
|
|
* unique enough.
|
|
|
|
*/
|
2005-12-14 07:25:44 +00:00
|
|
|
inet_bind_bucket_for_each(tb, node, &head->chain) {
|
2007-02-09 14:24:49 +00:00
|
|
|
if (tb->port == port) {
|
|
|
|
BUG_TRAP(!hlist_empty(&tb->owners));
|
|
|
|
if (tb->fastreuse >= 0)
|
|
|
|
goto next_port;
|
|
|
|
if (!__inet6_check_established(death_row,
|
2005-12-14 07:25:44 +00:00
|
|
|
sk, port,
|
|
|
|
&tw))
|
2007-02-09 14:24:49 +00:00
|
|
|
goto ok;
|
|
|
|
goto next_port;
|
|
|
|
}
|
|
|
|
}
|
2005-12-14 07:25:44 +00:00
|
|
|
|
2007-02-09 14:24:49 +00:00
|
|
|
tb = inet_bind_bucket_create(hinfo->bind_bucket_cachep,
|
2005-12-14 07:25:44 +00:00
|
|
|
head, port);
|
2007-02-09 14:24:49 +00:00
|
|
|
if (!tb) {
|
|
|
|
spin_unlock(&head->lock);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
tb->fastreuse = -1;
|
|
|
|
goto ok;
|
2005-12-14 07:25:44 +00:00
|
|
|
|
2007-02-09 14:24:49 +00:00
|
|
|
next_port:
|
|
|
|
spin_unlock(&head->lock);
|
|
|
|
}
|
|
|
|
local_bh_enable();
|
2005-12-14 07:25:44 +00:00
|
|
|
|
2007-02-09 14:24:49 +00:00
|
|
|
return -EADDRNOTAVAIL;
|
2005-12-14 07:25:44 +00:00
|
|
|
|
|
|
|
ok:
|
|
|
|
hint += i;
|
|
|
|
|
2007-02-09 14:24:49 +00:00
|
|
|
/* Head lock still held and bh's disabled */
|
|
|
|
inet_bind_hash(sk, tb, port);
|
2005-12-14 07:25:44 +00:00
|
|
|
if (sk_unhashed(sk)) {
|
2007-02-09 14:24:49 +00:00
|
|
|
inet_sk(sk)->sport = htons(port);
|
|
|
|
__inet6_hash(hinfo, sk);
|
|
|
|
}
|
|
|
|
spin_unlock(&head->lock);
|
2005-12-14 07:25:44 +00:00
|
|
|
|
2007-02-09 14:24:49 +00:00
|
|
|
if (tw) {
|
|
|
|
inet_twsk_deschedule(tw, death_row);
|
|
|
|
inet_twsk_put(tw);
|
|
|
|
}
|
2005-12-14 07:25:44 +00:00
|
|
|
|
|
|
|
ret = 0;
|
|
|
|
goto out;
|
2007-02-09 14:24:49 +00:00
|
|
|
}
|
2005-12-14 07:25:44 +00:00
|
|
|
|
2007-02-09 14:24:49 +00:00
|
|
|
head = &hinfo->bhash[inet_bhashfn(snum, hinfo->bhash_size)];
|
|
|
|
tb = inet_csk(sk)->icsk_bind_hash;
|
2005-12-14 07:25:44 +00:00
|
|
|
spin_lock_bh(&head->lock);
|
|
|
|
|
|
|
|
if (sk_head(&tb->owners) == sk && sk->sk_bind_node.next == NULL) {
|
|
|
|
__inet6_hash(hinfo, sk);
|
|
|
|
spin_unlock_bh(&head->lock);
|
|
|
|
return 0;
|
|
|
|
} else {
|
|
|
|
spin_unlock(&head->lock);
|
|
|
|
/* No definite answer... Walk to established hash table */
|
|
|
|
ret = __inet6_check_established(death_row, sk, snum, NULL);
|
|
|
|
out:
|
|
|
|
local_bh_enable();
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL_GPL(inet6_hash_connect);
|