mirror of
https://github.com/torvalds/linux.git
synced 2024-11-22 12:11:40 +00:00
9388637270
<tldr> skb network header of the single-tagged vlan packet continues to point the vlan payload (e.g. IP) after second vlan tag is pushed by tc act_vlan. This causes problem at the dissector which expects double-tagged packet network header to point to the inner vlan. The fix is to adjust network header in tcf_act_vlan.c but requires refactoring of skb_vlan_push function. </tldr> Consider the following shell script snippet configuring TC rules on the veth interface: ip link add veth0 type veth peer veth1 ip link set veth0 up ip link set veth1 up tc qdisc add dev veth0 clsact tc filter add dev veth0 ingress pref 10 chain 0 flower \ num_of_vlans 2 cvlan_ethtype 0x800 action goto chain 5 tc filter add dev veth0 ingress pref 20 chain 0 flower \ num_of_vlans 1 action vlan push id 100 \ protocol 0x8100 action goto chain 5 tc filter add dev veth0 ingress pref 30 chain 5 flower \ num_of_vlans 2 cvlan_ethtype 0x800 action simple sdata "success" Sending double-tagged vlan packet with the IP payload inside: cat <<ENDS | text2pcap - - | tcpreplay -i veth1 - 0000 00 00 00 00 00 11 00 00 00 00 00 22 81 00 00 64 ..........."...d 0010 81 00 00 14 08 00 45 04 00 26 04 d2 00 00 7f 11 ......E..&...... 0020 18 ef 0a 00 00 01 14 00 00 02 00 00 00 00 00 12 ................ 0030 e1 c7 00 00 00 00 00 00 00 00 00 00 ............ ENDS will match rule 10, goto rule 30 in chain 5 and correctly emit "success" to the dmesg. OTOH, sending single-tagged vlan packet: cat <<ENDS | text2pcap - - | tcpreplay -i veth1 - 0000 00 00 00 00 00 11 00 00 00 00 00 22 81 00 00 14 ...........".... 0010 08 00 45 04 00 2a 04 d2 00 00 7f 11 18 eb 0a 00 ..E..*.......... 0020 00 01 14 00 00 02 00 00 00 00 00 16 e1 bf 00 00 ................ 0030 00 00 00 00 00 00 00 00 00 00 00 00 ............ ENDS will match rule 20, will push the second vlan tag but will *not* match rule 30. IOW, the match at rule 30 fails if the second vlan was freshly pushed by the kernel. Lets look at __skb_flow_dissect working on the double-tagged vlan packet. Here is the relevant code from around net/core/flow_dissector.c:1277 copy-pasted here for convenience: if (dissector_vlan == FLOW_DISSECTOR_KEY_MAX && skb && skb_vlan_tag_present(skb)) { proto = skb->protocol; } else { vlan = __skb_header_pointer(skb, nhoff, sizeof(_vlan), data, hlen, &_vlan); if (!vlan) { fdret = FLOW_DISSECT_RET_OUT_BAD; break; } proto = vlan->h_vlan_encapsulated_proto; nhoff += sizeof(*vlan); } The "else" clause above gets the protocol of the encapsulated packet from the skb data at the network header location. printk debugging has showed that in the good double-tagged packet case proto is htons(0x800 == ETH_P_IP) as expected. However in the single-tagged packet case proto is garbage leading to the failure to match tc filter 30. proto is being set from the skb header pointed by nhoff parameter which is defined at the beginning of __skb_flow_dissect (net/core/flow_dissector.c:1055 in the current version): nhoff = skb_network_offset(skb); Therefore the culprit seems to be that the skb network offset is different between double-tagged packet received from the interface and single-tagged packet having its vlan tag pushed by TC. Lets look at the interesting points of the lifetime of the single/double tagged packets as they traverse our packet flow. Both of them will start at __netif_receive_skb_core where the first vlan tag will be stripped: if (eth_type_vlan(skb->protocol)) { skb = skb_vlan_untag(skb); if (unlikely(!skb)) goto out; } At this stage in double-tagged case skb->data points to the second vlan tag while in single-tagged case skb->data points to the network (eg. IP) header. Looking at TC vlan push action (net/sched/act_vlan.c) we have the following code at tcf_vlan_act (interesting points are in square brackets): if (skb_at_tc_ingress(skb)) [1] skb_push_rcsum(skb, skb->mac_len); .... case TCA_VLAN_ACT_PUSH: err = skb_vlan_push(skb, p->tcfv_push_proto, p->tcfv_push_vid | (p->tcfv_push_prio << VLAN_PRIO_SHIFT), 0); if (err) goto drop; break; .... out: if (skb_at_tc_ingress(skb)) [3] skb_pull_rcsum(skb, skb->mac_len); And skb_vlan_push (net/core/skbuff.c:6204) function does: err = __vlan_insert_tag(skb, skb->vlan_proto, skb_vlan_tag_get(skb)); if (err) return err; skb->protocol = skb->vlan_proto; [2] skb->mac_len += VLAN_HLEN; in the case of pushing the second tag. Lets look at what happens with skb->data of the single-tagged packet at each of the above points: 1. As a result of the skb_push_rcsum, skb->data is moved back to the start of the packet. 2. First VLAN tag is moved from the skb into packet buffer, skb->mac_len is incremented, skb->data still points to the start of the packet. 3. As a result of the skb_pull_rcsum, skb->data is moved forward by the modified skb->mac_len, thus pointing to the network header again. Then __skb_flow_dissect will get confused by having double-tagged vlan packet with the skb->data at the network header. The solution for the bug is to preserve "skb->data at second vlan header" semantics in the skb_vlan_push function. We do this by manipulating skb->network_header rather than skb->mac_len. skb_vlan_push callers are updated to do skb_reset_mac_len. Signed-off-by: Boris Sukholitko <boris.sukholitko@broadcom.com> Signed-off-by: Paolo Abeni <pabeni@redhat.com>
468 lines
12 KiB
C
468 lines
12 KiB
C
// SPDX-License-Identifier: GPL-2.0-or-later
|
|
/*
|
|
* Copyright (c) 2014 Jiri Pirko <jiri@resnulli.us>
|
|
*/
|
|
|
|
#include <linux/module.h>
|
|
#include <linux/init.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/skbuff.h>
|
|
#include <linux/rtnetlink.h>
|
|
#include <linux/if_vlan.h>
|
|
#include <net/netlink.h>
|
|
#include <net/pkt_sched.h>
|
|
#include <net/pkt_cls.h>
|
|
#include <net/tc_wrapper.h>
|
|
|
|
#include <linux/tc_act/tc_vlan.h>
|
|
#include <net/tc_act/tc_vlan.h>
|
|
|
|
static struct tc_action_ops act_vlan_ops;
|
|
|
|
TC_INDIRECT_SCOPE int tcf_vlan_act(struct sk_buff *skb,
|
|
const struct tc_action *a,
|
|
struct tcf_result *res)
|
|
{
|
|
struct tcf_vlan *v = to_vlan(a);
|
|
struct tcf_vlan_params *p;
|
|
int action;
|
|
int err;
|
|
u16 tci;
|
|
|
|
tcf_lastuse_update(&v->tcf_tm);
|
|
tcf_action_update_bstats(&v->common, skb);
|
|
|
|
/* Ensure 'data' points at mac_header prior calling vlan manipulating
|
|
* functions.
|
|
*/
|
|
if (skb_at_tc_ingress(skb))
|
|
skb_push_rcsum(skb, skb->mac_len);
|
|
|
|
action = READ_ONCE(v->tcf_action);
|
|
|
|
p = rcu_dereference_bh(v->vlan_p);
|
|
|
|
switch (p->tcfv_action) {
|
|
case TCA_VLAN_ACT_POP:
|
|
err = skb_vlan_pop(skb);
|
|
if (err)
|
|
goto drop;
|
|
break;
|
|
case TCA_VLAN_ACT_PUSH:
|
|
err = skb_vlan_push(skb, p->tcfv_push_proto, p->tcfv_push_vid |
|
|
(p->tcfv_push_prio << VLAN_PRIO_SHIFT));
|
|
if (err)
|
|
goto drop;
|
|
break;
|
|
case TCA_VLAN_ACT_MODIFY:
|
|
/* No-op if no vlan tag (either hw-accel or in-payload) */
|
|
if (!skb_vlan_tagged(skb))
|
|
goto out;
|
|
/* extract existing tag (and guarantee no hw-accel tag) */
|
|
if (skb_vlan_tag_present(skb)) {
|
|
tci = skb_vlan_tag_get(skb);
|
|
__vlan_hwaccel_clear_tag(skb);
|
|
} else {
|
|
/* in-payload vlan tag, pop it */
|
|
err = __skb_vlan_pop(skb, &tci);
|
|
if (err)
|
|
goto drop;
|
|
}
|
|
/* replace the vid */
|
|
tci = (tci & ~VLAN_VID_MASK) | p->tcfv_push_vid;
|
|
/* replace prio bits, if tcfv_push_prio specified */
|
|
if (p->tcfv_push_prio_exists) {
|
|
tci &= ~VLAN_PRIO_MASK;
|
|
tci |= p->tcfv_push_prio << VLAN_PRIO_SHIFT;
|
|
}
|
|
/* put updated tci as hwaccel tag */
|
|
__vlan_hwaccel_put_tag(skb, p->tcfv_push_proto, tci);
|
|
break;
|
|
case TCA_VLAN_ACT_POP_ETH:
|
|
err = skb_eth_pop(skb);
|
|
if (err)
|
|
goto drop;
|
|
break;
|
|
case TCA_VLAN_ACT_PUSH_ETH:
|
|
err = skb_eth_push(skb, p->tcfv_push_dst, p->tcfv_push_src);
|
|
if (err)
|
|
goto drop;
|
|
break;
|
|
default:
|
|
BUG();
|
|
}
|
|
|
|
out:
|
|
if (skb_at_tc_ingress(skb))
|
|
skb_pull_rcsum(skb, skb->mac_len);
|
|
|
|
skb_reset_mac_len(skb);
|
|
return action;
|
|
|
|
drop:
|
|
tcf_action_inc_drop_qstats(&v->common);
|
|
return TC_ACT_SHOT;
|
|
}
|
|
|
|
static const struct nla_policy vlan_policy[TCA_VLAN_MAX + 1] = {
|
|
[TCA_VLAN_UNSPEC] = { .strict_start_type = TCA_VLAN_PUSH_ETH_DST },
|
|
[TCA_VLAN_PARMS] = { .len = sizeof(struct tc_vlan) },
|
|
[TCA_VLAN_PUSH_VLAN_ID] = { .type = NLA_U16 },
|
|
[TCA_VLAN_PUSH_VLAN_PROTOCOL] = { .type = NLA_U16 },
|
|
[TCA_VLAN_PUSH_VLAN_PRIORITY] = { .type = NLA_U8 },
|
|
[TCA_VLAN_PUSH_ETH_DST] = NLA_POLICY_ETH_ADDR,
|
|
[TCA_VLAN_PUSH_ETH_SRC] = NLA_POLICY_ETH_ADDR,
|
|
};
|
|
|
|
static int tcf_vlan_init(struct net *net, struct nlattr *nla,
|
|
struct nlattr *est, struct tc_action **a,
|
|
struct tcf_proto *tp, u32 flags,
|
|
struct netlink_ext_ack *extack)
|
|
{
|
|
struct tc_action_net *tn = net_generic(net, act_vlan_ops.net_id);
|
|
bool bind = flags & TCA_ACT_FLAGS_BIND;
|
|
struct nlattr *tb[TCA_VLAN_MAX + 1];
|
|
struct tcf_chain *goto_ch = NULL;
|
|
bool push_prio_exists = false;
|
|
struct tcf_vlan_params *p;
|
|
struct tc_vlan *parm;
|
|
struct tcf_vlan *v;
|
|
int action;
|
|
u16 push_vid = 0;
|
|
__be16 push_proto = 0;
|
|
u8 push_prio = 0;
|
|
bool exists = false;
|
|
int ret = 0, err;
|
|
u32 index;
|
|
|
|
if (!nla)
|
|
return -EINVAL;
|
|
|
|
err = nla_parse_nested_deprecated(tb, TCA_VLAN_MAX, nla, vlan_policy,
|
|
NULL);
|
|
if (err < 0)
|
|
return err;
|
|
|
|
if (!tb[TCA_VLAN_PARMS])
|
|
return -EINVAL;
|
|
parm = nla_data(tb[TCA_VLAN_PARMS]);
|
|
index = parm->index;
|
|
err = tcf_idr_check_alloc(tn, &index, a, bind);
|
|
if (err < 0)
|
|
return err;
|
|
exists = err;
|
|
if (exists && bind)
|
|
return ACT_P_BOUND;
|
|
|
|
switch (parm->v_action) {
|
|
case TCA_VLAN_ACT_POP:
|
|
break;
|
|
case TCA_VLAN_ACT_PUSH:
|
|
case TCA_VLAN_ACT_MODIFY:
|
|
if (!tb[TCA_VLAN_PUSH_VLAN_ID]) {
|
|
if (exists)
|
|
tcf_idr_release(*a, bind);
|
|
else
|
|
tcf_idr_cleanup(tn, index);
|
|
return -EINVAL;
|
|
}
|
|
push_vid = nla_get_u16(tb[TCA_VLAN_PUSH_VLAN_ID]);
|
|
if (push_vid >= VLAN_VID_MASK) {
|
|
if (exists)
|
|
tcf_idr_release(*a, bind);
|
|
else
|
|
tcf_idr_cleanup(tn, index);
|
|
return -ERANGE;
|
|
}
|
|
|
|
if (tb[TCA_VLAN_PUSH_VLAN_PROTOCOL]) {
|
|
push_proto = nla_get_be16(tb[TCA_VLAN_PUSH_VLAN_PROTOCOL]);
|
|
switch (push_proto) {
|
|
case htons(ETH_P_8021Q):
|
|
case htons(ETH_P_8021AD):
|
|
break;
|
|
default:
|
|
if (exists)
|
|
tcf_idr_release(*a, bind);
|
|
else
|
|
tcf_idr_cleanup(tn, index);
|
|
return -EPROTONOSUPPORT;
|
|
}
|
|
} else {
|
|
push_proto = htons(ETH_P_8021Q);
|
|
}
|
|
|
|
push_prio_exists = !!tb[TCA_VLAN_PUSH_VLAN_PRIORITY];
|
|
if (push_prio_exists)
|
|
push_prio = nla_get_u8(tb[TCA_VLAN_PUSH_VLAN_PRIORITY]);
|
|
break;
|
|
case TCA_VLAN_ACT_POP_ETH:
|
|
break;
|
|
case TCA_VLAN_ACT_PUSH_ETH:
|
|
if (!tb[TCA_VLAN_PUSH_ETH_DST] || !tb[TCA_VLAN_PUSH_ETH_SRC]) {
|
|
if (exists)
|
|
tcf_idr_release(*a, bind);
|
|
else
|
|
tcf_idr_cleanup(tn, index);
|
|
return -EINVAL;
|
|
}
|
|
break;
|
|
default:
|
|
if (exists)
|
|
tcf_idr_release(*a, bind);
|
|
else
|
|
tcf_idr_cleanup(tn, index);
|
|
return -EINVAL;
|
|
}
|
|
action = parm->v_action;
|
|
|
|
if (!exists) {
|
|
ret = tcf_idr_create_from_flags(tn, index, est, a,
|
|
&act_vlan_ops, bind, flags);
|
|
if (ret) {
|
|
tcf_idr_cleanup(tn, index);
|
|
return ret;
|
|
}
|
|
|
|
ret = ACT_P_CREATED;
|
|
} else if (!(flags & TCA_ACT_FLAGS_REPLACE)) {
|
|
tcf_idr_release(*a, bind);
|
|
return -EEXIST;
|
|
}
|
|
|
|
err = tcf_action_check_ctrlact(parm->action, tp, &goto_ch, extack);
|
|
if (err < 0)
|
|
goto release_idr;
|
|
|
|
v = to_vlan(*a);
|
|
|
|
p = kzalloc(sizeof(*p), GFP_KERNEL);
|
|
if (!p) {
|
|
err = -ENOMEM;
|
|
goto put_chain;
|
|
}
|
|
|
|
p->tcfv_action = action;
|
|
p->tcfv_push_vid = push_vid;
|
|
p->tcfv_push_prio = push_prio;
|
|
p->tcfv_push_prio_exists = push_prio_exists || action == TCA_VLAN_ACT_PUSH;
|
|
p->tcfv_push_proto = push_proto;
|
|
|
|
if (action == TCA_VLAN_ACT_PUSH_ETH) {
|
|
nla_memcpy(&p->tcfv_push_dst, tb[TCA_VLAN_PUSH_ETH_DST],
|
|
ETH_ALEN);
|
|
nla_memcpy(&p->tcfv_push_src, tb[TCA_VLAN_PUSH_ETH_SRC],
|
|
ETH_ALEN);
|
|
}
|
|
|
|
spin_lock_bh(&v->tcf_lock);
|
|
goto_ch = tcf_action_set_ctrlact(*a, parm->action, goto_ch);
|
|
p = rcu_replace_pointer(v->vlan_p, p, lockdep_is_held(&v->tcf_lock));
|
|
spin_unlock_bh(&v->tcf_lock);
|
|
|
|
if (goto_ch)
|
|
tcf_chain_put_by_act(goto_ch);
|
|
if (p)
|
|
kfree_rcu(p, rcu);
|
|
|
|
return ret;
|
|
put_chain:
|
|
if (goto_ch)
|
|
tcf_chain_put_by_act(goto_ch);
|
|
release_idr:
|
|
tcf_idr_release(*a, bind);
|
|
return err;
|
|
}
|
|
|
|
static void tcf_vlan_cleanup(struct tc_action *a)
|
|
{
|
|
struct tcf_vlan *v = to_vlan(a);
|
|
struct tcf_vlan_params *p;
|
|
|
|
p = rcu_dereference_protected(v->vlan_p, 1);
|
|
if (p)
|
|
kfree_rcu(p, rcu);
|
|
}
|
|
|
|
static int tcf_vlan_dump(struct sk_buff *skb, struct tc_action *a,
|
|
int bind, int ref)
|
|
{
|
|
unsigned char *b = skb_tail_pointer(skb);
|
|
struct tcf_vlan *v = to_vlan(a);
|
|
struct tcf_vlan_params *p;
|
|
struct tc_vlan opt = {
|
|
.index = v->tcf_index,
|
|
.refcnt = refcount_read(&v->tcf_refcnt) - ref,
|
|
.bindcnt = atomic_read(&v->tcf_bindcnt) - bind,
|
|
};
|
|
struct tcf_t t;
|
|
|
|
spin_lock_bh(&v->tcf_lock);
|
|
opt.action = v->tcf_action;
|
|
p = rcu_dereference_protected(v->vlan_p, lockdep_is_held(&v->tcf_lock));
|
|
opt.v_action = p->tcfv_action;
|
|
if (nla_put(skb, TCA_VLAN_PARMS, sizeof(opt), &opt))
|
|
goto nla_put_failure;
|
|
|
|
if ((p->tcfv_action == TCA_VLAN_ACT_PUSH ||
|
|
p->tcfv_action == TCA_VLAN_ACT_MODIFY) &&
|
|
(nla_put_u16(skb, TCA_VLAN_PUSH_VLAN_ID, p->tcfv_push_vid) ||
|
|
nla_put_be16(skb, TCA_VLAN_PUSH_VLAN_PROTOCOL,
|
|
p->tcfv_push_proto) ||
|
|
(p->tcfv_push_prio_exists &&
|
|
nla_put_u8(skb, TCA_VLAN_PUSH_VLAN_PRIORITY, p->tcfv_push_prio))))
|
|
goto nla_put_failure;
|
|
|
|
if (p->tcfv_action == TCA_VLAN_ACT_PUSH_ETH) {
|
|
if (nla_put(skb, TCA_VLAN_PUSH_ETH_DST, ETH_ALEN,
|
|
p->tcfv_push_dst))
|
|
goto nla_put_failure;
|
|
if (nla_put(skb, TCA_VLAN_PUSH_ETH_SRC, ETH_ALEN,
|
|
p->tcfv_push_src))
|
|
goto nla_put_failure;
|
|
}
|
|
|
|
tcf_tm_dump(&t, &v->tcf_tm);
|
|
if (nla_put_64bit(skb, TCA_VLAN_TM, sizeof(t), &t, TCA_VLAN_PAD))
|
|
goto nla_put_failure;
|
|
spin_unlock_bh(&v->tcf_lock);
|
|
|
|
return skb->len;
|
|
|
|
nla_put_failure:
|
|
spin_unlock_bh(&v->tcf_lock);
|
|
nlmsg_trim(skb, b);
|
|
return -1;
|
|
}
|
|
|
|
static void tcf_vlan_stats_update(struct tc_action *a, u64 bytes, u64 packets,
|
|
u64 drops, u64 lastuse, bool hw)
|
|
{
|
|
struct tcf_vlan *v = to_vlan(a);
|
|
struct tcf_t *tm = &v->tcf_tm;
|
|
|
|
tcf_action_update_stats(a, bytes, packets, drops, hw);
|
|
tm->lastuse = max_t(u64, tm->lastuse, lastuse);
|
|
}
|
|
|
|
static size_t tcf_vlan_get_fill_size(const struct tc_action *act)
|
|
{
|
|
return nla_total_size(sizeof(struct tc_vlan))
|
|
+ nla_total_size(sizeof(u16)) /* TCA_VLAN_PUSH_VLAN_ID */
|
|
+ nla_total_size(sizeof(u16)) /* TCA_VLAN_PUSH_VLAN_PROTOCOL */
|
|
+ nla_total_size(sizeof(u8)); /* TCA_VLAN_PUSH_VLAN_PRIORITY */
|
|
}
|
|
|
|
static int tcf_vlan_offload_act_setup(struct tc_action *act, void *entry_data,
|
|
u32 *index_inc, bool bind,
|
|
struct netlink_ext_ack *extack)
|
|
{
|
|
if (bind) {
|
|
struct flow_action_entry *entry = entry_data;
|
|
|
|
switch (tcf_vlan_action(act)) {
|
|
case TCA_VLAN_ACT_PUSH:
|
|
entry->id = FLOW_ACTION_VLAN_PUSH;
|
|
entry->vlan.vid = tcf_vlan_push_vid(act);
|
|
entry->vlan.proto = tcf_vlan_push_proto(act);
|
|
entry->vlan.prio = tcf_vlan_push_prio(act);
|
|
break;
|
|
case TCA_VLAN_ACT_POP:
|
|
entry->id = FLOW_ACTION_VLAN_POP;
|
|
break;
|
|
case TCA_VLAN_ACT_MODIFY:
|
|
entry->id = FLOW_ACTION_VLAN_MANGLE;
|
|
entry->vlan.vid = tcf_vlan_push_vid(act);
|
|
entry->vlan.proto = tcf_vlan_push_proto(act);
|
|
entry->vlan.prio = tcf_vlan_push_prio(act);
|
|
break;
|
|
case TCA_VLAN_ACT_POP_ETH:
|
|
entry->id = FLOW_ACTION_VLAN_POP_ETH;
|
|
break;
|
|
case TCA_VLAN_ACT_PUSH_ETH:
|
|
entry->id = FLOW_ACTION_VLAN_PUSH_ETH;
|
|
tcf_vlan_push_eth(entry->vlan_push_eth.src, entry->vlan_push_eth.dst, act);
|
|
break;
|
|
default:
|
|
NL_SET_ERR_MSG_MOD(extack, "Unsupported vlan action mode offload");
|
|
return -EOPNOTSUPP;
|
|
}
|
|
*index_inc = 1;
|
|
} else {
|
|
struct flow_offload_action *fl_action = entry_data;
|
|
|
|
switch (tcf_vlan_action(act)) {
|
|
case TCA_VLAN_ACT_PUSH:
|
|
fl_action->id = FLOW_ACTION_VLAN_PUSH;
|
|
break;
|
|
case TCA_VLAN_ACT_POP:
|
|
fl_action->id = FLOW_ACTION_VLAN_POP;
|
|
break;
|
|
case TCA_VLAN_ACT_MODIFY:
|
|
fl_action->id = FLOW_ACTION_VLAN_MANGLE;
|
|
break;
|
|
case TCA_VLAN_ACT_POP_ETH:
|
|
fl_action->id = FLOW_ACTION_VLAN_POP_ETH;
|
|
break;
|
|
case TCA_VLAN_ACT_PUSH_ETH:
|
|
fl_action->id = FLOW_ACTION_VLAN_PUSH_ETH;
|
|
break;
|
|
default:
|
|
return -EOPNOTSUPP;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static struct tc_action_ops act_vlan_ops = {
|
|
.kind = "vlan",
|
|
.id = TCA_ID_VLAN,
|
|
.owner = THIS_MODULE,
|
|
.act = tcf_vlan_act,
|
|
.dump = tcf_vlan_dump,
|
|
.init = tcf_vlan_init,
|
|
.cleanup = tcf_vlan_cleanup,
|
|
.stats_update = tcf_vlan_stats_update,
|
|
.get_fill_size = tcf_vlan_get_fill_size,
|
|
.offload_act_setup = tcf_vlan_offload_act_setup,
|
|
.size = sizeof(struct tcf_vlan),
|
|
};
|
|
MODULE_ALIAS_NET_ACT("vlan");
|
|
|
|
static __net_init int vlan_init_net(struct net *net)
|
|
{
|
|
struct tc_action_net *tn = net_generic(net, act_vlan_ops.net_id);
|
|
|
|
return tc_action_net_init(net, tn, &act_vlan_ops);
|
|
}
|
|
|
|
static void __net_exit vlan_exit_net(struct list_head *net_list)
|
|
{
|
|
tc_action_net_exit(net_list, act_vlan_ops.net_id);
|
|
}
|
|
|
|
static struct pernet_operations vlan_net_ops = {
|
|
.init = vlan_init_net,
|
|
.exit_batch = vlan_exit_net,
|
|
.id = &act_vlan_ops.net_id,
|
|
.size = sizeof(struct tc_action_net),
|
|
};
|
|
|
|
static int __init vlan_init_module(void)
|
|
{
|
|
return tcf_register_action(&act_vlan_ops, &vlan_net_ops);
|
|
}
|
|
|
|
static void __exit vlan_cleanup_module(void)
|
|
{
|
|
tcf_unregister_action(&act_vlan_ops, &vlan_net_ops);
|
|
}
|
|
|
|
module_init(vlan_init_module);
|
|
module_exit(vlan_cleanup_module);
|
|
|
|
MODULE_AUTHOR("Jiri Pirko <jiri@resnulli.us>");
|
|
MODULE_DESCRIPTION("vlan manipulation actions");
|
|
MODULE_LICENSE("GPL v2");
|