You've already forked linux-apfs
mirror of
https://github.com/linux-apfs/linux-apfs.git
synced 2026-05-01 15:00:59 -07:00
Merge branch 'master' of git://git.kernel.org/pub/scm/linux/kernel/git/kaber/nf-next-2.6
This commit is contained in:
@@ -142,6 +142,12 @@ static inline int ebt_basic_match(struct ebt_entry *e, struct ethhdr *h,
|
||||
return 0;
|
||||
}
|
||||
|
||||
static inline __pure
|
||||
struct ebt_entry *ebt_next_entry(const struct ebt_entry *entry)
|
||||
{
|
||||
return (void *)entry + entry->next_offset;
|
||||
}
|
||||
|
||||
/* Do some firewalling */
|
||||
unsigned int ebt_do_table (unsigned int hook, struct sk_buff *skb,
|
||||
const struct net_device *in, const struct net_device *out,
|
||||
@@ -164,7 +170,7 @@ unsigned int ebt_do_table (unsigned int hook, struct sk_buff *skb,
|
||||
mtpar.in = tgpar.in = in;
|
||||
mtpar.out = tgpar.out = out;
|
||||
mtpar.hotdrop = &hotdrop;
|
||||
tgpar.hooknum = hook;
|
||||
mtpar.hooknum = tgpar.hooknum = hook;
|
||||
|
||||
read_lock_bh(&table->lock);
|
||||
private = table->private;
|
||||
@@ -249,8 +255,7 @@ letsreturn:
|
||||
/* jump to a udc */
|
||||
cs[sp].n = i + 1;
|
||||
cs[sp].chaininfo = chaininfo;
|
||||
cs[sp].e = (struct ebt_entry *)
|
||||
(((char *)point) + point->next_offset);
|
||||
cs[sp].e = ebt_next_entry(point);
|
||||
i = 0;
|
||||
chaininfo = (struct ebt_entries *) (base + verdict);
|
||||
#ifdef CONFIG_NETFILTER_DEBUG
|
||||
@@ -266,8 +271,7 @@ letsreturn:
|
||||
sp++;
|
||||
continue;
|
||||
letscontinue:
|
||||
point = (struct ebt_entry *)
|
||||
(((char *)point) + point->next_offset);
|
||||
point = ebt_next_entry(point);
|
||||
i++;
|
||||
}
|
||||
|
||||
@@ -787,7 +791,7 @@ static int check_chainloops(struct ebt_entries *chain, struct ebt_cl_stack *cl_s
|
||||
/* this can't be 0, so the loop test is correct */
|
||||
cl_s[i].cs.n = pos + 1;
|
||||
pos = 0;
|
||||
cl_s[i].cs.e = ((void *)e + e->next_offset);
|
||||
cl_s[i].cs.e = ebt_next_entry(e);
|
||||
e = (struct ebt_entry *)(hlp2->data);
|
||||
nentries = hlp2->nentries;
|
||||
cl_s[i].from = chain_nr;
|
||||
@@ -797,7 +801,7 @@ static int check_chainloops(struct ebt_entries *chain, struct ebt_cl_stack *cl_s
|
||||
continue;
|
||||
}
|
||||
letscontinue:
|
||||
e = (void *)e + e->next_offset;
|
||||
e = ebt_next_entry(e);
|
||||
pos++;
|
||||
}
|
||||
return 0;
|
||||
|
||||
@@ -231,6 +231,12 @@ static inline struct arpt_entry *get_entry(void *base, unsigned int offset)
|
||||
return (struct arpt_entry *)(base + offset);
|
||||
}
|
||||
|
||||
static inline __pure
|
||||
struct arpt_entry *arpt_next_entry(const struct arpt_entry *entry)
|
||||
{
|
||||
return (void *)entry + entry->next_offset;
|
||||
}
|
||||
|
||||
unsigned int arpt_do_table(struct sk_buff *skb,
|
||||
unsigned int hook,
|
||||
const struct net_device *in,
|
||||
@@ -267,67 +273,64 @@ unsigned int arpt_do_table(struct sk_buff *skb,
|
||||
|
||||
arp = arp_hdr(skb);
|
||||
do {
|
||||
if (arp_packet_match(arp, skb->dev, indev, outdev, &e->arp)) {
|
||||
struct arpt_entry_target *t;
|
||||
int hdr_len;
|
||||
struct arpt_entry_target *t;
|
||||
int hdr_len;
|
||||
|
||||
hdr_len = sizeof(*arp) + (2 * sizeof(struct in_addr)) +
|
||||
(2 * skb->dev->addr_len);
|
||||
|
||||
ADD_COUNTER(e->counters, hdr_len, 1);
|
||||
|
||||
t = arpt_get_target(e);
|
||||
|
||||
/* Standard target? */
|
||||
if (!t->u.kernel.target->target) {
|
||||
int v;
|
||||
|
||||
v = ((struct arpt_standard_target *)t)->verdict;
|
||||
if (v < 0) {
|
||||
/* Pop from stack? */
|
||||
if (v != ARPT_RETURN) {
|
||||
verdict = (unsigned)(-v) - 1;
|
||||
break;
|
||||
}
|
||||
e = back;
|
||||
back = get_entry(table_base,
|
||||
back->comefrom);
|
||||
continue;
|
||||
}
|
||||
if (table_base + v
|
||||
!= (void *)e + e->next_offset) {
|
||||
/* Save old back ptr in next entry */
|
||||
struct arpt_entry *next
|
||||
= (void *)e + e->next_offset;
|
||||
next->comefrom =
|
||||
(void *)back - table_base;
|
||||
|
||||
/* set back pointer to next entry */
|
||||
back = next;
|
||||
}
|
||||
|
||||
e = get_entry(table_base, v);
|
||||
} else {
|
||||
/* Targets which reenter must return
|
||||
* abs. verdicts
|
||||
*/
|
||||
tgpar.target = t->u.kernel.target;
|
||||
tgpar.targinfo = t->data;
|
||||
verdict = t->u.kernel.target->target(skb,
|
||||
&tgpar);
|
||||
|
||||
/* Target might have changed stuff. */
|
||||
arp = arp_hdr(skb);
|
||||
|
||||
if (verdict == ARPT_CONTINUE)
|
||||
e = (void *)e + e->next_offset;
|
||||
else
|
||||
/* Verdict */
|
||||
break;
|
||||
}
|
||||
} else {
|
||||
e = (void *)e + e->next_offset;
|
||||
if (!arp_packet_match(arp, skb->dev, indev, outdev, &e->arp)) {
|
||||
e = arpt_next_entry(e);
|
||||
continue;
|
||||
}
|
||||
|
||||
hdr_len = sizeof(*arp) + (2 * sizeof(struct in_addr)) +
|
||||
(2 * skb->dev->addr_len);
|
||||
ADD_COUNTER(e->counters, hdr_len, 1);
|
||||
|
||||
t = arpt_get_target(e);
|
||||
|
||||
/* Standard target? */
|
||||
if (!t->u.kernel.target->target) {
|
||||
int v;
|
||||
|
||||
v = ((struct arpt_standard_target *)t)->verdict;
|
||||
if (v < 0) {
|
||||
/* Pop from stack? */
|
||||
if (v != ARPT_RETURN) {
|
||||
verdict = (unsigned)(-v) - 1;
|
||||
break;
|
||||
}
|
||||
e = back;
|
||||
back = get_entry(table_base, back->comefrom);
|
||||
continue;
|
||||
}
|
||||
if (table_base + v
|
||||
!= arpt_next_entry(e)) {
|
||||
/* Save old back ptr in next entry */
|
||||
struct arpt_entry *next = arpt_next_entry(e);
|
||||
next->comefrom = (void *)back - table_base;
|
||||
|
||||
/* set back pointer to next entry */
|
||||
back = next;
|
||||
}
|
||||
|
||||
e = get_entry(table_base, v);
|
||||
continue;
|
||||
}
|
||||
|
||||
/* Targets which reenter must return
|
||||
* abs. verdicts
|
||||
*/
|
||||
tgpar.target = t->u.kernel.target;
|
||||
tgpar.targinfo = t->data;
|
||||
verdict = t->u.kernel.target->target(skb, &tgpar);
|
||||
|
||||
/* Target might have changed stuff. */
|
||||
arp = arp_hdr(skb);
|
||||
|
||||
if (verdict == ARPT_CONTINUE)
|
||||
e = arpt_next_entry(e);
|
||||
else
|
||||
/* Verdict */
|
||||
break;
|
||||
} while (!hotdrop);
|
||||
xt_info_rdunlock_bh();
|
||||
|
||||
|
||||
@@ -596,7 +596,7 @@ static int __init ip_queue_init(void)
|
||||
#ifdef CONFIG_SYSCTL
|
||||
ipq_sysctl_header = register_sysctl_paths(net_ipv4_ctl_path, ipq_table);
|
||||
#endif
|
||||
status = nf_register_queue_handler(PF_INET, &nfqh);
|
||||
status = nf_register_queue_handler(NFPROTO_IPV4, &nfqh);
|
||||
if (status < 0) {
|
||||
printk(KERN_ERR "ip_queue: failed to register queue handler\n");
|
||||
goto cleanup_sysctl;
|
||||
|
||||
@@ -238,8 +238,8 @@ static struct nf_loginfo trace_loginfo = {
|
||||
/* Mildly perf critical (only if packet tracing is on) */
|
||||
static inline int
|
||||
get_chainname_rulenum(struct ipt_entry *s, struct ipt_entry *e,
|
||||
char *hookname, char **chainname,
|
||||
char **comment, unsigned int *rulenum)
|
||||
const char *hookname, const char **chainname,
|
||||
const char **comment, unsigned int *rulenum)
|
||||
{
|
||||
struct ipt_standard_target *t = (void *)ipt_get_target(s);
|
||||
|
||||
@@ -257,8 +257,8 @@ get_chainname_rulenum(struct ipt_entry *s, struct ipt_entry *e,
|
||||
&& unconditional(&s->ip)) {
|
||||
/* Tail of chains: STANDARD target (return/policy) */
|
||||
*comment = *chainname == hookname
|
||||
? (char *)comments[NF_IP_TRACE_COMMENT_POLICY]
|
||||
: (char *)comments[NF_IP_TRACE_COMMENT_RETURN];
|
||||
? comments[NF_IP_TRACE_COMMENT_POLICY]
|
||||
: comments[NF_IP_TRACE_COMMENT_RETURN];
|
||||
}
|
||||
return 1;
|
||||
} else
|
||||
@@ -277,14 +277,14 @@ static void trace_packet(struct sk_buff *skb,
|
||||
{
|
||||
void *table_base;
|
||||
const struct ipt_entry *root;
|
||||
char *hookname, *chainname, *comment;
|
||||
const char *hookname, *chainname, *comment;
|
||||
unsigned int rulenum = 0;
|
||||
|
||||
table_base = (void *)private->entries[smp_processor_id()];
|
||||
table_base = private->entries[smp_processor_id()];
|
||||
root = get_entry(table_base, private->hook_entry[hook]);
|
||||
|
||||
hookname = chainname = (char *)hooknames[hook];
|
||||
comment = (char *)comments[NF_IP_TRACE_COMMENT_RULE];
|
||||
hookname = chainname = hooknames[hook];
|
||||
comment = comments[NF_IP_TRACE_COMMENT_RULE];
|
||||
|
||||
IPT_ENTRY_ITERATE(root,
|
||||
private->size - private->hook_entry[hook],
|
||||
@@ -297,6 +297,12 @@ static void trace_packet(struct sk_buff *skb,
|
||||
}
|
||||
#endif
|
||||
|
||||
static inline __pure
|
||||
struct ipt_entry *ipt_next_entry(const struct ipt_entry *entry)
|
||||
{
|
||||
return (void *)entry + entry->next_offset;
|
||||
}
|
||||
|
||||
/* Returns one of the generic firewall policies, like NF_ACCEPT. */
|
||||
unsigned int
|
||||
ipt_do_table(struct sk_buff *skb,
|
||||
@@ -305,6 +311,8 @@ ipt_do_table(struct sk_buff *skb,
|
||||
const struct net_device *out,
|
||||
struct xt_table *table)
|
||||
{
|
||||
#define tb_comefrom ((struct ipt_entry *)table_base)->comefrom
|
||||
|
||||
static const char nulldevname[IFNAMSIZ] __attribute__((aligned(sizeof(long))));
|
||||
const struct iphdr *ip;
|
||||
u_int16_t datalen;
|
||||
@@ -335,7 +343,7 @@ ipt_do_table(struct sk_buff *skb,
|
||||
mtpar.in = tgpar.in = in;
|
||||
mtpar.out = tgpar.out = out;
|
||||
mtpar.family = tgpar.family = NFPROTO_IPV4;
|
||||
tgpar.hooknum = hook;
|
||||
mtpar.hooknum = tgpar.hooknum = hook;
|
||||
|
||||
IP_NF_ASSERT(table->valid_hooks & (1 << hook));
|
||||
xt_info_rdlock_bh();
|
||||
@@ -348,92 +356,84 @@ ipt_do_table(struct sk_buff *skb,
|
||||
back = get_entry(table_base, private->underflow[hook]);
|
||||
|
||||
do {
|
||||
struct ipt_entry_target *t;
|
||||
|
||||
IP_NF_ASSERT(e);
|
||||
IP_NF_ASSERT(back);
|
||||
if (ip_packet_match(ip, indev, outdev,
|
||||
&e->ip, mtpar.fragoff)) {
|
||||
struct ipt_entry_target *t;
|
||||
if (!ip_packet_match(ip, indev, outdev,
|
||||
&e->ip, mtpar.fragoff) ||
|
||||
IPT_MATCH_ITERATE(e, do_match, skb, &mtpar) != 0) {
|
||||
e = ipt_next_entry(e);
|
||||
continue;
|
||||
}
|
||||
|
||||
if (IPT_MATCH_ITERATE(e, do_match, skb, &mtpar) != 0)
|
||||
goto no_match;
|
||||
ADD_COUNTER(e->counters, ntohs(ip->tot_len), 1);
|
||||
|
||||
ADD_COUNTER(e->counters, ntohs(ip->tot_len), 1);
|
||||
|
||||
t = ipt_get_target(e);
|
||||
IP_NF_ASSERT(t->u.kernel.target);
|
||||
t = ipt_get_target(e);
|
||||
IP_NF_ASSERT(t->u.kernel.target);
|
||||
|
||||
#if defined(CONFIG_NETFILTER_XT_TARGET_TRACE) || \
|
||||
defined(CONFIG_NETFILTER_XT_TARGET_TRACE_MODULE)
|
||||
/* The packet is traced: log it */
|
||||
if (unlikely(skb->nf_trace))
|
||||
trace_packet(skb, hook, in, out,
|
||||
table->name, private, e);
|
||||
/* The packet is traced: log it */
|
||||
if (unlikely(skb->nf_trace))
|
||||
trace_packet(skb, hook, in, out,
|
||||
table->name, private, e);
|
||||
#endif
|
||||
/* Standard target? */
|
||||
if (!t->u.kernel.target->target) {
|
||||
int v;
|
||||
/* Standard target? */
|
||||
if (!t->u.kernel.target->target) {
|
||||
int v;
|
||||
|
||||
v = ((struct ipt_standard_target *)t)->verdict;
|
||||
if (v < 0) {
|
||||
/* Pop from stack? */
|
||||
if (v != IPT_RETURN) {
|
||||
verdict = (unsigned)(-v) - 1;
|
||||
break;
|
||||
}
|
||||
e = back;
|
||||
back = get_entry(table_base,
|
||||
back->comefrom);
|
||||
continue;
|
||||
}
|
||||
if (table_base + v != (void *)e + e->next_offset
|
||||
&& !(e->ip.flags & IPT_F_GOTO)) {
|
||||
/* Save old back ptr in next entry */
|
||||
struct ipt_entry *next
|
||||
= (void *)e + e->next_offset;
|
||||
next->comefrom
|
||||
= (void *)back - table_base;
|
||||
/* set back pointer to next entry */
|
||||
back = next;
|
||||
}
|
||||
|
||||
e = get_entry(table_base, v);
|
||||
} else {
|
||||
/* Targets which reenter must return
|
||||
abs. verdicts */
|
||||
tgpar.target = t->u.kernel.target;
|
||||
tgpar.targinfo = t->data;
|
||||
#ifdef CONFIG_NETFILTER_DEBUG
|
||||
((struct ipt_entry *)table_base)->comefrom
|
||||
= 0xeeeeeeec;
|
||||
#endif
|
||||
verdict = t->u.kernel.target->target(skb,
|
||||
&tgpar);
|
||||
#ifdef CONFIG_NETFILTER_DEBUG
|
||||
if (((struct ipt_entry *)table_base)->comefrom
|
||||
!= 0xeeeeeeec
|
||||
&& verdict == IPT_CONTINUE) {
|
||||
printk("Target %s reentered!\n",
|
||||
t->u.kernel.target->name);
|
||||
verdict = NF_DROP;
|
||||
}
|
||||
((struct ipt_entry *)table_base)->comefrom
|
||||
= 0x57acc001;
|
||||
#endif
|
||||
/* Target might have changed stuff. */
|
||||
ip = ip_hdr(skb);
|
||||
datalen = skb->len - ip->ihl * 4;
|
||||
|
||||
if (verdict == IPT_CONTINUE)
|
||||
e = (void *)e + e->next_offset;
|
||||
else
|
||||
/* Verdict */
|
||||
v = ((struct ipt_standard_target *)t)->verdict;
|
||||
if (v < 0) {
|
||||
/* Pop from stack? */
|
||||
if (v != IPT_RETURN) {
|
||||
verdict = (unsigned)(-v) - 1;
|
||||
break;
|
||||
}
|
||||
e = back;
|
||||
back = get_entry(table_base, back->comefrom);
|
||||
continue;
|
||||
}
|
||||
if (table_base + v != ipt_next_entry(e)
|
||||
&& !(e->ip.flags & IPT_F_GOTO)) {
|
||||
/* Save old back ptr in next entry */
|
||||
struct ipt_entry *next = ipt_next_entry(e);
|
||||
next->comefrom = (void *)back - table_base;
|
||||
/* set back pointer to next entry */
|
||||
back = next;
|
||||
}
|
||||
} else {
|
||||
|
||||
no_match:
|
||||
e = (void *)e + e->next_offset;
|
||||
e = get_entry(table_base, v);
|
||||
continue;
|
||||
}
|
||||
|
||||
/* Targets which reenter must return
|
||||
abs. verdicts */
|
||||
tgpar.target = t->u.kernel.target;
|
||||
tgpar.targinfo = t->data;
|
||||
|
||||
|
||||
#ifdef CONFIG_NETFILTER_DEBUG
|
||||
tb_comefrom = 0xeeeeeeec;
|
||||
#endif
|
||||
verdict = t->u.kernel.target->target(skb, &tgpar);
|
||||
#ifdef CONFIG_NETFILTER_DEBUG
|
||||
if (tb_comefrom != 0xeeeeeeec && verdict == IPT_CONTINUE) {
|
||||
printk("Target %s reentered!\n",
|
||||
t->u.kernel.target->name);
|
||||
verdict = NF_DROP;
|
||||
}
|
||||
tb_comefrom = 0x57acc001;
|
||||
#endif
|
||||
/* Target might have changed stuff. */
|
||||
ip = ip_hdr(skb);
|
||||
datalen = skb->len - ip->ihl * 4;
|
||||
|
||||
if (verdict == IPT_CONTINUE)
|
||||
e = ipt_next_entry(e);
|
||||
else
|
||||
/* Verdict */
|
||||
break;
|
||||
} while (!hotdrop);
|
||||
xt_info_rdunlock_bh();
|
||||
|
||||
@@ -444,6 +444,8 @@ ipt_do_table(struct sk_buff *skb,
|
||||
return NF_DROP;
|
||||
else return verdict;
|
||||
#endif
|
||||
|
||||
#undef tb_comefrom
|
||||
}
|
||||
|
||||
/* Figures out from what hook each rule can be called: returns 0 if
|
||||
@@ -2158,7 +2160,7 @@ static bool icmp_checkentry(const struct xt_mtchk_param *par)
|
||||
static struct xt_target ipt_standard_target __read_mostly = {
|
||||
.name = IPT_STANDARD_TARGET,
|
||||
.targetsize = sizeof(int),
|
||||
.family = AF_INET,
|
||||
.family = NFPROTO_IPV4,
|
||||
#ifdef CONFIG_COMPAT
|
||||
.compatsize = sizeof(compat_int_t),
|
||||
.compat_from_user = compat_standard_from_user,
|
||||
@@ -2170,7 +2172,7 @@ static struct xt_target ipt_error_target __read_mostly = {
|
||||
.name = IPT_ERROR_TARGET,
|
||||
.target = ipt_error,
|
||||
.targetsize = IPT_FUNCTION_MAXNAMELEN,
|
||||
.family = AF_INET,
|
||||
.family = NFPROTO_IPV4,
|
||||
};
|
||||
|
||||
static struct nf_sockopt_ops ipt_sockopts = {
|
||||
@@ -2196,17 +2198,17 @@ static struct xt_match icmp_matchstruct __read_mostly = {
|
||||
.matchsize = sizeof(struct ipt_icmp),
|
||||
.checkentry = icmp_checkentry,
|
||||
.proto = IPPROTO_ICMP,
|
||||
.family = AF_INET,
|
||||
.family = NFPROTO_IPV4,
|
||||
};
|
||||
|
||||
static int __net_init ip_tables_net_init(struct net *net)
|
||||
{
|
||||
return xt_proto_init(net, AF_INET);
|
||||
return xt_proto_init(net, NFPROTO_IPV4);
|
||||
}
|
||||
|
||||
static void __net_exit ip_tables_net_exit(struct net *net)
|
||||
{
|
||||
xt_proto_fini(net, AF_INET);
|
||||
xt_proto_fini(net, NFPROTO_IPV4);
|
||||
}
|
||||
|
||||
static struct pernet_operations ip_tables_net_ops = {
|
||||
|
||||
@@ -27,9 +27,6 @@ MODULE_LICENSE("GPL");
|
||||
MODULE_AUTHOR("Netfilter Core Team <coreteam@netfilter.org>");
|
||||
MODULE_DESCRIPTION("Xtables: automatic-address SNAT");
|
||||
|
||||
/* Lock protects masq region inside conntrack */
|
||||
static DEFINE_RWLOCK(masq_lock);
|
||||
|
||||
/* FIXME: Multiple targets. --RR */
|
||||
static bool masquerade_tg_check(const struct xt_tgchk_param *par)
|
||||
{
|
||||
@@ -79,9 +76,7 @@ masquerade_tg(struct sk_buff *skb, const struct xt_target_param *par)
|
||||
return NF_DROP;
|
||||
}
|
||||
|
||||
write_lock_bh(&masq_lock);
|
||||
nat->masq_index = par->out->ifindex;
|
||||
write_unlock_bh(&masq_lock);
|
||||
|
||||
/* Transfer from original range. */
|
||||
newrange = ((struct nf_nat_range)
|
||||
@@ -97,16 +92,11 @@ static int
|
||||
device_cmp(struct nf_conn *i, void *ifindex)
|
||||
{
|
||||
const struct nf_conn_nat *nat = nfct_nat(i);
|
||||
int ret;
|
||||
|
||||
if (!nat)
|
||||
return 0;
|
||||
|
||||
read_lock_bh(&masq_lock);
|
||||
ret = (nat->masq_index == (int)(long)ifindex);
|
||||
read_unlock_bh(&masq_lock);
|
||||
|
||||
return ret;
|
||||
return nat->masq_index == (int)(long)ifindex;
|
||||
}
|
||||
|
||||
static int masq_device_event(struct notifier_block *this,
|
||||
|
||||
@@ -82,18 +82,10 @@ static int icmp_packet(struct nf_conn *ct,
|
||||
u_int8_t pf,
|
||||
unsigned int hooknum)
|
||||
{
|
||||
/* Try to delete connection immediately after all replies:
|
||||
won't actually vanish as we still have skb, and del_timer
|
||||
means this will only run once even if count hits zero twice
|
||||
(theoretically possible with SMP) */
|
||||
if (CTINFO2DIR(ctinfo) == IP_CT_DIR_REPLY) {
|
||||
if (atomic_dec_and_test(&ct->proto.icmp.count))
|
||||
nf_ct_kill_acct(ct, ctinfo, skb);
|
||||
} else {
|
||||
atomic_inc(&ct->proto.icmp.count);
|
||||
nf_conntrack_event_cache(IPCT_PROTOINFO_VOLATILE, ct);
|
||||
nf_ct_refresh_acct(ct, ctinfo, skb, nf_ct_icmp_timeout);
|
||||
}
|
||||
/* Do not immediately delete the connection after the first
|
||||
successful reply to avoid excessive conntrackd traffic
|
||||
and also to handle correctly ICMP echo reply duplicates. */
|
||||
nf_ct_refresh_acct(ct, ctinfo, skb, nf_ct_icmp_timeout);
|
||||
|
||||
return NF_ACCEPT;
|
||||
}
|
||||
@@ -117,7 +109,6 @@ static bool icmp_new(struct nf_conn *ct, const struct sk_buff *skb,
|
||||
nf_ct_dump_tuple_ip(&ct->tuplehash[0].tuple);
|
||||
return false;
|
||||
}
|
||||
atomic_set(&ct->proto.icmp.count, 0);
|
||||
return true;
|
||||
}
|
||||
|
||||
|
||||
@@ -598,7 +598,7 @@ static int __init ip6_queue_init(void)
|
||||
#ifdef CONFIG_SYSCTL
|
||||
ipq_sysctl_header = register_sysctl_paths(net_ipv6_ctl_path, ipq_table);
|
||||
#endif
|
||||
status = nf_register_queue_handler(PF_INET6, &nfqh);
|
||||
status = nf_register_queue_handler(NFPROTO_IPV6, &nfqh);
|
||||
if (status < 0) {
|
||||
printk(KERN_ERR "ip6_queue: failed to register queue handler\n");
|
||||
goto cleanup_sysctl;
|
||||
|
||||
@@ -270,8 +270,8 @@ static struct nf_loginfo trace_loginfo = {
|
||||
/* Mildly perf critical (only if packet tracing is on) */
|
||||
static inline int
|
||||
get_chainname_rulenum(struct ip6t_entry *s, struct ip6t_entry *e,
|
||||
char *hookname, char **chainname,
|
||||
char **comment, unsigned int *rulenum)
|
||||
const char *hookname, const char **chainname,
|
||||
const char **comment, unsigned int *rulenum)
|
||||
{
|
||||
struct ip6t_standard_target *t = (void *)ip6t_get_target(s);
|
||||
|
||||
@@ -289,8 +289,8 @@ get_chainname_rulenum(struct ip6t_entry *s, struct ip6t_entry *e,
|
||||
&& unconditional(&s->ipv6)) {
|
||||
/* Tail of chains: STANDARD target (return/policy) */
|
||||
*comment = *chainname == hookname
|
||||
? (char *)comments[NF_IP6_TRACE_COMMENT_POLICY]
|
||||
: (char *)comments[NF_IP6_TRACE_COMMENT_RETURN];
|
||||
? comments[NF_IP6_TRACE_COMMENT_POLICY]
|
||||
: comments[NF_IP6_TRACE_COMMENT_RETURN];
|
||||
}
|
||||
return 1;
|
||||
} else
|
||||
@@ -309,14 +309,14 @@ static void trace_packet(struct sk_buff *skb,
|
||||
{
|
||||
void *table_base;
|
||||
const struct ip6t_entry *root;
|
||||
char *hookname, *chainname, *comment;
|
||||
const char *hookname, *chainname, *comment;
|
||||
unsigned int rulenum = 0;
|
||||
|
||||
table_base = (void *)private->entries[smp_processor_id()];
|
||||
table_base = private->entries[smp_processor_id()];
|
||||
root = get_entry(table_base, private->hook_entry[hook]);
|
||||
|
||||
hookname = chainname = (char *)hooknames[hook];
|
||||
comment = (char *)comments[NF_IP6_TRACE_COMMENT_RULE];
|
||||
hookname = chainname = hooknames[hook];
|
||||
comment = comments[NF_IP6_TRACE_COMMENT_RULE];
|
||||
|
||||
IP6T_ENTRY_ITERATE(root,
|
||||
private->size - private->hook_entry[hook],
|
||||
@@ -329,6 +329,12 @@ static void trace_packet(struct sk_buff *skb,
|
||||
}
|
||||
#endif
|
||||
|
||||
static inline __pure struct ip6t_entry *
|
||||
ip6t_next_entry(const struct ip6t_entry *entry)
|
||||
{
|
||||
return (void *)entry + entry->next_offset;
|
||||
}
|
||||
|
||||
/* Returns one of the generic firewall policies, like NF_ACCEPT. */
|
||||
unsigned int
|
||||
ip6t_do_table(struct sk_buff *skb,
|
||||
@@ -337,6 +343,8 @@ ip6t_do_table(struct sk_buff *skb,
|
||||
const struct net_device *out,
|
||||
struct xt_table *table)
|
||||
{
|
||||
#define tb_comefrom ((struct ip6t_entry *)table_base)->comefrom
|
||||
|
||||
static const char nulldevname[IFNAMSIZ] __attribute__((aligned(sizeof(long))));
|
||||
bool hotdrop = false;
|
||||
/* Initializing verdict to NF_DROP keeps gcc happy. */
|
||||
@@ -361,7 +369,7 @@ ip6t_do_table(struct sk_buff *skb,
|
||||
mtpar.in = tgpar.in = in;
|
||||
mtpar.out = tgpar.out = out;
|
||||
mtpar.family = tgpar.family = NFPROTO_IPV6;
|
||||
tgpar.hooknum = hook;
|
||||
mtpar.hooknum = tgpar.hooknum = hook;
|
||||
|
||||
IP_NF_ASSERT(table->valid_hooks & (1 << hook));
|
||||
|
||||
@@ -375,96 +383,86 @@ ip6t_do_table(struct sk_buff *skb,
|
||||
back = get_entry(table_base, private->underflow[hook]);
|
||||
|
||||
do {
|
||||
struct ip6t_entry_target *t;
|
||||
|
||||
IP_NF_ASSERT(e);
|
||||
IP_NF_ASSERT(back);
|
||||
if (ip6_packet_match(skb, indev, outdev, &e->ipv6,
|
||||
&mtpar.thoff, &mtpar.fragoff, &hotdrop)) {
|
||||
struct ip6t_entry_target *t;
|
||||
if (!ip6_packet_match(skb, indev, outdev, &e->ipv6,
|
||||
&mtpar.thoff, &mtpar.fragoff, &hotdrop) ||
|
||||
IP6T_MATCH_ITERATE(e, do_match, skb, &mtpar) != 0) {
|
||||
e = ip6t_next_entry(e);
|
||||
continue;
|
||||
}
|
||||
|
||||
if (IP6T_MATCH_ITERATE(e, do_match, skb, &mtpar) != 0)
|
||||
goto no_match;
|
||||
ADD_COUNTER(e->counters,
|
||||
ntohs(ipv6_hdr(skb)->payload_len) +
|
||||
sizeof(struct ipv6hdr), 1);
|
||||
|
||||
ADD_COUNTER(e->counters,
|
||||
ntohs(ipv6_hdr(skb)->payload_len) +
|
||||
sizeof(struct ipv6hdr), 1);
|
||||
|
||||
t = ip6t_get_target(e);
|
||||
IP_NF_ASSERT(t->u.kernel.target);
|
||||
t = ip6t_get_target(e);
|
||||
IP_NF_ASSERT(t->u.kernel.target);
|
||||
|
||||
#if defined(CONFIG_NETFILTER_XT_TARGET_TRACE) || \
|
||||
defined(CONFIG_NETFILTER_XT_TARGET_TRACE_MODULE)
|
||||
/* The packet is traced: log it */
|
||||
if (unlikely(skb->nf_trace))
|
||||
trace_packet(skb, hook, in, out,
|
||||
table->name, private, e);
|
||||
/* The packet is traced: log it */
|
||||
if (unlikely(skb->nf_trace))
|
||||
trace_packet(skb, hook, in, out,
|
||||
table->name, private, e);
|
||||
#endif
|
||||
/* Standard target? */
|
||||
if (!t->u.kernel.target->target) {
|
||||
int v;
|
||||
/* Standard target? */
|
||||
if (!t->u.kernel.target->target) {
|
||||
int v;
|
||||
|
||||
v = ((struct ip6t_standard_target *)t)->verdict;
|
||||
if (v < 0) {
|
||||
/* Pop from stack? */
|
||||
if (v != IP6T_RETURN) {
|
||||
verdict = (unsigned)(-v) - 1;
|
||||
break;
|
||||
}
|
||||
e = back;
|
||||
back = get_entry(table_base,
|
||||
back->comefrom);
|
||||
continue;
|
||||
}
|
||||
if (table_base + v != (void *)e + e->next_offset
|
||||
&& !(e->ipv6.flags & IP6T_F_GOTO)) {
|
||||
/* Save old back ptr in next entry */
|
||||
struct ip6t_entry *next
|
||||
= (void *)e + e->next_offset;
|
||||
next->comefrom
|
||||
= (void *)back - table_base;
|
||||
/* set back pointer to next entry */
|
||||
back = next;
|
||||
}
|
||||
|
||||
e = get_entry(table_base, v);
|
||||
} else {
|
||||
/* Targets which reenter must return
|
||||
abs. verdicts */
|
||||
tgpar.target = t->u.kernel.target;
|
||||
tgpar.targinfo = t->data;
|
||||
|
||||
#ifdef CONFIG_NETFILTER_DEBUG
|
||||
((struct ip6t_entry *)table_base)->comefrom
|
||||
= 0xeeeeeeec;
|
||||
#endif
|
||||
verdict = t->u.kernel.target->target(skb,
|
||||
&tgpar);
|
||||
|
||||
#ifdef CONFIG_NETFILTER_DEBUG
|
||||
if (((struct ip6t_entry *)table_base)->comefrom
|
||||
!= 0xeeeeeeec
|
||||
&& verdict == IP6T_CONTINUE) {
|
||||
printk("Target %s reentered!\n",
|
||||
t->u.kernel.target->name);
|
||||
verdict = NF_DROP;
|
||||
}
|
||||
((struct ip6t_entry *)table_base)->comefrom
|
||||
= 0x57acc001;
|
||||
#endif
|
||||
if (verdict == IP6T_CONTINUE)
|
||||
e = (void *)e + e->next_offset;
|
||||
else
|
||||
/* Verdict */
|
||||
v = ((struct ip6t_standard_target *)t)->verdict;
|
||||
if (v < 0) {
|
||||
/* Pop from stack? */
|
||||
if (v != IP6T_RETURN) {
|
||||
verdict = (unsigned)(-v) - 1;
|
||||
break;
|
||||
}
|
||||
e = back;
|
||||
back = get_entry(table_base, back->comefrom);
|
||||
continue;
|
||||
}
|
||||
if (table_base + v != ip6t_next_entry(e)
|
||||
&& !(e->ipv6.flags & IP6T_F_GOTO)) {
|
||||
/* Save old back ptr in next entry */
|
||||
struct ip6t_entry *next = ip6t_next_entry(e);
|
||||
next->comefrom = (void *)back - table_base;
|
||||
/* set back pointer to next entry */
|
||||
back = next;
|
||||
}
|
||||
} else {
|
||||
|
||||
no_match:
|
||||
e = (void *)e + e->next_offset;
|
||||
e = get_entry(table_base, v);
|
||||
continue;
|
||||
}
|
||||
|
||||
/* Targets which reenter must return
|
||||
abs. verdicts */
|
||||
tgpar.target = t->u.kernel.target;
|
||||
tgpar.targinfo = t->data;
|
||||
|
||||
#ifdef CONFIG_NETFILTER_DEBUG
|
||||
tb_comefrom = 0xeeeeeeec;
|
||||
#endif
|
||||
verdict = t->u.kernel.target->target(skb, &tgpar);
|
||||
|
||||
#ifdef CONFIG_NETFILTER_DEBUG
|
||||
if (tb_comefrom != 0xeeeeeeec && verdict == IP6T_CONTINUE) {
|
||||
printk("Target %s reentered!\n",
|
||||
t->u.kernel.target->name);
|
||||
verdict = NF_DROP;
|
||||
}
|
||||
tb_comefrom = 0x57acc001;
|
||||
#endif
|
||||
if (verdict == IP6T_CONTINUE)
|
||||
e = ip6t_next_entry(e);
|
||||
else
|
||||
/* Verdict */
|
||||
break;
|
||||
} while (!hotdrop);
|
||||
|
||||
#ifdef CONFIG_NETFILTER_DEBUG
|
||||
((struct ip6t_entry *)table_base)->comefrom = NETFILTER_LINK_POISON;
|
||||
tb_comefrom = NETFILTER_LINK_POISON;
|
||||
#endif
|
||||
xt_info_rdunlock_bh();
|
||||
|
||||
@@ -475,6 +473,8 @@ ip6t_do_table(struct sk_buff *skb,
|
||||
return NF_DROP;
|
||||
else return verdict;
|
||||
#endif
|
||||
|
||||
#undef tb_comefrom
|
||||
}
|
||||
|
||||
/* Figures out from what hook each rule can be called: returns 0 if
|
||||
@@ -2191,7 +2191,7 @@ static bool icmp6_checkentry(const struct xt_mtchk_param *par)
|
||||
static struct xt_target ip6t_standard_target __read_mostly = {
|
||||
.name = IP6T_STANDARD_TARGET,
|
||||
.targetsize = sizeof(int),
|
||||
.family = AF_INET6,
|
||||
.family = NFPROTO_IPV6,
|
||||
#ifdef CONFIG_COMPAT
|
||||
.compatsize = sizeof(compat_int_t),
|
||||
.compat_from_user = compat_standard_from_user,
|
||||
@@ -2203,7 +2203,7 @@ static struct xt_target ip6t_error_target __read_mostly = {
|
||||
.name = IP6T_ERROR_TARGET,
|
||||
.target = ip6t_error,
|
||||
.targetsize = IP6T_FUNCTION_MAXNAMELEN,
|
||||
.family = AF_INET6,
|
||||
.family = NFPROTO_IPV6,
|
||||
};
|
||||
|
||||
static struct nf_sockopt_ops ip6t_sockopts = {
|
||||
@@ -2229,17 +2229,17 @@ static struct xt_match icmp6_matchstruct __read_mostly = {
|
||||
.matchsize = sizeof(struct ip6t_icmp),
|
||||
.checkentry = icmp6_checkentry,
|
||||
.proto = IPPROTO_ICMPV6,
|
||||
.family = AF_INET6,
|
||||
.family = NFPROTO_IPV6,
|
||||
};
|
||||
|
||||
static int __net_init ip6_tables_net_init(struct net *net)
|
||||
{
|
||||
return xt_proto_init(net, AF_INET6);
|
||||
return xt_proto_init(net, NFPROTO_IPV6);
|
||||
}
|
||||
|
||||
static void __net_exit ip6_tables_net_exit(struct net *net)
|
||||
{
|
||||
xt_proto_fini(net, AF_INET6);
|
||||
xt_proto_fini(net, NFPROTO_IPV6);
|
||||
}
|
||||
|
||||
static struct pernet_operations ip6_tables_net_ops = {
|
||||
|
||||
@@ -95,18 +95,10 @@ static int icmpv6_packet(struct nf_conn *ct,
|
||||
u_int8_t pf,
|
||||
unsigned int hooknum)
|
||||
{
|
||||
/* Try to delete connection immediately after all replies:
|
||||
won't actually vanish as we still have skb, and del_timer
|
||||
means this will only run once even if count hits zero twice
|
||||
(theoretically possible with SMP) */
|
||||
if (CTINFO2DIR(ctinfo) == IP_CT_DIR_REPLY) {
|
||||
if (atomic_dec_and_test(&ct->proto.icmp.count))
|
||||
nf_ct_kill_acct(ct, ctinfo, skb);
|
||||
} else {
|
||||
atomic_inc(&ct->proto.icmp.count);
|
||||
nf_conntrack_event_cache(IPCT_PROTOINFO_VOLATILE, ct);
|
||||
nf_ct_refresh_acct(ct, ctinfo, skb, nf_ct_icmpv6_timeout);
|
||||
}
|
||||
/* Do not immediately delete the connection after the first
|
||||
successful reply to avoid excessive conntrackd traffic
|
||||
and also to handle correctly ICMP echo reply duplicates. */
|
||||
nf_ct_refresh_acct(ct, ctinfo, skb, nf_ct_icmpv6_timeout);
|
||||
|
||||
return NF_ACCEPT;
|
||||
}
|
||||
@@ -132,7 +124,6 @@ static bool icmpv6_new(struct nf_conn *ct, const struct sk_buff *skb,
|
||||
type + 128);
|
||||
return false;
|
||||
}
|
||||
atomic_set(&ct->proto.icmp.count, 0);
|
||||
return true;
|
||||
}
|
||||
|
||||
|
||||
@@ -917,6 +917,19 @@ config NETFILTER_XT_MATCH_U32
|
||||
|
||||
Details and examples are in the kernel module source.
|
||||
|
||||
config NETFILTER_XT_MATCH_OSF
|
||||
tristate '"osf" Passive OS fingerprint match'
|
||||
depends on NETFILTER_ADVANCED && NETFILTER_NETLINK
|
||||
help
|
||||
This option selects the Passive OS Fingerprinting match module
|
||||
that allows to passively match the remote operating system by
|
||||
analyzing incoming TCP SYN packets.
|
||||
|
||||
Rules and loading software can be downloaded from
|
||||
http://www.ioremap.net/projects/osf
|
||||
|
||||
To compile it as a module, choose M here. If unsure, say N.
|
||||
|
||||
endif # NETFILTER_XTABLES
|
||||
|
||||
endmenu
|
||||
|
||||
@@ -77,6 +77,7 @@ obj-$(CONFIG_NETFILTER_XT_MATCH_LIMIT) += xt_limit.o
|
||||
obj-$(CONFIG_NETFILTER_XT_MATCH_MAC) += xt_mac.o
|
||||
obj-$(CONFIG_NETFILTER_XT_MATCH_MARK) += xt_mark.o
|
||||
obj-$(CONFIG_NETFILTER_XT_MATCH_MULTIPORT) += xt_multiport.o
|
||||
obj-$(CONFIG_NETFILTER_XT_MATCH_OSF) += xt_osf.o
|
||||
obj-$(CONFIG_NETFILTER_XT_MATCH_OWNER) += xt_owner.o
|
||||
obj-$(CONFIG_NETFILTER_XT_MATCH_PHYSDEV) += xt_physdev.o
|
||||
obj-$(CONFIG_NETFILTER_XT_MATCH_PKTTYPE) += xt_pkttype.o
|
||||
|
||||
@@ -398,11 +398,7 @@ __nf_conntrack_confirm(struct sk_buff *skb)
|
||||
help = nfct_help(ct);
|
||||
if (help && help->helper)
|
||||
nf_conntrack_event_cache(IPCT_HELPER, ct);
|
||||
#ifdef CONFIG_NF_NAT_NEEDED
|
||||
if (test_bit(IPS_SRC_NAT_DONE_BIT, &ct->status) ||
|
||||
test_bit(IPS_DST_NAT_DONE_BIT, &ct->status))
|
||||
nf_conntrack_event_cache(IPCT_NATINFO, ct);
|
||||
#endif
|
||||
|
||||
nf_conntrack_event_cache(master_ct(ct) ?
|
||||
IPCT_RELATED : IPCT_NEW, ct);
|
||||
return NF_ACCEPT;
|
||||
@@ -523,6 +519,7 @@ struct nf_conn *nf_conntrack_alloc(struct net *net,
|
||||
return ERR_PTR(-ENOMEM);
|
||||
}
|
||||
|
||||
spin_lock_init(&ct->lock);
|
||||
atomic_set(&ct->ct_general.use, 1);
|
||||
ct->tuplehash[IP_CT_DIR_ORIGINAL].tuple = *orig;
|
||||
ct->tuplehash[IP_CT_DIR_REPLY].tuple = *repl;
|
||||
@@ -807,8 +804,6 @@ void __nf_ct_refresh_acct(struct nf_conn *ct,
|
||||
unsigned long extra_jiffies,
|
||||
int do_acct)
|
||||
{
|
||||
int event = 0;
|
||||
|
||||
NF_CT_ASSERT(ct->timeout.data == (unsigned long)ct);
|
||||
NF_CT_ASSERT(skb);
|
||||
|
||||
@@ -821,7 +816,6 @@ void __nf_ct_refresh_acct(struct nf_conn *ct,
|
||||
/* If not in hash table, timer will not be active yet */
|
||||
if (!nf_ct_is_confirmed(ct)) {
|
||||
ct->timeout.expires = extra_jiffies;
|
||||
event = IPCT_REFRESH;
|
||||
} else {
|
||||
unsigned long newtime = jiffies + extra_jiffies;
|
||||
|
||||
@@ -832,7 +826,6 @@ void __nf_ct_refresh_acct(struct nf_conn *ct,
|
||||
&& del_timer(&ct->timeout)) {
|
||||
ct->timeout.expires = newtime;
|
||||
add_timer(&ct->timeout);
|
||||
event = IPCT_REFRESH;
|
||||
}
|
||||
}
|
||||
|
||||
@@ -849,10 +842,6 @@ acct:
|
||||
}
|
||||
|
||||
spin_unlock_bh(&nf_conntrack_lock);
|
||||
|
||||
/* must be unlocked when calling event cache */
|
||||
if (event)
|
||||
nf_conntrack_event_cache(event, ct);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(__nf_ct_refresh_acct);
|
||||
|
||||
@@ -1001,7 +990,7 @@ struct __nf_ct_flush_report {
|
||||
int report;
|
||||
};
|
||||
|
||||
static int kill_all(struct nf_conn *i, void *data)
|
||||
static int kill_report(struct nf_conn *i, void *data)
|
||||
{
|
||||
struct __nf_ct_flush_report *fr = (struct __nf_ct_flush_report *)data;
|
||||
|
||||
@@ -1013,6 +1002,11 @@ static int kill_all(struct nf_conn *i, void *data)
|
||||
return 1;
|
||||
}
|
||||
|
||||
static int kill_all(struct nf_conn *i, void *data)
|
||||
{
|
||||
return 1;
|
||||
}
|
||||
|
||||
void nf_ct_free_hashtable(void *hash, int vmalloced, unsigned int size)
|
||||
{
|
||||
if (vmalloced)
|
||||
@@ -1023,15 +1017,15 @@ void nf_ct_free_hashtable(void *hash, int vmalloced, unsigned int size)
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(nf_ct_free_hashtable);
|
||||
|
||||
void nf_conntrack_flush(struct net *net, u32 pid, int report)
|
||||
void nf_conntrack_flush_report(struct net *net, u32 pid, int report)
|
||||
{
|
||||
struct __nf_ct_flush_report fr = {
|
||||
.pid = pid,
|
||||
.report = report,
|
||||
};
|
||||
nf_ct_iterate_cleanup(net, kill_all, &fr);
|
||||
nf_ct_iterate_cleanup(net, kill_report, &fr);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(nf_conntrack_flush);
|
||||
EXPORT_SYMBOL_GPL(nf_conntrack_flush_report);
|
||||
|
||||
static void nf_conntrack_cleanup_init_net(void)
|
||||
{
|
||||
@@ -1045,7 +1039,7 @@ static void nf_conntrack_cleanup_net(struct net *net)
|
||||
nf_ct_event_cache_flush(net);
|
||||
nf_conntrack_ecache_fini(net);
|
||||
i_see_dead_people:
|
||||
nf_conntrack_flush(net, 0, 0);
|
||||
nf_ct_iterate_cleanup(net, kill_all, NULL);
|
||||
if (atomic_read(&net->ct.count) != 0) {
|
||||
schedule();
|
||||
goto i_see_dead_people;
|
||||
|
||||
@@ -16,24 +16,32 @@
|
||||
#include <linux/stddef.h>
|
||||
#include <linux/err.h>
|
||||
#include <linux/percpu.h>
|
||||
#include <linux/notifier.h>
|
||||
#include <linux/kernel.h>
|
||||
#include <linux/netdevice.h>
|
||||
|
||||
#include <net/netfilter/nf_conntrack.h>
|
||||
#include <net/netfilter/nf_conntrack_core.h>
|
||||
|
||||
ATOMIC_NOTIFIER_HEAD(nf_conntrack_chain);
|
||||
EXPORT_SYMBOL_GPL(nf_conntrack_chain);
|
||||
static DEFINE_MUTEX(nf_ct_ecache_mutex);
|
||||
|
||||
ATOMIC_NOTIFIER_HEAD(nf_ct_expect_chain);
|
||||
EXPORT_SYMBOL_GPL(nf_ct_expect_chain);
|
||||
struct nf_ct_event_notifier *nf_conntrack_event_cb __read_mostly;
|
||||
EXPORT_SYMBOL_GPL(nf_conntrack_event_cb);
|
||||
|
||||
struct nf_exp_event_notifier *nf_expect_event_cb __read_mostly;
|
||||
EXPORT_SYMBOL_GPL(nf_expect_event_cb);
|
||||
|
||||
/* deliver cached events and clear cache entry - must be called with locally
|
||||
* disabled softirqs */
|
||||
static inline void
|
||||
__nf_ct_deliver_cached_events(struct nf_conntrack_ecache *ecache)
|
||||
{
|
||||
struct nf_ct_event_notifier *notify;
|
||||
|
||||
rcu_read_lock();
|
||||
notify = rcu_dereference(nf_conntrack_event_cb);
|
||||
if (notify == NULL)
|
||||
goto out_unlock;
|
||||
|
||||
if (nf_ct_is_confirmed(ecache->ct) && !nf_ct_is_dying(ecache->ct)
|
||||
&& ecache->events) {
|
||||
struct nf_ct_event item = {
|
||||
@@ -42,14 +50,15 @@ __nf_ct_deliver_cached_events(struct nf_conntrack_ecache *ecache)
|
||||
.report = 0
|
||||
};
|
||||
|
||||
atomic_notifier_call_chain(&nf_conntrack_chain,
|
||||
ecache->events,
|
||||
&item);
|
||||
notify->fcn(ecache->events, &item);
|
||||
}
|
||||
|
||||
ecache->events = 0;
|
||||
nf_ct_put(ecache->ct);
|
||||
ecache->ct = NULL;
|
||||
|
||||
out_unlock:
|
||||
rcu_read_unlock();
|
||||
}
|
||||
|
||||
/* Deliver all cached events for a particular conntrack. This is called
|
||||
@@ -111,26 +120,68 @@ void nf_conntrack_ecache_fini(struct net *net)
|
||||
free_percpu(net->ct.ecache);
|
||||
}
|
||||
|
||||
int nf_conntrack_register_notifier(struct notifier_block *nb)
|
||||
int nf_conntrack_register_notifier(struct nf_ct_event_notifier *new)
|
||||
{
|
||||
return atomic_notifier_chain_register(&nf_conntrack_chain, nb);
|
||||
int ret = 0;
|
||||
struct nf_ct_event_notifier *notify;
|
||||
|
||||
mutex_lock(&nf_ct_ecache_mutex);
|
||||
notify = rcu_dereference(nf_conntrack_event_cb);
|
||||
if (notify != NULL) {
|
||||
ret = -EBUSY;
|
||||
goto out_unlock;
|
||||
}
|
||||
rcu_assign_pointer(nf_conntrack_event_cb, new);
|
||||
mutex_unlock(&nf_ct_ecache_mutex);
|
||||
return ret;
|
||||
|
||||
out_unlock:
|
||||
mutex_unlock(&nf_ct_ecache_mutex);
|
||||
return ret;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(nf_conntrack_register_notifier);
|
||||
|
||||
int nf_conntrack_unregister_notifier(struct notifier_block *nb)
|
||||
void nf_conntrack_unregister_notifier(struct nf_ct_event_notifier *new)
|
||||
{
|
||||
return atomic_notifier_chain_unregister(&nf_conntrack_chain, nb);
|
||||
struct nf_ct_event_notifier *notify;
|
||||
|
||||
mutex_lock(&nf_ct_ecache_mutex);
|
||||
notify = rcu_dereference(nf_conntrack_event_cb);
|
||||
BUG_ON(notify != new);
|
||||
rcu_assign_pointer(nf_conntrack_event_cb, NULL);
|
||||
mutex_unlock(&nf_ct_ecache_mutex);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(nf_conntrack_unregister_notifier);
|
||||
|
||||
int nf_ct_expect_register_notifier(struct notifier_block *nb)
|
||||
int nf_ct_expect_register_notifier(struct nf_exp_event_notifier *new)
|
||||
{
|
||||
return atomic_notifier_chain_register(&nf_ct_expect_chain, nb);
|
||||
int ret = 0;
|
||||
struct nf_exp_event_notifier *notify;
|
||||
|
||||
mutex_lock(&nf_ct_ecache_mutex);
|
||||
notify = rcu_dereference(nf_expect_event_cb);
|
||||
if (notify != NULL) {
|
||||
ret = -EBUSY;
|
||||
goto out_unlock;
|
||||
}
|
||||
rcu_assign_pointer(nf_expect_event_cb, new);
|
||||
mutex_unlock(&nf_ct_ecache_mutex);
|
||||
return ret;
|
||||
|
||||
out_unlock:
|
||||
mutex_unlock(&nf_ct_ecache_mutex);
|
||||
return ret;
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(nf_ct_expect_register_notifier);
|
||||
|
||||
int nf_ct_expect_unregister_notifier(struct notifier_block *nb)
|
||||
void nf_ct_expect_unregister_notifier(struct nf_exp_event_notifier *new)
|
||||
{
|
||||
return atomic_notifier_chain_unregister(&nf_ct_expect_chain, nb);
|
||||
struct nf_exp_event_notifier *notify;
|
||||
|
||||
mutex_lock(&nf_ct_ecache_mutex);
|
||||
notify = rcu_dereference(nf_expect_event_cb);
|
||||
BUG_ON(notify != new);
|
||||
rcu_assign_pointer(nf_expect_event_cb, NULL);
|
||||
mutex_unlock(&nf_ct_ecache_mutex);
|
||||
}
|
||||
EXPORT_SYMBOL_GPL(nf_ct_expect_unregister_notifier);
|
||||
|
||||
@@ -338,11 +338,9 @@ static void update_nl_seq(struct nf_conn *ct, u32 nl_seq,
|
||||
|
||||
if (info->seq_aft_nl_num[dir] < NUM_SEQ_TO_REMEMBER) {
|
||||
info->seq_aft_nl[dir][info->seq_aft_nl_num[dir]++] = nl_seq;
|
||||
nf_conntrack_event_cache(IPCT_HELPINFO_VOLATILE, ct);
|
||||
} else if (oldest != NUM_SEQ_TO_REMEMBER &&
|
||||
after(nl_seq, info->seq_aft_nl[dir][oldest])) {
|
||||
info->seq_aft_nl[dir][oldest] = nl_seq;
|
||||
nf_conntrack_event_cache(IPCT_HELPINFO_VOLATILE, ct);
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
File diff suppressed because it is too large
Load Diff
@@ -25,8 +25,6 @@
|
||||
#include <net/netfilter/nf_conntrack_ecache.h>
|
||||
#include <net/netfilter/nf_log.h>
|
||||
|
||||
static DEFINE_RWLOCK(dccp_lock);
|
||||
|
||||
/* Timeouts are based on values from RFC4340:
|
||||
*
|
||||
* - REQUEST:
|
||||
@@ -492,7 +490,7 @@ static int dccp_packet(struct nf_conn *ct, const struct sk_buff *skb,
|
||||
return NF_ACCEPT;
|
||||
}
|
||||
|
||||
write_lock_bh(&dccp_lock);
|
||||
spin_lock_bh(&ct->lock);
|
||||
|
||||
role = ct->proto.dccp.role[dir];
|
||||
old_state = ct->proto.dccp.state;
|
||||
@@ -536,13 +534,13 @@ static int dccp_packet(struct nf_conn *ct, const struct sk_buff *skb,
|
||||
ct->proto.dccp.last_dir = dir;
|
||||
ct->proto.dccp.last_pkt = type;
|
||||
|
||||
write_unlock_bh(&dccp_lock);
|
||||
spin_unlock_bh(&ct->lock);
|
||||
if (LOG_INVALID(net, IPPROTO_DCCP))
|
||||
nf_log_packet(pf, 0, skb, NULL, NULL, NULL,
|
||||
"nf_ct_dccp: invalid packet ignored ");
|
||||
return NF_ACCEPT;
|
||||
case CT_DCCP_INVALID:
|
||||
write_unlock_bh(&dccp_lock);
|
||||
spin_unlock_bh(&ct->lock);
|
||||
if (LOG_INVALID(net, IPPROTO_DCCP))
|
||||
nf_log_packet(pf, 0, skb, NULL, NULL, NULL,
|
||||
"nf_ct_dccp: invalid state transition ");
|
||||
@@ -552,7 +550,7 @@ static int dccp_packet(struct nf_conn *ct, const struct sk_buff *skb,
|
||||
ct->proto.dccp.last_dir = dir;
|
||||
ct->proto.dccp.last_pkt = type;
|
||||
ct->proto.dccp.state = new_state;
|
||||
write_unlock_bh(&dccp_lock);
|
||||
spin_unlock_bh(&ct->lock);
|
||||
|
||||
if (new_state != old_state)
|
||||
nf_conntrack_event_cache(IPCT_PROTOINFO, ct);
|
||||
@@ -621,36 +619,39 @@ static int dccp_print_tuple(struct seq_file *s,
|
||||
ntohs(tuple->dst.u.dccp.port));
|
||||
}
|
||||
|
||||
static int dccp_print_conntrack(struct seq_file *s, const struct nf_conn *ct)
|
||||
static int dccp_print_conntrack(struct seq_file *s, struct nf_conn *ct)
|
||||
{
|
||||
return seq_printf(s, "%s ", dccp_state_names[ct->proto.dccp.state]);
|
||||
}
|
||||
|
||||
#if defined(CONFIG_NF_CT_NETLINK) || defined(CONFIG_NF_CT_NETLINK_MODULE)
|
||||
static int dccp_to_nlattr(struct sk_buff *skb, struct nlattr *nla,
|
||||
const struct nf_conn *ct)
|
||||
struct nf_conn *ct)
|
||||
{
|
||||
struct nlattr *nest_parms;
|
||||
|
||||
read_lock_bh(&dccp_lock);
|
||||
spin_lock_bh(&ct->lock);
|
||||
nest_parms = nla_nest_start(skb, CTA_PROTOINFO_DCCP | NLA_F_NESTED);
|
||||
if (!nest_parms)
|
||||
goto nla_put_failure;
|
||||
NLA_PUT_U8(skb, CTA_PROTOINFO_DCCP_STATE, ct->proto.dccp.state);
|
||||
NLA_PUT_U8(skb, CTA_PROTOINFO_DCCP_ROLE,
|
||||
ct->proto.dccp.role[IP_CT_DIR_ORIGINAL]);
|
||||
NLA_PUT_BE64(skb, CTA_PROTOINFO_DCCP_HANDSHAKE_SEQ,
|
||||
cpu_to_be64(ct->proto.dccp.handshake_seq));
|
||||
nla_nest_end(skb, nest_parms);
|
||||
read_unlock_bh(&dccp_lock);
|
||||
spin_unlock_bh(&ct->lock);
|
||||
return 0;
|
||||
|
||||
nla_put_failure:
|
||||
read_unlock_bh(&dccp_lock);
|
||||
spin_unlock_bh(&ct->lock);
|
||||
return -1;
|
||||
}
|
||||
|
||||
static const struct nla_policy dccp_nla_policy[CTA_PROTOINFO_DCCP_MAX + 1] = {
|
||||
[CTA_PROTOINFO_DCCP_STATE] = { .type = NLA_U8 },
|
||||
[CTA_PROTOINFO_DCCP_ROLE] = { .type = NLA_U8 },
|
||||
[CTA_PROTOINFO_DCCP_HANDSHAKE_SEQ] = { .type = NLA_U64 },
|
||||
};
|
||||
|
||||
static int nlattr_to_dccp(struct nlattr *cda[], struct nf_conn *ct)
|
||||
@@ -674,7 +675,7 @@ static int nlattr_to_dccp(struct nlattr *cda[], struct nf_conn *ct)
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
write_lock_bh(&dccp_lock);
|
||||
spin_lock_bh(&ct->lock);
|
||||
ct->proto.dccp.state = nla_get_u8(tb[CTA_PROTOINFO_DCCP_STATE]);
|
||||
if (nla_get_u8(tb[CTA_PROTOINFO_DCCP_ROLE]) == CT_DCCP_ROLE_CLIENT) {
|
||||
ct->proto.dccp.role[IP_CT_DIR_ORIGINAL] = CT_DCCP_ROLE_CLIENT;
|
||||
@@ -683,7 +684,11 @@ static int nlattr_to_dccp(struct nlattr *cda[], struct nf_conn *ct)
|
||||
ct->proto.dccp.role[IP_CT_DIR_ORIGINAL] = CT_DCCP_ROLE_SERVER;
|
||||
ct->proto.dccp.role[IP_CT_DIR_REPLY] = CT_DCCP_ROLE_CLIENT;
|
||||
}
|
||||
write_unlock_bh(&dccp_lock);
|
||||
if (tb[CTA_PROTOINFO_DCCP_HANDSHAKE_SEQ]) {
|
||||
ct->proto.dccp.handshake_seq =
|
||||
be64_to_cpu(nla_get_be64(tb[CTA_PROTOINFO_DCCP_HANDSHAKE_SEQ]));
|
||||
}
|
||||
spin_unlock_bh(&ct->lock);
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
||||
@@ -219,8 +219,7 @@ static int gre_print_tuple(struct seq_file *s,
|
||||
}
|
||||
|
||||
/* print private data for conntrack */
|
||||
static int gre_print_conntrack(struct seq_file *s,
|
||||
const struct nf_conn *ct)
|
||||
static int gre_print_conntrack(struct seq_file *s, struct nf_conn *ct)
|
||||
{
|
||||
return seq_printf(s, "timeout=%u, stream_timeout=%u ",
|
||||
(ct->proto.gre.timeout / HZ),
|
||||
|
||||
@@ -25,9 +25,6 @@
|
||||
#include <net/netfilter/nf_conntrack_l4proto.h>
|
||||
#include <net/netfilter/nf_conntrack_ecache.h>
|
||||
|
||||
/* Protects ct->proto.sctp */
|
||||
static DEFINE_RWLOCK(sctp_lock);
|
||||
|
||||
/* FIXME: Examine ipfilter's timeouts and conntrack transitions more
|
||||
closely. They're more complex. --RR
|
||||
|
||||
@@ -164,13 +161,13 @@ static int sctp_print_tuple(struct seq_file *s,
|
||||
}
|
||||
|
||||
/* Print out the private part of the conntrack. */
|
||||
static int sctp_print_conntrack(struct seq_file *s, const struct nf_conn *ct)
|
||||
static int sctp_print_conntrack(struct seq_file *s, struct nf_conn *ct)
|
||||
{
|
||||
enum sctp_conntrack state;
|
||||
|
||||
read_lock_bh(&sctp_lock);
|
||||
spin_lock_bh(&ct->lock);
|
||||
state = ct->proto.sctp.state;
|
||||
read_unlock_bh(&sctp_lock);
|
||||
spin_unlock_bh(&ct->lock);
|
||||
|
||||
return seq_printf(s, "%s ", sctp_conntrack_names[state]);
|
||||
}
|
||||
@@ -318,7 +315,7 @@ static int sctp_packet(struct nf_conn *ct,
|
||||
}
|
||||
|
||||
old_state = new_state = SCTP_CONNTRACK_NONE;
|
||||
write_lock_bh(&sctp_lock);
|
||||
spin_lock_bh(&ct->lock);
|
||||
for_each_sctp_chunk (skb, sch, _sch, offset, dataoff, count) {
|
||||
/* Special cases of Verification tag check (Sec 8.5.1) */
|
||||
if (sch->type == SCTP_CID_INIT) {
|
||||
@@ -371,7 +368,7 @@ static int sctp_packet(struct nf_conn *ct,
|
||||
if (old_state != new_state)
|
||||
nf_conntrack_event_cache(IPCT_PROTOINFO, ct);
|
||||
}
|
||||
write_unlock_bh(&sctp_lock);
|
||||
spin_unlock_bh(&ct->lock);
|
||||
|
||||
nf_ct_refresh_acct(ct, ctinfo, skb, sctp_timeouts[new_state]);
|
||||
|
||||
@@ -386,7 +383,7 @@ static int sctp_packet(struct nf_conn *ct,
|
||||
return NF_ACCEPT;
|
||||
|
||||
out_unlock:
|
||||
write_unlock_bh(&sctp_lock);
|
||||
spin_unlock_bh(&ct->lock);
|
||||
out:
|
||||
return -NF_ACCEPT;
|
||||
}
|
||||
@@ -469,11 +466,11 @@ static bool sctp_new(struct nf_conn *ct, const struct sk_buff *skb,
|
||||
#include <linux/netfilter/nfnetlink_conntrack.h>
|
||||
|
||||
static int sctp_to_nlattr(struct sk_buff *skb, struct nlattr *nla,
|
||||
const struct nf_conn *ct)
|
||||
struct nf_conn *ct)
|
||||
{
|
||||
struct nlattr *nest_parms;
|
||||
|
||||
read_lock_bh(&sctp_lock);
|
||||
spin_lock_bh(&ct->lock);
|
||||
nest_parms = nla_nest_start(skb, CTA_PROTOINFO_SCTP | NLA_F_NESTED);
|
||||
if (!nest_parms)
|
||||
goto nla_put_failure;
|
||||
@@ -488,14 +485,14 @@ static int sctp_to_nlattr(struct sk_buff *skb, struct nlattr *nla,
|
||||
CTA_PROTOINFO_SCTP_VTAG_REPLY,
|
||||
ct->proto.sctp.vtag[IP_CT_DIR_REPLY]);
|
||||
|
||||
read_unlock_bh(&sctp_lock);
|
||||
spin_unlock_bh(&ct->lock);
|
||||
|
||||
nla_nest_end(skb, nest_parms);
|
||||
|
||||
return 0;
|
||||
|
||||
nla_put_failure:
|
||||
read_unlock_bh(&sctp_lock);
|
||||
spin_unlock_bh(&ct->lock);
|
||||
return -1;
|
||||
}
|
||||
|
||||
@@ -527,13 +524,13 @@ static int nlattr_to_sctp(struct nlattr *cda[], struct nf_conn *ct)
|
||||
!tb[CTA_PROTOINFO_SCTP_VTAG_REPLY])
|
||||
return -EINVAL;
|
||||
|
||||
write_lock_bh(&sctp_lock);
|
||||
spin_lock_bh(&ct->lock);
|
||||
ct->proto.sctp.state = nla_get_u8(tb[CTA_PROTOINFO_SCTP_STATE]);
|
||||
ct->proto.sctp.vtag[IP_CT_DIR_ORIGINAL] =
|
||||
nla_get_be32(tb[CTA_PROTOINFO_SCTP_VTAG_ORIGINAL]);
|
||||
ct->proto.sctp.vtag[IP_CT_DIR_REPLY] =
|
||||
nla_get_be32(tb[CTA_PROTOINFO_SCTP_VTAG_REPLY]);
|
||||
write_unlock_bh(&sctp_lock);
|
||||
spin_unlock_bh(&ct->lock);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
@@ -29,9 +29,6 @@
|
||||
#include <net/netfilter/ipv4/nf_conntrack_ipv4.h>
|
||||
#include <net/netfilter/ipv6/nf_conntrack_ipv6.h>
|
||||
|
||||
/* Protects ct->proto.tcp */
|
||||
static DEFINE_RWLOCK(tcp_lock);
|
||||
|
||||
/* "Be conservative in what you do,
|
||||
be liberal in what you accept from others."
|
||||
If it's non-zero, we mark only out of window RST segments as INVALID. */
|
||||
@@ -59,7 +56,7 @@ static const char *const tcp_conntrack_names[] = {
|
||||
"LAST_ACK",
|
||||
"TIME_WAIT",
|
||||
"CLOSE",
|
||||
"LISTEN"
|
||||
"SYN_SENT2",
|
||||
};
|
||||
|
||||
#define SECS * HZ
|
||||
@@ -82,6 +79,7 @@ static unsigned int tcp_timeouts[TCP_CONNTRACK_MAX] __read_mostly = {
|
||||
[TCP_CONNTRACK_LAST_ACK] = 30 SECS,
|
||||
[TCP_CONNTRACK_TIME_WAIT] = 2 MINS,
|
||||
[TCP_CONNTRACK_CLOSE] = 10 SECS,
|
||||
[TCP_CONNTRACK_SYN_SENT2] = 2 MINS,
|
||||
};
|
||||
|
||||
#define sNO TCP_CONNTRACK_NONE
|
||||
@@ -93,7 +91,7 @@ static unsigned int tcp_timeouts[TCP_CONNTRACK_MAX] __read_mostly = {
|
||||
#define sLA TCP_CONNTRACK_LAST_ACK
|
||||
#define sTW TCP_CONNTRACK_TIME_WAIT
|
||||
#define sCL TCP_CONNTRACK_CLOSE
|
||||
#define sLI TCP_CONNTRACK_LISTEN
|
||||
#define sS2 TCP_CONNTRACK_SYN_SENT2
|
||||
#define sIV TCP_CONNTRACK_MAX
|
||||
#define sIG TCP_CONNTRACK_IGNORE
|
||||
|
||||
@@ -123,6 +121,7 @@ enum tcp_bit_set {
|
||||
*
|
||||
* NONE: initial state
|
||||
* SYN_SENT: SYN-only packet seen
|
||||
* SYN_SENT2: SYN-only packet seen from reply dir, simultaneous open
|
||||
* SYN_RECV: SYN-ACK packet seen
|
||||
* ESTABLISHED: ACK packet seen
|
||||
* FIN_WAIT: FIN packet seen
|
||||
@@ -131,26 +130,24 @@ enum tcp_bit_set {
|
||||
* TIME_WAIT: last ACK seen
|
||||
* CLOSE: closed connection (RST)
|
||||
*
|
||||
* LISTEN state is not used.
|
||||
*
|
||||
* Packets marked as IGNORED (sIG):
|
||||
* if they may be either invalid or valid
|
||||
* and the receiver may send back a connection
|
||||
* closing RST or a SYN/ACK.
|
||||
*
|
||||
* Packets marked as INVALID (sIV):
|
||||
* if they are invalid
|
||||
* or we do not support the request (simultaneous open)
|
||||
* if we regard them as truly invalid packets
|
||||
*/
|
||||
static const u8 tcp_conntracks[2][6][TCP_CONNTRACK_MAX] = {
|
||||
{
|
||||
/* ORIGINAL */
|
||||
/* sNO, sSS, sSR, sES, sFW, sCW, sLA, sTW, sCL, sLI */
|
||||
/*syn*/ { sSS, sSS, sIG, sIG, sIG, sIG, sIG, sSS, sSS, sIV },
|
||||
/* sNO, sSS, sSR, sES, sFW, sCW, sLA, sTW, sCL, sS2 */
|
||||
/*syn*/ { sSS, sSS, sIG, sIG, sIG, sIG, sIG, sSS, sSS, sS2 },
|
||||
/*
|
||||
* sNO -> sSS Initialize a new connection
|
||||
* sSS -> sSS Retransmitted SYN
|
||||
* sSR -> sIG Late retransmitted SYN?
|
||||
* sS2 -> sS2 Late retransmitted SYN
|
||||
* sSR -> sIG
|
||||
* sES -> sIG Error: SYNs in window outside the SYN_SENT state
|
||||
* are errors. Receiver will reply with RST
|
||||
* and close the connection.
|
||||
@@ -161,22 +158,30 @@ static const u8 tcp_conntracks[2][6][TCP_CONNTRACK_MAX] = {
|
||||
* sTW -> sSS Reopened connection (RFC 1122).
|
||||
* sCL -> sSS
|
||||
*/
|
||||
/* sNO, sSS, sSR, sES, sFW, sCW, sLA, sTW, sCL, sLI */
|
||||
/*synack*/ { sIV, sIV, sIV, sIV, sIV, sIV, sIV, sIV, sIV, sIV },
|
||||
/* sNO, sSS, sSR, sES, sFW, sCW, sLA, sTW, sCL, sS2 */
|
||||
/*synack*/ { sIV, sIV, sIG, sIG, sIG, sIG, sIG, sIG, sIG, sSR },
|
||||
/*
|
||||
* A SYN/ACK from the client is always invalid:
|
||||
* - either it tries to set up a simultaneous open, which is
|
||||
* not supported;
|
||||
* - or the firewall has just been inserted between the two hosts
|
||||
* during the session set-up. The SYN will be retransmitted
|
||||
* by the true client (or it'll time out).
|
||||
* sNO -> sIV Too late and no reason to do anything
|
||||
* sSS -> sIV Client can't send SYN and then SYN/ACK
|
||||
* sS2 -> sSR SYN/ACK sent to SYN2 in simultaneous open
|
||||
* sSR -> sIG
|
||||
* sES -> sIG Error: SYNs in window outside the SYN_SENT state
|
||||
* are errors. Receiver will reply with RST
|
||||
* and close the connection.
|
||||
* Or we are not in sync and hold a dead connection.
|
||||
* sFW -> sIG
|
||||
* sCW -> sIG
|
||||
* sLA -> sIG
|
||||
* sTW -> sIG
|
||||
* sCL -> sIG
|
||||
*/
|
||||
/* sNO, sSS, sSR, sES, sFW, sCW, sLA, sTW, sCL, sLI */
|
||||
/* sNO, sSS, sSR, sES, sFW, sCW, sLA, sTW, sCL, sS2 */
|
||||
/*fin*/ { sIV, sIV, sFW, sFW, sLA, sLA, sLA, sTW, sCL, sIV },
|
||||
/*
|
||||
* sNO -> sIV Too late and no reason to do anything...
|
||||
* sSS -> sIV Client migth not send FIN in this state:
|
||||
* we enforce waiting for a SYN/ACK reply first.
|
||||
* sS2 -> sIV
|
||||
* sSR -> sFW Close started.
|
||||
* sES -> sFW
|
||||
* sFW -> sLA FIN seen in both directions, waiting for
|
||||
@@ -187,11 +192,12 @@ static const u8 tcp_conntracks[2][6][TCP_CONNTRACK_MAX] = {
|
||||
* sTW -> sTW
|
||||
* sCL -> sCL
|
||||
*/
|
||||
/* sNO, sSS, sSR, sES, sFW, sCW, sLA, sTW, sCL, sLI */
|
||||
/* sNO, sSS, sSR, sES, sFW, sCW, sLA, sTW, sCL, sS2 */
|
||||
/*ack*/ { sES, sIV, sES, sES, sCW, sCW, sTW, sTW, sCL, sIV },
|
||||
/*
|
||||
* sNO -> sES Assumed.
|
||||
* sSS -> sIV ACK is invalid: we haven't seen a SYN/ACK yet.
|
||||
* sS2 -> sIV
|
||||
* sSR -> sES Established state is reached.
|
||||
* sES -> sES :-)
|
||||
* sFW -> sCW Normal close request answered by ACK.
|
||||
@@ -200,29 +206,31 @@ static const u8 tcp_conntracks[2][6][TCP_CONNTRACK_MAX] = {
|
||||
* sTW -> sTW Retransmitted last ACK. Remain in the same state.
|
||||
* sCL -> sCL
|
||||
*/
|
||||
/* sNO, sSS, sSR, sES, sFW, sCW, sLA, sTW, sCL, sLI */
|
||||
/*rst*/ { sIV, sCL, sCL, sCL, sCL, sCL, sCL, sCL, sCL, sIV },
|
||||
/* sNO, sSS, sSR, sES, sFW, sCW, sLA, sTW, sCL, sS2 */
|
||||
/*rst*/ { sIV, sCL, sCL, sCL, sCL, sCL, sCL, sCL, sCL, sCL },
|
||||
/*none*/ { sIV, sIV, sIV, sIV, sIV, sIV, sIV, sIV, sIV, sIV }
|
||||
},
|
||||
{
|
||||
/* REPLY */
|
||||
/* sNO, sSS, sSR, sES, sFW, sCW, sLA, sTW, sCL, sLI */
|
||||
/*syn*/ { sIV, sIV, sIV, sIV, sIV, sIV, sIV, sIV, sIV, sIV },
|
||||
/* sNO, sSS, sSR, sES, sFW, sCW, sLA, sTW, sCL, sS2 */
|
||||
/*syn*/ { sIV, sS2, sIV, sIV, sIV, sIV, sIV, sIV, sIV, sS2 },
|
||||
/*
|
||||
* sNO -> sIV Never reached.
|
||||
* sSS -> sIV Simultaneous open, not supported
|
||||
* sSR -> sIV Simultaneous open, not supported.
|
||||
* sES -> sIV Server may not initiate a connection.
|
||||
* sSS -> sS2 Simultaneous open
|
||||
* sS2 -> sS2 Retransmitted simultaneous SYN
|
||||
* sSR -> sIV Invalid SYN packets sent by the server
|
||||
* sES -> sIV
|
||||
* sFW -> sIV
|
||||
* sCW -> sIV
|
||||
* sLA -> sIV
|
||||
* sTW -> sIV Reopened connection, but server may not do it.
|
||||
* sCL -> sIV
|
||||
*/
|
||||
/* sNO, sSS, sSR, sES, sFW, sCW, sLA, sTW, sCL, sLI */
|
||||
/*synack*/ { sIV, sSR, sSR, sIG, sIG, sIG, sIG, sIG, sIG, sIV },
|
||||
/* sNO, sSS, sSR, sES, sFW, sCW, sLA, sTW, sCL, sS2 */
|
||||
/*synack*/ { sIV, sSR, sSR, sIG, sIG, sIG, sIG, sIG, sIG, sSR },
|
||||
/*
|
||||
* sSS -> sSR Standard open.
|
||||
* sS2 -> sSR Simultaneous open
|
||||
* sSR -> sSR Retransmitted SYN/ACK.
|
||||
* sES -> sIG Late retransmitted SYN/ACK?
|
||||
* sFW -> sIG Might be SYN/ACK answering ignored SYN
|
||||
@@ -231,10 +239,11 @@ static const u8 tcp_conntracks[2][6][TCP_CONNTRACK_MAX] = {
|
||||
* sTW -> sIG
|
||||
* sCL -> sIG
|
||||
*/
|
||||
/* sNO, sSS, sSR, sES, sFW, sCW, sLA, sTW, sCL, sLI */
|
||||
/* sNO, sSS, sSR, sES, sFW, sCW, sLA, sTW, sCL, sS2 */
|
||||
/*fin*/ { sIV, sIV, sFW, sFW, sLA, sLA, sLA, sTW, sCL, sIV },
|
||||
/*
|
||||
* sSS -> sIV Server might not send FIN in this state.
|
||||
* sS2 -> sIV
|
||||
* sSR -> sFW Close started.
|
||||
* sES -> sFW
|
||||
* sFW -> sLA FIN seen in both directions.
|
||||
@@ -243,10 +252,11 @@ static const u8 tcp_conntracks[2][6][TCP_CONNTRACK_MAX] = {
|
||||
* sTW -> sTW
|
||||
* sCL -> sCL
|
||||
*/
|
||||
/* sNO, sSS, sSR, sES, sFW, sCW, sLA, sTW, sCL, sLI */
|
||||
/*ack*/ { sIV, sIG, sSR, sES, sCW, sCW, sTW, sTW, sCL, sIV },
|
||||
/* sNO, sSS, sSR, sES, sFW, sCW, sLA, sTW, sCL, sS2 */
|
||||
/*ack*/ { sIV, sIG, sSR, sES, sCW, sCW, sTW, sTW, sCL, sIG },
|
||||
/*
|
||||
* sSS -> sIG Might be a half-open connection.
|
||||
* sS2 -> sIG
|
||||
* sSR -> sSR Might answer late resent SYN.
|
||||
* sES -> sES :-)
|
||||
* sFW -> sCW Normal close request answered by ACK.
|
||||
@@ -255,8 +265,8 @@ static const u8 tcp_conntracks[2][6][TCP_CONNTRACK_MAX] = {
|
||||
* sTW -> sTW Retransmitted last ACK.
|
||||
* sCL -> sCL
|
||||
*/
|
||||
/* sNO, sSS, sSR, sES, sFW, sCW, sLA, sTW, sCL, sLI */
|
||||
/*rst*/ { sIV, sCL, sCL, sCL, sCL, sCL, sCL, sCL, sCL, sIV },
|
||||
/* sNO, sSS, sSR, sES, sFW, sCW, sLA, sTW, sCL, sS2 */
|
||||
/*rst*/ { sIV, sCL, sCL, sCL, sCL, sCL, sCL, sCL, sCL, sCL },
|
||||
/*none*/ { sIV, sIV, sIV, sIV, sIV, sIV, sIV, sIV, sIV, sIV }
|
||||
}
|
||||
};
|
||||
@@ -296,13 +306,13 @@ static int tcp_print_tuple(struct seq_file *s,
|
||||
}
|
||||
|
||||
/* Print out the private part of the conntrack. */
|
||||
static int tcp_print_conntrack(struct seq_file *s, const struct nf_conn *ct)
|
||||
static int tcp_print_conntrack(struct seq_file *s, struct nf_conn *ct)
|
||||
{
|
||||
enum tcp_conntrack state;
|
||||
|
||||
read_lock_bh(&tcp_lock);
|
||||
spin_lock_bh(&ct->lock);
|
||||
state = ct->proto.tcp.state;
|
||||
read_unlock_bh(&tcp_lock);
|
||||
spin_unlock_bh(&ct->lock);
|
||||
|
||||
return seq_printf(s, "%s ", tcp_conntrack_names[state]);
|
||||
}
|
||||
@@ -521,13 +531,14 @@ static bool tcp_in_window(const struct nf_conn *ct,
|
||||
receiver->td_end, receiver->td_maxend, receiver->td_maxwin,
|
||||
receiver->td_scale);
|
||||
|
||||
if (sender->td_end == 0) {
|
||||
if (sender->td_maxwin == 0) {
|
||||
/*
|
||||
* Initialize sender data.
|
||||
*/
|
||||
if (tcph->syn && tcph->ack) {
|
||||
if (tcph->syn) {
|
||||
/*
|
||||
* Outgoing SYN-ACK in reply to a SYN.
|
||||
* SYN-ACK in reply to a SYN
|
||||
* or SYN from reply direction in simultaneous open.
|
||||
*/
|
||||
sender->td_end =
|
||||
sender->td_maxend = end;
|
||||
@@ -543,6 +554,9 @@ static bool tcp_in_window(const struct nf_conn *ct,
|
||||
&& receiver->flags & IP_CT_TCP_FLAG_WINDOW_SCALE))
|
||||
sender->td_scale =
|
||||
receiver->td_scale = 0;
|
||||
if (!tcph->ack)
|
||||
/* Simultaneous open */
|
||||
return true;
|
||||
} else {
|
||||
/*
|
||||
* We are in the middle of a connection,
|
||||
@@ -716,14 +730,14 @@ void nf_conntrack_tcp_update(const struct sk_buff *skb,
|
||||
|
||||
end = segment_seq_plus_len(ntohl(tcph->seq), skb->len, dataoff, tcph);
|
||||
|
||||
write_lock_bh(&tcp_lock);
|
||||
spin_lock_bh(&ct->lock);
|
||||
/*
|
||||
* We have to worry for the ack in the reply packet only...
|
||||
*/
|
||||
if (after(end, ct->proto.tcp.seen[dir].td_end))
|
||||
ct->proto.tcp.seen[dir].td_end = end;
|
||||
ct->proto.tcp.last_end = end;
|
||||
write_unlock_bh(&tcp_lock);
|
||||
spin_unlock_bh(&ct->lock);
|
||||
pr_debug("tcp_update: sender end=%u maxend=%u maxwin=%u scale=%i "
|
||||
"receiver end=%u maxend=%u maxwin=%u scale=%i\n",
|
||||
sender->td_end, sender->td_maxend, sender->td_maxwin,
|
||||
@@ -832,7 +846,7 @@ static int tcp_packet(struct nf_conn *ct,
|
||||
th = skb_header_pointer(skb, dataoff, sizeof(_tcph), &_tcph);
|
||||
BUG_ON(th == NULL);
|
||||
|
||||
write_lock_bh(&tcp_lock);
|
||||
spin_lock_bh(&ct->lock);
|
||||
old_state = ct->proto.tcp.state;
|
||||
dir = CTINFO2DIR(ctinfo);
|
||||
index = get_conntrack_index(th);
|
||||
@@ -862,7 +876,7 @@ static int tcp_packet(struct nf_conn *ct,
|
||||
&& ct->proto.tcp.last_index == TCP_RST_SET)) {
|
||||
/* Attempt to reopen a closed/aborted connection.
|
||||
* Delete this connection and look up again. */
|
||||
write_unlock_bh(&tcp_lock);
|
||||
spin_unlock_bh(&ct->lock);
|
||||
|
||||
/* Only repeat if we can actually remove the timer.
|
||||
* Destruction may already be in progress in process
|
||||
@@ -898,7 +912,7 @@ static int tcp_packet(struct nf_conn *ct,
|
||||
* that the client cannot but retransmit its SYN and
|
||||
* thus initiate a clean new session.
|
||||
*/
|
||||
write_unlock_bh(&tcp_lock);
|
||||
spin_unlock_bh(&ct->lock);
|
||||
if (LOG_INVALID(net, IPPROTO_TCP))
|
||||
nf_log_packet(pf, 0, skb, NULL, NULL, NULL,
|
||||
"nf_ct_tcp: killing out of sync session ");
|
||||
@@ -911,7 +925,7 @@ static int tcp_packet(struct nf_conn *ct,
|
||||
ct->proto.tcp.last_end =
|
||||
segment_seq_plus_len(ntohl(th->seq), skb->len, dataoff, th);
|
||||
|
||||
write_unlock_bh(&tcp_lock);
|
||||
spin_unlock_bh(&ct->lock);
|
||||
if (LOG_INVALID(net, IPPROTO_TCP))
|
||||
nf_log_packet(pf, 0, skb, NULL, NULL, NULL,
|
||||
"nf_ct_tcp: invalid packet ignored ");
|
||||
@@ -920,7 +934,7 @@ static int tcp_packet(struct nf_conn *ct,
|
||||
/* Invalid packet */
|
||||
pr_debug("nf_ct_tcp: Invalid dir=%i index=%u ostate=%u\n",
|
||||
dir, get_conntrack_index(th), old_state);
|
||||
write_unlock_bh(&tcp_lock);
|
||||
spin_unlock_bh(&ct->lock);
|
||||
if (LOG_INVALID(net, IPPROTO_TCP))
|
||||
nf_log_packet(pf, 0, skb, NULL, NULL, NULL,
|
||||
"nf_ct_tcp: invalid state ");
|
||||
@@ -930,7 +944,7 @@ static int tcp_packet(struct nf_conn *ct,
|
||||
&& (ct->proto.tcp.seen[!dir].flags & IP_CT_TCP_FLAG_MAXACK_SET)
|
||||
&& before(ntohl(th->seq), ct->proto.tcp.seen[!dir].td_maxack)) {
|
||||
/* Invalid RST */
|
||||
write_unlock_bh(&tcp_lock);
|
||||
spin_unlock_bh(&ct->lock);
|
||||
if (LOG_INVALID(net, IPPROTO_TCP))
|
||||
nf_log_packet(pf, 0, skb, NULL, NULL, NULL,
|
||||
"nf_ct_tcp: invalid RST ");
|
||||
@@ -961,7 +975,7 @@ static int tcp_packet(struct nf_conn *ct,
|
||||
|
||||
if (!tcp_in_window(ct, &ct->proto.tcp, dir, index,
|
||||
skb, dataoff, th, pf)) {
|
||||
write_unlock_bh(&tcp_lock);
|
||||
spin_unlock_bh(&ct->lock);
|
||||
return -NF_ACCEPT;
|
||||
}
|
||||
in_window:
|
||||
@@ -990,9 +1004,8 @@ static int tcp_packet(struct nf_conn *ct,
|
||||
timeout = nf_ct_tcp_timeout_unacknowledged;
|
||||
else
|
||||
timeout = tcp_timeouts[new_state];
|
||||
write_unlock_bh(&tcp_lock);
|
||||
spin_unlock_bh(&ct->lock);
|
||||
|
||||
nf_conntrack_event_cache(IPCT_PROTOINFO_VOLATILE, ct);
|
||||
if (new_state != old_state)
|
||||
nf_conntrack_event_cache(IPCT_PROTOINFO, ct);
|
||||
|
||||
@@ -1086,7 +1099,7 @@ static bool tcp_new(struct nf_conn *ct, const struct sk_buff *skb,
|
||||
|
||||
ct->proto.tcp.seen[1].td_end = 0;
|
||||
ct->proto.tcp.seen[1].td_maxend = 0;
|
||||
ct->proto.tcp.seen[1].td_maxwin = 1;
|
||||
ct->proto.tcp.seen[1].td_maxwin = 0;
|
||||
ct->proto.tcp.seen[1].td_scale = 0;
|
||||
|
||||
/* tcp_packet will set them */
|
||||
@@ -1108,12 +1121,12 @@ static bool tcp_new(struct nf_conn *ct, const struct sk_buff *skb,
|
||||
#include <linux/netfilter/nfnetlink_conntrack.h>
|
||||
|
||||
static int tcp_to_nlattr(struct sk_buff *skb, struct nlattr *nla,
|
||||
const struct nf_conn *ct)
|
||||
struct nf_conn *ct)
|
||||
{
|
||||
struct nlattr *nest_parms;
|
||||
struct nf_ct_tcp_flags tmp = {};
|
||||
|
||||
read_lock_bh(&tcp_lock);
|
||||
spin_lock_bh(&ct->lock);
|
||||
nest_parms = nla_nest_start(skb, CTA_PROTOINFO_TCP | NLA_F_NESTED);
|
||||
if (!nest_parms)
|
||||
goto nla_put_failure;
|
||||
@@ -1133,14 +1146,14 @@ static int tcp_to_nlattr(struct sk_buff *skb, struct nlattr *nla,
|
||||
tmp.flags = ct->proto.tcp.seen[1].flags;
|
||||
NLA_PUT(skb, CTA_PROTOINFO_TCP_FLAGS_REPLY,
|
||||
sizeof(struct nf_ct_tcp_flags), &tmp);
|
||||
read_unlock_bh(&tcp_lock);
|
||||
spin_unlock_bh(&ct->lock);
|
||||
|
||||
nla_nest_end(skb, nest_parms);
|
||||
|
||||
return 0;
|
||||
|
||||
nla_put_failure:
|
||||
read_unlock_bh(&tcp_lock);
|
||||
spin_unlock_bh(&ct->lock);
|
||||
return -1;
|
||||
}
|
||||
|
||||
@@ -1171,7 +1184,7 @@ static int nlattr_to_tcp(struct nlattr *cda[], struct nf_conn *ct)
|
||||
nla_get_u8(tb[CTA_PROTOINFO_TCP_STATE]) >= TCP_CONNTRACK_MAX)
|
||||
return -EINVAL;
|
||||
|
||||
write_lock_bh(&tcp_lock);
|
||||
spin_lock_bh(&ct->lock);
|
||||
if (tb[CTA_PROTOINFO_TCP_STATE])
|
||||
ct->proto.tcp.state = nla_get_u8(tb[CTA_PROTOINFO_TCP_STATE]);
|
||||
|
||||
@@ -1198,7 +1211,7 @@ static int nlattr_to_tcp(struct nlattr *cda[], struct nf_conn *ct)
|
||||
ct->proto.tcp.seen[1].td_scale =
|
||||
nla_get_u8(tb[CTA_PROTOINFO_TCP_WSCALE_REPLY]);
|
||||
}
|
||||
write_unlock_bh(&tcp_lock);
|
||||
spin_unlock_bh(&ct->lock);
|
||||
|
||||
return 0;
|
||||
}
|
||||
@@ -1327,6 +1340,13 @@ static struct ctl_table tcp_compat_sysctl_table[] = {
|
||||
.mode = 0644,
|
||||
.proc_handler = proc_dointvec_jiffies,
|
||||
},
|
||||
{
|
||||
.procname = "ip_conntrack_tcp_timeout_syn_sent2",
|
||||
.data = &tcp_timeouts[TCP_CONNTRACK_SYN_SENT2],
|
||||
.maxlen = sizeof(unsigned int),
|
||||
.mode = 0644,
|
||||
.proc_handler = proc_dointvec_jiffies,
|
||||
},
|
||||
{
|
||||
.procname = "ip_conntrack_tcp_timeout_syn_recv",
|
||||
.data = &tcp_timeouts[TCP_CONNTRACK_SYN_RECV],
|
||||
|
||||
@@ -204,10 +204,10 @@ int nf_queue(struct sk_buff *skb,
|
||||
queuenum);
|
||||
|
||||
switch (pf) {
|
||||
case AF_INET:
|
||||
case NFPROTO_IPV4:
|
||||
skb->protocol = htons(ETH_P_IP);
|
||||
break;
|
||||
case AF_INET6:
|
||||
case NFPROTO_IPV6:
|
||||
skb->protocol = htons(ETH_P_IPV6);
|
||||
break;
|
||||
}
|
||||
|
||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user