Message ID | 1354613751-30154-1-git-send-email-fw@strlen.de |
---|---|
State | Superseded |
Headers | show |
Hi Florian, On Tue, Dec 04, 2012 at 10:35:51AM +0100, Florian Westphal wrote: > This brings the (per-conntrack) ecache extension back to 24 bytes in > size (was 112 byte on x86_64 with lockdep on). > > Instead we use a per-ns tasklet to re-trigger event delivery. When we > enqueue a ct entry into the dying list, the tasklet is scheduled. > > The tasklet will then deliver up to 20 entries. It will re-sched > itself unless all the pending events could be delivered. > > While at it, dying list handling is moved into ecache.c, since its only > revlevant if ct events are enabled. Just tested this. My testbed consists of two firewalls in HA running conntrackd with event reliable mode. I've got a client that generates lots of small TCP flows that goes through the firewalls and reach a benchmark server. This is my analysis: conntrack -C shows: 253244 254798 256754 258807 261548 <--- we hit table full, dropping packets 176849 <--- it seems the tasklet gets a chance to run given that we get less interruptions from the NIC 166449 <--- it slightly empty the dying list 131176 55602 28316 28317 28317 # hits hits/s ^h/s ^bytes kB/s errs rst tout mhtime 4796894 15727 16509 2393805 2227 0 0 0 0.005 4813038 15728 16144 2340880 2227 0 0 0 0.005 4828796 15728 15758 2284910 2227 0 0 0 0.005 4845279 15731 16483 2390035 2227 0 0 0 0.005 4860956 15731 15677 2273165 2227 0 0 0 0.005 4876826 15731 15870 2301150 2227 0 0 0 0.005 4883165 15701 6339 919155 2223 0 0 0 0.004 4883165 15651 0 0 2216 0 0 0 0.000 <--- table full 4883165 15601 0 0 2209 0 0 0 0.000 4894657 15588 11492 1666340 2207 0 0 0 3.008 4913408 15598 18751 2718895 2208 0 0 0 0.004 4931896 15607 18488 2680760 2210 0 0 0 0.004 So it seems the tasklet gets starved under heavy load. This happens on and on, so after some time we hit table full and again the dying list is empty. These are old HP proliant DL145G2 from 2005, that's why the maximum flows/s looks low. Looking at the number and the behaviour under heavy stress, I think we have to consider a different approach. -- To unsubscribe from this list: send the line "unsubscribe netfilter-devel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
Pablo Neira Ayuso <pablo@netfilter.org> wrote: > On Tue, Dec 04, 2012 at 10:35:51AM +0100, Florian Westphal wrote: > > This brings the (per-conntrack) ecache extension back to 24 bytes in > > size (was 112 byte on x86_64 with lockdep on). > > > > Instead we use a per-ns tasklet to re-trigger event delivery. When we > > enqueue a ct entry into the dying list, the tasklet is scheduled. > > > > The tasklet will then deliver up to 20 entries. It will re-sched > > itself unless all the pending events could be delivered. > > > > While at it, dying list handling is moved into ecache.c, since its only > > revlevant if ct events are enabled. > > Just tested this. My testbed consists of two firewalls in HA running > conntrackd with event reliable mode. I've got a client that generates > lots of small TCP flows that goes through the firewalls and reach a > benchmark server. > > This is my analysis: > > conntrack -C shows: [..] > 261548 <--- we hit table full, dropping packets > 176849 <--- it seems the tasklet gets a chance to run > given that we get less interruptions from the NIC > 166449 <--- it slightly empty the dying list > 131176 > 55602 > 28316 [..] > # hits hits/s ^h/s ^bytes kB/s errs rst tout mhtime > 4796894 15727 16509 2393805 2227 0 0 0 0.005 > 4813038 15728 16144 2340880 2227 0 0 0 0.005 > 4828796 15728 15758 2284910 2227 0 0 0 0.005 > 4845279 15731 16483 2390035 2227 0 0 0 0.005 > 4860956 15731 15677 2273165 2227 0 0 0 0.005 > 4876826 15731 15870 2301150 2227 0 0 0 0.005 > 4883165 15701 6339 919155 2223 0 0 0 0.004 > 4883165 15651 0 0 2216 0 0 0 0.000 <--- table full > 4883165 15601 0 0 2209 0 0 0 0.000 > 4894657 15588 11492 1666340 2207 0 0 0 3.008 > 4913408 15598 18751 2718895 2208 0 0 0 0.004 > 4931896 15607 18488 2680760 2210 0 0 0 0.004 > > So it seems the tasklet gets starved under heavy load. > > This happens on and on, so after some time we hit table full and again > the dying list is empty. > > These are old HP proliant DL145G2 from 2005, that's why the maximum > flows/s looks low. > > Looking at the number and the behaviour under heavy stress, I think we > have to consider a different approach. Thanks for testing. Is that a single cpu machine? If yes, I think this result might be because the tasklet busy-loop competes with conntrackd for cpu, so essentially we waste cycles on futile re-delivery instead of leaving the cpu to conntrackd, (which should process events). If thats true, then we might be able to improve this by avoiding the 'tasklet re-scheds itself'. This would also solve the 'softirqd eats 100% cpu' when conntrackd is stopped/suspended. I'll see if I can cook up a patch some time tomorrow. Thanks, Florian -- To unsubscribe from this list: send the line "unsubscribe netfilter-devel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
On Tue, Dec 04, 2012 at 04:41:18PM +0100, Florian Westphal wrote: > Pablo Neira Ayuso <pablo@netfilter.org> wrote: > > On Tue, Dec 04, 2012 at 10:35:51AM +0100, Florian Westphal wrote: > > > This brings the (per-conntrack) ecache extension back to 24 bytes in > > > size (was 112 byte on x86_64 with lockdep on). > > > > > > Instead we use a per-ns tasklet to re-trigger event delivery. When we > > > enqueue a ct entry into the dying list, the tasklet is scheduled. > > > > > > The tasklet will then deliver up to 20 entries. It will re-sched > > > itself unless all the pending events could be delivered. > > > > > > While at it, dying list handling is moved into ecache.c, since its only > > > revlevant if ct events are enabled. > > > > Just tested this. My testbed consists of two firewalls in HA running > > conntrackd with event reliable mode. I've got a client that generates > > lots of small TCP flows that goes through the firewalls and reach a > > benchmark server. > > > > This is my analysis: > > > > conntrack -C shows: > [..] > > 261548 <--- we hit table full, dropping packets > > 176849 <--- it seems the tasklet gets a chance to run > > given that we get less interruptions from the NIC > > 166449 <--- it slightly empty the dying list > > 131176 > > 55602 > > 28316 > [..] > > # hits hits/s ^h/s ^bytes kB/s errs rst tout mhtime > > 4796894 15727 16509 2393805 2227 0 0 0 0.005 > > 4813038 15728 16144 2340880 2227 0 0 0 0.005 > > 4828796 15728 15758 2284910 2227 0 0 0 0.005 > > 4845279 15731 16483 2390035 2227 0 0 0 0.005 > > 4860956 15731 15677 2273165 2227 0 0 0 0.005 > > 4876826 15731 15870 2301150 2227 0 0 0 0.005 > > 4883165 15701 6339 919155 2223 0 0 0 0.004 > > 4883165 15651 0 0 2216 0 0 0 0.000 <--- table full > > 4883165 15601 0 0 2209 0 0 0 0.000 > > 4894657 15588 11492 1666340 2207 0 0 0 3.008 > > 4913408 15598 18751 2718895 2208 0 0 0 0.004 > > 4931896 15607 18488 2680760 2210 0 0 0 0.004 > > > > So it seems the tasklet gets starved under heavy load. > > > > This happens on and on, so after some time we hit table full and again > > the dying list is empty. > > > > These are old HP proliant DL145G2 from 2005, that's why the maximum > > flows/s looks low. > > > > Looking at the number and the behaviour under heavy stress, I think we > > have to consider a different approach. > > Thanks for testing. Is that a single cpu machine? Single cpu with two cores. > If yes, I think this result might be because the tasklet busy-loop > competes with conntrackd for cpu, so essentially we waste cycles > on futile re-delivery instead of leaving the cpu to conntrackd, > (which should process events). Makes sense. > If thats true, then we might be able to improve this by avoiding the > 'tasklet re-scheds itself'. This would also solve the > 'softirqd eats 100% cpu' when conntrackd is stopped/suspended. > > I'll see if I can cook up a patch some time tomorrow. That's fine. Thanks. -- To unsubscribe from this list: send the line "unsubscribe netfilter-devel" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
diff --git a/include/net/netfilter/nf_conntrack.h b/include/net/netfilter/nf_conntrack.h index caca0c4..e1cc862 100644 --- a/include/net/netfilter/nf_conntrack.h +++ b/include/net/netfilter/nf_conntrack.h @@ -182,7 +182,6 @@ __nf_conntrack_find(struct net *net, u16 zone, extern int nf_conntrack_hash_check_insert(struct nf_conn *ct); extern void nf_ct_delete_from_lists(struct nf_conn *ct); -extern void nf_ct_dying_timeout(struct nf_conn *ct); extern void nf_conntrack_flush_report(struct net *net, u32 pid, int report); diff --git a/include/net/netfilter/nf_conntrack_ecache.h b/include/net/netfilter/nf_conntrack_ecache.h index 5654d29..07c71e9 100644 --- a/include/net/netfilter/nf_conntrack_ecache.h +++ b/include/net/netfilter/nf_conntrack_ecache.h @@ -18,7 +18,6 @@ struct nf_conntrack_ecache { u16 ctmask; /* bitmask of ct events to be delivered */ u16 expmask; /* bitmask of expect events to be delivered */ u32 portid; /* netlink portid of destroyer */ - struct timer_list timeout; }; static inline struct nf_conntrack_ecache * @@ -207,6 +206,7 @@ nf_ct_expect_event(enum ip_conntrack_expect_events event, nf_ct_expect_event_report(event, exp, 0, 0); } +extern void nf_ct_dying_schedule(struct net *net); extern int nf_conntrack_ecache_init(struct net *net); extern void nf_conntrack_ecache_fini(struct net *net); @@ -232,6 +232,8 @@ static inline void nf_ct_expect_event_report(enum ip_conntrack_expect_events e, u32 portid, int report) {} +static inline void nf_ct_dying_schedule(struct net *net) { } + static inline int nf_conntrack_ecache_init(struct net *net) { return 0; diff --git a/include/net/netns/conntrack.h b/include/net/netns/conntrack.h index a1d83cc..0cef968 100644 --- a/include/net/netns/conntrack.h +++ b/include/net/netns/conntrack.h @@ -4,6 +4,7 @@ #include <linux/list.h> #include <linux/list_nulls.h> #include <linux/atomic.h> +#include <linux/interrupt.h> #include <linux/netfilter/nf_conntrack_tcp.h> struct ctl_table_header; @@ -71,11 +72,13 @@ struct netns_ct { struct hlist_head *expect_hash; struct hlist_nulls_head unconfirmed; struct hlist_nulls_head dying; +#ifdef CONFIG_NF_CONNTRACK_EVENTS + struct tasklet_struct dying_tasklet; +#endif struct ip_conntrack_stat __percpu *stat; struct nf_ct_event_notifier __rcu *nf_conntrack_event_cb; struct nf_exp_event_notifier __rcu *nf_expect_event_cb; int sysctl_events; - unsigned int sysctl_events_retry_timeout; int sysctl_acct; int sysctl_tstamp; int sysctl_checksum; diff --git a/net/netfilter/nf_conntrack_core.c b/net/netfilter/nf_conntrack_core.c index af17516..a98a8a9 100644 --- a/net/netfilter/nf_conntrack_core.c +++ b/net/netfilter/nf_conntrack_core.c @@ -252,41 +252,6 @@ void nf_ct_delete_from_lists(struct nf_conn *ct) } EXPORT_SYMBOL_GPL(nf_ct_delete_from_lists); -static void death_by_event(unsigned long ul_conntrack) -{ - struct nf_conn *ct = (void *)ul_conntrack; - struct net *net = nf_ct_net(ct); - struct nf_conntrack_ecache *ecache = nf_ct_ecache_find(ct); - - BUG_ON(ecache == NULL); - - if (nf_conntrack_event(IPCT_DESTROY, ct) < 0) { - /* bad luck, let's retry again */ - ecache->timeout.expires = jiffies + - (random32() % net->ct.sysctl_events_retry_timeout); - add_timer(&ecache->timeout); - return; - } - /* we've got the event delivered, now it's dying */ - set_bit(IPS_DYING_BIT, &ct->status); - nf_ct_put(ct); -} - -void nf_ct_dying_timeout(struct nf_conn *ct) -{ - struct net *net = nf_ct_net(ct); - struct nf_conntrack_ecache *ecache = nf_ct_ecache_find(ct); - - BUG_ON(ecache == NULL); - - /* set a new timer to retry event delivery */ - setup_timer(&ecache->timeout, death_by_event, (unsigned long)ct); - ecache->timeout.expires = jiffies + - (random32() % net->ct.sysctl_events_retry_timeout); - add_timer(&ecache->timeout); -} -EXPORT_SYMBOL_GPL(nf_ct_dying_timeout); - static void death_by_timeout(unsigned long ul_conntrack) { struct nf_conn *ct = (void *)ul_conntrack; @@ -300,7 +265,7 @@ static void death_by_timeout(unsigned long ul_conntrack) unlikely(nf_conntrack_event(IPCT_DESTROY, ct) < 0)) { /* destroy event was not delivered */ nf_ct_delete_from_lists(ct); - nf_ct_dying_timeout(ct); + nf_ct_dying_schedule(nf_ct_net(ct)); return; } set_bit(IPS_DYING_BIT, &ct->status); @@ -1304,21 +1269,6 @@ void nf_conntrack_flush_report(struct net *net, u32 pid, int report) } EXPORT_SYMBOL_GPL(nf_conntrack_flush_report); -static void nf_ct_release_dying_list(struct net *net) -{ - struct nf_conntrack_tuple_hash *h; - struct nf_conn *ct; - struct hlist_nulls_node *n; - - spin_lock_bh(&nf_conntrack_lock); - hlist_nulls_for_each_entry(h, n, &net->ct.dying, hnnode) { - ct = nf_ct_tuplehash_to_ctrack(h); - /* never fails to remove them, no listeners at this point */ - nf_ct_kill(ct); - } - spin_unlock_bh(&nf_conntrack_lock); -} - static int untrack_refs(void) { int cnt = 0, cpu; @@ -1345,7 +1295,7 @@ static void nf_conntrack_cleanup_net(struct net *net) { i_see_dead_people: nf_ct_iterate_cleanup(net, kill_all, NULL); - nf_ct_release_dying_list(net); + nf_ct_dying_schedule(net); if (atomic_read(&net->ct.count) != 0) { schedule(); goto i_see_dead_people; diff --git a/net/netfilter/nf_conntrack_ecache.c b/net/netfilter/nf_conntrack_ecache.c index faa978f..b48efc0 100644 --- a/net/netfilter/nf_conntrack_ecache.c +++ b/net/netfilter/nf_conntrack_ecache.c @@ -27,6 +27,38 @@ static DEFINE_MUTEX(nf_ct_ecache_mutex); +static void dying_tasklet_retry_events(unsigned long ctnetns) +{ + struct netns_ct *ctnet = (void *) ctnetns; + struct nf_conntrack_tuple_hash *h; + struct hlist_nulls_node *n; + struct nf_conn *ct; + int err = 0; + int budget = 20; + + rcu_read_lock(); + + hlist_nulls_for_each_entry(h, n, &ctnet->dying, hnnode) { + ct = nf_ct_tuplehash_to_ctrack(h); + if (test_bit(IPS_DYING_BIT, &ct->status)) + continue; + err = nf_conntrack_event(IPCT_DESTROY, ct); + if (err) + break; + /* we've got the event delivered, now it's dying */ + set_bit(IPS_DYING_BIT, &ct->status); + nf_ct_put(ct); + if (--budget == 0) + break; + } + + rcu_read_unlock(); + + /* err or budget exhausted? -> entries with undelivered ct event -- resched. */ + if (err || budget == 0) + tasklet_schedule(&ctnet->dying_tasklet); +} + /* deliver cached events and clear cache entry - must be called with locally * disabled softirqs */ void nf_ct_deliver_cached_events(struct nf_conn *ct) @@ -81,6 +113,13 @@ out_unlock: } EXPORT_SYMBOL_GPL(nf_ct_deliver_cached_events); +void nf_ct_dying_schedule(struct net *net) +{ + /* retry event delivery */ + tasklet_schedule(&net->ct.dying_tasklet); +} +EXPORT_SYMBOL_GPL(nf_ct_dying_schedule); + int nf_conntrack_register_notifier(struct net *net, struct nf_ct_event_notifier *new) { @@ -155,7 +194,6 @@ EXPORT_SYMBOL_GPL(nf_ct_expect_unregister_notifier); #define NF_CT_EVENTS_DEFAULT 1 static int nf_ct_events __read_mostly = NF_CT_EVENTS_DEFAULT; -static int nf_ct_events_retry_timeout __read_mostly = 15*HZ; #ifdef CONFIG_SYSCTL static struct ctl_table event_sysctl_table[] = { @@ -166,13 +204,6 @@ static struct ctl_table event_sysctl_table[] = { .mode = 0644, .proc_handler = proc_dointvec, }, - { - .procname = "nf_conntrack_events_retry_timeout", - .data = &init_net.ct.sysctl_events_retry_timeout, - .maxlen = sizeof(unsigned int), - .mode = 0644, - .proc_handler = proc_dointvec_jiffies, - }, {} }; #endif /* CONFIG_SYSCTL */ @@ -194,7 +225,6 @@ static int nf_conntrack_event_init_sysctl(struct net *net) goto out; table[0].data = &net->ct.sysctl_events; - table[1].data = &net->ct.sysctl_events_retry_timeout; /* Don't export sysctls to unprivileged users */ if (net->user_ns != &init_user_ns) @@ -238,7 +268,9 @@ int nf_conntrack_ecache_init(struct net *net) int ret; net->ct.sysctl_events = nf_ct_events; - net->ct.sysctl_events_retry_timeout = nf_ct_events_retry_timeout; + + tasklet_init(&net->ct.dying_tasklet, dying_tasklet_retry_events, + (unsigned long) &net->ct); if (net_eq(net, &init_net)) { ret = nf_ct_extend_register(&event_extend); @@ -264,6 +296,8 @@ out_extend_register: void nf_conntrack_ecache_fini(struct net *net) { + tasklet_kill(&net->ct.dying_tasklet); + nf_conntrack_event_fini_sysctl(net); if (net_eq(net, &init_net)) nf_ct_extend_unregister(&event_extend); diff --git a/net/netfilter/nf_conntrack_netlink.c b/net/netfilter/nf_conntrack_netlink.c index 4e078cd..6e808ad 100644 --- a/net/netfilter/nf_conntrack_netlink.c +++ b/net/netfilter/nf_conntrack_netlink.c @@ -992,7 +992,7 @@ ctnetlink_del_conntrack(struct sock *ctnl, struct sk_buff *skb, nlmsg_report(nlh)) < 0) { nf_ct_delete_from_lists(ct); /* we failed to report the event, try later */ - nf_ct_dying_timeout(ct); + nf_ct_dying_schedule(net); nf_ct_put(ct); return 0; }
This brings the (per-conntrack) ecache extension back to 24 bytes in size (was 112 byte on x86_64 with lockdep on). Instead we use a per-ns tasklet to re-trigger event delivery. When we enqueue a ct entry into the dying list, the tasklet is scheduled. The tasklet will then deliver up to 20 entries. It will re-sched itself unless all the pending events could be delivered. While at it, dying list handling is moved into ecache.c, since its only revlevant if ct events are enabled. Signed-off-by: Florian Westphal <fw@strlen.de> --- Changes since V1: - rebase on top of nf-next master - remove nf_ct_release_dying_list (we can simply wait for the tasklet until all entries are gone. This should happen very quickly since all listeners are gone at this point). include/net/netfilter/nf_conntrack.h | 1 - include/net/netfilter/nf_conntrack_ecache.h | 4 +- include/net/netns/conntrack.h | 5 ++- net/netfilter/nf_conntrack_core.c | 54 +-------------------------- net/netfilter/nf_conntrack_ecache.c | 54 ++++++++++++++++++++++----- net/netfilter/nf_conntrack_netlink.c | 2 +- 6 files changed, 54 insertions(+), 66 deletions(-)