diff mbox

[1/2] tipc: avoid possible deadlock while remove link_timeout()

Message ID 520376BA.5040509@huawei.com
State Not Applicable, archived
Delegated to: David Miller
Headers show

Commit Message

Ding Tianhong Aug. 8, 2013, 10:45 a.m. UTC
We met lockdep warning when enable and disable the bearer for commands such as:

tipc-config -netid=1234 -addr=1.1.3 -be=eth:eth0
tipc-config -netid=1234 -addr=1.1.3 -bd=eth:eth0

[ 3001.445459] tipc: Established link <1.1.3:eth0-1.1.2:br0> on network plane A
[ 3029.457875] tipc: Disabling bearer <eth:eth0>
[ 3029.458066]
[ 3029.458071] ======================================================
[ 3029.458075] [ INFO: possible circular locking dependency detected ]
[ 3029.458080] 3.11.0-rc3-wwd-default #4 Not tainted
[ 3029.458084] -------------------------------------------------------
[ 3029.458088] rmmod/7092 is trying to acquire lock:
[ 3029.458092]  (((timer))#3){+.-...}, at: [<ffffffff8105be80>] del_timer_sync+0x0/0xd0
[ 3029.458107]
[ 3029.458107] but task is already holding lock:
[ 3029.458112]  (&(&b_ptr->lock)->rlock){+.-...}, at: [<ffffffffa02b94e3>] bearer_disable+0x33/0xd0 [tipc]
[ 3029.458126]
[ 3029.458126] which lock already depends on the new lock.
[ 3029.458126]
[ 3029.458132]
[ 3029.458132] the existing dependency chain (in reverse order) is:
[ 3029.458137]
[ 3029.458137] -> #2 (&(&b_ptr->lock)->rlock){+.-...}:
[ 3029.458143]        [<ffffffff810b3b4d>] validate_chain+0x6dd/0x870
[ 3029.458151]        [<ffffffff810b40bb>] __lock_acquire+0x3db/0x670
[ 3029.458156]        [<ffffffff810b4453>] lock_acquire+0x103/0x130
[ 3029.458161]        [<ffffffff814d65b1>] _raw_spin_lock_bh+0x41/0x80
[ 3029.458169]        [<ffffffffa02b9600>] tipc_bearer_blocked+0x20/0x40 [tipc]
[ 3029.458176]        [<ffffffffa02bf01b>] tipc_link_send_proto_msg+0x35b/0x520 tipc]
[ 3029.458184]        [<ffffffffa02bf83a>] link_state_event+0x33a/0x590 [tipc]
[ 3029.458191]        [<ffffffffa02bfab9>] link_start+0x29/0x40 [tipc]
[ 3029.458198]        [<ffffffffa02bb13f>] process_signal_queue+0x7f/0xc0 [tipc]
[ 3029.458206]        [<ffffffff8105304d>] tasklet_action+0x6d/0xf0
[ 3029.458214]        [<ffffffff8105379a>] __do_softirq+0x16a/0x2e0
[ 3029.458219]        [<ffffffff81053945>] run_ksoftirqd+0x35/0x50
[ 3029.458224]        [<ffffffff8107d042>] smpboot_thread_fn+0x1e2/0x2f0
[ 3029.458235]        [<ffffffff81073c5e>] kthread+0xde/0xf0
[ 3029.458242]        [<ffffffff814de7ac>] ret_from_fork+0x7c/0xb0
[ 3029.458250]
[ 3029.458250] -> #1 (&(&n_ptr->lock)->rlock){+.-...}:
[ 3029.458257]        [<ffffffff810b3b4d>] validate_chain+0x6dd/0x870
[ 3029.458262]        [<ffffffff810b40bb>] __lock_acquire+0x3db/0x670
[ 3029.458268]        [<ffffffff810b4453>] lock_acquire+0x103/0x130
[ 3029.458273]        [<ffffffff814d65b1>] _raw_spin_lock_bh+0x41/0x80
[ 3029.458279]        [<ffffffffa02bfaec>] link_timeout+0x1c/0x170 [tipc]
[ 3029.458287]        [<ffffffff8105b92a>] call_timer_fn+0xda/0x1e0
[ 3029.458292]        [<ffffffff8105bcd7>] run_timer_softirq+0x2a7/0x2d0
[ 3029.458298]        [<ffffffff8105379a>] __do_softirq+0x16a/0x2e0
[ 3029.458304]        [<ffffffff81053a35>] irq_exit+0xd5/0xe0
[ 3029.458309]        [<ffffffff81033005>] smp_apic_timer_interrupt+0x45/0x60
[ 3029.458319]        [<ffffffff814df4af>] apic_timer_interrupt+0x6f/0x80
[ 3029.458325]        [<ffffffff8100b70e>] arch_cpu_idle+0x1e/0x30
[ 3029.458332]        [<ffffffff810a039d>] cpu_idle_loop+0x1fd/0x280
[ 3029.458338]        [<ffffffff810a043e>] cpu_startup_entry+0x1e/0x20
[ 3029.458343]        [<ffffffff814c8841>] rest_init+0xc1/0xd0
[ 3029.458349]        [<ffffffff81c990fc>] start_kernel+0x3a3/0x451
[ 3029.458356]        [<ffffffff81c984d1>] x86_64_start_reservations+0x1b/0x32
[ 3029.458362]        [<ffffffff81c98622>] x86_64_start_kernel+0x13a/0x141
[ 3029.458368]
[ 3029.458368] -> #0 (((timer))#3){+.-...}:
[ 3029.458375]        [<ffffffff810b33fe>] check_prev_add+0x43e/0x4b0
[ 3029.458380]        [<ffffffff810b3b4d>] validate_chain+0x6dd/0x870
[ 3029.458386]        [<ffffffff810b40bb>] __lock_acquire+0x3db/0x670
[ 3029.458391]        [<ffffffff810b4453>] lock_acquire+0x103/0x130
[ 3029.458397]        [<ffffffff8105bebd>] del_timer_sync+0x3d/0xd0
[ 3029.458402]        [<ffffffffa02bcc4e>] tipc_link_delete+0x1e/0xb0 [tipc]
[ 3029.458410]        [<ffffffffa02b9528>] bearer_disable+0x78/0xd0 [tipc]
[ 3029.458417]        [<ffffffffa02b95b4>] tipc_bearer_stop+0x34/0x60 [tipc]
[ 3029.458423]        [<ffffffffa02c384b>] tipc_net_stop+0x2b/0x90 [tipc]
[ 3029.458432]        [<ffffffffa02caf49>] tipc_exit+0x9/0xc0 [tipc]
[ 3029.458439]        [<ffffffff810c1a58>] SyS_delete_module+0x198/0x290
[ 3029.458445]        [<ffffffff814de852>] system_call_fastpath+0x16/0x1b
[ 3029.458451]
[ 3029.458451] other info that might help us debug this:
[ 3029.458451]
[ 3029.458458] Chain exists of:
[ 3029.458458]   ((timer))#3 --> &(&n_ptr->lock)->rlock --> &(&b_ptr->lock)->rlock
[ 3029.458458]
[ 3029.458469]  Possible unsafe locking scenario:
[ 3029.458469]
[ 3029.458474]        CPU0                    CPU1
[ 3029.458478]        ----                    ----
[ 3029.458481]   lock(&(&b_ptr->lock)->rlock);
[ 3029.458486]                                lock(&(&n_ptr->lock)->rlock);
[ 3029.458492]                                lock(&(&b_ptr->lock)->rlock);
[ 3029.458497]   lock(((timer))#3);
[ 3029.458502]
[ 3029.458502]  *** DEADLOCK ***
[ 3029.458502]
[ 3029.458508] 2 locks held by rmmod/7092:
[ 3029.458511]  #0:  (tipc_net_lock){++.-..}, at: [<ffffffffa02c3846>] tipc_net_stop+0x26/0x90 [tipc]
[ 3029.458523]  #1:  (&(&b_ptr->lock)->rlock){+.-...}, at: [<ffffffffa02b94e3>]bearer_disable+0x33/0xd0 [tipc]
[ 3029.458535]
[ 3029.458535] stack backtrace:
[ 3029.458541] CPU: 3 PID: 7092 Comm: rmmod Not tainted 3.11.0-rc3-wwd-default #4
[ 3029.458546] Hardware name: Bochs Bochs, BIOS Bochs 01/01/2007
[ 3029.458550]  00000000ffffffff ffff88010fd09c08 ffffffff814d03dd 0000000000000000
[ 3029.458559]  ffffffff8205fca0 ffff88010fd09c48 ffffffff810b1c4f 000000000fd09c48
[ 3029.458566]  ffff88010fd09c68 ffff88010e4d4fc0 0000000000000000 ffff88010e4d56f0
[ 3029.458574] Call Trace:
[ 3029.458579]  [<ffffffff814d03dd>] dump_stack+0x4d/0xa0
[ 3029.458585]  [<ffffffff810b1c4f>] print_circular_bug+0x10f/0x120
[ 3029.458591]  [<ffffffff810b33fe>] check_prev_add+0x43e/0x4b0
[ 3029.458598]  [<ffffffff8100a226>] ? native_sched_clock+0x26/0x90
[ 3029.458604]  [<ffffffff810b3b4d>] validate_chain+0x6dd/0x870
[ 3029.458612]  [<ffffffff81087a28>] ? sched_clock_cpu+0xd8/0x110
[ 3029.458618]  [<ffffffff810b40bb>] __lock_acquire+0x3db/0x670
[ 3029.458624]  [<ffffffff810b4453>] lock_acquire+0x103/0x130
[ 3029.458629]  [<ffffffff8105be80>] ? try_to_del_timer_sync+0x70/0x70
[ 3029.458635]  [<ffffffff8105bebd>] del_timer_sync+0x3d/0xd0
[ 3029.458641]  [<ffffffff8105be80>] ? try_to_del_timer_sync+0x70/0x70
[ 3029.458649]  [<ffffffffa02bcc4e>] tipc_link_delete+0x1e/0xb0 [tipc]
[ 3029.458656]  [<ffffffffa02b9528>] bearer_disable+0x78/0xd0 [tipc]
[ 3029.458663]  [<ffffffffa02b95b4>] tipc_bearer_stop+0x34/0x60 [tipc]
[ 3029.458671]  [<ffffffffa02c384b>] tipc_net_stop+0x2b/0x90 [tipc]
[ 3029.458679]  [<ffffffffa02caf49>] tipc_exit+0x9/0xc0 [tipc]
[ 3029.458685]  [<ffffffff810c1a58>] SyS_delete_module+0x198/0x290
[ 3029.458691]  [<ffffffff814de852>] system_call_fastpath+0x16/0x1b

----------------------------------------------------------------------

The problem is that the tipc_link_delete() will cancel the timer l_ptr->timer when
the b_ptr->lock is hold, but the l_ptr->timer still call b_ptr->lock to finish the
work, so the dead lock occurs.

We should unlock the b_ptr->lock when del the l_ptr->timer.

Reported-by: Wang Weidong <wangweidong1@huawei.com>
Signed-off-by: Ding Tianhong <dingtianhong@huawei.com>
---
 net/tipc/bearer.c | 8 +++++++-
 net/tipc/link.c   | 2 ++
 2 files changed, 9 insertions(+), 1 deletion(-)

Comments

Ying Xue Aug. 9, 2013, 2:39 a.m. UTC | #1
On 08/08/2013 06:45 PM, Ding Tianhong wrote:
> We met lockdep warning when enable and disable the bearer for commands such as:
> 
> tipc-config -netid=1234 -addr=1.1.3 -be=eth:eth0
> tipc-config -netid=1234 -addr=1.1.3 -bd=eth:eth0
> 
> [ 3001.445459] tipc: Established link <1.1.3:eth0-1.1.2:br0> on network plane A
> [ 3029.457875] tipc: Disabling bearer <eth:eth0>
> [ 3029.458066]
> [ 3029.458071] ======================================================
> [ 3029.458075] [ INFO: possible circular locking dependency detected ]
> [ 3029.458080] 3.11.0-rc3-wwd-default #4 Not tainted
> [ 3029.458084] -------------------------------------------------------
> [ 3029.458088] rmmod/7092 is trying to acquire lock:
> [ 3029.458092]  (((timer))#3){+.-...}, at: [<ffffffff8105be80>] del_timer_sync+0x0/0xd0
> [ 3029.458107]
> [ 3029.458107] but task is already holding lock:
> [ 3029.458112]  (&(&b_ptr->lock)->rlock){+.-...}, at: [<ffffffffa02b94e3>] bearer_disable+0x33/0xd0 [tipc]
> [ 3029.458126]
> [ 3029.458126] which lock already depends on the new lock.
> [ 3029.458126]
> [ 3029.458132]
> [ 3029.458132] the existing dependency chain (in reverse order) is:
> [ 3029.458137]
> [ 3029.458137] -> #2 (&(&b_ptr->lock)->rlock){+.-...}:
> [ 3029.458143]        [<ffffffff810b3b4d>] validate_chain+0x6dd/0x870
> [ 3029.458151]        [<ffffffff810b40bb>] __lock_acquire+0x3db/0x670
> [ 3029.458156]        [<ffffffff810b4453>] lock_acquire+0x103/0x130
> [ 3029.458161]        [<ffffffff814d65b1>] _raw_spin_lock_bh+0x41/0x80
> [ 3029.458169]        [<ffffffffa02b9600>] tipc_bearer_blocked+0x20/0x40 [tipc]
> [ 3029.458176]        [<ffffffffa02bf01b>] tipc_link_send_proto_msg+0x35b/0x520 tipc]
> [ 3029.458184]        [<ffffffffa02bf83a>] link_state_event+0x33a/0x590 [tipc]
> [ 3029.458191]        [<ffffffffa02bfab9>] link_start+0x29/0x40 [tipc]
> [ 3029.458198]        [<ffffffffa02bb13f>] process_signal_queue+0x7f/0xc0 [tipc]
> [ 3029.458206]        [<ffffffff8105304d>] tasklet_action+0x6d/0xf0
> [ 3029.458214]        [<ffffffff8105379a>] __do_softirq+0x16a/0x2e0
> [ 3029.458219]        [<ffffffff81053945>] run_ksoftirqd+0x35/0x50
> [ 3029.458224]        [<ffffffff8107d042>] smpboot_thread_fn+0x1e2/0x2f0
> [ 3029.458235]        [<ffffffff81073c5e>] kthread+0xde/0xf0
> [ 3029.458242]        [<ffffffff814de7ac>] ret_from_fork+0x7c/0xb0
> [ 3029.458250]
> [ 3029.458250] -> #1 (&(&n_ptr->lock)->rlock){+.-...}:
> [ 3029.458257]        [<ffffffff810b3b4d>] validate_chain+0x6dd/0x870
> [ 3029.458262]        [<ffffffff810b40bb>] __lock_acquire+0x3db/0x670
> [ 3029.458268]        [<ffffffff810b4453>] lock_acquire+0x103/0x130
> [ 3029.458273]        [<ffffffff814d65b1>] _raw_spin_lock_bh+0x41/0x80
> [ 3029.458279]        [<ffffffffa02bfaec>] link_timeout+0x1c/0x170 [tipc]
> [ 3029.458287]        [<ffffffff8105b92a>] call_timer_fn+0xda/0x1e0
> [ 3029.458292]        [<ffffffff8105bcd7>] run_timer_softirq+0x2a7/0x2d0
> [ 3029.458298]        [<ffffffff8105379a>] __do_softirq+0x16a/0x2e0
> [ 3029.458304]        [<ffffffff81053a35>] irq_exit+0xd5/0xe0
> [ 3029.458309]        [<ffffffff81033005>] smp_apic_timer_interrupt+0x45/0x60
> [ 3029.458319]        [<ffffffff814df4af>] apic_timer_interrupt+0x6f/0x80
> [ 3029.458325]        [<ffffffff8100b70e>] arch_cpu_idle+0x1e/0x30
> [ 3029.458332]        [<ffffffff810a039d>] cpu_idle_loop+0x1fd/0x280
> [ 3029.458338]        [<ffffffff810a043e>] cpu_startup_entry+0x1e/0x20
> [ 3029.458343]        [<ffffffff814c8841>] rest_init+0xc1/0xd0
> [ 3029.458349]        [<ffffffff81c990fc>] start_kernel+0x3a3/0x451
> [ 3029.458356]        [<ffffffff81c984d1>] x86_64_start_reservations+0x1b/0x32
> [ 3029.458362]        [<ffffffff81c98622>] x86_64_start_kernel+0x13a/0x141
> [ 3029.458368]
> [ 3029.458368] -> #0 (((timer))#3){+.-...}:
> [ 3029.458375]        [<ffffffff810b33fe>] check_prev_add+0x43e/0x4b0
> [ 3029.458380]        [<ffffffff810b3b4d>] validate_chain+0x6dd/0x870
> [ 3029.458386]        [<ffffffff810b40bb>] __lock_acquire+0x3db/0x670
> [ 3029.458391]        [<ffffffff810b4453>] lock_acquire+0x103/0x130
> [ 3029.458397]        [<ffffffff8105bebd>] del_timer_sync+0x3d/0xd0
> [ 3029.458402]        [<ffffffffa02bcc4e>] tipc_link_delete+0x1e/0xb0 [tipc]
> [ 3029.458410]        [<ffffffffa02b9528>] bearer_disable+0x78/0xd0 [tipc]
> [ 3029.458417]        [<ffffffffa02b95b4>] tipc_bearer_stop+0x34/0x60 [tipc]
> [ 3029.458423]        [<ffffffffa02c384b>] tipc_net_stop+0x2b/0x90 [tipc]
> [ 3029.458432]        [<ffffffffa02caf49>] tipc_exit+0x9/0xc0 [tipc]
> [ 3029.458439]        [<ffffffff810c1a58>] SyS_delete_module+0x198/0x290
> [ 3029.458445]        [<ffffffff814de852>] system_call_fastpath+0x16/0x1b
> [ 3029.458451]
> [ 3029.458451] other info that might help us debug this:
> [ 3029.458451]
> [ 3029.458458] Chain exists of:
> [ 3029.458458]   ((timer))#3 --> &(&n_ptr->lock)->rlock --> &(&b_ptr->lock)->rlock
> [ 3029.458458]
> [ 3029.458469]  Possible unsafe locking scenario:
> [ 3029.458469]
> [ 3029.458474]        CPU0                    CPU1
> [ 3029.458478]        ----                    ----
> [ 3029.458481]   lock(&(&b_ptr->lock)->rlock);
> [ 3029.458486]                                lock(&(&n_ptr->lock)->rlock);
> [ 3029.458492]                                lock(&(&b_ptr->lock)->rlock);
> [ 3029.458497]   lock(((timer))#3);
> [ 3029.458502]
> [ 3029.458502]  *** DEADLOCK ***
> [ 3029.458502]
> [ 3029.458508] 2 locks held by rmmod/7092:
> [ 3029.458511]  #0:  (tipc_net_lock){++.-..}, at: [<ffffffffa02c3846>] tipc_net_stop+0x26/0x90 [tipc]
> [ 3029.458523]  #1:  (&(&b_ptr->lock)->rlock){+.-...}, at: [<ffffffffa02b94e3>]bearer_disable+0x33/0xd0 [tipc]
> [ 3029.458535]
> [ 3029.458535] stack backtrace:
> [ 3029.458541] CPU: 3 PID: 7092 Comm: rmmod Not tainted 3.11.0-rc3-wwd-default #4
> [ 3029.458546] Hardware name: Bochs Bochs, BIOS Bochs 01/01/2007
> [ 3029.458550]  00000000ffffffff ffff88010fd09c08 ffffffff814d03dd 0000000000000000
> [ 3029.458559]  ffffffff8205fca0 ffff88010fd09c48 ffffffff810b1c4f 000000000fd09c48
> [ 3029.458566]  ffff88010fd09c68 ffff88010e4d4fc0 0000000000000000 ffff88010e4d56f0
> [ 3029.458574] Call Trace:
> [ 3029.458579]  [<ffffffff814d03dd>] dump_stack+0x4d/0xa0
> [ 3029.458585]  [<ffffffff810b1c4f>] print_circular_bug+0x10f/0x120
> [ 3029.458591]  [<ffffffff810b33fe>] check_prev_add+0x43e/0x4b0
> [ 3029.458598]  [<ffffffff8100a226>] ? native_sched_clock+0x26/0x90
> [ 3029.458604]  [<ffffffff810b3b4d>] validate_chain+0x6dd/0x870
> [ 3029.458612]  [<ffffffff81087a28>] ? sched_clock_cpu+0xd8/0x110
> [ 3029.458618]  [<ffffffff810b40bb>] __lock_acquire+0x3db/0x670
> [ 3029.458624]  [<ffffffff810b4453>] lock_acquire+0x103/0x130
> [ 3029.458629]  [<ffffffff8105be80>] ? try_to_del_timer_sync+0x70/0x70
> [ 3029.458635]  [<ffffffff8105bebd>] del_timer_sync+0x3d/0xd0
> [ 3029.458641]  [<ffffffff8105be80>] ? try_to_del_timer_sync+0x70/0x70
> [ 3029.458649]  [<ffffffffa02bcc4e>] tipc_link_delete+0x1e/0xb0 [tipc]
> [ 3029.458656]  [<ffffffffa02b9528>] bearer_disable+0x78/0xd0 [tipc]
> [ 3029.458663]  [<ffffffffa02b95b4>] tipc_bearer_stop+0x34/0x60 [tipc]
> [ 3029.458671]  [<ffffffffa02c384b>] tipc_net_stop+0x2b/0x90 [tipc]
> [ 3029.458679]  [<ffffffffa02caf49>] tipc_exit+0x9/0xc0 [tipc]
> [ 3029.458685]  [<ffffffff810c1a58>] SyS_delete_module+0x198/0x290
> [ 3029.458691]  [<ffffffff814de852>] system_call_fastpath+0x16/0x1b
> 
> ----------------------------------------------------------------------
> 
> The problem is that the tipc_link_delete() will cancel the timer l_ptr->timer when
> the b_ptr->lock is hold, but the l_ptr->timer still call b_ptr->lock to finish the
> work, so the dead lock occurs.
> 

I guess the issue would disappear if you apply below patch:

http://article.gmane.org/gmane.network.tipc.general/4380

And the patch will be submitted into net-next tree soon.

Regards,
Ying

> We should unlock the b_ptr->lock when del the l_ptr->timer.
> 
> Reported-by: Wang Weidong <wangweidong1@huawei.com>
> Signed-off-by: Ding Tianhong <dingtianhong@huawei.com>
> ---
>  net/tipc/bearer.c | 8 +++++++-
>  net/tipc/link.c   | 2 ++
>  2 files changed, 9 insertions(+), 1 deletion(-)
> 
> diff --git a/net/tipc/bearer.c b/net/tipc/bearer.c
> index cb29ef7..7687211 100644
> --- a/net/tipc/bearer.c
> +++ b/net/tipc/bearer.c
> @@ -460,14 +460,20 @@ static void bearer_disable(struct tipc_bearer *b_ptr)
>  {
>  	struct tipc_link *l_ptr;
>  	struct tipc_link *temp_l_ptr;
> +	struct list_head list;
>  
>  	pr_info("Disabling bearer <%s>\n", b_ptr->name);
>  	spin_lock_bh(&b_ptr->lock);
>  	b_ptr->blocked = 1;
>  	b_ptr->media->disable_bearer(b_ptr);
> -	list_for_each_entry_safe(l_ptr, temp_l_ptr, &b_ptr->links, link_list) {
> +	list_replace_init(&b_ptr->links, &list);
> +	spin_unlock_bh(&b_ptr->lock);
> +
> +	list_for_each_entry_safe(l_ptr, temp_l_ptr, &list, link_list) {
>  		tipc_link_delete(l_ptr);
>  	}
> +
> +	spin_lock_bh(&b_ptr->lock);
>  	if (b_ptr->link_req)
>  		tipc_disc_delete(b_ptr->link_req);
>  	spin_unlock_bh(&b_ptr->lock);
> diff --git a/net/tipc/link.c b/net/tipc/link.c
> index 0cc3d90..a145718 100644
> --- a/net/tipc/link.c
> +++ b/net/tipc/link.c
> @@ -384,10 +384,12 @@ void tipc_link_delete(struct tipc_link *l_ptr)
>  	k_cancel_timer(&l_ptr->timer);
>  
>  	tipc_node_lock(l_ptr->owner);
> +	spin_lock_bh(&l_ptr->b_ptr->lock);
>  	tipc_link_reset(l_ptr);
>  	tipc_node_detach_link(l_ptr->owner, l_ptr);
>  	tipc_link_stop(l_ptr);
>  	list_del_init(&l_ptr->link_list);
> +	spin_unlock_bh(&l_ptr->b_ptr->lock);
>  	tipc_node_unlock(l_ptr->owner);
>  	k_term_timer(&l_ptr->timer);
>  	kfree(l_ptr);
> 

--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Ding Tianhong Aug. 9, 2013, 3:38 a.m. UTC | #2
On 2013/8/9 10:39, Ying Xue wrote:
> On 08/08/2013 06:45 PM, Ding Tianhong wrote:

>>
>> The problem is that the tipc_link_delete() will cancel the timer l_ptr->timer when
>> the b_ptr->lock is hold, but the l_ptr->timer still call b_ptr->lock to finish the
>> work, so the dead lock occurs.
>>
> 
> I guess the issue would disappear if you apply below patch:
> 
> http://article.gmane.org/gmane.network.tipc.general/4380
> 
> And the patch will be submitted into net-next tree soon.
> 
> Regards,
> Ying
> 

yes, I test the patch and fix the 1/2 problem, but the 2/2 problem still occurs, if 
the http://article.gmane.org/gmane.network.tipc.general/4380 and queue for stable, I will resend
2/2 patch again and cancel the 1/2 patch.


--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Ying Xue Aug. 9, 2013, 4:36 a.m. UTC | #3
On 08/09/2013 11:38 AM, Ding Tianhong wrote:
> On 2013/8/9 10:39, Ying Xue wrote:
>> On 08/08/2013 06:45 PM, Ding Tianhong wrote:
> 
>>>
>>> The problem is that the tipc_link_delete() will cancel the timer l_ptr->timer when
>>> the b_ptr->lock is hold, but the l_ptr->timer still call b_ptr->lock to finish the
>>> work, so the dead lock occurs.
>>>
>>
>> I guess the issue would disappear if you apply below patch:
>>
>> http://article.gmane.org/gmane.network.tipc.general/4380
>>
>> And the patch will be submitted into net-next tree soon.
>>
>> Regards,
>> Ying
>>
> 
> yes, I test the patch and fix the 1/2 problem, but the 2/2 problem still occurs, if 
> the http://article.gmane.org/gmane.network.tipc.general/4380 and queue for stable, I will resend
> 2/2 patch again and cancel the 1/2 patch.
> 

Yes, I know. Just send patch #2.

Regards,
Ying

> 
> 
> 

--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@vger.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
diff mbox

Patch

diff --git a/net/tipc/bearer.c b/net/tipc/bearer.c
index cb29ef7..7687211 100644
--- a/net/tipc/bearer.c
+++ b/net/tipc/bearer.c
@@ -460,14 +460,20 @@  static void bearer_disable(struct tipc_bearer *b_ptr)
 {
 	struct tipc_link *l_ptr;
 	struct tipc_link *temp_l_ptr;
+	struct list_head list;
 
 	pr_info("Disabling bearer <%s>\n", b_ptr->name);
 	spin_lock_bh(&b_ptr->lock);
 	b_ptr->blocked = 1;
 	b_ptr->media->disable_bearer(b_ptr);
-	list_for_each_entry_safe(l_ptr, temp_l_ptr, &b_ptr->links, link_list) {
+	list_replace_init(&b_ptr->links, &list);
+	spin_unlock_bh(&b_ptr->lock);
+
+	list_for_each_entry_safe(l_ptr, temp_l_ptr, &list, link_list) {
 		tipc_link_delete(l_ptr);
 	}
+
+	spin_lock_bh(&b_ptr->lock);
 	if (b_ptr->link_req)
 		tipc_disc_delete(b_ptr->link_req);
 	spin_unlock_bh(&b_ptr->lock);
diff --git a/net/tipc/link.c b/net/tipc/link.c
index 0cc3d90..a145718 100644
--- a/net/tipc/link.c
+++ b/net/tipc/link.c
@@ -384,10 +384,12 @@  void tipc_link_delete(struct tipc_link *l_ptr)
 	k_cancel_timer(&l_ptr->timer);
 
 	tipc_node_lock(l_ptr->owner);
+	spin_lock_bh(&l_ptr->b_ptr->lock);
 	tipc_link_reset(l_ptr);
 	tipc_node_detach_link(l_ptr->owner, l_ptr);
 	tipc_link_stop(l_ptr);
 	list_del_init(&l_ptr->link_list);
+	spin_unlock_bh(&l_ptr->b_ptr->lock);
 	tipc_node_unlock(l_ptr->owner);
 	k_term_timer(&l_ptr->timer);
 	kfree(l_ptr);