Message ID | 506FC4CD.7070509@oracle.com |
---|---|
State | Changes Requested, archived |
Delegated to: | David Miller |
Headers | show |
On 10/6/2012 12:42 AM, Jeff Liu wrote: > Hello, > > RDS ping/pong over TCP feature has broke for years(2.6.39 to 3.6.0) since we have to set TCP cork and > call kerenel_sendmsg() to reply a ping requirement which both need to lock "struct sock *sk". > However, this lock has already been hold before our rda_tcp_data_ready() callback is triggerred. > As a result, we always facing spinlock recursion which would resulting in system panic... > > Given that RDS ping is a special kind of message, we don't need to reply it as > soon as possible, IMHO, we can schedule it to work queue as a delayed response to > make TCP transport totally works. Also, I think we can using the system default > work queue to serve it to reduce the possible impact on general TCP transmit. > > With below patch, I have run rds-ping(run multiple ping between two hosts at the same time) and > rds-stress(hostA listen, hostB send packets) for half day, it works to me. > > Thanks, > -Jeff > > > Reported-by: Dan Carpenter<dan.carpenter@oracle.com> > CC: Venkat Venkatsubra<venkat.x.venkatsubra@oracle.com> > CC: David S. Miller<davem@davemloft.net> > CC: James Morris<james.l.morris@oracle.com> > Signed-off-by: Jie Liu<jeff.liu@oracle.com> > > --- > net/rds/send.c | 80 ++++++++++++++++++++++++++++++++++++++++++++++++++++---- > net/rds/tcp.h | 7 +++++ > 2 files changed, 82 insertions(+), 5 deletions(-) > > diff --git a/net/rds/send.c b/net/rds/send.c > index 96531d4..011006e 100644 > --- a/net/rds/send.c > +++ b/net/rds/send.c > @@ -38,8 +38,10 @@ > #include<linux/list.h> > #include<linux/ratelimit.h> > #include<linux/export.h> > +#include<linux/workqueue.h> > > #include "rds.h" > +#include "tcp.h" > > /* When transmitting messages in rds_send_xmit, we need to emerge from > * time to time and briefly release the CPU. Otherwise the softlock watchdog > @@ -55,6 +57,12 @@ static int send_batch_count = 64; > module_param(send_batch_count, int, 0444); > MODULE_PARM_DESC(send_batch_count, " batch factor when working the send queue"); > > +/* RDS over TCP ping/pong */ > +static void rds_tcp_pong_worker(struct work_struct *work); > +static DEFINE_SPINLOCK(rds_tcp_pong_lock); > +static LIST_HEAD(rds_tcp_pong_list); > +static DECLARE_WORK(rds_tcp_pong_work, rds_tcp_pong_worker); > + > static void rds_send_remove_from_sock(struct list_head *messages, int status); > > /* > @@ -1082,11 +1090,7 @@ out: > return ret; > } > > -/* > - * Reply to a ping packet. > - */ > -int > -rds_send_pong(struct rds_connection *conn, __be16 dport) > +static int rds_tcp_send_pong(struct rds_connection *conn, __be16 dport) > { > struct rds_message *rm; > unsigned long flags; > @@ -1132,3 +1136,69 @@ out: > rds_message_put(rm); > return ret; > } > + > +static void rds_tcp_pong_worker(struct work_struct *work) > +{ > + struct rds_tcp_pong *tp; > + struct rds_connection *conn; > + __be16 dport; > + > + spin_lock(&rds_tcp_pong_lock); > + if (list_empty(&rds_tcp_pong_list)) > + goto out_unlock; > + > + /* > + * Process on tcp pong once one time to reduce the possbile impact > + * on normal transmit. > + */ > + tp = list_entry(rds_tcp_pong_list.next, struct rds_tcp_pong, tp_node); > + conn = tp->tp_conn; > + dport = tp->tp_dport; > + list_del(&tp->tp_node); > + spin_unlock(&rds_tcp_pong_lock); > + > + kfree(tp); > + rds_tcp_send_pong(conn, dport); > + goto out; > + > +out_unlock: > + spin_unlock(&rds_tcp_pong_lock); > +out: > + return; > +} > + > +/* > + * RDS over TCP transport support ping/pong message. However, it > + * always resulting in sock spinlock recursion up to 3.7.0. To solve > + * this issue, we can shedule it to work queue as a delayed response. > + * Here we using the system default work queue. > + */ > +int rds_tcp_pong(struct rds_connection *conn, __be16 dport) > +{ > + struct rds_tcp_pong *tp; > + int ret = 0; > + > + tp = kmalloc(sizeof(*tp), GFP_ATOMIC); > + if (!tp) { > + ret = -ENOMEM; > + goto out; > + } > + > + tp->tp_conn = conn; > + tp->tp_dport = dport; > + spin_lock(&rds_tcp_pong_lock); > + list_add_tail(&tp->tp_node,&rds_tcp_pong_list); > + spin_unlock(&rds_tcp_pong_lock); > + schedule_work(&rds_tcp_pong_work); > + > +out: > + return ret; > +} > + > +/* > + * Reply to a ping package, TCP only. > + */ > +int rds_send_pong(struct rds_connection *conn, __be16 dport) > +{ > + return rds_tcp_pong(conn, dport); > +} > diff --git a/net/rds/tcp.h b/net/rds/tcp.h > index 9cf2927..c4c7e01 100644 > --- a/net/rds/tcp.h > +++ b/net/rds/tcp.h > @@ -3,6 +3,13 @@ > > #define RDS_TCP_PORT 16385 > > +/* RDS over TCP ping/pong message entry */ > +struct rds_tcp_pong { > + struct list_head tp_node; > + struct rds_connection *tp_conn; > + __be16 tp_dport; > +}; > + > struct rds_tcp_incoming { > struct rds_incoming ti_inc; > struct sk_buff_head ti_skb_list; Hi Jeff, I was looking at the history of changes to rds_send_pong. At one time rds_send_pong did this to transmit the pong message: queue_delayed_work(rds_wq, &conn->c_send_w, 0); instead of the current ret = rds_send_xmit(conn); i.e. the older versions did not have the deadlock problem and used to work once. ;-) I have suggestions for fixing it in a couple of other ways which you may want to consider to reduce the amount of code changes in a transport independent layer such as "send.c" for a specific underlying transport (tcp in this case). 1. One option is to move back to the old way for all transports (IB, tcp, loopback) since queuing delay shouldn't be an issue for a diagnostic tool like rds-ping which is typically used just to test the connectivity and not for serious performance measurements. 2. The underlying transport such as IB, loopback,TCP tells which method it wants to send the pong: queued way or send immediately. And the code change in rds_send_pong could then simply be: if (conn->c_flags & QUEUE_PONG) queue_delayed_work(rds_wq, &conn->c_send_w,0); else ret = rds_send_xmit(conn); (The above example codes are not complete. You will need to propagate this new flag to "conn" from "rds_transport", etc. at connection setup time) Venkat -- To unsubscribe from this list: send the line "unsubscribe netdev" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
Hi Venkat, On 10/09/12 00:47, Venkat Venkatsubra wrote: > On 10/6/2012 12:42 AM, Jeff Liu wrote: >> Hello, >> >> RDS ping/pong over TCP feature has broke for years(2.6.39 to 3.6.0) >> since we have to set TCP cork and >> call kerenel_sendmsg() to reply a ping requirement which both need to >> lock "struct sock *sk". >> However, this lock has already been hold before our >> rda_tcp_data_ready() callback is triggerred. >> As a result, we always facing spinlock recursion which would >> resulting in system panic... >> >> Given that RDS ping is a special kind of message, we don't need to >> reply it as >> soon as possible, IMHO, we can schedule it to work queue as a delayed >> response to >> make TCP transport totally works. Also, I think we can using the >> system default >> work queue to serve it to reduce the possible impact on general TCP >> transmit. >> > Hi Jeff, > > I was looking at the history of changes to rds_send_pong. > At one time rds_send_pong did this to transmit the pong message: > queue_delayed_work(rds_wq, &conn->c_send_w, 0); > instead of the current > ret = rds_send_xmit(conn); > i.e. the older versions did not have the deadlock problem and used to > work once. ;-) > > I have suggestions for fixing it in a couple of other ways which you > may want to consider > to reduce the amount of code changes in a transport independent layer > such as "send.c" for a specific underlying transport (tcp in this case). Thanks for the feedback! > > 1. One option is to move back to the old way for all transports (IB, > tcp, loopback) since > queuing delay shouldn't be an issue for a diagnostic tool like > rds-ping which is typically used just to test the connectivity > and not for serious performance measurements. So I prefer to your first suggestions since I have also tried to fix this issue in this way at that time, of course, it really works fine. :) And also, it could keep the code change as little as possible. I changed my mind to queue pong message to the system default queue as it might reduce the impact on general RDS TCP transmits, but I turned out to be a bit overkill and with more code changes. I'll send out the V2 patch for your review after a little while. Thanks, -Jeff > > 2. The underlying transport such as IB, loopback,TCP tells which > method it wants to send the pong: queued way or send immediately. > And the code change in rds_send_pong could then simply be: > if (conn->c_flags & QUEUE_PONG) > queue_delayed_work(rds_wq, &conn->c_send_w,0); > else > ret = rds_send_xmit(conn); > (The above example codes are not complete. You will need to propagate > this new flag to "conn" from "rds_transport", etc. at connection setup > time) > > Venkat -- To unsubscribe from this list: send the line "unsubscribe netdev" in the body of a message to majordomo@vger.kernel.org More majordomo info at http://vger.kernel.org/majordomo-info.html
diff --git a/net/rds/send.c b/net/rds/send.c index 96531d4..011006e 100644 --- a/net/rds/send.c +++ b/net/rds/send.c @@ -38,8 +38,10 @@ #include <linux/list.h> #include <linux/ratelimit.h> #include <linux/export.h> +#include <linux/workqueue.h> #include "rds.h" +#include "tcp.h" /* When transmitting messages in rds_send_xmit, we need to emerge from * time to time and briefly release the CPU. Otherwise the softlock watchdog @@ -55,6 +57,12 @@ static int send_batch_count = 64; module_param(send_batch_count, int, 0444); MODULE_PARM_DESC(send_batch_count, " batch factor when working the send queue"); +/* RDS over TCP ping/pong */ +static void rds_tcp_pong_worker(struct work_struct *work); +static DEFINE_SPINLOCK(rds_tcp_pong_lock); +static LIST_HEAD(rds_tcp_pong_list); +static DECLARE_WORK(rds_tcp_pong_work, rds_tcp_pong_worker); + static void rds_send_remove_from_sock(struct list_head *messages, int status); /* @@ -1082,11 +1090,7 @@ out: return ret; } -/* - * Reply to a ping packet. - */ -int -rds_send_pong(struct rds_connection *conn, __be16 dport) +static int rds_tcp_send_pong(struct rds_connection *conn, __be16 dport) { struct rds_message *rm; unsigned long flags; @@ -1132,3 +1136,69 @@ out: rds_message_put(rm); return ret; } + +static void rds_tcp_pong_worker(struct work_struct *work) +{ + struct rds_tcp_pong *tp; + struct rds_connection *conn; + __be16 dport; + + spin_lock(&rds_tcp_pong_lock); + if (list_empty(&rds_tcp_pong_list)) + goto out_unlock; + + /* + * Process on tcp pong once one time to reduce the possbile impact + * on normal transmit. + */ + tp = list_entry(rds_tcp_pong_list.next, struct rds_tcp_pong, tp_node); + conn = tp->tp_conn; + dport = tp->tp_dport; + list_del(&tp->tp_node); + spin_unlock(&rds_tcp_pong_lock); + + kfree(tp); + rds_tcp_send_pong(conn, dport); + goto out; + +out_unlock: + spin_unlock(&rds_tcp_pong_lock); +out: + return; +} + +/* + * RDS over TCP transport support ping/pong message. However, it + * always resulting in sock spinlock recursion up to 3.7.0. To solve + * this issue, we can shedule it to work queue as a delayed response. + * Here we using the system default work queue. + */ +int rds_tcp_pong(struct rds_connection *conn, __be16 dport) +{ + struct rds_tcp_pong *tp; + int ret = 0; + + tp = kmalloc(sizeof(*tp), GFP_ATOMIC); + if (!tp) { + ret = -ENOMEM; + goto out; + } + + tp->tp_conn = conn; + tp->tp_dport = dport; + spin_lock(&rds_tcp_pong_lock); + list_add_tail(&tp->tp_node, &rds_tcp_pong_list); + spin_unlock(&rds_tcp_pong_lock); + schedule_work(&rds_tcp_pong_work); + +out: + return ret; +} + +/* + * Reply to a ping package, TCP only. + */ +int rds_send_pong(struct rds_connection *conn, __be16 dport) +{ + return rds_tcp_pong(conn, dport); +} diff --git a/net/rds/tcp.h b/net/rds/tcp.h index 9cf2927..c4c7e01 100644 --- a/net/rds/tcp.h +++ b/net/rds/tcp.h @@ -3,6 +3,13 @@ #define RDS_TCP_PORT 16385 +/* RDS over TCP ping/pong message entry */ +struct rds_tcp_pong { + struct list_head tp_node; + struct rds_connection *tp_conn; + __be16 tp_dport; +}; + struct rds_tcp_incoming { struct rds_incoming ti_inc; struct sk_buff_head ti_skb_list;
Hello, RDS ping/pong over TCP feature has broke for years(2.6.39 to 3.6.0) since we have to set TCP cork and call kerenel_sendmsg() to reply a ping requirement which both need to lock "struct sock *sk". However, this lock has already been hold before our rda_tcp_data_ready() callback is triggerred. As a result, we always facing spinlock recursion which would resulting in system panic... Given that RDS ping is a special kind of message, we don't need to reply it as soon as possible, IMHO, we can schedule it to work queue as a delayed response to make TCP transport totally works. Also, I think we can using the system default work queue to serve it to reduce the possible impact on general TCP transmit. With below patch, I have run rds-ping(run multiple ping between two hosts at the same time) and rds-stress(hostA listen, hostB send packets) for half day, it works to me. Thanks, -Jeff Reported-by: Dan Carpenter <dan.carpenter@oracle.com> CC: Venkat Venkatsubra <venkat.x.venkatsubra@oracle.com> CC: David S. Miller <davem@davemloft.net> CC: James Morris <james.l.morris@oracle.com> Signed-off-by: Jie Liu <jeff.liu@oracle.com> --- net/rds/send.c | 80 ++++++++++++++++++++++++++++++++++++++++++++++++++++---- net/rds/tcp.h | 7 +++++ 2 files changed, 82 insertions(+), 5 deletions(-)