Message ID | 20190411.150207.1202792333614364121.davem@davemloft.net |
---|---|
State | Accepted |
Delegated to: | David Miller |
Headers | show |
Series | SCTP: Event skb list overhaul. | expand |
On Thu, Apr 11, 2019 at 03:02:07PM -0700, David Miller wrote: > > Now the SKB list implementation assumption can be removed. > > And now that we know that the list head is always non-NULL > we can remove the code blocks dealing with that as well. > > Signed-off-by: David S. Miller <davem@davemloft.net> Acked-by: Marcelo Ricardo Leitner <marcelo.leitner@gmail.com> Thanks! > --- > include/net/sctp/ulpqueue.h | 2 +- > net/sctp/stream_interleave.c | 2 +- > net/sctp/ulpqueue.c | 29 +++++++++++------------------ > 3 files changed, 13 insertions(+), 20 deletions(-) > > diff --git a/include/net/sctp/ulpqueue.h b/include/net/sctp/ulpqueue.h > index bb0ecba3db2b..f4ac7117ff29 100644 > --- a/include/net/sctp/ulpqueue.h > +++ b/include/net/sctp/ulpqueue.h > @@ -59,7 +59,7 @@ void sctp_ulpq_free(struct sctp_ulpq *); > int sctp_ulpq_tail_data(struct sctp_ulpq *, struct sctp_chunk *, gfp_t); > > /* Add a new event for propagation to the ULP. */ > -int sctp_ulpq_tail_event(struct sctp_ulpq *, struct sctp_ulpevent *ev); > +int sctp_ulpq_tail_event(struct sctp_ulpq *, struct sk_buff_head *skb_list); > > /* Renege previously received chunks. */ > void sctp_ulpq_renege(struct sctp_ulpq *, struct sctp_chunk *, gfp_t); > diff --git a/net/sctp/stream_interleave.c b/net/sctp/stream_interleave.c > index 2c50627bdfdb..25e0b7e5189c 100644 > --- a/net/sctp/stream_interleave.c > +++ b/net/sctp/stream_interleave.c > @@ -1317,7 +1317,7 @@ static int do_ulpq_tail_event(struct sctp_ulpq *ulpq, struct sctp_ulpevent *even > > skb_queue_head_init(&temp); > __skb_queue_tail(&temp, sctp_event2skb(event)); > - return sctp_ulpq_tail_event(ulpq, event); > + return sctp_ulpq_tail_event(ulpq, &temp); > } > > static struct sctp_stream_interleave sctp_stream_interleave_0 = { > diff --git a/net/sctp/ulpqueue.c b/net/sctp/ulpqueue.c > index a698f1a509bf..7cdc3623fa35 100644 > --- a/net/sctp/ulpqueue.c > +++ b/net/sctp/ulpqueue.c > @@ -130,7 +130,7 @@ int sctp_ulpq_tail_data(struct sctp_ulpq *ulpq, struct sctp_chunk *chunk, > */ > if (event) { > event_eor = (event->msg_flags & MSG_EOR) ? 1 : 0; > - sctp_ulpq_tail_event(ulpq, event); > + sctp_ulpq_tail_event(ulpq, &temp); > } > > return event_eor; > @@ -194,18 +194,17 @@ static int sctp_ulpq_clear_pd(struct sctp_ulpq *ulpq) > return sctp_clear_pd(ulpq->asoc->base.sk, ulpq->asoc); > } > > -/* If the SKB of 'event' is on a list, it is the first such member > - * of that list. > - */ > -int sctp_ulpq_tail_event(struct sctp_ulpq *ulpq, struct sctp_ulpevent *event) > +int sctp_ulpq_tail_event(struct sctp_ulpq *ulpq, struct sk_buff_head *skb_list) > { > struct sock *sk = ulpq->asoc->base.sk; > struct sctp_sock *sp = sctp_sk(sk); > - struct sk_buff_head *queue, *skb_list; > - struct sk_buff *skb = sctp_event2skb(event); > + struct sctp_ulpevent *event; > + struct sk_buff_head *queue; > + struct sk_buff *skb; > int clear_pd = 0; > > - skb_list = (struct sk_buff_head *) skb->prev; > + skb = __skb_peek(skb_list); > + event = sctp_skb2event(skb); > > /* If the socket is just going to throw this away, do not > * even try to deliver it. > @@ -258,13 +257,7 @@ int sctp_ulpq_tail_event(struct sctp_ulpq *ulpq, struct sctp_ulpevent *event) > } > } > > - /* If we are harvesting multiple skbs they will be > - * collected on a list. > - */ > - if (skb_list) > - skb_queue_splice_tail_init(skb_list, queue); > - else > - __skb_queue_tail(queue, skb); > + skb_queue_splice_tail_init(skb_list, queue); > > /* Did we just complete partial delivery and need to get > * rolling again? Move pending data to the receive > @@ -757,7 +750,7 @@ static void sctp_ulpq_reasm_drain(struct sctp_ulpq *ulpq) > * sctp_ulpevent for very first SKB on the temp' list. > */ > if (event) > - sctp_ulpq_tail_event(ulpq, event); > + sctp_ulpq_tail_event(ulpq, &temp); > } > } > > @@ -957,7 +950,7 @@ static void sctp_ulpq_reap_ordered(struct sctp_ulpq *ulpq, __u16 sid) > if (event) { > /* see if we have more ordered that we can deliver */ > sctp_ulpq_retrieve_ordered(ulpq, event); > - sctp_ulpq_tail_event(ulpq, event); > + sctp_ulpq_tail_event(ulpq, &temp); > } > } > > @@ -1087,7 +1080,7 @@ void sctp_ulpq_partial_delivery(struct sctp_ulpq *ulpq, > > skb_queue_head_init(&temp); > __skb_queue_tail(&temp, sctp_event2skb(event)); > - sctp_ulpq_tail_event(ulpq, event); > + sctp_ulpq_tail_event(ulpq, &temp); > sctp_ulpq_set_pd(ulpq); > return; > } > -- > 2.20.1 >
diff --git a/include/net/sctp/ulpqueue.h b/include/net/sctp/ulpqueue.h index bb0ecba3db2b..f4ac7117ff29 100644 --- a/include/net/sctp/ulpqueue.h +++ b/include/net/sctp/ulpqueue.h @@ -59,7 +59,7 @@ void sctp_ulpq_free(struct sctp_ulpq *); int sctp_ulpq_tail_data(struct sctp_ulpq *, struct sctp_chunk *, gfp_t); /* Add a new event for propagation to the ULP. */ -int sctp_ulpq_tail_event(struct sctp_ulpq *, struct sctp_ulpevent *ev); +int sctp_ulpq_tail_event(struct sctp_ulpq *, struct sk_buff_head *skb_list); /* Renege previously received chunks. */ void sctp_ulpq_renege(struct sctp_ulpq *, struct sctp_chunk *, gfp_t); diff --git a/net/sctp/stream_interleave.c b/net/sctp/stream_interleave.c index 2c50627bdfdb..25e0b7e5189c 100644 --- a/net/sctp/stream_interleave.c +++ b/net/sctp/stream_interleave.c @@ -1317,7 +1317,7 @@ static int do_ulpq_tail_event(struct sctp_ulpq *ulpq, struct sctp_ulpevent *even skb_queue_head_init(&temp); __skb_queue_tail(&temp, sctp_event2skb(event)); - return sctp_ulpq_tail_event(ulpq, event); + return sctp_ulpq_tail_event(ulpq, &temp); } static struct sctp_stream_interleave sctp_stream_interleave_0 = { diff --git a/net/sctp/ulpqueue.c b/net/sctp/ulpqueue.c index a698f1a509bf..7cdc3623fa35 100644 --- a/net/sctp/ulpqueue.c +++ b/net/sctp/ulpqueue.c @@ -130,7 +130,7 @@ int sctp_ulpq_tail_data(struct sctp_ulpq *ulpq, struct sctp_chunk *chunk, */ if (event) { event_eor = (event->msg_flags & MSG_EOR) ? 1 : 0; - sctp_ulpq_tail_event(ulpq, event); + sctp_ulpq_tail_event(ulpq, &temp); } return event_eor; @@ -194,18 +194,17 @@ static int sctp_ulpq_clear_pd(struct sctp_ulpq *ulpq) return sctp_clear_pd(ulpq->asoc->base.sk, ulpq->asoc); } -/* If the SKB of 'event' is on a list, it is the first such member - * of that list. - */ -int sctp_ulpq_tail_event(struct sctp_ulpq *ulpq, struct sctp_ulpevent *event) +int sctp_ulpq_tail_event(struct sctp_ulpq *ulpq, struct sk_buff_head *skb_list) { struct sock *sk = ulpq->asoc->base.sk; struct sctp_sock *sp = sctp_sk(sk); - struct sk_buff_head *queue, *skb_list; - struct sk_buff *skb = sctp_event2skb(event); + struct sctp_ulpevent *event; + struct sk_buff_head *queue; + struct sk_buff *skb; int clear_pd = 0; - skb_list = (struct sk_buff_head *) skb->prev; + skb = __skb_peek(skb_list); + event = sctp_skb2event(skb); /* If the socket is just going to throw this away, do not * even try to deliver it. @@ -258,13 +257,7 @@ int sctp_ulpq_tail_event(struct sctp_ulpq *ulpq, struct sctp_ulpevent *event) } } - /* If we are harvesting multiple skbs they will be - * collected on a list. - */ - if (skb_list) - skb_queue_splice_tail_init(skb_list, queue); - else - __skb_queue_tail(queue, skb); + skb_queue_splice_tail_init(skb_list, queue); /* Did we just complete partial delivery and need to get * rolling again? Move pending data to the receive @@ -757,7 +750,7 @@ static void sctp_ulpq_reasm_drain(struct sctp_ulpq *ulpq) * sctp_ulpevent for very first SKB on the temp' list. */ if (event) - sctp_ulpq_tail_event(ulpq, event); + sctp_ulpq_tail_event(ulpq, &temp); } } @@ -957,7 +950,7 @@ static void sctp_ulpq_reap_ordered(struct sctp_ulpq *ulpq, __u16 sid) if (event) { /* see if we have more ordered that we can deliver */ sctp_ulpq_retrieve_ordered(ulpq, event); - sctp_ulpq_tail_event(ulpq, event); + sctp_ulpq_tail_event(ulpq, &temp); } } @@ -1087,7 +1080,7 @@ void sctp_ulpq_partial_delivery(struct sctp_ulpq *ulpq, skb_queue_head_init(&temp); __skb_queue_tail(&temp, sctp_event2skb(event)); - sctp_ulpq_tail_event(ulpq, event); + sctp_ulpq_tail_event(ulpq, &temp); sctp_ulpq_set_pd(ulpq); return; }
Now the SKB list implementation assumption can be removed. And now that we know that the list head is always non-NULL we can remove the code blocks dealing with that as well. Signed-off-by: David S. Miller <davem@davemloft.net> --- include/net/sctp/ulpqueue.h | 2 +- net/sctp/stream_interleave.c | 2 +- net/sctp/ulpqueue.c | 29 +++++++++++------------------ 3 files changed, 13 insertions(+), 20 deletions(-)