From patchwork Wed Mar 4 15:02:09 2020 Content-Type: text/plain; charset="utf-8" MIME-Version: 1.0 Content-Transfer-Encoding: 7bit X-Patchwork-Submitter: Paolo Abeni X-Patchwork-Id: 1249060 X-Patchwork-Delegate: mathew.j.martineau@linux.intel.com Return-Path: X-Original-To: incoming@patchwork.ozlabs.org Delivered-To: patchwork-incoming@bilbo.ozlabs.org Authentication-Results: ozlabs.org; spf=none (no SPF record) smtp.mailfrom=lists.01.org (client-ip=2001:19d0:306:5::1; helo=ml01.01.org; envelope-from=mptcp-bounces@lists.01.org; receiver=) Authentication-Results: ozlabs.org; dmarc=fail (p=none dis=none) header.from=redhat.com Authentication-Results: ozlabs.org; dkim=fail reason="signature verification failed" (1024-bit key; unprotected) header.d=redhat.com header.i=@redhat.com header.a=rsa-sha256 header.s=mimecast20190719 header.b=gC4Sg6kQ; dkim-atps=neutral Received: from ml01.01.org (ml01.01.org [IPv6:2001:19d0:306:5::1]) (using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits) key-exchange X25519 server-signature RSA-PSS (4096 bits)) (No client certificate requested) by ozlabs.org (Postfix) with ESMTPS id 48XcYl1cmVz9sR4 for ; Thu, 5 Mar 2020 02:02:35 +1100 (AEDT) Received: from ml01.vlan13.01.org (localhost [IPv6:::1]) by ml01.01.org (Postfix) with ESMTP id EFA5710FC36FB; Wed, 4 Mar 2020 07:03:24 -0800 (PST) Received-SPF: Pass (mailfrom) identity=mailfrom; client-ip=207.211.31.120; helo=us-smtp-1.mimecast.com; envelope-from=pabeni@redhat.com; receiver= Received: from us-smtp-1.mimecast.com (us-smtp-delivery-1.mimecast.com [207.211.31.120]) (using TLSv1.2 with cipher ECDHE-RSA-AES256-SHA384 (256/256 bits)) (No client certificate requested) by ml01.01.org (Postfix) with ESMTPS id 9D26610FC36D7 for ; Wed, 4 Mar 2020 07:03:23 -0800 (PST) DKIM-Signature: v=1; a=rsa-sha256; c=relaxed/relaxed; d=redhat.com; s=mimecast20190719; t=1583334151; h=from:from:reply-to:subject:subject:date:date:message-id:message-id: to:to:cc:mime-version:mime-version:content-type:content-type: content-transfer-encoding:content-transfer-encoding: in-reply-to:in-reply-to:references:references; bh=rSr3rCgUrOXnpT3IY0bTKF5Ydbv0MQy7BxY7gJTKb9U=; b=gC4Sg6kQCRv73uHrpap2yvzWAstJByWSpWS9vkQSGOdYne5zU5PfVEWarpB2qEEsIl/n0x qr3llyrX5ltxuk02oqzESHbi38lF+TSGR6F6Vm3p38PxAVAXk81QJ7BLe151PV/T6mNmuQ nySS9PvdvEjkEVDbOVEclWX2YkkwNFw= Received: from mimecast-mx01.redhat.com (mimecast-mx01.redhat.com [209.132.183.4]) (Using TLS) by relay.mimecast.com with ESMTP id us-mta-21-Q1mz5Yc-Nie5pBfcxLqHPw-1; Wed, 04 Mar 2020 10:02:27 -0500 X-MC-Unique: Q1mz5Yc-Nie5pBfcxLqHPw-1 Received: from smtp.corp.redhat.com (int-mx01.intmail.prod.int.phx2.redhat.com [10.5.11.11]) (using TLSv1.2 with cipher AECDH-AES256-SHA (256/256 bits)) (No client certificate requested) by mimecast-mx01.redhat.com (Postfix) with ESMTPS id A96E6100550E for ; Wed, 4 Mar 2020 15:02:26 +0000 (UTC) Received: from localhost.localdomain.com (unknown [10.36.118.78]) by smtp.corp.redhat.com (Postfix) with ESMTP id 1E40519E9C for ; Wed, 4 Mar 2020 15:02:25 +0000 (UTC) From: Paolo Abeni To: mptcp@lists.01.org Date: Wed, 4 Mar 2020 16:02:09 +0100 Message-Id: In-Reply-To: References: MIME-Version: 1.0 X-Scanned-By: MIMEDefang 2.79 on 10.5.11.11 X-Mimecast-Spam-Score: 0 X-Mimecast-Originator: redhat.com Message-ID-Hash: YYOV4J4CRGIBPAKGJQOTBKR3RNQC2BVG X-Message-ID-Hash: YYOV4J4CRGIBPAKGJQOTBKR3RNQC2BVG X-MailFrom: pabeni@redhat.com X-Mailman-Rule-Misses: dmarc-mitigation; no-senders; approved; emergency; loop; banned-address; member-moderation; nonmember-moderation; administrivia; implicit-dest; max-recipients; max-size; news-moderation; no-subject; suspicious-header X-Mailman-Version: 3.1.1 Precedence: list Subject: [MPTCP] [PATCH v1 10/0010] Squash-to: "mptcp: add netlink based PM" List-Id: Discussions regarding MPTCP upstreaming Archived-At: List-Archive: List-Help: List-Post: List-Subscribe: List-Unsubscribe: Implement support for max subflow limits, updating the APIs. Be sure to avoid any pending work if the subflows limit will block that. Fix an issue in local address lookup: we must traverse also the join_list. Signed-off-by: Paolo Abeni --- include/uapi/linux/mptcp.h | 5 +- net/mptcp/pm_netlink.c | 93 +++++++++++++++++++++++++++----------- 2 files changed, 69 insertions(+), 29 deletions(-) diff --git a/include/uapi/linux/mptcp.h b/include/uapi/linux/mptcp.h index 92414f3d1463..d061c5a6ae05 100644 --- a/include/uapi/linux/mptcp.h +++ b/include/uapi/linux/mptcp.h @@ -45,6 +45,7 @@ enum mptcp_pm_attrs { MPTCP_PM_ATTR_ADDR, /* nested address */ MPTCP_PM_ATTR_RCV_ADD_ADDRS, /* u32 */ + MPTCP_PM_ATTR_SUBFLOWS, /* u32 */ __MPTCP_PM_ATTR_MAX }; @@ -78,8 +79,8 @@ enum { MPTCP_PM_CMD_DEL_ADDR, MPTCP_PM_CMD_GET_ADDR, MPTCP_PM_CMD_FLUSH_ADDRS, - MPTCP_PM_CMD_SET_RCV_ADD_ADDRS, - MPTCP_PM_CMD_GET_RCV_ADD_ADDRS, + MPTCP_PM_CMD_SET_LIMITS, + MPTCP_PM_CMD_GET_LIMITS, __MPTCP_PM_CMD_AFTER_LAST }; diff --git a/net/mptcp/pm_netlink.c b/net/mptcp/pm_netlink.c index 0887fae982fc..adfea3b85606 100644 --- a/net/mptcp/pm_netlink.c +++ b/net/mptcp/pm_netlink.c @@ -35,6 +35,7 @@ struct pm_nl_pernet { unsigned int add_addr_signal_max; unsigned int add_addr_accept_max; unsigned int local_addr_max; + unsigned int subflows_max; unsigned int next_id; }; @@ -89,14 +90,14 @@ static void remote_address(const struct sock_common *skc, #endif } -static bool lookup_subflow_by_saddr(const struct mptcp_sock *msk, +static bool lookup_subflow_by_saddr(const struct list_head *list, struct mptcp_addr_info *saddr) { struct mptcp_subflow_context *subflow; struct mptcp_addr_info cur; struct sock_common *skc; - list_for_each_entry(subflow, &msk->conn_list, node) { + list_for_each_entry(subflow, list, node) { skc = (struct sock_common *)mptcp_subflow_tcp_sock(subflow); local_address(skc, &cur); @@ -109,20 +110,27 @@ static bool lookup_subflow_by_saddr(const struct mptcp_sock *msk, static struct mptcp_pm_addr_entry * select_local_address(const struct pm_nl_pernet *pernet, - const struct mptcp_sock *msk) + struct mptcp_sock *msk) { struct mptcp_pm_addr_entry *entry, *ret = NULL; rcu_read_lock(); + spin_lock_bh(&msk->join_list_lock); list_for_each_entry_rcu(entry, &pernet->local_addr_list, list) { if (!(entry->flags & MPTCP_PM_ADDR_FLAG_SUBFLOW)) continue; + + /* avoid any address already in use by subflows and + * pending join + */ if (entry->addr.family == ((struct sock *)msk)->sk_family && - !lookup_subflow_by_saddr(msk, &entry->addr)) { + !lookup_subflow_by_saddr(&msk->conn_list, &entry->addr) && + !lookup_subflow_by_saddr(&msk->join_list, &entry->addr)) { ret = entry; break; } } + spin_unlock_bh(&msk->join_list_lock); rcu_read_unlock(); return ret; } @@ -154,7 +162,8 @@ select_signal_address(struct pm_nl_pernet *pernet, unsigned int pos) static void check_work_pending(struct mptcp_sock *msk) { if (msk->pm.add_addr_signaled == msk->pm.add_addr_signal_max && - msk->pm.local_addr_used == msk->pm.local_addr_max) + (msk->pm.local_addr_used == msk->pm.local_addr_max || + (msk->pm.subflows == msk->pm.subflows_max))) WRITE_ONCE(msk->pm.work_pending, false); } @@ -192,12 +201,14 @@ static void mptcp_pm_create_subflow_or_signal_addr(struct mptcp_sock *msk) } /* check if should create a new subflow */ - if (msk->pm.local_addr_used < msk->pm.local_addr_max) { + if (msk->pm.local_addr_used < msk->pm.local_addr_max && + msk->pm.subflows < msk->pm.subflows_max) { remote_address((struct sock_common *)sk, &remote); local = select_local_address(pernet, msk); if (local) { msk->pm.local_addr_used++; + msk->pm.subflows++; check_work_pending(msk); spin_unlock_bh(&msk->pm.lock); __mptcp_subflow_connect(sk, local->ifindex, @@ -239,7 +250,10 @@ void mptcp_pm_nl_add_addr_received(struct mptcp_sock *msk) pr_debug("accepted %d:%d remote family %d", msk->pm.add_addr_accepted, msk->pm.add_addr_accept_max, msk->pm.remote.family); - if (++msk->pm.add_addr_accepted >= msk->pm.add_addr_accept_max) + msk->pm.add_addr_accepted++; + msk->pm.subflows++; + if (msk->pm.add_addr_accepted >= msk->pm.add_addr_accept_max || + msk->pm.subflows >= msk->pm.subflows_max) WRITE_ONCE(msk->pm.accept_addr, false); /* connect to the specified remote address, using whatever @@ -352,16 +366,21 @@ int mptcp_pm_nl_get_local_id(struct mptcp_sock *msk, struct sock_common *skc) void mptcp_pm_nl_data_init(struct mptcp_sock *msk) { + struct mptcp_pm_data *pm = &msk->pm; struct pm_nl_pernet *pernet; + bool subflows; pernet = net_generic(sock_net((struct sock *)msk), pm_nl_pernet_id); - msk->pm.add_addr_signal_max = READ_ONCE(pernet->add_addr_signal_max); - msk->pm.add_addr_accept_max = READ_ONCE(pernet->add_addr_accept_max); - msk->pm.local_addr_max = READ_ONCE(pernet->local_addr_max); - WRITE_ONCE(msk->pm.work_pending, - !!msk->pm.local_addr_max || !!msk->pm.add_addr_signal_max); - WRITE_ONCE(msk->pm.accept_addr, !!msk->pm.add_addr_accept_max); + pm->add_addr_signal_max = READ_ONCE(pernet->add_addr_signal_max); + pm->add_addr_accept_max = READ_ONCE(pernet->add_addr_accept_max); + pm->local_addr_max = READ_ONCE(pernet->local_addr_max); + pm->subflows_max = READ_ONCE(pernet->subflows_max); + subflows = !!pm->subflows_max; + WRITE_ONCE(pm->work_pending, (!!pm->local_addr_max && subflows) || + !!pm->add_addr_signal_max); + WRITE_ONCE(pm->accept_addr, !!pm->add_addr_accept_max && subflows); + WRITE_ONCE(pm->accept_subflow, subflows); } #define MPTCP_PM_CMD_GRP_OFFSET 0 @@ -386,6 +405,7 @@ static const struct nla_policy mptcp_pm_policy[MPTCP_PM_ATTR_MAX + 1] = { [MPTCP_PM_ATTR_ADDR] = NLA_POLICY_NESTED(mptcp_pm_addr_policy), [MPTCP_PM_ATTR_RCV_ADD_ADDRS] = { .type = NLA_U32, }, + [MPTCP_PM_ATTR_SUBFLOWS] = { .type = NLA_U32, }, }; static int mptcp_pm_family_to_addr(int family) @@ -690,31 +710,46 @@ static int mptcp_nl_cmd_dump_addrs(struct sk_buff *msg, return msg->len; } -static int -mptcp_nl_cmd_set_rcv_add_addrs(struct sk_buff *skb, struct genl_info *info) +static int parse_limit(struct genl_info *info, int id, unsigned int *limit) { - struct nlattr *attr = info->attrs[MPTCP_PM_ATTR_RCV_ADD_ADDRS]; - struct pm_nl_pernet *pernet = genl_info_pm_nl(info); - int limit; + struct nlattr *attr = info->attrs[id]; if (!attr) { GENL_SET_ERR_MSG(info, "missing announce accept limit"); return -EINVAL; } - limit = nla_get_u32(attr); - if (limit > MPTCP_PM_ADDR_MAX) { + *limit = nla_get_u32(attr); + if (*limit > MPTCP_PM_ADDR_MAX) { GENL_SET_ERR_MSG(info, "announce accept limit greater than maximum"); return -EINVAL; } + return 0; +} - WRITE_ONCE(pernet->add_addr_accept_max, limit); +static int +mptcp_nl_cmd_set_limits(struct sk_buff *skb, struct genl_info *info) +{ + struct pm_nl_pernet *pernet = genl_info_pm_nl(info); + unsigned int rcv_addrs, subflows; + int ret; + + ret = parse_limit(info, MPTCP_PM_ATTR_RCV_ADD_ADDRS, &rcv_addrs); + if (ret) + return ret; + + ret = parse_limit(info, MPTCP_PM_ATTR_SUBFLOWS, &subflows); + if (ret) + return ret; + + WRITE_ONCE(pernet->add_addr_accept_max, rcv_addrs); + WRITE_ONCE(pernet->subflows_max, subflows); return 0; } static int -mptcp_nl_cmd_get_rcv_add_addrs(struct sk_buff *skb, struct genl_info *info) +mptcp_nl_cmd_get_limits(struct sk_buff *skb, struct genl_info *info) { struct pm_nl_pernet *pernet = genl_info_pm_nl(info); struct sk_buff *msg; @@ -725,7 +760,7 @@ mptcp_nl_cmd_get_rcv_add_addrs(struct sk_buff *skb, struct genl_info *info) return -ENOMEM; reply = genlmsg_put_reply(msg, info, &mptcp_genl_family, 0, - MPTCP_PM_CMD_GET_RCV_ADD_ADDRS); + MPTCP_PM_CMD_GET_LIMITS); if (!reply) goto fail; @@ -733,6 +768,10 @@ mptcp_nl_cmd_get_rcv_add_addrs(struct sk_buff *skb, struct genl_info *info) READ_ONCE(pernet->add_addr_accept_max))) goto fail; + if (nla_put_u32(msg, MPTCP_PM_ATTR_SUBFLOWS, + READ_ONCE(pernet->subflows_max))) + goto fail; + genlmsg_end(msg, reply); return genlmsg_reply(msg, info); @@ -765,13 +804,13 @@ static struct genl_ops mptcp_pm_ops[] = { .flags = GENL_ADMIN_PERM, }, { - .cmd = MPTCP_PM_CMD_SET_RCV_ADD_ADDRS, - .doit = mptcp_nl_cmd_set_rcv_add_addrs, + .cmd = MPTCP_PM_CMD_SET_LIMITS, + .doit = mptcp_nl_cmd_set_limits, .flags = GENL_ADMIN_PERM, }, { - .cmd = MPTCP_PM_CMD_GET_RCV_ADD_ADDRS, - .doit = mptcp_nl_cmd_get_rcv_add_addrs, + .cmd = MPTCP_PM_CMD_GET_LIMITS, + .doit = mptcp_nl_cmd_get_limits, .flags = GENL_ADMIN_PERM, }, };