Patch Detail
get:
Show a patch.
patch:
Update a patch.
put:
Update a patch.
GET /api/1.2/patches/2219970/?format=api
{ "id": 2219970, "url": "http://patchwork.ozlabs.org/api/1.2/patches/2219970/?format=api", "web_url": "http://patchwork.ozlabs.org/project/linuxppc-dev/patch/20260405125240.2558577-46-songmuchun@bytedance.com/", "project": { "id": 2, "url": "http://patchwork.ozlabs.org/api/1.2/projects/2/?format=api", "name": "Linux PPC development", "link_name": "linuxppc-dev", "list_id": "linuxppc-dev.lists.ozlabs.org", "list_email": "linuxppc-dev@lists.ozlabs.org", "web_url": "https://github.com/linuxppc/wiki/wiki", "scm_url": "https://git.kernel.org/pub/scm/linux/kernel/git/powerpc/linux.git", "webscm_url": "https://git.kernel.org/pub/scm/linux/kernel/git/powerpc/linux.git/", "list_archive_url": "https://lore.kernel.org/linuxppc-dev/", "list_archive_url_format": "https://lore.kernel.org/linuxppc-dev/{}/", "commit_url_format": "https://git.kernel.org/pub/scm/linux/kernel/git/powerpc/linux.git/commit/?id={}" }, "msgid": "<20260405125240.2558577-46-songmuchun@bytedance.com>", "list_archive_url": "https://lore.kernel.org/linuxppc-dev/20260405125240.2558577-46-songmuchun@bytedance.com/", "date": "2026-04-05T12:52:36", "name": "[45/49] mm/sparse-vmemmap: drop @pgmap parameter from vmemmap populate APIs", "commit_ref": null, "pull_url": null, "state": "new", "archived": false, "hash": "2d619ced1254faefce36613e045cbbf4e7a5e2e3", "submitter": { "id": 78930, "url": "http://patchwork.ozlabs.org/api/1.2/people/78930/?format=api", "name": "Muchun Song", "email": "songmuchun@bytedance.com" }, "delegate": null, "mbox": "http://patchwork.ozlabs.org/project/linuxppc-dev/patch/20260405125240.2558577-46-songmuchun@bytedance.com/mbox/", "series": [ { "id": 498783, "url": "http://patchwork.ozlabs.org/api/1.2/series/498783/?format=api", "web_url": "http://patchwork.ozlabs.org/project/linuxppc-dev/list/?series=498783", "date": "2026-04-05T12:51:51", "name": "mm: Generalize vmemmap optimization for DAX and HugeTLB", "version": 1, "mbox": "http://patchwork.ozlabs.org/series/498783/mbox/" } ], "comments": "http://patchwork.ozlabs.org/api/patches/2219970/comments/", "check": "pending", "checks": "http://patchwork.ozlabs.org/api/patches/2219970/checks/", "tags": {}, "related": [], "headers": { "Return-Path": "\n <linuxppc-dev+bounces-19375-incoming=patchwork.ozlabs.org@lists.ozlabs.org>", "X-Original-To": [ "incoming@patchwork.ozlabs.org", "linuxppc-dev@lists.ozlabs.org" ], "Delivered-To": "patchwork-incoming@legolas.ozlabs.org", "Authentication-Results": [ "legolas.ozlabs.org;\n\tdkim=pass (2048-bit key;\n unprotected) header.d=bytedance.com header.i=@bytedance.com\n header.a=rsa-sha256 header.s=google header.b=flYIdONV;\n\tdkim-atps=neutral", "legolas.ozlabs.org;\n spf=pass (sender SPF authorized) smtp.mailfrom=lists.ozlabs.org\n (client-ip=2404:9400:21b9:f100::1; helo=lists.ozlabs.org;\n envelope-from=linuxppc-dev+bounces-19375-incoming=patchwork.ozlabs.org@lists.ozlabs.org;\n receiver=patchwork.ozlabs.org)", "lists.ozlabs.org;\n arc=none smtp.remote-ip=\"2607:f8b0:4864:20::102e\"", "lists.ozlabs.org;\n dmarc=pass (p=quarantine dis=none) header.from=bytedance.com", "lists.ozlabs.org;\n\tdkim=pass (2048-bit key;\n unprotected) header.d=bytedance.com header.i=@bytedance.com\n header.a=rsa-sha256 header.s=google header.b=flYIdONV;\n\tdkim-atps=neutral", "lists.ozlabs.org;\n spf=pass (sender SPF authorized) smtp.mailfrom=bytedance.com\n (client-ip=2607:f8b0:4864:20::102e; helo=mail-pj1-x102e.google.com;\n envelope-from=songmuchun@bytedance.com; receiver=lists.ozlabs.org)" ], "Received": [ "from lists.ozlabs.org (lists.ozlabs.org\n [IPv6:2404:9400:21b9:f100::1])\n\t(using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits)\n\t key-exchange x25519)\n\t(No client certificate requested)\n\tby legolas.ozlabs.org (Postfix) with ESMTPS id 4fpXZj1HD7z1xy1\n\tfor <incoming@patchwork.ozlabs.org>; Sun, 05 Apr 2026 22:58:37 +1000 (AEST)", "from boromir.ozlabs.org (localhost [127.0.0.1])\n\tby lists.ozlabs.org (Postfix) with ESMTP id 4fpXZZ49bXz304h;\n\tSun, 05 Apr 2026 22:58:30 +1000 (AEST)", "from mail-pj1-x102e.google.com (mail-pj1-x102e.google.com\n [IPv6:2607:f8b0:4864:20::102e])\n\t(using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits)\n\t key-exchange x25519 server-signature RSA-PSS (2048 bits) server-digest\n SHA256)\n\t(No client certificate requested)\n\tby lists.ozlabs.org (Postfix) with ESMTPS id 4fpXZY1GWvz2ypk\n\tfor <linuxppc-dev@lists.ozlabs.org>; Sun, 05 Apr 2026 22:58:29 +1000 (AEST)", "by mail-pj1-x102e.google.com with SMTP id\n 98e67ed59e1d1-35d9c7bf9a1so2806004a91.3\n for <linuxppc-dev@lists.ozlabs.org>;\n Sun, 05 Apr 2026 05:58:28 -0700 (PDT)", "from n232-176-004.byted.org ([36.110.163.97])\n by smtp.gmail.com with ESMTPSA id\n 98e67ed59e1d1-35de66b4808sm3748505a91.2.2026.04.05.05.58.20\n (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256);\n Sun, 05 Apr 2026 05:58:26 -0700 (PDT)" ], "ARC-Seal": "i=1; a=rsa-sha256; d=lists.ozlabs.org; s=201707; t=1775393910;\n\tcv=none;\n b=jceIN6JOwsho0jQAjoyEg9wv9Sm/a+wp39q8QqFWjF+ICjFYrsIi/x5Jr8rFb0WPOmtEcLi/dyGkfew5eVwtRLa+gDhODZd/oe3Cq8f3Vk7AqT9kiyIpcVwuYf/FJ8OnCh5xI0sQOjC19QzcUeEExypxGl5OHEct52mwwV8g2L2pa+tGy/Hn+GoqDCk+RHy7A+Ad8v0fmycFFBsD3OwxMVzeUGNtQqlEAAHXqOFh9yXkT0jZuLgIVDLcwp8z1XyPUr5hmOMEarPkMWBd2yztzc9RwyhOhVpFlssTdwLRHrlxt4V9GVPaC7JoZhvwrf/MgBRjJt9tPilZoeXP9+Uh6Q==", "ARC-Message-Signature": "i=1; a=rsa-sha256; d=lists.ozlabs.org; s=201707;\n\tt=1775393910; c=relaxed/relaxed;\n\tbh=J13vgUIT6r3ooj2ZXZFUt9fRLd85j+bHxFhjbMlfMio=;\n\th=From:To:Cc:Subject:Date:Message-Id:In-Reply-To:References:\n\t MIME-Version;\n b=lhG2sDDEx1Hx/tD5VhA3tYeBRkcuvsJXXnltG6pgQGegXqKzIdslnvJRcQIaqN4Rb/ZiwaugNM+f9r10OUDueCzGkOv/Bi5XfdHBZaxGuJTJStuSJL7DN4cbsX/miF5Dvpgi7p5YmVfyMaqFuSI/fkLsRvKvX3bPedaY2aFvY/xFOja2e2aT5patF95Nvuz6gA/DK7aQGOMLh34F+d9YPUPHBR5zhgdbRqudzM/d4tO8GEui9Xy7mhlTLfoqWfIHhrk7OIcEw9uVQHF+z2fFCcUpAgCX/1+aEuZTjzjNXfZbRG9uCMWZKklVmKYBrQx+D4iaxFVNAwGsDLSlU+IkBg==", "ARC-Authentication-Results": "i=1; lists.ozlabs.org;\n dmarc=pass (p=quarantine dis=none) header.from=bytedance.com;\n dkim=pass (2048-bit key;\n unprotected) header.d=bytedance.com header.i=@bytedance.com\n header.a=rsa-sha256 header.s=google header.b=flYIdONV; dkim-atps=neutral;\n spf=pass (client-ip=2607:f8b0:4864:20::102e; helo=mail-pj1-x102e.google.com;\n envelope-from=songmuchun@bytedance.com;\n receiver=lists.ozlabs.org) smtp.mailfrom=bytedance.com", "DKIM-Signature": "v=1; a=rsa-sha256; c=relaxed/relaxed;\n d=bytedance.com; s=google; t=1775393907; x=1775998707;\n darn=lists.ozlabs.org;\n h=content-transfer-encoding:mime-version:references:in-reply-to\n :message-id:date:subject:cc:to:from:from:to:cc:subject:date\n :message-id:reply-to;\n bh=J13vgUIT6r3ooj2ZXZFUt9fRLd85j+bHxFhjbMlfMio=;\n b=flYIdONV8bkHrMZe/GAq84YUvU19VXyEiAHotBx7GxIDls8S0V9tx7ehChV3bkaxcS\n kNMKT2D8ddoI0WS8wE27UpOYiT2UgegYzwI8AWeyV7t4AMkdgv5ZkqOCkQH7XfhO8qi5\n w3FcR2bIBlh82JDxnw7TXHFp89uIApNr6q7Nv9XwB+sKB/7Ju6Bd/pIkzIqRQj/XUFwX\n ZoD7raXgUIptmXlSTk6Au1myVVVwBSjR+mnjY8D3qO7V24hFqfG6mv3TAluz9CgElqJr\n 4A+TkaB5juzIptAQo2MyloKNFfaJnatKLdxqTgde4kxeYoNpI9TyhcVRaGf0eDMXQvOO\n GCug==", "X-Google-DKIM-Signature": "v=1; a=rsa-sha256; c=relaxed/relaxed;\n d=1e100.net; s=20251104; t=1775393907; x=1775998707;\n h=content-transfer-encoding:mime-version:references:in-reply-to\n :message-id:date:subject:cc:to:from:x-gm-gg:x-gm-message-state:from\n :to:cc:subject:date:message-id:reply-to;\n bh=J13vgUIT6r3ooj2ZXZFUt9fRLd85j+bHxFhjbMlfMio=;\n b=oBYyqBzR7J4gYj40f82uqFUq6z4TQRpkXqixlj8K3vN0TSSUOJxniwtqhqhQahxkFD\n D0NMkBT4vvVUIFPUFcCfUOd3XvbPOpCJkje7aOLd8HEANQOIezjXPPYuoznZ84vfaSh6\n gLQ+dOyJSFH64s9qENE0yjMo132u2VrAautHrEM8shcBI2CjzD+b5p4ipQxG8WItDePh\n ZdzuLjJEYktKAHESnkPLfTm1BgNlZy4RghEu0TqA4+b0mZNTD97tyYVkDr46It/drmAj\n IT3oupkZS5Ma9Hb0QanGbiTg+wRLG9WJIekAkJlsU6uHuDYAD2V/uWu+4sHvPBz0Y6nX\n w30g==", "X-Forwarded-Encrypted": "i=1;\n AJvYcCUNaKUa8D4gLP5cLj9i8cxwyvQIgcMdOSzsREILDDkO0GuPw0kyWELznQCPCflUdms0fpnfHRz+F88yJhM=@lists.ozlabs.org", "X-Gm-Message-State": "AOJu0YwviEj3WABB55qfddWp23/ltagoNObyB9zYBJt3NrVAVCmy/4Py\n\t6ZO2PjoDHZs7ScpHjD+iaA1i1VCO8rhMvdAA7yEbz/5JiSxo4nmz7NVVnlRtrc5toUM=", "X-Gm-Gg": "AeBDieuqCFKt1brSV7UFg9CfRzZBtOHBGOVl/ZFX09+7Bzn+dGjJQj3Hw0WD/ULkaNA\n\t4aFh+/wyd+6nYsgf9snsjKLtdZb8Otptq4cTUA2AWMflILHkQPZ5hEEUI9zCq0mV7uMKAKIDks7\n\tOlE9vKaiL52JzgcMxYQ3SqS9HEnPbX8jTq2qqcjRNqsoDy/LxfjURzUo2gLDHbr4JyT9MVy2ZSP\n\tQp7SCDM1DGSKQ7D1UpAmNzGYJtzVGjP82H+34gkGnHmwoiI9aI4wYCCUotiy1MRK3dhX4U4Y4t/\n\tmillDcZPtXkYKOxug10rXGM9VL347qJT3+CDNe9wKtGpDqd/974M5lwxD/OR0n2C+Mw4BBMgPBG\n\tLo/h2tgxkcKJuq+1GsQlv2vF1QxZHweYTirDmB8tK6gZDhy7t+jZqlAfkc2y2hMOZFEWQnEU2gU\n\t7hrSEZ8RV5C/RspduwqYzYmfMVlJpc+8mSxwFPGy/98Qg=", "X-Received": "by 2002:a17:90b:314d:b0:359:f43d:4a6e with SMTP id\n 98e67ed59e1d1-35de662f67fmr9299980a91.0.1775393907102;\n Sun, 05 Apr 2026 05:58:27 -0700 (PDT)", "From": "Muchun Song <songmuchun@bytedance.com>", "To": "Andrew Morton <akpm@linux-foundation.org>,\n\tDavid Hildenbrand <david@kernel.org>,\n\tMuchun Song <muchun.song@linux.dev>,\n\tOscar Salvador <osalvador@suse.de>,\n\tMichael Ellerman <mpe@ellerman.id.au>,\n\tMadhavan Srinivasan <maddy@linux.ibm.com>", "Cc": "Lorenzo Stoakes <ljs@kernel.org>,\n\t\"Liam R . Howlett\" <Liam.Howlett@oracle.com>,\n\tVlastimil Babka <vbabka@kernel.org>,\n\tMike Rapoport <rppt@kernel.org>,\n\tSuren Baghdasaryan <surenb@google.com>,\n\tMichal Hocko <mhocko@suse.com>,\n\tNicholas Piggin <npiggin@gmail.com>,\n\tChristophe Leroy <chleroy@kernel.org>,\n\taneesh.kumar@linux.ibm.com,\n\tjoao.m.martins@oracle.com,\n\tlinux-mm@kvack.org,\n\tlinuxppc-dev@lists.ozlabs.org,\n\tlinux-kernel@vger.kernel.org,\n\tMuchun Song <songmuchun@bytedance.com>", "Subject": "[PATCH 45/49] mm/sparse-vmemmap: drop @pgmap parameter from vmemmap\n populate APIs", "Date": "Sun, 5 Apr 2026 20:52:36 +0800", "Message-Id": "<20260405125240.2558577-46-songmuchun@bytedance.com>", "X-Mailer": "git-send-email 2.20.1", "In-Reply-To": "<20260405125240.2558577-1-songmuchun@bytedance.com>", "References": "<20260405125240.2558577-1-songmuchun@bytedance.com>", "X-Mailing-List": "linuxppc-dev@lists.ozlabs.org", "List-Id": "<linuxppc-dev.lists.ozlabs.org>", "List-Help": "<mailto:linuxppc-dev+help@lists.ozlabs.org>", "List-Owner": "<mailto:linuxppc-dev+owner@lists.ozlabs.org>", "List-Post": "<mailto:linuxppc-dev@lists.ozlabs.org>", "List-Archive": "<https://lore.kernel.org/linuxppc-dev/>,\n <https://lists.ozlabs.org/pipermail/linuxppc-dev/>", "List-Subscribe": "<mailto:linuxppc-dev+subscribe@lists.ozlabs.org>,\n <mailto:linuxppc-dev+subscribe-digest@lists.ozlabs.org>,\n <mailto:linuxppc-dev+subscribe-nomail@lists.ozlabs.org>", "List-Unsubscribe": "<mailto:linuxppc-dev+unsubscribe@lists.ozlabs.org>", "Precedence": "list", "MIME-Version": "1.0", "Content-Transfer-Encoding": "8bit", "X-Spam-Status": "No, score=-0.2 required=3.0 tests=DKIM_SIGNED,DKIM_VALID,\n\tDKIM_VALID_AU,DKIM_VALID_EF,RCVD_IN_DNSWL_NONE,SPF_HELO_NONE,SPF_PASS\n\tautolearn=disabled version=4.0.1 OzLabs 8", "X-Spam-Checker-Version": "SpamAssassin 4.0.1 (2024-03-25) on lists.ozlabs.org" }, "content": "Since architecture-specific choices about vmemmap optimization are now\nhandled directly inside the vmemmap_populate() implementations, the\n@pgmap is no longer needed in the core memory hotplug APIs and most\nsparse section routines.\n\nRemove the pgmap parameter entirely from:\n- sparse_remove_section()\n- __remove_pages()\n- arch_remove_memory()\n- vmemmap_populate() and related functions\n\nThis simplifies the API a little.\n\nSigned-off-by: Muchun Song <songmuchun@bytedance.com>\n---\n arch/arm64/mm/mmu.c | 11 ++++----\n arch/loongarch/mm/init.c | 12 ++++----\n arch/powerpc/include/asm/book3s/64/radix.h | 4 +--\n arch/powerpc/mm/book3s64/radix_pgtable.c | 10 +++----\n arch/powerpc/mm/init_64.c | 4 +--\n arch/powerpc/mm/mem.c | 5 ++--\n arch/riscv/mm/init.c | 9 +++---\n arch/s390/mm/init.c | 5 ++--\n arch/s390/mm/vmem.c | 2 +-\n arch/sparc/mm/init_64.c | 5 ++--\n arch/x86/mm/init_64.c | 13 ++++-----\n include/linux/memory_hotplug.h | 8 ++----\n include/linux/mm.h | 11 +++-----\n mm/memory_hotplug.c | 12 ++++----\n mm/memremap.c | 4 +--\n mm/sparse-vmemmap.c | 33 +++++++++-------------\n mm/sparse.c | 6 ++--\n 17 files changed, 65 insertions(+), 89 deletions(-)", "diff": "diff --git a/arch/arm64/mm/mmu.c b/arch/arm64/mm/mmu.c\nindex 86162aab5185..ec1c6971a561 100644\n--- a/arch/arm64/mm/mmu.c\n+++ b/arch/arm64/mm/mmu.c\n@@ -1760,7 +1760,7 @@ int __meminit vmemmap_check_pmd(pmd_t *pmdp, int node,\n }\n \n int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node,\n-\t\tstruct vmem_altmap *altmap, struct dev_pagemap *pgmap)\n+\t\tstruct vmem_altmap *altmap)\n {\n \tWARN_ON((start < VMEMMAP_START) || (end > VMEMMAP_END));\n \t/* [start, end] should be within one section */\n@@ -1768,9 +1768,9 @@ int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node,\n \n \tif (!IS_ENABLED(CONFIG_ARM64_4K_PAGES) ||\n \t (end - start < PAGES_PER_SECTION * sizeof(struct page)))\n-\t\treturn vmemmap_populate_basepages(start, end, node, altmap, pgmap);\n+\t\treturn vmemmap_populate_basepages(start, end, node, altmap);\n \telse\n-\t\treturn vmemmap_populate_hugepages(start, end, node, altmap, pgmap);\n+\t\treturn vmemmap_populate_hugepages(start, end, node, altmap);\n }\n \n #ifdef CONFIG_MEMORY_HOTPLUG\n@@ -1994,13 +1994,12 @@ int arch_add_memory(int nid, u64 start, u64 size,\n \treturn ret;\n }\n \n-void arch_remove_memory(u64 start, u64 size, struct vmem_altmap *altmap,\n-\t\t\tstruct dev_pagemap *pgmap)\n+void arch_remove_memory(u64 start, u64 size, struct vmem_altmap *altmap)\n {\n \tunsigned long start_pfn = start >> PAGE_SHIFT;\n \tunsigned long nr_pages = size >> PAGE_SHIFT;\n \n-\t__remove_pages(start_pfn, nr_pages, altmap, pgmap);\n+\t__remove_pages(start_pfn, nr_pages, altmap);\n \t__remove_pgd_mapping(swapper_pg_dir, __phys_to_virt(start), size);\n }\n \ndiff --git a/arch/loongarch/mm/init.c b/arch/loongarch/mm/init.c\nindex d61c2e09caae..00f3822b6e47 100644\n--- a/arch/loongarch/mm/init.c\n+++ b/arch/loongarch/mm/init.c\n@@ -86,8 +86,7 @@ int arch_add_memory(int nid, u64 start, u64 size, struct mhp_params *params)\n \treturn ret;\n }\n \n-void arch_remove_memory(u64 start, u64 size, struct vmem_altmap *altmap,\n-\t\t\tstruct dev_pagemap *pgmap)\n+void arch_remove_memory(u64 start, u64 size, struct vmem_altmap *altmap)\n {\n \tunsigned long start_pfn = start >> PAGE_SHIFT;\n \tunsigned long nr_pages = size >> PAGE_SHIFT;\n@@ -96,7 +95,7 @@ void arch_remove_memory(u64 start, u64 size, struct vmem_altmap *altmap,\n \t/* With altmap the first mapped page is offset from @start */\n \tif (altmap)\n \t\tpage += vmem_altmap_offset(altmap);\n-\t__remove_pages(start_pfn, nr_pages, altmap, pgmap);\n+\t__remove_pages(start_pfn, nr_pages, altmap);\n }\n #endif\n \n@@ -123,13 +122,12 @@ int __meminit vmemmap_check_pmd(pmd_t *pmd, int node,\n }\n \n int __meminit vmemmap_populate(unsigned long start, unsigned long end,\n-\t\t\t int node, struct vmem_altmap *altmap,\n-\t\t\t struct dev_pagemap *pgmap)\n+\t\t\t int node, struct vmem_altmap *altmap)\n {\n #if CONFIG_PGTABLE_LEVELS == 2\n-\treturn vmemmap_populate_basepages(start, end, node, NULL, pgmap);\n+\treturn vmemmap_populate_basepages(start, end, node, NULL);\n #else\n-\treturn vmemmap_populate_hugepages(start, end, node, NULL, pgmap);\n+\treturn vmemmap_populate_hugepages(start, end, node, NULL);\n #endif\n }\n \ndiff --git a/arch/powerpc/include/asm/book3s/64/radix.h b/arch/powerpc/include/asm/book3s/64/radix.h\nindex 18e28deba255..0c9195dd50c9 100644\n--- a/arch/powerpc/include/asm/book3s/64/radix.h\n+++ b/arch/powerpc/include/asm/book3s/64/radix.h\n@@ -316,13 +316,11 @@ static inline int radix__has_transparent_pud_hugepage(void)\n #endif\n \n struct vmem_altmap;\n-struct dev_pagemap;\n extern int __meminit radix__vmemmap_create_mapping(unsigned long start,\n \t\t\t\t\t unsigned long page_size,\n \t\t\t\t\t unsigned long phys);\n int __meminit radix__vmemmap_populate(unsigned long start, unsigned long end,\n-\t\t\t\t int node, struct vmem_altmap *altmap,\n-\t\t\t\t struct dev_pagemap *pgmap);\n+\t\t\t\t int node, struct vmem_altmap *altmap);\n void __ref radix__vmemmap_free(unsigned long start, unsigned long end,\n \t\t\t struct vmem_altmap *altmap);\n extern void radix__vmemmap_remove_mapping(unsigned long start,\ndiff --git a/arch/powerpc/mm/book3s64/radix_pgtable.c b/arch/powerpc/mm/book3s64/radix_pgtable.c\nindex 36a69589fae4..190448a17119 100644\n--- a/arch/powerpc/mm/book3s64/radix_pgtable.c\n+++ b/arch/powerpc/mm/book3s64/radix_pgtable.c\n@@ -1101,11 +1101,10 @@ static inline pte_t *vmemmap_pte_alloc(pmd_t *pmdp, int node,\n \n static int __meminit vmemmap_populate_compound_pages(unsigned long start_pfn,\n \t\t\t\t\t\t unsigned long start,\n-\t\t\t\t\t\t unsigned long end, int node,\n-\t\t\t\t\t\t struct dev_pagemap *pgmap);\n+\t\t\t\t\t\t unsigned long end, int node);\n \n int __meminit radix__vmemmap_populate(unsigned long start, unsigned long end, int node,\n-\t\t\t\t struct vmem_altmap *altmap, struct dev_pagemap *pgmap)\n+\t\t\t\t struct vmem_altmap *altmap)\n {\n \tunsigned long addr;\n \tunsigned long next;\n@@ -1117,7 +1116,7 @@ int __meminit radix__vmemmap_populate(unsigned long start, unsigned long end, in\n \tunsigned long pfn = page_to_pfn((struct page *)start);\n \n \tif (section_vmemmap_optimizable(__pfn_to_section(pfn)))\n-\t\treturn vmemmap_populate_compound_pages(pfn, start, end, node, pgmap);\n+\t\treturn vmemmap_populate_compound_pages(pfn, start, end, node);\n \t/*\n \t * If altmap is present, Make sure we align the start vmemmap addr\n \t * to PAGE_SIZE so that we calculate the correct start_pfn in\n@@ -1248,8 +1247,7 @@ static pte_t * __meminit radix__vmemmap_populate_address(unsigned long addr, int\n \n static int __meminit vmemmap_populate_compound_pages(unsigned long start_pfn,\n \t\t\t\t\t\t unsigned long start,\n-\t\t\t\t\t\t unsigned long end, int node,\n-\t\t\t\t\t\t struct dev_pagemap *pgmap)\n+\t\t\t\t\t\t unsigned long end, int node)\n {\n \t/*\n \t * we want to map things as base page size mapping so that\ndiff --git a/arch/powerpc/mm/init_64.c b/arch/powerpc/mm/init_64.c\nindex 56cbea89d304..8e18ed427fdd 100644\n--- a/arch/powerpc/mm/init_64.c\n+++ b/arch/powerpc/mm/init_64.c\n@@ -275,12 +275,12 @@ static int __meminit __vmemmap_populate(unsigned long start, unsigned long end,\n }\n \n int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node,\n-\t\t\t struct vmem_altmap *altmap, struct dev_pagemap *pgmap)\n+\t\t\t struct vmem_altmap *altmap)\n {\n \n #ifdef CONFIG_PPC_BOOK3S_64\n \tif (radix_enabled())\n-\t\treturn radix__vmemmap_populate(start, end, node, altmap, pgmap);\n+\t\treturn radix__vmemmap_populate(start, end, node, altmap);\n #endif\n \n \tsection_set_order(__pfn_to_section(page_to_pfn((struct page *)start)), 0);\ndiff --git a/arch/powerpc/mm/mem.c b/arch/powerpc/mm/mem.c\nindex 4c1afab91996..648d0c5602ec 100644\n--- a/arch/powerpc/mm/mem.c\n+++ b/arch/powerpc/mm/mem.c\n@@ -158,13 +158,12 @@ int __ref arch_add_memory(int nid, u64 start, u64 size,\n \treturn rc;\n }\n \n-void __ref arch_remove_memory(u64 start, u64 size, struct vmem_altmap *altmap,\n-\t\t\t struct dev_pagemap *pgmap)\n+void __ref arch_remove_memory(u64 start, u64 size, struct vmem_altmap *altmap)\n {\n \tunsigned long start_pfn = start >> PAGE_SHIFT;\n \tunsigned long nr_pages = size >> PAGE_SHIFT;\n \n-\t__remove_pages(start_pfn, nr_pages, altmap, pgmap);\n+\t__remove_pages(start_pfn, nr_pages, altmap);\n \tarch_remove_linear_mapping(start, size);\n }\n #endif\ndiff --git a/arch/riscv/mm/init.c b/arch/riscv/mm/init.c\nindex 277c89661dff..5142ca80be6f 100644\n--- a/arch/riscv/mm/init.c\n+++ b/arch/riscv/mm/init.c\n@@ -1443,7 +1443,7 @@ int __meminit vmemmap_check_pmd(pmd_t *pmdp, int node,\n }\n \n int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node,\n-\t\t\t struct vmem_altmap *altmap, struct dev_pagemap *pgmap)\n+\t\t\t struct vmem_altmap *altmap)\n {\n \t/*\n \t * Note that SPARSEMEM_VMEMMAP is only selected for rv64 and that we\n@@ -1451,7 +1451,7 @@ int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node,\n \t * memory hotplug, we are not able to update all the page tables with\n \t * the new PMDs.\n \t */\n-\treturn vmemmap_populate_hugepages(start, end, node, altmap, pgmap);\n+\treturn vmemmap_populate_hugepages(start, end, node, altmap);\n }\n #endif\n \n@@ -1810,10 +1810,9 @@ int __ref arch_add_memory(int nid, u64 start, u64 size, struct mhp_params *param\n \treturn ret;\n }\n \n-void __ref arch_remove_memory(u64 start, u64 size, struct vmem_altmap *altmap,\n-\t\t\t struct dev_pagemap *pgmap)\n+void __ref arch_remove_memory(u64 start, u64 size, struct vmem_altmap *altmap)\n {\n-\t__remove_pages(start >> PAGE_SHIFT, size >> PAGE_SHIFT, altmap, pgmap);\n+\t__remove_pages(start >> PAGE_SHIFT, size >> PAGE_SHIFT, altmap);\n \tremove_linear_mapping(start, size);\n \tflush_tlb_all();\n }\ndiff --git a/arch/s390/mm/init.c b/arch/s390/mm/init.c\nindex 11a689423440..1f72efc2a579 100644\n--- a/arch/s390/mm/init.c\n+++ b/arch/s390/mm/init.c\n@@ -276,13 +276,12 @@ int arch_add_memory(int nid, u64 start, u64 size,\n \treturn rc;\n }\n \n-void arch_remove_memory(u64 start, u64 size, struct vmem_altmap *altmap,\n-\t\t\tstruct dev_pagemap *pgmap)\n+void arch_remove_memory(u64 start, u64 size, struct vmem_altmap *altmap)\n {\n \tunsigned long start_pfn = start >> PAGE_SHIFT;\n \tunsigned long nr_pages = size >> PAGE_SHIFT;\n \n-\t__remove_pages(start_pfn, nr_pages, altmap, pgmap);\n+\t__remove_pages(start_pfn, nr_pages, altmap);\n \tvmem_remove_mapping(start, size);\n }\n #endif /* CONFIG_MEMORY_HOTPLUG */\ndiff --git a/arch/s390/mm/vmem.c b/arch/s390/mm/vmem.c\nindex a7bf8d3d5601..eeadff45e0e1 100644\n--- a/arch/s390/mm/vmem.c\n+++ b/arch/s390/mm/vmem.c\n@@ -506,7 +506,7 @@ static void vmem_remove_range(unsigned long start, unsigned long size)\n * Add a backed mem_map array to the virtual mem_map array.\n */\n int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node,\n-\t\t\t struct vmem_altmap *altmap, struct dev_pagemap *pgmap)\n+\t\t\t struct vmem_altmap *altmap)\n {\n \tint ret;\n \ndiff --git a/arch/sparc/mm/init_64.c b/arch/sparc/mm/init_64.c\nindex f870ca330f9e..367c269305e5 100644\n--- a/arch/sparc/mm/init_64.c\n+++ b/arch/sparc/mm/init_64.c\n@@ -2591,10 +2591,9 @@ int __meminit vmemmap_check_pmd(pmd_t *pmdp, int node,\n }\n \n int __meminit vmemmap_populate(unsigned long vstart, unsigned long vend,\n-\t\t\t int node, struct vmem_altmap *altmap,\n-\t\t\t struct dev_pagemap *pgmap)\n+\t\t\t int node, struct vmem_altmap *altmap)\n {\n-\treturn vmemmap_populate_hugepages(vstart, vend, node, NULL, pgmap);\n+\treturn vmemmap_populate_hugepages(vstart, vend, node, NULL);\n }\n #endif /* CONFIG_SPARSEMEM_VMEMMAP */\n \ndiff --git a/arch/x86/mm/init_64.c b/arch/x86/mm/init_64.c\nindex e18cc81a30b4..df2261fa4f98 100644\n--- a/arch/x86/mm/init_64.c\n+++ b/arch/x86/mm/init_64.c\n@@ -1288,13 +1288,12 @@ kernel_physical_mapping_remove(unsigned long start, unsigned long end)\n \tremove_pagetable(start, end, true, NULL);\n }\n \n-void __ref arch_remove_memory(u64 start, u64 size, struct vmem_altmap *altmap,\n-\t\t\t struct dev_pagemap *pgmap)\n+void __ref arch_remove_memory(u64 start, u64 size, struct vmem_altmap *altmap)\n {\n \tunsigned long start_pfn = start >> PAGE_SHIFT;\n \tunsigned long nr_pages = size >> PAGE_SHIFT;\n \n-\t__remove_pages(start_pfn, nr_pages, altmap, pgmap);\n+\t__remove_pages(start_pfn, nr_pages, altmap);\n \tkernel_physical_mapping_remove(start, start + size);\n }\n #endif /* CONFIG_MEMORY_HOTPLUG */\n@@ -1557,7 +1556,7 @@ int __meminit vmemmap_check_pmd(pmd_t *pmd, int node,\n }\n \n int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node,\n-\t\tstruct vmem_altmap *altmap, struct dev_pagemap *pgmap)\n+\t\tstruct vmem_altmap *altmap)\n {\n \tint err;\n \n@@ -1565,15 +1564,15 @@ int __meminit vmemmap_populate(unsigned long start, unsigned long end, int node,\n \tVM_BUG_ON(!PAGE_ALIGNED(end));\n \n \tif (end - start < PAGES_PER_SECTION * sizeof(struct page))\n-\t\terr = vmemmap_populate_basepages(start, end, node, NULL, pgmap);\n+\t\terr = vmemmap_populate_basepages(start, end, node, NULL);\n \telse if (boot_cpu_has(X86_FEATURE_PSE))\n-\t\terr = vmemmap_populate_hugepages(start, end, node, altmap, pgmap);\n+\t\terr = vmemmap_populate_hugepages(start, end, node, altmap);\n \telse if (altmap) {\n \t\tpr_err_once(\"%s: no cpu support for altmap allocations\\n\",\n \t\t\t\t__func__);\n \t\terr = -ENOMEM;\n \t} else\n-\t\terr = vmemmap_populate_basepages(start, end, node, NULL, pgmap);\n+\t\terr = vmemmap_populate_basepages(start, end, node, NULL);\n \tif (!err)\n \t\tsync_global_pgds(start, end - 1);\n \treturn err;\ndiff --git a/include/linux/memory_hotplug.h b/include/linux/memory_hotplug.h\nindex 7c9d66729c60..815e908c4135 100644\n--- a/include/linux/memory_hotplug.h\n+++ b/include/linux/memory_hotplug.h\n@@ -135,10 +135,9 @@ static inline bool movable_node_is_enabled(void)\n \treturn movable_node_enabled;\n }\n \n-extern void arch_remove_memory(u64 start, u64 size, struct vmem_altmap *altmap,\n-\t\t\t struct dev_pagemap *pgmap);\n+extern void arch_remove_memory(u64 start, u64 size, struct vmem_altmap *altmap);\n extern void __remove_pages(unsigned long start_pfn, unsigned long nr_pages,\n-\t\t\t struct vmem_altmap *altmap, struct dev_pagemap *pgmap);\n+\t\t\t struct vmem_altmap *altmap);\n \n /* reasonably generic interface to expand the physical pages */\n extern int __add_pages(int nid, unsigned long start_pfn, unsigned long nr_pages,\n@@ -308,8 +307,7 @@ extern int sparse_add_section(int nid, unsigned long pfn,\n \t\tunsigned long nr_pages, struct vmem_altmap *altmap,\n \t\tstruct dev_pagemap *pgmap);\n extern void sparse_remove_section(unsigned long pfn, unsigned long nr_pages,\n-\t\t\t\t struct vmem_altmap *altmap,\n-\t\t\t\t struct dev_pagemap *pgmap);\n+\t\t\t\t struct vmem_altmap *altmap);\n extern struct zone *zone_for_pfn_range(enum mmop online_type,\n \t\tint nid, struct memory_group *group, unsigned long start_pfn,\n \t\tunsigned long nr_pages);\ndiff --git a/include/linux/mm.h b/include/linux/mm.h\nindex 8baa224444be..adca19a4b2c7 100644\n--- a/include/linux/mm.h\n+++ b/include/linux/mm.h\n@@ -4858,8 +4858,7 @@ static inline void print_vma_addr(char *prefix, unsigned long rip)\n void *sparse_buffer_alloc(unsigned long size);\n unsigned long section_map_size(void);\n struct page * __populate_section_memmap(unsigned long pfn,\n-\t\tunsigned long nr_pages, int nid, struct vmem_altmap *altmap,\n-\t\tstruct dev_pagemap *pgmap);\n+\t\tunsigned long nr_pages, int nid, struct vmem_altmap *altmap);\n void *vmemmap_alloc_block(unsigned long size, int node);\n struct vmem_altmap;\n void *vmemmap_alloc_block_buf(unsigned long size, int node,\n@@ -4870,13 +4869,11 @@ void vmemmap_set_pmd(pmd_t *pmd, void *p, int node,\n int vmemmap_check_pmd(pmd_t *pmd, int node,\n \t\t unsigned long addr, unsigned long next);\n int vmemmap_populate_basepages(unsigned long start, unsigned long end,\n-\t\t\t int node, struct vmem_altmap *altmap,\n-\t\t\t struct dev_pagemap *pgmap);\n+\t\t\t int node, struct vmem_altmap *altmap);\n int vmemmap_populate_hugepages(unsigned long start, unsigned long end,\n-\t\t\t int node, struct vmem_altmap *altmap,\n-\t\t\t struct dev_pagemap *pgmap);\n+\t\t\t int node, struct vmem_altmap *altmap);\n int vmemmap_populate(unsigned long start, unsigned long end, int node,\n-\t\tstruct vmem_altmap *altmap, struct dev_pagemap *pgmap);\n+\t\tstruct vmem_altmap *altmap);\n void vmemmap_populate_print_last(void);\n struct page *vmemmap_shared_tail_page(unsigned int order, struct zone *zone);\n #ifdef CONFIG_MEMORY_HOTPLUG\ndiff --git a/mm/memory_hotplug.c b/mm/memory_hotplug.c\nindex 28306196c0fe..68dd56dd9f74 100644\n--- a/mm/memory_hotplug.c\n+++ b/mm/memory_hotplug.c\n@@ -584,7 +584,7 @@ void remove_pfn_range_from_zone(struct zone *zone,\n * calling offline_pages().\n */\n void __remove_pages(unsigned long pfn, unsigned long nr_pages,\n-\t\t struct vmem_altmap *altmap, struct dev_pagemap *pgmap)\n+\t\t struct vmem_altmap *altmap)\n {\n \tconst unsigned long end_pfn = pfn + nr_pages;\n \tunsigned long cur_nr_pages;\n@@ -599,7 +599,7 @@ void __remove_pages(unsigned long pfn, unsigned long nr_pages,\n \t\t/* Select all remaining pages up to the next section boundary */\n \t\tcur_nr_pages = min(end_pfn - pfn,\n \t\t\t\t SECTION_ALIGN_UP(pfn + 1) - pfn);\n-\t\tsparse_remove_section(pfn, cur_nr_pages, altmap, pgmap);\n+\t\tsparse_remove_section(pfn, cur_nr_pages, altmap);\n \t}\n }\n \n@@ -1419,7 +1419,7 @@ static void remove_memory_blocks_and_altmaps(u64 start, u64 size)\n \n \t\tremove_memory_block_devices(cur_start, memblock_size);\n \n-\t\tarch_remove_memory(cur_start, memblock_size, altmap, NULL);\n+\t\tarch_remove_memory(cur_start, memblock_size, altmap);\n \n \t\t/* Verify that all vmemmap pages have actually been freed. */\n \t\tWARN(altmap->alloc, \"Altmap not fully unmapped\");\n@@ -1462,7 +1462,7 @@ static int create_altmaps_and_memory_blocks(int nid, struct memory_group *group,\n \t\tret = create_memory_block_devices(cur_start, memblock_size, nid,\n \t\t\t\t\t\t params.altmap, group);\n \t\tif (ret) {\n-\t\t\tarch_remove_memory(cur_start, memblock_size, NULL, NULL);\n+\t\t\tarch_remove_memory(cur_start, memblock_size, NULL);\n \t\t\tkfree(params.altmap);\n \t\t\tgoto out;\n \t\t}\n@@ -1548,7 +1548,7 @@ int add_memory_resource(int nid, struct resource *res, mhp_t mhp_flags)\n \t\t/* create memory block devices after memory was added */\n \t\tret = create_memory_block_devices(start, size, nid, NULL, group);\n \t\tif (ret) {\n-\t\t\tarch_remove_memory(start, size, params.altmap, NULL);\n+\t\t\tarch_remove_memory(start, size, params.altmap);\n \t\t\tgoto error;\n \t\t}\n \t}\n@@ -2247,7 +2247,7 @@ static int try_remove_memory(u64 start, u64 size)\n \t\t * No altmaps present, do the removal directly\n \t\t */\n \t\tremove_memory_block_devices(start, size);\n-\t\tarch_remove_memory(start, size, NULL, NULL);\n+\t\tarch_remove_memory(start, size, NULL);\n \t} else {\n \t\t/* all memblocks in the range have altmaps */\n \t\tremove_memory_blocks_and_altmaps(start, size);\ndiff --git a/mm/memremap.c b/mm/memremap.c\nindex c45b90f334ea..ac7be07e3361 100644\n--- a/mm/memremap.c\n+++ b/mm/memremap.c\n@@ -97,10 +97,10 @@ static void pageunmap_range(struct dev_pagemap *pgmap, int range_id)\n \t\t\t\t PHYS_PFN(range_len(range)));\n \tif (pgmap->type == MEMORY_DEVICE_PRIVATE) {\n \t\t__remove_pages(PHYS_PFN(range->start),\n-\t\t\t PHYS_PFN(range_len(range)), NULL, pgmap);\n+\t\t\t PHYS_PFN(range_len(range)), NULL);\n \t} else {\n \t\tarch_remove_memory(range->start, range_len(range),\n-\t\t\t\tpgmap_altmap(pgmap), pgmap);\n+\t\t\t\tpgmap_altmap(pgmap));\n \t\tkasan_remove_zero_shadow(__va(range->start), range_len(range));\n \t}\n \tmem_hotplug_done();\ndiff --git a/mm/sparse-vmemmap.c b/mm/sparse-vmemmap.c\nindex 752a48112504..68dcc52591d5 100644\n--- a/mm/sparse-vmemmap.c\n+++ b/mm/sparse-vmemmap.c\n@@ -281,8 +281,7 @@ static pte_t * __meminit vmemmap_populate_address(unsigned long addr, int node,\n }\n \n int __meminit vmemmap_populate_basepages(unsigned long start, unsigned long end,\n-\t\t\t\t\t int node, struct vmem_altmap *altmap,\n-\t\t\t\t\t struct dev_pagemap *pgmap)\n+\t\t\t\t\t int node, struct vmem_altmap *altmap)\n {\n \tunsigned long addr = start;\n \tpte_t *pte;\n@@ -342,8 +341,7 @@ int __weak __meminit vmemmap_check_pmd(pmd_t *pmd, int node,\n }\n \n int __meminit vmemmap_populate_hugepages(unsigned long start, unsigned long end,\n-\t\t\t\t\t int node, struct vmem_altmap *altmap,\n-\t\t\t\t\t struct dev_pagemap *pgmap)\n+\t\t\t\t\t int node, struct vmem_altmap *altmap)\n {\n \tunsigned long addr;\n \tunsigned long next;\n@@ -393,15 +391,14 @@ int __meminit vmemmap_populate_hugepages(unsigned long start, unsigned long end,\n \t\t\tVM_BUG_ON(section_vmemmap_optimizable(ms));\n \t\t\tcontinue;\n \t\t}\n-\t\tif (vmemmap_populate_basepages(addr, next, node, altmap, pgmap))\n+\t\tif (vmemmap_populate_basepages(addr, next, node, altmap))\n \t\t\treturn -ENOMEM;\n \t}\n \treturn 0;\n }\n \n struct page * __meminit __populate_section_memmap(unsigned long pfn,\n-\t\tunsigned long nr_pages, int nid, struct vmem_altmap *altmap,\n-\t\tstruct dev_pagemap *pgmap)\n+\t\tunsigned long nr_pages, int nid, struct vmem_altmap *altmap)\n {\n \tunsigned long start = (unsigned long) pfn_to_page(pfn);\n \tunsigned long end = start + nr_pages * sizeof(struct page);\n@@ -410,7 +407,7 @@ struct page * __meminit __populate_section_memmap(unsigned long pfn,\n \t\t!IS_ALIGNED(nr_pages, PAGES_PER_SUBSECTION)))\n \t\treturn NULL;\n \n-\tif (vmemmap_populate(start, end, nid, altmap, pgmap))\n+\tif (vmemmap_populate(start, end, nid, altmap))\n \t\treturn NULL;\n \n \treturn pfn_to_page(pfn);\n@@ -486,10 +483,9 @@ void offline_mem_sections(unsigned long start_pfn, unsigned long end_pfn)\n }\n \n static struct page * __meminit populate_section_memmap(unsigned long pfn,\n-\t\tunsigned long nr_pages, int nid, struct vmem_altmap *altmap,\n-\t\tstruct dev_pagemap *pgmap)\n+\t\tunsigned long nr_pages, int nid, struct vmem_altmap *altmap)\n {\n-\treturn __populate_section_memmap(pfn, nr_pages, nid, altmap, pgmap);\n+\treturn __populate_section_memmap(pfn, nr_pages, nid, altmap);\n }\n \n static void depopulate_section_memmap(unsigned long pfn, unsigned long nr_pages,\n@@ -570,7 +566,7 @@ static int fill_subsection_map(unsigned long pfn, unsigned long nr_pages)\n * usage map, but still need to free the vmemmap range.\n */\n static void section_deactivate(unsigned long pfn, unsigned long nr_pages,\n-\t\tstruct vmem_altmap *altmap, struct dev_pagemap *pgmap)\n+\t\tstruct vmem_altmap *altmap)\n {\n \tstruct mem_section *ms = __pfn_to_section(pfn);\n \tbool section_is_early = early_section(ms);\n@@ -622,8 +618,7 @@ static void section_deactivate(unsigned long pfn, unsigned long nr_pages,\n }\n \n static struct page * __meminit section_activate(int nid, unsigned long pfn,\n-\t\tunsigned long nr_pages, struct vmem_altmap *altmap,\n-\t\tstruct dev_pagemap *pgmap)\n+\t\tunsigned long nr_pages, struct vmem_altmap *altmap)\n {\n \tstruct mem_section *ms = __pfn_to_section(pfn);\n \tstruct mem_section_usage *usage = NULL;\n@@ -655,10 +650,10 @@ static struct page * __meminit section_activate(int nid, unsigned long pfn,\n \tif (nr_pages < PAGES_PER_SECTION && early_section(ms))\n \t\treturn pfn_to_page(pfn);\n \n-\tmemmap = populate_section_memmap(pfn, nr_pages, nid, altmap, pgmap);\n+\tmemmap = populate_section_memmap(pfn, nr_pages, nid, altmap);\n \tmemmap_pages_add(section_vmemmap_pages(pfn, nr_pages));\n \tif (!memmap) {\n-\t\tsection_deactivate(pfn, nr_pages, altmap, pgmap);\n+\t\tsection_deactivate(pfn, nr_pages, altmap);\n \t\treturn ERR_PTR(-ENOMEM);\n \t}\n \n@@ -704,7 +699,7 @@ int __meminit sparse_add_section(int nid, unsigned long start_pfn,\n \t\tsection_set_zone(ms, ZONE_DEVICE);\n #endif\n \t}\n-\tmemmap = section_activate(nid, start_pfn, nr_pages, altmap, pgmap);\n+\tmemmap = section_activate(nid, start_pfn, nr_pages, altmap);\n \tif (IS_ERR(memmap))\n \t\treturn PTR_ERR(memmap);\n \n@@ -726,13 +721,13 @@ int __meminit sparse_add_section(int nid, unsigned long start_pfn,\n }\n \n void sparse_remove_section(unsigned long pfn, unsigned long nr_pages,\n-\t\t\t struct vmem_altmap *altmap, struct dev_pagemap *pgmap)\n+\t\t\t struct vmem_altmap *altmap)\n {\n \tstruct mem_section *ms = __pfn_to_section(pfn);\n \n \tif (WARN_ON_ONCE(!valid_section(ms)))\n \t\treturn;\n \n-\tsection_deactivate(pfn, nr_pages, altmap, pgmap);\n+\tsection_deactivate(pfn, nr_pages, altmap);\n }\n #endif /* CONFIG_MEMORY_HOTPLUG */\ndiff --git a/mm/sparse.c b/mm/sparse.c\nindex 400542302ad4..77bb0113bac5 100644\n--- a/mm/sparse.c\n+++ b/mm/sparse.c\n@@ -237,8 +237,7 @@ unsigned long __init section_map_size(void)\n }\n \n struct page __init *__populate_section_memmap(unsigned long pfn,\n-\t\tunsigned long nr_pages, int nid, struct vmem_altmap *altmap,\n-\t\tstruct dev_pagemap *pgmap)\n+\t\tunsigned long nr_pages, int nid, struct vmem_altmap *altmap)\n {\n \tunsigned long size = section_map_size();\n \tstruct page *map = sparse_buffer_alloc(size);\n@@ -386,8 +385,7 @@ static void __init sparse_init_nid(int nid, unsigned long pnum_begin,\n \t\tif (pnum >= pnum_end)\n \t\t\tbreak;\n \n-\t\tmap = __populate_section_memmap(pfn, PAGES_PER_SECTION,\n-\t\t\t\t\t\tnid, NULL, NULL);\n+\t\tmap = __populate_section_memmap(pfn, PAGES_PER_SECTION, nid, NULL);\n \t\tif (!map)\n \t\t\tpanic(\"Populate section (%ld) on node[%d] failed\\n\", pnum, nid);\n \t\tmemmap_boot_pages_add(section_vmemmap_pages(pfn, PAGES_PER_SECTION));\n", "prefixes": [ "45/49" ] }