Patch Detail
get:
Show a patch.
patch:
Update a patch.
put:
Update a patch.
GET /api/patches/2216729/?format=api
{ "id": 2216729, "url": "http://patchwork.ozlabs.org/api/patches/2216729/?format=api", "web_url": "http://patchwork.ozlabs.org/project/linux-um/patch/20260327061704.3707577-21-hch@lst.de/", "project": { "id": 60, "url": "http://patchwork.ozlabs.org/api/projects/60/?format=api", "name": "User-mode Linux Development", "link_name": "linux-um", "list_id": "linux-um.lists.infradead.org", "list_email": "linux-um@lists.infradead.org", "web_url": "", "scm_url": "", "webscm_url": "", "list_archive_url": "", "list_archive_url_format": "", "commit_url_format": "" }, "msgid": "<20260327061704.3707577-21-hch@lst.de>", "list_archive_url": null, "date": "2026-03-27T06:16:52", "name": "[20/28] xor: avoid indirect calls for arm64-optimized ops", "commit_ref": null, "pull_url": null, "state": "not-applicable", "archived": false, "hash": "f19836a94c181d77c371c7df122d1b40c985c039", "submitter": { "id": 82, "url": "http://patchwork.ozlabs.org/api/people/82/?format=api", "name": "Christoph Hellwig", "email": "hch@lst.de" }, "delegate": null, "mbox": "http://patchwork.ozlabs.org/project/linux-um/patch/20260327061704.3707577-21-hch@lst.de/mbox/", "series": [ { "id": 497692, "url": "http://patchwork.ozlabs.org/api/series/497692/?format=api", "web_url": "http://patchwork.ozlabs.org/project/linux-um/list/?series=497692", "date": "2026-03-27T06:16:33", "name": "[01/28] xor: assert that xor_blocks is not call from interrupt context", "version": 1, "mbox": "http://patchwork.ozlabs.org/series/497692/mbox/" } ], "comments": "http://patchwork.ozlabs.org/api/patches/2216729/comments/", "check": "pending", "checks": "http://patchwork.ozlabs.org/api/patches/2216729/checks/", "tags": {}, "related": [], "headers": { "Return-Path": "\n <linux-um-bounces+incoming=patchwork.ozlabs.org@lists.infradead.org>", "X-Original-To": "incoming@patchwork.ozlabs.org", "Delivered-To": "patchwork-incoming@legolas.ozlabs.org", "Authentication-Results": [ "legolas.ozlabs.org;\n\tdkim=pass (2048-bit key;\n secure) header.d=lists.infradead.org header.i=@lists.infradead.org\n header.a=rsa-sha256 header.s=bombadil.20210309 header.b=NiUcdLh0;\n\tdkim-atps=neutral", "legolas.ozlabs.org;\n spf=none (no SPF record) smtp.mailfrom=lists.infradead.org\n (client-ip=2607:7c80:54:3::133; helo=bombadil.infradead.org;\n envelope-from=linux-um-bounces+incoming=patchwork.ozlabs.org@lists.infradead.org;\n receiver=patchwork.ozlabs.org)" ], "Received": [ "from bombadil.infradead.org (bombadil.infradead.org\n [IPv6:2607:7c80:54:3::133])\n\t(using TLSv1.3 with cipher TLS_AES_256_GCM_SHA384 (256/256 bits)\n\t key-exchange x25519 server-signature ECDSA (secp384r1) server-digest SHA384)\n\t(No client certificate requested)\n\tby legolas.ozlabs.org (Postfix) with ESMTPS id 4fhrCQ0fPKz1xy1\n\tfor <incoming@patchwork.ozlabs.org>; Fri, 27 Mar 2026 17:22:10 +1100 (AEDT)", "from localhost ([::1] helo=bombadil.infradead.org)\n\tby bombadil.infradead.org with esmtp (Exim 4.98.2 #2 (Red Hat Linux))\n\tid 1w60a8-00000006nQe-2G4B;\n\tFri, 27 Mar 2026 06:22:08 +0000", "from\n 2a02-8389-2341-5b80-d601-7564-c2e0-491c.cable.dynamic.v6.surfer.at\n ([2a02:8389:2341:5b80:d601:7564:c2e0:491c] helo=localhost)\n\tby bombadil.infradead.org with esmtpsa (Exim 4.98.2 #2 (Red Hat Linux))\n\tid 1w60a5-00000006nGi-1YFY;\n\tFri, 27 Mar 2026 06:22:05 +0000" ], "DKIM-Signature": "v=1; a=rsa-sha256; q=dns/txt; c=relaxed/relaxed;\n\td=lists.infradead.org; s=bombadil.20210309; h=Sender:List-Subscribe:List-Help\n\t:List-Post:List-Archive:List-Unsubscribe:List-Id:Content-Transfer-Encoding:\n\tMIME-Version:References:In-Reply-To:Message-ID:Date:Subject:Cc:To:From:\n\tReply-To:Content-Type:Content-ID:Content-Description:Resent-Date:Resent-From:\n\tResent-Sender:Resent-To:Resent-Cc:Resent-Message-ID:List-Owner;\n\tbh=vfRQfd6HX2UIy1GEtF4WxXFvmD5740u62PQfhFgAIN8=; b=NiUcdLh0fJAylGax/YA2p0H0NV\n\tG8oFQPjSTciRQbo2uEYWt5Uqbu/dXsEJbqQqS0tMQv2xt7d7xaelcEEL9yW2eHEIi3ZRIadDvhxX2\n\tRXuTofTTMoMncBlc1G9Vm5A2K64W1d8CITS8v10/tbqsu4RK+dVk2YXwjH4sH0Za3sV374RS2t3Wv\n\toqcloRjwU7enwqaPNCZ78LETuVmXRGP3cSon5zih0vSdnKvwRP+Zydl24PWV2kuCVgTE6qdMGMMlR\n\tuGuaDAaii2YWHejXj6XLwMRescFd1lHgARBzWYJ8z5tk3X+7NZ6jKYFkpPnLSgTtnGnul0gE3X/Zw\n\tsSsiN5tw==;", "From": "Christoph Hellwig <hch@lst.de>", "To": "Andrew Morton <akpm@linux-foundation.org>", "Cc": "Richard Henderson <richard.henderson@linaro.org>,\n\tMatt Turner <mattst88@gmail.com>,\n\tMagnus Lindholm <linmag7@gmail.com>,\n\tRussell King <linux@armlinux.org.uk>,\n\tCatalin Marinas <catalin.marinas@arm.com>,\n\tWill Deacon <will@kernel.org>,\n\tArd Biesheuvel <ardb@kernel.org>,\n\tHuacai Chen <chenhuacai@kernel.org>,\n\tWANG Xuerui <kernel@xen0n.name>,\n\tMadhavan Srinivasan <maddy@linux.ibm.com>,\n\tMichael Ellerman <mpe@ellerman.id.au>,\n\tNicholas Piggin <npiggin@gmail.com>,\n\t\"Christophe Leroy (CS GROUP)\" <chleroy@kernel.org>,\n\tPaul Walmsley <pjw@kernel.org>,\n\tPalmer Dabbelt <palmer@dabbelt.com>,\n\tAlbert Ou <aou@eecs.berkeley.edu>,\n\tAlexandre Ghiti <alex@ghiti.fr>,\n\tHeiko Carstens <hca@linux.ibm.com>,\n\tVasily Gorbik <gor@linux.ibm.com>,\n\tAlexander Gordeev <agordeev@linux.ibm.com>,\n\tChristian Borntraeger <borntraeger@linux.ibm.com>,\n\tSven Schnelle <svens@linux.ibm.com>,\n\t\"David S. Miller\" <davem@davemloft.net>,\n\tAndreas Larsson <andreas@gaisler.com>,\n\tRichard Weinberger <richard@nod.at>,\n\tAnton Ivanov <anton.ivanov@cambridgegreys.com>,\n\tJohannes Berg <johannes@sipsolutions.net>,\n\tThomas Gleixner <tglx@kernel.org>,\n\tIngo Molnar <mingo@redhat.com>,\n\tBorislav Petkov <bp@alien8.de>,\n\tDave Hansen <dave.hansen@linux.intel.com>,\n\tx86@kernel.org,\n\t\"H. Peter Anvin\" <hpa@zytor.com>,\n\tHerbert Xu <herbert@gondor.apana.org.au>,\n\tDan Williams <dan.j.williams@intel.com>,\n\tChris Mason <clm@fb.com>,\n\tDavid Sterba <dsterba@suse.com>,\n\tArnd Bergmann <arnd@arndb.de>,\n\tSong Liu <song@kernel.org>,\n\tYu Kuai <yukuai@fnnas.com>,\n\tLi Nan <linan122@huawei.com>,\n\t\"Theodore Ts'o\" <tytso@mit.edu>,\n\t\"Jason A. Donenfeld\" <Jason@zx2c4.com>,\n\tlinux-alpha@vger.kernel.org,\n\tlinux-kernel@vger.kernel.org,\n\tlinux-arm-kernel@lists.infradead.org,\n\tloongarch@lists.linux.dev,\n\tlinuxppc-dev@lists.ozlabs.org,\n\tlinux-riscv@lists.infradead.org,\n\tlinux-s390@vger.kernel.org,\n\tsparclinux@vger.kernel.org,\n\tlinux-um@lists.infradead.org,\n\tlinux-crypto@vger.kernel.org,\n\tlinux-btrfs@vger.kernel.org,\n\tlinux-arch@vger.kernel.org,\n\tlinux-raid@vger.kernel.org", "Subject": "[PATCH 20/28] xor: avoid indirect calls for arm64-optimized ops", "Date": "Fri, 27 Mar 2026 07:16:52 +0100", "Message-ID": "<20260327061704.3707577-21-hch@lst.de>", "X-Mailer": "git-send-email 2.47.3", "In-Reply-To": "<20260327061704.3707577-1-hch@lst.de>", "References": "<20260327061704.3707577-1-hch@lst.de>", "MIME-Version": "1.0", "Content-Transfer-Encoding": "8bit", "X-BeenThere": "linux-um@lists.infradead.org", "X-Mailman-Version": "2.1.34", "Precedence": "list", "List-Id": "<linux-um.lists.infradead.org>", "List-Unsubscribe": "<http://lists.infradead.org/mailman/options/linux-um>,\n <mailto:linux-um-request@lists.infradead.org?subject=unsubscribe>", "List-Archive": "<http://lists.infradead.org/pipermail/linux-um/>", "List-Post": "<mailto:linux-um@lists.infradead.org>", "List-Help": "<mailto:linux-um-request@lists.infradead.org?subject=help>", "List-Subscribe": "<http://lists.infradead.org/mailman/listinfo/linux-um>,\n <mailto:linux-um-request@lists.infradead.org?subject=subscribe>", "Sender": "\"linux-um\" <linux-um-bounces@lists.infradead.org>", "Errors-To": "linux-um-bounces+incoming=patchwork.ozlabs.org@lists.infradead.org" }, "content": "Remove the inner xor_block_templates, and instead have two separate\nactual template that call into the neon-enabled compilation unit.\n\nSigned-off-by: Christoph Hellwig <hch@lst.de>\n---\n arch/arm64/include/asm/xor.h | 13 ++--\n lib/raid/xor/arm64/xor-neon-glue.c | 95 +++++++++++++++---------------\n lib/raid/xor/arm64/xor-neon.c | 73 +++++++++--------------\n lib/raid/xor/arm64/xor-neon.h | 30 ++++++++++\n 4 files changed, 114 insertions(+), 97 deletions(-)\n create mode 100644 lib/raid/xor/arm64/xor-neon.h", "diff": "diff --git a/arch/arm64/include/asm/xor.h b/arch/arm64/include/asm/xor.h\nindex 81718f010761..4782c760bcac 100644\n--- a/arch/arm64/include/asm/xor.h\n+++ b/arch/arm64/include/asm/xor.h\n@@ -7,15 +7,18 @@\n #include <asm-generic/xor.h>\n #include <asm/simd.h>\n \n-extern struct xor_block_template xor_block_arm64;\n-void __init xor_neon_init(void);\n+extern struct xor_block_template xor_block_neon;\n+extern struct xor_block_template xor_block_eor3;\n \n #define arch_xor_init arch_xor_init\n static __always_inline void __init arch_xor_init(void)\n {\n-\txor_neon_init();\n \txor_register(&xor_block_8regs);\n \txor_register(&xor_block_32regs);\n-\tif (cpu_has_neon())\n-\t\txor_register(&xor_block_arm64);\n+\tif (cpu_has_neon()) {\n+\t\tif (cpu_have_named_feature(SHA3))\n+\t\t\txor_register(&xor_block_eor3);\n+\t\telse\n+\t\t\txor_register(&xor_block_neon);\n+\t}\n }\ndiff --git a/lib/raid/xor/arm64/xor-neon-glue.c b/lib/raid/xor/arm64/xor-neon-glue.c\nindex 067a2095659a..08c3e3573388 100644\n--- a/lib/raid/xor/arm64/xor-neon-glue.c\n+++ b/lib/raid/xor/arm64/xor-neon-glue.c\n@@ -7,51 +7,54 @@\n #include <linux/raid/xor_impl.h>\n #include <asm/simd.h>\n #include <asm/xor.h>\n+#include \"xor-neon.h\"\n \n-extern struct xor_block_template const xor_block_inner_neon;\n-\n-static void\n-xor_neon_2(unsigned long bytes, unsigned long * __restrict p1,\n-\t const unsigned long * __restrict p2)\n-{\n-\tscoped_ksimd()\n-\t\txor_block_inner_neon.do_2(bytes, p1, p2);\n-}\n-\n-static void\n-xor_neon_3(unsigned long bytes, unsigned long * __restrict p1,\n-\t const unsigned long * __restrict p2,\n-\t const unsigned long * __restrict p3)\n-{\n-\tscoped_ksimd()\n-\t\txor_block_inner_neon.do_3(bytes, p1, p2, p3);\n-}\n-\n-static void\n-xor_neon_4(unsigned long bytes, unsigned long * __restrict p1,\n-\t const unsigned long * __restrict p2,\n-\t const unsigned long * __restrict p3,\n-\t const unsigned long * __restrict p4)\n-{\n-\tscoped_ksimd()\n-\t\txor_block_inner_neon.do_4(bytes, p1, p2, p3, p4);\n-}\n-\n-static void\n-xor_neon_5(unsigned long bytes, unsigned long * __restrict p1,\n-\t const unsigned long * __restrict p2,\n-\t const unsigned long * __restrict p3,\n-\t const unsigned long * __restrict p4,\n-\t const unsigned long * __restrict p5)\n-{\n-\tscoped_ksimd()\n-\t\txor_block_inner_neon.do_5(bytes, p1, p2, p3, p4, p5);\n-}\n-\n-struct xor_block_template xor_block_arm64 = {\n-\t.name = \"arm64_neon\",\n-\t.do_2 = xor_neon_2,\n-\t.do_3 = xor_neon_3,\n-\t.do_4 = xor_neon_4,\n-\t.do_5\t= xor_neon_5\n+#define XOR_TEMPLATE(_name)\t\t\t\t\t\t\\\n+static void\t\t\t\t\t\t\t\t\\\n+xor_##_name##_2(unsigned long bytes, unsigned long * __restrict p1,\t\\\n+\t const unsigned long * __restrict p2)\t\t\t\t\\\n+{\t\t\t\t\t\t\t\t\t\\\n+\tscoped_ksimd()\t\t\t\t\t\t\t\\\n+\t\t__xor_##_name##_2(bytes, p1, p2);\t\t\t\\\n+}\t\t\t\t\t\t\t\t\t\\\n+\t\t\t\t\t\t\t\t\t\\\n+static void\t\t\t\t\t\t\t\t\\\n+xor_##_name##_3(unsigned long bytes, unsigned long * __restrict p1,\t\\\n+\t const unsigned long * __restrict p2,\t\t\t\t\\\n+\t const unsigned long * __restrict p3)\t\t\t\t\\\n+{\t\t\t\t\t\t\t\t\t\\\n+\tscoped_ksimd()\t\t\t\t\t\t\t\\\n+\t\t__xor_##_name##_3(bytes, p1, p2, p3);\t\t\t\\\n+}\t\t\t\t\t\t\t\t\t\\\n+\t\t\t\t\t\t\t\t\t\\\n+static void\t\t\t\t\t\t\t\t\\\n+xor_##_name##_4(unsigned long bytes, unsigned long * __restrict p1,\t\\\n+\t const unsigned long * __restrict p2,\t\t\t\t\\\n+\t const unsigned long * __restrict p3,\t\t\t\t\\\n+\t const unsigned long * __restrict p4)\t\t\t\t\\\n+{\t\t\t\t\t\t\t\t\t\\\n+\tscoped_ksimd()\t\t\t\t\t\t\t\\\n+\t\t__xor_##_name##_4(bytes, p1, p2, p3, p4);\t\t\\\n+}\t\t\t\t\t\t\t\t\t\\\n+\t\t\t\t\t\t\t\t\t\\\n+static void\t\t\t\t\t\t\t\t\\\n+xor_##_name##_5(unsigned long bytes, unsigned long * __restrict p1,\t\\\n+\t const unsigned long * __restrict p2,\t\t\t\t\\\n+\t const unsigned long * __restrict p3,\t\t\t\t\\\n+\t const unsigned long * __restrict p4,\t\t\t\t\\\n+\t const unsigned long * __restrict p5)\t\t\t\t\\\n+{\t\t\t\t\t\t\t\t\t\\\n+\tscoped_ksimd()\t\t\t\t\t\t\t\\\n+\t\t__xor_##_name##_5(bytes, p1, p2, p3, p4, p5);\t\t\\\n+}\t\t\t\t\t\t\t\t\t\\\n+\t\t\t\t\t\t\t\t\t\\\n+struct xor_block_template xor_block_##_name = {\t\t\t\t\\\n+\t.name = __stringify(_name),\t\t\t\t\t\\\n+\t.do_2 = xor_##_name##_2,\t\t\t\t\t\\\n+\t.do_3 = xor_##_name##_3,\t\t\t\t\t\\\n+\t.do_4 = xor_##_name##_4,\t\t\t\t\t\\\n+\t.do_5\t= xor_##_name##_5\t\t\t\t\t\\\n };\n+\n+XOR_TEMPLATE(neon);\n+XOR_TEMPLATE(eor3);\ndiff --git a/lib/raid/xor/arm64/xor-neon.c b/lib/raid/xor/arm64/xor-neon.c\nindex 8d2d185090db..61194c292917 100644\n--- a/lib/raid/xor/arm64/xor-neon.c\n+++ b/lib/raid/xor/arm64/xor-neon.c\n@@ -8,9 +8,10 @@\n #include <linux/cache.h>\n #include <asm/neon-intrinsics.h>\n #include <asm/xor.h>\n+#include \"xor-neon.h\"\n \n-static void xor_arm64_neon_2(unsigned long bytes, unsigned long * __restrict p1,\n-\tconst unsigned long * __restrict p2)\n+void __xor_neon_2(unsigned long bytes, unsigned long * __restrict p1,\n+\t\tconst unsigned long * __restrict p2)\n {\n \tuint64_t *dp1 = (uint64_t *)p1;\n \tuint64_t *dp2 = (uint64_t *)p2;\n@@ -36,9 +37,9 @@ static void xor_arm64_neon_2(unsigned long bytes, unsigned long * __restrict p1,\n \t} while (--lines > 0);\n }\n \n-static void xor_arm64_neon_3(unsigned long bytes, unsigned long * __restrict p1,\n-\tconst unsigned long * __restrict p2,\n-\tconst unsigned long * __restrict p3)\n+void __xor_neon_3(unsigned long bytes, unsigned long * __restrict p1,\n+\t\tconst unsigned long * __restrict p2,\n+\t\tconst unsigned long * __restrict p3)\n {\n \tuint64_t *dp1 = (uint64_t *)p1;\n \tuint64_t *dp2 = (uint64_t *)p2;\n@@ -72,10 +73,10 @@ static void xor_arm64_neon_3(unsigned long bytes, unsigned long * __restrict p1,\n \t} while (--lines > 0);\n }\n \n-static void xor_arm64_neon_4(unsigned long bytes, unsigned long * __restrict p1,\n-\tconst unsigned long * __restrict p2,\n-\tconst unsigned long * __restrict p3,\n-\tconst unsigned long * __restrict p4)\n+void __xor_neon_4(unsigned long bytes, unsigned long * __restrict p1,\n+\t\tconst unsigned long * __restrict p2,\n+\t\tconst unsigned long * __restrict p3,\n+\t\tconst unsigned long * __restrict p4)\n {\n \tuint64_t *dp1 = (uint64_t *)p1;\n \tuint64_t *dp2 = (uint64_t *)p2;\n@@ -117,11 +118,11 @@ static void xor_arm64_neon_4(unsigned long bytes, unsigned long * __restrict p1,\n \t} while (--lines > 0);\n }\n \n-static void xor_arm64_neon_5(unsigned long bytes, unsigned long * __restrict p1,\n-\tconst unsigned long * __restrict p2,\n-\tconst unsigned long * __restrict p3,\n-\tconst unsigned long * __restrict p4,\n-\tconst unsigned long * __restrict p5)\n+void __xor_neon_5(unsigned long bytes, unsigned long * __restrict p1,\n+\t\tconst unsigned long * __restrict p2,\n+\t\tconst unsigned long * __restrict p3,\n+\t\tconst unsigned long * __restrict p4,\n+\t\tconst unsigned long * __restrict p5)\n {\n \tuint64_t *dp1 = (uint64_t *)p1;\n \tuint64_t *dp2 = (uint64_t *)p2;\n@@ -171,14 +172,6 @@ static void xor_arm64_neon_5(unsigned long bytes, unsigned long * __restrict p1,\n \t} while (--lines > 0);\n }\n \n-struct xor_block_template xor_block_inner_neon __ro_after_init = {\n-\t.name\t= \"__inner_neon__\",\n-\t.do_2\t= xor_arm64_neon_2,\n-\t.do_3\t= xor_arm64_neon_3,\n-\t.do_4\t= xor_arm64_neon_4,\n-\t.do_5\t= xor_arm64_neon_5,\n-};\n-\n static inline uint64x2_t eor3(uint64x2_t p, uint64x2_t q, uint64x2_t r)\n {\n \tuint64x2_t res;\n@@ -189,10 +182,9 @@ static inline uint64x2_t eor3(uint64x2_t p, uint64x2_t q, uint64x2_t r)\n \treturn res;\n }\n \n-static void xor_arm64_eor3_3(unsigned long bytes,\n-\tunsigned long * __restrict p1,\n-\tconst unsigned long * __restrict p2,\n-\tconst unsigned long * __restrict p3)\n+void __xor_eor3_3(unsigned long bytes, unsigned long * __restrict p1,\n+\t\tconst unsigned long * __restrict p2,\n+\t\tconst unsigned long * __restrict p3)\n {\n \tuint64_t *dp1 = (uint64_t *)p1;\n \tuint64_t *dp2 = (uint64_t *)p2;\n@@ -224,11 +216,10 @@ static void xor_arm64_eor3_3(unsigned long bytes,\n \t} while (--lines > 0);\n }\n \n-static void xor_arm64_eor3_4(unsigned long bytes,\n-\tunsigned long * __restrict p1,\n-\tconst unsigned long * __restrict p2,\n-\tconst unsigned long * __restrict p3,\n-\tconst unsigned long * __restrict p4)\n+void __xor_eor3_4(unsigned long bytes, unsigned long * __restrict p1,\n+\t\tconst unsigned long * __restrict p2,\n+\t\tconst unsigned long * __restrict p3,\n+\t\tconst unsigned long * __restrict p4)\n {\n \tuint64_t *dp1 = (uint64_t *)p1;\n \tuint64_t *dp2 = (uint64_t *)p2;\n@@ -268,12 +259,11 @@ static void xor_arm64_eor3_4(unsigned long bytes,\n \t} while (--lines > 0);\n }\n \n-static void xor_arm64_eor3_5(unsigned long bytes,\n-\tunsigned long * __restrict p1,\n-\tconst unsigned long * __restrict p2,\n-\tconst unsigned long * __restrict p3,\n-\tconst unsigned long * __restrict p4,\n-\tconst unsigned long * __restrict p5)\n+void __xor_eor3_5(unsigned long bytes, unsigned long * __restrict p1,\n+\t\tconst unsigned long * __restrict p2,\n+\t\tconst unsigned long * __restrict p3,\n+\t\tconst unsigned long * __restrict p4,\n+\t\tconst unsigned long * __restrict p5)\n {\n \tuint64_t *dp1 = (uint64_t *)p1;\n \tuint64_t *dp2 = (uint64_t *)p2;\n@@ -314,12 +304,3 @@ static void xor_arm64_eor3_5(unsigned long bytes,\n \t\tdp5 += 8;\n \t} while (--lines > 0);\n }\n-\n-void __init xor_neon_init(void)\n-{\n-\tif (cpu_have_named_feature(SHA3)) {\n-\t\txor_block_inner_neon.do_3 = xor_arm64_eor3_3;\n-\t\txor_block_inner_neon.do_4 = xor_arm64_eor3_4;\n-\t\txor_block_inner_neon.do_5 = xor_arm64_eor3_5;\n-\t}\n-}\ndiff --git a/lib/raid/xor/arm64/xor-neon.h b/lib/raid/xor/arm64/xor-neon.h\nnew file mode 100644\nindex 000000000000..cec0ac846fea\n--- /dev/null\n+++ b/lib/raid/xor/arm64/xor-neon.h\n@@ -0,0 +1,30 @@\n+/* SPDX-License-Identifier: GPL-2.0-only */\n+\n+void __xor_neon_2(unsigned long bytes, unsigned long * __restrict p1,\n+\t\tconst unsigned long * __restrict p2);\n+void __xor_neon_3(unsigned long bytes, unsigned long * __restrict p1,\n+\t\tconst unsigned long * __restrict p2,\n+\t\tconst unsigned long * __restrict p3);\n+void __xor_neon_4(unsigned long bytes, unsigned long * __restrict p1,\n+\t\tconst unsigned long * __restrict p2,\n+\t\tconst unsigned long * __restrict p3,\n+\t\tconst unsigned long * __restrict p4);\n+void __xor_neon_5(unsigned long bytes, unsigned long * __restrict p1,\n+\t\tconst unsigned long * __restrict p2,\n+\t\tconst unsigned long * __restrict p3,\n+\t\tconst unsigned long * __restrict p4,\n+\t\tconst unsigned long * __restrict p5);\n+\n+#define __xor_eor3_2\t__xor_neon_2\n+void __xor_eor3_3(unsigned long bytes, unsigned long * __restrict p1,\n+\t\tconst unsigned long * __restrict p2,\n+\t\tconst unsigned long * __restrict p3);\n+void __xor_eor3_4(unsigned long bytes, unsigned long * __restrict p1,\n+\t\tconst unsigned long * __restrict p2,\n+\t\tconst unsigned long * __restrict p3,\n+\t\tconst unsigned long * __restrict p4);\n+void __xor_eor3_5(unsigned long bytes, unsigned long * __restrict p1,\n+\t\tconst unsigned long * __restrict p2,\n+\t\tconst unsigned long * __restrict p3,\n+\t\tconst unsigned long * __restrict p4,\n+\t\tconst unsigned long * __restrict p5);\n", "prefixes": [ "20/28" ] }