get:
Show a patch.

patch:
Update a patch.

put:
Update a patch.

GET /api/patches/2219155/?format=api
HTTP 200 OK
Allow: GET, PUT, PATCH, HEAD, OPTIONS
Content-Type: application/json
Vary: Accept

{
    "id": 2219155,
    "url": "http://patchwork.ozlabs.org/api/patches/2219155/?format=api",
    "web_url": "http://patchwork.ozlabs.org/project/qemu-devel/patch/20260402125234.1371897-9-max.chou@sifive.com/",
    "project": {
        "id": 14,
        "url": "http://patchwork.ozlabs.org/api/projects/14/?format=api",
        "name": "QEMU Development",
        "link_name": "qemu-devel",
        "list_id": "qemu-devel.nongnu.org",
        "list_email": "qemu-devel@nongnu.org",
        "web_url": "",
        "scm_url": "",
        "webscm_url": "",
        "list_archive_url": "",
        "list_archive_url_format": "",
        "commit_url_format": ""
    },
    "msgid": "<20260402125234.1371897-9-max.chou@sifive.com>",
    "list_archive_url": null,
    "date": "2026-04-02T12:52:33",
    "name": "[v6,8/9] target/riscv: rvv: Support Zvfbfa vector bf16 operations",
    "commit_ref": null,
    "pull_url": null,
    "state": "new",
    "archived": false,
    "hash": "f05112551b9ba4436716729cc3a6333e51d30974",
    "submitter": {
        "id": 86650,
        "url": "http://patchwork.ozlabs.org/api/people/86650/?format=api",
        "name": "Max Chou",
        "email": "max.chou@sifive.com"
    },
    "delegate": null,
    "mbox": "http://patchwork.ozlabs.org/project/qemu-devel/patch/20260402125234.1371897-9-max.chou@sifive.com/mbox/",
    "series": [
        {
            "id": 498485,
            "url": "http://patchwork.ozlabs.org/api/series/498485/?format=api",
            "web_url": "http://patchwork.ozlabs.org/project/qemu-devel/list/?series=498485",
            "date": "2026-04-02T12:52:26",
            "name": "Add Zvfbfa extension support",
            "version": 6,
            "mbox": "http://patchwork.ozlabs.org/series/498485/mbox/"
        }
    ],
    "comments": "http://patchwork.ozlabs.org/api/patches/2219155/comments/",
    "check": "pending",
    "checks": "http://patchwork.ozlabs.org/api/patches/2219155/checks/",
    "tags": {},
    "related": [],
    "headers": {
        "Return-Path": "<qemu-devel-bounces+incoming=patchwork.ozlabs.org@nongnu.org>",
        "X-Original-To": "incoming@patchwork.ozlabs.org",
        "Delivered-To": "patchwork-incoming@legolas.ozlabs.org",
        "Authentication-Results": [
            "legolas.ozlabs.org;\n\tdkim=pass (2048-bit key;\n unprotected) header.d=sifive.com header.i=@sifive.com header.a=rsa-sha256\n header.s=google header.b=A8GtBq5E;\n\tdkim-atps=neutral",
            "legolas.ozlabs.org;\n spf=pass (sender SPF authorized) smtp.mailfrom=nongnu.org\n (client-ip=209.51.188.17; helo=lists.gnu.org;\n envelope-from=qemu-devel-bounces+incoming=patchwork.ozlabs.org@nongnu.org;\n receiver=patchwork.ozlabs.org)"
        ],
        "Received": [
            "from lists.gnu.org (lists.gnu.org [209.51.188.17])\n\t(using TLSv1.2 with cipher ECDHE-ECDSA-AES256-GCM-SHA384 (256/256 bits))\n\t(No client certificate requested)\n\tby legolas.ozlabs.org (Postfix) with ESMTPS id 4fmjx51wWyz1yGY\n\tfor <incoming@patchwork.ozlabs.org>; Fri, 03 Apr 2026 00:53:13 +1100 (AEDT)",
            "from localhost ([::1] helo=lists1p.gnu.org)\n\tby lists.gnu.org with esmtp (Exim 4.90_1)\n\t(envelope-from <qemu-devel-bounces@nongnu.org>)\n\tid 1w8ILI-0008OL-KN; Thu, 02 Apr 2026 09:44:16 -0400",
            "from eggs.gnu.org ([2001:470:142:3::10])\n by lists.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_256_GCM_SHA384:256)\n (Exim 4.90_1) (envelope-from <max.chou@sifive.com>)\n id 1w8IKw-0005fW-8K\n for qemu-devel@nongnu.org; Thu, 02 Apr 2026 09:43:55 -0400",
            "from mail-pj1-x1034.google.com ([2607:f8b0:4864:20::1034])\n by eggs.gnu.org with esmtps (TLS1.2:ECDHE_RSA_AES_128_GCM_SHA256:128)\n (Exim 4.90_1) (envelope-from <max.chou@sifive.com>)\n id 1w8HXl-0008QW-JO\n for qemu-devel@nongnu.org; Thu, 02 Apr 2026 08:53:11 -0400",
            "by mail-pj1-x1034.google.com with SMTP id\n 98e67ed59e1d1-35c238f1063so512773a91.1\n for <qemu-devel@nongnu.org>; Thu, 02 Apr 2026 05:53:00 -0700 (PDT)",
            "from duncan.localdomain (114-35-142-126.hinet-ip.hinet.net.\n [114.35.142.126]) by smtp.gmail.com with ESMTPSA id\n 98e67ed59e1d1-35dd35364edsm2730032a91.0.2026.04.02.05.52.57\n (version=TLS1_3 cipher=TLS_AES_256_GCM_SHA384 bits=256/256);\n Thu, 02 Apr 2026 05:52:58 -0700 (PDT)"
        ],
        "DKIM-Signature": "v=1; a=rsa-sha256; c=relaxed/relaxed;\n d=sifive.com; s=google; t=1775134380; x=1775739180; darn=nongnu.org;\n h=content-transfer-encoding:mime-version:references:in-reply-to\n :message-id:date:subject:cc:to:from:from:to:cc:subject:date\n :message-id:reply-to;\n bh=I76XUCOqBZ0ZbD74c2BnAlaGpVQt1/0Xs9bS5ZC1iUE=;\n b=A8GtBq5ExAUqpcxe2t/DUWiDzgCVydJ+g4YNXV2CyIsqc2A4Bbg9uKryiLWPuufI/+\n lVflrfYyO7JmwhoKV0gG8l9wk5kahTIiK0lt/62lx52juldljec3lhsPY4vM7ZHHcCxr\n 6pdCLP/QG8a7DOc3FKINccc2RpEEHoUU7gTWp8HqDDNrf/ubMmu6yWyD2eYqyoXnLqZg\n KXYy9sv5LU/VuMxLcbZfqaDrcxsuBzhfCnXHde/8lPMh307m3N6DbiZ2W9KXzgxblLEV\n 9qNkNvh3INIjY2WPATQFpeKqn+PZVSiYiSlNu/lLOfkAmraua3nbB0c5ApUp4awPIIy9\n CNsg==",
        "X-Google-DKIM-Signature": "v=1; a=rsa-sha256; c=relaxed/relaxed;\n d=1e100.net; s=20251104; t=1775134380; x=1775739180;\n h=content-transfer-encoding:mime-version:references:in-reply-to\n :message-id:date:subject:cc:to:from:x-gm-gg:x-gm-message-state:from\n :to:cc:subject:date:message-id:reply-to;\n bh=I76XUCOqBZ0ZbD74c2BnAlaGpVQt1/0Xs9bS5ZC1iUE=;\n b=nPvSMejjhkq/l6YHyWJOxQgoC2VGU2GdK4Hllyfg4/tGYiNJP5xs7GL2+nbBtfXYtQ\n tLSC2Ozk4VE6huDWWJcgOYW+0BiG1PSIY+gecymguM45Nx45J3oW/8O9bYHYgWANbVhM\n mtEwRpTqmH3Bok+bWHK6XQ8wuP3wO+gQElsvCRTq0aOdxr5nb9v0Q2qyxQTDr08XF1N+\n py8ybgnMNss22n2y7qKONx8n7zBZmowuIFd7ByXHc9DH6WqJHzq02EaZRgaJxLwo/Rwy\n OdQ7JVVmyhr3wpywLVM5zgJc8cQUJzWIiTJBS/Qa5GEv3xZ9qRQjSZbma2btuF4KErtX\n IlHQ==",
        "X-Gm-Message-State": "AOJu0Yy2Xza63DHFD47jgjHSdNUkKB+ct3NuOT1FZQ7xF2CccZF8u/1B\n s5GLc3W4+IEbpGreDp8u5pPKk3oSRTUJXqrsV3uFjVq2a8aCZeVrdJKUJsYq8LeDbuhpPlIgiAJ\n Go+dCiQHb2PX3h/HMhnSR56/EFBUX34X4YsGoU18fHGJ5Sk8o5qxhHlaTfK5SpoK6PD8bR96tNq\n DuaAInZxAw03EeaH+UqmCajTHeQNdiuO4We10d9vA=",
        "X-Gm-Gg": "AeBDieu7OxoqvD9sOSxnBn+cpQTwiBNZZDUlfV/YmlNDsaN3EJ6pAeDoPjsXg3u5Xqw\n 8TEKbrxc1wXbbZwmiOxcg+uqIAGpXeIqN5RCClYb6F33yBddBrpD6LqufU3WTdhC2j6XNemTGme\n gEC6t+tkwQFMu97+Fm83lPMNPoeTGx3iVi96qe+wvlvQwcbJz0H04Wi8DKq8XAoFIGt0ysv56G0\n HrqC6U9RF7WchZtxahuAbQZ5KbGygOszGa/Y2uyOPknidS8BzpO5iR7vRV5Dg6UUJ4OhikaL3h7\n pBB+ZZkjF4X1LTUFgr/KTI7g4yQdOoPeHRsuOv6PXxKU2N7jDVAD3ANaZb/V/2zm7W62SnV28hE\n vDpICZX0xEM0ynpZNJASJBKyJp15Y676TXerah0Ou2wnUca5dbRP3NneTul+dAe+7wyMomVenPu\n n/nPgRp0u4MRwKt5nCPS/9J1kiq+ZIYXYJ26iRbt11Q/nGg1NupcGxVADQv+gx4XXjah/llZsgR\n gwH9IeAL/CyyT4qMNE7KFNIiT4=",
        "X-Received": "by 2002:a17:90b:1c11:b0:35b:c900:79a6 with SMTP id\n 98e67ed59e1d1-35dc6e77dc3mr7061864a91.4.1775134379353;\n Thu, 02 Apr 2026 05:52:59 -0700 (PDT)",
        "From": "Max Chou <max.chou@sifive.com>",
        "To": "qemu-devel@nongnu.org,\n\tqemu-riscv@nongnu.org",
        "Cc": "Palmer Dabbelt <palmer@dabbelt.com>,\n Alistair Francis <alistair.francis@wdc.com>,\n Weiwei Li <liwei1518@gmail.com>,\n Daniel Henrique Barboza <daniel.barboza@oss.qualcomm.com>,\n Liu Zhiwei <zhiwei_liu@linux.alibaba.com>,\n Chao Liu <chao.liu.zevorn@gmail.com>, Max Chou <max.chou@sifive.com>",
        "Subject": "[PATCH v6 8/9] target/riscv: rvv: Support Zvfbfa vector bf16\n operations",
        "Date": "Thu,  2 Apr 2026 20:52:33 +0800",
        "Message-ID": "<20260402125234.1371897-9-max.chou@sifive.com>",
        "X-Mailer": "git-send-email 2.43.0",
        "In-Reply-To": "<20260402125234.1371897-1-max.chou@sifive.com>",
        "References": "<20260402125234.1371897-1-max.chou@sifive.com>",
        "MIME-Version": "1.0",
        "Content-Transfer-Encoding": "8bit",
        "Received-SPF": "pass client-ip=2607:f8b0:4864:20::1034;\n envelope-from=max.chou@sifive.com; helo=mail-pj1-x1034.google.com",
        "X-Spam_score_int": "-20",
        "X-Spam_score": "-2.1",
        "X-Spam_bar": "--",
        "X-Spam_report": "(-2.1 / 5.0 requ) BAYES_00=-1.9, DKIM_SIGNED=0.1,\n DKIM_VALID=-0.1, DKIM_VALID_AU=-0.1, DKIM_VALID_EF=-0.1,\n RCVD_IN_DNSWL_NONE=-0.0001, SPF_HELO_NONE=0.001,\n SPF_PASS=-0.001 autolearn=unavailable autolearn_force=no",
        "X-Spam_action": "no action",
        "X-BeenThere": "qemu-devel@nongnu.org",
        "X-Mailman-Version": "2.1.29",
        "Precedence": "list",
        "List-Id": "qemu development <qemu-devel.nongnu.org>",
        "List-Unsubscribe": "<https://lists.nongnu.org/mailman/options/qemu-devel>,\n <mailto:qemu-devel-request@nongnu.org?subject=unsubscribe>",
        "List-Archive": "<https://lists.nongnu.org/archive/html/qemu-devel>",
        "List-Post": "<mailto:qemu-devel@nongnu.org>",
        "List-Help": "<mailto:qemu-devel-request@nongnu.org?subject=help>",
        "List-Subscribe": "<https://lists.nongnu.org/mailman/listinfo/qemu-devel>,\n <mailto:qemu-devel-request@nongnu.org?subject=subscribe>",
        "Errors-To": "qemu-devel-bounces+incoming=patchwork.ozlabs.org@nongnu.org",
        "Sender": "qemu-devel-bounces+incoming=patchwork.ozlabs.org@nongnu.org"
    },
    "content": "According to the Zvfbfa ISA spec v0.1, the following vector floating\npoint instructions have different behaviors depend on the ALTFMT and\nVSEW fields of VTYPE CSR.\n\nWhen altfmt=1 and SEW=8, all vector floating-point instructions become\nreserved, except for the following, which are redefined to use the\nBF16 format for any operand that would otherwise have used the FP16\nformat:\n- vfwcvt.f.x[u].v, vfncvt.x[u].f.w, vfncvt.rtz.x[u].f.w\n\nWhen altfmt=1 and SEW=16, all vector floating-point instructions become\nreserved, except for the following, which are redefined to use the\nBF16 format for any operand that would otherwise have used the FP16\nformat:\n- vfadd.v[vf], vfsub.v[vf], vfmin.v[vf], vfmax.v[vf], vmfeq.v[vf],\n  vmfle.v[vf], vmflt.v[vf], vmfne.v[vf], vmfgt.vf, vmfge.vf,\n  vfmul.v[vf], vfrsub.vf, vfmadd.v[vf], vfnmadd.v[vf], vfmsub.v[vf],\n  vfnmsub.v[vf], vfmacc.v[vf], vfnmacc.v[vf], vfmsac.v[vf],\n  vfnmsac.v[vf], vfwadd.v[vf], vfwsub.v[vf], vfwadd.w[vf],\n  vfwsub.w[vf], vfwmul.v[vf], vfwmacc.v[vf], vfwnmacc.v[vf],\n  vfwmsac.v[vf], vfwnmsac.v[vf], vfwcvt.f.f.v, vfncvt.f.f.w,\n  vfncvt.rod.f.f.w, vfrsqrt7.v, vfrec7.v, vfclass.v\n\nThe following instructions marked with * have the same semantics\nregardless of altfmt.\n*- vfmv.f.s,\n   vfwmaccbf16.v[vf] (only if Zvfbfwma is implemented)\n   vfwcvtbf16.f.f.v (only if Zvfbfmin is implemented)\n   vfncvtbf16.f.f.w (only if Zvfbfmin is implemented)\n\nThe following instructions marked with ** differ only in that\nimproperly NaN-boxed f-register operands must substitute the BF16\ncanonical NaN instead of the FP16 canonical NaN.\n**- vfsgnj.v[vf], vfsgnjn.v[vf], vfsgnjx.v[vf], vfslide1up.vf,\n    vfslide1down.vf, vfmv.v.f, vfmerge.vfm, vfmv.s.f\n\nReviewed-by: Daniel Henrique Barboza <daniel.barboza@oss.qualcomm.com>\nReviewed-by: Chao Liu <chao.liu.zevorn@gmail.com>\nSigned-off-by: Max Chou <max.chou@sifive.com>\n---\n target/riscv/helper.h                   |  60 ++\n target/riscv/insn_trans/trans_rvv.c.inc | 970 +++++++++++++++---------\n target/riscv/internals.h                |   1 +\n target/riscv/vector_helper.c            | 329 ++++++++\n 4 files changed, 989 insertions(+), 371 deletions(-)",
    "diff": "diff --git a/target/riscv/helper.h b/target/riscv/helper.h\nindex 7722c590bd..54d2331966 100644\n--- a/target/riscv/helper.h\n+++ b/target/riscv/helper.h\n@@ -768,45 +768,60 @@ DEF_HELPER_6(vnclip_wx_b, void, ptr, ptr, tl, ptr, env, i32)\n DEF_HELPER_6(vnclip_wx_h, void, ptr, ptr, tl, ptr, env, i32)\n DEF_HELPER_6(vnclip_wx_w, void, ptr, ptr, tl, ptr, env, i32)\n \n+DEF_HELPER_6(vfadd_vv_h_bf16, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfadd_vv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfadd_vv_w, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfadd_vv_d, void, ptr, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_6(vfsub_vv_h_bf16, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfsub_vv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfsub_vv_w, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfsub_vv_d, void, ptr, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_6(vfadd_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfadd_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfadd_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfadd_vf_d, void, ptr, ptr, i64, ptr, env, i32)\n+DEF_HELPER_6(vfsub_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfsub_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfsub_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfsub_vf_d, void, ptr, ptr, i64, ptr, env, i32)\n+DEF_HELPER_6(vfrsub_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfrsub_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfrsub_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfrsub_vf_d, void, ptr, ptr, i64, ptr, env, i32)\n \n+DEF_HELPER_6(vfwadd_vv_h_bf16, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfwadd_vv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfwadd_vv_w, void, ptr, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_6(vfwsub_vv_h_bf16, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfwsub_vv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfwsub_vv_w, void, ptr, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_6(vfwadd_wv_h_bf16, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfwadd_wv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfwadd_wv_w, void, ptr, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_6(vfwsub_wv_h_bf16, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfwsub_wv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfwsub_wv_w, void, ptr, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_6(vfwadd_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfwadd_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfwadd_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n+DEF_HELPER_6(vfwsub_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfwsub_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfwsub_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n+DEF_HELPER_6(vfwadd_wf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfwadd_wf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfwadd_wf_w, void, ptr, ptr, i64, ptr, env, i32)\n+DEF_HELPER_6(vfwsub_wf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfwsub_wf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfwsub_wf_w, void, ptr, ptr, i64, ptr, env, i32)\n \n+DEF_HELPER_6(vfmul_vv_h_bf16, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfmul_vv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfmul_vv_w, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfmul_vv_d, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfdiv_vv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfdiv_vv_w, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfdiv_vv_d, void, ptr, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_6(vfmul_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfmul_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfmul_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfmul_vf_d, void, ptr, ptr, i64, ptr, env, i32)\n@@ -817,74 +832,98 @@ DEF_HELPER_6(vfrdiv_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfrdiv_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfrdiv_vf_d, void, ptr, ptr, i64, ptr, env, i32)\n \n+DEF_HELPER_6(vfwmul_vv_h_bf16, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfwmul_vv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfwmul_vv_w, void, ptr, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_6(vfwmul_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfwmul_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfwmul_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n \n+DEF_HELPER_6(vfmacc_vv_h_bf16, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfmacc_vv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfmacc_vv_w, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfmacc_vv_d, void, ptr, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_6(vfnmacc_vv_h_bf16, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfnmacc_vv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfnmacc_vv_w, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfnmacc_vv_d, void, ptr, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_6(vfmsac_vv_h_bf16, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfmsac_vv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfmsac_vv_w, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfmsac_vv_d, void, ptr, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_6(vfnmsac_vv_h_bf16, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfnmsac_vv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfnmsac_vv_w, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfnmsac_vv_d, void, ptr, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_6(vfmadd_vv_h_bf16, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfmadd_vv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfmadd_vv_w, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfmadd_vv_d, void, ptr, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_6(vfnmadd_vv_h_bf16, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfnmadd_vv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfnmadd_vv_w, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfnmadd_vv_d, void, ptr, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_6(vfmsub_vv_h_bf16, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfmsub_vv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfmsub_vv_w, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfmsub_vv_d, void, ptr, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_6(vfnmsub_vv_h_bf16, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfnmsub_vv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfnmsub_vv_w, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfnmsub_vv_d, void, ptr, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_6(vfmacc_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfmacc_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfmacc_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfmacc_vf_d, void, ptr, ptr, i64, ptr, env, i32)\n+DEF_HELPER_6(vfnmacc_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfnmacc_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfnmacc_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfnmacc_vf_d, void, ptr, ptr, i64, ptr, env, i32)\n+DEF_HELPER_6(vfmsac_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfmsac_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfmsac_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfmsac_vf_d, void, ptr, ptr, i64, ptr, env, i32)\n+DEF_HELPER_6(vfnmsac_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfnmsac_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfnmsac_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfnmsac_vf_d, void, ptr, ptr, i64, ptr, env, i32)\n+DEF_HELPER_6(vfmadd_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfmadd_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfmadd_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfmadd_vf_d, void, ptr, ptr, i64, ptr, env, i32)\n+DEF_HELPER_6(vfnmadd_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfnmadd_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfnmadd_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfnmadd_vf_d, void, ptr, ptr, i64, ptr, env, i32)\n+DEF_HELPER_6(vfmsub_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfmsub_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfmsub_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfmsub_vf_d, void, ptr, ptr, i64, ptr, env, i32)\n+DEF_HELPER_6(vfnmsub_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfnmsub_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfnmsub_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfnmsub_vf_d, void, ptr, ptr, i64, ptr, env, i32)\n \n DEF_HELPER_6(vfwmacc_vv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfwmacc_vv_w, void, ptr, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_6(vfwnmacc_vv_h_bf16, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfwnmacc_vv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfwnmacc_vv_w, void, ptr, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_6(vfwmsac_vv_h_bf16, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfwmsac_vv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfwmsac_vv_w, void, ptr, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_6(vfwnmsac_vv_h_bf16, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfwnmsac_vv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfwnmsac_vv_w, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfwmacc_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfwmacc_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n+DEF_HELPER_6(vfwnmacc_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfwnmacc_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfwnmacc_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n+DEF_HELPER_6(vfwmsac_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfwmsac_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfwmsac_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n+DEF_HELPER_6(vfwnmsac_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfwnmsac_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfwnmsac_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n \n@@ -892,23 +931,29 @@ DEF_HELPER_5(vfsqrt_v_h, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfsqrt_v_w, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfsqrt_v_d, void, ptr, ptr, ptr, env, i32)\n \n+DEF_HELPER_5(vfrsqrt7_v_h_bf16, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfrsqrt7_v_h, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfrsqrt7_v_w, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfrsqrt7_v_d, void, ptr, ptr, ptr, env, i32)\n \n+DEF_HELPER_5(vfrec7_v_h_bf16, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfrec7_v_h, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfrec7_v_w, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfrec7_v_d, void, ptr, ptr, ptr, env, i32)\n \n+DEF_HELPER_6(vfmin_vv_h_bf16, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfmin_vv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfmin_vv_w, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfmin_vv_d, void, ptr, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_6(vfmax_vv_h_bf16, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfmax_vv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfmax_vv_w, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vfmax_vv_d, void, ptr, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_6(vfmin_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfmin_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfmin_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfmin_vf_d, void, ptr, ptr, i64, ptr, env, i32)\n+DEF_HELPER_6(vfmax_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfmax_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfmax_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfmax_vf_d, void, ptr, ptr, i64, ptr, env, i32)\n@@ -932,37 +977,48 @@ DEF_HELPER_6(vfsgnjx_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfsgnjx_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vfsgnjx_vf_d, void, ptr, ptr, i64, ptr, env, i32)\n \n+DEF_HELPER_6(vmfeq_vv_h_bf16, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vmfeq_vv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vmfeq_vv_w, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vmfeq_vv_d, void, ptr, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_6(vmfne_vv_h_bf16, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vmfne_vv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vmfne_vv_w, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vmfne_vv_d, void, ptr, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_6(vmflt_vv_h_bf16, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vmflt_vv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vmflt_vv_w, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vmflt_vv_d, void, ptr, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_6(vmfle_vv_h_bf16, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vmfle_vv_h, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vmfle_vv_w, void, ptr, ptr, ptr, ptr, env, i32)\n DEF_HELPER_6(vmfle_vv_d, void, ptr, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_6(vmfeq_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vmfeq_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vmfeq_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vmfeq_vf_d, void, ptr, ptr, i64, ptr, env, i32)\n+DEF_HELPER_6(vmfne_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vmfne_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vmfne_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vmfne_vf_d, void, ptr, ptr, i64, ptr, env, i32)\n+DEF_HELPER_6(vmflt_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vmflt_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vmflt_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vmflt_vf_d, void, ptr, ptr, i64, ptr, env, i32)\n+DEF_HELPER_6(vmfle_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vmfle_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vmfle_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vmfle_vf_d, void, ptr, ptr, i64, ptr, env, i32)\n+DEF_HELPER_6(vmfgt_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vmfgt_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vmfgt_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vmfgt_vf_d, void, ptr, ptr, i64, ptr, env, i32)\n+DEF_HELPER_6(vmfge_vf_h_bf16, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vmfge_vf_h, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vmfge_vf_w, void, ptr, ptr, i64, ptr, env, i32)\n DEF_HELPER_6(vmfge_vf_d, void, ptr, ptr, i64, ptr, env, i32)\n \n+DEF_HELPER_5(vfclass_v_h_bf16, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfclass_v_h, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfclass_v_w, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfclass_v_d, void, ptr, ptr, ptr, env, i32)\n@@ -989,18 +1045,22 @@ DEF_HELPER_5(vfwcvt_xu_f_v_w, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfwcvt_x_f_v_h, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfwcvt_x_f_v_w, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfwcvt_f_xu_v_b, void, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_5(vfwcvt_f_xu_v_b_bf16, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfwcvt_f_xu_v_h, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfwcvt_f_xu_v_w, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfwcvt_f_x_v_b, void, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_5(vfwcvt_f_x_v_b_bf16, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfwcvt_f_x_v_h, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfwcvt_f_x_v_w, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfwcvt_f_f_v_h, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfwcvt_f_f_v_w, void, ptr, ptr, ptr, env, i32)\n \n DEF_HELPER_5(vfncvt_xu_f_w_b, void, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_5(vfncvt_xu_f_w_b_bf16, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfncvt_xu_f_w_h, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfncvt_xu_f_w_w, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfncvt_x_f_w_b, void, ptr, ptr, ptr, env, i32)\n+DEF_HELPER_5(vfncvt_x_f_w_b_bf16, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfncvt_x_f_w_h, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfncvt_x_f_w_w, void, ptr, ptr, ptr, env, i32)\n DEF_HELPER_5(vfncvt_f_xu_w_h, void, ptr, ptr, ptr, env, i32)\ndiff --git a/target/riscv/insn_trans/trans_rvv.c.inc b/target/riscv/insn_trans/trans_rvv.c.inc\nindex 03ae85796a..5b72926b3c 100644\n--- a/target/riscv/insn_trans/trans_rvv.c.inc\n+++ b/target/riscv/insn_trans/trans_rvv.c.inc\n@@ -40,6 +40,9 @@ static bool require_rvf(DisasContext *s)\n \n     switch (s->sew) {\n     case MO_16:\n+        if (s->altfmt) {\n+            return s->cfg_ptr->ext_zvfbfa;\n+        }\n         return s->cfg_ptr->ext_zvfh;\n     case MO_32:\n         return s->cfg_ptr->ext_zve32f;\n@@ -58,6 +61,9 @@ static bool require_rvfmin(DisasContext *s)\n \n     switch (s->sew) {\n     case MO_16:\n+        if (s->altfmt) {\n+            return s->cfg_ptr->ext_zvfbfa;\n+        }\n         return s->cfg_ptr->ext_zvfhmin;\n     case MO_32:\n         return s->cfg_ptr->ext_zve32f;\n@@ -74,6 +80,9 @@ static bool require_scale_rvf(DisasContext *s)\n \n     switch (s->sew) {\n     case MO_8:\n+        if (s->altfmt) {\n+            return s->cfg_ptr->ext_zvfbfa;\n+        }\n         return s->cfg_ptr->ext_zvfh;\n     case MO_16:\n         return s->cfg_ptr->ext_zve32f;\n@@ -2334,25 +2343,39 @@ static void do_nanbox(DisasContext *s, TCGv_i64 out, TCGv_i64 in)\n     }\n }\n \n+/*\n+ * Check altfmt & sew combinations when Zvfbfa extension is enabled.\n+ */\n+static bool vext_check_altfmt(DisasContext *s, int8_t valid_vsew)\n+{\n+    if (s->cfg_ptr->ext_zvfbfa) {\n+        if (s->altfmt && (valid_vsew == -1 || s->sew != valid_vsew)) {\n+            return false;\n+        }\n+    }\n+    return true;\n+}\n+\n /* Vector Single-Width Floating-Point Add/Subtract Instructions */\n \n /*\n  * If the current SEW does not correspond to a supported IEEE floating-point\n  * type, an illegal instruction exception is raised.\n  */\n-static bool opfvv_check(DisasContext *s, arg_rmrr *a)\n+static bool opfvv_check(DisasContext *s, arg_rmrr *a, int8_t valid_bfa_vsew)\n {\n     return require_rvv(s) &&\n            require_rvf(s) &&\n            vext_check_isa_ill(s) &&\n-           vext_check_sss(s, a->rd, a->rs1, a->rs2, a->vm);\n+           vext_check_sss(s, a->rd, a->rs1, a->rs2, a->vm) &&\n+           vext_check_altfmt(s, valid_bfa_vsew);\n }\n \n /* OPFVV without GVEC IR */\n #define GEN_OPFVV_TRANS(NAME, CHECK)                               \\\n static bool trans_##NAME(DisasContext *s, arg_rmrr *a)             \\\n {                                                                  \\\n-    if (CHECK(s, a)) {                                             \\\n+    if (CHECK(s, a, -1)) {                                         \\\n         uint32_t data = 0;                                         \\\n         static gen_helper_gvec_4_ptr * const fns[3] = {            \\\n             gen_helper_##NAME##_h,                                 \\\n@@ -2378,8 +2401,41 @@ static bool trans_##NAME(DisasContext *s, arg_rmrr *a)             \\\n     }                                                              \\\n     return false;                                                  \\\n }\n-GEN_OPFVV_TRANS(vfadd_vv, opfvv_check)\n-GEN_OPFVV_TRANS(vfsub_vv, opfvv_check)\n+\n+#define GEN_OPFVV_BFA_TRANS(NAME, CHECK, BFA_HELPER)                    \\\n+static bool trans_##NAME(DisasContext *s, arg_rmrr *a)                  \\\n+{                                                                       \\\n+    if (CHECK(s, a, MO_16)) {                                           \\\n+        uint32_t data = 0;                                              \\\n+        static gen_helper_gvec_4_ptr * const fns[3] = {                 \\\n+            gen_helper_##NAME##_h,                                      \\\n+            gen_helper_##NAME##_w,                                      \\\n+            gen_helper_##NAME##_d                                       \\\n+        };                                                              \\\n+        gen_set_rm(s, RISCV_FRM_DYN);                                   \\\n+                                                                        \\\n+        data = FIELD_DP32(data, VDATA, VM, a->vm);                      \\\n+        data = FIELD_DP32(data, VDATA, LMUL, s->lmul);                  \\\n+        data = FIELD_DP32(data, VDATA, VTA, s->vta);                    \\\n+        data = FIELD_DP32(data, VDATA, VTA_ALL_1S, s->cfg_vta_all_1s);  \\\n+        data = FIELD_DP32(data, VDATA, VMA, s->vma);                    \\\n+        tcg_gen_gvec_4_ptr(vreg_ofs(s, a->rd), vreg_ofs(s, 0),          \\\n+                           vreg_ofs(s, a->rs1),                         \\\n+                           vreg_ofs(s, a->rs2), tcg_env,                \\\n+                           s->cfg_ptr->vlenb,                           \\\n+                           s->cfg_ptr->vlenb, data,                     \\\n+                           (s->altfmt ? gen_helper_##BFA_HELPER :       \\\n+                                        fns[s->sew - 1]));              \\\n+        tcg_gen_movi_tl(cpu_vstart, 0);                                 \\\n+        finalize_rvv_inst(s);                                           \\\n+                                                                        \\\n+        return true;                                                    \\\n+    }                                                                   \\\n+    return false;                                                       \\\n+}\n+\n+GEN_OPFVV_BFA_TRANS(vfadd_vv, opfvv_check, vfadd_vv_h_bf16)\n+GEN_OPFVV_BFA_TRANS(vfsub_vv, opfvv_check, vfsub_vv_h_bf16)\n \n typedef void gen_helper_opfvf(TCGv_ptr, TCGv_ptr, TCGv_i64, TCGv_ptr,\n                               TCGv_env, TCGv_i32);\n@@ -2415,244 +2471,316 @@ static bool opfvf_trans(uint32_t vd, uint32_t rs1, uint32_t vs2,\n  * If the current SEW does not correspond to a supported IEEE floating-point\n  * type, an illegal instruction exception is raised\n  */\n-static bool opfvf_check(DisasContext *s, arg_rmrr *a)\n+static bool opfvf_check(DisasContext *s, arg_rmrr *a, int8_t valid_bfa_vsew)\n {\n     return require_rvv(s) &&\n            require_rvf(s) &&\n            vext_check_isa_ill(s) &&\n-           vext_check_ss(s, a->rd, a->rs2, a->vm);\n+           vext_check_ss(s, a->rd, a->rs2, a->vm) &&\n+           vext_check_altfmt(s, valid_bfa_vsew);\n }\n \n /* OPFVF without GVEC IR */\n-#define GEN_OPFVF_TRANS(NAME, CHECK)                              \\\n-static bool trans_##NAME(DisasContext *s, arg_rmrr *a)            \\\n-{                                                                 \\\n-    if (CHECK(s, a)) {                                            \\\n-        uint32_t data = 0;                                        \\\n-        static gen_helper_opfvf *const fns[3] = {                 \\\n-            gen_helper_##NAME##_h,                                \\\n-            gen_helper_##NAME##_w,                                \\\n-            gen_helper_##NAME##_d,                                \\\n-        };                                                        \\\n-        gen_set_rm(s, RISCV_FRM_DYN);                             \\\n-        data = FIELD_DP32(data, VDATA, VM, a->vm);                \\\n-        data = FIELD_DP32(data, VDATA, LMUL, s->lmul);            \\\n-        data = FIELD_DP32(data, VDATA, VTA, s->vta);              \\\n-        data = FIELD_DP32(data, VDATA, VTA_ALL_1S,                \\\n-                          s->cfg_vta_all_1s);                     \\\n-        data = FIELD_DP32(data, VDATA, VMA, s->vma);              \\\n-        return opfvf_trans(a->rd, a->rs1, a->rs2, data,           \\\n-                           fns[s->sew - 1], s);                   \\\n-    }                                                             \\\n-    return false;                                                 \\\n-}\n-\n-GEN_OPFVF_TRANS(vfadd_vf,  opfvf_check)\n-GEN_OPFVF_TRANS(vfsub_vf,  opfvf_check)\n-GEN_OPFVF_TRANS(vfrsub_vf,  opfvf_check)\n+#define GEN_OPFVF_TRANS(NAME, CHECK)                    \\\n+static bool trans_##NAME(DisasContext *s, arg_rmrr *a)  \\\n+{                                                       \\\n+    if (CHECK(s, a, -1)) {                              \\\n+        uint32_t data = 0;                              \\\n+        static gen_helper_opfvf *const fns[3] = {       \\\n+            gen_helper_##NAME##_h,                      \\\n+            gen_helper_##NAME##_w,                      \\\n+            gen_helper_##NAME##_d,                      \\\n+        };                                              \\\n+        gen_set_rm(s, RISCV_FRM_DYN);                   \\\n+        data = FIELD_DP32(data, VDATA, VM, a->vm);      \\\n+        data = FIELD_DP32(data, VDATA, LMUL, s->lmul);  \\\n+        data = FIELD_DP32(data, VDATA, VTA, s->vta);    \\\n+        data = FIELD_DP32(data, VDATA, VTA_ALL_1S,      \\\n+                          s->cfg_vta_all_1s);           \\\n+        data = FIELD_DP32(data, VDATA, VMA, s->vma);    \\\n+        return opfvf_trans(a->rd, a->rs1, a->rs2, data, \\\n+                           fns[s->sew - 1], s);         \\\n+    }                                                   \\\n+    return false;                                       \\\n+}\n+\n+#define GEN_OPFVF_BFA_TRANS(NAME, CHECK, BFA_HELPER)                \\\n+static bool trans_##NAME(DisasContext *s, arg_rmrr *a)              \\\n+{                                                                   \\\n+    if (CHECK(s, a, MO_16)) {                                       \\\n+        uint32_t data = 0;                                          \\\n+        static gen_helper_opfvf *const fns[3] = {                   \\\n+            gen_helper_##NAME##_h,                                  \\\n+            gen_helper_##NAME##_w,                                  \\\n+            gen_helper_##NAME##_d,                                  \\\n+        };                                                          \\\n+        gen_set_rm(s, RISCV_FRM_DYN);                               \\\n+        data = FIELD_DP32(data, VDATA, VM, a->vm);                  \\\n+        data = FIELD_DP32(data, VDATA, LMUL, s->lmul);              \\\n+        data = FIELD_DP32(data, VDATA, VTA, s->vta);                \\\n+        data = FIELD_DP32(data, VDATA, VTA_ALL_1S,                  \\\n+                          s->cfg_vta_all_1s);                       \\\n+        data = FIELD_DP32(data, VDATA, VMA, s->vma);                \\\n+        return opfvf_trans(a->rd, a->rs1, a->rs2, data,             \\\n+                           (s->altfmt ? gen_helper_##BFA_HELPER :   \\\n+                                        fns[s->sew - 1]),           \\\n+                           s);                                      \\\n+    }                                                               \\\n+    return false;                                                   \\\n+}\n+\n+GEN_OPFVF_BFA_TRANS(vfadd_vf,  opfvf_check, vfadd_vf_h_bf16)\n+GEN_OPFVF_BFA_TRANS(vfsub_vf,  opfvf_check, vfsub_vf_h_bf16)\n+GEN_OPFVF_BFA_TRANS(vfrsub_vf,  opfvf_check, vfrsub_vf_h_bf16)\n \n /* Vector Widening Floating-Point Add/Subtract Instructions */\n-static bool opfvv_widen_check(DisasContext *s, arg_rmrr *a)\n+static bool opfvv_widen_check(DisasContext *s, arg_rmrr *a,\n+                              int8_t valid_bfa_vsew)\n {\n     return require_rvv(s) &&\n            require_rvf(s) &&\n            require_scale_rvf(s) &&\n            vext_check_isa_ill(s) &&\n-           vext_check_dss(s, a->rd, a->rs1, a->rs2, a->vm);\n+           vext_check_dss(s, a->rd, a->rs1, a->rs2, a->vm) &&\n+           vext_check_altfmt(s, valid_bfa_vsew);\n }\n \n-static bool opfvv_overwrite_widen_check(DisasContext *s, arg_rmrr *a)\n+static bool opfvv_overwrite_widen_check(DisasContext *s, arg_rmrr *a,\n+                                        int8_t valid_bfa_vsew)\n {\n-    return require_rvv(s) &&\n-           require_rvf(s) &&\n-           require_scale_rvf(s) &&\n-           vext_check_isa_ill(s) &&\n-           vext_check_dss(s, a->rd, a->rs1, a->rs2, a->vm) &&\n+    return opfvv_widen_check(s, a, valid_bfa_vsew) &&\n            vext_check_input_eew(s, a->rd, s->sew + 1, a->rs1, s->sew, a->vm) &&\n            vext_check_input_eew(s, a->rd, s->sew + 1, a->rs2, s->sew, a->vm);\n }\n \n /* OPFVV with WIDEN */\n-#define GEN_OPFVV_WIDEN_TRANS(NAME, CHECK)                       \\\n-static bool trans_##NAME(DisasContext *s, arg_rmrr *a)           \\\n-{                                                                \\\n-    if (CHECK(s, a)) {                                           \\\n-        uint32_t data = 0;                                       \\\n-        static gen_helper_gvec_4_ptr * const fns[2] = {          \\\n-            gen_helper_##NAME##_h, gen_helper_##NAME##_w,        \\\n-        };                                                       \\\n-        gen_set_rm(s, RISCV_FRM_DYN);                            \\\n-                                                                 \\\n-        data = FIELD_DP32(data, VDATA, VM, a->vm);               \\\n-        data = FIELD_DP32(data, VDATA, LMUL, s->lmul);           \\\n-        data = FIELD_DP32(data, VDATA, VTA, s->vta);             \\\n-        data = FIELD_DP32(data, VDATA, VMA, s->vma);             \\\n-        tcg_gen_gvec_4_ptr(vreg_ofs(s, a->rd), vreg_ofs(s, 0),   \\\n-                           vreg_ofs(s, a->rs1),                  \\\n-                           vreg_ofs(s, a->rs2), tcg_env,         \\\n-                           s->cfg_ptr->vlenb,                    \\\n-                           s->cfg_ptr->vlenb, data,              \\\n-                           fns[s->sew - 1]);                     \\\n-        finalize_rvv_inst(s);                                    \\\n-        return true;                                             \\\n-    }                                                            \\\n-    return false;                                                \\\n-}\n-\n-GEN_OPFVV_WIDEN_TRANS(vfwadd_vv, opfvv_widen_check)\n-GEN_OPFVV_WIDEN_TRANS(vfwsub_vv, opfvv_widen_check)\n-\n-static bool opfvf_widen_check(DisasContext *s, arg_rmrr *a)\n+#define GEN_OPFVV_WIDEN_TRANS(NAME, CHECK)                      \\\n+static bool trans_##NAME(DisasContext *s, arg_rmrr *a)          \\\n+{                                                               \\\n+    if (CHECK(s, a, -1)) {                                      \\\n+        uint32_t data = 0;                                      \\\n+        static gen_helper_gvec_4_ptr * const fns[2] = {         \\\n+            gen_helper_##NAME##_h, gen_helper_##NAME##_w,       \\\n+        };                                                      \\\n+        gen_set_rm(s, RISCV_FRM_DYN);                           \\\n+                                                                \\\n+        data = FIELD_DP32(data, VDATA, VM, a->vm);              \\\n+        data = FIELD_DP32(data, VDATA, LMUL, s->lmul);          \\\n+        data = FIELD_DP32(data, VDATA, VTA, s->vta);            \\\n+        data = FIELD_DP32(data, VDATA, VMA, s->vma);            \\\n+        tcg_gen_gvec_4_ptr(vreg_ofs(s, a->rd), vreg_ofs(s, 0),  \\\n+                           vreg_ofs(s, a->rs1),                 \\\n+                           vreg_ofs(s, a->rs2), tcg_env,        \\\n+                           s->cfg_ptr->vlenb,                   \\\n+                           s->cfg_ptr->vlenb, data,             \\\n+                           fns[s->sew - 1]);                    \\\n+        finalize_rvv_inst(s);                                   \\\n+        return true;                                            \\\n+    }                                                           \\\n+    return false;                                               \\\n+}\n+\n+#define GEN_OPFVV_WIDEN_BFA_TRANS(NAME, CHECK, BFA_HELPER)          \\\n+static bool trans_##NAME(DisasContext *s, arg_rmrr *a)              \\\n+{                                                                   \\\n+    if (CHECK(s, a, MO_16)) {                                       \\\n+        uint32_t data = 0;                                          \\\n+        static gen_helper_gvec_4_ptr * const fns[2] = {             \\\n+            gen_helper_##NAME##_h,                                  \\\n+            gen_helper_##NAME##_w                                   \\\n+        };                                                          \\\n+        gen_set_rm(s, RISCV_FRM_DYN);                               \\\n+                                                                    \\\n+        data = FIELD_DP32(data, VDATA, VM, a->vm);                  \\\n+        data = FIELD_DP32(data, VDATA, LMUL, s->lmul);              \\\n+        data = FIELD_DP32(data, VDATA, VTA, s->vta);                \\\n+        data = FIELD_DP32(data, VDATA, VMA, s->vma);                \\\n+        tcg_gen_gvec_4_ptr(vreg_ofs(s, a->rd), vreg_ofs(s, 0),      \\\n+                           vreg_ofs(s, a->rs1),                     \\\n+                           vreg_ofs(s, a->rs2), tcg_env,            \\\n+                           s->cfg_ptr->vlenb,                       \\\n+                           s->cfg_ptr->vlenb, data,                 \\\n+                           (s->altfmt ? gen_helper_##BFA_HELPER :   \\\n+                                        fns[s->sew - 1]));          \\\n+        finalize_rvv_inst(s);                                       \\\n+        return true;                                                \\\n+    }                                                               \\\n+    return false;                                                   \\\n+}\n+\n+GEN_OPFVV_WIDEN_BFA_TRANS(vfwadd_vv, opfvv_widen_check, vfwadd_vv_h_bf16)\n+GEN_OPFVV_WIDEN_BFA_TRANS(vfwsub_vv, opfvv_widen_check, vfwsub_vv_h_bf16)\n+\n+static bool opfvf_widen_check(DisasContext *s, arg_rmrr *a,\n+                              int8_t valid_bfa_vsew)\n {\n     return require_rvv(s) &&\n            require_rvf(s) &&\n            require_scale_rvf(s) &&\n            vext_check_isa_ill(s) &&\n-           vext_check_ds(s, a->rd, a->rs2, a->vm);\n+           vext_check_ds(s, a->rd, a->rs2, a->vm) &&\n+           vext_check_altfmt(s, valid_bfa_vsew);\n }\n \n-static bool opfvf_overwrite_widen_check(DisasContext *s, arg_rmrr *a)\n+static bool opfvf_overwrite_widen_check(DisasContext *s, arg_rmrr *a,\n+                                        int8_t valid_bfa_vsew)\n {\n-    return require_rvv(s) &&\n-           require_rvf(s) &&\n-           require_scale_rvf(s) &&\n-           vext_check_isa_ill(s) &&\n-           vext_check_ds(s, a->rd, a->rs2, a->vm) &&\n+    return opfvf_widen_check(s, a, valid_bfa_vsew) &&\n            vext_check_input_eew(s, a->rd, s->sew + 1, a->rs2, s->sew, a->vm);\n }\n \n /* OPFVF with WIDEN */\n-#define GEN_OPFVF_WIDEN_TRANS(NAME, CHECK)                       \\\n-static bool trans_##NAME(DisasContext *s, arg_rmrr *a)           \\\n-{                                                                \\\n-    if (CHECK(s, a)) {                                           \\\n-        uint32_t data = 0;                                       \\\n-        static gen_helper_opfvf *const fns[2] = {                \\\n-            gen_helper_##NAME##_h, gen_helper_##NAME##_w,        \\\n-        };                                                       \\\n-        gen_set_rm(s, RISCV_FRM_DYN);                            \\\n-        data = FIELD_DP32(data, VDATA, VM, a->vm);               \\\n-        data = FIELD_DP32(data, VDATA, LMUL, s->lmul);           \\\n-        data = FIELD_DP32(data, VDATA, VTA, s->vta);             \\\n-        data = FIELD_DP32(data, VDATA, VMA, s->vma);             \\\n-        return opfvf_trans(a->rd, a->rs1, a->rs2, data,          \\\n-                           fns[s->sew - 1], s);                  \\\n-    }                                                            \\\n-    return false;                                                \\\n-}\n-\n-GEN_OPFVF_WIDEN_TRANS(vfwadd_vf, opfvf_widen_check)\n-GEN_OPFVF_WIDEN_TRANS(vfwsub_vf, opfvf_widen_check)\n-\n-static bool opfwv_widen_check(DisasContext *s, arg_rmrr *a)\n+#define GEN_OPFVF_WIDEN_BFA_TRANS(NAME, CHECK, BFA_HELPER)          \\\n+static bool trans_##NAME(DisasContext *s, arg_rmrr *a)              \\\n+{                                                                   \\\n+    if (CHECK(s, a, MO_16)) {                                       \\\n+        uint32_t data = 0;                                          \\\n+        static gen_helper_opfvf *const fns[2] = {                   \\\n+            gen_helper_##NAME##_h,                                  \\\n+            gen_helper_##NAME##_w,                                  \\\n+        };                                                          \\\n+        gen_set_rm(s, RISCV_FRM_DYN);                               \\\n+        data = FIELD_DP32(data, VDATA, VM, a->vm);                  \\\n+        data = FIELD_DP32(data, VDATA, LMUL, s->lmul);              \\\n+        data = FIELD_DP32(data, VDATA, VTA, s->vta);                \\\n+        data = FIELD_DP32(data, VDATA, VMA, s->vma);                \\\n+        return opfvf_trans(a->rd, a->rs1, a->rs2, data,             \\\n+                           (s->altfmt ? gen_helper_##BFA_HELPER :   \\\n+                                        fns[s->sew - 1]),           \\\n+                           s);                                      \\\n+    }                                                               \\\n+    return false;                                                   \\\n+}\n+\n+GEN_OPFVF_WIDEN_BFA_TRANS(vfwadd_vf, opfvf_widen_check, vfwadd_vf_h_bf16)\n+GEN_OPFVF_WIDEN_BFA_TRANS(vfwsub_vf, opfvf_widen_check, vfwsub_vf_h_bf16)\n+\n+static bool opfwv_widen_check(DisasContext *s, arg_rmrr *a,\n+                              int8_t valid_bfa_vsew)\n {\n     return require_rvv(s) &&\n            require_rvf(s) &&\n            require_scale_rvf(s) &&\n            vext_check_isa_ill(s) &&\n-           vext_check_dds(s, a->rd, a->rs1, a->rs2, a->vm);\n+           vext_check_dds(s, a->rd, a->rs1, a->rs2, a->vm) &&\n+           vext_check_altfmt(s, valid_bfa_vsew);\n }\n \n /* WIDEN OPFVV with WIDEN */\n-#define GEN_OPFWV_WIDEN_TRANS(NAME)                                \\\n-static bool trans_##NAME(DisasContext *s, arg_rmrr *a)             \\\n-{                                                                  \\\n-    if (opfwv_widen_check(s, a)) {                                 \\\n-        uint32_t data = 0;                                         \\\n-        static gen_helper_gvec_4_ptr * const fns[2] = {            \\\n-            gen_helper_##NAME##_h, gen_helper_##NAME##_w,          \\\n-        };                                                         \\\n-        gen_set_rm(s, RISCV_FRM_DYN);                              \\\n-                                                                   \\\n-        data = FIELD_DP32(data, VDATA, VM, a->vm);                 \\\n-        data = FIELD_DP32(data, VDATA, LMUL, s->lmul);             \\\n-        data = FIELD_DP32(data, VDATA, VTA, s->vta);               \\\n-        data = FIELD_DP32(data, VDATA, VMA, s->vma);               \\\n-        tcg_gen_gvec_4_ptr(vreg_ofs(s, a->rd), vreg_ofs(s, 0),     \\\n-                           vreg_ofs(s, a->rs1),                    \\\n-                           vreg_ofs(s, a->rs2), tcg_env,           \\\n-                           s->cfg_ptr->vlenb,                      \\\n-                           s->cfg_ptr->vlenb, data,                \\\n-                           fns[s->sew - 1]);                       \\\n-        finalize_rvv_inst(s);                                      \\\n-        return true;                                               \\\n-    }                                                              \\\n-    return false;                                                  \\\n+#define GEN_OPFWV_WIDEN_BFA_TRANS(NAME)                                 \\\n+static bool trans_##NAME(DisasContext *s, arg_rmrr *a)                  \\\n+{                                                                       \\\n+    if (opfwv_widen_check(s, a, MO_16)) {                               \\\n+        uint32_t data = 0;                                              \\\n+        static gen_helper_gvec_4_ptr * const fns[2] = {                 \\\n+            gen_helper_##NAME##_h,                                      \\\n+            gen_helper_##NAME##_w                                       \\\n+        };                                                              \\\n+        gen_set_rm(s, RISCV_FRM_DYN);                                   \\\n+                                                                        \\\n+        data = FIELD_DP32(data, VDATA, VM, a->vm);                      \\\n+        data = FIELD_DP32(data, VDATA, LMUL, s->lmul);                  \\\n+        data = FIELD_DP32(data, VDATA, VTA, s->vta);                    \\\n+        data = FIELD_DP32(data, VDATA, VMA, s->vma);                    \\\n+        tcg_gen_gvec_4_ptr(vreg_ofs(s, a->rd), vreg_ofs(s, 0),          \\\n+                           vreg_ofs(s, a->rs1),                         \\\n+                           vreg_ofs(s, a->rs2), tcg_env,                \\\n+                           s->cfg_ptr->vlenb,                           \\\n+                           s->cfg_ptr->vlenb, data,                     \\\n+                           (s->altfmt ? gen_helper_##NAME##_h_bf16 :    \\\n+                                        fns[s->sew - 1]));              \\\n+        finalize_rvv_inst(s);                                           \\\n+        return true;                                                    \\\n+    }                                                                   \\\n+    return false;                                                       \\\n }\n \n-GEN_OPFWV_WIDEN_TRANS(vfwadd_wv)\n-GEN_OPFWV_WIDEN_TRANS(vfwsub_wv)\n+GEN_OPFWV_WIDEN_BFA_TRANS(vfwadd_wv)\n+GEN_OPFWV_WIDEN_BFA_TRANS(vfwsub_wv)\n \n-static bool opfwf_widen_check(DisasContext *s, arg_rmrr *a)\n+static bool opfwf_widen_check(DisasContext *s, arg_rmrr *a,\n+                              int8_t valid_bfa_vsew)\n {\n     return require_rvv(s) &&\n            require_rvf(s) &&\n            require_scale_rvf(s) &&\n            vext_check_isa_ill(s) &&\n-           vext_check_dd(s, a->rd, a->rs2, a->vm);\n+           vext_check_dd(s, a->rd, a->rs2, a->vm) &&\n+           vext_check_altfmt(s, valid_bfa_vsew);\n }\n \n /* WIDEN OPFVF with WIDEN */\n-#define GEN_OPFWF_WIDEN_TRANS(NAME)                              \\\n-static bool trans_##NAME(DisasContext *s, arg_rmrr *a)           \\\n-{                                                                \\\n-    if (opfwf_widen_check(s, a)) {                               \\\n-        uint32_t data = 0;                                       \\\n-        static gen_helper_opfvf *const fns[2] = {                \\\n-            gen_helper_##NAME##_h, gen_helper_##NAME##_w,        \\\n-        };                                                       \\\n-        gen_set_rm(s, RISCV_FRM_DYN);                            \\\n-        data = FIELD_DP32(data, VDATA, VM, a->vm);               \\\n-        data = FIELD_DP32(data, VDATA, LMUL, s->lmul);           \\\n-        data = FIELD_DP32(data, VDATA, VTA, s->vta);             \\\n-        data = FIELD_DP32(data, VDATA, VMA, s->vma);             \\\n-        return opfvf_trans(a->rd, a->rs1, a->rs2, data,          \\\n-                           fns[s->sew - 1], s);                  \\\n-    }                                                            \\\n-    return false;                                                \\\n-}\n-\n-GEN_OPFWF_WIDEN_TRANS(vfwadd_wf)\n-GEN_OPFWF_WIDEN_TRANS(vfwsub_wf)\n+#define GEN_OPFWF_WIDEN_BFA_TRANS(NAME)                                 \\\n+static bool trans_##NAME(DisasContext *s, arg_rmrr *a)                  \\\n+{                                                                       \\\n+    if (opfwf_widen_check(s, a, MO_16)) {                               \\\n+        uint32_t data = 0;                                              \\\n+        static gen_helper_opfvf *const fns[2] = {                       \\\n+            gen_helper_##NAME##_h,                                      \\\n+            gen_helper_##NAME##_w                                       \\\n+        };                                                              \\\n+        gen_set_rm(s, RISCV_FRM_DYN);                                   \\\n+        data = FIELD_DP32(data, VDATA, VM, a->vm);                      \\\n+        data = FIELD_DP32(data, VDATA, LMUL, s->lmul);                  \\\n+        data = FIELD_DP32(data, VDATA, VTA, s->vta);                    \\\n+        data = FIELD_DP32(data, VDATA, VMA, s->vma);                    \\\n+        return opfvf_trans(a->rd, a->rs1, a->rs2, data,                 \\\n+                           (s->altfmt ? gen_helper_##NAME##_h_bf16 :    \\\n+                                        fns[s->sew - 1]),               \\\n+                           s);                                          \\\n+    }                                                                   \\\n+    return false;                                                       \\\n+}\n+\n+GEN_OPFWF_WIDEN_BFA_TRANS(vfwadd_wf)\n+GEN_OPFWF_WIDEN_BFA_TRANS(vfwsub_wf)\n \n /* Vector Single-Width Floating-Point Multiply/Divide Instructions */\n-GEN_OPFVV_TRANS(vfmul_vv, opfvv_check)\n+GEN_OPFVV_BFA_TRANS(vfmul_vv, opfvv_check, vfmul_vv_h_bf16)\n GEN_OPFVV_TRANS(vfdiv_vv, opfvv_check)\n-GEN_OPFVF_TRANS(vfmul_vf,  opfvf_check)\n+GEN_OPFVF_BFA_TRANS(vfmul_vf,  opfvf_check, vfmul_vf_h_bf16)\n GEN_OPFVF_TRANS(vfdiv_vf,  opfvf_check)\n GEN_OPFVF_TRANS(vfrdiv_vf,  opfvf_check)\n \n /* Vector Widening Floating-Point Multiply */\n-GEN_OPFVV_WIDEN_TRANS(vfwmul_vv, opfvv_widen_check)\n-GEN_OPFVF_WIDEN_TRANS(vfwmul_vf, opfvf_widen_check)\n+GEN_OPFVV_WIDEN_BFA_TRANS(vfwmul_vv, opfvv_widen_check, vfwmul_vv_h_bf16)\n+GEN_OPFVF_WIDEN_BFA_TRANS(vfwmul_vf, opfvf_widen_check, vfwmul_vf_h_bf16)\n \n /* Vector Single-Width Floating-Point Fused Multiply-Add Instructions */\n-GEN_OPFVV_TRANS(vfmacc_vv, opfvv_check)\n-GEN_OPFVV_TRANS(vfnmacc_vv, opfvv_check)\n-GEN_OPFVV_TRANS(vfmsac_vv, opfvv_check)\n-GEN_OPFVV_TRANS(vfnmsac_vv, opfvv_check)\n-GEN_OPFVV_TRANS(vfmadd_vv, opfvv_check)\n-GEN_OPFVV_TRANS(vfnmadd_vv, opfvv_check)\n-GEN_OPFVV_TRANS(vfmsub_vv, opfvv_check)\n-GEN_OPFVV_TRANS(vfnmsub_vv, opfvv_check)\n-GEN_OPFVF_TRANS(vfmacc_vf, opfvf_check)\n-GEN_OPFVF_TRANS(vfnmacc_vf, opfvf_check)\n-GEN_OPFVF_TRANS(vfmsac_vf, opfvf_check)\n-GEN_OPFVF_TRANS(vfnmsac_vf, opfvf_check)\n-GEN_OPFVF_TRANS(vfmadd_vf, opfvf_check)\n-GEN_OPFVF_TRANS(vfnmadd_vf, opfvf_check)\n-GEN_OPFVF_TRANS(vfmsub_vf, opfvf_check)\n-GEN_OPFVF_TRANS(vfnmsub_vf, opfvf_check)\n+GEN_OPFVV_BFA_TRANS(vfmacc_vv, opfvv_check, vfmacc_vv_h_bf16)\n+GEN_OPFVV_BFA_TRANS(vfnmacc_vv, opfvv_check, vfnmacc_vv_h_bf16)\n+GEN_OPFVV_BFA_TRANS(vfmsac_vv, opfvv_check, vfmsac_vv_h_bf16)\n+GEN_OPFVV_BFA_TRANS(vfnmsac_vv, opfvv_check, vfnmsac_vv_h_bf16)\n+GEN_OPFVV_BFA_TRANS(vfmadd_vv, opfvv_check, vfmadd_vv_h_bf16)\n+GEN_OPFVV_BFA_TRANS(vfnmadd_vv, opfvv_check, vfnmadd_vv_h_bf16)\n+GEN_OPFVV_BFA_TRANS(vfmsub_vv, opfvv_check, vfmsub_vv_h_bf16)\n+GEN_OPFVV_BFA_TRANS(vfnmsub_vv, opfvv_check, vfnmsub_vv_h_bf16)\n+GEN_OPFVF_BFA_TRANS(vfmacc_vf, opfvf_check, vfmacc_vf_h_bf16)\n+GEN_OPFVF_BFA_TRANS(vfnmacc_vf, opfvf_check, vfnmacc_vf_h_bf16)\n+GEN_OPFVF_BFA_TRANS(vfmsac_vf, opfvf_check, vfmsac_vf_h_bf16)\n+GEN_OPFVF_BFA_TRANS(vfnmsac_vf, opfvf_check, vfnmsac_vf_h_bf16)\n+GEN_OPFVF_BFA_TRANS(vfmadd_vf, opfvf_check, vfmadd_vf_h_bf16)\n+GEN_OPFVF_BFA_TRANS(vfnmadd_vf, opfvf_check, vfnmadd_vf_h_bf16)\n+GEN_OPFVF_BFA_TRANS(vfmsub_vf, opfvf_check, vfmsub_vf_h_bf16)\n+GEN_OPFVF_BFA_TRANS(vfnmsub_vf, opfvf_check, vfnmsub_vf_h_bf16)\n \n /* Vector Widening Floating-Point Fused Multiply-Add Instructions */\n-GEN_OPFVV_WIDEN_TRANS(vfwmacc_vv, opfvv_overwrite_widen_check)\n-GEN_OPFVV_WIDEN_TRANS(vfwnmacc_vv, opfvv_overwrite_widen_check)\n-GEN_OPFVV_WIDEN_TRANS(vfwmsac_vv, opfvv_overwrite_widen_check)\n-GEN_OPFVV_WIDEN_TRANS(vfwnmsac_vv, opfvv_overwrite_widen_check)\n-GEN_OPFVF_WIDEN_TRANS(vfwmacc_vf, opfvf_overwrite_widen_check)\n-GEN_OPFVF_WIDEN_TRANS(vfwnmacc_vf, opfvf_overwrite_widen_check)\n-GEN_OPFVF_WIDEN_TRANS(vfwmsac_vf, opfvf_overwrite_widen_check)\n-GEN_OPFVF_WIDEN_TRANS(vfwnmsac_vf, opfvf_overwrite_widen_check)\n+GEN_OPFVV_WIDEN_BFA_TRANS(vfwmacc_vv, opfvv_overwrite_widen_check,\n+                          vfwmaccbf16_vv)\n+GEN_OPFVV_WIDEN_BFA_TRANS(vfwnmacc_vv, opfvv_overwrite_widen_check,\n+                          vfwnmacc_vv_h_bf16)\n+GEN_OPFVV_WIDEN_BFA_TRANS(vfwmsac_vv, opfvv_overwrite_widen_check,\n+                          vfwmsac_vv_h_bf16)\n+GEN_OPFVV_WIDEN_BFA_TRANS(vfwnmsac_vv, opfvv_overwrite_widen_check,\n+                          vfwnmsac_vv_h_bf16)\n+GEN_OPFVF_WIDEN_BFA_TRANS(vfwmacc_vf, opfvf_overwrite_widen_check,\n+                          vfwmaccbf16_vf)\n+GEN_OPFVF_WIDEN_BFA_TRANS(vfwnmacc_vf, opfvf_overwrite_widen_check,\n+                          vfwnmacc_vf_h_bf16)\n+GEN_OPFVF_WIDEN_BFA_TRANS(vfwmsac_vf, opfvf_overwrite_widen_check,\n+                          vfwmsac_vf_h_bf16)\n+GEN_OPFVF_WIDEN_BFA_TRANS(vfwnmsac_vf, opfvf_overwrite_widen_check,\n+                          vfwnmsac_vf_h_bf16)\n \n /* Vector Floating-Point Square-Root Instruction */\n \n@@ -2660,21 +2788,23 @@ GEN_OPFVF_WIDEN_TRANS(vfwnmsac_vf, opfvf_overwrite_widen_check)\n  * If the current SEW does not correspond to a supported IEEE floating-point\n  * type, an illegal instruction exception is raised\n  */\n-static bool opfv_check(DisasContext *s, arg_rmr *a)\n+static bool opfv_check(DisasContext *s, arg_rmr *a, int8_t valid_bfa_vsew)\n {\n     return require_rvv(s) &&\n            require_rvf(s) &&\n            vext_check_isa_ill(s) &&\n            /* OPFV instructions ignore vs1 check */\n-           vext_check_ss(s, a->rd, a->rs2, a->vm);\n+           vext_check_ss(s, a->rd, a->rs2, a->vm) &&\n+           vext_check_altfmt(s, valid_bfa_vsew);\n }\n \n static bool do_opfv(DisasContext *s, arg_rmr *a,\n                     gen_helper_gvec_3_ptr *fn,\n-                    bool (*checkfn)(DisasContext *, arg_rmr *),\n-                    int rm)\n+                    bool (*checkfn)(DisasContext *, arg_rmr *, int8_t),\n+                    int rm,\n+                    int8_t valid_bfa_vsew)\n {\n-    if (checkfn(s, a)) {\n+    if (checkfn(s, a, valid_bfa_vsew)) {\n         uint32_t data = 0;\n         gen_set_rm_chkfrm(s, rm);\n \n@@ -2692,76 +2822,95 @@ static bool do_opfv(DisasContext *s, arg_rmr *a,\n     return false;\n }\n \n-#define GEN_OPFV_TRANS(NAME, CHECK, FRM)               \\\n-static bool trans_##NAME(DisasContext *s, arg_rmr *a)  \\\n-{                                                      \\\n-    static gen_helper_gvec_3_ptr * const fns[3] = {    \\\n-        gen_helper_##NAME##_h,                         \\\n-        gen_helper_##NAME##_w,                         \\\n-        gen_helper_##NAME##_d                          \\\n-    };                                                 \\\n-    return do_opfv(s, a, fns[s->sew - 1], CHECK, FRM); \\\n+#define GEN_OPFV_TRANS(NAME, CHECK, FRM)                    \\\n+static bool trans_##NAME(DisasContext *s, arg_rmr *a)       \\\n+{                                                           \\\n+    static gen_helper_gvec_3_ptr * const fns[3] = {         \\\n+        gen_helper_##NAME##_h,                              \\\n+        gen_helper_##NAME##_w,                              \\\n+        gen_helper_##NAME##_d                               \\\n+    };                                                      \\\n+    return do_opfv(s, a, fns[s->sew - 1], CHECK, FRM, -1);  \\\n+}\n+\n+#define GEN_OPFV_BFA_TRANS(NAME, CHECK, FRM)                    \\\n+static bool trans_##NAME(DisasContext *s, arg_rmr *a)           \\\n+{                                                               \\\n+    static gen_helper_gvec_3_ptr * const fns[3] = {             \\\n+        gen_helper_##NAME##_h,                                  \\\n+        gen_helper_##NAME##_w,                                  \\\n+        gen_helper_##NAME##_d                                   \\\n+    };                                                          \\\n+    return do_opfv(s, a,                                        \\\n+                   (s->altfmt ? gen_helper_##NAME##_h_bf16 :    \\\n+                                fns[s->sew - 1]),               \\\n+                   CHECK, FRM, MO_16);                          \\\n }\n \n GEN_OPFV_TRANS(vfsqrt_v, opfv_check, RISCV_FRM_DYN)\n-GEN_OPFV_TRANS(vfrsqrt7_v, opfv_check, RISCV_FRM_DYN)\n-GEN_OPFV_TRANS(vfrec7_v, opfv_check, RISCV_FRM_DYN)\n+GEN_OPFV_BFA_TRANS(vfrsqrt7_v, opfv_check, RISCV_FRM_DYN)\n+GEN_OPFV_BFA_TRANS(vfrec7_v, opfv_check, RISCV_FRM_DYN)\n \n /* Vector Floating-Point MIN/MAX Instructions */\n-GEN_OPFVV_TRANS(vfmin_vv, opfvv_check)\n-GEN_OPFVV_TRANS(vfmax_vv, opfvv_check)\n-GEN_OPFVF_TRANS(vfmin_vf, opfvf_check)\n-GEN_OPFVF_TRANS(vfmax_vf, opfvf_check)\n+GEN_OPFVV_BFA_TRANS(vfmin_vv, opfvv_check, vfmin_vv_h_bf16)\n+GEN_OPFVV_BFA_TRANS(vfmax_vv, opfvv_check, vfmax_vv_h_bf16)\n+GEN_OPFVF_BFA_TRANS(vfmin_vf, opfvf_check, vfmin_vf_h_bf16)\n+GEN_OPFVF_BFA_TRANS(vfmax_vf, opfvf_check, vfmax_vf_h_bf16)\n \n /* Vector Floating-Point Sign-Injection Instructions */\n-GEN_OPFVV_TRANS(vfsgnj_vv, opfvv_check)\n-GEN_OPFVV_TRANS(vfsgnjn_vv, opfvv_check)\n-GEN_OPFVV_TRANS(vfsgnjx_vv, opfvv_check)\n-GEN_OPFVF_TRANS(vfsgnj_vf, opfvf_check)\n-GEN_OPFVF_TRANS(vfsgnjn_vf, opfvf_check)\n-GEN_OPFVF_TRANS(vfsgnjx_vf, opfvf_check)\n+GEN_OPFVV_BFA_TRANS(vfsgnj_vv, opfvv_check, vfsgnj_vv_h)\n+GEN_OPFVV_BFA_TRANS(vfsgnjn_vv, opfvv_check, vfsgnjn_vv_h)\n+GEN_OPFVV_BFA_TRANS(vfsgnjx_vv, opfvv_check, vfsgnjx_vv_h)\n+GEN_OPFVF_BFA_TRANS(vfsgnj_vf, opfvf_check, vfsgnj_vf_h)\n+GEN_OPFVF_BFA_TRANS(vfsgnjn_vf, opfvf_check, vfsgnjn_vf_h)\n+GEN_OPFVF_BFA_TRANS(vfsgnjx_vf, opfvf_check, vfsgnjx_vf_h)\n \n /* Vector Floating-Point Compare Instructions */\n-static bool opfvv_cmp_check(DisasContext *s, arg_rmrr *a)\n+static bool opfvv_cmp_check(DisasContext *s, arg_rmrr *a,\n+                            int8_t valid_bfa_vsew)\n {\n     return require_rvv(s) &&\n            require_rvf(s) &&\n            vext_check_isa_ill(s) &&\n-           vext_check_mss(s, a->rd, a->rs1, a->rs2);\n+           vext_check_mss(s, a->rd, a->rs1, a->rs2) &&\n+           vext_check_altfmt(s, valid_bfa_vsew);\n }\n \n-GEN_OPFVV_TRANS(vmfeq_vv, opfvv_cmp_check)\n-GEN_OPFVV_TRANS(vmfne_vv, opfvv_cmp_check)\n-GEN_OPFVV_TRANS(vmflt_vv, opfvv_cmp_check)\n-GEN_OPFVV_TRANS(vmfle_vv, opfvv_cmp_check)\n+GEN_OPFVV_BFA_TRANS(vmfeq_vv, opfvv_cmp_check, vmfeq_vv_h_bf16)\n+GEN_OPFVV_BFA_TRANS(vmfne_vv, opfvv_cmp_check, vmfne_vv_h_bf16)\n+GEN_OPFVV_BFA_TRANS(vmflt_vv, opfvv_cmp_check, vmflt_vv_h_bf16)\n+GEN_OPFVV_BFA_TRANS(vmfle_vv, opfvv_cmp_check, vmfle_vv_h_bf16)\n \n-static bool opfvf_cmp_check(DisasContext *s, arg_rmrr *a)\n+static bool opfvf_cmp_check(DisasContext *s, arg_rmrr *a,\n+                            int8_t valid_bfa_vsew)\n {\n     return require_rvv(s) &&\n            require_rvf(s) &&\n            vext_check_isa_ill(s) &&\n-           vext_check_ms(s, a->rd, a->rs2);\n+           vext_check_ms(s, a->rd, a->rs2) &&\n+           vext_check_altfmt(s, valid_bfa_vsew);\n }\n \n-GEN_OPFVF_TRANS(vmfeq_vf, opfvf_cmp_check)\n-GEN_OPFVF_TRANS(vmfne_vf, opfvf_cmp_check)\n-GEN_OPFVF_TRANS(vmflt_vf, opfvf_cmp_check)\n-GEN_OPFVF_TRANS(vmfle_vf, opfvf_cmp_check)\n-GEN_OPFVF_TRANS(vmfgt_vf, opfvf_cmp_check)\n-GEN_OPFVF_TRANS(vmfge_vf, opfvf_cmp_check)\n+GEN_OPFVF_BFA_TRANS(vmfeq_vf, opfvf_cmp_check, vmfeq_vf_h_bf16)\n+GEN_OPFVF_BFA_TRANS(vmfne_vf, opfvf_cmp_check, vmfne_vf_h_bf16)\n+GEN_OPFVF_BFA_TRANS(vmflt_vf, opfvf_cmp_check, vmflt_vf_h_bf16)\n+GEN_OPFVF_BFA_TRANS(vmfle_vf, opfvf_cmp_check, vmfle_vf_h_bf16)\n+GEN_OPFVF_BFA_TRANS(vmfgt_vf, opfvf_cmp_check, vmfgt_vf_h_bf16)\n+GEN_OPFVF_BFA_TRANS(vmfge_vf, opfvf_cmp_check, vmfge_vf_h_bf16)\n \n /* Vector Floating-Point Classify Instruction */\n-GEN_OPFV_TRANS(vfclass_v, opfv_check, RISCV_FRM_DYN)\n+GEN_OPFV_BFA_TRANS(vfclass_v, opfv_check, RISCV_FRM_DYN)\n \n /* Vector Floating-Point Merge Instruction */\n-GEN_OPFVF_TRANS(vfmerge_vfm,  opfvf_check)\n+GEN_OPFVF_BFA_TRANS(vfmerge_vfm,  opfvf_check, vfmerge_vfm_h)\n \n static bool trans_vfmv_v_f(DisasContext *s, arg_vfmv_v_f *a)\n {\n     if (require_rvv(s) &&\n         require_rvf(s) &&\n         vext_check_isa_ill(s) &&\n-        require_align(a->rd, s->lmul)) {\n+        require_align(a->rd, s->lmul) &&\n+        vext_check_altfmt(s, MO_16)) {\n         gen_set_rm(s, RISCV_FRM_DYN);\n \n         TCGv_i64 t1;\n@@ -2782,7 +2931,7 @@ static bool trans_vfmv_v_f(DisasContext *s, arg_vfmv_v_f *a)\n             static gen_helper_vmv_vx * const fns[3] = {\n                 gen_helper_vmv_v_x_h,\n                 gen_helper_vmv_v_x_w,\n-                gen_helper_vmv_v_x_d,\n+                gen_helper_vmv_v_x_d\n             };\n \n             t1 = tcg_temp_new_i64();\n@@ -2803,15 +2952,15 @@ static bool trans_vfmv_v_f(DisasContext *s, arg_vfmv_v_f *a)\n }\n \n /* Single-Width Floating-Point/Integer Type-Convert Instructions */\n-#define GEN_OPFV_CVT_TRANS(NAME, HELPER, FRM)               \\\n-static bool trans_##NAME(DisasContext *s, arg_rmr *a)       \\\n-{                                                           \\\n-    static gen_helper_gvec_3_ptr * const fns[3] = {         \\\n-        gen_helper_##HELPER##_h,                            \\\n-        gen_helper_##HELPER##_w,                            \\\n-        gen_helper_##HELPER##_d                             \\\n-    };                                                      \\\n-    return do_opfv(s, a, fns[s->sew - 1], opfv_check, FRM); \\\n+#define GEN_OPFV_CVT_TRANS(NAME, HELPER, FRM)                   \\\n+static bool trans_##NAME(DisasContext *s, arg_rmr *a)           \\\n+{                                                               \\\n+    static gen_helper_gvec_3_ptr * const fns[3] = {             \\\n+        gen_helper_##HELPER##_h,                                \\\n+        gen_helper_##HELPER##_w,                                \\\n+        gen_helper_##HELPER##_d                                 \\\n+    };                                                          \\\n+    return do_opfv(s, a, fns[s->sew - 1], opfv_check, FRM, -1); \\\n }\n \n GEN_OPFV_CVT_TRANS(vfcvt_xu_f_v, vfcvt_xu_f_v, RISCV_FRM_DYN)\n@@ -2835,95 +2984,129 @@ static bool opfv_widen_check(DisasContext *s, arg_rmr *a)\n            vext_check_ds(s, a->rd, a->rs2, a->vm);\n }\n \n-static bool opxfv_widen_check(DisasContext *s, arg_rmr *a)\n+static bool opxfv_widen_check(DisasContext *s, arg_rmr *a,\n+                              int8_t valid_bfa_vsew)\n {\n     return opfv_widen_check(s, a) &&\n-           require_rvf(s);\n+           require_rvf(s) &&\n+           vext_check_altfmt(s, valid_bfa_vsew);\n }\n \n-static bool opffv_widen_check(DisasContext *s, arg_rmr *a)\n+static bool opffv_widen_check(DisasContext *s, arg_rmr *a,\n+                              int8_t valid_bfa_vsew)\n {\n     return opfv_widen_check(s, a) &&\n            require_rvfmin(s) &&\n-           require_scale_rvfmin(s);\n-}\n-\n-#define GEN_OPFV_WIDEN_TRANS(NAME, CHECK, HELPER, FRM)             \\\n-static bool trans_##NAME(DisasContext *s, arg_rmr *a)              \\\n-{                                                                  \\\n-    if (CHECK(s, a)) {                                             \\\n-        uint32_t data = 0;                                         \\\n-        static gen_helper_gvec_3_ptr * const fns[2] = {            \\\n-            gen_helper_##HELPER##_h,                               \\\n-            gen_helper_##HELPER##_w,                               \\\n-        };                                                         \\\n-        gen_set_rm_chkfrm(s, FRM);                                 \\\n-                                                                   \\\n-        data = FIELD_DP32(data, VDATA, VM, a->vm);                 \\\n-        data = FIELD_DP32(data, VDATA, LMUL, s->lmul);             \\\n-        data = FIELD_DP32(data, VDATA, VTA, s->vta);               \\\n-        data = FIELD_DP32(data, VDATA, VMA, s->vma);               \\\n-        tcg_gen_gvec_3_ptr(vreg_ofs(s, a->rd), vreg_ofs(s, 0),     \\\n-                           vreg_ofs(s, a->rs2), tcg_env,           \\\n-                           s->cfg_ptr->vlenb,                      \\\n-                           s->cfg_ptr->vlenb, data,                \\\n-                           fns[s->sew - 1]);                       \\\n-        finalize_rvv_inst(s);                                      \\\n-        return true;                                               \\\n-    }                                                              \\\n-    return false;                                                  \\\n+           require_scale_rvfmin(s) &&\n+           vext_check_altfmt(s, valid_bfa_vsew);\n+}\n+\n+#define GEN_OPFV_WIDEN_TRANS(NAME, CHECK, HELPER, FRM)          \\\n+static bool trans_##NAME(DisasContext *s, arg_rmr *a)           \\\n+{                                                               \\\n+    if (CHECK(s, a, -1)) {                                      \\\n+        uint32_t data = 0;                                      \\\n+        static gen_helper_gvec_3_ptr * const fns[2] = {         \\\n+            gen_helper_##HELPER##_h,                            \\\n+            gen_helper_##HELPER##_w,                            \\\n+        };                                                      \\\n+        gen_set_rm_chkfrm(s, FRM);                              \\\n+                                                                \\\n+        data = FIELD_DP32(data, VDATA, VM, a->vm);              \\\n+        data = FIELD_DP32(data, VDATA, LMUL, s->lmul);          \\\n+        data = FIELD_DP32(data, VDATA, VTA, s->vta);            \\\n+        data = FIELD_DP32(data, VDATA, VMA, s->vma);            \\\n+        tcg_gen_gvec_3_ptr(vreg_ofs(s, a->rd), vreg_ofs(s, 0),  \\\n+                           vreg_ofs(s, a->rs2), tcg_env,        \\\n+                           s->cfg_ptr->vlenb,                   \\\n+                           s->cfg_ptr->vlenb, data,             \\\n+                           fns[s->sew - 1]);                    \\\n+        finalize_rvv_inst(s);                                   \\\n+        return true;                                            \\\n+    }                                                           \\\n+    return false;                                               \\\n+}\n+\n+#define GEN_OPFV_WIDEN_BFA_TRANS(NAME, CHECK, HELPER, FRM, BFA_HELPER)  \\\n+static bool trans_##NAME(DisasContext *s, arg_rmr *a)                   \\\n+{                                                                       \\\n+    if (CHECK(s, a, MO_16)) {                                           \\\n+        uint32_t data = 0;                                              \\\n+        static gen_helper_gvec_3_ptr * const fns[2] = {                 \\\n+            gen_helper_##HELPER##_h,                                    \\\n+            gen_helper_##HELPER##_w,                                    \\\n+        };                                                              \\\n+        gen_set_rm_chkfrm(s, FRM);                                      \\\n+                                                                        \\\n+        data = FIELD_DP32(data, VDATA, VM, a->vm);                      \\\n+        data = FIELD_DP32(data, VDATA, LMUL, s->lmul);                  \\\n+        data = FIELD_DP32(data, VDATA, VTA, s->vta);                    \\\n+        data = FIELD_DP32(data, VDATA, VMA, s->vma);                    \\\n+        tcg_gen_gvec_3_ptr(vreg_ofs(s, a->rd), vreg_ofs(s, 0),          \\\n+                           vreg_ofs(s, a->rs2), tcg_env,                \\\n+                           s->cfg_ptr->vlenb,                           \\\n+                           s->cfg_ptr->vlenb, data,                     \\\n+                           (s->altfmt ? gen_helper_##BFA_HELPER :       \\\n+                                        fns[s->sew - 1]));              \\\n+        finalize_rvv_inst(s);                                           \\\n+        return true;                                                    \\\n+    }                                                                   \\\n+    return false;                                                       \\\n }\n \n GEN_OPFV_WIDEN_TRANS(vfwcvt_xu_f_v, opxfv_widen_check, vfwcvt_xu_f_v,\n                      RISCV_FRM_DYN)\n GEN_OPFV_WIDEN_TRANS(vfwcvt_x_f_v, opxfv_widen_check, vfwcvt_x_f_v,\n                      RISCV_FRM_DYN)\n-GEN_OPFV_WIDEN_TRANS(vfwcvt_f_f_v, opffv_widen_check, vfwcvt_f_f_v,\n-                     RISCV_FRM_DYN)\n+GEN_OPFV_WIDEN_BFA_TRANS(vfwcvt_f_f_v, opffv_widen_check, vfwcvt_f_f_v,\n+                         RISCV_FRM_DYN, vfwcvtbf16_f_f_v)\n /* Reuse the helper functions from vfwcvt.xu.f.v and vfwcvt.x.f.v */\n GEN_OPFV_WIDEN_TRANS(vfwcvt_rtz_xu_f_v, opxfv_widen_check, vfwcvt_xu_f_v,\n                      RISCV_FRM_RTZ)\n GEN_OPFV_WIDEN_TRANS(vfwcvt_rtz_x_f_v, opxfv_widen_check, vfwcvt_x_f_v,\n                      RISCV_FRM_RTZ)\n \n-static bool opfxv_widen_check(DisasContext *s, arg_rmr *a)\n+static bool opfxv_widen_check(DisasContext *s, arg_rmr *a,\n+                              int8_t valid_bfa_vsew)\n {\n     return require_rvv(s) &&\n            require_scale_rvf(s) &&\n            vext_check_isa_ill(s) &&\n            /* OPFV widening instructions ignore vs1 check */\n-           vext_check_ds(s, a->rd, a->rs2, a->vm);\n+           vext_check_ds(s, a->rd, a->rs2, a->vm) &&\n+           vext_check_altfmt(s, valid_bfa_vsew);\n }\n \n-#define GEN_OPFXV_WIDEN_TRANS(NAME)                                \\\n-static bool trans_##NAME(DisasContext *s, arg_rmr *a)              \\\n-{                                                                  \\\n-    if (opfxv_widen_check(s, a)) {                                 \\\n-        uint32_t data = 0;                                         \\\n-        static gen_helper_gvec_3_ptr * const fns[3] = {            \\\n-            gen_helper_##NAME##_b,                                 \\\n-            gen_helper_##NAME##_h,                                 \\\n-            gen_helper_##NAME##_w,                                 \\\n-        };                                                         \\\n-        gen_set_rm(s, RISCV_FRM_DYN);                              \\\n-                                                                   \\\n-        data = FIELD_DP32(data, VDATA, VM, a->vm);                 \\\n-        data = FIELD_DP32(data, VDATA, LMUL, s->lmul);             \\\n-        data = FIELD_DP32(data, VDATA, VTA, s->vta);               \\\n-        data = FIELD_DP32(data, VDATA, VMA, s->vma);               \\\n-        tcg_gen_gvec_3_ptr(vreg_ofs(s, a->rd), vreg_ofs(s, 0),     \\\n-                           vreg_ofs(s, a->rs2), tcg_env,           \\\n-                           s->cfg_ptr->vlenb,                      \\\n-                           s->cfg_ptr->vlenb, data,                \\\n-                           fns[s->sew]);                           \\\n-        finalize_rvv_inst(s);                                      \\\n-        return true;                                               \\\n-    }                                                              \\\n-    return false;                                                  \\\n+#define GEN_OPFXV_WIDEN_BFA_TRANS(NAME)                                 \\\n+static bool trans_##NAME(DisasContext *s, arg_rmr *a)                   \\\n+{                                                                       \\\n+    if (opfxv_widen_check(s, a, MO_8)) {                                \\\n+        uint32_t data = 0;                                              \\\n+        static gen_helper_gvec_3_ptr * const fns[3] = {                 \\\n+            gen_helper_##NAME##_b,                                      \\\n+            gen_helper_##NAME##_h,                                      \\\n+            gen_helper_##NAME##_w                                       \\\n+        };                                                              \\\n+        gen_set_rm(s, RISCV_FRM_DYN);                                   \\\n+                                                                        \\\n+        data = FIELD_DP32(data, VDATA, VM, a->vm);                      \\\n+        data = FIELD_DP32(data, VDATA, LMUL, s->lmul);                  \\\n+        data = FIELD_DP32(data, VDATA, VTA, s->vta);                    \\\n+        data = FIELD_DP32(data, VDATA, VMA, s->vma);                    \\\n+        tcg_gen_gvec_3_ptr(vreg_ofs(s, a->rd), vreg_ofs(s, 0),          \\\n+                           vreg_ofs(s, a->rs2), tcg_env,                \\\n+                           s->cfg_ptr->vlenb,                           \\\n+                           s->cfg_ptr->vlenb, data,                     \\\n+                           (s->altfmt ? gen_helper_##NAME##_b_bf16 :    \\\n+                                        fns[s->sew]));                  \\\n+        finalize_rvv_inst(s);                                           \\\n+        return true;                                                    \\\n+    }                                                                   \\\n+    return false;                                                       \\\n }\n \n-GEN_OPFXV_WIDEN_TRANS(vfwcvt_f_xu_v)\n-GEN_OPFXV_WIDEN_TRANS(vfwcvt_f_x_v)\n+GEN_OPFXV_WIDEN_BFA_TRANS(vfwcvt_f_xu_v)\n+GEN_OPFXV_WIDEN_BFA_TRANS(vfwcvt_f_x_v)\n \n /* Narrowing Floating-Point/Integer Type-Convert Instructions */\n \n@@ -2939,104 +3122,140 @@ static bool opfv_narrow_check(DisasContext *s, arg_rmr *a)\n            vext_check_sd(s, a->rd, a->rs2, a->vm);\n }\n \n-static bool opfxv_narrow_check(DisasContext *s, arg_rmr *a)\n+static bool opfxv_narrow_check(DisasContext *s, arg_rmr *a,\n+                               int8_t valid_bfa_vsew)\n {\n     return opfv_narrow_check(s, a) &&\n            require_rvf(s) &&\n-           (s->sew != MO_64);\n+           (s->sew != MO_64) &&\n+           vext_check_altfmt(s, valid_bfa_vsew);\n }\n \n-static bool opffv_narrow_check(DisasContext *s, arg_rmr *a)\n+static bool opffv_narrow_check(DisasContext *s, arg_rmr *a,\n+                               int8_t valid_bfa_vsew)\n {\n     return opfv_narrow_check(s, a) &&\n            require_rvfmin(s) &&\n-           require_scale_rvfmin(s);\n+           require_scale_rvfmin(s) &&\n+           vext_check_altfmt(s, valid_bfa_vsew);\n }\n \n-static bool opffv_rod_narrow_check(DisasContext *s, arg_rmr *a)\n+static bool opffv_rod_narrow_check(DisasContext *s, arg_rmr *a,\n+                                   int8_t valid_bfa_vsew)\n {\n     return opfv_narrow_check(s, a) &&\n            require_rvf(s) &&\n-           require_scale_rvf(s);\n-}\n-\n-#define GEN_OPFV_NARROW_TRANS(NAME, CHECK, HELPER, FRM)            \\\n-static bool trans_##NAME(DisasContext *s, arg_rmr *a)              \\\n-{                                                                  \\\n-    if (CHECK(s, a)) {                                             \\\n-        uint32_t data = 0;                                         \\\n-        static gen_helper_gvec_3_ptr * const fns[2] = {            \\\n-            gen_helper_##HELPER##_h,                               \\\n-            gen_helper_##HELPER##_w,                               \\\n-        };                                                         \\\n-        gen_set_rm_chkfrm(s, FRM);                                 \\\n-                                                                   \\\n-        data = FIELD_DP32(data, VDATA, VM, a->vm);                 \\\n-        data = FIELD_DP32(data, VDATA, LMUL, s->lmul);             \\\n-        data = FIELD_DP32(data, VDATA, VTA, s->vta);               \\\n-        data = FIELD_DP32(data, VDATA, VMA, s->vma);               \\\n-        tcg_gen_gvec_3_ptr(vreg_ofs(s, a->rd), vreg_ofs(s, 0),     \\\n-                           vreg_ofs(s, a->rs2), tcg_env,           \\\n-                           s->cfg_ptr->vlenb,                      \\\n-                           s->cfg_ptr->vlenb, data,                \\\n-                           fns[s->sew - 1]);                       \\\n-        finalize_rvv_inst(s);                                      \\\n-        return true;                                               \\\n-    }                                                              \\\n-    return false;                                                  \\\n+           require_scale_rvf(s) &&\n+           vext_check_altfmt(s, valid_bfa_vsew);\n+}\n+\n+#define GEN_OPFV_NARROW_TRANS(NAME, CHECK, HELPER, FRM)         \\\n+static bool trans_##NAME(DisasContext *s, arg_rmr *a)           \\\n+{                                                               \\\n+    if (CHECK(s, a, -1)) {                                      \\\n+        uint32_t data = 0;                                      \\\n+        static gen_helper_gvec_3_ptr * const fns[2] = {         \\\n+            gen_helper_##HELPER##_h,                            \\\n+            gen_helper_##HELPER##_w,                            \\\n+        };                                                      \\\n+        gen_set_rm_chkfrm(s, FRM);                              \\\n+                                                                \\\n+        data = FIELD_DP32(data, VDATA, VM, a->vm);              \\\n+        data = FIELD_DP32(data, VDATA, LMUL, s->lmul);          \\\n+        data = FIELD_DP32(data, VDATA, VTA, s->vta);            \\\n+        data = FIELD_DP32(data, VDATA, VMA, s->vma);            \\\n+        tcg_gen_gvec_3_ptr(vreg_ofs(s, a->rd), vreg_ofs(s, 0),  \\\n+                           vreg_ofs(s, a->rs2), tcg_env,        \\\n+                           s->cfg_ptr->vlenb,                   \\\n+                           s->cfg_ptr->vlenb, data,             \\\n+                           fns[s->sew - 1]);                    \\\n+        finalize_rvv_inst(s);                                   \\\n+        return true;                                            \\\n+    }                                                           \\\n+    return false;                                               \\\n+}\n+\n+#define GEN_OPFV_NARROW_BFA_TRANS(NAME, CHECK, HELPER, FRM, BFA_HELPER) \\\n+static bool trans_##NAME(DisasContext *s, arg_rmr *a)                   \\\n+{                                                                       \\\n+    if (CHECK(s, a, MO_16)) {                                           \\\n+        uint32_t data = 0;                                              \\\n+        static gen_helper_gvec_3_ptr * const fns[2] = {                 \\\n+            gen_helper_##HELPER##_h,                                    \\\n+            gen_helper_##HELPER##_w,                                    \\\n+        };                                                              \\\n+        gen_set_rm_chkfrm(s, FRM);                                      \\\n+                                                                        \\\n+        data = FIELD_DP32(data, VDATA, VM, a->vm);                      \\\n+        data = FIELD_DP32(data, VDATA, LMUL, s->lmul);                  \\\n+        data = FIELD_DP32(data, VDATA, VTA, s->vta);                    \\\n+        data = FIELD_DP32(data, VDATA, VMA, s->vma);                    \\\n+        tcg_gen_gvec_3_ptr(vreg_ofs(s, a->rd), vreg_ofs(s, 0),          \\\n+                           vreg_ofs(s, a->rs2), tcg_env,                \\\n+                           s->cfg_ptr->vlenb,                           \\\n+                           s->cfg_ptr->vlenb, data,                     \\\n+                           (s->altfmt ? gen_helper_##BFA_HELPER :       \\\n+                                        fns[s->sew - 1]));              \\\n+        finalize_rvv_inst(s);                                           \\\n+        return true;                                                    \\\n+    }                                                                   \\\n+    return false;                                                       \\\n }\n \n GEN_OPFV_NARROW_TRANS(vfncvt_f_xu_w, opfxv_narrow_check, vfncvt_f_xu_w,\n                       RISCV_FRM_DYN)\n GEN_OPFV_NARROW_TRANS(vfncvt_f_x_w, opfxv_narrow_check, vfncvt_f_x_w,\n                       RISCV_FRM_DYN)\n-GEN_OPFV_NARROW_TRANS(vfncvt_f_f_w, opffv_narrow_check, vfncvt_f_f_w,\n-                      RISCV_FRM_DYN)\n+GEN_OPFV_NARROW_BFA_TRANS(vfncvt_f_f_w, opffv_narrow_check, vfncvt_f_f_w,\n+                          RISCV_FRM_DYN, vfncvtbf16_f_f_w)\n /* Reuse the helper function from vfncvt.f.f.w */\n-GEN_OPFV_NARROW_TRANS(vfncvt_rod_f_f_w, opffv_rod_narrow_check, vfncvt_f_f_w,\n-                      RISCV_FRM_ROD)\n+GEN_OPFV_NARROW_BFA_TRANS(vfncvt_rod_f_f_w, opffv_rod_narrow_check,\n+                          vfncvt_f_f_w, RISCV_FRM_ROD, vfncvtbf16_f_f_w)\n \n-static bool opxfv_narrow_check(DisasContext *s, arg_rmr *a)\n+static bool opxfv_narrow_check(DisasContext *s, arg_rmr *a,\n+                               int8_t valid_bfa_vsew)\n {\n     return require_rvv(s) &&\n            require_scale_rvf(s) &&\n            vext_check_isa_ill(s) &&\n            /* OPFV narrowing instructions ignore vs1 check */\n-           vext_check_sd(s, a->rd, a->rs2, a->vm);\n+           vext_check_sd(s, a->rd, a->rs2, a->vm) &&\n+           vext_check_altfmt(s, valid_bfa_vsew);\n }\n \n-#define GEN_OPXFV_NARROW_TRANS(NAME, HELPER, FRM)                  \\\n-static bool trans_##NAME(DisasContext *s, arg_rmr *a)              \\\n-{                                                                  \\\n-    if (opxfv_narrow_check(s, a)) {                                \\\n-        uint32_t data = 0;                                         \\\n-        static gen_helper_gvec_3_ptr * const fns[3] = {            \\\n-            gen_helper_##HELPER##_b,                               \\\n-            gen_helper_##HELPER##_h,                               \\\n-            gen_helper_##HELPER##_w,                               \\\n-        };                                                         \\\n-        gen_set_rm_chkfrm(s, FRM);                                 \\\n-                                                                   \\\n-        data = FIELD_DP32(data, VDATA, VM, a->vm);                 \\\n-        data = FIELD_DP32(data, VDATA, LMUL, s->lmul);             \\\n-        data = FIELD_DP32(data, VDATA, VTA, s->vta);               \\\n-        data = FIELD_DP32(data, VDATA, VMA, s->vma);               \\\n-        tcg_gen_gvec_3_ptr(vreg_ofs(s, a->rd), vreg_ofs(s, 0),     \\\n-                           vreg_ofs(s, a->rs2), tcg_env,           \\\n-                           s->cfg_ptr->vlenb,                      \\\n-                           s->cfg_ptr->vlenb, data,                \\\n-                           fns[s->sew]);                           \\\n-        finalize_rvv_inst(s);                                      \\\n-        return true;                                               \\\n-    }                                                              \\\n-    return false;                                                  \\\n+#define GEN_OPXFV_NARROW_BFA_TRANS(NAME, HELPER, FRM)                   \\\n+static bool trans_##NAME(DisasContext *s, arg_rmr *a)                   \\\n+{                                                                       \\\n+    if (opxfv_narrow_check(s, a, MO_8)) {                               \\\n+        uint32_t data = 0;                                              \\\n+        static gen_helper_gvec_3_ptr * const fns[3] = {                 \\\n+            gen_helper_##HELPER##_b,                                    \\\n+            gen_helper_##HELPER##_h,                                    \\\n+            gen_helper_##HELPER##_w                                     \\\n+        };                                                              \\\n+        gen_set_rm_chkfrm(s, FRM);                                      \\\n+                                                                        \\\n+        data = FIELD_DP32(data, VDATA, VM, a->vm);                      \\\n+        data = FIELD_DP32(data, VDATA, LMUL, s->lmul);                  \\\n+        data = FIELD_DP32(data, VDATA, VTA, s->vta);                    \\\n+        data = FIELD_DP32(data, VDATA, VMA, s->vma);                    \\\n+        tcg_gen_gvec_3_ptr(vreg_ofs(s, a->rd), vreg_ofs(s, 0),          \\\n+                           vreg_ofs(s, a->rs2), tcg_env,                \\\n+                           s->cfg_ptr->vlenb,                           \\\n+                           s->cfg_ptr->vlenb, data,                     \\\n+                           (s->altfmt ? gen_helper_##HELPER##_b_bf16 :  \\\n+                                        fns[s->sew]));                  \\\n+        finalize_rvv_inst(s);                                           \\\n+        return true;                                                    \\\n+    }                                                                   \\\n+    return false;                                                       \\\n }\n \n-GEN_OPXFV_NARROW_TRANS(vfncvt_xu_f_w, vfncvt_xu_f_w, RISCV_FRM_DYN)\n-GEN_OPXFV_NARROW_TRANS(vfncvt_x_f_w, vfncvt_x_f_w, RISCV_FRM_DYN)\n+GEN_OPXFV_NARROW_BFA_TRANS(vfncvt_xu_f_w, vfncvt_xu_f_w, RISCV_FRM_DYN)\n+GEN_OPXFV_NARROW_BFA_TRANS(vfncvt_x_f_w, vfncvt_x_f_w, RISCV_FRM_DYN)\n /* Reuse the helper functions from vfncvt.xu.f.w and vfncvt.x.f.w */\n-GEN_OPXFV_NARROW_TRANS(vfncvt_rtz_xu_f_w, vfncvt_xu_f_w, RISCV_FRM_RTZ)\n-GEN_OPXFV_NARROW_TRANS(vfncvt_rtz_x_f_w, vfncvt_x_f_w, RISCV_FRM_RTZ)\n+GEN_OPXFV_NARROW_BFA_TRANS(vfncvt_rtz_xu_f_w, vfncvt_xu_f_w, RISCV_FRM_RTZ)\n+GEN_OPXFV_NARROW_BFA_TRANS(vfncvt_rtz_x_f_w, vfncvt_x_f_w, RISCV_FRM_RTZ)\n \n /*\n  *** Vector Reduction Operations\n@@ -3069,10 +3288,12 @@ GEN_OPIVV_WIDEN_TRANS(vwredsum_vs, reduction_widen_check)\n GEN_OPIVV_WIDEN_TRANS(vwredsumu_vs, reduction_widen_check)\n \n /* Vector Single-Width Floating-Point Reduction Instructions */\n-static bool freduction_check(DisasContext *s, arg_rmrr *a)\n+static bool freduction_check(DisasContext *s, arg_rmrr *a,\n+                             int8_t valid_bfa_vsew)\n {\n     return reduction_check(s, a) &&\n-           require_rvf(s);\n+           require_rvf(s) &&\n+           vext_check_altfmt(s, valid_bfa_vsew);\n }\n \n GEN_OPFVV_TRANS(vfredusum_vs, freduction_check)\n@@ -3081,11 +3302,13 @@ GEN_OPFVV_TRANS(vfredmax_vs, freduction_check)\n GEN_OPFVV_TRANS(vfredmin_vs, freduction_check)\n \n /* Vector Widening Floating-Point Reduction Instructions */\n-static bool freduction_widen_check(DisasContext *s, arg_rmrr *a)\n+static bool freduction_widen_check(DisasContext *s, arg_rmrr *a,\n+                                   int8_t valid_bfa_vsew)\n {\n     return reduction_widen_check(s, a) &&\n            require_rvf(s) &&\n-           require_scale_rvf(s);\n+           require_scale_rvf(s) &&\n+           vext_check_altfmt(s, valid_bfa_vsew);\n }\n \n GEN_OPFVV_WIDEN_TRANS(vfwredusum_vs, freduction_widen_check)\n@@ -3500,7 +3723,8 @@ static bool trans_vfmv_s_f(DisasContext *s, arg_vfmv_s_f *a)\n {\n     if (require_rvv(s) &&\n         require_rvf(s) &&\n-        vext_check_isa_ill(s)) {\n+        vext_check_isa_ill(s) &&\n+        vext_check_altfmt(s, MO_16)) {\n         gen_set_rm(s, RISCV_FRM_DYN);\n \n         /* The instructions ignore LMUL and vector register group. */\n@@ -3594,20 +3818,24 @@ GEN_OPIVX_VSLIDE1_TRANS(vslide1up_vx, slideup_check)\n GEN_OPIVX_VSLIDE1_TRANS(vslide1down_vx, slidedown_check)\n \n /* Vector Floating-Point Slide Instructions */\n-static bool fslideup_check(DisasContext *s, arg_rmrr *a)\n+static bool fslideup_check(DisasContext *s, arg_rmrr *a,\n+                           int8_t valid_bfa_vsew)\n {\n     return slideup_check(s, a) &&\n-           require_rvf(s);\n+           require_rvf(s) &&\n+           vext_check_altfmt(s, valid_bfa_vsew);\n }\n \n-static bool fslidedown_check(DisasContext *s, arg_rmrr *a)\n+static bool fslidedown_check(DisasContext *s, arg_rmrr *a,\n+                             int8_t valid_bfa_vsew)\n {\n     return slidedown_check(s, a) &&\n-           require_rvf(s);\n+           require_rvf(s) &&\n+           vext_check_altfmt(s, valid_bfa_vsew);\n }\n \n-GEN_OPFVF_TRANS(vfslide1up_vf, fslideup_check)\n-GEN_OPFVF_TRANS(vfslide1down_vf, fslidedown_check)\n+GEN_OPFVF_BFA_TRANS(vfslide1up_vf, fslideup_check, vfslide1up_vf_h)\n+GEN_OPFVF_BFA_TRANS(vfslide1down_vf, fslidedown_check, vfslide1down_vf_h)\n \n /* Vector Register Gather Instruction */\n static bool vrgather_vv_check(DisasContext *s, arg_rmrr *a)\ndiff --git a/target/riscv/internals.h b/target/riscv/internals.h\nindex 460346dd6d..b001cbc080 100644\n--- a/target/riscv/internals.h\n+++ b/target/riscv/internals.h\n@@ -84,6 +84,7 @@ FIELD(VDATA, NF, 7, 4)\n FIELD(VDATA, WD, 7, 1)\n \n /* float point classify helpers */\n+target_ulong fclass_h_bf16(uint64_t frs1);\n target_ulong fclass_h(uint64_t frs1);\n target_ulong fclass_s(uint64_t frs1);\n target_ulong fclass_d(uint64_t frs1);\ndiff --git a/target/riscv/vector_helper.c b/target/riscv/vector_helper.c\nindex 1e0cce5ae5..538168efc9 100644\n--- a/target/riscv/vector_helper.c\n+++ b/target/riscv/vector_helper.c\n@@ -3168,9 +3168,11 @@ void HELPER(NAME)(void *vd, void *v0, void *vs1,          \\\n                       total_elems * ESZ);                 \\\n }\n \n+RVVCALL(OPFVV2, vfadd_vv_h_bf16, OP_UUU_H, H2, H2, H2, bfloat16_add)\n RVVCALL(OPFVV2, vfadd_vv_h, OP_UUU_H, H2, H2, H2, float16_add)\n RVVCALL(OPFVV2, vfadd_vv_w, OP_UUU_W, H4, H4, H4, float32_add)\n RVVCALL(OPFVV2, vfadd_vv_d, OP_UUU_D, H8, H8, H8, float64_add)\n+GEN_VEXT_VV_ENV(vfadd_vv_h_bf16, 2)\n GEN_VEXT_VV_ENV(vfadd_vv_h, 2)\n GEN_VEXT_VV_ENV(vfadd_vv_w, 4)\n GEN_VEXT_VV_ENV(vfadd_vv_d, 8)\n@@ -3213,26 +3215,37 @@ void HELPER(NAME)(void *vd, void *v0, uint64_t s1,        \\\n                       total_elems * ESZ);                 \\\n }\n \n+RVVCALL(OPFVF2, vfadd_vf_h_bf16, OP_UUU_H, H2, H2, bfloat16_add)\n RVVCALL(OPFVF2, vfadd_vf_h, OP_UUU_H, H2, H2, float16_add)\n RVVCALL(OPFVF2, vfadd_vf_w, OP_UUU_W, H4, H4, float32_add)\n RVVCALL(OPFVF2, vfadd_vf_d, OP_UUU_D, H8, H8, float64_add)\n+GEN_VEXT_VF(vfadd_vf_h_bf16, 2)\n GEN_VEXT_VF(vfadd_vf_h, 2)\n GEN_VEXT_VF(vfadd_vf_w, 4)\n GEN_VEXT_VF(vfadd_vf_d, 8)\n \n+RVVCALL(OPFVV2, vfsub_vv_h_bf16, OP_UUU_H, H2, H2, H2, bfloat16_sub)\n RVVCALL(OPFVV2, vfsub_vv_h, OP_UUU_H, H2, H2, H2, float16_sub)\n RVVCALL(OPFVV2, vfsub_vv_w, OP_UUU_W, H4, H4, H4, float32_sub)\n RVVCALL(OPFVV2, vfsub_vv_d, OP_UUU_D, H8, H8, H8, float64_sub)\n+GEN_VEXT_VV_ENV(vfsub_vv_h_bf16, 2)\n GEN_VEXT_VV_ENV(vfsub_vv_h, 2)\n GEN_VEXT_VV_ENV(vfsub_vv_w, 4)\n GEN_VEXT_VV_ENV(vfsub_vv_d, 8)\n+RVVCALL(OPFVF2, vfsub_vf_h_bf16, OP_UUU_H, H2, H2, bfloat16_sub)\n RVVCALL(OPFVF2, vfsub_vf_h, OP_UUU_H, H2, H2, float16_sub)\n RVVCALL(OPFVF2, vfsub_vf_w, OP_UUU_W, H4, H4, float32_sub)\n RVVCALL(OPFVF2, vfsub_vf_d, OP_UUU_D, H8, H8, float64_sub)\n+GEN_VEXT_VF(vfsub_vf_h_bf16, 2)\n GEN_VEXT_VF(vfsub_vf_h, 2)\n GEN_VEXT_VF(vfsub_vf_w, 4)\n GEN_VEXT_VF(vfsub_vf_d, 8)\n \n+static uint16_t bfloat16_rsub(uint16_t a, uint16_t b, float_status * s)\n+{\n+    return bfloat16_sub(b, a, s);\n+}\n+\n static uint16_t float16_rsub(uint16_t a, uint16_t b, float_status *s)\n {\n     return float16_sub(b, a, s);\n@@ -3248,14 +3261,22 @@ static uint64_t float64_rsub(uint64_t a, uint64_t b, float_status *s)\n     return float64_sub(b, a, s);\n }\n \n+RVVCALL(OPFVF2, vfrsub_vf_h_bf16, OP_UUU_H, H2, H2, bfloat16_rsub)\n RVVCALL(OPFVF2, vfrsub_vf_h, OP_UUU_H, H2, H2, float16_rsub)\n RVVCALL(OPFVF2, vfrsub_vf_w, OP_UUU_W, H4, H4, float32_rsub)\n RVVCALL(OPFVF2, vfrsub_vf_d, OP_UUU_D, H8, H8, float64_rsub)\n+GEN_VEXT_VF(vfrsub_vf_h_bf16, 2)\n GEN_VEXT_VF(vfrsub_vf_h, 2)\n GEN_VEXT_VF(vfrsub_vf_w, 4)\n GEN_VEXT_VF(vfrsub_vf_d, 8)\n \n /* Vector Widening Floating-Point Add/Subtract Instructions */\n+static uint32_t vfwadd16_bf16(uint16_t a, uint16_t b, float_status *s)\n+{\n+    return float32_add(bfloat16_to_float32(a, s),\n+                       bfloat16_to_float32(b, s), s);\n+}\n+\n static uint32_t vfwadd16(uint16_t a, uint16_t b, float_status *s)\n {\n     return float32_add(float16_to_float32(a, true, s),\n@@ -3269,15 +3290,25 @@ static uint64_t vfwadd32(uint32_t a, uint32_t b, float_status *s)\n \n }\n \n+RVVCALL(OPFVV2, vfwadd_vv_h_bf16, WOP_UUU_H, H4, H2, H2, vfwadd16_bf16)\n RVVCALL(OPFVV2, vfwadd_vv_h, WOP_UUU_H, H4, H2, H2, vfwadd16)\n RVVCALL(OPFVV2, vfwadd_vv_w, WOP_UUU_W, H8, H4, H4, vfwadd32)\n+GEN_VEXT_VV_ENV(vfwadd_vv_h_bf16, 4)\n GEN_VEXT_VV_ENV(vfwadd_vv_h, 4)\n GEN_VEXT_VV_ENV(vfwadd_vv_w, 8)\n+RVVCALL(OPFVF2, vfwadd_vf_h_bf16, WOP_UUU_H, H4, H2, vfwadd16_bf16)\n RVVCALL(OPFVF2, vfwadd_vf_h, WOP_UUU_H, H4, H2, vfwadd16)\n RVVCALL(OPFVF2, vfwadd_vf_w, WOP_UUU_W, H8, H4, vfwadd32)\n+GEN_VEXT_VF(vfwadd_vf_h_bf16, 4)\n GEN_VEXT_VF(vfwadd_vf_h, 4)\n GEN_VEXT_VF(vfwadd_vf_w, 8)\n \n+static uint32_t vfwsub16_bf16(uint16_t a, uint16_t b, float_status *s)\n+{\n+    return float32_sub(bfloat16_to_float32(a, s),\n+                       bfloat16_to_float32(b, s), s);\n+}\n+\n static uint32_t vfwsub16(uint16_t a, uint16_t b, float_status *s)\n {\n     return float32_sub(float16_to_float32(a, true, s),\n@@ -3291,15 +3322,24 @@ static uint64_t vfwsub32(uint32_t a, uint32_t b, float_status *s)\n \n }\n \n+RVVCALL(OPFVV2, vfwsub_vv_h_bf16, WOP_UUU_H, H4, H2, H2, vfwsub16_bf16)\n RVVCALL(OPFVV2, vfwsub_vv_h, WOP_UUU_H, H4, H2, H2, vfwsub16)\n RVVCALL(OPFVV2, vfwsub_vv_w, WOP_UUU_W, H8, H4, H4, vfwsub32)\n+GEN_VEXT_VV_ENV(vfwsub_vv_h_bf16, 4)\n GEN_VEXT_VV_ENV(vfwsub_vv_h, 4)\n GEN_VEXT_VV_ENV(vfwsub_vv_w, 8)\n+RVVCALL(OPFVF2, vfwsub_vf_h_bf16, WOP_UUU_H, H4, H2, vfwsub16_bf16)\n RVVCALL(OPFVF2, vfwsub_vf_h, WOP_UUU_H, H4, H2, vfwsub16)\n RVVCALL(OPFVF2, vfwsub_vf_w, WOP_UUU_W, H8, H4, vfwsub32)\n+GEN_VEXT_VF(vfwsub_vf_h_bf16, 4)\n GEN_VEXT_VF(vfwsub_vf_h, 4)\n GEN_VEXT_VF(vfwsub_vf_w, 8)\n \n+static uint32_t vfwaddw16_bf16(uint32_t a, uint16_t b, float_status *s)\n+{\n+    return float32_add(a, bfloat16_to_float32(b, s), s);\n+}\n+\n static uint32_t vfwaddw16(uint32_t a, uint16_t b, float_status *s)\n {\n     return float32_add(a, float16_to_float32(b, true, s), s);\n@@ -3310,15 +3350,24 @@ static uint64_t vfwaddw32(uint64_t a, uint32_t b, float_status *s)\n     return float64_add(a, float32_to_float64(b, s), s);\n }\n \n+RVVCALL(OPFVV2, vfwadd_wv_h_bf16, WOP_WUUU_H, H4, H2, H2, vfwaddw16_bf16)\n RVVCALL(OPFVV2, vfwadd_wv_h, WOP_WUUU_H, H4, H2, H2, vfwaddw16)\n RVVCALL(OPFVV2, vfwadd_wv_w, WOP_WUUU_W, H8, H4, H4, vfwaddw32)\n+GEN_VEXT_VV_ENV(vfwadd_wv_h_bf16, 4)\n GEN_VEXT_VV_ENV(vfwadd_wv_h, 4)\n GEN_VEXT_VV_ENV(vfwadd_wv_w, 8)\n+RVVCALL(OPFVF2, vfwadd_wf_h_bf16, WOP_WUUU_H, H4, H2, vfwaddw16_bf16)\n RVVCALL(OPFVF2, vfwadd_wf_h, WOP_WUUU_H, H4, H2, vfwaddw16)\n RVVCALL(OPFVF2, vfwadd_wf_w, WOP_WUUU_W, H8, H4, vfwaddw32)\n+GEN_VEXT_VF(vfwadd_wf_h_bf16, 4)\n GEN_VEXT_VF(vfwadd_wf_h, 4)\n GEN_VEXT_VF(vfwadd_wf_w, 8)\n \n+static uint32_t vfwsubw16_bf16(uint32_t a, uint16_t b, float_status *s)\n+{\n+    return float32_sub(a, bfloat16_to_float32(b, s), s);\n+}\n+\n static uint32_t vfwsubw16(uint32_t a, uint16_t b, float_status *s)\n {\n     return float32_sub(a, float16_to_float32(b, true, s), s);\n@@ -3329,25 +3378,33 @@ static uint64_t vfwsubw32(uint64_t a, uint32_t b, float_status *s)\n     return float64_sub(a, float32_to_float64(b, s), s);\n }\n \n+RVVCALL(OPFVV2, vfwsub_wv_h_bf16, WOP_WUUU_H, H4, H2, H2, vfwsubw16_bf16)\n RVVCALL(OPFVV2, vfwsub_wv_h, WOP_WUUU_H, H4, H2, H2, vfwsubw16)\n RVVCALL(OPFVV2, vfwsub_wv_w, WOP_WUUU_W, H8, H4, H4, vfwsubw32)\n+GEN_VEXT_VV_ENV(vfwsub_wv_h_bf16, 4)\n GEN_VEXT_VV_ENV(vfwsub_wv_h, 4)\n GEN_VEXT_VV_ENV(vfwsub_wv_w, 8)\n+RVVCALL(OPFVF2, vfwsub_wf_h_bf16, WOP_WUUU_H, H4, H2, vfwsubw16_bf16)\n RVVCALL(OPFVF2, vfwsub_wf_h, WOP_WUUU_H, H4, H2, vfwsubw16)\n RVVCALL(OPFVF2, vfwsub_wf_w, WOP_WUUU_W, H8, H4, vfwsubw32)\n+GEN_VEXT_VF(vfwsub_wf_h_bf16, 4)\n GEN_VEXT_VF(vfwsub_wf_h, 4)\n GEN_VEXT_VF(vfwsub_wf_w, 8)\n \n /* Vector Single-Width Floating-Point Multiply/Divide Instructions */\n+RVVCALL(OPFVV2, vfmul_vv_h_bf16, OP_UUU_H, H2, H2, H2, bfloat16_mul)\n RVVCALL(OPFVV2, vfmul_vv_h, OP_UUU_H, H2, H2, H2, float16_mul)\n RVVCALL(OPFVV2, vfmul_vv_w, OP_UUU_W, H4, H4, H4, float32_mul)\n RVVCALL(OPFVV2, vfmul_vv_d, OP_UUU_D, H8, H8, H8, float64_mul)\n+GEN_VEXT_VV_ENV(vfmul_vv_h_bf16, 2)\n GEN_VEXT_VV_ENV(vfmul_vv_h, 2)\n GEN_VEXT_VV_ENV(vfmul_vv_w, 4)\n GEN_VEXT_VV_ENV(vfmul_vv_d, 8)\n+RVVCALL(OPFVF2, vfmul_vf_h_bf16, OP_UUU_H, H2, H2, bfloat16_mul)\n RVVCALL(OPFVF2, vfmul_vf_h, OP_UUU_H, H2, H2, float16_mul)\n RVVCALL(OPFVF2, vfmul_vf_w, OP_UUU_W, H4, H4, float32_mul)\n RVVCALL(OPFVF2, vfmul_vf_d, OP_UUU_D, H8, H8, float64_mul)\n+GEN_VEXT_VF(vfmul_vf_h_bf16, 2)\n GEN_VEXT_VF(vfmul_vf_h, 2)\n GEN_VEXT_VF(vfmul_vf_w, 4)\n GEN_VEXT_VF(vfmul_vf_d, 8)\n@@ -3388,6 +3445,12 @@ GEN_VEXT_VF(vfrdiv_vf_w, 4)\n GEN_VEXT_VF(vfrdiv_vf_d, 8)\n \n /* Vector Widening Floating-Point Multiply */\n+static uint32_t vfwmul16_bf16(uint16_t a, uint16_t b, float_status *s)\n+{\n+    return float32_mul(bfloat16_to_float32(a, s),\n+                       bfloat16_to_float32(b, s), s);\n+}\n+\n static uint32_t vfwmul16(uint16_t a, uint16_t b, float_status *s)\n {\n     return float32_mul(float16_to_float32(a, true, s),\n@@ -3400,12 +3463,17 @@ static uint64_t vfwmul32(uint32_t a, uint32_t b, float_status *s)\n                        float32_to_float64(b, s), s);\n \n }\n+\n+RVVCALL(OPFVV2, vfwmul_vv_h_bf16, WOP_UUU_H, H4, H2, H2, vfwmul16_bf16)\n RVVCALL(OPFVV2, vfwmul_vv_h, WOP_UUU_H, H4, H2, H2, vfwmul16)\n RVVCALL(OPFVV2, vfwmul_vv_w, WOP_UUU_W, H8, H4, H4, vfwmul32)\n+GEN_VEXT_VV_ENV(vfwmul_vv_h_bf16, 4)\n GEN_VEXT_VV_ENV(vfwmul_vv_h, 4)\n GEN_VEXT_VV_ENV(vfwmul_vv_w, 8)\n+RVVCALL(OPFVF2, vfwmul_vf_h_bf16, WOP_UUU_H, H4, H2, vfwmul16_bf16)\n RVVCALL(OPFVF2, vfwmul_vf_h, WOP_UUU_H, H4, H2, vfwmul16)\n RVVCALL(OPFVF2, vfwmul_vf_w, WOP_UUU_W, H8, H4, vfwmul32)\n+GEN_VEXT_VF(vfwmul_vf_h_bf16, 4)\n GEN_VEXT_VF(vfwmul_vf_h, 4)\n GEN_VEXT_VF(vfwmul_vf_w, 8)\n \n@@ -3420,6 +3488,12 @@ static void do_##NAME(void *vd, void *vs1, void *vs2, int i,       \\\n     *((TD *)vd + HD(i)) = OP(s2, s1, d, &env->fp_status);          \\\n }\n \n+static uint16_t fmacc16_bf16(uint16_t a, uint16_t b, uint16_t d,\n+                             float_status *s)\n+{\n+    return bfloat16_muladd(a, b, d, 0, s);\n+}\n+\n static uint16_t fmacc16(uint16_t a, uint16_t b, uint16_t d, float_status *s)\n {\n     return float16_muladd(a, b, d, 0, s);\n@@ -3435,9 +3509,11 @@ static uint64_t fmacc64(uint64_t a, uint64_t b, uint64_t d, float_status *s)\n     return float64_muladd(a, b, d, 0, s);\n }\n \n+RVVCALL(OPFVV3, vfmacc_vv_h_bf16, OP_UUU_H, H2, H2, H2, fmacc16_bf16)\n RVVCALL(OPFVV3, vfmacc_vv_h, OP_UUU_H, H2, H2, H2, fmacc16)\n RVVCALL(OPFVV3, vfmacc_vv_w, OP_UUU_W, H4, H4, H4, fmacc32)\n RVVCALL(OPFVV3, vfmacc_vv_d, OP_UUU_D, H8, H8, H8, fmacc64)\n+GEN_VEXT_VV_ENV(vfmacc_vv_h_bf16, 2)\n GEN_VEXT_VV_ENV(vfmacc_vv_h, 2)\n GEN_VEXT_VV_ENV(vfmacc_vv_w, 4)\n GEN_VEXT_VV_ENV(vfmacc_vv_d, 8)\n@@ -3451,13 +3527,22 @@ static void do_##NAME(void *vd, uint64_t s1, void *vs2, int i,    \\\n     *((TD *)vd + HD(i)) = OP(s2, (TX1)(T1)s1, d, &env->fp_status);\\\n }\n \n+RVVCALL(OPFVF3, vfmacc_vf_h_bf16, OP_UUU_H, H2, H2, fmacc16_bf16)\n RVVCALL(OPFVF3, vfmacc_vf_h, OP_UUU_H, H2, H2, fmacc16)\n RVVCALL(OPFVF3, vfmacc_vf_w, OP_UUU_W, H4, H4, fmacc32)\n RVVCALL(OPFVF3, vfmacc_vf_d, OP_UUU_D, H8, H8, fmacc64)\n+GEN_VEXT_VF(vfmacc_vf_h_bf16, 2)\n GEN_VEXT_VF(vfmacc_vf_h, 2)\n GEN_VEXT_VF(vfmacc_vf_w, 4)\n GEN_VEXT_VF(vfmacc_vf_d, 8)\n \n+static uint16_t fnmacc16_bf16(uint16_t a, uint16_t b, uint16_t d,\n+                              float_status *s)\n+{\n+    return bfloat16_muladd(a, b, d, float_muladd_negate_c |\n+                                    float_muladd_negate_product, s);\n+}\n+\n static uint16_t fnmacc16(uint16_t a, uint16_t b, uint16_t d, float_status *s)\n {\n     return float16_muladd(a, b, d, float_muladd_negate_c |\n@@ -3476,19 +3561,29 @@ static uint64_t fnmacc64(uint64_t a, uint64_t b, uint64_t d, float_status *s)\n                                    float_muladd_negate_product, s);\n }\n \n+RVVCALL(OPFVV3, vfnmacc_vv_h_bf16, OP_UUU_H, H2, H2, H2, fnmacc16_bf16)\n RVVCALL(OPFVV3, vfnmacc_vv_h, OP_UUU_H, H2, H2, H2, fnmacc16)\n RVVCALL(OPFVV3, vfnmacc_vv_w, OP_UUU_W, H4, H4, H4, fnmacc32)\n RVVCALL(OPFVV3, vfnmacc_vv_d, OP_UUU_D, H8, H8, H8, fnmacc64)\n+GEN_VEXT_VV_ENV(vfnmacc_vv_h_bf16, 2)\n GEN_VEXT_VV_ENV(vfnmacc_vv_h, 2)\n GEN_VEXT_VV_ENV(vfnmacc_vv_w, 4)\n GEN_VEXT_VV_ENV(vfnmacc_vv_d, 8)\n+RVVCALL(OPFVF3, vfnmacc_vf_h_bf16, OP_UUU_H, H2, H2, fnmacc16_bf16)\n RVVCALL(OPFVF3, vfnmacc_vf_h, OP_UUU_H, H2, H2, fnmacc16)\n RVVCALL(OPFVF3, vfnmacc_vf_w, OP_UUU_W, H4, H4, fnmacc32)\n RVVCALL(OPFVF3, vfnmacc_vf_d, OP_UUU_D, H8, H8, fnmacc64)\n+GEN_VEXT_VF(vfnmacc_vf_h_bf16, 2)\n GEN_VEXT_VF(vfnmacc_vf_h, 2)\n GEN_VEXT_VF(vfnmacc_vf_w, 4)\n GEN_VEXT_VF(vfnmacc_vf_d, 8)\n \n+static uint16_t fmsac16_bf16(uint16_t a, uint16_t b, uint16_t d,\n+                             float_status *s)\n+{\n+    return bfloat16_muladd(a, b, d, float_muladd_negate_c, s);\n+}\n+\n static uint16_t fmsac16(uint16_t a, uint16_t b, uint16_t d, float_status *s)\n {\n     return float16_muladd(a, b, d, float_muladd_negate_c, s);\n@@ -3504,19 +3599,29 @@ static uint64_t fmsac64(uint64_t a, uint64_t b, uint64_t d, float_status *s)\n     return float64_muladd(a, b, d, float_muladd_negate_c, s);\n }\n \n+RVVCALL(OPFVV3, vfmsac_vv_h_bf16, OP_UUU_H, H2, H2, H2, fmsac16_bf16)\n RVVCALL(OPFVV3, vfmsac_vv_h, OP_UUU_H, H2, H2, H2, fmsac16)\n RVVCALL(OPFVV3, vfmsac_vv_w, OP_UUU_W, H4, H4, H4, fmsac32)\n RVVCALL(OPFVV3, vfmsac_vv_d, OP_UUU_D, H8, H8, H8, fmsac64)\n+GEN_VEXT_VV_ENV(vfmsac_vv_h_bf16, 2)\n GEN_VEXT_VV_ENV(vfmsac_vv_h, 2)\n GEN_VEXT_VV_ENV(vfmsac_vv_w, 4)\n GEN_VEXT_VV_ENV(vfmsac_vv_d, 8)\n+RVVCALL(OPFVF3, vfmsac_vf_h_bf16, OP_UUU_H, H2, H2, fmsac16_bf16)\n RVVCALL(OPFVF3, vfmsac_vf_h, OP_UUU_H, H2, H2, fmsac16)\n RVVCALL(OPFVF3, vfmsac_vf_w, OP_UUU_W, H4, H4, fmsac32)\n RVVCALL(OPFVF3, vfmsac_vf_d, OP_UUU_D, H8, H8, fmsac64)\n+GEN_VEXT_VF(vfmsac_vf_h_bf16, 2)\n GEN_VEXT_VF(vfmsac_vf_h, 2)\n GEN_VEXT_VF(vfmsac_vf_w, 4)\n GEN_VEXT_VF(vfmsac_vf_d, 8)\n \n+static uint16_t fnmsac16_bf16(uint16_t a, uint16_t b, uint16_t d,\n+                              float_status *s)\n+{\n+    return bfloat16_muladd(a, b, d, float_muladd_negate_product, s);\n+}\n+\n static uint16_t fnmsac16(uint16_t a, uint16_t b, uint16_t d, float_status *s)\n {\n     return float16_muladd(a, b, d, float_muladd_negate_product, s);\n@@ -3532,19 +3637,29 @@ static uint64_t fnmsac64(uint64_t a, uint64_t b, uint64_t d, float_status *s)\n     return float64_muladd(a, b, d, float_muladd_negate_product, s);\n }\n \n+RVVCALL(OPFVV3, vfnmsac_vv_h_bf16, OP_UUU_H, H2, H2, H2, fnmsac16_bf16)\n RVVCALL(OPFVV3, vfnmsac_vv_h, OP_UUU_H, H2, H2, H2, fnmsac16)\n RVVCALL(OPFVV3, vfnmsac_vv_w, OP_UUU_W, H4, H4, H4, fnmsac32)\n RVVCALL(OPFVV3, vfnmsac_vv_d, OP_UUU_D, H8, H8, H8, fnmsac64)\n+GEN_VEXT_VV_ENV(vfnmsac_vv_h_bf16, 2)\n GEN_VEXT_VV_ENV(vfnmsac_vv_h, 2)\n GEN_VEXT_VV_ENV(vfnmsac_vv_w, 4)\n GEN_VEXT_VV_ENV(vfnmsac_vv_d, 8)\n+RVVCALL(OPFVF3, vfnmsac_vf_h_bf16, OP_UUU_H, H2, H2, fnmsac16_bf16)\n RVVCALL(OPFVF3, vfnmsac_vf_h, OP_UUU_H, H2, H2, fnmsac16)\n RVVCALL(OPFVF3, vfnmsac_vf_w, OP_UUU_W, H4, H4, fnmsac32)\n RVVCALL(OPFVF3, vfnmsac_vf_d, OP_UUU_D, H8, H8, fnmsac64)\n+GEN_VEXT_VF(vfnmsac_vf_h_bf16, 2)\n GEN_VEXT_VF(vfnmsac_vf_h, 2)\n GEN_VEXT_VF(vfnmsac_vf_w, 4)\n GEN_VEXT_VF(vfnmsac_vf_d, 8)\n \n+static uint16_t fmadd16_bf16(uint16_t a, uint16_t b, uint16_t d,\n+                             float_status *s)\n+{\n+    return bfloat16_muladd(d, b, a, 0, s);\n+}\n+\n static uint16_t fmadd16(uint16_t a, uint16_t b, uint16_t d, float_status *s)\n {\n     return float16_muladd(d, b, a, 0, s);\n@@ -3560,19 +3675,30 @@ static uint64_t fmadd64(uint64_t a, uint64_t b, uint64_t d, float_status *s)\n     return float64_muladd(d, b, a, 0, s);\n }\n \n+RVVCALL(OPFVV3, vfmadd_vv_h_bf16, OP_UUU_H, H2, H2, H2, fmadd16_bf16)\n RVVCALL(OPFVV3, vfmadd_vv_h, OP_UUU_H, H2, H2, H2, fmadd16)\n RVVCALL(OPFVV3, vfmadd_vv_w, OP_UUU_W, H4, H4, H4, fmadd32)\n RVVCALL(OPFVV3, vfmadd_vv_d, OP_UUU_D, H8, H8, H8, fmadd64)\n+GEN_VEXT_VV_ENV(vfmadd_vv_h_bf16, 2)\n GEN_VEXT_VV_ENV(vfmadd_vv_h, 2)\n GEN_VEXT_VV_ENV(vfmadd_vv_w, 4)\n GEN_VEXT_VV_ENV(vfmadd_vv_d, 8)\n+RVVCALL(OPFVF3, vfmadd_vf_h_bf16, OP_UUU_H, H2, H2, fmadd16_bf16)\n RVVCALL(OPFVF3, vfmadd_vf_h, OP_UUU_H, H2, H2, fmadd16)\n RVVCALL(OPFVF3, vfmadd_vf_w, OP_UUU_W, H4, H4, fmadd32)\n RVVCALL(OPFVF3, vfmadd_vf_d, OP_UUU_D, H8, H8, fmadd64)\n+GEN_VEXT_VF(vfmadd_vf_h_bf16, 2)\n GEN_VEXT_VF(vfmadd_vf_h, 2)\n GEN_VEXT_VF(vfmadd_vf_w, 4)\n GEN_VEXT_VF(vfmadd_vf_d, 8)\n \n+static uint16_t fnmadd16_bf16(uint16_t a, uint16_t b, uint16_t d,\n+                              float_status *s)\n+{\n+    return bfloat16_muladd(d, b, a, float_muladd_negate_c |\n+                                    float_muladd_negate_product, s);\n+}\n+\n static uint16_t fnmadd16(uint16_t a, uint16_t b, uint16_t d, float_status *s)\n {\n     return float16_muladd(d, b, a, float_muladd_negate_c |\n@@ -3591,19 +3717,29 @@ static uint64_t fnmadd64(uint64_t a, uint64_t b, uint64_t d, float_status *s)\n                                    float_muladd_negate_product, s);\n }\n \n+RVVCALL(OPFVV3, vfnmadd_vv_h_bf16, OP_UUU_H, H2, H2, H2, fnmadd16_bf16)\n RVVCALL(OPFVV3, vfnmadd_vv_h, OP_UUU_H, H2, H2, H2, fnmadd16)\n RVVCALL(OPFVV3, vfnmadd_vv_w, OP_UUU_W, H4, H4, H4, fnmadd32)\n RVVCALL(OPFVV3, vfnmadd_vv_d, OP_UUU_D, H8, H8, H8, fnmadd64)\n+GEN_VEXT_VV_ENV(vfnmadd_vv_h_bf16, 2)\n GEN_VEXT_VV_ENV(vfnmadd_vv_h, 2)\n GEN_VEXT_VV_ENV(vfnmadd_vv_w, 4)\n GEN_VEXT_VV_ENV(vfnmadd_vv_d, 8)\n+RVVCALL(OPFVF3, vfnmadd_vf_h_bf16, OP_UUU_H, H2, H2, fnmadd16_bf16)\n RVVCALL(OPFVF3, vfnmadd_vf_h, OP_UUU_H, H2, H2, fnmadd16)\n RVVCALL(OPFVF3, vfnmadd_vf_w, OP_UUU_W, H4, H4, fnmadd32)\n RVVCALL(OPFVF3, vfnmadd_vf_d, OP_UUU_D, H8, H8, fnmadd64)\n+GEN_VEXT_VF(vfnmadd_vf_h_bf16, 2)\n GEN_VEXT_VF(vfnmadd_vf_h, 2)\n GEN_VEXT_VF(vfnmadd_vf_w, 4)\n GEN_VEXT_VF(vfnmadd_vf_d, 8)\n \n+static uint16_t fmsub16_bf16(uint16_t a, uint16_t b, uint16_t d,\n+                             float_status *s)\n+{\n+    return bfloat16_muladd(d, b, a, float_muladd_negate_c, s);\n+}\n+\n static uint16_t fmsub16(uint16_t a, uint16_t b, uint16_t d, float_status *s)\n {\n     return float16_muladd(d, b, a, float_muladd_negate_c, s);\n@@ -3619,19 +3755,29 @@ static uint64_t fmsub64(uint64_t a, uint64_t b, uint64_t d, float_status *s)\n     return float64_muladd(d, b, a, float_muladd_negate_c, s);\n }\n \n+RVVCALL(OPFVV3, vfmsub_vv_h_bf16, OP_UUU_H, H2, H2, H2, fmsub16_bf16)\n RVVCALL(OPFVV3, vfmsub_vv_h, OP_UUU_H, H2, H2, H2, fmsub16)\n RVVCALL(OPFVV3, vfmsub_vv_w, OP_UUU_W, H4, H4, H4, fmsub32)\n RVVCALL(OPFVV3, vfmsub_vv_d, OP_UUU_D, H8, H8, H8, fmsub64)\n+GEN_VEXT_VV_ENV(vfmsub_vv_h_bf16, 2)\n GEN_VEXT_VV_ENV(vfmsub_vv_h, 2)\n GEN_VEXT_VV_ENV(vfmsub_vv_w, 4)\n GEN_VEXT_VV_ENV(vfmsub_vv_d, 8)\n+RVVCALL(OPFVF3, vfmsub_vf_h_bf16, OP_UUU_H, H2, H2, fmsub16_bf16)\n RVVCALL(OPFVF3, vfmsub_vf_h, OP_UUU_H, H2, H2, fmsub16)\n RVVCALL(OPFVF3, vfmsub_vf_w, OP_UUU_W, H4, H4, fmsub32)\n RVVCALL(OPFVF3, vfmsub_vf_d, OP_UUU_D, H8, H8, fmsub64)\n+GEN_VEXT_VF(vfmsub_vf_h_bf16, 2)\n GEN_VEXT_VF(vfmsub_vf_h, 2)\n GEN_VEXT_VF(vfmsub_vf_w, 4)\n GEN_VEXT_VF(vfmsub_vf_d, 8)\n \n+static uint16_t fnmsub16_bf16(uint16_t a, uint16_t b, uint16_t d,\n+                              float_status *s)\n+{\n+    return bfloat16_muladd(d, b, a, float_muladd_negate_product, s);\n+}\n+\n static uint16_t fnmsub16(uint16_t a, uint16_t b, uint16_t d, float_status *s)\n {\n     return float16_muladd(d, b, a, float_muladd_negate_product, s);\n@@ -3647,15 +3793,19 @@ static uint64_t fnmsub64(uint64_t a, uint64_t b, uint64_t d, float_status *s)\n     return float64_muladd(d, b, a, float_muladd_negate_product, s);\n }\n \n+RVVCALL(OPFVV3, vfnmsub_vv_h_bf16, OP_UUU_H, H2, H2, H2, fnmsub16_bf16)\n RVVCALL(OPFVV3, vfnmsub_vv_h, OP_UUU_H, H2, H2, H2, fnmsub16)\n RVVCALL(OPFVV3, vfnmsub_vv_w, OP_UUU_W, H4, H4, H4, fnmsub32)\n RVVCALL(OPFVV3, vfnmsub_vv_d, OP_UUU_D, H8, H8, H8, fnmsub64)\n+GEN_VEXT_VV_ENV(vfnmsub_vv_h_bf16, 2)\n GEN_VEXT_VV_ENV(vfnmsub_vv_h, 2)\n GEN_VEXT_VV_ENV(vfnmsub_vv_w, 4)\n GEN_VEXT_VV_ENV(vfnmsub_vv_d, 8)\n+RVVCALL(OPFVF3, vfnmsub_vf_h_bf16, OP_UUU_H, H2, H2, fnmsub16_bf16)\n RVVCALL(OPFVF3, vfnmsub_vf_h, OP_UUU_H, H2, H2, fnmsub16)\n RVVCALL(OPFVF3, vfnmsub_vf_w, OP_UUU_W, H4, H4, fnmsub32)\n RVVCALL(OPFVF3, vfnmsub_vf_d, OP_UUU_D, H8, H8, fnmsub64)\n+GEN_VEXT_VF(vfnmsub_vf_h_bf16, 2)\n GEN_VEXT_VF(vfnmsub_vf_h, 2)\n GEN_VEXT_VF(vfnmsub_vf_w, 4)\n GEN_VEXT_VF(vfnmsub_vf_d, 8)\n@@ -3693,6 +3843,15 @@ GEN_VEXT_VV_ENV(vfwmaccbf16_vv, 4)\n RVVCALL(OPFVF3, vfwmaccbf16_vf, WOP_UUU_H, H4, H2, fwmaccbf16)\n GEN_VEXT_VF(vfwmaccbf16_vf, 4)\n \n+static uint32_t fwnmacc16_bf16(uint16_t a, uint16_t b, uint32_t d,\n+                               float_status *s)\n+{\n+    return float32_muladd(bfloat16_to_float32(a, s),\n+                          bfloat16_to_float32(b, s), d,\n+                          float_muladd_negate_c | float_muladd_negate_product,\n+                          s);\n+}\n+\n static uint32_t fwnmacc16(uint16_t a, uint16_t b, uint32_t d, float_status *s)\n {\n     return float32_muladd(float16_to_float32(a, true, s),\n@@ -3708,15 +3867,27 @@ static uint64_t fwnmacc32(uint32_t a, uint32_t b, uint64_t d, float_status *s)\n                              float_muladd_negate_product, s);\n }\n \n+RVVCALL(OPFVV3, vfwnmacc_vv_h_bf16, WOP_UUU_H, H4, H2, H2, fwnmacc16_bf16)\n RVVCALL(OPFVV3, vfwnmacc_vv_h, WOP_UUU_H, H4, H2, H2, fwnmacc16)\n RVVCALL(OPFVV3, vfwnmacc_vv_w, WOP_UUU_W, H8, H4, H4, fwnmacc32)\n+GEN_VEXT_VV_ENV(vfwnmacc_vv_h_bf16, 4)\n GEN_VEXT_VV_ENV(vfwnmacc_vv_h, 4)\n GEN_VEXT_VV_ENV(vfwnmacc_vv_w, 8)\n+RVVCALL(OPFVF3, vfwnmacc_vf_h_bf16, WOP_UUU_H, H4, H2, fwnmacc16_bf16)\n RVVCALL(OPFVF3, vfwnmacc_vf_h, WOP_UUU_H, H4, H2, fwnmacc16)\n RVVCALL(OPFVF3, vfwnmacc_vf_w, WOP_UUU_W, H8, H4, fwnmacc32)\n+GEN_VEXT_VF(vfwnmacc_vf_h_bf16, 4)\n GEN_VEXT_VF(vfwnmacc_vf_h, 4)\n GEN_VEXT_VF(vfwnmacc_vf_w, 8)\n \n+static uint32_t fwmsac16_bf16(uint16_t a, uint16_t b, uint32_t d,\n+                              float_status *s)\n+{\n+    return float32_muladd(bfloat16_to_float32(a, s),\n+                          bfloat16_to_float32(b, s), d,\n+                          float_muladd_negate_c, s);\n+}\n+\n static uint32_t fwmsac16(uint16_t a, uint16_t b, uint32_t d, float_status *s)\n {\n     return float32_muladd(float16_to_float32(a, true, s),\n@@ -3731,15 +3902,27 @@ static uint64_t fwmsac32(uint32_t a, uint32_t b, uint64_t d, float_status *s)\n                           float_muladd_negate_c, s);\n }\n \n+RVVCALL(OPFVV3, vfwmsac_vv_h_bf16, WOP_UUU_H, H4, H2, H2, fwmsac16_bf16)\n RVVCALL(OPFVV3, vfwmsac_vv_h, WOP_UUU_H, H4, H2, H2, fwmsac16)\n RVVCALL(OPFVV3, vfwmsac_vv_w, WOP_UUU_W, H8, H4, H4, fwmsac32)\n+GEN_VEXT_VV_ENV(vfwmsac_vv_h_bf16, 4)\n GEN_VEXT_VV_ENV(vfwmsac_vv_h, 4)\n GEN_VEXT_VV_ENV(vfwmsac_vv_w, 8)\n+RVVCALL(OPFVF3, vfwmsac_vf_h_bf16, WOP_UUU_H, H4, H2, fwmsac16_bf16)\n RVVCALL(OPFVF3, vfwmsac_vf_h, WOP_UUU_H, H4, H2, fwmsac16)\n RVVCALL(OPFVF3, vfwmsac_vf_w, WOP_UUU_W, H8, H4, fwmsac32)\n+GEN_VEXT_VF(vfwmsac_vf_h_bf16, 4)\n GEN_VEXT_VF(vfwmsac_vf_h, 4)\n GEN_VEXT_VF(vfwmsac_vf_w, 8)\n \n+static uint32_t fwnmsac16_bf16(uint16_t a, uint16_t b, uint32_t d,\n+                               float_status *s)\n+{\n+    return float32_muladd(bfloat16_to_float32(a, s),\n+                          bfloat16_to_float32(b, s), d,\n+                          float_muladd_negate_product, s);\n+}\n+\n static uint32_t fwnmsac16(uint16_t a, uint16_t b, uint32_t d, float_status *s)\n {\n     return float32_muladd(float16_to_float32(a, true, s),\n@@ -3754,12 +3937,16 @@ static uint64_t fwnmsac32(uint32_t a, uint32_t b, uint64_t d, float_status *s)\n                           float_muladd_negate_product, s);\n }\n \n+RVVCALL(OPFVV3, vfwnmsac_vv_h_bf16, WOP_UUU_H, H4, H2, H2, fwnmsac16_bf16)\n RVVCALL(OPFVV3, vfwnmsac_vv_h, WOP_UUU_H, H4, H2, H2, fwnmsac16)\n RVVCALL(OPFVV3, vfwnmsac_vv_w, WOP_UUU_W, H8, H4, H4, fwnmsac32)\n+GEN_VEXT_VV_ENV(vfwnmsac_vv_h_bf16, 4)\n GEN_VEXT_VV_ENV(vfwnmsac_vv_h, 4)\n GEN_VEXT_VV_ENV(vfwnmsac_vv_w, 8)\n+RVVCALL(OPFVF3, vfwnmsac_vf_h_bf16, WOP_UUU_H, H4, H2, fwnmsac16_bf16)\n RVVCALL(OPFVF3, vfwnmsac_vf_h, WOP_UUU_H, H4, H2, fwnmsac16)\n RVVCALL(OPFVF3, vfwnmsac_vf_w, WOP_UUU_W, H8, H4, fwnmsac32)\n+GEN_VEXT_VF(vfwnmsac_vf_h_bf16, 4)\n GEN_VEXT_VF(vfwnmsac_vf_h, 4)\n GEN_VEXT_VF(vfwnmsac_vf_w, 8)\n \n@@ -3865,6 +4052,46 @@ static uint64_t frsqrt7(uint64_t f, int exp_size, int frac_size)\n     return val;\n }\n \n+static bfloat16 frsqrt7_h_bf16(bfloat16 f, float_status *s)\n+{\n+    int exp_size = 8, frac_size = 7;\n+    bool sign = bfloat16_is_neg(f);\n+\n+    /*\n+     * frsqrt7(sNaN) = canonical NaN\n+     * frsqrt7(-inf) = canonical NaN\n+     * frsqrt7(-normal) = canonical NaN\n+     * frsqrt7(-subnormal) = canonical NaN\n+     */\n+    if (bfloat16_is_signaling_nan(f, s) ||\n+        (bfloat16_is_infinity(f) && sign) ||\n+        (bfloat16_is_normal(f) && sign) ||\n+        (bfloat16_is_zero_or_denormal(f) && !bfloat16_is_zero(f) && sign)) {\n+        s->float_exception_flags |= float_flag_invalid;\n+        return bfloat16_default_nan(s);\n+    }\n+\n+    /* frsqrt7(qNaN) = canonical NaN */\n+    if (bfloat16_is_quiet_nan(f, s)) {\n+        return bfloat16_default_nan(s);\n+    }\n+\n+    /* frsqrt7(+-0) = +-inf */\n+    if (bfloat16_is_zero(f)) {\n+        s->float_exception_flags |= float_flag_divbyzero;\n+        return bfloat16_set_sign(bfloat16_infinity, sign);\n+    }\n+\n+    /* frsqrt7(+inf) = +0 */\n+    if (bfloat16_is_infinity(f) && !sign) {\n+        return bfloat16_set_sign(bfloat16_zero, sign);\n+    }\n+\n+    /* +normal, +subnormal */\n+    uint64_t val = frsqrt7(f, exp_size, frac_size);\n+    return make_float16(val);\n+}\n+\n static float16 frsqrt7_h(float16 f, float_status *s)\n {\n     int exp_size = 5, frac_size = 10;\n@@ -3985,9 +4212,11 @@ static float64 frsqrt7_d(float64 f, float_status *s)\n     return make_float64(val);\n }\n \n+RVVCALL(OPFVV1, vfrsqrt7_v_h_bf16, OP_UU_H, H2, H2, frsqrt7_h_bf16)\n RVVCALL(OPFVV1, vfrsqrt7_v_h, OP_UU_H, H2, H2, frsqrt7_h)\n RVVCALL(OPFVV1, vfrsqrt7_v_w, OP_UU_W, H4, H4, frsqrt7_s)\n RVVCALL(OPFVV1, vfrsqrt7_v_d, OP_UU_D, H8, H8, frsqrt7_d)\n+GEN_VEXT_V_ENV(vfrsqrt7_v_h_bf16, 2)\n GEN_VEXT_V_ENV(vfrsqrt7_v_h, 2)\n GEN_VEXT_V_ENV(vfrsqrt7_v_w, 4)\n GEN_VEXT_V_ENV(vfrsqrt7_v_d, 8)\n@@ -4080,6 +4309,38 @@ static uint64_t frec7(uint64_t f, int exp_size, int frac_size,\n     return val;\n }\n \n+static bfloat16 frec7_h_bf16(bfloat16 f, float_status *s)\n+{\n+    int exp_size = 8, frac_size = 7;\n+    bool sign = bfloat16_is_neg(f);\n+\n+    /* frec7(+-inf) = +-0 */\n+    if (bfloat16_is_infinity(f)) {\n+        return bfloat16_set_sign(bfloat16_zero, sign);\n+    }\n+\n+    /* frec7(+-0) = +-inf */\n+    if (bfloat16_is_zero(f)) {\n+        s->float_exception_flags |= float_flag_divbyzero;\n+        return bfloat16_set_sign(bfloat16_infinity, sign);\n+    }\n+\n+    /* frec7(sNaN) = canonical NaN */\n+    if (bfloat16_is_signaling_nan(f, s)) {\n+        s->float_exception_flags |= float_flag_invalid;\n+        return bfloat16_default_nan(s);\n+    }\n+\n+    /* frec7(qNaN) = canonical NaN */\n+    if (bfloat16_is_quiet_nan(f, s)) {\n+        return bfloat16_default_nan(s);\n+    }\n+\n+    /* +-normal, +-subnormal */\n+    uint64_t val = frec7(f, exp_size, frac_size, s);\n+    return make_float16(val);\n+}\n+\n static float16 frec7_h(float16 f, float_status *s)\n {\n     int exp_size = 5, frac_size = 10;\n@@ -4176,36 +4437,46 @@ static float64 frec7_d(float64 f, float_status *s)\n     return make_float64(val);\n }\n \n+RVVCALL(OPFVV1, vfrec7_v_h_bf16, OP_UU_H, H2, H2, frec7_h_bf16)\n RVVCALL(OPFVV1, vfrec7_v_h, OP_UU_H, H2, H2, frec7_h)\n RVVCALL(OPFVV1, vfrec7_v_w, OP_UU_W, H4, H4, frec7_s)\n RVVCALL(OPFVV1, vfrec7_v_d, OP_UU_D, H8, H8, frec7_d)\n+GEN_VEXT_V_ENV(vfrec7_v_h_bf16, 2)\n GEN_VEXT_V_ENV(vfrec7_v_h, 2)\n GEN_VEXT_V_ENV(vfrec7_v_w, 4)\n GEN_VEXT_V_ENV(vfrec7_v_d, 8)\n \n /* Vector Floating-Point MIN/MAX Instructions */\n+RVVCALL(OPFVV2, vfmin_vv_h_bf16, OP_UUU_H, H2, H2, H2, bfloat16_minimum_number)\n RVVCALL(OPFVV2, vfmin_vv_h, OP_UUU_H, H2, H2, H2, float16_minimum_number)\n RVVCALL(OPFVV2, vfmin_vv_w, OP_UUU_W, H4, H4, H4, float32_minimum_number)\n RVVCALL(OPFVV2, vfmin_vv_d, OP_UUU_D, H8, H8, H8, float64_minimum_number)\n+GEN_VEXT_VV_ENV(vfmin_vv_h_bf16, 2)\n GEN_VEXT_VV_ENV(vfmin_vv_h, 2)\n GEN_VEXT_VV_ENV(vfmin_vv_w, 4)\n GEN_VEXT_VV_ENV(vfmin_vv_d, 8)\n+RVVCALL(OPFVF2, vfmin_vf_h_bf16, OP_UUU_H, H2, H2, bfloat16_minimum_number)\n RVVCALL(OPFVF2, vfmin_vf_h, OP_UUU_H, H2, H2, float16_minimum_number)\n RVVCALL(OPFVF2, vfmin_vf_w, OP_UUU_W, H4, H4, float32_minimum_number)\n RVVCALL(OPFVF2, vfmin_vf_d, OP_UUU_D, H8, H8, float64_minimum_number)\n+GEN_VEXT_VF(vfmin_vf_h_bf16, 2)\n GEN_VEXT_VF(vfmin_vf_h, 2)\n GEN_VEXT_VF(vfmin_vf_w, 4)\n GEN_VEXT_VF(vfmin_vf_d, 8)\n \n+RVVCALL(OPFVV2, vfmax_vv_h_bf16, OP_UUU_H, H2, H2, H2, bfloat16_maximum_number)\n RVVCALL(OPFVV2, vfmax_vv_h, OP_UUU_H, H2, H2, H2, float16_maximum_number)\n RVVCALL(OPFVV2, vfmax_vv_w, OP_UUU_W, H4, H4, H4, float32_maximum_number)\n RVVCALL(OPFVV2, vfmax_vv_d, OP_UUU_D, H8, H8, H8, float64_maximum_number)\n+GEN_VEXT_VV_ENV(vfmax_vv_h_bf16, 2)\n GEN_VEXT_VV_ENV(vfmax_vv_h, 2)\n GEN_VEXT_VV_ENV(vfmax_vv_w, 4)\n GEN_VEXT_VV_ENV(vfmax_vv_d, 8)\n+RVVCALL(OPFVF2, vfmax_vf_h_bf16, OP_UUU_H, H2, H2, bfloat16_maximum_number)\n RVVCALL(OPFVF2, vfmax_vf_h, OP_UUU_H, H2, H2, float16_maximum_number)\n RVVCALL(OPFVF2, vfmax_vf_w, OP_UUU_W, H4, H4, float32_maximum_number)\n RVVCALL(OPFVF2, vfmax_vf_d, OP_UUU_D, H8, H8, float64_maximum_number)\n+GEN_VEXT_VF(vfmax_vf_h_bf16, 2)\n GEN_VEXT_VF(vfmax_vf_h, 2)\n GEN_VEXT_VF(vfmax_vf_w, 4)\n GEN_VEXT_VF(vfmax_vf_d, 8)\n@@ -4334,6 +4605,7 @@ void HELPER(NAME)(void *vd, void *v0, void *vs1, void *vs2,   \\\n     }                                                         \\\n }\n \n+GEN_VEXT_CMP_VV_ENV(vmfeq_vv_h_bf16, uint16_t, H2, bfloat16_eq_quiet)\n GEN_VEXT_CMP_VV_ENV(vmfeq_vv_h, uint16_t, H2, float16_eq_quiet)\n GEN_VEXT_CMP_VV_ENV(vmfeq_vv_w, uint32_t, H4, float32_eq_quiet)\n GEN_VEXT_CMP_VV_ENV(vmfeq_vv_d, uint64_t, H8, float64_eq_quiet)\n@@ -4375,10 +4647,17 @@ void HELPER(NAME)(void *vd, void *v0, uint64_t s1, void *vs2,       \\\n     }                                                               \\\n }\n \n+GEN_VEXT_CMP_VF(vmfeq_vf_h_bf16, uint16_t, H2, bfloat16_eq_quiet)\n GEN_VEXT_CMP_VF(vmfeq_vf_h, uint16_t, H2, float16_eq_quiet)\n GEN_VEXT_CMP_VF(vmfeq_vf_w, uint32_t, H4, float32_eq_quiet)\n GEN_VEXT_CMP_VF(vmfeq_vf_d, uint64_t, H8, float64_eq_quiet)\n \n+static bool vmfne16_bf16(uint16_t a, uint16_t b, float_status *s)\n+{\n+    FloatRelation compare = bfloat16_compare_quiet(a, b, s);\n+    return compare != float_relation_equal;\n+}\n+\n static bool vmfne16(uint16_t a, uint16_t b, float_status *s)\n {\n     FloatRelation compare = float16_compare_quiet(a, b, s);\n@@ -4397,27 +4676,39 @@ static bool vmfne64(uint64_t a, uint64_t b, float_status *s)\n     return compare != float_relation_equal;\n }\n \n+GEN_VEXT_CMP_VV_ENV(vmfne_vv_h_bf16, uint16_t, H2, vmfne16_bf16)\n GEN_VEXT_CMP_VV_ENV(vmfne_vv_h, uint16_t, H2, vmfne16)\n GEN_VEXT_CMP_VV_ENV(vmfne_vv_w, uint32_t, H4, vmfne32)\n GEN_VEXT_CMP_VV_ENV(vmfne_vv_d, uint64_t, H8, vmfne64)\n+GEN_VEXT_CMP_VF(vmfne_vf_h_bf16, uint16_t, H2, vmfne16_bf16)\n GEN_VEXT_CMP_VF(vmfne_vf_h, uint16_t, H2, vmfne16)\n GEN_VEXT_CMP_VF(vmfne_vf_w, uint32_t, H4, vmfne32)\n GEN_VEXT_CMP_VF(vmfne_vf_d, uint64_t, H8, vmfne64)\n \n+GEN_VEXT_CMP_VV_ENV(vmflt_vv_h_bf16, uint16_t, H2, bfloat16_lt)\n GEN_VEXT_CMP_VV_ENV(vmflt_vv_h, uint16_t, H2, float16_lt)\n GEN_VEXT_CMP_VV_ENV(vmflt_vv_w, uint32_t, H4, float32_lt)\n GEN_VEXT_CMP_VV_ENV(vmflt_vv_d, uint64_t, H8, float64_lt)\n+GEN_VEXT_CMP_VF(vmflt_vf_h_bf16, uint16_t, H2, bfloat16_lt)\n GEN_VEXT_CMP_VF(vmflt_vf_h, uint16_t, H2, float16_lt)\n GEN_VEXT_CMP_VF(vmflt_vf_w, uint32_t, H4, float32_lt)\n GEN_VEXT_CMP_VF(vmflt_vf_d, uint64_t, H8, float64_lt)\n \n+GEN_VEXT_CMP_VV_ENV(vmfle_vv_h_bf16, uint16_t, H2, bfloat16_le)\n GEN_VEXT_CMP_VV_ENV(vmfle_vv_h, uint16_t, H2, float16_le)\n GEN_VEXT_CMP_VV_ENV(vmfle_vv_w, uint32_t, H4, float32_le)\n GEN_VEXT_CMP_VV_ENV(vmfle_vv_d, uint64_t, H8, float64_le)\n+GEN_VEXT_CMP_VF(vmfle_vf_h_bf16, uint16_t, H2, bfloat16_le)\n GEN_VEXT_CMP_VF(vmfle_vf_h, uint16_t, H2, float16_le)\n GEN_VEXT_CMP_VF(vmfle_vf_w, uint32_t, H4, float32_le)\n GEN_VEXT_CMP_VF(vmfle_vf_d, uint64_t, H8, float64_le)\n \n+static bool vmfgt16_bf16(uint16_t a, uint16_t b, float_status *s)\n+{\n+    FloatRelation compare = bfloat16_compare(a, b, s);\n+    return compare == float_relation_greater;\n+}\n+\n static bool vmfgt16(uint16_t a, uint16_t b, float_status *s)\n {\n     FloatRelation compare = float16_compare(a, b, s);\n@@ -4436,10 +4727,18 @@ static bool vmfgt64(uint64_t a, uint64_t b, float_status *s)\n     return compare == float_relation_greater;\n }\n \n+GEN_VEXT_CMP_VF(vmfgt_vf_h_bf16, uint16_t, H2, vmfgt16_bf16)\n GEN_VEXT_CMP_VF(vmfgt_vf_h, uint16_t, H2, vmfgt16)\n GEN_VEXT_CMP_VF(vmfgt_vf_w, uint32_t, H4, vmfgt32)\n GEN_VEXT_CMP_VF(vmfgt_vf_d, uint64_t, H8, vmfgt64)\n \n+static bool vmfge16_bf16(uint16_t a, uint16_t b, float_status *s)\n+{\n+    FloatRelation compare = bfloat16_compare(a, b, s);\n+    return compare == float_relation_greater ||\n+           compare == float_relation_equal;\n+}\n+\n static bool vmfge16(uint16_t a, uint16_t b, float_status *s)\n {\n     FloatRelation compare = float16_compare(a, b, s);\n@@ -4461,11 +4760,31 @@ static bool vmfge64(uint64_t a, uint64_t b, float_status *s)\n            compare == float_relation_equal;\n }\n \n+GEN_VEXT_CMP_VF(vmfge_vf_h_bf16, uint16_t, H2, vmfge16_bf16)\n GEN_VEXT_CMP_VF(vmfge_vf_h, uint16_t, H2, vmfge16)\n GEN_VEXT_CMP_VF(vmfge_vf_w, uint32_t, H4, vmfge32)\n GEN_VEXT_CMP_VF(vmfge_vf_d, uint64_t, H8, vmfge64)\n \n /* Vector Floating-Point Classify Instruction */\n+target_ulong fclass_h_bf16(uint64_t frs1)\n+{\n+    bfloat16 f = frs1;\n+    bool sign = bfloat16_is_neg(f);\n+\n+    if (bfloat16_is_infinity(f)) {\n+        return sign ? 1 << 0 : 1 << 7;\n+    } else if (bfloat16_is_zero(f)) {\n+        return sign ? 1 << 3 : 1 << 4;\n+    } else if (bfloat16_is_zero_or_denormal(f)) {\n+        return sign ? 1 << 2 : 1 << 5;\n+    } else if (bfloat16_is_any_nan(f)) {\n+        float_status s = { }; /* for snan_bit_is_one */\n+        return bfloat16_is_quiet_nan(f, &s) ? 1 << 9 : 1 << 8;\n+    } else {\n+        return sign ? 1 << 1 : 1 << 6;\n+    }\n+}\n+\n target_ulong fclass_h(uint64_t frs1)\n {\n     float16 f = frs1;\n@@ -4523,9 +4842,11 @@ target_ulong fclass_d(uint64_t frs1)\n     }\n }\n \n+RVVCALL(OPIVV1, vfclass_v_h_bf16, OP_UU_H, H2, H2, fclass_h_bf16)\n RVVCALL(OPIVV1, vfclass_v_h, OP_UU_H, H2, H2, fclass_h)\n RVVCALL(OPIVV1, vfclass_v_w, OP_UU_W, H4, H4, fclass_s)\n RVVCALL(OPIVV1, vfclass_v_d, OP_UU_D, H8, H8, fclass_d)\n+GEN_VEXT_V(vfclass_v_h_bf16, 2)\n GEN_VEXT_V(vfclass_v_h, 2)\n GEN_VEXT_V(vfclass_v_w, 4)\n GEN_VEXT_V(vfclass_v_d, 8)\n@@ -4616,17 +4937,21 @@ GEN_VEXT_V_ENV(vfwcvt_x_f_v_w, 8)\n  * vfwcvt.f.xu.v vd, vs2, vm # Convert unsigned integer to double-width float.\n  */\n RVVCALL(OPFVV1, vfwcvt_f_xu_v_b, WOP_UU_B, H2, H1, uint8_to_float16)\n+RVVCALL(OPFVV1, vfwcvt_f_xu_v_b_bf16, WOP_UU_B, H2, H1, uint8_to_bfloat16)\n RVVCALL(OPFVV1, vfwcvt_f_xu_v_h, WOP_UU_H, H4, H2, uint16_to_float32)\n RVVCALL(OPFVV1, vfwcvt_f_xu_v_w, WOP_UU_W, H8, H4, uint32_to_float64)\n GEN_VEXT_V_ENV(vfwcvt_f_xu_v_b, 2)\n+GEN_VEXT_V_ENV(vfwcvt_f_xu_v_b_bf16, 2)\n GEN_VEXT_V_ENV(vfwcvt_f_xu_v_h, 4)\n GEN_VEXT_V_ENV(vfwcvt_f_xu_v_w, 8)\n \n /* vfwcvt.f.x.v vd, vs2, vm # Convert integer to double-width float. */\n RVVCALL(OPFVV1, vfwcvt_f_x_v_b, WOP_UU_B, H2, H1, int8_to_float16)\n+RVVCALL(OPFVV1, vfwcvt_f_x_v_b_bf16, WOP_UU_B, H2, H1, int8_to_bfloat16)\n RVVCALL(OPFVV1, vfwcvt_f_x_v_h, WOP_UU_H, H4, H2, int16_to_float32)\n RVVCALL(OPFVV1, vfwcvt_f_x_v_w, WOP_UU_W, H8, H4, int32_to_float64)\n GEN_VEXT_V_ENV(vfwcvt_f_x_v_b, 2)\n+GEN_VEXT_V_ENV(vfwcvt_f_x_v_b_bf16, 2)\n GEN_VEXT_V_ENV(vfwcvt_f_x_v_h, 4)\n GEN_VEXT_V_ENV(vfwcvt_f_x_v_w, 8)\n \n@@ -4653,17 +4978,21 @@ GEN_VEXT_V_ENV(vfwcvtbf16_f_f_v, 4)\n #define NOP_UU_W uint32_t, uint64_t, uint64_t\n /* vfncvt.xu.f.v vd, vs2, vm # Convert float to unsigned integer. */\n RVVCALL(OPFVV1, vfncvt_xu_f_w_b, NOP_UU_B, H1, H2, float16_to_uint8)\n+RVVCALL(OPFVV1, vfncvt_xu_f_w_b_bf16, NOP_UU_B, H1, H2, bfloat16_to_uint8)\n RVVCALL(OPFVV1, vfncvt_xu_f_w_h, NOP_UU_H, H2, H4, float32_to_uint16)\n RVVCALL(OPFVV1, vfncvt_xu_f_w_w, NOP_UU_W, H4, H8, float64_to_uint32)\n GEN_VEXT_V_ENV(vfncvt_xu_f_w_b, 1)\n+GEN_VEXT_V_ENV(vfncvt_xu_f_w_b_bf16, 1)\n GEN_VEXT_V_ENV(vfncvt_xu_f_w_h, 2)\n GEN_VEXT_V_ENV(vfncvt_xu_f_w_w, 4)\n \n /* vfncvt.x.f.v vd, vs2, vm # Convert double-width float to signed integer. */\n RVVCALL(OPFVV1, vfncvt_x_f_w_b, NOP_UU_B, H1, H2, float16_to_int8)\n+RVVCALL(OPFVV1, vfncvt_x_f_w_b_bf16, NOP_UU_B, H1, H2, bfloat16_to_int8)\n RVVCALL(OPFVV1, vfncvt_x_f_w_h, NOP_UU_H, H2, H4, float32_to_int16)\n RVVCALL(OPFVV1, vfncvt_x_f_w_w, NOP_UU_W, H4, H8, float64_to_int32)\n GEN_VEXT_V_ENV(vfncvt_x_f_w_b, 1)\n+GEN_VEXT_V_ENV(vfncvt_x_f_w_b_bf16, 1)\n GEN_VEXT_V_ENV(vfncvt_x_f_w_h, 2)\n GEN_VEXT_V_ENV(vfncvt_x_f_w_w, 4)\n \n",
    "prefixes": [
        "v6",
        "8/9"
    ]
}