From 228fabd5be82ac15d14109b90c7ee1cd682b4254 Mon Sep 17 00:00:00 2001 From: Brandon Wu Date: Sat, 25 Apr 2026 22:35:26 +0900 Subject: [PATCH] [Clang][RISCV] Introduce OFP8(E4M3, E5M2) RISC-V vector types (#191349) Currently there's no OFP8 scalar type supported in both clang and llvm type system, the vector OFP8 RVV types are lowered to i8 llvm types for now. The reason to support only clang type is because of intrinsics definition capability. If we make the clang type also using uint8 vector types, it's not able to distinguish between E4M3 type and E5M2 type so that we have to append additional type suffix to it. intrinsic spec update pr: https://github.com/riscv-non-isa/riscv-rvv-intrinsic-doc/pull/432 vreinterpret intrinsic PR: https://github.com/llvm/llvm-project/pull/191626 DONT MERGE: We have to get the intrinsic spec merged first to be able to make zvfofp8min change --- clang/include/clang/AST/TypeBase.h | 2 +- clang/include/clang/Basic/RISCVVTypes.def | 36 + clang/include/clang/Basic/riscv_vector.td | 20 +- .../include/clang/Serialization/ASTBitCodes.h | 2 +- clang/lib/Sema/SemaRISCV.cpp | 37 +- clang/lib/Support/RISCVVIntrinsicUtils.cpp | 16 +- .../non-policy/non-overloaded/vfncvt.c | 160 +-- .../non-policy/non-overloaded/vfncvtbf16.c | 192 ++-- .../non-policy/non-overloaded/vfwcvtbf16.c | 48 +- .../zvfofp8min/non-policy/overloaded/vfncvt.c | 160 +-- .../non-policy/overloaded/vfncvtbf16.c | 384 +++---- .../non-policy/overloaded/vfwcvtbf16.c | 96 +- .../zvfofp8min/policy/non-overloaded/vfncvt.c | 332 +++--- .../policy/non-overloaded/vfncvtbf16.c | 560 +++++------ .../policy/non-overloaded/vfwcvtbf16.c | 96 +- .../zvfofp8min/policy/overloaded/vfncvt.c | 332 +++--- .../zvfofp8min/policy/overloaded/vfncvtbf16.c | 944 +++++++++--------- .../zvfofp8min/policy/overloaded/vfwcvtbf16.c | 192 ++-- .../rvv-intrinsic-datatypes.cpp | 38 +- .../CodeGenCXX/riscv-mangle-rvv-vectors.cpp | 62 ++ clang/test/Sema/riscv-types.c | 45 +- clang/utils/TableGen/RISCVVEmitter.cpp | 9 + 22 files changed, 1993 insertions(+), 1770 deletions(-) create mode 100644 clang/test/CodeGenCXX/riscv-mangle-rvv-vectors.cpp diff --git a/clang/include/clang/AST/TypeBase.h b/clang/include/clang/AST/TypeBase.h index 53fec5e2e7a4..b2887bcc3624 100644 --- a/clang/include/clang/AST/TypeBase.h +++ b/clang/include/clang/AST/TypeBase.h @@ -1967,7 +1967,7 @@ protected: unsigned : NumTypeBits; /// The kind (BuiltinType::Kind) of builtin type this is. - static constexpr unsigned NumOfBuiltinTypeBits = 9; + static constexpr unsigned NumOfBuiltinTypeBits = 10; unsigned Kind : NumOfBuiltinTypeBits; }; diff --git a/clang/include/clang/Basic/RISCVVTypes.def b/clang/include/clang/Basic/RISCVVTypes.def index ccb8cb39068e..3cac9522c96a 100644 --- a/clang/include/clang/Basic/RISCVVTypes.def +++ b/clang/include/clang/Basic/RISCVVTypes.def @@ -73,6 +73,11 @@ RVV_VECTOR_TYPE(Name, Id, SingletonId, NumEls, ElBits, NF, false, false, true) #endif +#ifndef RVV_VECTOR_TYPE_OFP8 +#define RVV_VECTOR_TYPE_OFP8(Name, Id, SingletonId, NumEls, E5m2) \ + RVV_VECTOR_TYPE_INT(Name, Id, SingletonId, NumEls, 8, 1, false) +#endif + //===- Vector types -------------------------------------------------------===// RVV_VECTOR_TYPE_INT("__rvv_int8mf8_t", RvvInt8mf8, RvvInt8mf8Ty, 1, 8, 1, true) @@ -127,6 +132,36 @@ RVV_VECTOR_TYPE_INT("__rvv_uint64m2_t",RvvUint64m2,RvvUint64m2Ty,2, 64, 1, fals RVV_VECTOR_TYPE_INT("__rvv_uint64m4_t",RvvUint64m4,RvvUint64m4Ty,4, 64, 1, false) RVV_VECTOR_TYPE_INT("__rvv_uint64m8_t",RvvUint64m8,RvvUint64m8Ty,8, 64, 1, false) +RVV_VECTOR_TYPE_OFP8("__rvv_float8e4m3mf8_t", RvvFloat8E4M3mf8, + RvvFloat8E4M3mf8Ty, 1, false) +RVV_VECTOR_TYPE_OFP8("__rvv_float8e4m3mf4_t", RvvFloat8E4M3mf4, + RvvFloat8E4M3mf4Ty, 2, false) +RVV_VECTOR_TYPE_OFP8("__rvv_float8e4m3mf2_t", RvvFloat8E4M3mf2, + RvvFloat8E4M3mf2Ty, 4, false) +RVV_VECTOR_TYPE_OFP8("__rvv_float8e4m3m1_t", RvvFloat8E4M3m1, + RvvFloat8E4M3m1Ty, 8, false) +RVV_VECTOR_TYPE_OFP8("__rvv_float8e4m3m2_t", RvvFloat8E4M3m2, + RvvFloat8E4M3m2Ty, 16, false) +RVV_VECTOR_TYPE_OFP8("__rvv_float8e4m3m4_t", RvvFloat8E4M3m4, + RvvFloat8E4M3m4Ty, 32, false) +RVV_VECTOR_TYPE_OFP8("__rvv_float8e4m3m8_t", RvvFloat8E4M3m8, + RvvFloat8E4M3m8Ty, 64, false) + +RVV_VECTOR_TYPE_OFP8("__rvv_float8e5m2mf8_t", RvvFloat8E5M2mf8, + RvvFloat8E5M2mf8Ty, 1, true) +RVV_VECTOR_TYPE_OFP8("__rvv_float8e5m2mf4_t", RvvFloat8E5M2mf4, + RvvFloat8E5M2mf4Ty, 2, true) +RVV_VECTOR_TYPE_OFP8("__rvv_float8e5m2mf2_t", RvvFloat8E5M2mf2, + RvvFloat8E5M2mf2Ty, 4, true) +RVV_VECTOR_TYPE_OFP8("__rvv_float8e5m2m1_t", RvvFloat8E5M2m1, + RvvFloat8E5M2m1Ty, 8, true) +RVV_VECTOR_TYPE_OFP8("__rvv_float8e5m2m2_t", RvvFloat8E5M2m2, + RvvFloat8E5M2m2Ty, 16, true) +RVV_VECTOR_TYPE_OFP8("__rvv_float8e5m2m4_t", RvvFloat8E5M2m4, + RvvFloat8E5M2m4Ty, 32, true) +RVV_VECTOR_TYPE_OFP8("__rvv_float8e5m2m8_t", RvvFloat8E5M2m8, + RvvFloat8E5M2m8Ty, 64, true) + RVV_VECTOR_TYPE_FLOAT("__rvv_float16mf4_t",RvvFloat16mf4,RvvFloat16mf4Ty,1, 16, 1) RVV_VECTOR_TYPE_FLOAT("__rvv_float16mf2_t",RvvFloat16mf2,RvvFloat16mf2Ty,2, 16, 1) RVV_VECTOR_TYPE_FLOAT("__rvv_float16m1_t", RvvFloat16m1, RvvFloat16m1Ty, 4, 16, 1) @@ -508,6 +543,7 @@ RVV_VECTOR_TYPE_BFLOAT("__rvv_bfloat16m2x4_t", RvvBFloat16m2x4, RvvBFloat16m2x4T RVV_VECTOR_TYPE_BFLOAT("__rvv_bfloat16m4x2_t", RvvBFloat16m4x2, RvvBFloat16m4x2Ty, 16, 16, 2) +#undef RVV_VECTOR_TYPE_OFP8 #undef RVV_VECTOR_TYPE_BFLOAT #undef RVV_VECTOR_TYPE_FLOAT #undef RVV_VECTOR_TYPE_INT diff --git a/clang/include/clang/Basic/riscv_vector.td b/clang/include/clang/Basic/riscv_vector.td index cc20d7459b24..414769e2429c 100644 --- a/clang/include/clang/Basic/riscv_vector.td +++ b/clang/include/clang/Basic/riscv_vector.td @@ -1360,9 +1360,9 @@ def vfwcvtbf16_f_f_v : RVVConvBuiltin<"Fw", "Fwv", "y", "vfwcvtbf16_f">; let Log2LMUL = [-3, -2, -1, 0, 1, 2], RequiredFeatures = ["zvfofp8min"], UnMaskedPolicyScheme = HasPassthruOperand in { - let OverloadedName = "vfwcvt_f_f8e4m3_bf16" in + let OverloadedName = "vfwcvt_f_bf16" in defm : RVVConvBuiltinSet<"vfwcvt_f_f_v", "a", [["vw", "wv"]]>; - let OverloadedName = "vfwcvt_f_f8e5m2_bf16", + let OverloadedName = "vfwcvt_f_bf16", IRName = "vfwcvt_f_f_v_alt", MaskedIRName = "vfwcvt_f_f_v_alt_mask", AltFmt = 1 in @@ -1451,16 +1451,16 @@ let ManualCodegen = [{ // Zvfofp8min let RequiredFeatures = ["zvfofp8min"] in { let Log2LMUL = [-3, -2, -1, 0, 1, 2] in { - let OverloadedName = "vfncvt_f_bf16_f8e4m3" in + let OverloadedName = "vfncvt_f_f8e4m3" in defm : RVVConvBuiltinSet<"vfncvt_f_f_w", "a", [["wv", "vwu"]]>; - let OverloadedName = "vfncvt_f_bf16_f8e5m2", + let OverloadedName = "vfncvt_f_f8e5m2", IRName = "vfncvt_f_f_w_alt", MaskedIRName = "vfncvt_f_f_w_alt_mask", AltFmt = 1 in defm : RVVConvBuiltinSet<"vfncvt_f_f_w", "b", [["wv", "vwu"]]>; - let OverloadedName = "vfncvt_sat_f_bf16_f8e4m3" in + let OverloadedName = "vfncvt_sat_f_f8e4m3" in defm : RVVConvBuiltinSet<"vfncvt_sat_f_f_w", "a", [["wv", "vwu"]]>; - let OverloadedName = "vfncvt_sat_f_bf16_f8e5m2", + let OverloadedName = "vfncvt_sat_f_f8e5m2", IRName = "vfncvt_sat_f_f_w_alt", MaskedIRName = "vfncvt_sat_f_f_w_alt_mask", AltFmt = 1 in @@ -1559,16 +1559,16 @@ let ManualCodegen = [{ // Zvfofp8min let RequiredFeatures = ["zvfofp8min"] in { let Log2LMUL = [-3, -2, -1, 0, 1, 2] in { - let OverloadedName = "vfncvt_f_bf16_f8e4m3" in + let OverloadedName = "vfncvt_f_f8e4m3" in defm : RVVConvBuiltinSet<"vfncvt_f_f_w", "a", [["wv", "vw"]]>; - let OverloadedName = "vfncvt_f_bf16_f8e5m2", + let OverloadedName = "vfncvt_f_f8e5m2", IRName = "vfncvt_f_f_w_alt", MaskedIRName = "vfncvt_f_f_w_alt_mask", AltFmt = 1 in defm : RVVConvBuiltinSet<"vfncvt_f_f_w", "b", [["wv", "vw"]]>; - let OverloadedName = "vfncvt_sat_f_bf16_f8e4m3" in + let OverloadedName = "vfncvt_sat_f_f8e4m3" in defm : RVVConvBuiltinSet<"vfncvt_sat_f_f_w", "a", [["wv", "vw"]]>; - let OverloadedName = "vfncvt_sat_f_bf16_f8e5m2", + let OverloadedName = "vfncvt_sat_f_f8e5m2", IRName = "vfncvt_sat_f_f_w_alt", MaskedIRName = "vfncvt_sat_f_f_w_alt_mask", AltFmt = 1 in diff --git a/clang/include/clang/Serialization/ASTBitCodes.h b/clang/include/clang/Serialization/ASTBitCodes.h index dcffe84df05f..9a41f9e89df9 100644 --- a/clang/include/clang/Serialization/ASTBitCodes.h +++ b/clang/include/clang/Serialization/ASTBitCodes.h @@ -1166,7 +1166,7 @@ enum PredefinedTypeIDs { /// /// Type IDs for non-predefined types will start at /// NUM_PREDEF_TYPE_IDs. -const unsigned NUM_PREDEF_TYPE_IDS = 515; +const unsigned NUM_PREDEF_TYPE_IDS = 529; // Ensure we do not overrun the predefined types we reserved // in the enum PredefinedTypeIDs above. diff --git a/clang/lib/Sema/SemaRISCV.cpp b/clang/lib/Sema/SemaRISCV.cpp index 3cf5a0963b39..56555b931b8d 100644 --- a/clang/lib/Sema/SemaRISCV.cpp +++ b/clang/lib/Sema/SemaRISCV.cpp @@ -137,9 +137,24 @@ static QualType RVVType2Qual(ASTContext &Context, const RVVType *Type) { QT = Context.getIntTypeForBitwidth(Type->getElementBitwidth(), false); break; case ScalarTypeKind::FloatE4M3: - case ScalarTypeKind::FloatE5M2: - QT = Context.getIntTypeForBitwidth(8, false); - break; + case ScalarTypeKind::FloatE5M2: { + // TODO: This is a workaround code to only support OP8 RVV types without + // supporting scalar OFP8 types. We need to refactor after scalar types are + // supported. + assert(Type->isVector() && "Only support vector of OFP8 types."); + bool IsE5M2 = Type->getScalarType() == ScalarTypeKind::FloatE5M2; + unsigned Scale = *Type->getScale(); +#define RVV_VECTOR_TYPE_OFP8(Name, Id, SingletonId, NumEls, E5m2) \ + if (IsE5M2 == E5m2 && Scale == NumEls) \ + QT = Context.SingletonId; +#include "clang/Basic/RISCVVTypes.def" + assert(!QT.isNull() && "Unsupported OFP8 vector type"); + if (Type->isConstant()) + QT = Context.getConstType(QT); + if (Type->isPointer()) + QT = Context.getPointerType(QT); + return QT; + } case ScalarTypeKind::BFloat: QT = Context.BFloat16Ty; break; @@ -1513,11 +1528,23 @@ bool SemaRISCV::CheckBuiltinFunctionCall(const TargetInfo &TI, void SemaRISCV::checkRVVTypeSupport(QualType Ty, SourceLocation Loc, Decl *D, const llvm::StringMap &FeatureMap) { + const BuiltinType *BT = Ty->castAs(); ASTContext::BuiltinVectorTypeInfo Info = - SemaRef.Context.getBuiltinVectorTypeInfo(Ty->castAs()); + SemaRef.Context.getBuiltinVectorTypeInfo(BT); unsigned EltSize = SemaRef.Context.getTypeSize(Info.ElementType); unsigned MinElts = Info.EC.getKnownMinValue(); + auto IsOFP8Type = [](const BuiltinType *BT) { + switch (BT->getKind()) { +#define RVV_VECTOR_TYPE_OFP8(Name, Id, SingletonId, NumEls, E5m2) \ + case BuiltinType::Id: +#include "clang/Basic/RISCVVTypes.def" + return true; + default: + return false; + } + }; + if (Info.ElementType->isSpecificBuiltinType(BuiltinType::Double) && !FeatureMap.lookup("zve64d")) Diag(Loc, diag::err_riscv_type_requires_extension) << Ty << "zve64d"; @@ -1554,6 +1581,8 @@ void SemaRISCV::checkRVVTypeSupport(QualType Ty, SourceLocation Loc, Decl *D, // if we don't have at least zve32x supported, then we need to emit error. else if (!FeatureMap.lookup("zve32x")) Diag(Loc, diag::err_riscv_type_requires_extension) << Ty << "zve32x"; + else if (IsOFP8Type(BT) && !FeatureMap.lookup("experimental-zvfofp8min")) + Diag(Loc, diag::err_riscv_type_requires_extension) << Ty << "zvfofp8min"; } /// Are the two types RVV-bitcast-compatible types? I.e. is bitcasting from the diff --git a/clang/lib/Support/RISCVVIntrinsicUtils.cpp b/clang/lib/Support/RISCVVIntrinsicUtils.cpp index 8f562235d605..866b54f6e5b1 100644 --- a/clang/lib/Support/RISCVVIntrinsicUtils.cpp +++ b/clang/lib/Support/RISCVVIntrinsicUtils.cpp @@ -250,10 +250,14 @@ void RVVType::initClangBuiltinStr() { ClangBuiltinStr += "int"; break; case ScalarTypeKind::UnsignedInteger: - case ScalarTypeKind::FloatE4M3: - case ScalarTypeKind::FloatE5M2: ClangBuiltinStr += "uint"; break; + case ScalarTypeKind::FloatE4M3: + ClangBuiltinStr += "float8e4m3" + LMUL.str() + "_t"; + return; + case ScalarTypeKind::FloatE5M2: + ClangBuiltinStr += "float8e5m2" + LMUL.str() + "_t"; + return; default: llvm_unreachable("ScalarTypeKind is invalid"); } @@ -327,10 +331,14 @@ void RVVType::initTypeStr() { Str += getTypeString("int"); break; case ScalarTypeKind::UnsignedInteger: - case ScalarTypeKind::FloatE4M3: - case ScalarTypeKind::FloatE5M2: Str += getTypeString("uint"); break; + case ScalarTypeKind::FloatE4M3: + Str += "vfloat8e4m3" + LMUL.str() + "_t"; + break; + case ScalarTypeKind::FloatE5M2: + Str += "vfloat8e5m2" + LMUL.str() + "_t"; + break; default: llvm_unreachable("ScalarType is invalid!"); } diff --git a/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/non-policy/non-overloaded/vfncvt.c b/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/non-policy/non-overloaded/vfncvt.c index 3daae3806afd..ec8b8c0f9c1e 100644 --- a/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/non-policy/non-overloaded/vfncvt.c +++ b/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/non-policy/non-overloaded/vfncvt.c @@ -13,7 +13,7 @@ // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv1i8.nxv1f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8(vfloat32mf2_t vs2, size_t vl) { +vfloat8e4m3mf8_t test_vfncvt_f_f_q_f8e4m3mf8(vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf8(vs2, vl); } @@ -23,7 +23,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8(vfloat32mf2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv1i8.nxv1f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8(vfloat32mf2_t vs2, size_t vl) { +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8(vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf8(vs2, vl); } @@ -33,7 +33,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8(vfloat32mf2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv2i8.nxv2f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4(vfloat32m1_t vs2, size_t vl) { +vfloat8e4m3mf4_t test_vfncvt_f_f_q_f8e4m3mf4(vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf4(vs2, vl); } @@ -43,7 +43,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4(vfloat32m1_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv2i8.nxv2f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4(vfloat32m1_t vs2, size_t vl) { +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4(vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf4(vs2, vl); } @@ -53,7 +53,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4(vfloat32m1_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv4i8.nxv4f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2(vfloat32m2_t vs2, size_t vl) { +vfloat8e4m3mf2_t test_vfncvt_f_f_q_f8e4m3mf2(vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf2(vs2, vl); } @@ -63,7 +63,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2(vfloat32m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv4i8.nxv4f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2(vfloat32m2_t vs2, size_t vl) { +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2(vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf2(vs2, vl); } @@ -73,7 +73,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2(vfloat32m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv8i8.nxv8f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1(vfloat32m4_t vs2, size_t vl) { +vfloat8e4m3m1_t test_vfncvt_f_f_q_f8e4m3m1(vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3m1(vs2, vl); } @@ -83,7 +83,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1(vfloat32m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv8i8.nxv8f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1(vfloat32m4_t vs2, size_t vl) { +vfloat8e4m3m1_t test_vfncvt_sat_f_f_q_f8e4m3m1(vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3m1(vs2, vl); } @@ -93,7 +93,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1(vfloat32m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv16i8.nxv16f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2(vfloat32m8_t vs2, size_t vl) { +vfloat8e4m3m2_t test_vfncvt_f_f_q_f8e4m3m2(vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3m2(vs2, vl); } @@ -103,7 +103,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2(vfloat32m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv16i8.nxv16f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2(vfloat32m8_t vs2, size_t vl) { +vfloat8e4m3m2_t test_vfncvt_sat_f_f_q_f8e4m3m2(vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3m2(vs2, vl); } @@ -113,7 +113,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2(vfloat32m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv1i8.nxv1f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_m(vbool64_t vm, vfloat32mf2_t vs2, +vfloat8e4m3mf8_t test_vfncvt_f_f_q_f8e4m3mf8_m(vbool64_t vm, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf8_m(vm, vs2, vl); } @@ -124,7 +124,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_m(vbool64_t vm, vfloat32mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv1i8.nxv1f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_m(vbool64_t vm, vfloat32mf2_t vs2, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_m(vbool64_t vm, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf8_m(vm, vs2, vl); } @@ -135,7 +135,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_m(vbool64_t vm, vfloat32mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv2i8.nxv2f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_m(vbool32_t vm, vfloat32m1_t vs2, +vfloat8e4m3mf4_t test_vfncvt_f_f_q_f8e4m3mf4_m(vbool32_t vm, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf4_m(vm, vs2, vl); } @@ -146,7 +146,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_m(vbool32_t vm, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv2i8.nxv2f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_m(vbool32_t vm, vfloat32m1_t vs2, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_m(vbool32_t vm, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf4_m(vm, vs2, vl); } @@ -157,7 +157,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_m(vbool32_t vm, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv4i8.nxv4f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_m(vbool16_t vm, vfloat32m2_t vs2, +vfloat8e4m3mf2_t test_vfncvt_f_f_q_f8e4m3mf2_m(vbool16_t vm, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf2_m(vm, vs2, vl); } @@ -168,7 +168,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_m(vbool16_t vm, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv4i8.nxv4f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_m(vbool16_t vm, vfloat32m2_t vs2, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_m(vbool16_t vm, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf2_m(vm, vs2, vl); } @@ -179,7 +179,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_m(vbool16_t vm, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv8i8.nxv8f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_m(vbool8_t vm, vfloat32m4_t vs2, +vfloat8e4m3m1_t test_vfncvt_f_f_q_f8e4m3m1_m(vbool8_t vm, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3m1_m(vm, vs2, vl); } @@ -190,7 +190,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_m(vbool8_t vm, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv8i8.nxv8f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_m(vbool8_t vm, vfloat32m4_t vs2, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_m(vbool8_t vm, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3m1_m(vm, vs2, vl); } @@ -201,7 +201,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_m(vbool8_t vm, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv16i8.nxv16f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_m(vbool4_t vm, vfloat32m8_t vs2, +vfloat8e4m3m2_t test_vfncvt_f_f_q_f8e4m3m2_m(vbool4_t vm, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3m2_m(vm, vs2, vl); } @@ -212,7 +212,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_m(vbool4_t vm, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv16i8.nxv16f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_m(vbool4_t vm, vfloat32m8_t vs2, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_m(vbool4_t vm, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3m2_m(vm, vs2, vl); } @@ -223,7 +223,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_m(vbool4_t vm, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv1i8.nxv1f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm(vfloat32mf2_t vs2, size_t vl) { +vfloat8e4m3mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm(vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf8_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -233,7 +233,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm(vfloat32mf2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv1i8.nxv1f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm(vfloat32mf2_t vs2, size_t vl) { +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm(vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf8_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -243,7 +243,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm(vfloat32mf2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv2i8.nxv2f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm(vfloat32m1_t vs2, size_t vl) { +vfloat8e4m3mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm(vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf4_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -253,7 +253,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm(vfloat32m1_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv2i8.nxv2f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm(vfloat32m1_t vs2, size_t vl) { +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm(vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf4_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -263,7 +263,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm(vfloat32m1_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv4i8.nxv4f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm(vfloat32m2_t vs2, size_t vl) { +vfloat8e4m3mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm(vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf2_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -273,7 +273,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm(vfloat32m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv4i8.nxv4f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm(vfloat32m2_t vs2, size_t vl) { +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm(vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf2_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -283,7 +283,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm(vfloat32m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv8i8.nxv8f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_rm(vfloat32m4_t vs2, size_t vl) { +vfloat8e4m3m1_t test_vfncvt_f_f_q_f8e4m3m1_rm(vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3m1_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -293,7 +293,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_rm(vfloat32m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv8i8.nxv8f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm(vfloat32m4_t vs2, size_t vl) { +vfloat8e4m3m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm(vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3m1_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -303,7 +303,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm(vfloat32m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv16i8.nxv16f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_rm(vfloat32m8_t vs2, size_t vl) { +vfloat8e4m3m2_t test_vfncvt_f_f_q_f8e4m3m2_rm(vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3m2_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -313,7 +313,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_rm(vfloat32m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv16i8.nxv16f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm(vfloat32m8_t vs2, size_t vl) { +vfloat8e4m3m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm(vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3m2_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -323,7 +323,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm(vfloat32m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv1i8.nxv1f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_m(vbool64_t vm, vfloat32mf2_t vs2, +vfloat8e4m3mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_m(vbool64_t vm, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf8_rm_m(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -334,7 +334,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_m(vbool64_t vm, vfloat32mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv1i8.nxv1f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_m(vbool64_t vm, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_m(vbool64_t vm, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf8_rm_m(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -345,7 +345,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_m(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv2i8.nxv2f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_m(vbool32_t vm, vfloat32m1_t vs2, +vfloat8e4m3mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_m(vbool32_t vm, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf4_rm_m(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -356,7 +356,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_m(vbool32_t vm, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv2i8.nxv2f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_m(vbool32_t vm, vfloat32m1_t vs2, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_m(vbool32_t vm, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf4_rm_m(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -367,7 +367,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_m(vbool32_t vm, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv4i8.nxv4f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_m(vbool16_t vm, vfloat32m2_t vs2, +vfloat8e4m3mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_m(vbool16_t vm, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf2_rm_m(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -378,7 +378,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_m(vbool16_t vm, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv4i8.nxv4f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_m(vbool16_t vm, vfloat32m2_t vs2, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_m(vbool16_t vm, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf2_rm_m(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -389,7 +389,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_m(vbool16_t vm, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv8i8.nxv8f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_m(vbool8_t vm, vfloat32m4_t vs2, +vfloat8e4m3m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_m(vbool8_t vm, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3m1_rm_m(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -400,7 +400,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_m(vbool8_t vm, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv8i8.nxv8f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_m(vbool8_t vm, vfloat32m4_t vs2, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_m(vbool8_t vm, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3m1_rm_m(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -411,7 +411,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_m(vbool8_t vm, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv16i8.nxv16f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_m(vbool4_t vm, vfloat32m8_t vs2, +vfloat8e4m3m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_m(vbool4_t vm, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3m2_rm_m(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -422,7 +422,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_m(vbool4_t vm, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv16i8.nxv16f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_m(vbool4_t vm, vfloat32m8_t vs2, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_m(vbool4_t vm, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3m2_rm_m(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -433,7 +433,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_m(vbool4_t vm, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv1i8.nxv1f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8(vfloat32mf2_t vs2, size_t vl) { +vfloat8e5m2mf8_t test_vfncvt_f_f_q_f8e5m2mf8(vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf8(vs2, vl); } @@ -443,7 +443,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8(vfloat32mf2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv1i8.nxv1f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8(vfloat32mf2_t vs2, size_t vl) { +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8(vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf8(vs2, vl); } @@ -453,7 +453,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8(vfloat32mf2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv2i8.nxv2f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4(vfloat32m1_t vs2, size_t vl) { +vfloat8e5m2mf4_t test_vfncvt_f_f_q_f8e5m2mf4(vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf4(vs2, vl); } @@ -463,7 +463,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4(vfloat32m1_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv2i8.nxv2f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4(vfloat32m1_t vs2, size_t vl) { +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4(vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf4(vs2, vl); } @@ -473,7 +473,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4(vfloat32m1_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv4i8.nxv4f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2(vfloat32m2_t vs2, size_t vl) { +vfloat8e5m2mf2_t test_vfncvt_f_f_q_f8e5m2mf2(vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf2(vs2, vl); } @@ -483,7 +483,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2(vfloat32m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv4i8.nxv4f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2(vfloat32m2_t vs2, size_t vl) { +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2(vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf2(vs2, vl); } @@ -493,7 +493,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2(vfloat32m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv8i8.nxv8f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1(vfloat32m4_t vs2, size_t vl) { +vfloat8e5m2m1_t test_vfncvt_f_f_q_f8e5m2m1(vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2m1(vs2, vl); } @@ -503,7 +503,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1(vfloat32m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv8i8.nxv8f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1(vfloat32m4_t vs2, size_t vl) { +vfloat8e5m2m1_t test_vfncvt_sat_f_f_q_f8e5m2m1(vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2m1(vs2, vl); } @@ -513,7 +513,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1(vfloat32m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv16i8.nxv16f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2(vfloat32m8_t vs2, size_t vl) { +vfloat8e5m2m2_t test_vfncvt_f_f_q_f8e5m2m2(vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2m2(vs2, vl); } @@ -523,7 +523,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2(vfloat32m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv16i8.nxv16f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2(vfloat32m8_t vs2, size_t vl) { +vfloat8e5m2m2_t test_vfncvt_sat_f_f_q_f8e5m2m2(vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2m2(vs2, vl); } @@ -533,7 +533,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2(vfloat32m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_m(vbool64_t vm, vfloat32mf2_t vs2, +vfloat8e5m2mf8_t test_vfncvt_f_f_q_f8e5m2mf8_m(vbool64_t vm, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf8_m(vm, vs2, vl); } @@ -544,7 +544,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_m(vbool64_t vm, vfloat32mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_m(vbool64_t vm, vfloat32mf2_t vs2, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_m(vbool64_t vm, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf8_m(vm, vs2, vl); } @@ -555,7 +555,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_m(vbool64_t vm, vfloat32mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_m(vbool32_t vm, vfloat32m1_t vs2, +vfloat8e5m2mf4_t test_vfncvt_f_f_q_f8e5m2mf4_m(vbool32_t vm, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf4_m(vm, vs2, vl); } @@ -566,7 +566,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_m(vbool32_t vm, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_m(vbool32_t vm, vfloat32m1_t vs2, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_m(vbool32_t vm, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf4_m(vm, vs2, vl); } @@ -577,7 +577,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_m(vbool32_t vm, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_m(vbool16_t vm, vfloat32m2_t vs2, +vfloat8e5m2mf2_t test_vfncvt_f_f_q_f8e5m2mf2_m(vbool16_t vm, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf2_m(vm, vs2, vl); } @@ -588,7 +588,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_m(vbool16_t vm, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_m(vbool16_t vm, vfloat32m2_t vs2, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_m(vbool16_t vm, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf2_m(vm, vs2, vl); } @@ -599,7 +599,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_m(vbool16_t vm, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_m(vbool8_t vm, vfloat32m4_t vs2, +vfloat8e5m2m1_t test_vfncvt_f_f_q_f8e5m2m1_m(vbool8_t vm, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2m1_m(vm, vs2, vl); } @@ -610,7 +610,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_m(vbool8_t vm, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_m(vbool8_t vm, vfloat32m4_t vs2, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_m(vbool8_t vm, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2m1_m(vm, vs2, vl); } @@ -621,7 +621,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_m(vbool8_t vm, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_m(vbool4_t vm, vfloat32m8_t vs2, +vfloat8e5m2m2_t test_vfncvt_f_f_q_f8e5m2m2_m(vbool4_t vm, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2m2_m(vm, vs2, vl); } @@ -632,7 +632,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_m(vbool4_t vm, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_m(vbool4_t vm, vfloat32m8_t vs2, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_m(vbool4_t vm, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2m2_m(vm, vs2, vl); } @@ -643,7 +643,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_m(vbool4_t vm, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv1i8.nxv1f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm(vfloat32mf2_t vs2, size_t vl) { +vfloat8e5m2mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm(vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf8_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -653,7 +653,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm(vfloat32mf2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv1i8.nxv1f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm(vfloat32mf2_t vs2, size_t vl) { +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm(vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf8_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -663,7 +663,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm(vfloat32mf2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv2i8.nxv2f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm(vfloat32m1_t vs2, size_t vl) { +vfloat8e5m2mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm(vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf4_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -673,7 +673,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm(vfloat32m1_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv2i8.nxv2f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm(vfloat32m1_t vs2, size_t vl) { +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm(vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf4_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -683,7 +683,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm(vfloat32m1_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv4i8.nxv4f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm(vfloat32m2_t vs2, size_t vl) { +vfloat8e5m2mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm(vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf2_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -693,7 +693,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm(vfloat32m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv4i8.nxv4f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm(vfloat32m2_t vs2, size_t vl) { +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm(vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf2_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -703,7 +703,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm(vfloat32m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv8i8.nxv8f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_rm(vfloat32m4_t vs2, size_t vl) { +vfloat8e5m2m1_t test_vfncvt_f_f_q_f8e5m2m1_rm(vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2m1_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -713,7 +713,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_rm(vfloat32m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv8i8.nxv8f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm(vfloat32m4_t vs2, size_t vl) { +vfloat8e5m2m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm(vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2m1_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -723,7 +723,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm(vfloat32m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv16i8.nxv16f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_rm(vfloat32m8_t vs2, size_t vl) { +vfloat8e5m2m2_t test_vfncvt_f_f_q_f8e5m2m2_rm(vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2m2_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -733,7 +733,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_rm(vfloat32m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv16i8.nxv16f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm(vfloat32m8_t vs2, size_t vl) { +vfloat8e5m2m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm(vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2m2_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -743,7 +743,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm(vfloat32m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_m(vbool64_t vm, vfloat32mf2_t vs2, +vfloat8e5m2mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_m(vbool64_t vm, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf8_rm_m(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -754,7 +754,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_m(vbool64_t vm, vfloat32mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_m(vbool64_t vm, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_m(vbool64_t vm, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf8_rm_m(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -765,7 +765,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_m(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_m(vbool32_t vm, vfloat32m1_t vs2, +vfloat8e5m2mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_m(vbool32_t vm, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf4_rm_m(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -776,7 +776,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_m(vbool32_t vm, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_m(vbool32_t vm, vfloat32m1_t vs2, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_m(vbool32_t vm, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf4_rm_m(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -787,7 +787,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_m(vbool32_t vm, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_m(vbool16_t vm, vfloat32m2_t vs2, +vfloat8e5m2mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_m(vbool16_t vm, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf2_rm_m(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -798,7 +798,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_m(vbool16_t vm, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_m(vbool16_t vm, vfloat32m2_t vs2, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_m(vbool16_t vm, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf2_rm_m(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -809,7 +809,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_m(vbool16_t vm, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_m(vbool8_t vm, vfloat32m4_t vs2, +vfloat8e5m2m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_m(vbool8_t vm, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2m1_rm_m(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -820,7 +820,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_m(vbool8_t vm, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_m(vbool8_t vm, vfloat32m4_t vs2, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_m(vbool8_t vm, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2m1_rm_m(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -831,7 +831,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_m(vbool8_t vm, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_m(vbool4_t vm, vfloat32m8_t vs2, +vfloat8e5m2m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_m(vbool4_t vm, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2m2_rm_m(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -842,7 +842,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_m(vbool4_t vm, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_m(vbool4_t vm, vfloat32m8_t vs2, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_m(vbool4_t vm, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2m2_rm_m(vm, vs2, __RISCV_FRM_RNE, vl); } diff --git a/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/non-policy/non-overloaded/vfncvtbf16.c b/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/non-policy/non-overloaded/vfncvtbf16.c index 158a93b9757f..499e9dc093bb 100644 --- a/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/non-policy/non-overloaded/vfncvtbf16.c +++ b/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/non-policy/non-overloaded/vfncvtbf16.c @@ -13,7 +13,7 @@ // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv1i8.nxv1bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8(vbfloat16mf4_t vs2, size_t vl) { +vfloat8e4m3mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8(vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf4_f8e4m3mf8(vs2, vl); } @@ -23,7 +23,7 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8(vbfloat16mf4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv1i8.nxv1bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8(vbfloat16mf4_t vs2, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8(vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8(vs2, vl); } @@ -34,7 +34,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8(vbfloat16mf4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv2i8.nxv2bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4(vbfloat16mf2_t vs2, size_t vl) { +vfloat8e4m3mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4(vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf2_f8e4m3mf4(vs2, vl); } @@ -44,7 +44,7 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4(vbfloat16mf2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv2i8.nxv2bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4(vbfloat16mf2_t vs2, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4(vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4(vs2, vl); } @@ -55,7 +55,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4(vbfloat16mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv4i8.nxv4bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2(vbfloat16m1_t vs2, size_t vl) { +vfloat8e4m3mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2(vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m1_f8e4m3mf2(vs2, vl); } @@ -65,7 +65,7 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2(vbfloat16m1_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv4i8.nxv4bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2(vbfloat16m1_t vs2, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2(vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2(vs2, vl); } @@ -76,7 +76,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2(vbfloat16m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv8i8.nxv8bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1(vbfloat16m2_t vs2, size_t vl) { +vfloat8e4m3m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1(vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m2_f8e4m3m1(vs2, vl); } @@ -86,7 +86,7 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1(vbfloat16m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv8i8.nxv8bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1(vbfloat16m2_t vs2, size_t vl) { +vfloat8e4m3m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1(vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1(vs2, vl); } @@ -96,7 +96,7 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1(vbfloat16m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv16i8.nxv16bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2(vbfloat16m4_t vs2, size_t vl) { +vfloat8e4m3m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2(vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m4_f8e4m3m2(vs2, vl); } @@ -106,7 +106,7 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2(vbfloat16m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv16i8.nxv16bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2(vbfloat16m4_t vs2, size_t vl) { +vfloat8e4m3m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2(vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2(vs2, vl); } @@ -116,7 +116,7 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2(vbfloat16m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv32i8.nxv32bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4(vbfloat16m8_t vs2, size_t vl) { +vfloat8e4m3m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4(vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m8_f8e4m3m4(vs2, vl); } @@ -126,7 +126,7 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4(vbfloat16m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv32i8.nxv32bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4(vbfloat16m8_t vs2, size_t vl) { +vfloat8e4m3m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4(vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4(vs2, vl); } @@ -136,7 +136,7 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4(vbfloat16m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv1i8.nxv1bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_m(vbool64_t vm, +vfloat8e4m3mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_m(vbool64_t vm, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_m(vm, vs2, vl); @@ -148,7 +148,7 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_m(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv1i8.nxv1bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_m(vbool64_t vm, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_m(vbool64_t vm, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_m(vm, vs2, vl); @@ -160,7 +160,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_m(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv2i8.nxv2bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_m(vbool32_t vm, +vfloat8e4m3mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_m(vbool32_t vm, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_m(vm, vs2, vl); @@ -172,7 +172,7 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_m(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv2i8.nxv2bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_m(vbool32_t vm, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_m(vbool32_t vm, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_m(vm, vs2, vl); @@ -184,7 +184,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_m(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv4i8.nxv4bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_m(vbool16_t vm, +vfloat8e4m3mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_m(vbool16_t vm, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m1_f8e4m3mf2_m(vm, vs2, vl); } @@ -195,7 +195,7 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_m(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv4i8.nxv4bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_m(vbool16_t vm, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_m(vbool16_t vm, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_m(vm, vs2, vl); @@ -207,7 +207,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_m(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv8i8.nxv8bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_m(vbool8_t vm, vbfloat16m2_t vs2, +vfloat8e4m3m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_m(vbool8_t vm, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m2_f8e4m3m1_m(vm, vs2, vl); } @@ -218,7 +218,7 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_m(vbool8_t vm, vbfloat16m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv8i8.nxv8bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_m(vbool8_t vm, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_m(vbool8_t vm, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_m(vm, vs2, vl); @@ -230,7 +230,7 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_m(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv16i8.nxv16bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_m(vbool4_t vm, vbfloat16m4_t vs2, +vfloat8e4m3m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_m(vbool4_t vm, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m4_f8e4m3m2_m(vm, vs2, vl); } @@ -241,7 +241,7 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_m(vbool4_t vm, vbfloat16m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv16i8.nxv16bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_m(vbool4_t vm, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_m(vbool4_t vm, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_m(vm, vs2, vl); @@ -253,7 +253,7 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_m(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv32i8.nxv32bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_m(vbool2_t vm, vbfloat16m8_t vs2, +vfloat8e4m3m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_m(vbool2_t vm, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m8_f8e4m3m4_m(vm, vs2, vl); } @@ -264,7 +264,7 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_m(vbool2_t vm, vbfloat16m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv32i8.nxv32bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_m(vbool2_t vm, +vfloat8e4m3m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_m(vbool2_t vm, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_m(vm, vs2, vl); @@ -276,7 +276,7 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_m(vbool2_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv1i8.nxv1bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm(vbfloat16mf4_t vs2, +vfloat8e4m3mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm(vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -287,7 +287,7 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm(vbfloat16mf4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv1i8.nxv1bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm(vbfloat16mf4_t vs2, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm(vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm(vs2, __RISCV_FRM_RNE, vl); @@ -299,7 +299,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm(vbfloat16mf4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv2i8.nxv2bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm(vbfloat16mf2_t vs2, +vfloat8e4m3mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm(vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -310,7 +310,7 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm(vbfloat16mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv2i8.nxv2bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm(vbfloat16mf2_t vs2, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm(vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm(vs2, __RISCV_FRM_RNE, vl); @@ -322,7 +322,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm(vbfloat16mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv4i8.nxv4bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm(vbfloat16m1_t vs2, +vfloat8e4m3mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm(vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -333,7 +333,7 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm(vbfloat16m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv4i8.nxv4bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm(vbfloat16m1_t vs2, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm(vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -344,7 +344,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm(vbfloat16m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv8i8.nxv8bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm(vbfloat16m2_t vs2, size_t vl) { +vfloat8e4m3m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm(vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -354,7 +354,7 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm(vbfloat16m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv8i8.nxv8bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm(vbfloat16m2_t vs2, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm(vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -365,7 +365,7 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm(vbfloat16m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv16i8.nxv16bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm(vbfloat16m4_t vs2, size_t vl) { +vfloat8e4m3m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm(vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -375,7 +375,7 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm(vbfloat16m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv16i8.nxv16bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm(vbfloat16m4_t vs2, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm(vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -386,7 +386,7 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm(vbfloat16m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv32i8.nxv32bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm(vbfloat16m8_t vs2, size_t vl) { +vfloat8e4m3m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm(vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -396,7 +396,7 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm(vbfloat16m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv32i8.nxv32bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm(vbfloat16m8_t vs2, +vfloat8e4m3m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm(vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -407,7 +407,7 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm(vbfloat16m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv1i8.nxv1bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_m(vbool64_t vm, +vfloat8e4m3mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_m(vbool64_t vm, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_m(vm, vs2, __RISCV_FRM_RNE, @@ -420,7 +420,7 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_m(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv1i8.nxv1bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_m(vbool64_t vm, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_m(vbool64_t vm, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_m(vm, vs2, @@ -433,7 +433,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_m(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv2i8.nxv2bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_m(vbool32_t vm, +vfloat8e4m3mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_m(vbool32_t vm, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_m(vm, vs2, __RISCV_FRM_RNE, @@ -446,7 +446,7 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_m(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv2i8.nxv2bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_m(vbool32_t vm, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_m(vbool32_t vm, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_m(vm, vs2, @@ -459,7 +459,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_m(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv4i8.nxv4bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_m(vbool16_t vm, +vfloat8e4m3mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_m(vbool16_t vm, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_m(vm, vs2, __RISCV_FRM_RNE, @@ -472,7 +472,7 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_m(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv4i8.nxv4bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_m(vbool16_t vm, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_m(vbool16_t vm, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_m(vm, vs2, @@ -485,7 +485,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_m(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv8i8.nxv8bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_m(vbool8_t vm, +vfloat8e4m3m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_m(vbool8_t vm, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_m(vm, vs2, __RISCV_FRM_RNE, @@ -498,7 +498,7 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_m(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv8i8.nxv8bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_m(vbool8_t vm, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_m(vbool8_t vm, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_m(vm, vs2, __RISCV_FRM_RNE, @@ -511,7 +511,7 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_m(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv16i8.nxv16bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_m(vbool4_t vm, +vfloat8e4m3m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_m(vbool4_t vm, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_m(vm, vs2, __RISCV_FRM_RNE, @@ -524,7 +524,7 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_m(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv16i8.nxv16bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_m(vbool4_t vm, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_m(vbool4_t vm, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_m(vm, vs2, __RISCV_FRM_RNE, @@ -537,7 +537,7 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_m(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv32i8.nxv32bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_m(vbool2_t vm, +vfloat8e4m3m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_m(vbool2_t vm, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_m(vm, vs2, __RISCV_FRM_RNE, @@ -550,7 +550,7 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_m(vbool2_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv32i8.nxv32bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_m(vbool2_t vm, +vfloat8e4m3m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_m(vbool2_t vm, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_m(vm, vs2, __RISCV_FRM_RNE, @@ -563,7 +563,7 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_m(vbool2_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv1i8.nxv1bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8(vbfloat16mf4_t vs2, size_t vl) { +vfloat8e5m2mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8(vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf4_f8e5m2mf8(vs2, vl); } @@ -573,7 +573,7 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8(vbfloat16mf4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv1i8.nxv1bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8(vbfloat16mf4_t vs2, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8(vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8(vs2, vl); } @@ -584,7 +584,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8(vbfloat16mf4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv2i8.nxv2bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4(vbfloat16mf2_t vs2, size_t vl) { +vfloat8e5m2mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4(vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf2_f8e5m2mf4(vs2, vl); } @@ -594,7 +594,7 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4(vbfloat16mf2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv2i8.nxv2bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4(vbfloat16mf2_t vs2, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4(vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4(vs2, vl); } @@ -605,7 +605,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4(vbfloat16mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv4i8.nxv4bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2(vbfloat16m1_t vs2, size_t vl) { +vfloat8e5m2mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2(vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m1_f8e5m2mf2(vs2, vl); } @@ -615,7 +615,7 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2(vbfloat16m1_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv4i8.nxv4bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2(vbfloat16m1_t vs2, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2(vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2(vs2, vl); } @@ -626,7 +626,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2(vbfloat16m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv8i8.nxv8bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1(vbfloat16m2_t vs2, size_t vl) { +vfloat8e5m2m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1(vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m2_f8e5m2m1(vs2, vl); } @@ -636,7 +636,7 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1(vbfloat16m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv8i8.nxv8bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1(vbfloat16m2_t vs2, size_t vl) { +vfloat8e5m2m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1(vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1(vs2, vl); } @@ -646,7 +646,7 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1(vbfloat16m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv16i8.nxv16bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2(vbfloat16m4_t vs2, size_t vl) { +vfloat8e5m2m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2(vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m4_f8e5m2m2(vs2, vl); } @@ -656,7 +656,7 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2(vbfloat16m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv16i8.nxv16bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2(vbfloat16m4_t vs2, size_t vl) { +vfloat8e5m2m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2(vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2(vs2, vl); } @@ -666,7 +666,7 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2(vbfloat16m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv32i8.nxv32bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4(vbfloat16m8_t vs2, size_t vl) { +vfloat8e5m2m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4(vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m8_f8e5m2m4(vs2, vl); } @@ -676,7 +676,7 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4(vbfloat16m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv32i8.nxv32bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4(vbfloat16m8_t vs2, size_t vl) { +vfloat8e5m2m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4(vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4(vs2, vl); } @@ -686,7 +686,7 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4(vbfloat16m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_m(vbool64_t vm, +vfloat8e5m2mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_m(vbool64_t vm, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_m(vm, vs2, vl); @@ -698,7 +698,7 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_m(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_m(vbool64_t vm, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_m(vbool64_t vm, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_m(vm, vs2, vl); @@ -710,7 +710,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_m(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_m(vbool32_t vm, +vfloat8e5m2mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_m(vbool32_t vm, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_m(vm, vs2, vl); @@ -722,7 +722,7 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_m(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_m(vbool32_t vm, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_m(vbool32_t vm, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_m(vm, vs2, vl); @@ -734,7 +734,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_m(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_m(vbool16_t vm, +vfloat8e5m2mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_m(vbool16_t vm, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m1_f8e5m2mf2_m(vm, vs2, vl); } @@ -745,7 +745,7 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_m(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_m(vbool16_t vm, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_m(vbool16_t vm, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_m(vm, vs2, vl); @@ -757,7 +757,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_m(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_m(vbool8_t vm, vbfloat16m2_t vs2, +vfloat8e5m2m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_m(vbool8_t vm, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m2_f8e5m2m1_m(vm, vs2, vl); } @@ -768,7 +768,7 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_m(vbool8_t vm, vbfloat16m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_m(vbool8_t vm, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_m(vbool8_t vm, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_m(vm, vs2, vl); @@ -780,7 +780,7 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_m(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_m(vbool4_t vm, vbfloat16m4_t vs2, +vfloat8e5m2m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_m(vbool4_t vm, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m4_f8e5m2m2_m(vm, vs2, vl); } @@ -791,7 +791,7 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_m(vbool4_t vm, vbfloat16m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_m(vbool4_t vm, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_m(vbool4_t vm, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_m(vm, vs2, vl); @@ -803,7 +803,7 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_m(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_m(vbool2_t vm, vbfloat16m8_t vs2, +vfloat8e5m2m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_m(vbool2_t vm, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m8_f8e5m2m4_m(vm, vs2, vl); } @@ -814,7 +814,7 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_m(vbool2_t vm, vbfloat16m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_m(vbool2_t vm, +vfloat8e5m2m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_m(vbool2_t vm, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_m(vm, vs2, vl); @@ -826,7 +826,7 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_m(vbool2_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv1i8.nxv1bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm(vbfloat16mf4_t vs2, +vfloat8e5m2mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm(vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -837,7 +837,7 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm(vbfloat16mf4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv1i8.nxv1bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm(vbfloat16mf4_t vs2, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm(vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm(vs2, __RISCV_FRM_RNE, vl); @@ -849,7 +849,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm(vbfloat16mf4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv2i8.nxv2bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm(vbfloat16mf2_t vs2, +vfloat8e5m2mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm(vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -860,7 +860,7 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm(vbfloat16mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv2i8.nxv2bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm(vbfloat16mf2_t vs2, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm(vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm(vs2, __RISCV_FRM_RNE, vl); @@ -872,7 +872,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm(vbfloat16mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv4i8.nxv4bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm(vbfloat16m1_t vs2, +vfloat8e5m2mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm(vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -883,7 +883,7 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm(vbfloat16m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv4i8.nxv4bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm(vbfloat16m1_t vs2, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm(vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -894,7 +894,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm(vbfloat16m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv8i8.nxv8bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm(vbfloat16m2_t vs2, size_t vl) { +vfloat8e5m2m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm(vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -904,7 +904,7 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm(vbfloat16m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv8i8.nxv8bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm(vbfloat16m2_t vs2, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm(vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -915,7 +915,7 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm(vbfloat16m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv16i8.nxv16bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm(vbfloat16m4_t vs2, size_t vl) { +vfloat8e5m2m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm(vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -925,7 +925,7 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm(vbfloat16m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv16i8.nxv16bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm(vbfloat16m4_t vs2, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm(vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -936,7 +936,7 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm(vbfloat16m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv32i8.nxv32bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm(vbfloat16m8_t vs2, size_t vl) { +vfloat8e5m2m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm(vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -946,7 +946,7 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm(vbfloat16m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv32i8.nxv32bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm(vbfloat16m8_t vs2, +vfloat8e5m2m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm(vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm(vs2, __RISCV_FRM_RNE, vl); } @@ -957,7 +957,7 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm(vbfloat16m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_m(vbool64_t vm, +vfloat8e5m2mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_m(vbool64_t vm, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_m(vm, vs2, __RISCV_FRM_RNE, @@ -970,7 +970,7 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_m(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_m(vbool64_t vm, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_m(vbool64_t vm, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_m(vm, vs2, @@ -983,7 +983,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_m(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_m(vbool32_t vm, +vfloat8e5m2mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_m(vbool32_t vm, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_m(vm, vs2, __RISCV_FRM_RNE, @@ -996,7 +996,7 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_m(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_m(vbool32_t vm, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_m(vbool32_t vm, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_m(vm, vs2, @@ -1009,7 +1009,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_m(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_m(vbool16_t vm, +vfloat8e5m2mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_m(vbool16_t vm, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_m(vm, vs2, __RISCV_FRM_RNE, @@ -1022,7 +1022,7 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_m(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_m(vbool16_t vm, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_m(vbool16_t vm, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_m(vm, vs2, @@ -1035,7 +1035,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_m(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_m(vbool8_t vm, +vfloat8e5m2m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_m(vbool8_t vm, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_m(vm, vs2, __RISCV_FRM_RNE, @@ -1048,7 +1048,7 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_m(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_m(vbool8_t vm, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_m(vbool8_t vm, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_m(vm, vs2, __RISCV_FRM_RNE, @@ -1061,7 +1061,7 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_m(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_m(vbool4_t vm, +vfloat8e5m2m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_m(vbool4_t vm, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_m(vm, vs2, __RISCV_FRM_RNE, @@ -1074,7 +1074,7 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_m(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_m(vbool4_t vm, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_m(vbool4_t vm, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_m(vm, vs2, __RISCV_FRM_RNE, @@ -1087,7 +1087,7 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_m(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_m(vbool2_t vm, +vfloat8e5m2m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_m(vbool2_t vm, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_m(vm, vs2, __RISCV_FRM_RNE, @@ -1100,7 +1100,7 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_m(vbool2_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_m(vbool2_t vm, +vfloat8e5m2m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_m(vbool2_t vm, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_m(vm, vs2, __RISCV_FRM_RNE, diff --git a/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/non-policy/non-overloaded/vfwcvtbf16.c b/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/non-policy/non-overloaded/vfwcvtbf16.c index 168ce81e3936..fb0a67b6f539 100644 --- a/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/non-policy/non-overloaded/vfwcvtbf16.c +++ b/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/non-policy/non-overloaded/vfwcvtbf16.c @@ -13,7 +13,7 @@ // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.nxv1bf16.nxv1i8.i64( poison, [[VS2]], i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4(vuint8mf8_t vs2, size_t vl) { +vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4(vfloat8e4m3mf8_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4(vs2, vl); } @@ -23,7 +23,7 @@ vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4(vuint8mf8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.nxv2bf16.nxv2i8.i64( poison, [[VS2]], i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2(vuint8mf4_t vs2, size_t vl) { +vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2(vfloat8e4m3mf4_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2(vs2, vl); } @@ -33,7 +33,7 @@ vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2(vuint8mf4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.nxv4bf16.nxv4i8.i64( poison, [[VS2]], i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1(vuint8mf2_t vs2, size_t vl) { +vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1(vfloat8e4m3mf2_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3mf2_bf16m1(vs2, vl); } @@ -43,7 +43,7 @@ vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1(vuint8mf2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.nxv8bf16.nxv8i8.i64( poison, [[VS2]], i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2(vuint8m1_t vs2, size_t vl) { +vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2(vfloat8e4m3m1_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3m1_bf16m2(vs2, vl); } @@ -53,7 +53,7 @@ vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2(vuint8m1_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.nxv16bf16.nxv16i8.i64( poison, [[VS2]], i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4(vuint8m2_t vs2, size_t vl) { +vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4(vfloat8e4m3m2_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3m2_bf16m4(vs2, vl); } @@ -63,7 +63,7 @@ vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4(vuint8m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.nxv32bf16.nxv32i8.i64( poison, [[VS2]], i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8(vuint8m4_t vs2, size_t vl) { +vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8(vfloat8e4m3m4_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3m4_bf16m8(vs2, vl); } @@ -74,7 +74,7 @@ vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8(vuint8m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_m(vbool64_t vm, - vuint8mf8_t vs2, + vfloat8e4m3mf8_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_m(vm, vs2, vl); } @@ -86,7 +86,7 @@ vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_m(vbool64_t vm, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_m(vbool32_t vm, - vuint8mf4_t vs2, + vfloat8e4m3mf4_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_m(vm, vs2, vl); } @@ -98,7 +98,7 @@ vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_m(vbool32_t vm, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_m(vbool16_t vm, - vuint8mf2_t vs2, size_t vl) { + vfloat8e4m3mf2_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_m(vm, vs2, vl); } @@ -108,7 +108,7 @@ vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_m(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.mask.nxv8bf16.nxv8i8.i64( poison, [[VS2]], [[VM]], i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_m(vbool8_t vm, vuint8m1_t vs2, +vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_m(vbool8_t vm, vfloat8e4m3m1_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3m1_bf16m2_m(vm, vs2, vl); } @@ -119,7 +119,7 @@ vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_m(vbool8_t vm, vuint8m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.mask.nxv16bf16.nxv16i8.i64( poison, [[VS2]], [[VM]], i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_m(vbool4_t vm, vuint8m2_t vs2, +vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_m(vbool4_t vm, vfloat8e4m3m2_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3m2_bf16m4_m(vm, vs2, vl); } @@ -130,7 +130,7 @@ vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_m(vbool4_t vm, vuint8m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.mask.nxv32bf16.nxv32i8.i64( poison, [[VS2]], [[VM]], i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_m(vbool2_t vm, vuint8m4_t vs2, +vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_m(vbool2_t vm, vfloat8e4m3m4_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3m4_bf16m8_m(vm, vs2, vl); } @@ -141,7 +141,7 @@ vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_m(vbool2_t vm, vuint8m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.alt.nxv1bf16.nxv1i8.i64( poison, [[VS2]], i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4(vuint8mf8_t vs2, size_t vl) { +vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4(vfloat8e5m2mf8_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4(vs2, vl); } @@ -151,7 +151,7 @@ vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4(vuint8mf8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.alt.nxv2bf16.nxv2i8.i64( poison, [[VS2]], i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2(vuint8mf4_t vs2, size_t vl) { +vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2(vfloat8e5m2mf4_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2(vs2, vl); } @@ -161,7 +161,7 @@ vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2(vuint8mf4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.alt.nxv4bf16.nxv4i8.i64( poison, [[VS2]], i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1(vuint8mf2_t vs2, size_t vl) { +vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1(vfloat8e5m2mf2_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2mf2_bf16m1(vs2, vl); } @@ -171,7 +171,7 @@ vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1(vuint8mf2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.alt.nxv8bf16.nxv8i8.i64( poison, [[VS2]], i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2(vuint8m1_t vs2, size_t vl) { +vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2(vfloat8e5m2m1_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2m1_bf16m2(vs2, vl); } @@ -181,7 +181,7 @@ vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2(vuint8m1_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.alt.nxv16bf16.nxv16i8.i64( poison, [[VS2]], i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4(vuint8m2_t vs2, size_t vl) { +vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4(vfloat8e5m2m2_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2m2_bf16m4(vs2, vl); } @@ -191,7 +191,7 @@ vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4(vuint8m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.alt.nxv32bf16.nxv32i8.i64( poison, [[VS2]], i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m8_t test_vfwcvt_f_f_v_f8e5m2m4_bf16m8(vuint8m4_t vs2, size_t vl) { +vbfloat16m8_t test_vfwcvt_f_f_v_f8e5m2m4_bf16m8(vfloat8e5m2m4_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2m4_bf16m8(vs2, vl); } @@ -202,7 +202,7 @@ vbfloat16m8_t test_vfwcvt_f_f_v_f8e5m2m4_bf16m8(vuint8m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_m(vbool64_t vm, - vuint8mf8_t vs2, + vfloat8e5m2mf8_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_m(vm, vs2, vl); } @@ -214,7 +214,7 @@ vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_m(vbool64_t vm, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_m(vbool32_t vm, - vuint8mf4_t vs2, + vfloat8e5m2mf4_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_m(vm, vs2, vl); } @@ -226,7 +226,7 @@ vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_m(vbool32_t vm, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_m(vbool16_t vm, - vuint8mf2_t vs2, size_t vl) { + vfloat8e5m2mf2_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_m(vm, vs2, vl); } @@ -236,7 +236,7 @@ vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_m(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.alt.mask.nxv8bf16.nxv8i8.i64( poison, [[VS2]], [[VM]], i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_m(vbool8_t vm, vuint8m1_t vs2, +vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_m(vbool8_t vm, vfloat8e5m2m1_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2m1_bf16m2_m(vm, vs2, vl); } @@ -247,7 +247,7 @@ vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_m(vbool8_t vm, vuint8m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.alt.mask.nxv16bf16.nxv16i8.i64( poison, [[VS2]], [[VM]], i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_m(vbool4_t vm, vuint8m2_t vs2, +vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_m(vbool4_t vm, vfloat8e5m2m2_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2m2_bf16m4_m(vm, vs2, vl); } @@ -258,7 +258,7 @@ vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_m(vbool4_t vm, vuint8m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.alt.mask.nxv32bf16.nxv32i8.i64( poison, [[VS2]], [[VM]], i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m8_t test_vfwcvt_f_f_v_f8e5m2m4_bf16m8_m(vbool2_t vm, vuint8m4_t vs2, +vbfloat16m8_t test_vfwcvt_f_f_v_f8e5m2m4_bf16m8_m(vbool2_t vm, vfloat8e5m2m4_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2m4_bf16m8_m(vm, vs2, vl); } diff --git a/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/non-policy/overloaded/vfncvt.c b/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/non-policy/overloaded/vfncvt.c index 100a7714bfd5..b511f6325cba 100644 --- a/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/non-policy/overloaded/vfncvt.c +++ b/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/non-policy/overloaded/vfncvt.c @@ -13,7 +13,7 @@ // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv1i8.nxv1f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8(vfloat32mf2_t vs2, size_t vl) { +vfloat8e4m3mf8_t test_vfncvt_f_f_q_f8e4m3mf8(vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3(vs2, vl); } @@ -23,7 +23,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8(vfloat32mf2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv1i8.nxv1f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8(vfloat32mf2_t vs2, size_t vl) { +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8(vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3(vs2, vl); } @@ -33,7 +33,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8(vfloat32mf2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv2i8.nxv2f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4(vfloat32m1_t vs2, size_t vl) { +vfloat8e4m3mf4_t test_vfncvt_f_f_q_f8e4m3mf4(vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3(vs2, vl); } @@ -43,7 +43,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4(vfloat32m1_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv2i8.nxv2f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4(vfloat32m1_t vs2, size_t vl) { +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4(vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3(vs2, vl); } @@ -53,7 +53,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4(vfloat32m1_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv4i8.nxv4f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2(vfloat32m2_t vs2, size_t vl) { +vfloat8e4m3mf2_t test_vfncvt_f_f_q_f8e4m3mf2(vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3(vs2, vl); } @@ -63,7 +63,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2(vfloat32m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv4i8.nxv4f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2(vfloat32m2_t vs2, size_t vl) { +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2(vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3(vs2, vl); } @@ -73,7 +73,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2(vfloat32m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv8i8.nxv8f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1(vfloat32m4_t vs2, size_t vl) { +vfloat8e4m3m1_t test_vfncvt_f_f_q_f8e4m3m1(vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3(vs2, vl); } @@ -83,7 +83,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1(vfloat32m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv8i8.nxv8f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1(vfloat32m4_t vs2, size_t vl) { +vfloat8e4m3m1_t test_vfncvt_sat_f_f_q_f8e4m3m1(vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3(vs2, vl); } @@ -93,7 +93,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1(vfloat32m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv16i8.nxv16f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2(vfloat32m8_t vs2, size_t vl) { +vfloat8e4m3m2_t test_vfncvt_f_f_q_f8e4m3m2(vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3(vs2, vl); } @@ -103,7 +103,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2(vfloat32m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv16i8.nxv16f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2(vfloat32m8_t vs2, size_t vl) { +vfloat8e4m3m2_t test_vfncvt_sat_f_f_q_f8e4m3m2(vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3(vs2, vl); } @@ -113,7 +113,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2(vfloat32m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv1i8.nxv1f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_m(vbool64_t vm, vfloat32mf2_t vs2, +vfloat8e4m3mf8_t test_vfncvt_f_f_q_f8e4m3mf8_m(vbool64_t vm, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3(vm, vs2, vl); } @@ -124,7 +124,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_m(vbool64_t vm, vfloat32mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv1i8.nxv1f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_m(vbool64_t vm, vfloat32mf2_t vs2, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_m(vbool64_t vm, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3(vm, vs2, vl); } @@ -135,7 +135,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_m(vbool64_t vm, vfloat32mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv2i8.nxv2f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_m(vbool32_t vm, vfloat32m1_t vs2, +vfloat8e4m3mf4_t test_vfncvt_f_f_q_f8e4m3mf4_m(vbool32_t vm, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3(vm, vs2, vl); } @@ -146,7 +146,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_m(vbool32_t vm, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv2i8.nxv2f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_m(vbool32_t vm, vfloat32m1_t vs2, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_m(vbool32_t vm, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3(vm, vs2, vl); } @@ -157,7 +157,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_m(vbool32_t vm, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv4i8.nxv4f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_m(vbool16_t vm, vfloat32m2_t vs2, +vfloat8e4m3mf2_t test_vfncvt_f_f_q_f8e4m3mf2_m(vbool16_t vm, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3(vm, vs2, vl); } @@ -168,7 +168,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_m(vbool16_t vm, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv4i8.nxv4f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_m(vbool16_t vm, vfloat32m2_t vs2, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_m(vbool16_t vm, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3(vm, vs2, vl); } @@ -179,7 +179,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_m(vbool16_t vm, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv8i8.nxv8f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_m(vbool8_t vm, vfloat32m4_t vs2, +vfloat8e4m3m1_t test_vfncvt_f_f_q_f8e4m3m1_m(vbool8_t vm, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3(vm, vs2, vl); } @@ -190,7 +190,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_m(vbool8_t vm, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv8i8.nxv8f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_m(vbool8_t vm, vfloat32m4_t vs2, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_m(vbool8_t vm, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3(vm, vs2, vl); } @@ -201,7 +201,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_m(vbool8_t vm, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv16i8.nxv16f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_m(vbool4_t vm, vfloat32m8_t vs2, +vfloat8e4m3m2_t test_vfncvt_f_f_q_f8e4m3m2_m(vbool4_t vm, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3(vm, vs2, vl); } @@ -212,7 +212,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_m(vbool4_t vm, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv16i8.nxv16f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_m(vbool4_t vm, vfloat32m8_t vs2, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_m(vbool4_t vm, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3(vm, vs2, vl); } @@ -223,7 +223,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_m(vbool4_t vm, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv1i8.nxv1f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm(vfloat32mf2_t vs2, size_t vl) { +vfloat8e4m3mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm(vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3(vs2, __RISCV_FRM_RNE, vl); } @@ -233,7 +233,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm(vfloat32mf2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv1i8.nxv1f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm(vfloat32mf2_t vs2, size_t vl) { +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm(vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3(vs2, __RISCV_FRM_RNE, vl); } @@ -243,7 +243,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm(vfloat32mf2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv2i8.nxv2f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm(vfloat32m1_t vs2, size_t vl) { +vfloat8e4m3mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm(vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3(vs2, __RISCV_FRM_RNE, vl); } @@ -253,7 +253,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm(vfloat32m1_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv2i8.nxv2f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm(vfloat32m1_t vs2, size_t vl) { +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm(vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3(vs2, __RISCV_FRM_RNE, vl); } @@ -263,7 +263,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm(vfloat32m1_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv4i8.nxv4f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm(vfloat32m2_t vs2, size_t vl) { +vfloat8e4m3mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm(vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3(vs2, __RISCV_FRM_RNE, vl); } @@ -273,7 +273,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm(vfloat32m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv4i8.nxv4f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm(vfloat32m2_t vs2, size_t vl) { +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm(vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3(vs2, __RISCV_FRM_RNE, vl); } @@ -283,7 +283,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm(vfloat32m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv8i8.nxv8f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_rm(vfloat32m4_t vs2, size_t vl) { +vfloat8e4m3m1_t test_vfncvt_f_f_q_f8e4m3m1_rm(vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3(vs2, __RISCV_FRM_RNE, vl); } @@ -293,7 +293,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_rm(vfloat32m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv8i8.nxv8f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm(vfloat32m4_t vs2, size_t vl) { +vfloat8e4m3m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm(vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3(vs2, __RISCV_FRM_RNE, vl); } @@ -303,7 +303,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm(vfloat32m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv16i8.nxv16f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_rm(vfloat32m8_t vs2, size_t vl) { +vfloat8e4m3m2_t test_vfncvt_f_f_q_f8e4m3m2_rm(vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3(vs2, __RISCV_FRM_RNE, vl); } @@ -313,7 +313,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_rm(vfloat32m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv16i8.nxv16f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm(vfloat32m8_t vs2, size_t vl) { +vfloat8e4m3m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm(vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3(vs2, __RISCV_FRM_RNE, vl); } @@ -323,7 +323,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm(vfloat32m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv1i8.nxv1f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_m(vbool64_t vm, vfloat32mf2_t vs2, +vfloat8e4m3mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_m(vbool64_t vm, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -334,7 +334,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_m(vbool64_t vm, vfloat32mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv1i8.nxv1f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_m(vbool64_t vm, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_m(vbool64_t vm, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -345,7 +345,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_m(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv2i8.nxv2f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_m(vbool32_t vm, vfloat32m1_t vs2, +vfloat8e4m3mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_m(vbool32_t vm, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -356,7 +356,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_m(vbool32_t vm, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv2i8.nxv2f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_m(vbool32_t vm, vfloat32m1_t vs2, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_m(vbool32_t vm, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -367,7 +367,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_m(vbool32_t vm, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv4i8.nxv4f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_m(vbool16_t vm, vfloat32m2_t vs2, +vfloat8e4m3mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_m(vbool16_t vm, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -378,7 +378,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_m(vbool16_t vm, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv4i8.nxv4f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_m(vbool16_t vm, vfloat32m2_t vs2, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_m(vbool16_t vm, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -389,7 +389,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_m(vbool16_t vm, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv8i8.nxv8f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_m(vbool8_t vm, vfloat32m4_t vs2, +vfloat8e4m3m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_m(vbool8_t vm, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -400,7 +400,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_m(vbool8_t vm, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv8i8.nxv8f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_m(vbool8_t vm, vfloat32m4_t vs2, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_m(vbool8_t vm, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -411,7 +411,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_m(vbool8_t vm, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv16i8.nxv16f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_m(vbool4_t vm, vfloat32m8_t vs2, +vfloat8e4m3m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_m(vbool4_t vm, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -422,7 +422,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_m(vbool4_t vm, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv16i8.nxv16f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_m(vbool4_t vm, vfloat32m8_t vs2, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_m(vbool4_t vm, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -433,7 +433,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_m(vbool4_t vm, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv1i8.nxv1f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8(vfloat32mf2_t vs2, size_t vl) { +vfloat8e5m2mf8_t test_vfncvt_f_f_q_f8e5m2mf8(vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2(vs2, vl); } @@ -443,7 +443,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8(vfloat32mf2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv1i8.nxv1f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8(vfloat32mf2_t vs2, size_t vl) { +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8(vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2(vs2, vl); } @@ -453,7 +453,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8(vfloat32mf2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv2i8.nxv2f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4(vfloat32m1_t vs2, size_t vl) { +vfloat8e5m2mf4_t test_vfncvt_f_f_q_f8e5m2mf4(vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2(vs2, vl); } @@ -463,7 +463,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4(vfloat32m1_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv2i8.nxv2f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4(vfloat32m1_t vs2, size_t vl) { +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4(vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2(vs2, vl); } @@ -473,7 +473,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4(vfloat32m1_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv4i8.nxv4f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2(vfloat32m2_t vs2, size_t vl) { +vfloat8e5m2mf2_t test_vfncvt_f_f_q_f8e5m2mf2(vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2(vs2, vl); } @@ -483,7 +483,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2(vfloat32m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv4i8.nxv4f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2(vfloat32m2_t vs2, size_t vl) { +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2(vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2(vs2, vl); } @@ -493,7 +493,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2(vfloat32m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv8i8.nxv8f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1(vfloat32m4_t vs2, size_t vl) { +vfloat8e5m2m1_t test_vfncvt_f_f_q_f8e5m2m1(vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2(vs2, vl); } @@ -503,7 +503,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1(vfloat32m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv8i8.nxv8f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1(vfloat32m4_t vs2, size_t vl) { +vfloat8e5m2m1_t test_vfncvt_sat_f_f_q_f8e5m2m1(vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2(vs2, vl); } @@ -513,7 +513,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1(vfloat32m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv16i8.nxv16f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2(vfloat32m8_t vs2, size_t vl) { +vfloat8e5m2m2_t test_vfncvt_f_f_q_f8e5m2m2(vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2(vs2, vl); } @@ -523,7 +523,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2(vfloat32m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv16i8.nxv16f32.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2(vfloat32m8_t vs2, size_t vl) { +vfloat8e5m2m2_t test_vfncvt_sat_f_f_q_f8e5m2m2(vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2(vs2, vl); } @@ -533,7 +533,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2(vfloat32m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_m(vbool64_t vm, vfloat32mf2_t vs2, +vfloat8e5m2mf8_t test_vfncvt_f_f_q_f8e5m2mf8_m(vbool64_t vm, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2(vm, vs2, vl); } @@ -544,7 +544,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_m(vbool64_t vm, vfloat32mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_m(vbool64_t vm, vfloat32mf2_t vs2, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_m(vbool64_t vm, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2(vm, vs2, vl); } @@ -555,7 +555,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_m(vbool64_t vm, vfloat32mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_m(vbool32_t vm, vfloat32m1_t vs2, +vfloat8e5m2mf4_t test_vfncvt_f_f_q_f8e5m2mf4_m(vbool32_t vm, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2(vm, vs2, vl); } @@ -566,7 +566,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_m(vbool32_t vm, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_m(vbool32_t vm, vfloat32m1_t vs2, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_m(vbool32_t vm, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2(vm, vs2, vl); } @@ -577,7 +577,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_m(vbool32_t vm, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_m(vbool16_t vm, vfloat32m2_t vs2, +vfloat8e5m2mf2_t test_vfncvt_f_f_q_f8e5m2mf2_m(vbool16_t vm, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2(vm, vs2, vl); } @@ -588,7 +588,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_m(vbool16_t vm, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_m(vbool16_t vm, vfloat32m2_t vs2, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_m(vbool16_t vm, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2(vm, vs2, vl); } @@ -599,7 +599,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_m(vbool16_t vm, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_m(vbool8_t vm, vfloat32m4_t vs2, +vfloat8e5m2m1_t test_vfncvt_f_f_q_f8e5m2m1_m(vbool8_t vm, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2(vm, vs2, vl); } @@ -610,7 +610,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_m(vbool8_t vm, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_m(vbool8_t vm, vfloat32m4_t vs2, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_m(vbool8_t vm, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2(vm, vs2, vl); } @@ -621,7 +621,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_m(vbool8_t vm, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_m(vbool4_t vm, vfloat32m8_t vs2, +vfloat8e5m2m2_t test_vfncvt_f_f_q_f8e5m2m2_m(vbool4_t vm, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2(vm, vs2, vl); } @@ -632,7 +632,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_m(vbool4_t vm, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_m(vbool4_t vm, vfloat32m8_t vs2, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_m(vbool4_t vm, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2(vm, vs2, vl); } @@ -643,7 +643,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_m(vbool4_t vm, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv1i8.nxv1f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm(vfloat32mf2_t vs2, size_t vl) { +vfloat8e5m2mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm(vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2(vs2, __RISCV_FRM_RNE, vl); } @@ -653,7 +653,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm(vfloat32mf2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv1i8.nxv1f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm(vfloat32mf2_t vs2, size_t vl) { +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm(vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2(vs2, __RISCV_FRM_RNE, vl); } @@ -663,7 +663,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm(vfloat32mf2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv2i8.nxv2f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm(vfloat32m1_t vs2, size_t vl) { +vfloat8e5m2mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm(vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2(vs2, __RISCV_FRM_RNE, vl); } @@ -673,7 +673,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm(vfloat32m1_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv2i8.nxv2f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm(vfloat32m1_t vs2, size_t vl) { +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm(vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2(vs2, __RISCV_FRM_RNE, vl); } @@ -683,7 +683,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm(vfloat32m1_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv4i8.nxv4f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm(vfloat32m2_t vs2, size_t vl) { +vfloat8e5m2mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm(vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2(vs2, __RISCV_FRM_RNE, vl); } @@ -693,7 +693,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm(vfloat32m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv4i8.nxv4f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm(vfloat32m2_t vs2, size_t vl) { +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm(vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2(vs2, __RISCV_FRM_RNE, vl); } @@ -703,7 +703,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm(vfloat32m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv8i8.nxv8f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_rm(vfloat32m4_t vs2, size_t vl) { +vfloat8e5m2m1_t test_vfncvt_f_f_q_f8e5m2m1_rm(vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2(vs2, __RISCV_FRM_RNE, vl); } @@ -713,7 +713,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_rm(vfloat32m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv8i8.nxv8f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm(vfloat32m4_t vs2, size_t vl) { +vfloat8e5m2m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm(vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2(vs2, __RISCV_FRM_RNE, vl); } @@ -723,7 +723,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm(vfloat32m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv16i8.nxv16f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_rm(vfloat32m8_t vs2, size_t vl) { +vfloat8e5m2m2_t test_vfncvt_f_f_q_f8e5m2m2_rm(vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2(vs2, __RISCV_FRM_RNE, vl); } @@ -733,7 +733,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_rm(vfloat32m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv16i8.nxv16f32.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm(vfloat32m8_t vs2, size_t vl) { +vfloat8e5m2m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm(vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2(vs2, __RISCV_FRM_RNE, vl); } @@ -743,7 +743,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm(vfloat32m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_m(vbool64_t vm, vfloat32mf2_t vs2, +vfloat8e5m2mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_m(vbool64_t vm, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -754,7 +754,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_m(vbool64_t vm, vfloat32mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_m(vbool64_t vm, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_m(vbool64_t vm, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -765,7 +765,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_m(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_m(vbool32_t vm, vfloat32m1_t vs2, +vfloat8e5m2mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_m(vbool32_t vm, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -776,7 +776,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_m(vbool32_t vm, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_m(vbool32_t vm, vfloat32m1_t vs2, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_m(vbool32_t vm, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -787,7 +787,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_m(vbool32_t vm, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_m(vbool16_t vm, vfloat32m2_t vs2, +vfloat8e5m2mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_m(vbool16_t vm, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -798,7 +798,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_m(vbool16_t vm, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_m(vbool16_t vm, vfloat32m2_t vs2, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_m(vbool16_t vm, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -809,7 +809,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_m(vbool16_t vm, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_m(vbool8_t vm, vfloat32m4_t vs2, +vfloat8e5m2m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_m(vbool8_t vm, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -820,7 +820,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_m(vbool8_t vm, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_m(vbool8_t vm, vfloat32m4_t vs2, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_m(vbool8_t vm, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -831,7 +831,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_m(vbool8_t vm, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_m(vbool4_t vm, vfloat32m8_t vs2, +vfloat8e5m2m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_m(vbool4_t vm, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); } @@ -842,7 +842,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_m(vbool4_t vm, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_m(vbool4_t vm, vfloat32m8_t vs2, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_m(vbool4_t vm, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); } diff --git a/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/non-policy/overloaded/vfncvtbf16.c b/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/non-policy/overloaded/vfncvtbf16.c index b13b59b40c1b..5cf3bf6de0f1 100644 --- a/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/non-policy/overloaded/vfncvtbf16.c +++ b/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/non-policy/overloaded/vfncvtbf16.c @@ -13,8 +13,8 @@ // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv1i8.nxv1bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8(vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3(vs2, vl); +vfloat8e4m3mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8(vbfloat16mf4_t vs2, size_t vl) { + return __riscv_vfncvt_f_f8e4m3(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8( @@ -23,9 +23,9 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8(vbfloat16mf4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv1i8.nxv1bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8(vbfloat16mf4_t vs2, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8(vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3(vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4( @@ -34,8 +34,8 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8(vbfloat16mf4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv2i8.nxv2bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4(vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3(vs2, vl); +vfloat8e4m3mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4(vbfloat16mf2_t vs2, size_t vl) { + return __riscv_vfncvt_f_f8e4m3(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4( @@ -44,9 +44,9 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4(vbfloat16mf2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv2i8.nxv2bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4(vbfloat16mf2_t vs2, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4(vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3(vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m1_f8e4m3mf2( @@ -55,8 +55,8 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4(vbfloat16mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv4i8.nxv4bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2(vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3(vs2, vl); +vfloat8e4m3mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2(vbfloat16m1_t vs2, size_t vl) { + return __riscv_vfncvt_f_f8e4m3(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2( @@ -65,9 +65,9 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2(vbfloat16m1_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv4i8.nxv4bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2(vbfloat16m1_t vs2, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2(vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3(vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m2_f8e4m3m1( @@ -76,8 +76,8 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2(vbfloat16m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv8i8.nxv8bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1(vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3(vs2, vl); +vfloat8e4m3m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1(vbfloat16m2_t vs2, size_t vl) { + return __riscv_vfncvt_f_f8e4m3(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1( @@ -86,8 +86,8 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1(vbfloat16m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv8i8.nxv8bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1(vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3(vs2, vl); +vfloat8e4m3m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1(vbfloat16m2_t vs2, size_t vl) { + return __riscv_vfncvt_sat_f_f8e4m3(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m4_f8e4m3m2( @@ -96,8 +96,8 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1(vbfloat16m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv16i8.nxv16bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2(vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3(vs2, vl); +vfloat8e4m3m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2(vbfloat16m4_t vs2, size_t vl) { + return __riscv_vfncvt_f_f8e4m3(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2( @@ -106,8 +106,8 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2(vbfloat16m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv16i8.nxv16bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2(vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3(vs2, vl); +vfloat8e4m3m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2(vbfloat16m4_t vs2, size_t vl) { + return __riscv_vfncvt_sat_f_f8e4m3(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m8_f8e4m3m4( @@ -116,8 +116,8 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2(vbfloat16m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv32i8.nxv32bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4(vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3(vs2, vl); +vfloat8e4m3m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4(vbfloat16m8_t vs2, size_t vl) { + return __riscv_vfncvt_f_f8e4m3(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4( @@ -126,8 +126,8 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4(vbfloat16m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv32i8.nxv32bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4(vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3(vs2, vl); +vfloat8e4m3m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4(vbfloat16m8_t vs2, size_t vl) { + return __riscv_vfncvt_sat_f_f8e4m3(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_m( @@ -136,10 +136,10 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4(vbfloat16m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv1i8.nxv1bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_m(vbool64_t vm, +vfloat8e4m3mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_m(vbool64_t vm, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3(vm, vs2, vl); + return __riscv_vfncvt_f_f8e4m3(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_m( @@ -148,10 +148,10 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_m(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv1i8.nxv1bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_m(vbool64_t vm, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_m(vbool64_t vm, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3(vm, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_m( @@ -160,10 +160,10 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_m(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv2i8.nxv2bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_m(vbool32_t vm, +vfloat8e4m3mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_m(vbool32_t vm, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3(vm, vs2, vl); + return __riscv_vfncvt_f_f8e4m3(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_m( @@ -172,10 +172,10 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_m(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv2i8.nxv2bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_m(vbool32_t vm, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_m(vbool32_t vm, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3(vm, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_m( @@ -184,9 +184,9 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_m(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv4i8.nxv4bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_m(vbool16_t vm, +vfloat8e4m3mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_m(vbool16_t vm, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3(vm, vs2, vl); + return __riscv_vfncvt_f_f8e4m3(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_m( @@ -195,10 +195,10 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_m(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv4i8.nxv4bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_m(vbool16_t vm, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_m(vbool16_t vm, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3(vm, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m2_f8e4m3m1_m( @@ -207,9 +207,9 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_m(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv8i8.nxv8bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_m(vbool8_t vm, vbfloat16m2_t vs2, +vfloat8e4m3m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_m(vbool8_t vm, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3(vm, vs2, vl); + return __riscv_vfncvt_f_f8e4m3(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_m( @@ -218,10 +218,10 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_m(vbool8_t vm, vbfloat16m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv8i8.nxv8bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_m(vbool8_t vm, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_m(vbool8_t vm, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3(vm, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m4_f8e4m3m2_m( @@ -230,9 +230,9 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_m(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv16i8.nxv16bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_m(vbool4_t vm, vbfloat16m4_t vs2, +vfloat8e4m3m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_m(vbool4_t vm, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3(vm, vs2, vl); + return __riscv_vfncvt_f_f8e4m3(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_m( @@ -241,10 +241,10 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_m(vbool4_t vm, vbfloat16m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv16i8.nxv16bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_m(vbool4_t vm, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_m(vbool4_t vm, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3(vm, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m8_f8e4m3m4_m( @@ -253,9 +253,9 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_m(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv32i8.nxv32bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_m(vbool2_t vm, vbfloat16m8_t vs2, +vfloat8e4m3m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_m(vbool2_t vm, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3(vm, vs2, vl); + return __riscv_vfncvt_f_f8e4m3(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_m( @@ -264,10 +264,10 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_m(vbool2_t vm, vbfloat16m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv32i8.nxv32bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_m(vbool2_t vm, +vfloat8e4m3m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_m(vbool2_t vm, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3(vm, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm( @@ -276,9 +276,9 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_m(vbool2_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv1i8.nxv1bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm(vbfloat16mf4_t vs2, +vfloat8e4m3mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm(vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3(vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3(vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm( @@ -287,9 +287,9 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm(vbfloat16mf4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv1i8.nxv1bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm(vbfloat16mf4_t vs2, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm(vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3(vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3(vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm( @@ -298,9 +298,9 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm(vbfloat16mf4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv2i8.nxv2bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm(vbfloat16mf2_t vs2, +vfloat8e4m3mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm(vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3(vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3(vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm( @@ -309,9 +309,9 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm(vbfloat16mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv2i8.nxv2bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm(vbfloat16mf2_t vs2, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm(vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3(vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3(vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm( @@ -320,9 +320,9 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm(vbfloat16mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv4i8.nxv4bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm(vbfloat16m1_t vs2, +vfloat8e4m3mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm(vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3(vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3(vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm( @@ -331,9 +331,9 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm(vbfloat16m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv4i8.nxv4bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm(vbfloat16m1_t vs2, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm(vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3(vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3(vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm( @@ -342,8 +342,8 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm(vbfloat16m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv8i8.nxv8bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm(vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3(vs2, __RISCV_FRM_RNE, vl); +vfloat8e4m3m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm(vbfloat16m2_t vs2, size_t vl) { + return __riscv_vfncvt_f_f8e4m3(vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm( @@ -352,9 +352,9 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm(vbfloat16m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv8i8.nxv8bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm(vbfloat16m2_t vs2, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm(vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3(vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3(vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm( @@ -363,8 +363,8 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm(vbfloat16m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv16i8.nxv16bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm(vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3(vs2, __RISCV_FRM_RNE, vl); +vfloat8e4m3m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm(vbfloat16m4_t vs2, size_t vl) { + return __riscv_vfncvt_f_f8e4m3(vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm( @@ -373,9 +373,9 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm(vbfloat16m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv16i8.nxv16bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm(vbfloat16m4_t vs2, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm(vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3(vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3(vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm( @@ -384,8 +384,8 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm(vbfloat16m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv32i8.nxv32bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm(vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3(vs2, __RISCV_FRM_RNE, vl); +vfloat8e4m3m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm(vbfloat16m8_t vs2, size_t vl) { + return __riscv_vfncvt_f_f8e4m3(vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm( @@ -394,9 +394,9 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm(vbfloat16m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv32i8.nxv32bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm(vbfloat16m8_t vs2, +vfloat8e4m3m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm(vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3(vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3(vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_m( @@ -405,10 +405,10 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm(vbfloat16m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv1i8.nxv1bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_m(vbool64_t vm, +vfloat8e4m3mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_m(vbool64_t vm, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_m( @@ -417,10 +417,10 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_m(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv1i8.nxv1bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_m(vbool64_t vm, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_m(vbool64_t vm, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_m( @@ -429,10 +429,10 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_m(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv2i8.nxv2bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_m(vbool32_t vm, +vfloat8e4m3mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_m(vbool32_t vm, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_m( @@ -441,10 +441,10 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_m(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv2i8.nxv2bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_m(vbool32_t vm, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_m(vbool32_t vm, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_m( @@ -453,10 +453,10 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_m(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv4i8.nxv4bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_m(vbool16_t vm, +vfloat8e4m3mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_m(vbool16_t vm, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_m( @@ -465,10 +465,10 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_m(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv4i8.nxv4bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_m(vbool16_t vm, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_m(vbool16_t vm, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_m( @@ -477,10 +477,10 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_m(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv8i8.nxv8bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_m(vbool8_t vm, +vfloat8e4m3m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_m(vbool8_t vm, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_m( @@ -489,10 +489,10 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_m(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv8i8.nxv8bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_m(vbool8_t vm, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_m(vbool8_t vm, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_m( @@ -501,10 +501,10 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_m(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv16i8.nxv16bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_m(vbool4_t vm, +vfloat8e4m3m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_m(vbool4_t vm, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_m( @@ -513,10 +513,10 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_m(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv16i8.nxv16bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_m(vbool4_t vm, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_m(vbool4_t vm, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_m( @@ -525,10 +525,10 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_m(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv32i8.nxv32bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_m(vbool2_t vm, +vfloat8e4m3m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_m(vbool2_t vm, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_m( @@ -537,10 +537,10 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_m(vbool2_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv32i8.nxv32bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_m(vbool2_t vm, +vfloat8e4m3m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_m(vbool2_t vm, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3(vm, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8( @@ -549,8 +549,8 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_m(vbool2_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv1i8.nxv1bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8(vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2(vs2, vl); +vfloat8e5m2mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8(vbfloat16mf4_t vs2, size_t vl) { + return __riscv_vfncvt_f_f8e5m2(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8( @@ -559,9 +559,9 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8(vbfloat16mf4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv1i8.nxv1bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8(vbfloat16mf4_t vs2, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8(vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2(vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4( @@ -570,8 +570,8 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8(vbfloat16mf4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv2i8.nxv2bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4(vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2(vs2, vl); +vfloat8e5m2mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4(vbfloat16mf2_t vs2, size_t vl) { + return __riscv_vfncvt_f_f8e5m2(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4( @@ -580,9 +580,9 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4(vbfloat16mf2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv2i8.nxv2bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4(vbfloat16mf2_t vs2, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4(vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2(vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m1_f8e5m2mf2( @@ -591,8 +591,8 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4(vbfloat16mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv4i8.nxv4bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2(vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2(vs2, vl); +vfloat8e5m2mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2(vbfloat16m1_t vs2, size_t vl) { + return __riscv_vfncvt_f_f8e5m2(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2( @@ -601,9 +601,9 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2(vbfloat16m1_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv4i8.nxv4bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2(vbfloat16m1_t vs2, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2(vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2(vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m2_f8e5m2m1( @@ -612,8 +612,8 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2(vbfloat16m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv8i8.nxv8bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1(vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2(vs2, vl); +vfloat8e5m2m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1(vbfloat16m2_t vs2, size_t vl) { + return __riscv_vfncvt_f_f8e5m2(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1( @@ -622,8 +622,8 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1(vbfloat16m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv8i8.nxv8bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1(vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2(vs2, vl); +vfloat8e5m2m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1(vbfloat16m2_t vs2, size_t vl) { + return __riscv_vfncvt_sat_f_f8e5m2(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m4_f8e5m2m2( @@ -632,8 +632,8 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1(vbfloat16m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv16i8.nxv16bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2(vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2(vs2, vl); +vfloat8e5m2m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2(vbfloat16m4_t vs2, size_t vl) { + return __riscv_vfncvt_f_f8e5m2(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2( @@ -642,8 +642,8 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2(vbfloat16m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv16i8.nxv16bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2(vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2(vs2, vl); +vfloat8e5m2m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2(vbfloat16m4_t vs2, size_t vl) { + return __riscv_vfncvt_sat_f_f8e5m2(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m8_f8e5m2m4( @@ -652,8 +652,8 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2(vbfloat16m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv32i8.nxv32bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4(vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2(vs2, vl); +vfloat8e5m2m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4(vbfloat16m8_t vs2, size_t vl) { + return __riscv_vfncvt_f_f8e5m2(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4( @@ -662,8 +662,8 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4(vbfloat16m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv32i8.nxv32bf16.i64( poison, [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4(vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2(vs2, vl); +vfloat8e5m2m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4(vbfloat16m8_t vs2, size_t vl) { + return __riscv_vfncvt_sat_f_f8e5m2(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_m( @@ -672,10 +672,10 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4(vbfloat16m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_m(vbool64_t vm, +vfloat8e5m2mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_m(vbool64_t vm, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2(vm, vs2, vl); + return __riscv_vfncvt_f_f8e5m2(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_m( @@ -684,10 +684,10 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_m(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_m(vbool64_t vm, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_m(vbool64_t vm, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2(vm, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_m( @@ -696,10 +696,10 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_m(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_m(vbool32_t vm, +vfloat8e5m2mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_m(vbool32_t vm, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2(vm, vs2, vl); + return __riscv_vfncvt_f_f8e5m2(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_m( @@ -708,10 +708,10 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_m(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_m(vbool32_t vm, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_m(vbool32_t vm, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2(vm, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_m( @@ -720,9 +720,9 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_m(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_m(vbool16_t vm, +vfloat8e5m2mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_m(vbool16_t vm, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2(vm, vs2, vl); + return __riscv_vfncvt_f_f8e5m2(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_m( @@ -731,10 +731,10 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_m(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_m(vbool16_t vm, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_m(vbool16_t vm, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2(vm, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m2_f8e5m2m1_m( @@ -743,9 +743,9 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_m(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_m(vbool8_t vm, vbfloat16m2_t vs2, +vfloat8e5m2m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_m(vbool8_t vm, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2(vm, vs2, vl); + return __riscv_vfncvt_f_f8e5m2(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_m( @@ -754,10 +754,10 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_m(vbool8_t vm, vbfloat16m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_m(vbool8_t vm, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_m(vbool8_t vm, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2(vm, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m4_f8e5m2m2_m( @@ -766,9 +766,9 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_m(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_m(vbool4_t vm, vbfloat16m4_t vs2, +vfloat8e5m2m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_m(vbool4_t vm, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2(vm, vs2, vl); + return __riscv_vfncvt_f_f8e5m2(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_m( @@ -777,10 +777,10 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_m(vbool4_t vm, vbfloat16m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_m(vbool4_t vm, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_m(vbool4_t vm, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2(vm, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m8_f8e5m2m4_m( @@ -789,9 +789,9 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_m(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_m(vbool2_t vm, vbfloat16m8_t vs2, +vfloat8e5m2m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_m(vbool2_t vm, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2(vm, vs2, vl); + return __riscv_vfncvt_f_f8e5m2(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_m( @@ -800,10 +800,10 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_m(vbool2_t vm, vbfloat16m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( poison, [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_m(vbool2_t vm, +vfloat8e5m2m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_m(vbool2_t vm, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2(vm, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm( @@ -812,9 +812,9 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_m(vbool2_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv1i8.nxv1bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm(vbfloat16mf4_t vs2, +vfloat8e5m2mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm(vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2(vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2(vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm( @@ -823,9 +823,9 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm(vbfloat16mf4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv1i8.nxv1bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm(vbfloat16mf4_t vs2, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm(vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2(vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2(vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm( @@ -834,9 +834,9 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm(vbfloat16mf4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv2i8.nxv2bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm(vbfloat16mf2_t vs2, +vfloat8e5m2mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm(vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2(vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2(vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm( @@ -845,9 +845,9 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm(vbfloat16mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv2i8.nxv2bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm(vbfloat16mf2_t vs2, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm(vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2(vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2(vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm( @@ -856,9 +856,9 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm(vbfloat16mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv4i8.nxv4bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm(vbfloat16m1_t vs2, +vfloat8e5m2mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm(vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2(vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2(vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm( @@ -867,9 +867,9 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm(vbfloat16m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv4i8.nxv4bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm(vbfloat16m1_t vs2, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm(vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2(vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2(vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm( @@ -878,8 +878,8 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm(vbfloat16m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv8i8.nxv8bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm(vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2(vs2, __RISCV_FRM_RNE, vl); +vfloat8e5m2m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm(vbfloat16m2_t vs2, size_t vl) { + return __riscv_vfncvt_f_f8e5m2(vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm( @@ -888,9 +888,9 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm(vbfloat16m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv8i8.nxv8bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm(vbfloat16m2_t vs2, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm(vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2(vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2(vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm( @@ -899,8 +899,8 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm(vbfloat16m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv16i8.nxv16bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm(vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2(vs2, __RISCV_FRM_RNE, vl); +vfloat8e5m2m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm(vbfloat16m4_t vs2, size_t vl) { + return __riscv_vfncvt_f_f8e5m2(vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm( @@ -909,9 +909,9 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm(vbfloat16m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv16i8.nxv16bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm(vbfloat16m4_t vs2, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm(vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2(vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2(vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm( @@ -920,8 +920,8 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm(vbfloat16m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv32i8.nxv32bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm(vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2(vs2, __RISCV_FRM_RNE, vl); +vfloat8e5m2m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm(vbfloat16m8_t vs2, size_t vl) { + return __riscv_vfncvt_f_f8e5m2(vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm( @@ -930,9 +930,9 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm(vbfloat16m8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv32i8.nxv32bf16.i64( poison, [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm(vbfloat16m8_t vs2, +vfloat8e5m2m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm(vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2(vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2(vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_m( @@ -941,10 +941,10 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm(vbfloat16m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_m(vbool64_t vm, +vfloat8e5m2mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_m(vbool64_t vm, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_m( @@ -953,10 +953,10 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_m(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_m(vbool64_t vm, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_m(vbool64_t vm, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_m( @@ -965,10 +965,10 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_m(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_m(vbool32_t vm, +vfloat8e5m2mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_m(vbool32_t vm, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_m( @@ -977,10 +977,10 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_m(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_m(vbool32_t vm, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_m(vbool32_t vm, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_m( @@ -989,10 +989,10 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_m(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_m(vbool16_t vm, +vfloat8e5m2mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_m(vbool16_t vm, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_m( @@ -1001,10 +1001,10 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_m(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_m(vbool16_t vm, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_m(vbool16_t vm, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_m( @@ -1013,10 +1013,10 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_m(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_m(vbool8_t vm, +vfloat8e5m2m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_m(vbool8_t vm, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_m( @@ -1025,10 +1025,10 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_m(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_m(vbool8_t vm, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_m(vbool8_t vm, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_m( @@ -1037,10 +1037,10 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_m(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_m(vbool4_t vm, +vfloat8e5m2m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_m(vbool4_t vm, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_m( @@ -1049,10 +1049,10 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_m(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_m(vbool4_t vm, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_m(vbool4_t vm, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_m( @@ -1061,10 +1061,10 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_m(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_m(vbool2_t vm, +vfloat8e5m2m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_m(vbool2_t vm, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_m( @@ -1073,8 +1073,8 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_m(vbool2_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( poison, [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_m(vbool2_t vm, +vfloat8e5m2m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_m(vbool2_t vm, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2(vm, vs2, __RISCV_FRM_RNE, vl); } diff --git a/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/non-policy/overloaded/vfwcvtbf16.c b/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/non-policy/overloaded/vfwcvtbf16.c index e909d8b3bd1a..ad00532931d8 100644 --- a/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/non-policy/overloaded/vfwcvtbf16.c +++ b/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/non-policy/overloaded/vfwcvtbf16.c @@ -13,8 +13,8 @@ // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.nxv1bf16.nxv1i8.i64( poison, [[VS2]], i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4(vuint8mf8_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16(vs2, vl); +vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4(vfloat8e4m3mf8_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2( @@ -23,8 +23,8 @@ vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4(vuint8mf8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.nxv2bf16.nxv2i8.i64( poison, [[VS2]], i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2(vuint8mf4_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16(vs2, vl); +vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2(vfloat8e4m3mf4_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1( @@ -33,8 +33,8 @@ vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2(vuint8mf4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.nxv4bf16.nxv4i8.i64( poison, [[VS2]], i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1(vuint8mf2_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16(vs2, vl); +vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1(vfloat8e4m3mf2_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3m1_bf16m2( @@ -43,8 +43,8 @@ vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1(vuint8mf2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.nxv8bf16.nxv8i8.i64( poison, [[VS2]], i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2(vuint8m1_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16(vs2, vl); +vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2(vfloat8e4m3m1_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3m2_bf16m4( @@ -53,8 +53,8 @@ vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2(vuint8m1_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.nxv16bf16.nxv16i8.i64( poison, [[VS2]], i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4(vuint8m2_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16(vs2, vl); +vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4(vfloat8e4m3m2_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3m4_bf16m8( @@ -63,8 +63,8 @@ vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4(vuint8m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.nxv32bf16.nxv32i8.i64( poison, [[VS2]], i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8(vuint8m4_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16(vs2, vl); +vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8(vfloat8e4m3m4_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_m( @@ -74,9 +74,9 @@ vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8(vuint8m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_m(vbool64_t vm, - vuint8mf8_t vs2, + vfloat8e4m3mf8_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16(vm, vs2, vl); + return __riscv_vfwcvt_f_bf16(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_m( @@ -86,9 +86,9 @@ vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_m(vbool64_t vm, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_m(vbool32_t vm, - vuint8mf4_t vs2, + vfloat8e4m3mf4_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16(vm, vs2, vl); + return __riscv_vfwcvt_f_bf16(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_m( @@ -98,8 +98,8 @@ vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_m(vbool32_t vm, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_m(vbool16_t vm, - vuint8mf2_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16(vm, vs2, vl); + vfloat8e4m3mf2_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_m( @@ -108,9 +108,9 @@ vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_m(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.mask.nxv8bf16.nxv8i8.i64( poison, [[VS2]], [[VM]], i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_m(vbool8_t vm, vuint8m1_t vs2, +vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_m(vbool8_t vm, vfloat8e4m3m1_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16(vm, vs2, vl); + return __riscv_vfwcvt_f_bf16(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_m( @@ -119,9 +119,9 @@ vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_m(vbool8_t vm, vuint8m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.mask.nxv16bf16.nxv16i8.i64( poison, [[VS2]], [[VM]], i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_m(vbool4_t vm, vuint8m2_t vs2, +vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_m(vbool4_t vm, vfloat8e4m3m2_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16(vm, vs2, vl); + return __riscv_vfwcvt_f_bf16(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_m( @@ -130,9 +130,9 @@ vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_m(vbool4_t vm, vuint8m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.mask.nxv32bf16.nxv32i8.i64( poison, [[VS2]], [[VM]], i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_m(vbool2_t vm, vuint8m4_t vs2, +vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_m(vbool2_t vm, vfloat8e4m3m4_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16(vm, vs2, vl); + return __riscv_vfwcvt_f_bf16(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4( @@ -141,8 +141,8 @@ vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_m(vbool2_t vm, vuint8m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.alt.nxv1bf16.nxv1i8.i64( poison, [[VS2]], i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4(vuint8mf8_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16(vs2, vl); +vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4(vfloat8e5m2mf8_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2( @@ -151,8 +151,8 @@ vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4(vuint8mf8_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.alt.nxv2bf16.nxv2i8.i64( poison, [[VS2]], i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2(vuint8mf4_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16(vs2, vl); +vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2(vfloat8e5m2mf4_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1( @@ -161,8 +161,8 @@ vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2(vuint8mf4_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.alt.nxv4bf16.nxv4i8.i64( poison, [[VS2]], i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1(vuint8mf2_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16(vs2, vl); +vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1(vfloat8e5m2mf2_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2m1_bf16m2( @@ -171,8 +171,8 @@ vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1(vuint8mf2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.alt.nxv8bf16.nxv8i8.i64( poison, [[VS2]], i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2(vuint8m1_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16(vs2, vl); +vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2(vfloat8e5m2m1_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2m2_bf16m4( @@ -181,8 +181,8 @@ vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2(vuint8m1_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.alt.nxv16bf16.nxv16i8.i64( poison, [[VS2]], i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4(vuint8m2_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16(vs2, vl); +vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4(vfloat8e5m2m2_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2m4_bf16m8( @@ -191,8 +191,8 @@ vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4(vuint8m2_t vs2, size_t vl) { // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.alt.nxv32bf16.nxv32i8.i64( poison, [[VS2]], i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m8_t test_vfwcvt_f_f_v_f8e5m2m4_bf16m8(vuint8m4_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16(vs2, vl); +vbfloat16m8_t test_vfwcvt_f_f_v_f8e5m2m4_bf16m8(vfloat8e5m2m4_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16(vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_m( @@ -202,9 +202,9 @@ vbfloat16m8_t test_vfwcvt_f_f_v_f8e5m2m4_bf16m8(vuint8m4_t vs2, size_t vl) { // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_m(vbool64_t vm, - vuint8mf8_t vs2, + vfloat8e5m2mf8_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16(vm, vs2, vl); + return __riscv_vfwcvt_f_bf16(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_m( @@ -214,9 +214,9 @@ vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_m(vbool64_t vm, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_m(vbool32_t vm, - vuint8mf4_t vs2, + vfloat8e5m2mf4_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16(vm, vs2, vl); + return __riscv_vfwcvt_f_bf16(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_m( @@ -226,8 +226,8 @@ vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_m(vbool32_t vm, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_m(vbool16_t vm, - vuint8mf2_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16(vm, vs2, vl); + vfloat8e5m2mf2_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_m( @@ -236,9 +236,9 @@ vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_m(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.alt.mask.nxv8bf16.nxv8i8.i64( poison, [[VS2]], [[VM]], i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_m(vbool8_t vm, vuint8m1_t vs2, +vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_m(vbool8_t vm, vfloat8e5m2m1_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16(vm, vs2, vl); + return __riscv_vfwcvt_f_bf16(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_m( @@ -247,9 +247,9 @@ vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_m(vbool8_t vm, vuint8m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.alt.mask.nxv16bf16.nxv16i8.i64( poison, [[VS2]], [[VM]], i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_m(vbool4_t vm, vuint8m2_t vs2, +vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_m(vbool4_t vm, vfloat8e5m2m2_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16(vm, vs2, vl); + return __riscv_vfwcvt_f_bf16(vm, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2m4_bf16m8_m( @@ -258,7 +258,7 @@ vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_m(vbool4_t vm, vuint8m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfwcvt.f.f.v.alt.mask.nxv32bf16.nxv32i8.i64( poison, [[VS2]], [[VM]], i64 [[VL]], i64 3) // CHECK-RV64-NEXT: ret [[TMP0]] // -vbfloat16m8_t test_vfwcvt_f_f_v_f8e5m2m4_bf16m8_m(vbool2_t vm, vuint8m4_t vs2, +vbfloat16m8_t test_vfwcvt_f_f_v_f8e5m2m4_bf16m8_m(vbool2_t vm, vfloat8e5m2m4_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16(vm, vs2, vl); + return __riscv_vfwcvt_f_bf16(vm, vs2, vl); } diff --git a/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/policy/non-overloaded/vfncvt.c b/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/policy/non-overloaded/vfncvt.c index c173502415b5..2341db180df0 100644 --- a/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/policy/non-overloaded/vfncvt.c +++ b/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/policy/non-overloaded/vfncvt.c @@ -13,7 +13,7 @@ // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_tu(vuint8mf8_t vd, vfloat32mf2_t vs2, +vfloat8e4m3mf8_t test_vfncvt_f_f_q_f8e4m3mf8_tu(vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf8_tu(vd, vs2, vl); } @@ -24,7 +24,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_tu(vuint8mf8_t vd, vfloat32mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_tu(vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_tu(vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf8_tu(vd, vs2, vl); } @@ -35,7 +35,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_tu(vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_tu(vuint8mf4_t vd, vfloat32m1_t vs2, +vfloat8e4m3mf4_t test_vfncvt_f_f_q_f8e4m3mf4_tu(vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf4_tu(vd, vs2, vl); } @@ -46,7 +46,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_tu(vuint8mf4_t vd, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_tu(vuint8mf4_t vd, vfloat32m1_t vs2, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_tu(vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf4_tu(vd, vs2, vl); } @@ -57,7 +57,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_tu(vuint8mf4_t vd, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_tu(vuint8mf2_t vd, vfloat32m2_t vs2, +vfloat8e4m3mf2_t test_vfncvt_f_f_q_f8e4m3mf2_tu(vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf2_tu(vd, vs2, vl); } @@ -68,7 +68,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_tu(vuint8mf2_t vd, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_tu(vuint8mf2_t vd, vfloat32m2_t vs2, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_tu(vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf2_tu(vd, vs2, vl); } @@ -79,7 +79,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_tu(vuint8mf2_t vd, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_tu(vuint8m1_t vd, vfloat32m4_t vs2, +vfloat8e4m3m1_t test_vfncvt_f_f_q_f8e4m3m1_tu(vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3m1_tu(vd, vs2, vl); } @@ -90,7 +90,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_tu(vuint8m1_t vd, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_tu(vuint8m1_t vd, vfloat32m4_t vs2, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_tu(vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3m1_tu(vd, vs2, vl); } @@ -101,7 +101,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_tu(vuint8m1_t vd, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_tu(vuint8m2_t vd, vfloat32m8_t vs2, +vfloat8e4m3m2_t test_vfncvt_f_f_q_f8e4m3m2_tu(vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3m2_tu(vd, vs2, vl); } @@ -112,7 +112,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_tu(vuint8m2_t vd, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_tu(vuint8m2_t vd, vfloat32m8_t vs2, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_tu(vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3m2_tu(vd, vs2, vl); } @@ -123,7 +123,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_tu(vuint8m2_t vd, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_tum(vbool64_t vm, vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_q_f8e4m3mf8_tum(vbool64_t vm, vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf8_tum(vm, vd, vs2, vl); } @@ -134,7 +134,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_tum(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_tum(vbool64_t vm, vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_tum(vbool64_t vm, vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf8_tum(vm, vd, vs2, vl); } @@ -145,7 +145,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_tum(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_tum(vbool32_t vm, vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_q_f8e4m3mf4_tum(vbool32_t vm, vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf4_tum(vm, vd, vs2, vl); } @@ -156,7 +156,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_tum(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_tum(vbool32_t vm, vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_tum(vbool32_t vm, vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf4_tum(vm, vd, vs2, vl); } @@ -167,7 +167,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_tum(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_tum(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_q_f8e4m3mf2_tum(vbool16_t vm, vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf2_tum(vm, vd, vs2, vl); } @@ -178,7 +178,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_tum(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_tum(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_tum(vbool16_t vm, vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf2_tum(vm, vd, vs2, vl); } @@ -189,7 +189,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_tum(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_q_f8e4m3m1_tum(vbool8_t vm, vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3m1_tum(vm, vd, vs2, vl); } @@ -200,7 +200,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_tum(vbool8_t vm, vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3m1_tum(vm, vd, vs2, vl); } @@ -211,7 +211,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_q_f8e4m3m2_tum(vbool4_t vm, vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3m2_tum(vm, vd, vs2, vl); } @@ -222,7 +222,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_tum(vbool4_t vm, vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3m2_tum(vm, vd, vs2, vl); } @@ -233,7 +233,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_tumu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_q_f8e4m3mf8_tumu(vbool64_t vm, vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf8_tumu(vm, vd, vs2, vl); } @@ -244,7 +244,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_tumu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_tumu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_tumu(vbool64_t vm, vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf8_tumu(vm, vd, vs2, vl); } @@ -255,7 +255,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_tumu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_tumu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_q_f8e4m3mf4_tumu(vbool32_t vm, vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf4_tumu(vm, vd, vs2, vl); } @@ -266,7 +266,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_tumu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_tumu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_tumu(vbool32_t vm, vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf4_tumu(vm, vd, vs2, vl); } @@ -277,7 +277,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_tumu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_tumu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_q_f8e4m3mf2_tumu(vbool16_t vm, vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf2_tumu(vm, vd, vs2, vl); } @@ -288,7 +288,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_tumu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_tumu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_tumu(vbool16_t vm, vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf2_tumu(vm, vd, vs2, vl); } @@ -299,7 +299,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_tumu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_tumu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_q_f8e4m3m1_tumu(vbool8_t vm, vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3m1_tumu(vm, vd, vs2, vl); } @@ -310,7 +310,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_tumu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_tumu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_tumu(vbool8_t vm, vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3m1_tumu(vm, vd, vs2, vl); } @@ -321,7 +321,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_tumu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_tumu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_q_f8e4m3m2_tumu(vbool4_t vm, vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3m2_tumu(vm, vd, vs2, vl); } @@ -332,7 +332,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_tumu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_tumu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_tumu(vbool4_t vm, vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3m2_tumu(vm, vd, vs2, vl); } @@ -343,7 +343,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_tumu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_mu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_q_f8e4m3mf8_mu(vbool64_t vm, vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf8_mu(vm, vd, vs2, vl); } @@ -354,7 +354,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_mu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_mu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_mu(vbool64_t vm, vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf8_mu(vm, vd, vs2, vl); } @@ -365,7 +365,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_mu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_mu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_q_f8e4m3mf4_mu(vbool32_t vm, vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf4_mu(vm, vd, vs2, vl); } @@ -376,7 +376,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_mu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_mu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_mu(vbool32_t vm, vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf4_mu(vm, vd, vs2, vl); } @@ -387,7 +387,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_mu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_mu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_q_f8e4m3mf2_mu(vbool16_t vm, vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf2_mu(vm, vd, vs2, vl); } @@ -398,7 +398,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_mu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_mu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_mu(vbool16_t vm, vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf2_mu(vm, vd, vs2, vl); } @@ -409,7 +409,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_mu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_q_f8e4m3m1_mu(vbool8_t vm, vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3m1_mu(vm, vd, vs2, vl); } @@ -420,7 +420,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_mu(vbool8_t vm, vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3m1_mu(vm, vd, vs2, vl); } @@ -431,7 +431,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_q_f8e4m3m2_mu(vbool4_t vm, vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3m2_mu(vm, vd, vs2, vl); } @@ -442,7 +442,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_mu(vbool4_t vm, vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3m2_mu(vm, vd, vs2, vl); } @@ -453,7 +453,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_tu(vuint8mf8_t vd, vfloat32mf2_t vs2, +vfloat8e4m3mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_tu(vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf8_rm_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -464,7 +464,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_tu(vuint8mf8_t vd, vfloat32mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_tu(vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_tu(vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf8_rm_tu(vd, vs2, __RISCV_FRM_RNE, vl); @@ -476,7 +476,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_tu(vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_tu(vuint8mf4_t vd, vfloat32m1_t vs2, +vfloat8e4m3mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_tu(vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf4_rm_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -487,7 +487,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_tu(vuint8mf4_t vd, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_tu(vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_tu(vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf4_rm_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -498,7 +498,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_tu(vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_tu(vuint8mf2_t vd, vfloat32m2_t vs2, +vfloat8e4m3mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_tu(vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf2_rm_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -509,7 +509,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_tu(vuint8mf2_t vd, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_tu(vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_tu(vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf2_rm_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -520,7 +520,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_tu(vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_tu(vuint8m1_t vd, vfloat32m4_t vs2, +vfloat8e4m3m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_tu(vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3m1_rm_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -531,7 +531,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_tu(vuint8m1_t vd, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_tu(vuint8m1_t vd, vfloat32m4_t vs2, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_tu(vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3m1_rm_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -542,7 +542,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_tu(vuint8m1_t vd, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_tu(vuint8m2_t vd, vfloat32m8_t vs2, +vfloat8e4m3m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_tu(vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3m2_rm_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -553,7 +553,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_tu(vuint8m2_t vd, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_tu(vuint8m2_t vd, vfloat32m8_t vs2, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_tu(vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3m2_rm_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -564,7 +564,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_tu(vuint8m2_t vd, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_tum(vbool64_t vm, vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_tum(vbool64_t vm, vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf8_rm_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -576,7 +576,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_tum(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_tum(vbool64_t vm, vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_tum(vbool64_t vm, vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf8_rm_tum(vm, vd, vs2, __RISCV_FRM_RNE, @@ -589,7 +589,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_tum(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_tum(vbool32_t vm, vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_tum(vbool32_t vm, vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf4_rm_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -601,7 +601,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_tum(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_tum(vbool32_t vm, vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_tum(vbool32_t vm, vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf4_rm_tum(vm, vd, vs2, __RISCV_FRM_RNE, @@ -614,7 +614,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_tum(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_tum(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_tum(vbool16_t vm, vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf2_rm_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -626,7 +626,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_tum(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_tum(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_tum(vbool16_t vm, vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf2_rm_tum(vm, vd, vs2, __RISCV_FRM_RNE, @@ -639,7 +639,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_tum(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_tum(vbool8_t vm, vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3m1_rm_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -650,7 +650,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_tum(vbool8_t vm, vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3m1_rm_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -662,7 +662,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_tum(vbool4_t vm, vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3m2_rm_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -673,7 +673,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_tum(vbool4_t vm, vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3m2_rm_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -685,7 +685,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_tumu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_tumu(vbool64_t vm, vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf8_rm_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -697,8 +697,8 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_tumu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_tumu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_tumu(vbool64_t vm, + vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf8_rm_tumu(vm, vd, vs2, @@ -711,7 +711,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_tumu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_tumu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_tumu(vbool32_t vm, vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf4_rm_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -723,8 +723,8 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_tumu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_tumu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_tumu(vbool32_t vm, + vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf4_rm_tumu(vm, vd, vs2, @@ -737,7 +737,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_tumu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_tumu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_tumu(vbool16_t vm, vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf2_rm_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -749,8 +749,8 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_tumu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_tumu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_tumu(vbool16_t vm, + vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf2_rm_tumu(vm, vd, vs2, @@ -763,7 +763,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_tumu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_tumu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_tumu(vbool8_t vm, vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3m1_rm_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -775,7 +775,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_tumu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_tumu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_tumu(vbool8_t vm, vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3m1_rm_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -787,7 +787,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_tumu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_tumu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_tumu(vbool4_t vm, vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3m2_rm_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -799,7 +799,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_tumu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_tumu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_tumu(vbool4_t vm, vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3m2_rm_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -811,7 +811,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_tumu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_mu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_mu(vbool64_t vm, vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf8_rm_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -822,7 +822,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_mu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_mu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_mu(vbool64_t vm, vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf8_rm_mu(vm, vd, vs2, __RISCV_FRM_RNE, @@ -835,7 +835,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_mu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_mu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_mu(vbool32_t vm, vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf4_rm_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -846,7 +846,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_mu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_mu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_mu(vbool32_t vm, vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf4_rm_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -858,7 +858,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_mu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_mu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_mu(vbool16_t vm, vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3mf2_rm_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -869,7 +869,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_mu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_mu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_mu(vbool16_t vm, vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3mf2_rm_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -881,7 +881,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_mu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_mu(vbool8_t vm, vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3m1_rm_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -892,7 +892,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_mu(vbool8_t vm, vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3m1_rm_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -904,7 +904,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_mu(vbool4_t vm, vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e4m3m2_rm_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -915,7 +915,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_mu(vbool4_t vm, vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e4m3m2_rm_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -927,7 +927,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_tu(vuint8mf8_t vd, vfloat32mf2_t vs2, +vfloat8e5m2mf8_t test_vfncvt_f_f_q_f8e5m2mf8_tu(vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf8_tu(vd, vs2, vl); } @@ -938,7 +938,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_tu(vuint8mf8_t vd, vfloat32mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_tu(vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_tu(vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf8_tu(vd, vs2, vl); } @@ -949,7 +949,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_tu(vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_tu(vuint8mf4_t vd, vfloat32m1_t vs2, +vfloat8e5m2mf4_t test_vfncvt_f_f_q_f8e5m2mf4_tu(vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf4_tu(vd, vs2, vl); } @@ -960,7 +960,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_tu(vuint8mf4_t vd, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_tu(vuint8mf4_t vd, vfloat32m1_t vs2, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_tu(vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf4_tu(vd, vs2, vl); } @@ -971,7 +971,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_tu(vuint8mf4_t vd, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_tu(vuint8mf2_t vd, vfloat32m2_t vs2, +vfloat8e5m2mf2_t test_vfncvt_f_f_q_f8e5m2mf2_tu(vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf2_tu(vd, vs2, vl); } @@ -982,7 +982,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_tu(vuint8mf2_t vd, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_tu(vuint8mf2_t vd, vfloat32m2_t vs2, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_tu(vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf2_tu(vd, vs2, vl); } @@ -993,7 +993,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_tu(vuint8mf2_t vd, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_tu(vuint8m1_t vd, vfloat32m4_t vs2, +vfloat8e5m2m1_t test_vfncvt_f_f_q_f8e5m2m1_tu(vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2m1_tu(vd, vs2, vl); } @@ -1004,7 +1004,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_tu(vuint8m1_t vd, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_tu(vuint8m1_t vd, vfloat32m4_t vs2, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_tu(vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2m1_tu(vd, vs2, vl); } @@ -1015,7 +1015,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_tu(vuint8m1_t vd, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_tu(vuint8m2_t vd, vfloat32m8_t vs2, +vfloat8e5m2m2_t test_vfncvt_f_f_q_f8e5m2m2_tu(vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2m2_tu(vd, vs2, vl); } @@ -1026,7 +1026,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_tu(vuint8m2_t vd, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_tu(vuint8m2_t vd, vfloat32m8_t vs2, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_tu(vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2m2_tu(vd, vs2, vl); } @@ -1037,7 +1037,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_tu(vuint8m2_t vd, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_tum(vbool64_t vm, vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_q_f8e5m2mf8_tum(vbool64_t vm, vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf8_tum(vm, vd, vs2, vl); } @@ -1048,7 +1048,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_tum(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_tum(vbool64_t vm, vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_tum(vbool64_t vm, vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf8_tum(vm, vd, vs2, vl); } @@ -1059,7 +1059,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_tum(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_tum(vbool32_t vm, vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_q_f8e5m2mf4_tum(vbool32_t vm, vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf4_tum(vm, vd, vs2, vl); } @@ -1070,7 +1070,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_tum(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_tum(vbool32_t vm, vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_tum(vbool32_t vm, vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf4_tum(vm, vd, vs2, vl); } @@ -1081,7 +1081,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_tum(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_tum(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_q_f8e5m2mf2_tum(vbool16_t vm, vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf2_tum(vm, vd, vs2, vl); } @@ -1092,7 +1092,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_tum(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_tum(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_tum(vbool16_t vm, vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf2_tum(vm, vd, vs2, vl); } @@ -1103,7 +1103,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_tum(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_q_f8e5m2m1_tum(vbool8_t vm, vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2m1_tum(vm, vd, vs2, vl); } @@ -1114,7 +1114,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_tum(vbool8_t vm, vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2m1_tum(vm, vd, vs2, vl); } @@ -1125,7 +1125,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_q_f8e5m2m2_tum(vbool4_t vm, vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2m2_tum(vm, vd, vs2, vl); } @@ -1136,7 +1136,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_tum(vbool4_t vm, vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2m2_tum(vm, vd, vs2, vl); } @@ -1147,7 +1147,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_tumu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_q_f8e5m2mf8_tumu(vbool64_t vm, vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf8_tumu(vm, vd, vs2, vl); } @@ -1158,7 +1158,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_tumu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_tumu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_tumu(vbool64_t vm, vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf8_tumu(vm, vd, vs2, vl); } @@ -1169,7 +1169,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_tumu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_tumu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_q_f8e5m2mf4_tumu(vbool32_t vm, vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf4_tumu(vm, vd, vs2, vl); } @@ -1180,7 +1180,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_tumu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_tumu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_tumu(vbool32_t vm, vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf4_tumu(vm, vd, vs2, vl); } @@ -1191,7 +1191,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_tumu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_tumu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_q_f8e5m2mf2_tumu(vbool16_t vm, vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf2_tumu(vm, vd, vs2, vl); } @@ -1202,7 +1202,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_tumu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_tumu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_tumu(vbool16_t vm, vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf2_tumu(vm, vd, vs2, vl); } @@ -1213,7 +1213,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_tumu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_tumu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_q_f8e5m2m1_tumu(vbool8_t vm, vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2m1_tumu(vm, vd, vs2, vl); } @@ -1224,7 +1224,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_tumu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_tumu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_tumu(vbool8_t vm, vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2m1_tumu(vm, vd, vs2, vl); } @@ -1235,7 +1235,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_tumu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_tumu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_q_f8e5m2m2_tumu(vbool4_t vm, vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2m2_tumu(vm, vd, vs2, vl); } @@ -1246,7 +1246,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_tumu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_tumu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_tumu(vbool4_t vm, vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2m2_tumu(vm, vd, vs2, vl); } @@ -1257,7 +1257,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_tumu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_mu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_q_f8e5m2mf8_mu(vbool64_t vm, vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf8_mu(vm, vd, vs2, vl); } @@ -1268,7 +1268,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_mu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_mu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_mu(vbool64_t vm, vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf8_mu(vm, vd, vs2, vl); } @@ -1279,7 +1279,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_mu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_mu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_q_f8e5m2mf4_mu(vbool32_t vm, vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf4_mu(vm, vd, vs2, vl); } @@ -1290,7 +1290,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_mu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_mu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_mu(vbool32_t vm, vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf4_mu(vm, vd, vs2, vl); } @@ -1301,7 +1301,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_mu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_mu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_q_f8e5m2mf2_mu(vbool16_t vm, vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf2_mu(vm, vd, vs2, vl); } @@ -1312,7 +1312,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_mu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_mu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_mu(vbool16_t vm, vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf2_mu(vm, vd, vs2, vl); } @@ -1323,7 +1323,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_mu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_q_f8e5m2m1_mu(vbool8_t vm, vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2m1_mu(vm, vd, vs2, vl); } @@ -1334,7 +1334,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_mu(vbool8_t vm, vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2m1_mu(vm, vd, vs2, vl); } @@ -1345,7 +1345,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_q_f8e5m2m2_mu(vbool4_t vm, vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2m2_mu(vm, vd, vs2, vl); } @@ -1356,7 +1356,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_mu(vbool4_t vm, vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2m2_mu(vm, vd, vs2, vl); } @@ -1367,7 +1367,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_tu(vuint8mf8_t vd, vfloat32mf2_t vs2, +vfloat8e5m2mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_tu(vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf8_rm_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1378,7 +1378,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_tu(vuint8mf8_t vd, vfloat32mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_tu(vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_tu(vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf8_rm_tu(vd, vs2, __RISCV_FRM_RNE, vl); @@ -1390,7 +1390,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_tu(vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_tu(vuint8mf4_t vd, vfloat32m1_t vs2, +vfloat8e5m2mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_tu(vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf4_rm_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1401,7 +1401,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_tu(vuint8mf4_t vd, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_tu(vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_tu(vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf4_rm_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1412,7 +1412,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_tu(vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_tu(vuint8mf2_t vd, vfloat32m2_t vs2, +vfloat8e5m2mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_tu(vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf2_rm_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1423,7 +1423,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_tu(vuint8mf2_t vd, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_tu(vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_tu(vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf2_rm_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1434,7 +1434,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_tu(vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_tu(vuint8m1_t vd, vfloat32m4_t vs2, +vfloat8e5m2m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_tu(vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2m1_rm_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1445,7 +1445,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_tu(vuint8m1_t vd, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_tu(vuint8m1_t vd, vfloat32m4_t vs2, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_tu(vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2m1_rm_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1456,7 +1456,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_tu(vuint8m1_t vd, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_tu(vuint8m2_t vd, vfloat32m8_t vs2, +vfloat8e5m2m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_tu(vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2m2_rm_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1467,7 +1467,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_tu(vuint8m2_t vd, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_tu(vuint8m2_t vd, vfloat32m8_t vs2, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_tu(vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2m2_rm_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1478,7 +1478,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_tu(vuint8m2_t vd, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_tum(vbool64_t vm, vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_tum(vbool64_t vm, vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf8_rm_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -1490,7 +1490,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_tum(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_tum(vbool64_t vm, vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_tum(vbool64_t vm, vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf8_rm_tum(vm, vd, vs2, __RISCV_FRM_RNE, @@ -1503,7 +1503,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_tum(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_tum(vbool32_t vm, vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_tum(vbool32_t vm, vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf4_rm_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -1515,7 +1515,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_tum(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_tum(vbool32_t vm, vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_tum(vbool32_t vm, vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf4_rm_tum(vm, vd, vs2, __RISCV_FRM_RNE, @@ -1528,7 +1528,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_tum(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_tum(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_tum(vbool16_t vm, vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf2_rm_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -1540,7 +1540,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_tum(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_tum(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_tum(vbool16_t vm, vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf2_rm_tum(vm, vd, vs2, __RISCV_FRM_RNE, @@ -1553,7 +1553,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_tum(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_tum(vbool8_t vm, vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2m1_rm_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1564,7 +1564,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_tum(vbool8_t vm, vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2m1_rm_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -1576,7 +1576,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_tum(vbool4_t vm, vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2m2_rm_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1587,7 +1587,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_tum(vbool4_t vm, vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2m2_rm_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -1599,7 +1599,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_tumu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_tumu(vbool64_t vm, vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf8_rm_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -1611,8 +1611,8 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_tumu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_tumu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_tumu(vbool64_t vm, + vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf8_rm_tumu(vm, vd, vs2, @@ -1625,7 +1625,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_tumu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_tumu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_tumu(vbool32_t vm, vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf4_rm_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -1637,8 +1637,8 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_tumu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_tumu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_tumu(vbool32_t vm, + vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf4_rm_tumu(vm, vd, vs2, @@ -1651,7 +1651,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_tumu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_tumu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_tumu(vbool16_t vm, vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf2_rm_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -1663,8 +1663,8 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_tumu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_tumu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_tumu(vbool16_t vm, + vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf2_rm_tumu(vm, vd, vs2, @@ -1677,7 +1677,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_tumu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_tumu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_tumu(vbool8_t vm, vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2m1_rm_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -1689,7 +1689,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_tumu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_tumu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_tumu(vbool8_t vm, vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2m1_rm_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -1701,7 +1701,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_tumu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_tumu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_tumu(vbool4_t vm, vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2m2_rm_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -1713,7 +1713,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_tumu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_tumu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_tumu(vbool4_t vm, vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2m2_rm_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -1725,7 +1725,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_tumu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_mu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_mu(vbool64_t vm, vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf8_rm_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1736,7 +1736,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_mu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_mu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_mu(vbool64_t vm, vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf8_rm_mu(vm, vd, vs2, __RISCV_FRM_RNE, @@ -1749,7 +1749,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_mu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_mu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_mu(vbool32_t vm, vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf4_rm_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1760,7 +1760,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_mu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_mu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_mu(vbool32_t vm, vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf4_rm_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -1772,7 +1772,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_mu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_mu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_mu(vbool16_t vm, vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2mf2_rm_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1783,7 +1783,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_mu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_mu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_mu(vbool16_t vm, vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2mf2_rm_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -1795,7 +1795,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_mu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_mu(vbool8_t vm, vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2m1_rm_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1806,7 +1806,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_mu(vbool8_t vm, vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2m1_rm_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -1818,7 +1818,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_mu(vbool4_t vm, vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_q_f8e5m2m2_rm_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1829,7 +1829,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_mu(vbool4_t vm, vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_q_f8e5m2m2_rm_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); diff --git a/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/policy/non-overloaded/vfncvtbf16.c b/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/policy/non-overloaded/vfncvtbf16.c index 9f6d985528fb..7279dd244701 100644 --- a/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/policy/non-overloaded/vfncvtbf16.c +++ b/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/policy/non-overloaded/vfncvtbf16.c @@ -13,7 +13,7 @@ // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_tu(vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_tu(vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_tu(vd, vs2, vl); @@ -25,7 +25,7 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_tu(vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_tu(vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_tu(vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_tu(vd, vs2, vl); @@ -37,7 +37,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_tu(vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_tu(vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_tu(vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_tu(vd, vs2, vl); @@ -49,7 +49,7 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_tu(vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_tu(vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_tu(vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_tu(vd, vs2, vl); @@ -61,7 +61,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_tu(vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_tu(vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_tu(vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m1_f8e4m3mf2_tu(vd, vs2, vl); @@ -73,7 +73,7 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_tu(vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_tu(vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_tu(vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_tu(vd, vs2, vl); @@ -85,7 +85,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_tu(vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_tu(vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_tu(vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m2_f8e4m3m1_tu(vd, vs2, vl); } @@ -96,7 +96,7 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_tu(vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_tu(vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_tu(vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_tu(vd, vs2, vl); @@ -108,7 +108,7 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_tu(vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_tu(vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_tu(vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m4_f8e4m3m2_tu(vd, vs2, vl); } @@ -119,7 +119,7 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_tu(vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_tu(vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_tu(vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_tu(vd, vs2, vl); @@ -131,7 +131,7 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_tu(vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_tu(vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_tu(vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m8_f8e4m3m4_tu(vd, vs2, vl); } @@ -142,7 +142,7 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_tu(vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_tu(vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_tu(vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_tu(vd, vs2, vl); @@ -154,8 +154,8 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_tu(vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_tum(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_tum(vbool64_t vm, + vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_tum(vm, vd, vs2, vl); @@ -167,8 +167,8 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_tum(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_tum(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_tum(vbool64_t vm, + vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_tum(vm, vd, vs2, vl); @@ -180,8 +180,8 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_tum(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_tum(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_tum(vbool32_t vm, + vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_tum(vm, vd, vs2, vl); @@ -193,8 +193,8 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_tum(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_tum(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_tum(vbool32_t vm, + vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_tum(vm, vd, vs2, vl); @@ -206,7 +206,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_tum(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_tum(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_tum(vbool16_t vm, vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m1_f8e4m3mf2_tum(vm, vd, vs2, vl); @@ -218,8 +218,8 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_tum(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_tum(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_tum(vbool16_t vm, + vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_tum(vm, vd, vs2, vl); @@ -231,7 +231,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_tum(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_tum(vbool8_t vm, vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m2_f8e4m3m1_tum(vm, vd, vs2, vl); } @@ -242,7 +242,7 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_tum(vbool8_t vm, vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_tum(vm, vd, vs2, vl); @@ -254,7 +254,7 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_tum(vbool4_t vm, vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m4_f8e4m3m2_tum(vm, vd, vs2, vl); } @@ -265,7 +265,7 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_tum(vbool4_t vm, vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_tum(vm, vd, vs2, vl); @@ -277,7 +277,7 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_tum(vbool2_t vm, vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_tum(vbool2_t vm, vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m8_f8e4m3m4_tum(vm, vd, vs2, vl); } @@ -288,7 +288,7 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_tum(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_tum(vbool2_t vm, vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_tum(vbool2_t vm, vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_tum(vm, vd, vs2, vl); @@ -300,8 +300,8 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_tum(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_tumu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_tumu(vbool64_t vm, + vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_tumu(vm, vd, vs2, vl); @@ -313,8 +313,8 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_tumu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_tumu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_tumu(vbool64_t vm, + vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_tumu(vm, vd, vs2, vl); @@ -326,8 +326,8 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_tumu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_tumu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_tumu(vbool32_t vm, + vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_tumu(vm, vd, vs2, vl); @@ -339,8 +339,8 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_tumu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_tumu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_tumu(vbool32_t vm, + vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_tumu(vm, vd, vs2, vl); @@ -352,8 +352,8 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_tumu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_tumu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_tumu(vbool16_t vm, + vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m1_f8e4m3mf2_tumu(vm, vd, vs2, vl); @@ -365,8 +365,8 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_tumu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_tumu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_tumu(vbool16_t vm, + vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_tumu(vm, vd, vs2, vl); @@ -378,7 +378,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_tumu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_tumu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_tumu(vbool8_t vm, vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m2_f8e4m3m1_tumu(vm, vd, vs2, vl); @@ -390,8 +390,8 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_tumu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_tumu(vbool8_t vm, - vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_tumu(vbool8_t vm, + vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_tumu(vm, vd, vs2, vl); @@ -403,7 +403,7 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_tumu(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_tumu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_tumu(vbool4_t vm, vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m4_f8e4m3m2_tumu(vm, vd, vs2, vl); @@ -415,8 +415,8 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_tumu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_tumu(vbool4_t vm, - vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_tumu(vbool4_t vm, + vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_tumu(vm, vd, vs2, vl); @@ -428,7 +428,7 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_tumu(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_tumu(vbool2_t vm, vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_tumu(vbool2_t vm, vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m8_f8e4m3m4_tumu(vm, vd, vs2, vl); @@ -440,8 +440,8 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_tumu(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_tumu(vbool2_t vm, - vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_tumu(vbool2_t vm, + vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_tumu(vm, vd, vs2, vl); @@ -453,7 +453,7 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_tumu(vbool2_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_mu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_mu(vbool64_t vm, vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_mu(vm, vd, vs2, vl); @@ -465,8 +465,8 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_mu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_mu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_mu(vbool64_t vm, + vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_mu(vm, vd, vs2, vl); @@ -478,7 +478,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_mu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_mu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_mu(vbool32_t vm, vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_mu(vm, vd, vs2, vl); @@ -490,8 +490,8 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_mu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_mu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_mu(vbool32_t vm, + vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_mu(vm, vd, vs2, vl); @@ -503,7 +503,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_mu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_mu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_mu(vbool16_t vm, vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m1_f8e4m3mf2_mu(vm, vd, vs2, vl); @@ -515,8 +515,8 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_mu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_mu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_mu(vbool16_t vm, + vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_mu(vm, vd, vs2, vl); @@ -528,7 +528,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_mu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_mu(vbool8_t vm, vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m2_f8e4m3m1_mu(vm, vd, vs2, vl); } @@ -539,7 +539,7 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_mu(vbool8_t vm, vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_mu(vm, vd, vs2, vl); @@ -551,7 +551,7 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_mu(vbool4_t vm, vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m4_f8e4m3m2_mu(vm, vd, vs2, vl); } @@ -562,7 +562,7 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_mu(vbool4_t vm, vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_mu(vm, vd, vs2, vl); @@ -574,7 +574,7 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_mu(vbool2_t vm, vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_mu(vbool2_t vm, vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m8_f8e4m3m4_mu(vm, vd, vs2, vl); } @@ -585,7 +585,7 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_mu(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_mu(vbool2_t vm, vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_mu(vbool2_t vm, vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_mu(vm, vd, vs2, vl); @@ -597,7 +597,7 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_mu(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_tu(vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_tu(vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_tu(vd, vs2, __RISCV_FRM_RNE, @@ -610,7 +610,7 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_tu(vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_tu(vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_tu(vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_tu(vd, vs2, @@ -623,7 +623,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_tu(vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_tu(vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_tu(vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_tu(vd, vs2, __RISCV_FRM_RNE, @@ -636,7 +636,7 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_tu(vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_tu(vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_tu(vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_tu(vd, vs2, @@ -649,7 +649,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_tu(vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_tu(vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_tu(vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_tu(vd, vs2, __RISCV_FRM_RNE, @@ -662,7 +662,7 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_tu(vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_tu(vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_tu(vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_tu(vd, vs2, @@ -675,7 +675,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_tu(vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_tu(vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_tu(vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_tu(vd, vs2, __RISCV_FRM_RNE, @@ -688,7 +688,7 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_tu(vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_tu(vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_tu(vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_tu(vd, vs2, @@ -701,7 +701,7 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_tu(vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_tu(vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_tu(vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_tu(vd, vs2, __RISCV_FRM_RNE, @@ -714,7 +714,7 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_tu(vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_tu(vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_tu(vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_tu(vd, vs2, @@ -727,7 +727,7 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_tu(vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_tu(vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_tu(vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_tu(vd, vs2, __RISCV_FRM_RNE, @@ -740,7 +740,7 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_tu(vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_tu(vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_tu(vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_tu(vd, vs2, @@ -753,8 +753,8 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_tu(vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_tum(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_tum(vbool64_t vm, + vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_tum(vm, vd, vs2, @@ -767,8 +767,8 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_tum(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_tum(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_tum(vbool64_t vm, + vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_tum(vm, vd, vs2, @@ -781,8 +781,8 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_tum(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_tum(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_tum(vbool32_t vm, + vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_tum(vm, vd, vs2, @@ -795,8 +795,8 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_tum(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_tum(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_tum(vbool32_t vm, + vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_tum(vm, vd, vs2, @@ -809,8 +809,8 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_tum(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_tum(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_tum(vbool16_t vm, + vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_tum(vm, vd, vs2, @@ -823,8 +823,8 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_tum(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_tum(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_tum(vbool16_t vm, + vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_tum(vm, vd, vs2, @@ -837,7 +837,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_tum(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_tum(vbool8_t vm, vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_tum(vm, vd, vs2, @@ -850,8 +850,8 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_tum(vbool8_t vm, - vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_tum(vbool8_t vm, + vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_tum(vm, vd, vs2, @@ -864,7 +864,7 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_tum(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_tum(vbool4_t vm, vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_tum(vm, vd, vs2, @@ -877,8 +877,8 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_tum(vbool4_t vm, - vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_tum(vbool4_t vm, + vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_tum(vm, vd, vs2, @@ -891,7 +891,7 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_tum(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_tum(vbool2_t vm, vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_tum(vbool2_t vm, vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_tum(vm, vd, vs2, @@ -904,8 +904,8 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_tum(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_tum(vbool2_t vm, - vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_tum(vbool2_t vm, + vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_tum(vm, vd, vs2, @@ -918,8 +918,8 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_tum(vbool2_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_tumu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_tumu(vbool64_t vm, + vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_tumu(vm, vd, vs2, @@ -932,8 +932,8 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_tumu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_tumu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_tumu(vbool64_t vm, + vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_tumu( @@ -946,8 +946,8 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_tumu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_tumu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_tumu(vbool32_t vm, + vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_tumu(vm, vd, vs2, @@ -960,8 +960,8 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_tumu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_tumu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_tumu(vbool32_t vm, + vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_tumu( @@ -974,8 +974,8 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_tumu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_tumu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_tumu(vbool16_t vm, + vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_tumu(vm, vd, vs2, @@ -988,8 +988,8 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_tumu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_tumu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_tumu(vbool16_t vm, + vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_tumu(vm, vd, vs2, @@ -1002,7 +1002,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_tumu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_tumu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_tumu(vbool8_t vm, vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_tumu(vm, vd, vs2, @@ -1015,8 +1015,8 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_tumu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_tumu(vbool8_t vm, - vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_tumu(vbool8_t vm, + vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_tumu(vm, vd, vs2, @@ -1029,7 +1029,7 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_tumu(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_tumu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_tumu(vbool4_t vm, vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_tumu(vm, vd, vs2, @@ -1042,8 +1042,8 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_tumu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_tumu(vbool4_t vm, - vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_tumu(vbool4_t vm, + vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_tumu(vm, vd, vs2, @@ -1056,7 +1056,7 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_tumu(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_tumu(vbool2_t vm, vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_tumu(vbool2_t vm, vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_tumu(vm, vd, vs2, @@ -1069,8 +1069,8 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_tumu(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_tumu(vbool2_t vm, - vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_tumu(vbool2_t vm, + vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_tumu(vm, vd, vs2, @@ -1083,8 +1083,8 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_tumu(vbool2_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_mu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_mu(vbool64_t vm, + vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_mu(vm, vd, vs2, @@ -1097,8 +1097,8 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_mu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_mu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_mu(vbool64_t vm, + vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_mu(vm, vd, vs2, @@ -1111,8 +1111,8 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_mu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_mu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_mu(vbool32_t vm, + vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_mu(vm, vd, vs2, @@ -1125,8 +1125,8 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_mu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_mu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_mu(vbool32_t vm, + vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_mu(vm, vd, vs2, @@ -1139,8 +1139,8 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_mu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_mu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_mu(vbool16_t vm, + vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_mu(vm, vd, vs2, @@ -1153,8 +1153,8 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_mu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_mu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_mu(vbool16_t vm, + vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_mu(vm, vd, vs2, @@ -1167,7 +1167,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_mu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_mu(vbool8_t vm, vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_mu(vm, vd, vs2, @@ -1180,8 +1180,8 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_mu(vbool8_t vm, - vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_mu(vbool8_t vm, + vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_mu(vm, vd, vs2, @@ -1194,7 +1194,7 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_mu(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_mu(vbool4_t vm, vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_mu(vm, vd, vs2, @@ -1207,8 +1207,8 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_mu(vbool4_t vm, - vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_mu(vbool4_t vm, + vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_mu(vm, vd, vs2, @@ -1221,7 +1221,7 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_mu(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_mu(vbool2_t vm, vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_mu(vbool2_t vm, vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_mu(vm, vd, vs2, @@ -1234,8 +1234,8 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_mu(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_mu(vbool2_t vm, - vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_mu(vbool2_t vm, + vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_mu(vm, vd, vs2, @@ -1248,7 +1248,7 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_mu(vbool2_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_tu(vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_tu(vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_tu(vd, vs2, vl); @@ -1260,7 +1260,7 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_tu(vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_tu(vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_tu(vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_tu(vd, vs2, vl); @@ -1272,7 +1272,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_tu(vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_tu(vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_tu(vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_tu(vd, vs2, vl); @@ -1284,7 +1284,7 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_tu(vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_tu(vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_tu(vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_tu(vd, vs2, vl); @@ -1296,7 +1296,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_tu(vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_tu(vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_tu(vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m1_f8e5m2mf2_tu(vd, vs2, vl); @@ -1308,7 +1308,7 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_tu(vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_tu(vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_tu(vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_tu(vd, vs2, vl); @@ -1320,7 +1320,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_tu(vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_tu(vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_tu(vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m2_f8e5m2m1_tu(vd, vs2, vl); } @@ -1331,7 +1331,7 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_tu(vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_tu(vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_tu(vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_tu(vd, vs2, vl); @@ -1343,7 +1343,7 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_tu(vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_tu(vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_tu(vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m4_f8e5m2m2_tu(vd, vs2, vl); } @@ -1354,7 +1354,7 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_tu(vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_tu(vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_tu(vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_tu(vd, vs2, vl); @@ -1366,7 +1366,7 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_tu(vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_tu(vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_tu(vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m8_f8e5m2m4_tu(vd, vs2, vl); } @@ -1377,7 +1377,7 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_tu(vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_tu(vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_tu(vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_tu(vd, vs2, vl); @@ -1389,8 +1389,8 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_tu(vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_tum(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_tum(vbool64_t vm, + vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_tum(vm, vd, vs2, vl); @@ -1402,8 +1402,8 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_tum(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_tum(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_tum(vbool64_t vm, + vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_tum(vm, vd, vs2, vl); @@ -1415,8 +1415,8 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_tum(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_tum(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_tum(vbool32_t vm, + vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_tum(vm, vd, vs2, vl); @@ -1428,8 +1428,8 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_tum(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_tum(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_tum(vbool32_t vm, + vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_tum(vm, vd, vs2, vl); @@ -1441,7 +1441,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_tum(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_tum(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_tum(vbool16_t vm, vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m1_f8e5m2mf2_tum(vm, vd, vs2, vl); @@ -1453,8 +1453,8 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_tum(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_tum(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_tum(vbool16_t vm, + vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_tum(vm, vd, vs2, vl); @@ -1466,7 +1466,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_tum(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_tum(vbool8_t vm, vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m2_f8e5m2m1_tum(vm, vd, vs2, vl); } @@ -1477,7 +1477,7 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_tum(vbool8_t vm, vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_tum(vm, vd, vs2, vl); @@ -1489,7 +1489,7 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_tum(vbool4_t vm, vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m4_f8e5m2m2_tum(vm, vd, vs2, vl); } @@ -1500,7 +1500,7 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_tum(vbool4_t vm, vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_tum(vm, vd, vs2, vl); @@ -1512,7 +1512,7 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_tum(vbool2_t vm, vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_tum(vbool2_t vm, vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m8_f8e5m2m4_tum(vm, vd, vs2, vl); } @@ -1523,7 +1523,7 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_tum(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_tum(vbool2_t vm, vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_tum(vbool2_t vm, vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_tum(vm, vd, vs2, vl); @@ -1535,8 +1535,8 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_tum(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_tumu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_tumu(vbool64_t vm, + vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_tumu(vm, vd, vs2, vl); @@ -1548,8 +1548,8 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_tumu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_tumu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_tumu(vbool64_t vm, + vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_tumu(vm, vd, vs2, vl); @@ -1561,8 +1561,8 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_tumu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_tumu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_tumu(vbool32_t vm, + vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_tumu(vm, vd, vs2, vl); @@ -1574,8 +1574,8 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_tumu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_tumu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_tumu(vbool32_t vm, + vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_tumu(vm, vd, vs2, vl); @@ -1587,8 +1587,8 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_tumu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_tumu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_tumu(vbool16_t vm, + vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m1_f8e5m2mf2_tumu(vm, vd, vs2, vl); @@ -1600,8 +1600,8 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_tumu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_tumu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_tumu(vbool16_t vm, + vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_tumu(vm, vd, vs2, vl); @@ -1613,7 +1613,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_tumu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_tumu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_tumu(vbool8_t vm, vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m2_f8e5m2m1_tumu(vm, vd, vs2, vl); @@ -1625,8 +1625,8 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_tumu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_tumu(vbool8_t vm, - vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_tumu(vbool8_t vm, + vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_tumu(vm, vd, vs2, vl); @@ -1638,7 +1638,7 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_tumu(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_tumu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_tumu(vbool4_t vm, vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m4_f8e5m2m2_tumu(vm, vd, vs2, vl); @@ -1650,8 +1650,8 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_tumu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_tumu(vbool4_t vm, - vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_tumu(vbool4_t vm, + vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_tumu(vm, vd, vs2, vl); @@ -1663,7 +1663,7 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_tumu(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_tumu(vbool2_t vm, vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_tumu(vbool2_t vm, vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m8_f8e5m2m4_tumu(vm, vd, vs2, vl); @@ -1675,8 +1675,8 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_tumu(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_tumu(vbool2_t vm, - vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_tumu(vbool2_t vm, + vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_tumu(vm, vd, vs2, vl); @@ -1688,7 +1688,7 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_tumu(vbool2_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_mu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_mu(vbool64_t vm, vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_mu(vm, vd, vs2, vl); @@ -1700,8 +1700,8 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_mu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_mu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_mu(vbool64_t vm, + vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_mu(vm, vd, vs2, vl); @@ -1713,7 +1713,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_mu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_mu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_mu(vbool32_t vm, vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_mu(vm, vd, vs2, vl); @@ -1725,8 +1725,8 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_mu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_mu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_mu(vbool32_t vm, + vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_mu(vm, vd, vs2, vl); @@ -1738,7 +1738,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_mu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_mu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_mu(vbool16_t vm, vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m1_f8e5m2mf2_mu(vm, vd, vs2, vl); @@ -1750,8 +1750,8 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_mu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_mu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_mu(vbool16_t vm, + vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_mu(vm, vd, vs2, vl); @@ -1763,7 +1763,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_mu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_mu(vbool8_t vm, vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m2_f8e5m2m1_mu(vm, vd, vs2, vl); } @@ -1774,7 +1774,7 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_mu(vbool8_t vm, vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_mu(vm, vd, vs2, vl); @@ -1786,7 +1786,7 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_mu(vbool4_t vm, vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m4_f8e5m2m2_mu(vm, vd, vs2, vl); } @@ -1797,7 +1797,7 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_mu(vbool4_t vm, vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_mu(vm, vd, vs2, vl); @@ -1809,7 +1809,7 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_mu(vbool2_t vm, vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_mu(vbool2_t vm, vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m8_f8e5m2m4_mu(vm, vd, vs2, vl); } @@ -1820,7 +1820,7 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_mu(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_mu(vbool2_t vm, vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_mu(vbool2_t vm, vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_mu(vm, vd, vs2, vl); @@ -1832,7 +1832,7 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_mu(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_tu(vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_tu(vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_tu(vd, vs2, __RISCV_FRM_RNE, @@ -1845,7 +1845,7 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_tu(vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_tu(vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_tu(vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_tu(vd, vs2, @@ -1858,7 +1858,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_tu(vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_tu(vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_tu(vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_tu(vd, vs2, __RISCV_FRM_RNE, @@ -1871,7 +1871,7 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_tu(vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_tu(vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_tu(vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_tu(vd, vs2, @@ -1884,7 +1884,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_tu(vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_tu(vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_tu(vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_tu(vd, vs2, __RISCV_FRM_RNE, @@ -1897,7 +1897,7 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_tu(vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_tu(vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_tu(vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_tu(vd, vs2, @@ -1910,7 +1910,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_tu(vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_tu(vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_tu(vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_tu(vd, vs2, __RISCV_FRM_RNE, @@ -1923,7 +1923,7 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_tu(vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_tu(vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_tu(vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_tu(vd, vs2, @@ -1936,7 +1936,7 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_tu(vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_tu(vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_tu(vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_tu(vd, vs2, __RISCV_FRM_RNE, @@ -1949,7 +1949,7 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_tu(vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_tu(vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_tu(vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_tu(vd, vs2, @@ -1962,7 +1962,7 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_tu(vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_tu(vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_tu(vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_tu(vd, vs2, __RISCV_FRM_RNE, @@ -1975,7 +1975,7 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_tu(vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_tu(vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_tu(vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_tu(vd, vs2, @@ -1988,8 +1988,8 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_tu(vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_tum(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_tum(vbool64_t vm, + vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_tum(vm, vd, vs2, @@ -2002,8 +2002,8 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_tum(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_tum(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_tum(vbool64_t vm, + vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_tum(vm, vd, vs2, @@ -2016,8 +2016,8 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_tum(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_tum(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_tum(vbool32_t vm, + vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_tum(vm, vd, vs2, @@ -2030,8 +2030,8 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_tum(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_tum(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_tum(vbool32_t vm, + vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_tum(vm, vd, vs2, @@ -2044,8 +2044,8 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_tum(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_tum(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_tum(vbool16_t vm, + vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_tum(vm, vd, vs2, @@ -2058,8 +2058,8 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_tum(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_tum(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_tum(vbool16_t vm, + vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_tum(vm, vd, vs2, @@ -2072,7 +2072,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_tum(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_tum(vbool8_t vm, vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_tum(vm, vd, vs2, @@ -2085,8 +2085,8 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_tum(vbool8_t vm, - vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_tum(vbool8_t vm, + vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_tum(vm, vd, vs2, @@ -2099,7 +2099,7 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_tum(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_tum(vbool4_t vm, vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_tum(vm, vd, vs2, @@ -2112,8 +2112,8 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_tum(vbool4_t vm, - vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_tum(vbool4_t vm, + vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_tum(vm, vd, vs2, @@ -2126,7 +2126,7 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_tum(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_tum(vbool2_t vm, vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_tum(vbool2_t vm, vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_tum(vm, vd, vs2, @@ -2139,8 +2139,8 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_tum(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_tum(vbool2_t vm, - vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_tum(vbool2_t vm, + vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_tum(vm, vd, vs2, @@ -2153,8 +2153,8 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_tum(vbool2_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_tumu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_tumu(vbool64_t vm, + vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_tumu(vm, vd, vs2, @@ -2167,8 +2167,8 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_tumu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_tumu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_tumu(vbool64_t vm, + vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_tumu( @@ -2181,8 +2181,8 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_tumu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_tumu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_tumu(vbool32_t vm, + vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_tumu(vm, vd, vs2, @@ -2195,8 +2195,8 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_tumu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_tumu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_tumu(vbool32_t vm, + vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_tumu( @@ -2209,8 +2209,8 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_tumu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_tumu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_tumu(vbool16_t vm, + vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_tumu(vm, vd, vs2, @@ -2223,8 +2223,8 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_tumu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_tumu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_tumu(vbool16_t vm, + vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_tumu(vm, vd, vs2, @@ -2237,7 +2237,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_tumu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_tumu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_tumu(vbool8_t vm, vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_tumu(vm, vd, vs2, @@ -2250,8 +2250,8 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_tumu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_tumu(vbool8_t vm, - vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_tumu(vbool8_t vm, + vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_tumu(vm, vd, vs2, @@ -2264,7 +2264,7 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_tumu(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_tumu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_tumu(vbool4_t vm, vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_tumu(vm, vd, vs2, @@ -2277,8 +2277,8 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_tumu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_tumu(vbool4_t vm, - vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_tumu(vbool4_t vm, + vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_tumu(vm, vd, vs2, @@ -2291,7 +2291,7 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_tumu(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_tumu(vbool2_t vm, vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_tumu(vbool2_t vm, vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_tumu(vm, vd, vs2, @@ -2304,8 +2304,8 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_tumu(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_tumu(vbool2_t vm, - vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_tumu(vbool2_t vm, + vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_tumu(vm, vd, vs2, @@ -2318,8 +2318,8 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_tumu(vbool2_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_mu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_mu(vbool64_t vm, + vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_mu(vm, vd, vs2, @@ -2332,8 +2332,8 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_mu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_mu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_mu(vbool64_t vm, + vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_mu(vm, vd, vs2, @@ -2346,8 +2346,8 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_mu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_mu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_mu(vbool32_t vm, + vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_mu(vm, vd, vs2, @@ -2360,8 +2360,8 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_mu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_mu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_mu(vbool32_t vm, + vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_mu(vm, vd, vs2, @@ -2374,8 +2374,8 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_mu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_mu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_mu(vbool16_t vm, + vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_mu(vm, vd, vs2, @@ -2388,8 +2388,8 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_mu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_mu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_mu(vbool16_t vm, + vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_mu(vm, vd, vs2, @@ -2402,7 +2402,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_mu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_mu(vbool8_t vm, vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_mu(vm, vd, vs2, @@ -2415,8 +2415,8 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_mu(vbool8_t vm, - vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_mu(vbool8_t vm, + vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_mu(vm, vd, vs2, @@ -2429,7 +2429,7 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_mu(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_mu(vbool4_t vm, vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_mu(vm, vd, vs2, @@ -2442,8 +2442,8 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_mu(vbool4_t vm, - vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_mu(vbool4_t vm, + vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_mu(vm, vd, vs2, @@ -2456,7 +2456,7 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_mu(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_mu(vbool2_t vm, vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_mu(vbool2_t vm, vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_mu(vm, vd, vs2, @@ -2469,8 +2469,8 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_mu(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_mu(vbool2_t vm, - vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_mu(vbool2_t vm, + vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_mu(vm, vd, vs2, diff --git a/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/policy/non-overloaded/vfwcvtbf16.c b/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/policy/non-overloaded/vfwcvtbf16.c index a114e043a926..6a0a8254ba91 100644 --- a/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/policy/non-overloaded/vfwcvtbf16.c +++ b/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/policy/non-overloaded/vfwcvtbf16.c @@ -14,7 +14,7 @@ // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_tu(vbfloat16mf4_t vd, - vuint8mf8_t vs2, + vfloat8e4m3mf8_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_tu(vd, vs2, vl); } @@ -26,7 +26,7 @@ vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_tu(vbfloat16mf4_t vd, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_tu(vbfloat16mf2_t vd, - vuint8mf4_t vs2, + vfloat8e4m3mf4_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_tu(vd, vs2, vl); } @@ -38,7 +38,7 @@ vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_tu(vbfloat16mf2_t vd, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_tu(vbfloat16m1_t vd, - vuint8mf2_t vs2, + vfloat8e4m3mf2_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_tu(vd, vs2, vl); } @@ -50,7 +50,7 @@ vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_tu(vbfloat16m1_t vd, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_tu(vbfloat16m2_t vd, - vuint8m1_t vs2, size_t vl) { + vfloat8e4m3m1_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3m1_bf16m2_tu(vd, vs2, vl); } @@ -61,7 +61,7 @@ vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_tu(vbfloat16m2_t vd, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_tu(vbfloat16m4_t vd, - vuint8m2_t vs2, size_t vl) { + vfloat8e4m3m2_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3m2_bf16m4_tu(vd, vs2, vl); } @@ -72,7 +72,7 @@ vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_tu(vbfloat16m4_t vd, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_tu(vbfloat16m8_t vd, - vuint8m4_t vs2, size_t vl) { + vfloat8e4m3m4_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3m4_bf16m8_tu(vd, vs2, vl); } @@ -84,7 +84,7 @@ vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_tu(vbfloat16m8_t vd, // vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_tum(vbool64_t vm, vbfloat16mf4_t vd, - vuint8mf8_t vs2, + vfloat8e4m3mf8_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_tum(vm, vd, vs2, vl); } @@ -97,7 +97,7 @@ vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_tum(vbool64_t vm, // vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_tum(vbool32_t vm, vbfloat16mf2_t vd, - vuint8mf4_t vs2, + vfloat8e4m3mf4_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_tum(vm, vd, vs2, vl); } @@ -110,7 +110,7 @@ vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_tum(vbool32_t vm, // vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_tum(vbool16_t vm, vbfloat16m1_t vd, - vuint8mf2_t vs2, + vfloat8e4m3mf2_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_tum(vm, vd, vs2, vl); } @@ -123,7 +123,7 @@ vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_tum(vbool16_t vm, // vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_tum(vbool8_t vm, vbfloat16m2_t vd, - vuint8m1_t vs2, size_t vl) { + vfloat8e4m3m1_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3m1_bf16m2_tum(vm, vd, vs2, vl); } @@ -135,7 +135,7 @@ vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_tum(vbool8_t vm, // vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_tum(vbool4_t vm, vbfloat16m4_t vd, - vuint8m2_t vs2, size_t vl) { + vfloat8e4m3m2_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3m2_bf16m4_tum(vm, vd, vs2, vl); } @@ -147,7 +147,7 @@ vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_tum(vbool4_t vm, // vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_tum(vbool2_t vm, vbfloat16m8_t vd, - vuint8m4_t vs2, size_t vl) { + vfloat8e4m3m4_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3m4_bf16m8_tum(vm, vd, vs2, vl); } @@ -159,7 +159,7 @@ vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_tum(vbool2_t vm, // vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_tumu(vbool64_t vm, vbfloat16mf4_t vd, - vuint8mf8_t vs2, + vfloat8e4m3mf8_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_tumu(vm, vd, vs2, vl); } @@ -172,7 +172,7 @@ vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_tumu(vbool64_t vm, // vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_tumu(vbool32_t vm, vbfloat16mf2_t vd, - vuint8mf4_t vs2, + vfloat8e4m3mf4_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_tumu(vm, vd, vs2, vl); } @@ -185,7 +185,7 @@ vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_tumu(vbool32_t vm, // vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_tumu(vbool16_t vm, vbfloat16m1_t vd, - vuint8mf2_t vs2, + vfloat8e4m3mf2_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_tumu(vm, vd, vs2, vl); } @@ -198,7 +198,7 @@ vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_tumu(vbool16_t vm, // vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_tumu(vbool8_t vm, vbfloat16m2_t vd, - vuint8m1_t vs2, + vfloat8e4m3m1_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3m1_bf16m2_tumu(vm, vd, vs2, vl); } @@ -211,7 +211,7 @@ vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_tumu(vbool8_t vm, // vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_tumu(vbool4_t vm, vbfloat16m4_t vd, - vuint8m2_t vs2, + vfloat8e4m3m2_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3m2_bf16m4_tumu(vm, vd, vs2, vl); } @@ -224,7 +224,7 @@ vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_tumu(vbool4_t vm, // vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_tumu(vbool2_t vm, vbfloat16m8_t vd, - vuint8m4_t vs2, + vfloat8e4m3m4_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3m4_bf16m8_tumu(vm, vd, vs2, vl); } @@ -237,7 +237,7 @@ vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_tumu(vbool2_t vm, // vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_mu(vbool64_t vm, vbfloat16mf4_t vd, - vuint8mf8_t vs2, + vfloat8e4m3mf8_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_mu(vm, vd, vs2, vl); } @@ -250,7 +250,7 @@ vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_mu(vbool64_t vm, // vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_mu(vbool32_t vm, vbfloat16mf2_t vd, - vuint8mf4_t vs2, + vfloat8e4m3mf4_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_mu(vm, vd, vs2, vl); } @@ -263,7 +263,7 @@ vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_mu(vbool32_t vm, // vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_mu(vbool16_t vm, vbfloat16m1_t vd, - vuint8mf2_t vs2, + vfloat8e4m3mf2_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_mu(vm, vd, vs2, vl); } @@ -276,7 +276,7 @@ vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_mu(vbool16_t vm, // vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_mu(vbool8_t vm, vbfloat16m2_t vd, - vuint8m1_t vs2, size_t vl) { + vfloat8e4m3m1_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3m1_bf16m2_mu(vm, vd, vs2, vl); } @@ -288,7 +288,7 @@ vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_mu(vbool8_t vm, // vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_mu(vbool4_t vm, vbfloat16m4_t vd, - vuint8m2_t vs2, size_t vl) { + vfloat8e4m3m2_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3m2_bf16m4_mu(vm, vd, vs2, vl); } @@ -300,7 +300,7 @@ vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_mu(vbool4_t vm, // vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_mu(vbool2_t vm, vbfloat16m8_t vd, - vuint8m4_t vs2, size_t vl) { + vfloat8e4m3m4_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e4m3m4_bf16m8_mu(vm, vd, vs2, vl); } @@ -311,7 +311,7 @@ vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_mu(vbool2_t vm, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_tu(vbfloat16mf4_t vd, - vuint8mf8_t vs2, + vfloat8e5m2mf8_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_tu(vd, vs2, vl); } @@ -323,7 +323,7 @@ vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_tu(vbfloat16mf4_t vd, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_tu(vbfloat16mf2_t vd, - vuint8mf4_t vs2, + vfloat8e5m2mf4_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_tu(vd, vs2, vl); } @@ -335,7 +335,7 @@ vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_tu(vbfloat16mf2_t vd, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_tu(vbfloat16m1_t vd, - vuint8mf2_t vs2, + vfloat8e5m2mf2_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_tu(vd, vs2, vl); } @@ -347,7 +347,7 @@ vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_tu(vbfloat16m1_t vd, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_tu(vbfloat16m2_t vd, - vuint8m1_t vs2, size_t vl) { + vfloat8e5m2m1_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2m1_bf16m2_tu(vd, vs2, vl); } @@ -358,7 +358,7 @@ vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_tu(vbfloat16m2_t vd, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_tu(vbfloat16m4_t vd, - vuint8m2_t vs2, size_t vl) { + vfloat8e5m2m2_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2m2_bf16m4_tu(vd, vs2, vl); } @@ -369,7 +369,7 @@ vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_tu(vbfloat16m4_t vd, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16m8_t test_vfwcvt_f_f_v_f8e5m2m4_bf16m8_tu(vbfloat16m8_t vd, - vuint8m4_t vs2, size_t vl) { + vfloat8e5m2m4_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2m4_bf16m8_tu(vd, vs2, vl); } @@ -381,7 +381,7 @@ vbfloat16m8_t test_vfwcvt_f_f_v_f8e5m2m4_bf16m8_tu(vbfloat16m8_t vd, // vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_tum(vbool64_t vm, vbfloat16mf4_t vd, - vuint8mf8_t vs2, + vfloat8e5m2mf8_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_tum(vm, vd, vs2, vl); } @@ -394,7 +394,7 @@ vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_tum(vbool64_t vm, // vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_tum(vbool32_t vm, vbfloat16mf2_t vd, - vuint8mf4_t vs2, + vfloat8e5m2mf4_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_tum(vm, vd, vs2, vl); } @@ -407,7 +407,7 @@ vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_tum(vbool32_t vm, // vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_tum(vbool16_t vm, vbfloat16m1_t vd, - vuint8mf2_t vs2, + vfloat8e5m2mf2_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_tum(vm, vd, vs2, vl); } @@ -420,7 +420,7 @@ vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_tum(vbool16_t vm, // vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_tum(vbool8_t vm, vbfloat16m2_t vd, - vuint8m1_t vs2, size_t vl) { + vfloat8e5m2m1_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2m1_bf16m2_tum(vm, vd, vs2, vl); } @@ -432,7 +432,7 @@ vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_tum(vbool8_t vm, // vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_tum(vbool4_t vm, vbfloat16m4_t vd, - vuint8m2_t vs2, size_t vl) { + vfloat8e5m2m2_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2m2_bf16m4_tum(vm, vd, vs2, vl); } @@ -444,7 +444,7 @@ vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_tum(vbool4_t vm, // vbfloat16m8_t test_vfwcvt_f_f_v_f8e5m2m4_bf16m8_tum(vbool2_t vm, vbfloat16m8_t vd, - vuint8m4_t vs2, size_t vl) { + vfloat8e5m2m4_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2m4_bf16m8_tum(vm, vd, vs2, vl); } @@ -456,7 +456,7 @@ vbfloat16m8_t test_vfwcvt_f_f_v_f8e5m2m4_bf16m8_tum(vbool2_t vm, // vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_tumu(vbool64_t vm, vbfloat16mf4_t vd, - vuint8mf8_t vs2, + vfloat8e5m2mf8_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_tumu(vm, vd, vs2, vl); } @@ -469,7 +469,7 @@ vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_tumu(vbool64_t vm, // vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_tumu(vbool32_t vm, vbfloat16mf2_t vd, - vuint8mf4_t vs2, + vfloat8e5m2mf4_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_tumu(vm, vd, vs2, vl); } @@ -482,7 +482,7 @@ vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_tumu(vbool32_t vm, // vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_tumu(vbool16_t vm, vbfloat16m1_t vd, - vuint8mf2_t vs2, + vfloat8e5m2mf2_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_tumu(vm, vd, vs2, vl); } @@ -495,7 +495,7 @@ vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_tumu(vbool16_t vm, // vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_tumu(vbool8_t vm, vbfloat16m2_t vd, - vuint8m1_t vs2, + vfloat8e5m2m1_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2m1_bf16m2_tumu(vm, vd, vs2, vl); } @@ -508,7 +508,7 @@ vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_tumu(vbool8_t vm, // vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_tumu(vbool4_t vm, vbfloat16m4_t vd, - vuint8m2_t vs2, + vfloat8e5m2m2_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2m2_bf16m4_tumu(vm, vd, vs2, vl); } @@ -521,7 +521,7 @@ vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_tumu(vbool4_t vm, // vbfloat16m8_t test_vfwcvt_f_f_v_f8e5m2m4_bf16m8_tumu(vbool2_t vm, vbfloat16m8_t vd, - vuint8m4_t vs2, + vfloat8e5m2m4_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2m4_bf16m8_tumu(vm, vd, vs2, vl); } @@ -534,7 +534,7 @@ vbfloat16m8_t test_vfwcvt_f_f_v_f8e5m2m4_bf16m8_tumu(vbool2_t vm, // vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_mu(vbool64_t vm, vbfloat16mf4_t vd, - vuint8mf8_t vs2, + vfloat8e5m2mf8_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_mu(vm, vd, vs2, vl); } @@ -547,7 +547,7 @@ vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_mu(vbool64_t vm, // vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_mu(vbool32_t vm, vbfloat16mf2_t vd, - vuint8mf4_t vs2, + vfloat8e5m2mf4_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_mu(vm, vd, vs2, vl); } @@ -560,7 +560,7 @@ vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_mu(vbool32_t vm, // vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_mu(vbool16_t vm, vbfloat16m1_t vd, - vuint8mf2_t vs2, + vfloat8e5m2mf2_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_mu(vm, vd, vs2, vl); } @@ -573,7 +573,7 @@ vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_mu(vbool16_t vm, // vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_mu(vbool8_t vm, vbfloat16m2_t vd, - vuint8m1_t vs2, size_t vl) { + vfloat8e5m2m1_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2m1_bf16m2_mu(vm, vd, vs2, vl); } @@ -585,7 +585,7 @@ vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_mu(vbool8_t vm, // vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_mu(vbool4_t vm, vbfloat16m4_t vd, - vuint8m2_t vs2, size_t vl) { + vfloat8e5m2m2_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2m2_bf16m4_mu(vm, vd, vs2, vl); } @@ -597,6 +597,6 @@ vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_mu(vbool4_t vm, // vbfloat16m8_t test_vfwcvt_f_f_v_f8e5m2m4_bf16m8_mu(vbool2_t vm, vbfloat16m8_t vd, - vuint8m4_t vs2, size_t vl) { + vfloat8e5m2m4_t vs2, size_t vl) { return __riscv_vfwcvt_f_f_v_f8e5m2m4_bf16m8_mu(vm, vd, vs2, vl); } diff --git a/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/policy/overloaded/vfncvt.c b/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/policy/overloaded/vfncvt.c index 30583aa61a70..f295b23303df 100644 --- a/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/policy/overloaded/vfncvt.c +++ b/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/policy/overloaded/vfncvt.c @@ -13,7 +13,7 @@ // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_tu(vuint8mf8_t vd, vfloat32mf2_t vs2, +vfloat8e4m3mf8_t test_vfncvt_f_f_q_f8e4m3mf8_tu(vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tu(vd, vs2, vl); } @@ -24,7 +24,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_tu(vuint8mf8_t vd, vfloat32mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_tu(vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_tu(vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tu(vd, vs2, vl); } @@ -35,7 +35,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_tu(vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_tu(vuint8mf4_t vd, vfloat32m1_t vs2, +vfloat8e4m3mf4_t test_vfncvt_f_f_q_f8e4m3mf4_tu(vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tu(vd, vs2, vl); } @@ -46,7 +46,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_tu(vuint8mf4_t vd, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_tu(vuint8mf4_t vd, vfloat32m1_t vs2, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_tu(vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tu(vd, vs2, vl); } @@ -57,7 +57,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_tu(vuint8mf4_t vd, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_tu(vuint8mf2_t vd, vfloat32m2_t vs2, +vfloat8e4m3mf2_t test_vfncvt_f_f_q_f8e4m3mf2_tu(vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tu(vd, vs2, vl); } @@ -68,7 +68,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_tu(vuint8mf2_t vd, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_tu(vuint8mf2_t vd, vfloat32m2_t vs2, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_tu(vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tu(vd, vs2, vl); } @@ -79,7 +79,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_tu(vuint8mf2_t vd, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_tu(vuint8m1_t vd, vfloat32m4_t vs2, +vfloat8e4m3m1_t test_vfncvt_f_f_q_f8e4m3m1_tu(vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tu(vd, vs2, vl); } @@ -90,7 +90,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_tu(vuint8m1_t vd, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_tu(vuint8m1_t vd, vfloat32m4_t vs2, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_tu(vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tu(vd, vs2, vl); } @@ -101,7 +101,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_tu(vuint8m1_t vd, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_tu(vuint8m2_t vd, vfloat32m8_t vs2, +vfloat8e4m3m2_t test_vfncvt_f_f_q_f8e4m3m2_tu(vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tu(vd, vs2, vl); } @@ -112,7 +112,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_tu(vuint8m2_t vd, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_tu(vuint8m2_t vd, vfloat32m8_t vs2, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_tu(vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tu(vd, vs2, vl); } @@ -123,7 +123,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_tu(vuint8m2_t vd, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_tum(vbool64_t vm, vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_q_f8e4m3mf8_tum(vbool64_t vm, vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tum(vm, vd, vs2, vl); } @@ -134,7 +134,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_tum(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_tum(vbool64_t vm, vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_tum(vbool64_t vm, vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tum(vm, vd, vs2, vl); } @@ -145,7 +145,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_tum(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_tum(vbool32_t vm, vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_q_f8e4m3mf4_tum(vbool32_t vm, vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tum(vm, vd, vs2, vl); } @@ -156,7 +156,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_tum(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_tum(vbool32_t vm, vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_tum(vbool32_t vm, vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tum(vm, vd, vs2, vl); } @@ -167,7 +167,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_tum(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_tum(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_q_f8e4m3mf2_tum(vbool16_t vm, vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tum(vm, vd, vs2, vl); } @@ -178,7 +178,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_tum(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_tum(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_tum(vbool16_t vm, vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tum(vm, vd, vs2, vl); } @@ -189,7 +189,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_tum(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_q_f8e4m3m1_tum(vbool8_t vm, vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tum(vm, vd, vs2, vl); } @@ -200,7 +200,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_tum(vbool8_t vm, vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tum(vm, vd, vs2, vl); } @@ -211,7 +211,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_q_f8e4m3m2_tum(vbool4_t vm, vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tum(vm, vd, vs2, vl); } @@ -222,7 +222,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_tum(vbool4_t vm, vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tum(vm, vd, vs2, vl); } @@ -233,7 +233,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_tumu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_q_f8e4m3mf8_tumu(vbool64_t vm, vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tumu(vm, vd, vs2, vl); } @@ -244,7 +244,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_tumu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_tumu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_tumu(vbool64_t vm, vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tumu(vm, vd, vs2, vl); } @@ -255,7 +255,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_tumu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_tumu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_q_f8e4m3mf4_tumu(vbool32_t vm, vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tumu(vm, vd, vs2, vl); } @@ -266,7 +266,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_tumu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_tumu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_tumu(vbool32_t vm, vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tumu(vm, vd, vs2, vl); } @@ -277,7 +277,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_tumu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_tumu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_q_f8e4m3mf2_tumu(vbool16_t vm, vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tumu(vm, vd, vs2, vl); } @@ -288,7 +288,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_tumu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_tumu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_tumu(vbool16_t vm, vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tumu(vm, vd, vs2, vl); } @@ -299,7 +299,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_tumu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_tumu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_q_f8e4m3m1_tumu(vbool8_t vm, vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tumu(vm, vd, vs2, vl); } @@ -310,7 +310,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_tumu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_tumu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_tumu(vbool8_t vm, vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tumu(vm, vd, vs2, vl); } @@ -321,7 +321,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_tumu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_tumu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_q_f8e4m3m2_tumu(vbool4_t vm, vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tumu(vm, vd, vs2, vl); } @@ -332,7 +332,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_tumu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_tumu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_tumu(vbool4_t vm, vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tumu(vm, vd, vs2, vl); } @@ -343,7 +343,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_tumu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_mu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_q_f8e4m3mf8_mu(vbool64_t vm, vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_mu(vm, vd, vs2, vl); } @@ -354,7 +354,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_mu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_mu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_mu(vbool64_t vm, vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_mu(vm, vd, vs2, vl); } @@ -365,7 +365,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_mu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_mu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_q_f8e4m3mf4_mu(vbool32_t vm, vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_mu(vm, vd, vs2, vl); } @@ -376,7 +376,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_mu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_mu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_mu(vbool32_t vm, vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_mu(vm, vd, vs2, vl); } @@ -387,7 +387,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_mu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_mu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_q_f8e4m3mf2_mu(vbool16_t vm, vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_mu(vm, vd, vs2, vl); } @@ -398,7 +398,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_mu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_mu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_mu(vbool16_t vm, vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_mu(vm, vd, vs2, vl); } @@ -409,7 +409,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_mu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_q_f8e4m3m1_mu(vbool8_t vm, vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_mu(vm, vd, vs2, vl); } @@ -420,7 +420,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_mu(vbool8_t vm, vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_mu(vm, vd, vs2, vl); } @@ -431,7 +431,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_q_f8e4m3m2_mu(vbool4_t vm, vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_mu(vm, vd, vs2, vl); } @@ -442,7 +442,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_mu(vbool4_t vm, vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_mu(vm, vd, vs2, vl); } @@ -453,7 +453,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_tu(vuint8mf8_t vd, vfloat32mf2_t vs2, +vfloat8e4m3mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_tu(vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -464,7 +464,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_tu(vuint8mf8_t vd, vfloat32mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_tu(vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_tu(vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); @@ -476,7 +476,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_tu(vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_tu(vuint8mf4_t vd, vfloat32m1_t vs2, +vfloat8e4m3mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_tu(vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -487,7 +487,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_tu(vuint8mf4_t vd, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_tu(vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_tu(vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -498,7 +498,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_tu(vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_tu(vuint8mf2_t vd, vfloat32m2_t vs2, +vfloat8e4m3mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_tu(vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -509,7 +509,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_tu(vuint8mf2_t vd, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_tu(vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_tu(vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -520,7 +520,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_tu(vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_tu(vuint8m1_t vd, vfloat32m4_t vs2, +vfloat8e4m3m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_tu(vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -531,7 +531,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_tu(vuint8m1_t vd, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_tu(vuint8m1_t vd, vfloat32m4_t vs2, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_tu(vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -542,7 +542,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_tu(vuint8m1_t vd, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_tu(vuint8m2_t vd, vfloat32m8_t vs2, +vfloat8e4m3m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_tu(vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -553,7 +553,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_tu(vuint8m2_t vd, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_tu(vuint8m2_t vd, vfloat32m8_t vs2, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_tu(vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -564,7 +564,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_tu(vuint8m2_t vd, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_tum(vbool64_t vm, vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_tum(vbool64_t vm, vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -575,7 +575,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_tum(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_tum(vbool64_t vm, vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_tum(vbool64_t vm, vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -587,7 +587,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_tum(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_tum(vbool32_t vm, vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_tum(vbool32_t vm, vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -598,7 +598,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_tum(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_tum(vbool32_t vm, vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_tum(vbool32_t vm, vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -610,7 +610,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_tum(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_tum(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_tum(vbool16_t vm, vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -621,7 +621,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_tum(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_tum(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_tum(vbool16_t vm, vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -633,7 +633,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_tum(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_tum(vbool8_t vm, vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -644,7 +644,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_tum(vbool8_t vm, vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -655,7 +655,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_tum(vbool4_t vm, vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -666,7 +666,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_tum(vbool4_t vm, vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -677,7 +677,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_tumu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_tumu(vbool64_t vm, vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -688,8 +688,8 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_tumu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_tumu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_tumu(vbool64_t vm, + vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -701,7 +701,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_tumu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_tumu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_tumu(vbool32_t vm, vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -712,8 +712,8 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_tumu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_tumu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_tumu(vbool32_t vm, + vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -725,7 +725,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_tumu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_tumu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_tumu(vbool16_t vm, vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -736,8 +736,8 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_tumu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_tumu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_tumu(vbool16_t vm, + vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -749,7 +749,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_tumu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_tumu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_tumu(vbool8_t vm, vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -760,7 +760,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_tumu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_tumu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_tumu(vbool8_t vm, vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -771,7 +771,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_tumu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_tumu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_tumu(vbool4_t vm, vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -782,7 +782,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_tumu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_tumu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_tumu(vbool4_t vm, vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -793,7 +793,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_tumu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_mu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_mu(vbool64_t vm, vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -804,7 +804,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e4m3mf8_rm_mu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_mu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_mu(vbool64_t vm, vfloat8e4m3mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -816,7 +816,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e4m3mf8_rm_mu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_mu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_mu(vbool32_t vm, vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -827,7 +827,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e4m3mf4_rm_mu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_mu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_mu(vbool32_t vm, vfloat8e4m3mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -838,7 +838,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e4m3mf4_rm_mu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_mu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_mu(vbool16_t vm, vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -849,7 +849,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e4m3mf2_rm_mu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_mu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_mu(vbool16_t vm, vfloat8e4m3mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -860,7 +860,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e4m3mf2_rm_mu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_mu(vbool8_t vm, vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -871,7 +871,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e4m3m1_rm_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_mu(vbool8_t vm, vfloat8e4m3m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -882,7 +882,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e4m3m1_rm_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_mu(vbool4_t vm, vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -893,7 +893,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e4m3m2_rm_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_mu(vbool4_t vm, vfloat8e4m3m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -904,7 +904,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e4m3m2_rm_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_tu(vuint8mf8_t vd, vfloat32mf2_t vs2, +vfloat8e5m2mf8_t test_vfncvt_f_f_q_f8e5m2mf8_tu(vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tu(vd, vs2, vl); } @@ -915,7 +915,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_tu(vuint8mf8_t vd, vfloat32mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_tu(vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_tu(vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tu(vd, vs2, vl); } @@ -926,7 +926,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_tu(vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_tu(vuint8mf4_t vd, vfloat32m1_t vs2, +vfloat8e5m2mf4_t test_vfncvt_f_f_q_f8e5m2mf4_tu(vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tu(vd, vs2, vl); } @@ -937,7 +937,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_tu(vuint8mf4_t vd, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_tu(vuint8mf4_t vd, vfloat32m1_t vs2, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_tu(vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tu(vd, vs2, vl); } @@ -948,7 +948,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_tu(vuint8mf4_t vd, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_tu(vuint8mf2_t vd, vfloat32m2_t vs2, +vfloat8e5m2mf2_t test_vfncvt_f_f_q_f8e5m2mf2_tu(vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tu(vd, vs2, vl); } @@ -959,7 +959,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_tu(vuint8mf2_t vd, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_tu(vuint8mf2_t vd, vfloat32m2_t vs2, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_tu(vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tu(vd, vs2, vl); } @@ -970,7 +970,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_tu(vuint8mf2_t vd, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_tu(vuint8m1_t vd, vfloat32m4_t vs2, +vfloat8e5m2m1_t test_vfncvt_f_f_q_f8e5m2m1_tu(vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tu(vd, vs2, vl); } @@ -981,7 +981,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_tu(vuint8m1_t vd, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_tu(vuint8m1_t vd, vfloat32m4_t vs2, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_tu(vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tu(vd, vs2, vl); } @@ -992,7 +992,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_tu(vuint8m1_t vd, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_tu(vuint8m2_t vd, vfloat32m8_t vs2, +vfloat8e5m2m2_t test_vfncvt_f_f_q_f8e5m2m2_tu(vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tu(vd, vs2, vl); } @@ -1003,7 +1003,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_tu(vuint8m2_t vd, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_tu(vuint8m2_t vd, vfloat32m8_t vs2, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_tu(vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tu(vd, vs2, vl); } @@ -1014,7 +1014,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_tu(vuint8m2_t vd, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_tum(vbool64_t vm, vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_q_f8e5m2mf8_tum(vbool64_t vm, vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tum(vm, vd, vs2, vl); } @@ -1025,7 +1025,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_tum(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_tum(vbool64_t vm, vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_tum(vbool64_t vm, vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tum(vm, vd, vs2, vl); } @@ -1036,7 +1036,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_tum(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_tum(vbool32_t vm, vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_q_f8e5m2mf4_tum(vbool32_t vm, vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tum(vm, vd, vs2, vl); } @@ -1047,7 +1047,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_tum(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_tum(vbool32_t vm, vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_tum(vbool32_t vm, vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tum(vm, vd, vs2, vl); } @@ -1058,7 +1058,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_tum(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_tum(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_q_f8e5m2mf2_tum(vbool16_t vm, vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tum(vm, vd, vs2, vl); } @@ -1069,7 +1069,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_tum(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_tum(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_tum(vbool16_t vm, vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tum(vm, vd, vs2, vl); } @@ -1080,7 +1080,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_tum(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_q_f8e5m2m1_tum(vbool8_t vm, vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tum(vm, vd, vs2, vl); } @@ -1091,7 +1091,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_tum(vbool8_t vm, vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tum(vm, vd, vs2, vl); } @@ -1102,7 +1102,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_q_f8e5m2m2_tum(vbool4_t vm, vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tum(vm, vd, vs2, vl); } @@ -1113,7 +1113,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_tum(vbool4_t vm, vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tum(vm, vd, vs2, vl); } @@ -1124,7 +1124,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_tumu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_q_f8e5m2mf8_tumu(vbool64_t vm, vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tumu(vm, vd, vs2, vl); } @@ -1135,7 +1135,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_tumu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_tumu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_tumu(vbool64_t vm, vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tumu(vm, vd, vs2, vl); } @@ -1146,7 +1146,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_tumu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_tumu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_q_f8e5m2mf4_tumu(vbool32_t vm, vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tumu(vm, vd, vs2, vl); } @@ -1157,7 +1157,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_tumu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_tumu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_tumu(vbool32_t vm, vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tumu(vm, vd, vs2, vl); } @@ -1168,7 +1168,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_tumu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_tumu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_q_f8e5m2mf2_tumu(vbool16_t vm, vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tumu(vm, vd, vs2, vl); } @@ -1179,7 +1179,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_tumu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_tumu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_tumu(vbool16_t vm, vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tumu(vm, vd, vs2, vl); } @@ -1190,7 +1190,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_tumu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_tumu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_q_f8e5m2m1_tumu(vbool8_t vm, vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tumu(vm, vd, vs2, vl); } @@ -1201,7 +1201,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_tumu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_tumu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_tumu(vbool8_t vm, vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tumu(vm, vd, vs2, vl); } @@ -1212,7 +1212,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_tumu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_tumu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_q_f8e5m2m2_tumu(vbool4_t vm, vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tumu(vm, vd, vs2, vl); } @@ -1223,7 +1223,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_tumu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_tumu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_tumu(vbool4_t vm, vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tumu(vm, vd, vs2, vl); } @@ -1234,7 +1234,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_tumu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_mu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_q_f8e5m2mf8_mu(vbool64_t vm, vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_mu(vm, vd, vs2, vl); } @@ -1245,7 +1245,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_mu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_mu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_mu(vbool64_t vm, vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_mu(vm, vd, vs2, vl); } @@ -1256,7 +1256,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_mu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_mu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_q_f8e5m2mf4_mu(vbool32_t vm, vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_mu(vm, vd, vs2, vl); } @@ -1267,7 +1267,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_mu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_mu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_mu(vbool32_t vm, vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_mu(vm, vd, vs2, vl); } @@ -1278,7 +1278,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_mu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_mu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_q_f8e5m2mf2_mu(vbool16_t vm, vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_mu(vm, vd, vs2, vl); } @@ -1289,7 +1289,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_mu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_mu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_mu(vbool16_t vm, vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_mu(vm, vd, vs2, vl); } @@ -1300,7 +1300,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_mu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_q_f8e5m2m1_mu(vbool8_t vm, vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_mu(vm, vd, vs2, vl); } @@ -1311,7 +1311,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_mu(vbool8_t vm, vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_mu(vm, vd, vs2, vl); } @@ -1322,7 +1322,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_q_f8e5m2m2_mu(vbool4_t vm, vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_mu(vm, vd, vs2, vl); } @@ -1333,7 +1333,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_mu(vbool4_t vm, vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_mu(vm, vd, vs2, vl); } @@ -1344,7 +1344,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_tu(vuint8mf8_t vd, vfloat32mf2_t vs2, +vfloat8e5m2mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_tu(vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1355,7 +1355,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_tu(vuint8mf8_t vd, vfloat32mf2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_tu(vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_tu(vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); @@ -1367,7 +1367,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_tu(vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_tu(vuint8mf4_t vd, vfloat32m1_t vs2, +vfloat8e5m2mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_tu(vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1378,7 +1378,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_tu(vuint8mf4_t vd, vfloat32m1_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_tu(vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_tu(vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1389,7 +1389,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_tu(vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_tu(vuint8mf2_t vd, vfloat32m2_t vs2, +vfloat8e5m2mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_tu(vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1400,7 +1400,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_tu(vuint8mf2_t vd, vfloat32m2_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_tu(vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_tu(vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1411,7 +1411,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_tu(vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_tu(vuint8m1_t vd, vfloat32m4_t vs2, +vfloat8e5m2m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_tu(vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1422,7 +1422,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_tu(vuint8m1_t vd, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_tu(vuint8m1_t vd, vfloat32m4_t vs2, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_tu(vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1433,7 +1433,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_tu(vuint8m1_t vd, vfloat32m4_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_tu(vuint8m2_t vd, vfloat32m8_t vs2, +vfloat8e5m2m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_tu(vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1444,7 +1444,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_tu(vuint8m2_t vd, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_tu(vuint8m2_t vd, vfloat32m8_t vs2, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_tu(vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1455,7 +1455,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_tu(vuint8m2_t vd, vfloat32m8_t vs2, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_tum(vbool64_t vm, vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_tum(vbool64_t vm, vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1466,7 +1466,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_tum(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_tum(vbool64_t vm, vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_tum(vbool64_t vm, vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -1478,7 +1478,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_tum(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_tum(vbool32_t vm, vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_tum(vbool32_t vm, vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1489,7 +1489,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_tum(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_tum(vbool32_t vm, vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_tum(vbool32_t vm, vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -1501,7 +1501,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_tum(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_tum(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_tum(vbool16_t vm, vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1512,7 +1512,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_tum(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_tum(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_tum(vbool16_t vm, vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -1524,7 +1524,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_tum(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_tum(vbool8_t vm, vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1535,7 +1535,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_tum(vbool8_t vm, vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1546,7 +1546,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_tum(vbool4_t vm, vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1557,7 +1557,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_tum(vbool4_t vm, vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1568,7 +1568,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_tumu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_tumu(vbool64_t vm, vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1579,8 +1579,8 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_tumu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_tumu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_tumu(vbool64_t vm, + vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -1592,7 +1592,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_tumu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_tumu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_tumu(vbool32_t vm, vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1603,8 +1603,8 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_tumu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_tumu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_tumu(vbool32_t vm, + vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -1616,7 +1616,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_tumu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_tumu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_tumu(vbool16_t vm, vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1627,8 +1627,8 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_tumu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_tumu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_tumu(vbool16_t vm, + vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -1640,7 +1640,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_tumu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_tumu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_tumu(vbool8_t vm, vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1651,7 +1651,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_tumu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_tumu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_tumu(vbool8_t vm, vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1662,7 +1662,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_tumu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_tumu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_tumu(vbool4_t vm, vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1673,7 +1673,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_tumu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_tumu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_tumu(vbool4_t vm, vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1684,7 +1684,7 @@ vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_tumu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_mu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_mu(vbool64_t vm, vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1695,7 +1695,7 @@ vuint8mf8_t test_vfncvt_f_f_q_f8e5m2mf8_rm_mu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv1i8.nxv1f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_mu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_mu(vbool64_t vm, vfloat8e5m2mf8_t vd, vfloat32mf2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); @@ -1707,7 +1707,7 @@ vuint8mf8_t test_vfncvt_sat_f_f_q_f8e5m2mf8_rm_mu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_mu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_mu(vbool32_t vm, vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1718,7 +1718,7 @@ vuint8mf4_t test_vfncvt_f_f_q_f8e5m2mf4_rm_mu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv2i8.nxv2f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_mu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_mu(vbool32_t vm, vfloat8e5m2mf4_t vd, vfloat32m1_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1729,7 +1729,7 @@ vuint8mf4_t test_vfncvt_sat_f_f_q_f8e5m2mf4_rm_mu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_mu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_mu(vbool16_t vm, vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1740,7 +1740,7 @@ vuint8mf2_t test_vfncvt_f_f_q_f8e5m2mf2_rm_mu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv4i8.nxv4f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_mu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_mu(vbool16_t vm, vfloat8e5m2mf2_t vd, vfloat32m2_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1751,7 +1751,7 @@ vuint8mf2_t test_vfncvt_sat_f_f_q_f8e5m2mf2_rm_mu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_mu(vbool8_t vm, vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1762,7 +1762,7 @@ vuint8m1_t test_vfncvt_f_f_q_f8e5m2m1_rm_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv8i8.nxv8f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_mu(vbool8_t vm, vfloat8e5m2m1_t vd, vfloat32m4_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1773,7 +1773,7 @@ vuint8m1_t test_vfncvt_sat_f_f_q_f8e5m2m1_rm_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_mu(vbool4_t vm, vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_f_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1784,7 +1784,7 @@ vuint8m2_t test_vfncvt_f_f_q_f8e5m2m2_rm_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.q.alt.mask.nxv16i8.nxv16f32.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_q_f8e5m2m2_rm_mu(vbool4_t vm, vfloat8e5m2m2_t vd, vfloat32m8_t vs2, size_t vl) { return __riscv_vfncvt_sat_f_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } diff --git a/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/policy/overloaded/vfncvtbf16.c b/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/policy/overloaded/vfncvtbf16.c index 6507e0bc9bd2..22db786450ac 100644 --- a/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/policy/overloaded/vfncvtbf16.c +++ b/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/policy/overloaded/vfncvtbf16.c @@ -13,10 +13,10 @@ // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_tu(vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_tu(vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tu(vd, vs2, vl); + return __riscv_vfncvt_f_f8e4m3_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_tu( @@ -25,10 +25,10 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_tu(vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_tu(vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_tu(vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tu(vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_tu( @@ -37,10 +37,10 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_tu(vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_tu(vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_tu(vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tu(vd, vs2, vl); + return __riscv_vfncvt_f_f8e4m3_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_tu( @@ -49,10 +49,10 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_tu(vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_tu(vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_tu(vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tu(vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_tu( @@ -61,10 +61,10 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_tu(vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_tu(vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_tu(vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tu(vd, vs2, vl); + return __riscv_vfncvt_f_f8e4m3_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_tu( @@ -73,10 +73,10 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_tu(vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_tu(vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_tu(vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tu(vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m2_f8e4m3m1_tu( @@ -85,9 +85,9 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_tu(vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_tu(vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_tu(vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tu(vd, vs2, vl); + return __riscv_vfncvt_f_f8e4m3_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_tu( @@ -96,10 +96,10 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_tu(vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_tu(vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_tu(vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tu(vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m4_f8e4m3m2_tu( @@ -108,9 +108,9 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_tu(vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_tu(vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_tu(vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tu(vd, vs2, vl); + return __riscv_vfncvt_f_f8e4m3_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_tu( @@ -119,10 +119,10 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_tu(vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_tu(vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_tu(vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tu(vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m8_f8e4m3m4_tu( @@ -131,9 +131,9 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_tu(vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_tu(vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_tu(vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tu(vd, vs2, vl); + return __riscv_vfncvt_f_f8e4m3_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_tu( @@ -142,10 +142,10 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_tu(vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_tu(vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_tu(vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tu(vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_tum( @@ -154,11 +154,11 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_tu(vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_tum(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_tum(vbool64_t vm, + vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tum(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e4m3_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_tum( @@ -167,11 +167,11 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_tum(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_tum(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_tum(vbool64_t vm, + vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tum(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_tum( @@ -180,11 +180,11 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_tum(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_tum(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_tum(vbool32_t vm, + vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tum(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e4m3_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_tum( @@ -193,11 +193,11 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_tum(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_tum(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_tum(vbool32_t vm, + vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tum(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_tum( @@ -206,10 +206,10 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_tum(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_tum(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_tum(vbool16_t vm, vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tum(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e4m3_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_tum( @@ -218,11 +218,11 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_tum(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_tum(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_tum(vbool16_t vm, + vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tum(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m2_f8e4m3m1_tum( @@ -231,9 +231,9 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_tum(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_tum(vbool8_t vm, vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tum(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e4m3_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_tum( @@ -242,10 +242,10 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_tum(vbool8_t vm, vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tum(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m4_f8e4m3m2_tum( @@ -254,9 +254,9 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_tum(vbool4_t vm, vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tum(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e4m3_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_tum( @@ -265,10 +265,10 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_tum(vbool4_t vm, vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tum(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m8_f8e4m3m4_tum( @@ -277,9 +277,9 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_tum(vbool2_t vm, vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_tum(vbool2_t vm, vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tum(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e4m3_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_tum( @@ -288,10 +288,10 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_tum(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_tum(vbool2_t vm, vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_tum(vbool2_t vm, vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tum(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_tumu( @@ -300,11 +300,11 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_tum(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_tumu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_tumu(vbool64_t vm, + vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tumu(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e4m3_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_tumu( @@ -313,11 +313,11 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_tumu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_tumu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_tumu(vbool64_t vm, + vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tumu(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_tumu( @@ -326,11 +326,11 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_tumu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_tumu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_tumu(vbool32_t vm, + vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tumu(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e4m3_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_tumu( @@ -339,11 +339,11 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_tumu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_tumu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_tumu(vbool32_t vm, + vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tumu(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_tumu( @@ -352,11 +352,11 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_tumu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_tumu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_tumu(vbool16_t vm, + vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tumu(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e4m3_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_tumu( @@ -365,11 +365,11 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_tumu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_tumu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_tumu(vbool16_t vm, + vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tumu(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m2_f8e4m3m1_tumu( @@ -378,10 +378,10 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_tumu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_tumu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_tumu(vbool8_t vm, vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tumu(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e4m3_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_tumu( @@ -390,11 +390,11 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_tumu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_tumu(vbool8_t vm, - vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_tumu(vbool8_t vm, + vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tumu(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m4_f8e4m3m2_tumu( @@ -403,10 +403,10 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_tumu(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_tumu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_tumu(vbool4_t vm, vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tumu(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e4m3_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_tumu( @@ -415,11 +415,11 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_tumu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_tumu(vbool4_t vm, - vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_tumu(vbool4_t vm, + vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tumu(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m8_f8e4m3m4_tumu( @@ -428,10 +428,10 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_tumu(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_tumu(vbool2_t vm, vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_tumu(vbool2_t vm, vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tumu(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e4m3_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_tumu( @@ -440,11 +440,11 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_tumu(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_tumu(vbool2_t vm, - vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_tumu(vbool2_t vm, + vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tumu(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_mu( @@ -453,10 +453,10 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_tumu(vbool2_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_mu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_mu(vbool64_t vm, vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_mu(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e4m3_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_mu( @@ -465,11 +465,11 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_mu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_mu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_mu(vbool64_t vm, + vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_mu(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_mu( @@ -478,10 +478,10 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_mu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_mu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_mu(vbool32_t vm, vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_mu(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e4m3_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_mu( @@ -490,11 +490,11 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_mu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_mu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_mu(vbool32_t vm, + vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_mu(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_mu( @@ -503,10 +503,10 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_mu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_mu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_mu(vbool16_t vm, vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_mu(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e4m3_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_mu( @@ -515,11 +515,11 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_mu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_mu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_mu(vbool16_t vm, + vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_mu(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m2_f8e4m3m1_mu( @@ -528,9 +528,9 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_mu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_mu(vbool8_t vm, vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_mu(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e4m3_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_mu( @@ -539,10 +539,10 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_mu(vbool8_t vm, vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_mu(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m4_f8e4m3m2_mu( @@ -551,9 +551,9 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_mu(vbool4_t vm, vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_mu(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e4m3_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_mu( @@ -562,10 +562,10 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_mu(vbool4_t vm, vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_mu(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m8_f8e4m3m4_mu( @@ -574,9 +574,9 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_mu(vbool2_t vm, vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_mu(vbool2_t vm, vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_mu(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e4m3_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_mu( @@ -585,10 +585,10 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_mu(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_mu(vbool2_t vm, vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_mu(vbool2_t vm, vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_mu(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e4m3_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_tu( @@ -597,10 +597,10 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_mu(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_tu(vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_tu(vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_tu( @@ -609,10 +609,10 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_tu(vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_tu(vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_tu(vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_tu( @@ -621,10 +621,10 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_tu(vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_tu(vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_tu(vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_tu( @@ -633,10 +633,10 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_tu(vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_tu(vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_tu(vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_tu( @@ -645,10 +645,10 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_tu(vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_tu(vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_tu(vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_tu( @@ -657,10 +657,10 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_tu(vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_tu(vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_tu(vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_tu( @@ -669,10 +669,10 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_tu(vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_tu(vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_tu(vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_tu( @@ -681,10 +681,10 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_tu(vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_tu(vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_tu(vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_tu( @@ -693,10 +693,10 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_tu(vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_tu(vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_tu(vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_tu( @@ -705,10 +705,10 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_tu(vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_tu(vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_tu(vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_tu( @@ -717,10 +717,10 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_tu(vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_tu(vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_tu(vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_tu( @@ -729,10 +729,10 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_tu(vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_tu(vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_tu(vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tu(vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_tum( @@ -741,11 +741,11 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_tu(vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_tum(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_tum(vbool64_t vm, + vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_tum( @@ -754,11 +754,11 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_tum(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_tum(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_tum(vbool64_t vm, + vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_tum( @@ -767,11 +767,11 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_tum(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_tum(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_tum(vbool32_t vm, + vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_tum( @@ -780,11 +780,11 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_tum(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_tum(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_tum(vbool32_t vm, + vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_tum( @@ -793,11 +793,11 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_tum(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_tum(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_tum(vbool16_t vm, + vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_tum( @@ -806,11 +806,11 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_tum(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_tum(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_tum(vbool16_t vm, + vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_tum( @@ -819,10 +819,10 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_tum(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_tum(vbool8_t vm, vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_tum( @@ -831,11 +831,11 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_tum(vbool8_t vm, - vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_tum(vbool8_t vm, + vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_tum( @@ -844,10 +844,10 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_tum(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_tum(vbool4_t vm, vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_tum( @@ -856,11 +856,11 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_tum(vbool4_t vm, - vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_tum(vbool4_t vm, + vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_tum( @@ -869,10 +869,10 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_tum(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_tum(vbool2_t vm, vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_tum(vbool2_t vm, vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_tum( @@ -881,11 +881,11 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_tum(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_tum(vbool2_t vm, - vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_tum(vbool2_t vm, + vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_tumu( @@ -894,11 +894,11 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_tum(vbool2_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_tumu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_tumu(vbool64_t vm, + vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_tumu( @@ -907,11 +907,11 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_tumu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_tumu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_tumu(vbool64_t vm, + vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, + return __riscv_vfncvt_sat_f_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -921,11 +921,11 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_tumu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_tumu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_tumu(vbool32_t vm, + vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_tumu( @@ -934,11 +934,11 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_tumu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_tumu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_tumu(vbool32_t vm, + vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, + return __riscv_vfncvt_sat_f_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -948,11 +948,11 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_tumu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_tumu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_tumu(vbool16_t vm, + vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_tumu( @@ -961,11 +961,11 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_tumu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_tumu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_tumu(vbool16_t vm, + vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, + return __riscv_vfncvt_sat_f_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -975,10 +975,10 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_tumu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_tumu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_tumu(vbool8_t vm, vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_tumu( @@ -987,11 +987,11 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_tumu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_tumu(vbool8_t vm, - vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_tumu(vbool8_t vm, + vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, + return __riscv_vfncvt_sat_f_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1001,10 +1001,10 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_tumu(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_tumu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_tumu(vbool4_t vm, vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_tumu( @@ -1013,11 +1013,11 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_tumu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_tumu(vbool4_t vm, - vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_tumu(vbool4_t vm, + vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, + return __riscv_vfncvt_sat_f_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1027,10 +1027,10 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_tumu(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_tumu(vbool2_t vm, vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_tumu(vbool2_t vm, vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_tumu( @@ -1039,11 +1039,11 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_tumu(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_tumu(vbool2_t vm, - vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_tumu(vbool2_t vm, + vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, + return __riscv_vfncvt_sat_f_f8e4m3_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -1053,11 +1053,11 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_tumu(vbool2_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_mu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_mu(vbool64_t vm, + vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_mu( @@ -1066,11 +1066,11 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e4m3mf8_rm_mu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_mu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e4m3mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_mu(vbool64_t vm, + vfloat8e4m3mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_mu( @@ -1079,11 +1079,11 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e4m3mf8_rm_mu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_mu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_mu(vbool32_t vm, + vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_mu( @@ -1092,11 +1092,11 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e4m3mf4_rm_mu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_mu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e4m3mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_mu(vbool32_t vm, + vfloat8e4m3mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_mu( @@ -1105,11 +1105,11 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e4m3mf4_rm_mu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_mu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_mu(vbool16_t vm, + vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_mu( @@ -1118,11 +1118,11 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e4m3mf2_rm_mu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_mu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e4m3mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_mu(vbool16_t vm, + vfloat8e4m3mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_mu( @@ -1131,10 +1131,10 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e4m3mf2_rm_mu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_mu(vbool8_t vm, vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_mu( @@ -1143,11 +1143,11 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e4m3m1_rm_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_mu(vbool8_t vm, - vuint8m1_t vd, +vfloat8e4m3m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_mu(vbool8_t vm, + vfloat8e4m3m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_mu( @@ -1156,10 +1156,10 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e4m3m1_rm_mu(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_mu(vbool4_t vm, vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_mu( @@ -1168,11 +1168,11 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e4m3m2_rm_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_mu(vbool4_t vm, - vuint8m2_t vd, +vfloat8e4m3m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_mu(vbool4_t vm, + vfloat8e4m3m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_mu( @@ -1181,10 +1181,10 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e4m3m2_rm_mu(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_mu(vbool2_t vm, vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_mu(vbool2_t vm, vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_mu( @@ -1193,11 +1193,11 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e4m3m4_rm_mu(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_mu(vbool2_t vm, - vuint8m4_t vd, +vfloat8e4m3m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_mu(vbool2_t vm, + vfloat8e4m3m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e4m3_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_tu( @@ -1206,10 +1206,10 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e4m3m4_rm_mu(vbool2_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_tu(vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_tu(vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tu(vd, vs2, vl); + return __riscv_vfncvt_f_f8e5m2_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_tu( @@ -1218,10 +1218,10 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_tu(vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_tu(vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_tu(vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tu(vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_tu( @@ -1230,10 +1230,10 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_tu(vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_tu(vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_tu(vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tu(vd, vs2, vl); + return __riscv_vfncvt_f_f8e5m2_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_tu( @@ -1242,10 +1242,10 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_tu(vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_tu(vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_tu(vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tu(vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_tu( @@ -1254,10 +1254,10 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_tu(vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_tu(vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_tu(vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tu(vd, vs2, vl); + return __riscv_vfncvt_f_f8e5m2_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_tu( @@ -1266,10 +1266,10 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_tu(vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_tu(vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_tu(vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tu(vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m2_f8e5m2m1_tu( @@ -1278,9 +1278,9 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_tu(vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_tu(vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_tu(vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tu(vd, vs2, vl); + return __riscv_vfncvt_f_f8e5m2_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_tu( @@ -1289,10 +1289,10 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_tu(vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_tu(vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_tu(vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tu(vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m4_f8e5m2m2_tu( @@ -1301,9 +1301,9 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_tu(vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_tu(vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_tu(vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tu(vd, vs2, vl); + return __riscv_vfncvt_f_f8e5m2_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_tu( @@ -1312,10 +1312,10 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_tu(vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_tu(vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_tu(vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tu(vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m8_f8e5m2m4_tu( @@ -1324,9 +1324,9 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_tu(vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_tu(vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_tu(vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tu(vd, vs2, vl); + return __riscv_vfncvt_f_f8e5m2_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_tu( @@ -1335,10 +1335,10 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_tu(vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], i64 7, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_tu(vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_tu(vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tu(vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_tum( @@ -1347,11 +1347,11 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_tu(vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_tum(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_tum(vbool64_t vm, + vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tum(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e5m2_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_tum( @@ -1360,11 +1360,11 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_tum(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_tum(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_tum(vbool64_t vm, + vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tum(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_tum( @@ -1373,11 +1373,11 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_tum(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_tum(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_tum(vbool32_t vm, + vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tum(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e5m2_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_tum( @@ -1386,11 +1386,11 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_tum(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_tum(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_tum(vbool32_t vm, + vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tum(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_tum( @@ -1399,10 +1399,10 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_tum(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_tum(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_tum(vbool16_t vm, vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tum(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e5m2_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_tum( @@ -1411,11 +1411,11 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_tum(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_tum(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_tum(vbool16_t vm, + vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tum(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m2_f8e5m2m1_tum( @@ -1424,9 +1424,9 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_tum(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_tum(vbool8_t vm, vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tum(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e5m2_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_tum( @@ -1435,10 +1435,10 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_tum(vbool8_t vm, vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tum(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m4_f8e5m2m2_tum( @@ -1447,9 +1447,9 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_tum(vbool4_t vm, vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tum(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e5m2_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_tum( @@ -1458,10 +1458,10 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_tum(vbool4_t vm, vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tum(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m8_f8e5m2m4_tum( @@ -1470,9 +1470,9 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_tum(vbool2_t vm, vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_tum(vbool2_t vm, vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tum(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e5m2_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_tum( @@ -1481,10 +1481,10 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_tum(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_tum(vbool2_t vm, vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_tum(vbool2_t vm, vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tum(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_tumu( @@ -1493,11 +1493,11 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_tum(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_tumu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_tumu(vbool64_t vm, + vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tumu(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e5m2_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_tumu( @@ -1506,11 +1506,11 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_tumu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_tumu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_tumu(vbool64_t vm, + vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tumu(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_tumu( @@ -1519,11 +1519,11 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_tumu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_tumu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_tumu(vbool32_t vm, + vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tumu(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e5m2_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_tumu( @@ -1532,11 +1532,11 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_tumu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_tumu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_tumu(vbool32_t vm, + vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tumu(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_tumu( @@ -1545,11 +1545,11 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_tumu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_tumu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_tumu(vbool16_t vm, + vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tumu(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e5m2_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_tumu( @@ -1558,11 +1558,11 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_tumu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_tumu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_tumu(vbool16_t vm, + vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tumu(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m2_f8e5m2m1_tumu( @@ -1571,10 +1571,10 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_tumu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_tumu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_tumu(vbool8_t vm, vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tumu(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e5m2_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_tumu( @@ -1583,11 +1583,11 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_tumu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_tumu(vbool8_t vm, - vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_tumu(vbool8_t vm, + vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tumu(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m4_f8e5m2m2_tumu( @@ -1596,10 +1596,10 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_tumu(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_tumu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_tumu(vbool4_t vm, vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tumu(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e5m2_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_tumu( @@ -1608,11 +1608,11 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_tumu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_tumu(vbool4_t vm, - vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_tumu(vbool4_t vm, + vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tumu(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m8_f8e5m2m4_tumu( @@ -1621,10 +1621,10 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_tumu(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_tumu(vbool2_t vm, vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_tumu(vbool2_t vm, vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tumu(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e5m2_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_tumu( @@ -1633,11 +1633,11 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_tumu(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_tumu(vbool2_t vm, - vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_tumu(vbool2_t vm, + vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tumu(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_mu( @@ -1646,10 +1646,10 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_tumu(vbool2_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_mu(vbool64_t vm, vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_mu(vbool64_t vm, vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_mu(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e5m2_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_mu( @@ -1658,11 +1658,11 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_mu(vbool64_t vm, vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_mu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_mu(vbool64_t vm, + vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_mu(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_mu( @@ -1671,10 +1671,10 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_mu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_mu(vbool32_t vm, vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_mu(vbool32_t vm, vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_mu(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e5m2_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_mu( @@ -1683,11 +1683,11 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_mu(vbool32_t vm, vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_mu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_mu(vbool32_t vm, + vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_mu(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_mu( @@ -1696,10 +1696,10 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_mu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_mu(vbool16_t vm, vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_mu(vbool16_t vm, vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_mu(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e5m2_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_mu( @@ -1708,11 +1708,11 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_mu(vbool16_t vm, vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_mu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_mu(vbool16_t vm, + vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_mu(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m2_f8e5m2m1_mu( @@ -1721,9 +1721,9 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_mu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_mu(vbool8_t vm, vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_mu(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e5m2_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_mu( @@ -1732,10 +1732,10 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_mu(vbool8_t vm, vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_mu(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m4_f8e5m2m2_mu( @@ -1744,9 +1744,9 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_mu(vbool4_t vm, vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_mu(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e5m2_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_mu( @@ -1755,10 +1755,10 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_mu(vbool4_t vm, vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_mu(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m8_f8e5m2m4_mu( @@ -1767,9 +1767,9 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_mu(vbool2_t vm, vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_mu(vbool2_t vm, vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_mu(vm, vd, vs2, vl); + return __riscv_vfncvt_f_f8e5m2_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_mu( @@ -1778,10 +1778,10 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_mu(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 7, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_mu(vbool2_t vm, vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_mu(vbool2_t vm, vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_mu(vm, vd, vs2, vl); + return __riscv_vfncvt_sat_f_f8e5m2_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_tu( @@ -1790,10 +1790,10 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_mu(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_tu(vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_tu(vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_tu( @@ -1802,10 +1802,10 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_tu(vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_tu(vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_tu(vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_tu( @@ -1814,10 +1814,10 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_tu(vuint8mf8_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_tu(vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_tu(vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_tu( @@ -1826,10 +1826,10 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_tu(vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_tu(vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_tu(vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_tu( @@ -1838,10 +1838,10 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_tu(vuint8mf4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_tu(vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_tu(vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_tu( @@ -1850,10 +1850,10 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_tu(vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_tu(vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_tu(vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_tu( @@ -1862,10 +1862,10 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_tu(vuint8mf2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_tu(vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_tu(vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_tu( @@ -1874,10 +1874,10 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_tu(vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_tu(vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_tu(vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_tu( @@ -1886,10 +1886,10 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_tu(vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_tu(vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_tu(vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_tu( @@ -1898,10 +1898,10 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_tu(vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_tu(vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_tu(vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_tu( @@ -1910,10 +1910,10 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_tu(vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_tu(vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_tu(vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_tu( @@ -1922,10 +1922,10 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_tu(vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], i64 0, i64 [[VL]]) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_tu(vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_tu(vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tu(vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_tum( @@ -1934,11 +1934,11 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_tu(vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_tum(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_tum(vbool64_t vm, + vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_tum( @@ -1947,11 +1947,11 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_tum(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_tum(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_tum(vbool64_t vm, + vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_tum( @@ -1960,11 +1960,11 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_tum(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_tum(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_tum(vbool32_t vm, + vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_tum( @@ -1973,11 +1973,11 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_tum(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_tum(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_tum(vbool32_t vm, + vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_tum( @@ -1986,11 +1986,11 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_tum(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_tum(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_tum(vbool16_t vm, + vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_tum( @@ -1999,11 +1999,11 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_tum(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_tum(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_tum(vbool16_t vm, + vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_tum( @@ -2012,10 +2012,10 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_tum(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_tum(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_tum(vbool8_t vm, vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_tum( @@ -2024,11 +2024,11 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_tum(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_tum(vbool8_t vm, - vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_tum(vbool8_t vm, + vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_tum( @@ -2037,10 +2037,10 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_tum(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_tum(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_tum(vbool4_t vm, vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_tum( @@ -2049,11 +2049,11 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_tum(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_tum(vbool4_t vm, - vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_tum(vbool4_t vm, + vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_tum( @@ -2062,10 +2062,10 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_tum(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_tum(vbool2_t vm, vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_tum(vbool2_t vm, vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_tum( @@ -2074,11 +2074,11 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_tum(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 2) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_tum(vbool2_t vm, - vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_tum(vbool2_t vm, + vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2_tum(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_tumu( @@ -2087,11 +2087,11 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_tum(vbool2_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_tumu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_tumu(vbool64_t vm, + vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_tumu( @@ -2100,11 +2100,11 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_tumu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_tumu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_tumu(vbool64_t vm, + vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, + return __riscv_vfncvt_sat_f_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -2114,11 +2114,11 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_tumu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_tumu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_tumu(vbool32_t vm, + vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_tumu( @@ -2127,11 +2127,11 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_tumu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_tumu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_tumu(vbool32_t vm, + vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, + return __riscv_vfncvt_sat_f_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -2141,11 +2141,11 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_tumu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_tumu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_tumu(vbool16_t vm, + vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_tumu( @@ -2154,11 +2154,11 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_tumu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_tumu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_tumu(vbool16_t vm, + vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, + return __riscv_vfncvt_sat_f_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -2168,10 +2168,10 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_tumu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_tumu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_tumu(vbool8_t vm, vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_tumu( @@ -2180,11 +2180,11 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_tumu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_tumu(vbool8_t vm, - vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_tumu(vbool8_t vm, + vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, + return __riscv_vfncvt_sat_f_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -2194,10 +2194,10 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_tumu(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_tumu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_tumu(vbool4_t vm, vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_tumu( @@ -2206,11 +2206,11 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_tumu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_tumu(vbool4_t vm, - vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_tumu(vbool4_t vm, + vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, + return __riscv_vfncvt_sat_f_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -2220,10 +2220,10 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_tumu(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_tumu(vbool2_t vm, vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_tumu(vbool2_t vm, vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_tumu( @@ -2232,11 +2232,11 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_tumu(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 0) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_tumu(vbool2_t vm, - vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_tumu(vbool2_t vm, + vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, + return __riscv_vfncvt_sat_f_f8e5m2_tumu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } @@ -2246,11 +2246,11 @@ vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_tumu(vbool2_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_mu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_mu(vbool64_t vm, + vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_mu( @@ -2259,11 +2259,11 @@ vuint8mf8_t test_vfncvt_f_f_w_bf16mf4_f8e5m2mf8_rm_mu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv1i8.nxv1bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_mu(vbool64_t vm, - vuint8mf8_t vd, +vfloat8e5m2mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_mu(vbool64_t vm, + vfloat8e5m2mf8_t vd, vbfloat16mf4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_mu( @@ -2272,11 +2272,11 @@ vuint8mf8_t test_vfncvt_sat_f_f_w_bf16mf4_f8e5m2mf8_rm_mu(vbool64_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_mu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_mu(vbool32_t vm, + vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_mu( @@ -2285,11 +2285,11 @@ vuint8mf4_t test_vfncvt_f_f_w_bf16mf2_f8e5m2mf4_rm_mu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv2i8.nxv2bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_mu(vbool32_t vm, - vuint8mf4_t vd, +vfloat8e5m2mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_mu(vbool32_t vm, + vfloat8e5m2mf4_t vd, vbfloat16mf2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_mu( @@ -2298,11 +2298,11 @@ vuint8mf4_t test_vfncvt_sat_f_f_w_bf16mf2_f8e5m2mf4_rm_mu(vbool32_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_mu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_mu(vbool16_t vm, + vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_mu( @@ -2311,11 +2311,11 @@ vuint8mf2_t test_vfncvt_f_f_w_bf16m1_f8e5m2mf2_rm_mu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv4i8.nxv4bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_mu(vbool16_t vm, - vuint8mf2_t vd, +vfloat8e5m2mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_mu(vbool16_t vm, + vfloat8e5m2mf2_t vd, vbfloat16m1_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_mu( @@ -2324,10 +2324,10 @@ vuint8mf2_t test_vfncvt_sat_f_f_w_bf16m1_f8e5m2mf2_rm_mu(vbool16_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_mu(vbool8_t vm, vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_mu(vbool8_t vm, vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_mu( @@ -2336,11 +2336,11 @@ vuint8m1_t test_vfncvt_f_f_w_bf16m2_f8e5m2m1_rm_mu(vbool8_t vm, vuint8m1_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv8i8.nxv8bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_mu(vbool8_t vm, - vuint8m1_t vd, +vfloat8e5m2m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_mu(vbool8_t vm, + vfloat8e5m2m1_t vd, vbfloat16m2_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_mu( @@ -2349,10 +2349,10 @@ vuint8m1_t test_vfncvt_sat_f_f_w_bf16m2_f8e5m2m1_rm_mu(vbool8_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_mu(vbool4_t vm, vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_mu(vbool4_t vm, vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_mu( @@ -2361,11 +2361,11 @@ vuint8m2_t test_vfncvt_f_f_w_bf16m4_f8e5m2m2_rm_mu(vbool4_t vm, vuint8m2_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv16i8.nxv16bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_mu(vbool4_t vm, - vuint8m2_t vd, +vfloat8e5m2m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_mu(vbool4_t vm, + vfloat8e5m2m2_t vd, vbfloat16m4_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_mu( @@ -2374,10 +2374,10 @@ vuint8m2_t test_vfncvt_sat_f_f_w_bf16m4_f8e5m2m2_rm_mu(vbool4_t vm, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_mu(vbool2_t vm, vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_mu(vbool2_t vm, vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_f_bf16_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_f_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_mu( @@ -2386,9 +2386,9 @@ vuint8m4_t test_vfncvt_f_f_w_bf16m8_f8e5m2m4_rm_mu(vbool2_t vm, vuint8m4_t vd, // CHECK-RV64-NEXT: [[TMP0:%.*]] = call @llvm.riscv.vfncvt.sat.f.f.w.alt.mask.nxv32i8.nxv32bf16.i64( [[VD]], [[VS2]], [[VM]], i64 0, i64 [[VL]], i64 1) // CHECK-RV64-NEXT: ret [[TMP0]] // -vuint8m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_mu(vbool2_t vm, - vuint8m4_t vd, +vfloat8e5m2m4_t test_vfncvt_sat_f_f_w_bf16m8_f8e5m2m4_rm_mu(vbool2_t vm, + vfloat8e5m2m4_t vd, vbfloat16m8_t vs2, size_t vl) { - return __riscv_vfncvt_sat_f_bf16_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); + return __riscv_vfncvt_sat_f_f8e5m2_mu(vm, vd, vs2, __RISCV_FRM_RNE, vl); } diff --git a/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/policy/overloaded/vfwcvtbf16.c b/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/policy/overloaded/vfwcvtbf16.c index 278642931b58..a4aae9573431 100644 --- a/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/policy/overloaded/vfwcvtbf16.c +++ b/clang/test/CodeGen/RISCV/rvv-intrinsics-autogenerated/zvfofp8min/policy/overloaded/vfwcvtbf16.c @@ -14,9 +14,9 @@ // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_tu(vbfloat16mf4_t vd, - vuint8mf8_t vs2, + vfloat8e4m3mf8_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16_tu(vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_tu( @@ -26,9 +26,9 @@ vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_tu(vbfloat16mf4_t vd, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_tu(vbfloat16mf2_t vd, - vuint8mf4_t vs2, + vfloat8e4m3mf4_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16_tu(vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_tu( @@ -38,9 +38,9 @@ vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_tu(vbfloat16mf2_t vd, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_tu(vbfloat16m1_t vd, - vuint8mf2_t vs2, + vfloat8e4m3mf2_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16_tu(vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_tu( @@ -50,8 +50,8 @@ vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_tu(vbfloat16m1_t vd, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_tu(vbfloat16m2_t vd, - vuint8m1_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16_tu(vd, vs2, vl); + vfloat8e4m3m1_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_tu( @@ -61,8 +61,8 @@ vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_tu(vbfloat16m2_t vd, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_tu(vbfloat16m4_t vd, - vuint8m2_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16_tu(vd, vs2, vl); + vfloat8e4m3m2_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_tu( @@ -72,8 +72,8 @@ vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_tu(vbfloat16m4_t vd, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_tu(vbfloat16m8_t vd, - vuint8m4_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16_tu(vd, vs2, vl); + vfloat8e4m3m4_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_tum( @@ -84,9 +84,9 @@ vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_tu(vbfloat16m8_t vd, // vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_tum(vbool64_t vm, vbfloat16mf4_t vd, - vuint8mf8_t vs2, + vfloat8e4m3mf8_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16_tum(vm, vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_tum( @@ -97,9 +97,9 @@ vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_tum(vbool64_t vm, // vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_tum(vbool32_t vm, vbfloat16mf2_t vd, - vuint8mf4_t vs2, + vfloat8e4m3mf4_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16_tum(vm, vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_tum( @@ -110,9 +110,9 @@ vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_tum(vbool32_t vm, // vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_tum(vbool16_t vm, vbfloat16m1_t vd, - vuint8mf2_t vs2, + vfloat8e4m3mf2_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16_tum(vm, vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_tum( @@ -123,8 +123,8 @@ vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_tum(vbool16_t vm, // vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_tum(vbool8_t vm, vbfloat16m2_t vd, - vuint8m1_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16_tum(vm, vd, vs2, vl); + vfloat8e4m3m1_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_tum( @@ -135,8 +135,8 @@ vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_tum(vbool8_t vm, // vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_tum(vbool4_t vm, vbfloat16m4_t vd, - vuint8m2_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16_tum(vm, vd, vs2, vl); + vfloat8e4m3m2_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_tum( @@ -147,8 +147,8 @@ vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_tum(vbool4_t vm, // vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_tum(vbool2_t vm, vbfloat16m8_t vd, - vuint8m4_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16_tum(vm, vd, vs2, vl); + vfloat8e4m3m4_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_tumu( @@ -159,9 +159,9 @@ vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_tum(vbool2_t vm, // vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_tumu(vbool64_t vm, vbfloat16mf4_t vd, - vuint8mf8_t vs2, + vfloat8e4m3mf8_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16_tumu(vm, vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_tumu( @@ -172,9 +172,9 @@ vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_tumu(vbool64_t vm, // vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_tumu(vbool32_t vm, vbfloat16mf2_t vd, - vuint8mf4_t vs2, + vfloat8e4m3mf4_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16_tumu(vm, vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_tumu( @@ -185,9 +185,9 @@ vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_tumu(vbool32_t vm, // vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_tumu(vbool16_t vm, vbfloat16m1_t vd, - vuint8mf2_t vs2, + vfloat8e4m3mf2_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16_tumu(vm, vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_tumu( @@ -198,9 +198,9 @@ vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_tumu(vbool16_t vm, // vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_tumu(vbool8_t vm, vbfloat16m2_t vd, - vuint8m1_t vs2, + vfloat8e4m3m1_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16_tumu(vm, vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_tumu( @@ -211,9 +211,9 @@ vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_tumu(vbool8_t vm, // vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_tumu(vbool4_t vm, vbfloat16m4_t vd, - vuint8m2_t vs2, + vfloat8e4m3m2_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16_tumu(vm, vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_tumu( @@ -224,9 +224,9 @@ vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_tumu(vbool4_t vm, // vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_tumu(vbool2_t vm, vbfloat16m8_t vd, - vuint8m4_t vs2, + vfloat8e4m3m4_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16_tumu(vm, vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_mu( @@ -237,9 +237,9 @@ vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_tumu(vbool2_t vm, // vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_mu(vbool64_t vm, vbfloat16mf4_t vd, - vuint8mf8_t vs2, + vfloat8e4m3mf8_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16_mu(vm, vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_mu( @@ -250,9 +250,9 @@ vbfloat16mf4_t test_vfwcvt_f_f_v_f8e4m3mf8_bf16mf4_mu(vbool64_t vm, // vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_mu(vbool32_t vm, vbfloat16mf2_t vd, - vuint8mf4_t vs2, + vfloat8e4m3mf4_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16_mu(vm, vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_mu( @@ -263,9 +263,9 @@ vbfloat16mf2_t test_vfwcvt_f_f_v_f8e4m3mf4_bf16mf2_mu(vbool32_t vm, // vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_mu(vbool16_t vm, vbfloat16m1_t vd, - vuint8mf2_t vs2, + vfloat8e4m3mf2_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16_mu(vm, vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_mu( @@ -276,8 +276,8 @@ vbfloat16m1_t test_vfwcvt_f_f_v_f8e4m3mf2_bf16m1_mu(vbool16_t vm, // vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_mu(vbool8_t vm, vbfloat16m2_t vd, - vuint8m1_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16_mu(vm, vd, vs2, vl); + vfloat8e4m3m1_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_mu( @@ -288,8 +288,8 @@ vbfloat16m2_t test_vfwcvt_f_f_v_f8e4m3m1_bf16m2_mu(vbool8_t vm, // vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_mu(vbool4_t vm, vbfloat16m4_t vd, - vuint8m2_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16_mu(vm, vd, vs2, vl); + vfloat8e4m3m2_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_mu( @@ -300,8 +300,8 @@ vbfloat16m4_t test_vfwcvt_f_f_v_f8e4m3m2_bf16m4_mu(vbool4_t vm, // vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_mu(vbool2_t vm, vbfloat16m8_t vd, - vuint8m4_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e4m3_bf16_mu(vm, vd, vs2, vl); + vfloat8e4m3m4_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_tu( @@ -311,9 +311,9 @@ vbfloat16m8_t test_vfwcvt_f_f_v_f8e4m3m4_bf16m8_mu(vbool2_t vm, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_tu(vbfloat16mf4_t vd, - vuint8mf8_t vs2, + vfloat8e5m2mf8_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16_tu(vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_tu( @@ -323,9 +323,9 @@ vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_tu(vbfloat16mf4_t vd, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_tu(vbfloat16mf2_t vd, - vuint8mf4_t vs2, + vfloat8e5m2mf4_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16_tu(vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_tu( @@ -335,9 +335,9 @@ vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_tu(vbfloat16mf2_t vd, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_tu(vbfloat16m1_t vd, - vuint8mf2_t vs2, + vfloat8e5m2mf2_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16_tu(vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_tu( @@ -347,8 +347,8 @@ vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_tu(vbfloat16m1_t vd, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_tu(vbfloat16m2_t vd, - vuint8m1_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16_tu(vd, vs2, vl); + vfloat8e5m2m1_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_tu( @@ -358,8 +358,8 @@ vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_tu(vbfloat16m2_t vd, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_tu(vbfloat16m4_t vd, - vuint8m2_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16_tu(vd, vs2, vl); + vfloat8e5m2m2_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2m4_bf16m8_tu( @@ -369,8 +369,8 @@ vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_tu(vbfloat16m4_t vd, // CHECK-RV64-NEXT: ret [[TMP0]] // vbfloat16m8_t test_vfwcvt_f_f_v_f8e5m2m4_bf16m8_tu(vbfloat16m8_t vd, - vuint8m4_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16_tu(vd, vs2, vl); + vfloat8e5m2m4_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16_tu(vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_tum( @@ -381,9 +381,9 @@ vbfloat16m8_t test_vfwcvt_f_f_v_f8e5m2m4_bf16m8_tu(vbfloat16m8_t vd, // vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_tum(vbool64_t vm, vbfloat16mf4_t vd, - vuint8mf8_t vs2, + vfloat8e5m2mf8_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16_tum(vm, vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_tum( @@ -394,9 +394,9 @@ vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_tum(vbool64_t vm, // vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_tum(vbool32_t vm, vbfloat16mf2_t vd, - vuint8mf4_t vs2, + vfloat8e5m2mf4_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16_tum(vm, vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_tum( @@ -407,9 +407,9 @@ vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_tum(vbool32_t vm, // vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_tum(vbool16_t vm, vbfloat16m1_t vd, - vuint8mf2_t vs2, + vfloat8e5m2mf2_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16_tum(vm, vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_tum( @@ -420,8 +420,8 @@ vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_tum(vbool16_t vm, // vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_tum(vbool8_t vm, vbfloat16m2_t vd, - vuint8m1_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16_tum(vm, vd, vs2, vl); + vfloat8e5m2m1_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_tum( @@ -432,8 +432,8 @@ vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_tum(vbool8_t vm, // vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_tum(vbool4_t vm, vbfloat16m4_t vd, - vuint8m2_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16_tum(vm, vd, vs2, vl); + vfloat8e5m2m2_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2m4_bf16m8_tum( @@ -444,8 +444,8 @@ vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_tum(vbool4_t vm, // vbfloat16m8_t test_vfwcvt_f_f_v_f8e5m2m4_bf16m8_tum(vbool2_t vm, vbfloat16m8_t vd, - vuint8m4_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16_tum(vm, vd, vs2, vl); + vfloat8e5m2m4_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16_tum(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_tumu( @@ -456,9 +456,9 @@ vbfloat16m8_t test_vfwcvt_f_f_v_f8e5m2m4_bf16m8_tum(vbool2_t vm, // vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_tumu(vbool64_t vm, vbfloat16mf4_t vd, - vuint8mf8_t vs2, + vfloat8e5m2mf8_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16_tumu(vm, vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_tumu( @@ -469,9 +469,9 @@ vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_tumu(vbool64_t vm, // vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_tumu(vbool32_t vm, vbfloat16mf2_t vd, - vuint8mf4_t vs2, + vfloat8e5m2mf4_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16_tumu(vm, vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_tumu( @@ -482,9 +482,9 @@ vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_tumu(vbool32_t vm, // vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_tumu(vbool16_t vm, vbfloat16m1_t vd, - vuint8mf2_t vs2, + vfloat8e5m2mf2_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16_tumu(vm, vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_tumu( @@ -495,9 +495,9 @@ vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_tumu(vbool16_t vm, // vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_tumu(vbool8_t vm, vbfloat16m2_t vd, - vuint8m1_t vs2, + vfloat8e5m2m1_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16_tumu(vm, vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_tumu( @@ -508,9 +508,9 @@ vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_tumu(vbool8_t vm, // vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_tumu(vbool4_t vm, vbfloat16m4_t vd, - vuint8m2_t vs2, + vfloat8e5m2m2_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16_tumu(vm, vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2m4_bf16m8_tumu( @@ -521,9 +521,9 @@ vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_tumu(vbool4_t vm, // vbfloat16m8_t test_vfwcvt_f_f_v_f8e5m2m4_bf16m8_tumu(vbool2_t vm, vbfloat16m8_t vd, - vuint8m4_t vs2, + vfloat8e5m2m4_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16_tumu(vm, vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_tumu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_mu( @@ -534,9 +534,9 @@ vbfloat16m8_t test_vfwcvt_f_f_v_f8e5m2m4_bf16m8_tumu(vbool2_t vm, // vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_mu(vbool64_t vm, vbfloat16mf4_t vd, - vuint8mf8_t vs2, + vfloat8e5m2mf8_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16_mu(vm, vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_mu( @@ -547,9 +547,9 @@ vbfloat16mf4_t test_vfwcvt_f_f_v_f8e5m2mf8_bf16mf4_mu(vbool64_t vm, // vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_mu(vbool32_t vm, vbfloat16mf2_t vd, - vuint8mf4_t vs2, + vfloat8e5m2mf4_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16_mu(vm, vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_mu( @@ -560,9 +560,9 @@ vbfloat16mf2_t test_vfwcvt_f_f_v_f8e5m2mf4_bf16mf2_mu(vbool32_t vm, // vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_mu(vbool16_t vm, vbfloat16m1_t vd, - vuint8mf2_t vs2, + vfloat8e5m2mf2_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16_mu(vm, vd, vs2, vl); + return __riscv_vfwcvt_f_bf16_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_mu( @@ -573,8 +573,8 @@ vbfloat16m1_t test_vfwcvt_f_f_v_f8e5m2mf2_bf16m1_mu(vbool16_t vm, // vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_mu(vbool8_t vm, vbfloat16m2_t vd, - vuint8m1_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16_mu(vm, vd, vs2, vl); + vfloat8e5m2m1_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_mu( @@ -585,8 +585,8 @@ vbfloat16m2_t test_vfwcvt_f_f_v_f8e5m2m1_bf16m2_mu(vbool8_t vm, // vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_mu(vbool4_t vm, vbfloat16m4_t vd, - vuint8m2_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16_mu(vm, vd, vs2, vl); + vfloat8e5m2m2_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16_mu(vm, vd, vs2, vl); } // CHECK-RV64-LABEL: define dso_local @test_vfwcvt_f_f_v_f8e5m2m4_bf16m8_mu( @@ -597,6 +597,6 @@ vbfloat16m4_t test_vfwcvt_f_f_v_f8e5m2m2_bf16m4_mu(vbool4_t vm, // vbfloat16m8_t test_vfwcvt_f_f_v_f8e5m2m4_bf16m8_mu(vbool2_t vm, vbfloat16m8_t vd, - vuint8m4_t vs2, size_t vl) { - return __riscv_vfwcvt_f_f8e5m2_bf16_mu(vm, vd, vs2, vl); + vfloat8e5m2m4_t vs2, size_t vl) { + return __riscv_vfwcvt_f_bf16_mu(vm, vd, vs2, vl); } diff --git a/clang/test/CodeGen/RISCV/rvv-intrinsics-handcrafted/rvv-intrinsic-datatypes.cpp b/clang/test/CodeGen/RISCV/rvv-intrinsics-handcrafted/rvv-intrinsic-datatypes.cpp index 3511f21c063d..3e2ee71d989c 100644 --- a/clang/test/CodeGen/RISCV/rvv-intrinsics-handcrafted/rvv-intrinsic-datatypes.cpp +++ b/clang/test/CodeGen/RISCV/rvv-intrinsics-handcrafted/rvv-intrinsic-datatypes.cpp @@ -1,7 +1,7 @@ // NOTE: Assertions have been autogenerated by utils/update_cc_test_checks.py UTC_ARGS: --version 2 // REQUIRES: riscv-registered-target // RUN: %clang_cc1 -triple riscv64 -target-feature +v -target-feature +zvfh -target-feature +zvfbfmin \ -// RUN: -O0 -emit-llvm %s -o - | FileCheck %s +// RUN: -target-feature +experimental-zvfofp8min -O0 -emit-llvm %s -o - | FileCheck %s #include @@ -720,3 +720,39 @@ void foo () { vbfloat16m4x2_t bf16m4x2; } + +// CHECK-LABEL: define dso_local void @_Z14test_ofp8_typev +// CHECK-NEXT: entry: +// CHECK-NEXT: [[E4M3MF8:%.*]] = alloca , align 1 +// CHECK-NEXT: [[E4M3MF4:%.*]] = alloca , align 1 +// CHECK-NEXT: [[E4M3MF2:%.*]] = alloca , align 1 +// CHECK-NEXT: [[E4M3M1:%.*]] = alloca , align 1 +// CHECK-NEXT: [[E4M3M2:%.*]] = alloca , align 1 +// CHECK-NEXT: [[E4M3M4:%.*]] = alloca , align 1 +// CHECK-NEXT: [[E4M3M8:%.*]] = alloca , align 1 +// CHECK-NEXT: [[E5M2MF8:%.*]] = alloca , align 1 +// CHECK-NEXT: [[E5M2MF4:%.*]] = alloca , align 1 +// CHECK-NEXT: [[E5M2MF2:%.*]] = alloca , align 1 +// CHECK-NEXT: [[E5M2M1:%.*]] = alloca , align 1 +// CHECK-NEXT: [[E5M2M2:%.*]] = alloca , align 1 +// CHECK-NEXT: [[E5M2M4:%.*]] = alloca , align 1 +// CHECK-NEXT: [[E5M2M8:%.*]] = alloca , align 1 +// CHECK-NEXT: ret void +void test_ofp8_type() { + // ofp8 e4m3 + vfloat8e4m3mf8_t e4m3mf8; + vfloat8e4m3mf4_t e4m3mf4; + vfloat8e4m3mf2_t e4m3mf2; + vfloat8e4m3m1_t e4m3m1; + vfloat8e4m3m2_t e4m3m2; + vfloat8e4m3m4_t e4m3m4; + vfloat8e4m3m8_t e4m3m8; + // ofp8 e5m2 + vfloat8e5m2mf8_t e5m2mf8; + vfloat8e5m2mf4_t e5m2mf4; + vfloat8e5m2mf2_t e5m2mf2; + vfloat8e5m2m1_t e5m2m1; + vfloat8e5m2m2_t e5m2m2; + vfloat8e5m2m4_t e5m2m4; + vfloat8e5m2m8_t e5m2m8; +} diff --git a/clang/test/CodeGenCXX/riscv-mangle-rvv-vectors.cpp b/clang/test/CodeGenCXX/riscv-mangle-rvv-vectors.cpp new file mode 100644 index 000000000000..ddb5237adc23 --- /dev/null +++ b/clang/test/CodeGenCXX/riscv-mangle-rvv-vectors.cpp @@ -0,0 +1,62 @@ +// RUN: %clang_cc1 -triple riscv32-none-linux-gnu %s -emit-llvm -o - \ +// RUN: -target-feature +experimental-zvfofp8min -target-feature +zve64x | FileCheck %s +// RUN: %clang_cc1 -triple riscv64-none-linux-gnu %s -emit-llvm -o - \ +// RUN: -target-feature +experimental-zvfofp8min -target-feature +zve64x | FileCheck %s + +typedef __rvv_float8e4m3mf8_t vfloat8e4m3mf8_t; +typedef __rvv_float8e4m3mf4_t vfloat8e4m3mf4_t; +typedef __rvv_float8e4m3mf2_t vfloat8e4m3mf2_t; +typedef __rvv_float8e4m3m1_t vfloat8e4m3m1_t; +typedef __rvv_float8e4m3m2_t vfloat8e4m3m2_t; +typedef __rvv_float8e4m3m4_t vfloat8e4m3m4_t; +typedef __rvv_float8e4m3m8_t vfloat8e4m3m8_t; + +typedef __rvv_float8e5m2mf8_t vfloat8e5m2mf8_t; +typedef __rvv_float8e5m2mf4_t vfloat8e5m2mf4_t; +typedef __rvv_float8e5m2mf2_t vfloat8e5m2mf2_t; +typedef __rvv_float8e5m2m1_t vfloat8e5m2m1_t; +typedef __rvv_float8e5m2m2_t vfloat8e5m2m2_t; +typedef __rvv_float8e5m2m4_t vfloat8e5m2m4_t; +typedef __rvv_float8e5m2m8_t vfloat8e5m2m8_t; + +// CHECK: _Z8fe4m3mf8u21__rvv_float8e4m3mf8_t +void fe4m3mf8(vfloat8e4m3mf8_t) {} + +// CHECK: _Z8fe4m3mf4u21__rvv_float8e4m3mf4_t +void fe4m3mf4(vfloat8e4m3mf4_t) {} + +// CHECK: _Z8fe4m3mf2u21__rvv_float8e4m3mf2_t +void fe4m3mf2(vfloat8e4m3mf2_t) {} + +// CHECK: _Z7fe4m3m1u20__rvv_float8e4m3m1_t +void fe4m3m1(vfloat8e4m3m1_t) {} + +// CHECK: _Z7fe4m3m2u20__rvv_float8e4m3m2_t +void fe4m3m2(vfloat8e4m3m2_t) {} + +// CHECK: _Z7fe4m3m4u20__rvv_float8e4m3m4_t +void fe4m3m4(vfloat8e4m3m4_t) {} + +// CHECK: _Z7fe4m3m8u20__rvv_float8e4m3m8_t +void fe4m3m8(vfloat8e4m3m8_t) {} + +// CHECK: _Z8fe5m2mf8u21__rvv_float8e5m2mf8_t +void fe5m2mf8(vfloat8e5m2mf8_t) {} + +// CHECK: _Z8fe5m2mf4u21__rvv_float8e5m2mf4_t +void fe5m2mf4(vfloat8e5m2mf4_t) {} + +// CHECK: _Z8fe5m2mf2u21__rvv_float8e5m2mf2_t +void fe5m2mf2(vfloat8e5m2mf2_t) {} + +// CHECK: _Z7fe5m2m1u20__rvv_float8e5m2m1_t +void fe5m2m1(vfloat8e5m2m1_t) {} + +// CHECK: _Z7fe5m2m2u20__rvv_float8e5m2m2_t +void fe5m2m2(vfloat8e5m2m2_t) {} + +// CHECK: _Z7fe5m2m4u20__rvv_float8e5m2m4_t +void fe5m2m4(vfloat8e5m2m4_t) {} + +// CHECK: _Z7fe5m2m8u20__rvv_float8e5m2m8_t +void fe5m2m8(vfloat8e5m2m8_t) {} diff --git a/clang/test/Sema/riscv-types.c b/clang/test/Sema/riscv-types.c index 81d1a011dbd8..26807882d37f 100644 --- a/clang/test/Sema/riscv-types.c +++ b/clang/test/Sema/riscv-types.c @@ -1,5 +1,6 @@ // RUN: %clang_cc1 -triple riscv64 -target-feature +v \ -// RUN: -target-feature +zvfh -target-feature +zvfbfmin -ast-print %s | FileCheck %s +// RUN: -target-feature +zvfh -target-feature +zvfbfmin \ +// RUN: -target-feature +experimental-zvfofp8min -ast-print %s | FileCheck %s void bar(void) { // CHECK: __rvv_int64m1_t x0; @@ -155,6 +156,48 @@ void bar(void) { // CHECK: __rvv_bfloat16mf2_t x50; __rvv_bfloat16mf2_t x50; + // CHECK: __rvv_float8e4m3mf8_t x51; + __rvv_float8e4m3mf8_t x51; + + // CHECK: __rvv_float8e4m3mf4_t x52; + __rvv_float8e4m3mf4_t x52; + + // CHECK: __rvv_float8e4m3mf2_t x53; + __rvv_float8e4m3mf2_t x53; + + // CHECK: __rvv_float8e4m3m1_t x54; + __rvv_float8e4m3m1_t x54; + + // CHECK: __rvv_float8e4m3m2_t x55; + __rvv_float8e4m3m2_t x55; + + // CHECK: __rvv_float8e4m3m4_t x56; + __rvv_float8e4m3m4_t x56; + + // CHECK: __rvv_float8e4m3m8_t x57; + __rvv_float8e4m3m8_t x57; + + // CHECK: __rvv_float8e5m2mf8_t x58; + __rvv_float8e5m2mf8_t x58; + + // CHECK: __rvv_float8e5m2mf4_t x59; + __rvv_float8e5m2mf4_t x59; + + // CHECK: __rvv_float8e5m2mf2_t x60; + __rvv_float8e5m2mf2_t x60; + + // CHECK: __rvv_float8e5m2m1_t x61; + __rvv_float8e5m2m1_t x61; + + // CHECK: __rvv_float8e5m2m2_t x62; + __rvv_float8e5m2m2_t x62; + + // CHECK: __rvv_float8e5m2m4_t x63; + __rvv_float8e5m2m4_t x63; + + // CHECK: __rvv_float8e5m2m8_t x64; + __rvv_float8e5m2m8_t x64; + } typedef __rvv_bool4_t vbool4_t; diff --git a/clang/utils/TableGen/RISCVVEmitter.cpp b/clang/utils/TableGen/RISCVVEmitter.cpp index 970132d85d5b..f5b787ce5a06 100644 --- a/clang/utils/TableGen/RISCVVEmitter.cpp +++ b/clang/utils/TableGen/RISCVVEmitter.cpp @@ -496,6 +496,15 @@ void RVVEmitter::createHeader(raw_ostream &OS) { } } + // TODO: Support tuple types for ofp8? + for (BasicType BT : {BasicType::F8E4M3, BasicType::F8E5M2}) { + for (int Log2LMUL : Log2LMULs) { + auto T = TypeCache.computeType(BT, Log2LMUL, PrototypeDescriptor::Vector); + if (T) + printType(*T); + } + } + OS << "\n#ifdef __cplusplus\n"; OS << "}\n"; OS << "#endif // __cplusplus\n";