diff --git a/src/coreclr/jit/CMakeLists.txt b/src/coreclr/jit/CMakeLists.txt index 288edf637a6dd4..883814eea8a191 100644 --- a/src/coreclr/jit/CMakeLists.txt +++ b/src/coreclr/jit/CMakeLists.txt @@ -69,6 +69,8 @@ function(create_standalone_jit) target_compile_definitions(${TARGETDETAILS_TARGET} PRIVATE FEATURE_SIMD) target_compile_definitions(${TARGETDETAILS_TARGET} PRIVATE FEATURE_HW_INTRINSICS) target_compile_definitions(${TARGETDETAILS_TARGET} PRIVATE FEATURE_MASKED_HW_INTRINSICS) + elseif (TARGETDETAILS_ARCH STREQUAL "riscv64") + target_compile_definitions(${TARGETDETAILS_TARGET} PRIVATE FEATURE_HW_INTRINSICS) endif () endfunction() @@ -76,6 +78,8 @@ if (CLR_CMAKE_TARGET_ARCH_AMD64 OR CLR_CMAKE_TARGET_ARCH_ARM64 OR (CLR_CMAKE_TAR add_compile_definitions($<$>>:FEATURE_SIMD>) add_compile_definitions($<$>>:FEATURE_HW_INTRINSICS>) add_compile_definitions($<$>>:FEATURE_MASKED_HW_INTRINSICS>) +elseif (CLR_CMAKE_TARGET_ARCH_RISCV64) + add_compile_definitions($<$>>:FEATURE_HW_INTRINSICS>) endif () # JIT_BUILD disables certain PAL_TRY debugging features @@ -274,6 +278,8 @@ set( JIT_RISCV64_SOURCES lsrariscv64.cpp targetriscv64.cpp unwindriscv64.cpp + hwintrinsicriscv64.cpp + hwintrinsiccodegenriscv64.cpp ) # We include the headers here for better experience in IDEs. diff --git a/src/coreclr/jit/assertionprop.cpp b/src/coreclr/jit/assertionprop.cpp index eee23f2df7cf19..0010850ac0d795 100644 --- a/src/coreclr/jit/assertionprop.cpp +++ b/src/coreclr/jit/assertionprop.cpp @@ -322,6 +322,8 @@ bool IntegralRange::Contains(int64_t value) const // Note: No advantage in using a precise range for IntegralRange. // Example: IntCns = 42 gives [0..127] with a non -precise range, [42,42] with a precise range. return {SymbolicIntegerValue::Zero, SymbolicIntegerValue::ByteMax}; +#elif defined(TARGET_RISCV64) + // No integral ranges so far #else #error Unsupported platform #endif @@ -3121,12 +3123,13 @@ bool Compiler::optIsProfitableToSubstitute(GenTree* dest, BasicBlock* destBlock, { return false; } - +#ifdef FEATURE_SIMD // For several of the scenarios we may skip the costing logic // since we know that the operand is always containable and therefore // is always cost effective to propagate. return parent->ShouldConstantProp(dest, value->AsVecCon()); +#endif // FEATURE_SIMD } #endif // FEATURE_HW_INTRINSICS } diff --git a/src/coreclr/jit/compiler.h b/src/coreclr/jit/compiler.h index 4571efd26adb0b..6a78f11aa3936c 100644 --- a/src/coreclr/jit/compiler.h +++ b/src/coreclr/jit/compiler.h @@ -3077,9 +3077,7 @@ class Compiler #ifdef FEATURE_SIMD void SetOpLclRelatedToSIMDIntrinsic(GenTree* op); -#endif -#ifdef FEATURE_HW_INTRINSICS GenTreeHWIntrinsic* gtNewSimdHWIntrinsicNode(var_types type, NamedIntrinsic hwIntrinsicID, CorInfoType simdBaseJitType, @@ -3418,6 +3416,7 @@ class Compiler GenTree* op2, CorInfoType simdBaseJitType, unsigned simdSize); +#endif // FEATURE_SIMD GenTreeHWIntrinsic* gtNewScalarHWIntrinsicNode(var_types type, NamedIntrinsic hwIntrinsicID); GenTreeHWIntrinsic* gtNewScalarHWIntrinsicNode(var_types type, GenTree* op1, NamedIntrinsic hwIntrinsicID); @@ -3435,7 +3434,6 @@ class Compiler GenTreeFieldList* gtConvertTableOpToFieldList(GenTree* op, unsigned fieldCount); GenTreeFieldList* gtConvertParamOpToFieldList(GenTree* op, unsigned fieldCount, CORINFO_CLASS_HANDLE clsHnd); #endif -#endif // FEATURE_HW_INTRINSICS GenTree* gtNewMemoryBarrier(BarrierKind barrierKind); @@ -4677,12 +4675,13 @@ class Compiler bool mustExpand); #ifdef FEATURE_HW_INTRINSICS +#ifdef FEATURE_SIMD bool IsValidForShuffle(GenTree* indices, unsigned simdSize, var_types simdBaseType, bool* canBecomeValid, bool isShuffleNative) const; - +#endif GenTree* impHWIntrinsic(NamedIntrinsic intrinsic, CORINFO_CLASS_HANDLE clsHnd, CORINFO_METHOD_HANDLE method, @@ -9639,7 +9638,7 @@ class Compiler // support/nonsupport for an instruction set bool compIsaSupportedDebugOnly(CORINFO_InstructionSet isa) const { -#if defined(TARGET_XARCH) || defined(TARGET_ARM64) +#if defined(TARGET_XARCH) || defined(TARGET_ARM64) || defined(TARGET_RISCV64) return opts.compSupportsISA.HasInstructionSet(isa); #else return false; diff --git a/src/coreclr/jit/emitriscv64.cpp b/src/coreclr/jit/emitriscv64.cpp index 0c5e32c9dbda59..9c822cbadc6ea3 100644 --- a/src/coreclr/jit/emitriscv64.cpp +++ b/src/coreclr/jit/emitriscv64.cpp @@ -992,7 +992,30 @@ void emitter::emitIns_R_R_I_I( void emitter::emitIns_R_R_R_R( instruction ins, emitAttr attr, regNumber reg1, regNumber reg2, regNumber reg3, regNumber reg4) { - NYI_RISCV64("emitIns_R_R_R_R-----unimplemented/unused on RISCV64 yet----"); + NYI_IF(!(INS_fmadd_s <= ins && ins <= INS_fnmadd_s) && !(INS_fmadd_d <= ins && ins <= INS_fnmadd_d), + "illegal ins within emitIns_R_R_R_R!"); + + assert(isFloatReg(reg1)); + assert(isFloatReg(reg2)); + assert(isFloatReg(reg3)); + assert(isFloatReg(reg4)); + + code_t code = emitInsCode(ins); + code |= (reg1 & 0x1f) << 7; + code |= (reg2 & 0x1f) << 15; + code |= (reg3 & 0x1f) << 20; + code |= (reg4 & 0x1f) << 27; + code |= 0b111 << 12; + + instrDesc* id = emitNewInstr(attr); + id->idIns(ins); + id->idReg1(reg1); + id->idReg2(reg2); + id->idReg3(reg3); + id->idReg4(reg4); + id->idAddr()->iiaSetInstrEncode(code); + id->idCodeSize(4); + appendToCurIG(id); } /***************************************************************************** @@ -3512,7 +3535,8 @@ void emitter::emitDispInsName( printf(" "); - switch (GetMajorOpcode(code)) + MajorOpcode opcode = GetMajorOpcode(code); + switch (opcode) { case MajorOpcode::Lui: { @@ -4351,6 +4375,33 @@ void emitter::emitDispInsName( } return; } + case MajorOpcode::MAdd: + case MajorOpcode::MSub: + case MajorOpcode::NmSub: + case MajorOpcode::NmAdd: + { + unsigned int opcode2 = (code >> 25) & 0b11; + if (opcode2 > 1) + return emitDispIllegalInstruction(code); + char width = "sdhq"[opcode2]; + + unsigned int opcode4 = (code >> 12) & 0b111; + if (opcode4 != 0b111) + return emitDispIllegalInstruction(code); + + static const char* names[] = {"fmadd", "fmsub", "fnmsub", "fnmadd"}; + + unsigned idx = (unsigned)opcode & 0b11; + const char* name = names[idx]; + const char* pad = (idx < 2) ? " " : ""; + + const char* fd = RegNames[((code >> 7) & 0x1f) | 0x20]; + const char* fs1 = RegNames[((code >> 15) & 0x1f) | 0x20]; + const char* fs2 = RegNames[((code >> 20) & 0x1f) | 0x20]; + const char* fs3 = RegNames[((code >> 27) & 0x1f) | 0x20]; + printf("%s.%c%s %s, %s, %s, %s\n", name, width, pad, fd, fs1, fs2, fs3); + return; + } case MajorOpcode::StoreFp: { unsigned int opcode2 = (code >> 12) & 0x7; diff --git a/src/coreclr/jit/fgbasic.cpp b/src/coreclr/jit/fgbasic.cpp index f6dda4d6a85764..20ee2de17fec6a 100644 --- a/src/coreclr/jit/fgbasic.cpp +++ b/src/coreclr/jit/fgbasic.cpp @@ -1211,9 +1211,11 @@ void Compiler::fgFindJumpTargets(const BYTE* codeAddr, IL_OFFSET codeSize, Fixed case NI_Vector64_CreateScalar: case NI_Vector64_CreateScalarUnsafe: #endif // TARGET_ARM64 +#if defined(FEATURE_SIMD) case NI_Vector128_Create: case NI_Vector128_CreateScalar: case NI_Vector128_CreateScalarUnsafe: +#endif // FEATURE_SIMD #if defined(TARGET_XARCH) case NI_BMI1_TrailingZeroCount: case NI_BMI1_X64_TrailingZeroCount: @@ -1453,6 +1455,7 @@ void Compiler::fgFindJumpTargets(const BYTE* codeAddr, IL_OFFSET codeSize, Fixed case NI_Vector64_AsUInt64: case NI_Vector64_op_UnaryPlus: #endif // TARGET_XARCH +#if defined(FEATURE_SIMD) case NI_Vector128_As: case NI_Vector128_AsByte: case NI_Vector128_AsDouble: @@ -1468,6 +1471,7 @@ void Compiler::fgFindJumpTargets(const BYTE* codeAddr, IL_OFFSET codeSize, Fixed case NI_Vector128_AsUInt64: case NI_Vector128_AsVector4: case NI_Vector128_op_UnaryPlus: +#endif // FEATURE_SIMD #if defined(TARGET_XARCH) case NI_Vector256_As: case NI_Vector256_AsByte: @@ -1513,7 +1517,7 @@ void Compiler::fgFindJumpTargets(const BYTE* codeAddr, IL_OFFSET codeSize, Fixed break; } -#if defined(FEATURE_HW_INTRINSICS) +#if defined(FEATURE_SIMD) #if defined(TARGET_ARM64) case NI_Vector64_get_AllBitsSet: case NI_Vector64_get_One: @@ -1530,7 +1534,6 @@ void Compiler::fgFindJumpTargets(const BYTE* codeAddr, IL_OFFSET codeSize, Fixed case NI_Vector512_get_One: case NI_Vector512_get_Zero: #endif // TARGET_XARCH -#endif // FEATURE_HW_INTRINSICS { // These always produce a vector constant @@ -1543,6 +1546,7 @@ void Compiler::fgFindJumpTargets(const BYTE* codeAddr, IL_OFFSET codeSize, Fixed break; } +#endif // FEATURE_SIMD case NI_SRCS_UNSAFE_NullRef: case NI_SRCS_UNSAFE_SizeOf: diff --git a/src/coreclr/jit/gentree.cpp b/src/coreclr/jit/gentree.cpp index 6a062b02f2b12d..d4cc9e1d283ff1 100644 --- a/src/coreclr/jit/gentree.cpp +++ b/src/coreclr/jit/gentree.cpp @@ -3386,9 +3386,11 @@ unsigned Compiler::gtHashValue(GenTree* tree) #ifdef FEATURE_HW_INTRINSICS case GT_HWINTRINSIC: hash += tree->AsHWIntrinsic()->GetHWIntrinsicId(); +#ifdef FEATURE_SIMD hash += tree->AsHWIntrinsic()->GetSimdBaseType(); hash += tree->AsHWIntrinsic()->GetSimdSize(); hash += tree->AsHWIntrinsic()->GetAuxiliaryType(); +#endif // FEATURE_SIMD hash += tree->AsHWIntrinsic()->GetRegByIndex(1); break; #endif // FEATURE_HW_INTRINSICS @@ -3851,7 +3853,7 @@ unsigned Compiler::gtSetMultiOpOrder(GenTreeMultiOp* multiOp) bool optsEnabled = opts.OptimizationEnabled(); -#if defined(FEATURE_HW_INTRINSICS) +#if defined(FEATURE_SIMD) if (multiOp->OperIs(GT_HWINTRINSIC) && optsEnabled) { GenTreeHWIntrinsic* hwTree = multiOp->AsHWIntrinsic(); @@ -3913,7 +3915,7 @@ unsigned Compiler::gtSetMultiOpOrder(GenTreeMultiOp* multiOp) break; } } -#endif // defined(FEATURE_SIMD) || defined(FEATURE_HW_INTRINSICS) +#endif // defined(FEATURE_SIMD) // The binary case is special because of GTF_REVERSE_OPS. if (multiOp->GetOperandCount() == 2) @@ -4003,7 +4005,7 @@ unsigned Compiler::gtSetMultiOpOrder(GenTreeMultiOp* multiOp) } return level; } -#endif +#endif // defined(FEATURE_SIMD) || defined(FEATURE_HW_INTRINSICS) //----------------------------------------------------------------------------- // gtWalkOp: Traverse and mark an address expression @@ -18392,7 +18394,7 @@ bool GenTreeIntConCommon::AddrNeedsReloc(Compiler* comp) } #endif // TARGET_X86 -#if defined(FEATURE_HW_INTRINSICS) +#if defined(FEATURE_SIMD) unsigned GenTreeVecCon::ElementCount(unsigned simdSize, var_types simdBaseType) { return simdSize / genTypeSize(simdBaseType); @@ -18835,7 +18837,7 @@ void GenTreeMskCon::EvaluateBinaryInPlace( unreached(); #endif // FEATURE_MASKED_HW_INTRINSICS } -#endif // FEATURE_HW_INTRINSICS*/ +#endif // FEATURE_SIMD*/ //------------------------------------------------------------------------ // IsFieldAddr: Is "this" a static or class field address? @@ -20083,6 +20085,7 @@ FieldSeq::FieldSeq(CORINFO_FIELD_HANDLE fieldHnd, ssize_t offset, FieldKind fiel } } +#ifdef FEATURE_HW_INTRINSICS #ifdef FEATURE_SIMD //------------------------------------------------------------------- // SetOpLclRelatedToSIMDIntrinsic: Determine if the tree has a local var that needs to be set @@ -20098,7 +20101,7 @@ void Compiler::SetOpLclRelatedToSIMDIntrinsic(GenTree* op) setLclRelatedToSIMDIntrinsic(op); } } - +#endif // FEATURE_SIMD void GenTreeMultiOp::ResetOperandArray(size_t newOperandCount, Compiler* compiler, GenTree** inlineOperands, @@ -20238,7 +20241,7 @@ var_types GenTreeJitIntrinsic::GetSimdBaseType() const } return JitType2PreciseVarType(simdBaseJitType); } -#endif // FEATURE_SIMD +#endif // FEATURE_HW_INTRINSICS #ifdef FEATURE_HW_INTRINSICS //------------------------------------------------------------------------ @@ -20643,6 +20646,7 @@ bool GenTree::isEmbeddedMaskingCompatibleHWIntrinsic() const return false; } +#ifdef FEATURE_SIMD GenTreeHWIntrinsic* Compiler::gtNewSimdHWIntrinsicNode(var_types type, NamedIntrinsic hwIntrinsicID, CorInfoType simdBaseJitType, @@ -27899,6 +27903,7 @@ GenTree* Compiler::gtNewSimdWithElementNode( return gtNewSimdHWIntrinsicNode(type, op1, op2, op3, hwIntrinsicID, simdBaseJitType, simdSize); } +#endif // FEATURE_SIMD #ifdef TARGET_ARM64 //------------------------------------------------------------------------ @@ -27968,7 +27973,7 @@ GenTreeFieldList* Compiler::gtConvertParamOpToFieldList(GenTree* op, unsigned fi return fieldList; } #endif // TARGET_ARM64 - +#ifdef FEATURE_SIMD GenTree* Compiler::gtNewSimdWithLowerNode( var_types type, GenTree* op1, GenTree* op2, CorInfoType simdBaseJitType, unsigned simdSize) { @@ -28026,6 +28031,7 @@ GenTree* Compiler::gtNewSimdWithUpperNode( return gtNewSimdHWIntrinsicNode(type, op1, op2, intrinsicId, simdBaseJitType, simdSize); } +#endif // FEATURE_SIMD GenTreeHWIntrinsic* Compiler::gtNewScalarHWIntrinsicNode(var_types type, NamedIntrinsic hwIntrinsicID) { @@ -28035,7 +28041,9 @@ GenTreeHWIntrinsic* Compiler::gtNewScalarHWIntrinsicNode(var_types type, NamedIn GenTreeHWIntrinsic* Compiler::gtNewScalarHWIntrinsicNode(var_types type, GenTree* op1, NamedIntrinsic hwIntrinsicID) { +#ifdef FEATURE_SIMD SetOpLclRelatedToSIMDIntrinsic(op1); +#endif return new (this, GT_HWINTRINSIC) GenTreeHWIntrinsic(type, getAllocator(CMK_ASTNode), hwIntrinsicID, CORINFO_TYPE_UNDEF, 0, op1); @@ -28046,9 +28054,10 @@ GenTreeHWIntrinsic* Compiler::gtNewScalarHWIntrinsicNode(var_types type, GenTree* op2, NamedIntrinsic hwIntrinsicID) { +#ifdef FEATURE_SIMD SetOpLclRelatedToSIMDIntrinsic(op1); SetOpLclRelatedToSIMDIntrinsic(op2); - +#endif return new (this, GT_HWINTRINSIC) GenTreeHWIntrinsic(type, getAllocator(CMK_ASTNode), hwIntrinsicID, CORINFO_TYPE_UNDEF, 0, op1, op2); } @@ -28056,9 +28065,11 @@ GenTreeHWIntrinsic* Compiler::gtNewScalarHWIntrinsicNode(var_types type, GenTreeHWIntrinsic* Compiler::gtNewScalarHWIntrinsicNode( var_types type, GenTree* op1, GenTree* op2, GenTree* op3, NamedIntrinsic hwIntrinsicID) { +#ifdef FEATURE_SIMD SetOpLclRelatedToSIMDIntrinsic(op1); SetOpLclRelatedToSIMDIntrinsic(op2); SetOpLclRelatedToSIMDIntrinsic(op3); +#endif return new (this, GT_HWINTRINSIC) GenTreeHWIntrinsic(type, getAllocator(CMK_ASTNode), hwIntrinsicID, CORINFO_TYPE_UNDEF, 0, op1, op2, op3); @@ -29551,7 +29562,7 @@ genTreeOps GenTreeHWIntrinsic::GetOperForHWIntrinsicId(NamedIntrinsic id, var_ty } } } - +#ifdef FEATURE_SIMD //------------------------------------------------------------------------------ // GetHWIntrinsicIdForUnOp: Returns intrinsic ID based on the oper, base type, and simd size // @@ -29668,6 +29679,7 @@ NamedIntrinsic GenTreeHWIntrinsic::GetHWIntrinsicIdForBinOp(Compiler* comp, unsigned simdSize, bool isScalar) { + var_types simdType = comp->getSIMDTypeForSize(simdSize); assert(varTypeIsArithmetic(simdBaseType)); @@ -30990,6 +31002,7 @@ bool GenTreeHWIntrinsic::ShouldConstantProp(GenTree* operand, GenTreeVecCon* vec return false; } +#endif // FEATURE_SIMD #endif // FEATURE_HW_INTRINSICS //--------------------------------------------------------------------------------------- @@ -32069,8 +32082,8 @@ GenTree* Compiler::gtFoldExprHWIntrinsic(GenTreeHWIntrinsic* tree) CorInfoType simdBaseJitType = tree->GetSimdBaseJitType(); unsigned int simdSize = tree->GetSimdSize(); +#ifdef FEATURE_SIMD simd_t simdVal = {}; - if (GenTreeVecCon::IsHWIntrinsicCreateConstant(tree, simdVal)) { GenTreeVecCon* vecCon = gtNewVconNode(retType); @@ -32085,6 +32098,7 @@ GenTree* Compiler::gtFoldExprHWIntrinsic(GenTreeHWIntrinsic* tree) fgUpdateConstTreeValueNumber(vecCon); return vecCon; } +#endif // FEATURE_SIMD GenTree* op1 = nullptr; GenTree* op2 = nullptr; @@ -32116,7 +32130,6 @@ GenTree* Compiler::gtFoldExprHWIntrinsic(GenTreeHWIntrinsic* tree) return tree; } } - #if defined(FEATURE_MASKED_HW_INTRINSICS) if (tree->OperIsConvertMaskToVector()) { @@ -32226,6 +32239,7 @@ GenTree* Compiler::gtFoldExprHWIntrinsic(GenTreeHWIntrinsic* tree) assert(op2 == nullptr); assert(op3 == nullptr); +#if defined(FEATURE_MASKED_HW_INTRINSICS) if (oper != GT_NONE) { if (varTypeIsMask(retType)) @@ -32238,7 +32252,6 @@ GenTree* Compiler::gtFoldExprHWIntrinsic(GenTreeHWIntrinsic* tree) } resultNode = cnsNode; } -#if defined(FEATURE_MASKED_HW_INTRINSICS) else if (tree->OperIsConvertMaskToVector()) { GenTreeMskCon* mskCon = cnsNode->AsMskCon(); @@ -32295,11 +32308,12 @@ GenTree* Compiler::gtFoldExprHWIntrinsic(GenTreeHWIntrinsic* tree) resultNode = mskCon; } -#endif // FEATURE_MASKED_HW_INTRINSICS else +#endif // FEATURE_MASKED_HW_INTRINSICS { switch (ni) { +#if defined(TARGET_ARM64) || defined(TARGET_XARCH) #ifdef TARGET_ARM64 case NI_ArmBase_LeadingZeroCount: #else @@ -32315,7 +32329,7 @@ GenTree* Compiler::gtFoldExprHWIntrinsic(GenTreeHWIntrinsic* tree) resultNode = cnsNode; break; } - +#endif // defined(TARGET_ARM64) || defined(TARGET_XARCH) #ifdef TARGET_ARM64 case NI_ArmBase_Arm64_LeadingZeroCount: { @@ -32330,7 +32344,7 @@ GenTree* Compiler::gtFoldExprHWIntrinsic(GenTreeHWIntrinsic* tree) resultNode = cnsNode; break; } -#else +#elif defined(TARGET_XARCH) case NI_LZCNT_X64_LeadingZeroCount: { assert(varTypeIsLong(retType)); @@ -32343,7 +32357,7 @@ GenTree* Compiler::gtFoldExprHWIntrinsic(GenTreeHWIntrinsic* tree) break; } #endif - +#ifdef FEATURE_SIMD case NI_Vector128_AsVector3: case NI_Vector128_AsVector128Unsafe: #ifdef TARGET_ARM64 @@ -32481,7 +32495,7 @@ GenTree* Compiler::gtFoldExprHWIntrinsic(GenTreeHWIntrinsic* tree) } break; } - +#endif // FEATURE_SIMD default: { break; @@ -32491,6 +32505,7 @@ GenTree* Compiler::gtFoldExprHWIntrinsic(GenTreeHWIntrinsic* tree) } else if (otherNode->OperIsConst()) { +#ifdef FEATURE_SIMD if (oper != GT_NONE) { assert(op3 == nullptr); @@ -32767,6 +32782,7 @@ GenTree* Compiler::gtFoldExprHWIntrinsic(GenTreeHWIntrinsic* tree) } } } +#endif // FEATURE_SIMD } else if (op3 == nullptr) { @@ -32778,6 +32794,7 @@ GenTree* Compiler::gtFoldExprHWIntrinsic(GenTreeHWIntrinsic* tree) oper = GT_NONE; } +#ifdef FEATURE_SIMD switch (oper) { case GT_ADD: @@ -32856,7 +32873,6 @@ GenTree* Compiler::gtFoldExprHWIntrinsic(GenTreeHWIntrinsic* tree) { break; } - if (cnsNode->AsVecCon()->IsScalarOne(simdBaseType)) { resultNode = otherNode; @@ -33256,8 +33272,10 @@ GenTree* Compiler::gtFoldExprHWIntrinsic(GenTreeHWIntrinsic* tree) break; } } +#endif // FEATURE_SIMD } +#ifdef FEATURE_SIMD if (op3 != nullptr) { switch (ni) @@ -33312,7 +33330,6 @@ GenTree* Compiler::gtFoldExprHWIntrinsic(GenTreeHWIntrinsic* tree) } break; } - default: { break; @@ -33325,6 +33342,7 @@ GenTree* Compiler::gtFoldExprHWIntrinsic(GenTreeHWIntrinsic* tree) resultNode = gtNewSimdCvtVectorToMaskNode(retType, resultNode, simdBaseJitType, simdSize); return gtFoldExprHWIntrinsic(resultNode->AsHWIntrinsic()); } +#endif // FEATURE_SIMD if (resultNode != tree) { diff --git a/src/coreclr/jit/gentree.h b/src/coreclr/jit/gentree.h index e19de5dad0c844..9e9d446882ef14 100644 --- a/src/coreclr/jit/gentree.h +++ b/src/coreclr/jit/gentree.h @@ -6570,7 +6570,7 @@ struct GenTreeHWIntrinsic : public GenTreeJitIntrinsic } static bool Equals(GenTreeHWIntrinsic* op1, GenTreeHWIntrinsic* op2); - +#ifdef FEATURE_SIMD static NamedIntrinsic GetHWIntrinsicIdForUnOp( Compiler* comp, genTreeOps oper, GenTree* op1, var_types simdBaseType, unsigned simdSize, bool isScalar); @@ -6593,7 +6593,7 @@ struct GenTreeHWIntrinsic : public GenTreeJitIntrinsic static var_types GetLookupTypeForCmpOp( Compiler* comp, genTreeOps oper, var_types type, var_types simdBaseType, unsigned simdSize); - +#endif // FEATURE_SIMD static genTreeOps GetOperForHWIntrinsicId(NamedIntrinsic id, var_types simdBaseType, bool* isScalar); genTreeOps GetOperForHWIntrinsicId(bool* isScalar) const @@ -6601,7 +6601,9 @@ struct GenTreeHWIntrinsic : public GenTreeJitIntrinsic return GetOperForHWIntrinsicId(GetHWIntrinsicId(), GetSimdBaseType(), isScalar); } +#ifdef FEATURE_SIMD bool ShouldConstantProp(GenTree* operand, GenTreeVecCon* vecCon); +#endif void NormalizeJitBaseTypeToInt(NamedIntrinsic id, var_types simdBaseType) { @@ -9645,7 +9647,7 @@ inline bool GenTree::IsVectorNaN(var_types simdBaseType) const // inline bool GenTree::IsVectorCreate() const { -#ifdef FEATURE_HW_INTRINSICS +#if defined(FEATURE_SIMD) if (OperIs(GT_HWINTRINSIC)) { switch (AsHWIntrinsic()->GetHWIntrinsicId()) @@ -9663,7 +9665,7 @@ inline bool GenTree::IsVectorCreate() const return false; } } -#endif // FEATURE_HW_INTRINSICS +#endif // FEATURE_SIMD return false; } @@ -9763,7 +9765,7 @@ inline bool GenTree::IsMaskZero() const // inline uint64_t GenTree::GetIntegralVectorConstElement(size_t index, var_types simdBaseType) { -#ifdef FEATURE_HW_INTRINSICS +#ifdef FEATURE_SIMD if (IsCnsVec()) { const GenTreeVecCon* node = AsVecCon(); @@ -9818,7 +9820,7 @@ inline uint64_t GenTree::GetIntegralVectorConstElement(size_t index, var_types s } } } -#endif // FEATURE_HW_INTRINSICS +#endif // FEATURE_SIMD return false; } diff --git a/src/coreclr/jit/hwintrinsic.cpp b/src/coreclr/jit/hwintrinsic.cpp index a00d57962d757b..73a8b6ee043598 100644 --- a/src/coreclr/jit/hwintrinsic.cpp +++ b/src/coreclr/jit/hwintrinsic.cpp @@ -34,6 +34,19 @@ static const HWIntrinsicInfo hwIntrinsicInfoArray[] = { /* category */ category \ }, #include "hwintrinsiclistarm64.h" +#elif defined (TARGET_RISCV64) +#define HARDWARE_INTRINSIC(isa, name, size, numarg, t1, t2, t3, t4, t5, t6, t7, t8, t9, t10, category, flag) \ + { \ + /* name */ #name, \ + /* flags */ static_cast(flag), \ + /* id */ NI_##isa##_##name, \ + /* ins */ t1, t2, t3, t4, t5, t6, t7, t8, t9, t10, \ + /* isa */ InstructionSet_##isa, \ + /* simdSize */ size, \ + /* numArgs */ numarg, \ + /* category */ category \ + }, +#include "hwintrinsiclistriscv64.h" #else #error Unsupported platform #endif @@ -712,7 +725,7 @@ uint8_t TernaryLogicInfo::GetTernaryControlByte(const TernaryLogicInfo& info, ui return oper3Result; } #endif // TARGET_XARCH - +#ifdef FEATURE_SIMD //------------------------------------------------------------------------ // getBaseJitTypeFromArgIfNeeded: Get simdBaseJitType of intrinsic from 1st or 2nd argument depending on the flag // @@ -756,6 +769,7 @@ CorInfoType Compiler::getBaseJitTypeFromArgIfNeeded(NamedIntrinsic intrinsic, return simdBaseJitType; } +#endif // FEATURE_SIMD struct HWIntrinsicIsaRange { @@ -866,6 +880,10 @@ static const HWIntrinsicIsaRange hwintrinsicIsaRangeArray[] = { { NI_Illegal, NI_Illegal }, // Sha1_Arm64 { NI_Illegal, NI_Illegal }, // Sha256_Arm64 { NI_Illegal, NI_Illegal }, // Sve_Arm64 +#elif defined (TARGET_RISCV64) + { FIRST_NI_RiscV64Base, LAST_NI_RiscV64Base }, + { NI_Illegal, NI_Illegal }, + { NI_Illegal, NI_Illegal }, #else #error Unsupported platform #endif @@ -888,6 +906,8 @@ static void ValidateHWIntrinsicInfo(CORINFO_InstructionSet isa, NamedIntrinsic n assert((info.simdSize == 8) || (info.simdSize == 16)); #elif defined(TARGET_XARCH) assert((info.simdSize == 16) || (info.simdSize == 32) || (info.simdSize == 64)); +#elif defined(TARGET_RISCV64) + unreached(); // no SIMD intrinsics yet #else unreached(); #endif @@ -898,6 +918,8 @@ static void ValidateHWIntrinsicInfo(CORINFO_InstructionSet isa, NamedIntrinsic n // We should only have an expected number of arguments #if defined(TARGET_ARM64) || defined(TARGET_XARCH) assert((info.numArgs >= 0) && (info.numArgs <= 5)); +#elif defined(TARGET_RISCV64) + assert((info.numArgs >= 1) && (info.numArgs <= 3)); #else unreached(); #endif @@ -918,6 +940,11 @@ static void ValidateHWIntrinsicIsaRange(CORINFO_InstructionSet isa, const HWIntr assert(isaRange.LastId != NI_Illegal); // Both entries should belong to the expected ISA + if (HWIntrinsicInfo::lookupIsa(isaRange.FirstId) != isa) + { + printf("HWIntrinsicInfo::lookupIsa(isaRange.FirstId):%i == isa:%i\n", + HWIntrinsicInfo::lookupIsa(isaRange.FirstId), isa); + } assert(HWIntrinsicInfo::lookupIsa(isaRange.FirstId) == isa); assert(HWIntrinsicInfo::lookupIsa(isaRange.LastId) == isa); @@ -1064,6 +1091,7 @@ NamedIntrinsic HWIntrinsicInfo::lookupId(Compiler* comp, if (isSupportedProp || isHardwareAcceleratedProp) { +#ifdef FEATURE_SIMD // The `compSupportsHWIntrinsic` above validates `compSupportsIsa` indicating // that the compiler can emit instructions for the ISA but not whether the // hardware supports them. @@ -1095,14 +1123,14 @@ NamedIntrinsic HWIntrinsicInfo::lookupId(Compiler* comp, return NI_IsSupported_Dynamic; } } - +#endif // FEATURE_SIMD return NI_IsSupported_False; } else if (!isIsaSupported) { return NI_Throw_PlatformNotSupportedException; } - +#ifdef FEATURE_SIMD // Special case: For Vector64/128/256 we currently don't accelerate any of the methods when // IsHardwareAccelerated reports false. For Vector64 and Vector128 this is when the baseline // ISA is unsupported. For Vector256 this is when AVX2 is unsupported since integer types @@ -1149,6 +1177,7 @@ NamedIntrinsic HWIntrinsicInfo::lookupId(Compiler* comp, } } #endif +#endif // FEATURE_SIMD size_t isaIndex = static_cast(isa) - 1; assert(isaIndex < ARRAY_SIZE(hwintrinsicIsaRangeArray)); @@ -1202,7 +1231,7 @@ NamedIntrinsic HWIntrinsicInfo::lookupId(Compiler* comp, // Those intrinsics will hit this code path and need to return NI_Illegal return NI_Illegal; } - +#ifdef FEATURE_SIMD //------------------------------------------------------------------------ // lookupSimdSize: Gets the SimdSize for a given HWIntrinsic and signature // @@ -1246,6 +1275,7 @@ unsigned HWIntrinsicInfo::lookupSimdSize(Compiler* comp, NamedIntrinsic id, CORI assert((simdSize > 0) && (simdBaseJitType != CORINFO_TYPE_UNDEF)); return simdSize; } +#endif // FEATURE_SIMD //------------------------------------------------------------------------ // isImmOp: Checks whether the HWIntrinsic node has an imm operand @@ -1273,6 +1303,8 @@ bool HWIntrinsicInfo::isImmOp(NamedIntrinsic id, const GenTree* op) { return false; } +#elif defined(TARGET_RISCV64) + return HWIntrinsicInfo::HasImmediateOperand(id); #else #error Unsupported platform #endif @@ -1298,7 +1330,7 @@ bool HWIntrinsicInfo::isImmOp(NamedIntrinsic id, const GenTree* op) GenTree* Compiler::getArgForHWIntrinsic(var_types argType, CORINFO_CLASS_HANDLE argClass) { GenTree* arg = nullptr; - +#ifdef FEATURE_SIMD if (varTypeIsStruct(argType)) { if (!varTypeIsSIMD(argType)) @@ -1313,6 +1345,7 @@ GenTree* Compiler::getArgForHWIntrinsic(var_types argType, CORINFO_CLASS_HANDLE assert(varTypeIsSIMDOrMask(arg)); } else +#endif // FEATURE_SIMD { assert(varTypeIsArithmetic(argType) || (argType == TYP_BYREF)); @@ -1438,7 +1471,11 @@ bool Compiler::compSupportsHWIntrinsic(CORINFO_InstructionSet isa) // static bool impIsTableDrivenHWIntrinsic(NamedIntrinsic intrinsicId, HWIntrinsicCategory category) { +#ifdef FEATURE_SIMD return (category != HW_Category_Special) && !HWIntrinsicInfo::HasSpecialImport(intrinsicId); +#else + return false; +#endif } //------------------------------------------------------------------------ @@ -1693,7 +1730,7 @@ GenTree* Compiler::impHWIntrinsic(NamedIntrinsic intrinsic, var_types retType = genActualType(JITtype2varType(sig->retType)); CorInfoType simdBaseJitType = CORINFO_TYPE_UNDEF; GenTree* retNode = nullptr; - +#ifdef FEATURE_SIMD if (retType == TYP_STRUCT) { unsigned int sizeBytes; @@ -1764,7 +1801,7 @@ GenTree* Compiler::impHWIntrinsic(NamedIntrinsic intrinsic, simdBaseJitType = getBaseJitTypeFromArgIfNeeded(intrinsic, sig, simdBaseJitType); unsigned simdSize = 0; - +#endif // FEATURE_SIMD if (simdBaseJitType == CORINFO_TYPE_UNDEF) { if ((category == HW_Category_Scalar) || HWIntrinsicInfo::isScalarIsa(isa)) @@ -1778,6 +1815,7 @@ GenTree* Compiler::impHWIntrinsic(NamedIntrinsic intrinsic, } else { +#ifdef FEATURE_SIMD unsigned int sizeBytes; simdBaseJitType = getBaseJitTypeAndSizeOfSIMDType(clsHnd, &sizeBytes); @@ -1802,6 +1840,9 @@ GenTree* Compiler::impHWIntrinsic(NamedIntrinsic intrinsic, { assert((category == HW_Category_Special) || (category == HW_Category_Helper) || (sizeBytes != 0)); } +#else + assert(false); +#endif // FEATURE_SIMD } } #ifdef TARGET_ARM64 @@ -1831,14 +1872,15 @@ GenTree* Compiler::impHWIntrinsic(NamedIntrinsic intrinsic, } #endif // TARGET_ARM64 + var_types simdBaseType = TYP_UNKNOWN; +#ifdef FEATURE_SIMD // Immediately return if the category is other than scalar/special and this is not a supported base type. if ((category != HW_Category_Special) && (category != HW_Category_Scalar) && !HWIntrinsicInfo::isScalarIsa(isa) && !isSupportedBaseType(intrinsic, simdBaseJitType)) { return nullptr; } - - var_types simdBaseType = TYP_UNKNOWN; +#endif // FEATURE_SIMD if (simdBaseJitType != CORINFO_TYPE_UNDEF) { @@ -1851,10 +1893,11 @@ GenTree* Compiler::impHWIntrinsic(NamedIntrinsic intrinsic, } #endif // TARGET_XARCH } - +#ifdef FEATURE_SIMD // We may have already determined simdSize for intrinsics that require special handling. // If so, skip the lookup. simdSize = (simdSize == 0) ? HWIntrinsicInfo::lookupSimdSize(this, intrinsic, sig) : simdSize; +#endif // FEATURE_SIMD HWIntrinsicSignatureReader sigReader; sigReader.Read(info.compCompHnd, sig); @@ -1926,7 +1969,7 @@ GenTree* Compiler::impHWIntrinsic(NamedIntrinsic intrinsic, getHWIntrinsicImmTypes(intrinsic, sig, 1, simdBaseType, simdBaseJitType, sigReader.op1ClsHnd, sigReader.op2ClsHnd, sigReader.op3ClsHnd, &immSimdSize, &immSimdBaseType); HWIntrinsicInfo::lookupImmBounds(intrinsic, immSimdSize, immSimdBaseType, 1, &immLowerBound, &immUpperBound); -#else +#elif defined(TARGET_XARCH) immUpperBound = HWIntrinsicInfo::lookupImmUpperBound(intrinsic); hasFullRangeImm = HWIntrinsicInfo::HasFullRangeImm(intrinsic); #endif @@ -2045,15 +2088,17 @@ GenTree* Compiler::impHWIntrinsic(NamedIntrinsic intrinsic, switch (numArgs) { +#ifdef FEATURE_SIMD case 0: { assert(!isScalar); retNode = gtNewSimdHWIntrinsicNode(nodeRetType, intrinsic, simdBaseJitType, simdSize); break; } - +#endif // FEATURE_SIMD case 1: { +#if !defined(TARGET_RISCV64) if ((category == HW_Category_MemoryLoad) && op1->OperIs(GT_CAST)) { // Although the API specifies a pointer, if what we have is a BYREF, that's what @@ -2063,9 +2108,19 @@ GenTree* Compiler::impHWIntrinsic(NamedIntrinsic intrinsic, op1 = op1->gtGetOp1(); } } - - retNode = isScalar ? gtNewScalarHWIntrinsicNode(nodeRetType, op1, intrinsic) - : gtNewSimdHWIntrinsicNode(nodeRetType, op1, intrinsic, simdBaseJitType, simdSize); +#endif // !TARGET_RISCV64 + if (isScalar) + { + retNode = gtNewScalarHWIntrinsicNode(nodeRetType, op1, intrinsic); + } + else + { +#ifdef FEATURE_SIMD + retNode = gtNewSimdHWIntrinsicNode(nodeRetType, op1, intrinsic, simdBaseJitType, simdSize); +#else + assert(false); +#endif + } #if defined(TARGET_XARCH) switch (intrinsic) @@ -2126,9 +2181,18 @@ GenTree* Compiler::impHWIntrinsic(NamedIntrinsic intrinsic, case 2: { - retNode = isScalar - ? gtNewScalarHWIntrinsicNode(nodeRetType, op1, op2, intrinsic) - : gtNewSimdHWIntrinsicNode(nodeRetType, op1, op2, intrinsic, simdBaseJitType, simdSize); + if (isScalar) + { + retNode = gtNewScalarHWIntrinsicNode(nodeRetType, op1, op2, intrinsic); + } + else + { +#ifdef FEATURE_SIMD + retNode = gtNewSimdHWIntrinsicNode(nodeRetType, op1, op2, intrinsic, simdBaseJitType, simdSize); +#else + assert(false); +#endif + } #ifdef TARGET_XARCH if ((intrinsic == NI_SSE42_Crc32) || (intrinsic == NI_SSE42_X64_Crc32)) @@ -2221,9 +2285,19 @@ GenTree* Compiler::impHWIntrinsic(NamedIntrinsic intrinsic, op3 = addRangeCheckIfNeeded(intrinsic, op3, immLowerBound, immUpperBound); } - retNode = isScalar ? gtNewScalarHWIntrinsicNode(nodeRetType, op1, op2, op3, intrinsic) - : gtNewSimdHWIntrinsicNode(nodeRetType, op1, op2, op3, intrinsic, simdBaseJitType, - simdSize); + if (isScalar) + { + retNode = gtNewScalarHWIntrinsicNode(nodeRetType, op1, op2, op3, intrinsic); + } + else + { +#ifdef FEATURE_SIMD + retNode = + gtNewSimdHWIntrinsicNode(nodeRetType, op1, op2, op3, intrinsic, simdBaseJitType, simdSize); +#else + assert(false); +#endif + } switch (intrinsic) { @@ -2274,7 +2348,7 @@ GenTree* Compiler::impHWIntrinsic(NamedIntrinsic intrinsic, break; } - +#ifdef FEATURE_SIMD case 4: { assert(!isScalar); @@ -2299,15 +2373,17 @@ GenTree* Compiler::impHWIntrinsic(NamedIntrinsic intrinsic, } break; } - +#endif // FEATURE_SIMD default: break; } } else { +#ifdef FEATURE_SIMD retNode = impSpecialIntrinsic(intrinsic, clsHnd, method, sig R2RARG(entryPoint), simdBaseJitType, nodeRetType, simdSize, mustExpand); +#endif // FEATURE_SIMD } if (setMethodHandle && (retNode != nullptr)) diff --git a/src/coreclr/jit/hwintrinsic.h b/src/coreclr/jit/hwintrinsic.h index d936d579d8e25a..17d055786597ae 100644 --- a/src/coreclr/jit/hwintrinsic.h +++ b/src/coreclr/jit/hwintrinsic.h @@ -67,6 +67,32 @@ enum HWIntrinsicCategory : uint8_t // - have to be addressed specially HW_Category_Special }; +#elif defined(TARGET_RISCV64) +enum HWIntrinsicCategory : uint8_t +{ + // Most of the Arm64 intrinsic fall into SIMD category: + // - vector or scalar intrinsics that operate on one-or-many SIMD registers + // HW_Category_SIMD, + + // Scalar intrinsics operate on general purpose registers (e.g. cls, clz, rbit) + HW_Category_Scalar, + + // // Memory access intrinsics + // HW_Category_MemoryLoad, + // HW_Category_MemoryStore, + + // // These are Arm64 that share some features in a given category (e.g. immediate operand value range) + // HW_Category_ShiftLeftByImmediate, + // HW_Category_ShiftRightByImmediate, + + // // Helper intrinsics + // // - do not directly correspond to a instruction, such as Vector64.AllBitsSet + // HW_Category_Helper, + + // // Special intrinsics + // // - have to be addressed specially + // HW_Category_Special +}; #else #error Unsupported platform #endif @@ -233,6 +259,67 @@ enum HWIntrinsicFlag : unsigned int // The intrinsic is a reduce operation. HW_Flag_ReduceOperation = 0x2000000, +#elif defined(TARGET_RISCV64) + // // The intrinsic has an immediate operand + // // - the value can be (and should be) encoded in a corresponding instruction when the operand value is constant + // HW_Flag_HasImmediateOperand = 0x400, + + // // The intrinsic has read/modify/write semantics in multiple-operands form. + // HW_Flag_HasRMWSemantics = 0x800, + + // // The intrinsic operates on the lower part of a SIMD register + // // - the upper part of the source registers are ignored + // // - the upper part of the destination register is zeroed + // HW_Flag_SIMDScalar = 0x1000, + + // // The intrinsic supports some sort of containment analysis + // HW_Flag_SupportsContainment = 0x2000, + + // // The intrinsic needs consecutive registers + // HW_Flag_NeedsConsecutiveRegisters = 0x4000, + + // // The intrinsic uses scalable registers + // HW_Flag_Scalable = 0x8000, + + // // Returns Per-Element Mask + // // the intrinsic returns a vector containing elements that are either "all bits set" or "all bits clear" + // // this output can be used as a per-element mask + // HW_Flag_ReturnsPerElementMask = 0x10000, + + // // The intrinsic uses a mask in arg1 to select elements present in the result + // HW_Flag_ExplicitMaskedOperation = 0x20000, + + // // The intrinsic uses a mask in arg1 (either explicitly, embedded or optionally embedded) to select elements + // present + // // in the result, and must use a low register. + // HW_Flag_LowMaskedOperation = 0x40000, + + // // The intrinsic can optionally use a mask in arg1 to select elements present in the result, which is not present + // in + // // the API call + // HW_Flag_OptionalEmbeddedMaskedOperation = 0x80000, + + // // The intrinsic uses a mask in arg1 to select elements present in the result, which is not present in the API + // call HW_Flag_EmbeddedMaskedOperation = 0x100000, + + // // The intrinsic comes in both vector and scalar variants. During the import stage if the basetype is scalar, + // // then the intrinsic should be switched to a scalar only version. + // HW_Flag_HasScalarInputVariant = 0x200000, + + // // The intrinsic uses a mask in arg1 to select elements present in the result, and must use a low vector + // register. HW_Flag_LowVectorOperation = 0x400000, + + // // The intrinsic uses a mask in arg1 to select elements present in the result, which zeros inactive elements + // // (instead of merging). + // HW_Flag_ZeroingMaskedOperation = 0x800000, + + // // The intrinsic has an overload where the base type is extracted from a ValueTuple of SIMD types + // // (HW_Flag_BaseTypeFrom{First, Second}Arg must also be set to denote the position of the ValueTuple) + // HW_Flag_BaseTypeFromValueTupleArg = 0x1000000, + + // // The intrinsic is a reduce operation. + // HW_Flag_ReduceOperation = 0x2000000, + #else #error Unsupported platform #endif @@ -529,14 +616,17 @@ struct HWIntrinsicInfo static CORINFO_InstructionSet lookupIsa(const char* className, const char* innerEnclosingClassName, const char* outerEnclosingClassName); - +#ifdef FEATURE_SIMD static unsigned lookupSimdSize(Compiler* comp, NamedIntrinsic id, CORINFO_SIG_INFO* sig); +#endif #if defined(TARGET_XARCH) static int lookupImmUpperBound(NamedIntrinsic intrinsic); #elif defined(TARGET_ARM64) static void lookupImmBounds( NamedIntrinsic intrinsic, int simdSize, var_types baseType, int immNumber, int* lowerBound, int* upperBound); +#elif defined(TARGET_RISCV64) + // No HW intrinsics with immediates for now #else #error Unsupported platform #endif @@ -674,7 +764,7 @@ struct HWIntrinsicInfo HWIntrinsicFlag flags = lookupFlags(id); #if defined(TARGET_XARCH) return (flags & HW_Flag_MaybeCommutative) != 0; -#elif defined(TARGET_ARM64) +#elif defined(TARGET_ARM64) || defined(TARGET_RISCV64) return false; #else #error Unsupported platform @@ -694,6 +784,8 @@ struct HWIntrinsicInfo return (flags & HW_Flag_NoContainment) == 0; #elif defined(TARGET_ARM64) return (flags & HW_Flag_SupportsContainment) != 0; +#elif defined(TARGET_RISCV64) + return false; #else #error Unsupported platform #endif @@ -704,6 +796,8 @@ struct HWIntrinsicInfo HWIntrinsicFlag flags = lookupFlags(id); #if defined(TARGET_XARCH) || defined(TARGET_ARM64) return (flags & HW_Flag_ReturnsPerElementMask) != 0; +#elif defined(TARGET_RISCV64) + return false; #else #error Unsupported platform #endif @@ -800,6 +894,8 @@ struct HWIntrinsicInfo return (flags & HW_Flag_NoRMWSemantics) == 0; #elif defined(TARGET_ARM64) return (flags & HW_Flag_HasRMWSemantics) != 0; +#elif defined(TARGET_RISCV64) + return false; #else #error Unsupported platform #endif @@ -927,6 +1023,7 @@ struct HWIntrinsicInfo { switch (id) { +#ifdef FEATURE_SIMD #if defined(TARGET_ARM64) case NI_Vector64_Create: #endif // TARGET_ARM64 @@ -936,6 +1033,7 @@ struct HWIntrinsicInfo case NI_Vector512_Create: #endif // TARGET_XARCH return true; +#endif // FEATURE_SIMD default: return false; } @@ -945,6 +1043,7 @@ struct HWIntrinsicInfo { switch (id) { +#ifdef FEATURE_SIMD #if defined(TARGET_ARM64) case NI_Vector64_CreateScalar: #endif // TARGET_ARM64 @@ -954,6 +1053,7 @@ struct HWIntrinsicInfo case NI_Vector512_CreateScalar: #endif // TARGET_XARCH return true; +#endif // FEATURE_SIMD default: return false; } @@ -963,6 +1063,7 @@ struct HWIntrinsicInfo { switch (id) { +#ifdef FEATURE_SIMD #if defined(TARGET_ARM64) case NI_Vector64_CreateScalarUnsafe: #endif // TARGET_ARM64 @@ -972,6 +1073,7 @@ struct HWIntrinsicInfo case NI_Vector512_CreateScalarUnsafe: #endif // TARGET_XARCH return true; +#endif // FEATURE_SIMD default: return false; } @@ -981,6 +1083,7 @@ struct HWIntrinsicInfo { switch (id) { +#ifdef FEATURE_SIMD #if defined(TARGET_ARM64) case NI_Vector64_GetElement: #endif // TARGET_ARM64 @@ -990,6 +1093,7 @@ struct HWIntrinsicInfo case NI_Vector512_GetElement: #endif // TARGET_XARCH return true; +#endif // FEATURE_SIMD default: return false; } @@ -999,6 +1103,7 @@ struct HWIntrinsicInfo { switch (id) { +#ifdef FEATURE_SIMD #if defined(TARGET_ARM64) case NI_Vector64_ToScalar: #endif // TARGET_ARM64 @@ -1008,6 +1113,7 @@ struct HWIntrinsicInfo case NI_Vector512_ToScalar: #endif // TARGET_XARCH return true; +#endif // FEATURE_SIMD default: return false; } diff --git a/src/coreclr/jit/hwintrinsiccodegenriscv64.cpp b/src/coreclr/jit/hwintrinsiccodegenriscv64.cpp new file mode 100644 index 00000000000000..9903b726242e29 --- /dev/null +++ b/src/coreclr/jit/hwintrinsiccodegenriscv64.cpp @@ -0,0 +1,65 @@ +// Licensed to the .NET Foundation under one or more agreements. +// The .NET Foundation licenses this file to you under the MIT license. + +#include "jitpch.h" +#ifdef _MSC_VER +#pragma hdrstop +#endif + +#ifdef FEATURE_HW_INTRINSICS + +#include "codegen.h" + +//------------------------------------------------------------------------ +// genHWIntrinsic: Generates the code for a given hardware intrinsic node. +// +// Arguments: +// node - The hardware intrinsic node +// +void CodeGen::genHWIntrinsic(GenTreeHWIntrinsic* node) +{ + NamedIntrinsic id = node->GetHWIntrinsicId(); + + HWIntrinsicCategory category = HWIntrinsicInfo::lookupCategory(id); + assert(category == HW_Category_Scalar); + + // We need to validate that other phases of the compiler haven't introduced unsupported intrinsics + assert(compiler->compIsaSupportedDebugOnly(HWIntrinsicInfo::lookupIsa(id))); + + size_t opCount = node->GetOperandCount(); + assert(opCount <= 3); + + regNumber destReg = node->GetRegNum(); + regNumber op1Reg = (opCount >= 1) ? node->Op(1)->GetRegNum() : REG_NA; + regNumber op2Reg = (opCount >= 2) ? node->Op(2)->GetRegNum() : REG_NA; + regNumber op3Reg = (opCount >= 3) ? node->Op(3)->GetRegNum() : REG_NA; + + var_types type = genActualType(node); + emitAttr emitSize = emitActualTypeSize(type); + + assert(!node->isRMWHWIntrinsic(compiler)); + assert(!HWIntrinsicInfo::HasImmediateOperand(id)); + + genConsumeMultiOpOperands(node); + + instruction ins = HWIntrinsicInfo::lookupIns(id, type); + assert(ins != INS_invalid); + + switch (id) + { + case NI_RiscV64Base_FusedMultiplyAddScalar: + case NI_RiscV64Base_FusedMultiplySubtractScalar: + case NI_RiscV64Base_FusedNegatedMultiplyAddScalar: + case NI_RiscV64Base_FusedNegatedMultiplySubtractScalar: + assert(opCount == 3); + GetEmitter()->emitIns_R_R_R_R(ins, emitSize, destReg, op1Reg, op2Reg, op3Reg); + break; + + default: + unreached(); + } + + genProduceReg(node); +} + +#endif // FEATURE_HW_INTRINSICS diff --git a/src/coreclr/jit/hwintrinsiclistriscv64.h b/src/coreclr/jit/hwintrinsiclistriscv64.h new file mode 100644 index 00000000000000..bd2efad038ec62 --- /dev/null +++ b/src/coreclr/jit/hwintrinsiclistriscv64.h @@ -0,0 +1,34 @@ +// Licensed to the .NET Foundation under one or more agreements. +// The .NET Foundation licenses this file to you under the MIT license. + +/*****************************************************************************/ +#ifndef HARDWARE_INTRINSIC +#error Define HARDWARE_INTRINSIC before including this file +#endif +/*****************************************************************************/ + +// clang-format off + +#ifdef FEATURE_HW_INTRINSICS +// ISA Function name SIMD size NumArg Instructions Category Flags +// {TYP_BYTE, TYP_UBYTE, TYP_SHORT, TYP_USHORT, TYP_INT, TYP_UINT, TYP_LONG, TYP_ULONG, TYP_FLOAT, TYP_DOUBLE} +// *************************************************************************************************************************************************************************************************************************************************************************************************************************************************************************************************************************************** +#define FIRST_NI_RiscV64Base NI_RiscV64Base_AbsScalar +// Base Intrinsics +HARDWARE_INTRINSIC(RiscV64Base, AbsScalar, 0, 1, {INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_fsgnjx_s, INS_fsgnjx_d}, HW_Category_Scalar, HW_Flag_NoFlag) +HARDWARE_INTRINSIC(RiscV64Base, FusedMultiplyAddScalar, 0, 3, {INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_fmadd_s, INS_fmadd_d}, HW_Category_Scalar, HW_Flag_NoFlag) +HARDWARE_INTRINSIC(RiscV64Base, FusedMultiplySubtractScalar, 0, 3, {INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_fmsub_s, INS_fmsub_d}, HW_Category_Scalar, HW_Flag_NoFlag) +HARDWARE_INTRINSIC(RiscV64Base, FusedNegatedMultiplyAddScalar, 0, 3, {INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_fnmadd_s, INS_fnmadd_d}, HW_Category_Scalar, HW_Flag_NoFlag) +HARDWARE_INTRINSIC(RiscV64Base, FusedNegatedMultiplySubtractScalar, 0, 3, {INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_fnmsub_s, INS_fnmsub_d}, HW_Category_Scalar, HW_Flag_NoFlag) +// HARDWARE_INTRINSIC(RiscV64Base, MultiplyHigh, 0, 2, {INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_smulh, INS_umulh, INS_invalid, INS_invalid}, HW_Category_Scalar, HW_Flag_NoFloatingPointUsed) +#define LAST_NI_RiscV64Base NI_RiscV64Base_FusedNegatedMultiplySubtractScalar + +// #define FIRST_NI_Zbb NI_Zbb_LeadingZeroCount +// HARDWARE_INTRINSIC(RiscV64Base, LeadingZeroCount, 0, 1, {INS_invalid, INS_invalid, INS_invalid, INS_invalid, INS_clzw, INS_clzw, INS_clz, INS_clz, INS_invalid, INS_invalid}, HW_Category_Scalar, HW_Flag_BaseTypeFromFirstArg|HW_Flag_NoFloatingPointUsed) +// #define LAST_NI_Zbb NI_Zbb_LeadingZeroCount + +#endif // FEATURE_HW_INTRINSIC + +#undef HARDWARE_INTRINSIC + +// clang-format on diff --git a/src/coreclr/jit/hwintrinsicriscv64.cpp b/src/coreclr/jit/hwintrinsicriscv64.cpp new file mode 100644 index 00000000000000..461dc1528d69ad --- /dev/null +++ b/src/coreclr/jit/hwintrinsicriscv64.cpp @@ -0,0 +1,201 @@ +// Licensed to the .NET Foundation under one or more agreements. +// The .NET Foundation licenses this file to you under the MIT license. + +#include "jitpch.h" +#include "hwintrinsic.h" + +#ifdef FEATURE_HW_INTRINSICS + +//------------------------------------------------------------------------ +// lookupInstructionSet: Gets the InstructionSet for a given class name +// +// Arguments: +// className -- The name of the class associated with the InstructionSet to lookup +// +// Return Value: +// The InstructionSet associated with className +static CORINFO_InstructionSet lookupInstructionSet(const char* className) +{ + assert(className != nullptr); + + if (className[0] == 'R') + { + if (strcmp(className, "RiscV64Base") == 0) + { + return InstructionSet_RiscV64Base; + } + } + else if (className[0] == 'Z') + { + if (strcmp(className, "Zba") == 0) + { + return InstructionSet_Zba; + } + if (strcmp(className, "Zbb") == 0) + { + return InstructionSet_Zbb; + } + } + + return InstructionSet_NONE; +} + +//------------------------------------------------------------------------ +// lookupIsa: Gets the InstructionSet for a given class name and enclsoing class name +// +// Arguments: +// className -- The name of the class associated with the InstructionSet to lookup +// innerEnclosingClassName -- The name of the inner enclosing class or nullptr if one doesn't exist +// outerEnclosingClassName -- The name of the outer enclosing class or nullptr if one doesn't exist +// +// Return Value: +// The InstructionSet associated with className and enclosingClassName +// +CORINFO_InstructionSet HWIntrinsicInfo::lookupIsa(const char* className, + const char* innerEnclosingClassName, + const char* outerEnclosingClassName) +{ + assert(className != nullptr); + + if (innerEnclosingClassName == nullptr) + { + // No nested class is the most common, so fast path it + return lookupInstructionSet(className); + } + + // Since lookupId is only called for the xplat intrinsics + // or intrinsics in the platform specific namespace, we assume + // that it will be one we can handle and don't try to early out. + return lookupIsa(innerEnclosingClassName, outerEnclosingClassName, nullptr); +} + +//------------------------------------------------------------------------ +// isFullyImplementedIsa: Gets a value that indicates whether the InstructionSet is fully implemented +// +// Arguments: +// isa - The InstructionSet to check +// +// Return Value: +// true if isa is supported; otherwise, false +bool HWIntrinsicInfo::isFullyImplementedIsa(CORINFO_InstructionSet isa) +{ + switch (isa) + { + case InstructionSet_RiscV64Base: + return true; + + default: + return false; + } +} + +//------------------------------------------------------------------------ +// isScalarIsa: Gets a value that indicates whether the InstructionSet is scalar +// +// Arguments: +// isa - The InstructionSet to check +// +// Return Value: +// true if isa is scalar; otherwise, false +bool HWIntrinsicInfo::isScalarIsa(CORINFO_InstructionSet isa) +{ + switch (isa) + { + case InstructionSet_RiscV64Base: + case InstructionSet_Zba: + case InstructionSet_Zbb: + return true; + + default: + return false; + } +} + +//------------------------------------------------------------------------ +// getHWIntrinsicImmOps: Gets the immediate Ops for an intrinsic +// +// Arguments: +// intrinsic -- NamedIntrinsic associated with the HWIntrinsic to lookup +// sig -- signature of the intrinsic call. +// immOp1Ptr [OUT] -- The first immediate Op +// immOp2Ptr [OUT] -- The second immediate Op, if any. Otherwise unchanged. +// +void Compiler::getHWIntrinsicImmOps(NamedIntrinsic intrinsic, + CORINFO_SIG_INFO* sig, + GenTree** immOp1Ptr, + GenTree** immOp2Ptr) +{ + assert(!HWIntrinsicInfo::HasImmediateOperand(intrinsic)); + return; +} + +//------------------------------------------------------------------------ +// impNonConstFallback: generate alternate code when the imm-arg is not a compile-time constant +// +// Arguments: +// intrinsic -- intrinsic ID +// simdType -- Vector type +// simdBaseJitType -- base JIT type of the Vector64/128 +// +// Return Value: +// return the IR of semantic alternative on non-const imm-arg +// +GenTree* Compiler::impNonConstFallback(NamedIntrinsic intrinsic, var_types simdType, CorInfoType simdBaseJitType) +{ + return nullptr; +} + +//------------------------------------------------------------------------ +// impSpecialIntrinsic: Import a hardware intrinsic that requires special handling as a GT_HWINTRINSIC node if possible +// +// Arguments: +// intrinsic -- id of the intrinsic function. +// clsHnd -- class handle containing the intrinsic function. +// method -- method handle of the intrinsic function. +// sig -- signature of the intrinsic call. +// entryPoint -- The entry point information required for R2R scenarios +// simdBaseJitType -- generic argument of the intrinsic. +// retType -- return type of the intrinsic. +// mustExpand -- true if the intrinsic must return a GenTree*; otherwise, false +// +// Return Value: +// The GT_HWINTRINSIC node, or nullptr if not a supported intrinsic +// +GenTree* Compiler::impSpecialIntrinsic(NamedIntrinsic intrinsic, + CORINFO_CLASS_HANDLE clsHnd, + CORINFO_METHOD_HANDLE method, + CORINFO_SIG_INFO* sig R2RARG(CORINFO_CONST_LOOKUP* entryPoint), + CorInfoType simdBaseJitType, + var_types retType, + unsigned simdSize, + bool mustExpand) +{ + const HWIntrinsicCategory category = HWIntrinsicInfo::lookupCategory(intrinsic); + const int numArgs = sig->numArgs; + + bool isScalar = (category == HW_Category_Scalar); + assert(!HWIntrinsicInfo::isScalarIsa(HWIntrinsicInfo::lookupIsa(intrinsic))); + + assert(numArgs >= 0); + + var_types simdBaseType = JitType2PreciseVarType(simdBaseJitType); + assert(varTypeIsArithmetic(simdBaseType)); + + GenTree* retNode = nullptr; + GenTree* op1 = nullptr; + GenTree* op2 = nullptr; + GenTree* op3 = nullptr; + GenTree* op4 = nullptr; + +#ifdef DEBUG + bool isValidScalarIntrinsic = false; +#endif + + switch (intrinsic) + { + default: + return nullptr; + } +} + +#endif // FEATURE_HW_INTRINSICS diff --git a/src/coreclr/jit/importercalls.cpp b/src/coreclr/jit/importercalls.cpp index cbc2c89f136ad3..6f786fa35468df 100644 --- a/src/coreclr/jit/importercalls.cpp +++ b/src/coreclr/jit/importercalls.cpp @@ -3241,6 +3241,8 @@ GenTree* Compiler::impIntrinsic(CORINFO_CLASS_HANDLE clsHnd, assert((LAST_NI_Vector128 + 1) == FIRST_NI_AdvSimd); if (ni < LAST_NI_Vector128) +#elif defined(TARGET_RISCV64) + // No vectors yet #else #error Unsupported platform #endif @@ -4307,6 +4309,17 @@ GenTree* Compiler::impIntrinsic(CORINFO_CLASS_HANDLE clsHnd, retNode = gtNewSimdToScalarNode(callType, retNode, callJitType, 8); break; } +#elif defined(TARGET_RISCV64) + if (compOpportunisticallyDependsOn(InstructionSet_RiscV64Base)) + { + assert(varTypeIsFloating(callType)); + GenTree* op3 = impImplicitR4orR8Cast(impPopStack().val, callType); + GenTree* op2 = impImplicitR4orR8Cast(impPopStack().val, callType); + GenTree* op1 = impImplicitR4orR8Cast(impPopStack().val, callType); + + retNode = + gtNewScalarHWIntrinsicNode(callType, op1, op2, op3, NI_RiscV64Base_FusedMultiplyAddScalar); + } #endif // TODO-CQ-XArch: Ideally we would create a GT_INTRINSIC node for fma, however, that currently @@ -7920,6 +7933,7 @@ bool Compiler::IsTargetIntrinsic(NamedIntrinsic intrinsicName) return false; } + case NI_System_Math_FusedMultiplyAdd: case NI_System_Math_MultiplyAddEstimate: case NI_System_Math_ReciprocalEstimate: return true; @@ -9376,7 +9390,16 @@ GenTree* Compiler::impEstimateIntrinsic(CORINFO_METHOD_HANDLE method, swapOp1AndOp3 = true; } -#endif // TARGET_ARM64 +#elif defined(TARGET_RISCV64) + if (compOpportunisticallyDependsOn(InstructionSet_RiscV64Base)) + { + assert(varTypeIsFloating(callType)); + GenTree* op3 = impImplicitR4orR8Cast(impPopStack().val, callType); + GenTree* op2 = impImplicitR4orR8Cast(impPopStack().val, callType); + GenTree* op1 = impImplicitR4orR8Cast(impPopStack().val, callType); + return gtNewScalarHWIntrinsicNode(callType, op1, op2, op3, NI_RiscV64Base_FusedMultiplyAddScalar); + } +#endif // TARGET_RISCV64 break; } @@ -9492,7 +9515,7 @@ GenTree* Compiler::impEstimateIntrinsic(CORINFO_METHOD_HANDLE method, } } -#if defined(FEATURE_HW_INTRINSICS) +#if defined(FEATURE_SIMD) if (intrinsicId != NI_Illegal) { unsigned simdSize; @@ -9542,7 +9565,7 @@ GenTree* Compiler::impEstimateIntrinsic(CORINFO_METHOD_HANDLE method, } assert(!swapOp1AndOp3); -#endif // FEATURE_HW_INTRINSICS +#endif // FEATURE_SIMD callType = genActualType(callType); @@ -10674,7 +10697,7 @@ NamedIntrinsic Compiler::lookupNamedIntrinsic(CORINFO_METHOD_HANDLE method) } else { -#ifdef FEATURE_HW_INTRINSICS +#ifdef FEATURE_SIMD bool isVectorT = strcmp(className, "Vector`1") == 0; if (isVectorT || (strcmp(className, "Vector") == 0)) @@ -10797,12 +10820,12 @@ NamedIntrinsic Compiler::lookupNamedIntrinsic(CORINFO_METHOD_HANDLE method) enclosingClassNames[0], enclosingClassNames[1]); } } -#endif // FEATURE_HW_INTRINSICS +#endif // FEATURE_SIMD if (result == NI_Illegal) { // This allows the relevant code paths to be dropped as dead code even - // on platforms where FEATURE_HW_INTRINSICS is not supported. + // on platforms where FEATURE_SIMD is not supported. if (strcmp(methodName, "get_IsSupported") == 0) { @@ -11018,10 +11041,12 @@ NamedIntrinsic Compiler::lookupNamedIntrinsic(CORINFO_METHOD_HANDLE method) platformNamespaceName = ".X86"; #elif defined(TARGET_ARM64) platformNamespaceName = ".Arm"; +#elif defined(TARGET_RISCV64) + platformNamespaceName = ".RiscV"; #else #error Unsupported platform #endif - +#ifdef FEATURE_SIMD if (strncmp(methodName, "System.Runtime.Intrinsics.ISimdVector GenTree* { -#ifdef FEATURE_HW_INTRINSICS +#ifdef FEATURE_SIMD if (varTypeIsSIMD(type)) { return gtNewSimdBinOpNode(oper, type, op1, op2, CORINFO_TYPE_NATIVEUINT, genTypeSize(type)); diff --git a/src/coreclr/jit/jitconfigvalues.h b/src/coreclr/jit/jitconfigvalues.h index 6988d11debe332..5c000b334f96e3 100644 --- a/src/coreclr/jit/jitconfigvalues.h +++ b/src/coreclr/jit/jitconfigvalues.h @@ -449,7 +449,7 @@ RELEASE_CONFIG_INTEGER(EnableApxNDD, "EnableApxNDD", // clang-format on -#ifdef FEATURE_SIMD +#ifdef FEATURE_HW_INTRINSICS // Default 0, ValueNumbering of SIMD nodes and HW Intrinsic nodes enabled // If 1, then disable ValueNumbering of SIMD nodes // If 2, then disable ValueNumbering of HW Intrinsic nodes diff --git a/src/coreclr/jit/lclmorph.cpp b/src/coreclr/jit/lclmorph.cpp index 89d06d24505cbc..84988b1a3ccc68 100644 --- a/src/coreclr/jit/lclmorph.cpp +++ b/src/coreclr/jit/lclmorph.cpp @@ -1680,7 +1680,7 @@ class LocalAddressVisitor final : public GenTreeVisitor *use = m_compiler->gtNewCastNode(genActualType(indir), lclNode, false, indir->TypeGet()); break; -#ifdef FEATURE_HW_INTRINSICS +#ifdef FEATURE_SIMD // We have three cases we want to handle: // 1. Vector2/3/4 and Quaternion where we have 2-4x float fields // 2. Plane where we have 1x Vector3 and 1x float field @@ -1819,7 +1819,7 @@ class LocalAddressVisitor final : public GenTreeVisitor lclNode = indir->AsLclVarCommon(); } break; -#endif // FEATURE_HW_INTRINSICS +#endif // FEATURE_SIMD case IndirTransform::LclVar: // TODO-ADDR: use "BashToLclVar" here. @@ -1933,7 +1933,7 @@ class LocalAddressVisitor final : public GenTreeVisitor return IndirTransform::LclFld; } -#ifdef FEATURE_HW_INTRINSICS +#ifdef FEATURE_SIMD if (varTypeIsSIMD(varDsc)) { // We have three cases we want to handle: @@ -1976,7 +1976,7 @@ class LocalAddressVisitor final : public GenTreeVisitor } #endif // FEATURE_SIMD && TARGET_XARCH } -#endif // FEATURE_HW_INTRINSICS +#endif // FEATURE_SIMD if ((!isDef) && (offset == 0)) { diff --git a/src/coreclr/jit/lower.cpp b/src/coreclr/jit/lower.cpp index a859ade2eb780e..172dc0e2bd1b29 100644 --- a/src/coreclr/jit/lower.cpp +++ b/src/coreclr/jit/lower.cpp @@ -11056,7 +11056,7 @@ bool Lowering::TryLowerAndNegativeOne(GenTreeOp* node, GenTree** nextNode) return true; } -#if defined(FEATURE_HW_INTRINSICS) +#if defined(FEATURE_SIMD) //---------------------------------------------------------------------------------------------- // Lowering::InsertNewSimdCreateScalarUnsafeNode: Inserts a new simd CreateScalarUnsafe node // @@ -11088,7 +11088,7 @@ GenTree* Lowering::InsertNewSimdCreateScalarUnsafeNode(var_types simdType, } return result; } -#endif // FEATURE_HW_INTRINSICS +#endif // FEATURE_SIMD //---------------------------------------------------------------------------------------------- // Lowering::RequireOutgoingArgSpace: Record that the compilation will require diff --git a/src/coreclr/jit/lower.h b/src/coreclr/jit/lower.h index d44880bd947554..3dd1d37256835c 100644 --- a/src/coreclr/jit/lower.h +++ b/src/coreclr/jit/lower.h @@ -436,10 +436,12 @@ class Lowering final : public Phase #ifdef FEATURE_HW_INTRINSICS GenTree* LowerHWIntrinsic(GenTreeHWIntrinsic* node); void LowerHWIntrinsicCC(GenTreeHWIntrinsic* node, NamedIntrinsic newIntrinsicId, GenCondition condition); +#ifdef FEATURE_SIMD GenTree* LowerHWIntrinsicCmpOp(GenTreeHWIntrinsic* node, genTreeOps cmpOp); GenTree* LowerHWIntrinsicCreate(GenTreeHWIntrinsic* node); GenTree* LowerHWIntrinsicDot(GenTreeHWIntrinsic* node); GenTree* LowerHWIntrinsicCndSel(GenTreeHWIntrinsic* node); +#endif // FEATURE_SIMD #if defined(TARGET_XARCH) void LowerFusedMultiplyAdd(GenTreeHWIntrinsic* node); GenTree* LowerHWIntrinsicToScalar(GenTreeHWIntrinsic* node); @@ -458,10 +460,12 @@ class Lowering final : public Phase bool TryLowerAddForPossibleContainment(GenTreeOp* node, GenTree** next); void StoreFFRValue(GenTreeHWIntrinsic* node); #endif // !TARGET_XARCH && !TARGET_ARM64 +#ifdef FEATURE_SIMD GenTree* InsertNewSimdCreateScalarUnsafeNode(var_types type, GenTree* op1, CorInfoType simdBaseJitType, unsigned simdSize); +#endif // FEATURE_SIMD #endif // FEATURE_HW_INTRINSICS // Utility functions diff --git a/src/coreclr/jit/lowerriscv64.cpp b/src/coreclr/jit/lowerriscv64.cpp index 5120fe34807ac0..6ba3ba4dfccab8 100644 --- a/src/coreclr/jit/lowerriscv64.cpp +++ b/src/coreclr/jit/lowerriscv64.cpp @@ -511,28 +511,50 @@ void Lowering::LowerSIMD(GenTreeSIMD* simdNode) // Arguments: // node - The hardware intrinsic node. // -void Lowering::LowerHWIntrinsic(GenTreeHWIntrinsic* node) +GenTree* Lowering::LowerHWIntrinsic(GenTreeHWIntrinsic* node) { - NYI_RISCV64("LowerHWIntrinsic"); -} + switch (node->GetHWIntrinsicId()) + { + case NI_RiscV64Base_FusedMultiplyAddScalar: + { + auto removeNegation = [this, node](size_t operandIndex) { + GenTree*& operand = node->Op(operandIndex); + bool isNegated = operand->OperIs(GT_NEG); + if (isNegated) + { + BlockRange().Remove(operand); + operand = operand->gtGetOp1(); + } + return isNegated; + }; + + // "Negated" on RISC-V refers to the whole (op1 * op2 + op3) expression so: + bool isNegated = (removeNegation(1) != removeNegation(2)); // product sign determines negation, + bool isSubtract = (removeNegation(3) != isNegated); // negation flips add/subtract op3. + assert(node->GetOperandCount() == 3); + + static const NamedIntrinsic base = NI_RiscV64Base_FusedMultiplyAddScalar; + static_assert((base + (0 << 1) + 0) == NI_RiscV64Base_FusedMultiplyAddScalar, ""); + static_assert((base + (0 << 1) + 1) == NI_RiscV64Base_FusedMultiplySubtractScalar, ""); + static_assert((base + (1 << 1) + 0) == NI_RiscV64Base_FusedNegatedMultiplyAddScalar, ""); + static_assert((base + (1 << 1) + 1) == NI_RiscV64Base_FusedNegatedMultiplySubtractScalar, ""); + int index = ((int)isNegated << 1) + (int)isSubtract; + if (index != 0) + { + node->ChangeHWIntrinsicId((NamedIntrinsic)(base + index)); + } + break; + } -//---------------------------------------------------------------------------------------------- -// Lowering::IsValidConstForMovImm: Determines if the given node can be replaced by a mov/fmov immediate instruction -// -// Arguments: -// node - The hardware intrinsic node. -// -// Returns: -// true if the node can be replaced by a mov/fmov immediate instruction; otherwise, false -// -// IMPORTANT: -// This check may end up modifying node->gtOp1 if it is a cast node that can be removed -bool Lowering::IsValidConstForMovImm(GenTreeHWIntrinsic* node) -{ - NYI_RISCV64("IsValidConstForMovImm"); - return false; + default: + break; + } + + ContainCheckHWIntrinsic(node); + return node->gtNext; } +#ifdef FEATURE_SIMD //---------------------------------------------------------------------------------------------- // Lowering::LowerHWIntrinsicCmpOp: Lowers a Vector128 or Vector256 comparison intrinsic // @@ -540,9 +562,10 @@ bool Lowering::IsValidConstForMovImm(GenTreeHWIntrinsic* node) // node - The hardware intrinsic node. // cmpOp - The comparison operation, currently must be GT_EQ or GT_NE // -void Lowering::LowerHWIntrinsicCmpOp(GenTreeHWIntrinsic* node, genTreeOps cmpOp) +GenTree* Lowering::LowerHWIntrinsicCmpOp(GenTreeHWIntrinsic* node, genTreeOps cmpOp) { NYI_RISCV64("LowerHWIntrinsicCmpOp"); + return node->gtNext; } //---------------------------------------------------------------------------------------------- @@ -551,9 +574,10 @@ void Lowering::LowerHWIntrinsicCmpOp(GenTreeHWIntrinsic* node, genTreeOps cmpOp) // Arguments: // node - The hardware intrinsic node. // -void Lowering::LowerHWIntrinsicCreate(GenTreeHWIntrinsic* node) +GenTree* Lowering::LowerHWIntrinsicCreate(GenTreeHWIntrinsic* node) { NYI_RISCV64("LowerHWIntrinsicCreate"); + return node->gtNext; } //---------------------------------------------------------------------------------------------- @@ -562,11 +586,12 @@ void Lowering::LowerHWIntrinsicCreate(GenTreeHWIntrinsic* node) // Arguments: // node - The hardware intrinsic node. // -void Lowering::LowerHWIntrinsicDot(GenTreeHWIntrinsic* node) +GenTree* Lowering::LowerHWIntrinsicDot(GenTreeHWIntrinsic* node) { NYI_RISCV64("LowerHWIntrinsicDot"); + return node->gtNext; } - +#endif // FEATURE_SIMD #endif // FEATURE_HW_INTRINSICS //------------------------------------------------------------------------ @@ -817,7 +842,9 @@ void Lowering::ContainCheckSIMD(GenTreeSIMD* simdNode) // void Lowering::ContainCheckHWIntrinsic(GenTreeHWIntrinsic* node) { - NYI_RISCV64("ContainCheckHWIntrinsic"); + NamedIntrinsic id = node->GetHWIntrinsicId(); + assert(!HWIntrinsicInfo::HasImmediateOperand(id)); + assert(!HWIntrinsicInfo::SupportsContainment(id)); } #endif // FEATURE_HW_INTRINSICS diff --git a/src/coreclr/jit/lsrariscv64.cpp b/src/coreclr/jit/lsrariscv64.cpp index 2d67e5ce33e50b..57cee972040393 100644 --- a/src/coreclr/jit/lsrariscv64.cpp +++ b/src/coreclr/jit/lsrariscv64.cpp @@ -781,10 +781,24 @@ int LinearScan::BuildSIMD(GenTreeSIMD* simdTree) // Return Value: // The number of sources consumed by this node. // -int LinearScan::BuildHWIntrinsic(GenTreeHWIntrinsic* intrinsicTree) +int LinearScan::BuildHWIntrinsic(GenTreeHWIntrinsic* intrinsic, int* pDstCount) { - NYI_RISCV64("-----unimplemented on RISCV64 yet----"); - return 0; + for (GenTree* op : intrinsic->Operands()) + { + assert(!op->isContained()); + int srcRegs = BuildOperandUses(op); + assert(srcRegs == 1); + } + + buildInternalRegisterUses(); + + assert(!HWIntrinsicInfo::IsMultiReg(intrinsic->GetHWIntrinsicId())); + assert(pDstCount != nullptr); + assert(intrinsic->IsValue()); + *pDstCount = 1; + BuildDef(intrinsic); + + return intrinsic->GetOperandCount(); } #endif diff --git a/src/coreclr/jit/morph.cpp b/src/coreclr/jit/morph.cpp index 79b9d2743c9375..ecbb1767dfbb85 100644 --- a/src/coreclr/jit/morph.cpp +++ b/src/coreclr/jit/morph.cpp @@ -9226,6 +9226,7 @@ GenTree* Compiler::fgOptimizeHWIntrinsic(GenTreeHWIntrinsic* node) switch (intrinsicId) { +#ifdef FEATURE_SIMD #if defined(TARGET_ARM64) case NI_Vector64_Create: #endif // TARGET_ARM64 @@ -9338,6 +9339,7 @@ GenTree* Compiler::fgOptimizeHWIntrinsic(GenTreeHWIntrinsic* node) node->SetMorphed(this); return node; } +#endif // FEATURE_SIMD default: { @@ -9766,6 +9768,7 @@ GenTree* Compiler::fgOptimizeHWIntrinsicAssociative(GenTreeHWIntrinsic* tree) return nullptr; } +#ifdef FEATURE_SIMD GenTreeVecCon* cns1 = intrinOp1->Op(2)->AsVecCon(); GenTreeVecCon* cns2 = tree->Op(2)->AsVecCon(); @@ -9801,6 +9804,9 @@ GenTree* Compiler::fgOptimizeHWIntrinsicAssociative(GenTreeHWIntrinsic* tree) assert(tree->Op(2) == cns1); return tree; } +#else + return nullptr; +#endif // FEATURE_SIMD } #endif // FEATURE_HW_INTRINSICS @@ -10788,6 +10794,7 @@ GenTree* Compiler::fgMorphHWIntrinsic(GenTreeHWIntrinsic* tree) { operand->SetDoNotCSE(); } +#ifdef FEATURE_SIMD else if (canBenefitFromConstantProp && operand->IsCnsVec()) { if (tree->ShouldConstantProp(operand, operand->AsVecCon())) @@ -10795,6 +10802,7 @@ GenTree* Compiler::fgMorphHWIntrinsic(GenTreeHWIntrinsic* tree) operand->SetDoNotCSE(); } } +#endif // FEATURE_SIMD } else { @@ -10852,6 +10860,7 @@ GenTree* Compiler::fgMorphHWIntrinsic(GenTreeHWIntrinsic* tree) std::swap(op1, tree->Op(2)); } } +#ifdef FEATURE_SIMD else { bool isScalar = false; @@ -10889,6 +10898,7 @@ GenTree* Compiler::fgMorphHWIntrinsic(GenTreeHWIntrinsic* tree) } } } +#endif // FEATURE_SIMD // Try to fold it, maybe we get lucky, GenTree* morphedTree = gtFoldExpr(tree); @@ -10912,13 +10922,14 @@ GenTree* Compiler::fgMorphHWIntrinsic(GenTreeHWIntrinsic* tree) morphedTree = fgOptimizeHWIntrinsic(tree); } - +#ifdef FEATURE_SIMD if (retType != morphedTree->TypeGet()) { assert(varTypeIsMask(morphedTree)); morphedTree = gtNewSimdCvtMaskToVectorNode(retType, morphedTree, simdBaseJitType, simdSize); morphedTree = gtFoldExpr(morphedTree); } +#endif return morphedTree; } diff --git a/src/coreclr/jit/namedintrinsiclist.h b/src/coreclr/jit/namedintrinsiclist.h index ddee451a200bea..109ab9ce2fc2c2 100644 --- a/src/coreclr/jit/namedintrinsiclist.h +++ b/src/coreclr/jit/namedintrinsiclist.h @@ -164,7 +164,13 @@ enum NamedIntrinsic : unsigned short #define HARDWARE_INTRINSIC(isa, name, size, numarg, t1, t2, t3, t4, t5, t6, t7, t8, t9, t10, category, flag) \ NI_##isa##_##name, #include "hwintrinsiclistarm64.h" -#endif // !defined(TARGET_XARCH) && !defined(TARGET_ARM64) +#elif defined(TARGET_RISCV64) +#define HARDWARE_INTRINSIC(isa, name, size, numarg, t1, t2, t3, t4, t5, t6, t7, t8, t9, t10, category, flag) \ + NI_##isa##_##name, +#include "hwintrinsiclistriscv64.h" +#else +#error "Platform with FEATURE_HW_INTRINSICS but no hwintrinsiclist" +#endif // !defined(TARGET_XARCH) && !defined(TARGET_ARM64) && !defined(TARGET_RISCV64) NI_HW_INTRINSIC_END, #endif // FEATURE_HW_INTRINSICS diff --git a/src/coreclr/jit/optcse.cpp b/src/coreclr/jit/optcse.cpp index 9afb109802e85f..6613eee272a228 100644 --- a/src/coreclr/jit/optcse.cpp +++ b/src/coreclr/jit/optcse.cpp @@ -1866,11 +1866,15 @@ bool CSE_HeuristicCommon::CanConsiderTree(GenTree* tree, bool isReturn) case HW_Category_Scalar: case HW_Category_Helper: break; +#elif defined(TARGET_RISCV64) + case HW_Category_Scalar: + break; #endif - +#if defined(TARGET_XARCH) || defined(TARGET_ARM64) case HW_Category_MemoryLoad: case HW_Category_MemoryStore: case HW_Category_Special: +#endif // TARGET_XARCH || TARGET_ARM64 default: return false; } diff --git a/src/coreclr/jit/rationalize.cpp b/src/coreclr/jit/rationalize.cpp index be2e6f619447b0..7467085ba8d42d 100644 --- a/src/coreclr/jit/rationalize.cpp +++ b/src/coreclr/jit/rationalize.cpp @@ -340,6 +340,7 @@ void Rationalizer::RewriteHWIntrinsicAsUserCall(GenTree** use, ArrayStackIsVNConstant(argVN)); @@ -7777,6 +7778,7 @@ ValueNum EvaluateSimdCvtVectorToMask(ValueNumStore* vns, var_types simdType, var return vns->VNForSimdMaskCon(result); } +#endif // FEATURE_SIMD ValueNum ValueNumStore::EvalHWIntrinsicFunUnary(GenTreeHWIntrinsic* tree, VNFunc func, @@ -7793,6 +7795,7 @@ ValueNum ValueNumStore::EvalHWIntrinsicFunUnary(GenTreeHWIntrinsic* tree, bool isScalar = false; genTreeOps oper = tree->GetOperForHWIntrinsicId(&isScalar); +#ifdef FEATURE_SIMD if (oper != GT_NONE) { if (varTypeIsMask(type)) @@ -7814,9 +7817,11 @@ ValueNum ValueNumStore::EvalHWIntrinsicFunUnary(GenTreeHWIntrinsic* tree, var_types simdType = Compiler::getSIMDTypeForSize(simdSize); return EvaluateSimdCvtVectorToMask(this, simdType, baseType, arg0VN); } +#endif // FEATURE_SIMD switch (ni) { +#if defined(TARGET_ARM64) || defined(TARGET_XARCH) #ifdef TARGET_ARM64 case NI_ArmBase_LeadingZeroCount: #else @@ -7830,6 +7835,7 @@ ValueNum ValueNumStore::EvalHWIntrinsicFunUnary(GenTreeHWIntrinsic* tree, return VNForIntCon(static_cast(result)); } +#endif // TARGET_ARM64 || TARGET_XARCH #ifdef TARGET_ARM64 case NI_ArmBase_Arm64_LeadingZeroCount: @@ -7841,7 +7847,7 @@ ValueNum ValueNumStore::EvalHWIntrinsicFunUnary(GenTreeHWIntrinsic* tree, return VNForIntCon(static_cast(result)); } -#else +#elif defined(TARGET_XARCH) case NI_LZCNT_X64_LeadingZeroCount: { assert(varTypeIsLong(type)); @@ -8063,7 +8069,7 @@ ValueNum ValueNumStore::EvalHWIntrinsicFunUnary(GenTreeHWIntrinsic* tree, return VNForSimd16Con(result); } #endif // TARGET_XARCH - +#ifdef FEATURE_SIMD case NI_Vector128_AsVector3: { simd12_t result = {}; @@ -8109,7 +8115,7 @@ ValueNum ValueNumStore::EvalHWIntrinsicFunUnary(GenTreeHWIntrinsic* tree, { return EvaluateSimdGetElement(this, TypeOfVN(arg0VN), baseType, arg0VN, 0); } - +#endif // FEATURE_SIMD default: break; } @@ -8155,11 +8161,11 @@ ValueNum ValueNumStore::EvalHWIntrinsicFunBinary( bool isScalar = false; genTreeOps oper = tree->GetOperForHWIntrinsicId(&isScalar); +#ifdef FEATURE_SIMD if (oper != GT_NONE) { // We shouldn't find AND_NOT, OR_NOT or XOR_NOT nodes since it should only be produced in lowering assert((oper != GT_AND_NOT) && (oper != GT_OR_NOT) && (oper != GT_XOR_NOT)); - if (varTypeIsMask(type)) { if (varTypeIsMask(TypeOfVN(arg0VN))) @@ -8294,6 +8300,7 @@ ValueNum ValueNumStore::EvalHWIntrinsicFunBinary( default: break; } +#endif // FEATURE_SIMD } else if (cnsVN != NoVN) { @@ -8320,6 +8327,7 @@ ValueNum ValueNumStore::EvalHWIntrinsicFunBinary( // Handle `x + NaN == NaN` and `NaN + x == NaN` // This is safe for all floats since we do not fault for sNaN +#ifdef FEATURE_SIMD if (VNIsVectorNaN(type, baseType, cnsVN)) { return cnsVN; @@ -8331,8 +8339,8 @@ ValueNum ValueNumStore::EvalHWIntrinsicFunBinary( { return argVN; } - - // We cannot handle `x + 0 == x` or `0 + x == x` since `-0 + 0 == 0` +#endif // FEATURE_SIMD + // We cannot handle `x + 0 == x` or `0 + x == x` since `-0 + 0 == 0` break; } @@ -8368,6 +8376,7 @@ ValueNum ValueNumStore::EvalHWIntrinsicFunBinary( case GT_DIV: { +#ifdef FEATURE_SIMD if (varTypeIsFloating(baseType)) { // Handle `x / NaN == NaN` and `NaN / x == NaN` @@ -8378,16 +8387,19 @@ ValueNum ValueNumStore::EvalHWIntrinsicFunBinary( return cnsVN; } } +#endif // FEATURE_SIMD // Handle `x / 1 == x`. // This is safe for all floats since we do not fault for sNaN ValueNum oneVN; +#ifdef FEATURE_SIMD if (varTypeIsSIMD(TypeOfVN(arg1VN))) { oneVN = VNOneForSimdType(type, baseType); } else +#endif // FEATURE_SIMD { oneVN = VNOneForType(baseType); } @@ -8399,6 +8411,7 @@ ValueNum ValueNumStore::EvalHWIntrinsicFunBinary( break; } +#ifdef FEATURE_SIMD case GT_EQ: { if (varTypeIsFloating(baseType)) @@ -8505,6 +8518,7 @@ ValueNum ValueNumStore::EvalHWIntrinsicFunBinary( } break; } +#endif // FEATURE_SIMD case GT_MUL: { @@ -8523,25 +8537,26 @@ ValueNum ValueNumStore::EvalHWIntrinsicFunBinary( { // Handle `x * NaN == NaN` and `NaN * x == NaN` // This is safe for all floats since we do not fault for sNaN - +#ifdef FEATURE_SIMD if (VNIsVectorNaN(type, baseType, cnsVN)) { return cnsVN; } - - // We cannot handle `x * 0 == 0` or ` 0 * x == 0` since `-0 * 0 == -0` - // We cannot handle `x * -0 == -0` or `-0 * x == -0` since `-0 * -0 == 0` +#endif // FEATURE_SIMD + // We cannot handle `x * 0 == 0` or ` 0 * x == 0` since `-0 * 0 == -0` + // We cannot handle `x * -0 == -0` or `-0 * x == -0` since `-0 * -0 == 0` } // Handle `x * 1 == x` and `1 * x == x` // This is safe for all floats since we do not fault for sNaN ValueNum oneVN; - +#ifdef FEATURE_SIMD if (varTypeIsSIMD(TypeOfVN(cnsVN))) { oneVN = VNOneForSimdType(type, baseType); } else +#endif // FEATURE_SIMD { oneVN = VNOneForType(baseType); } @@ -8553,6 +8568,7 @@ ValueNum ValueNumStore::EvalHWIntrinsicFunBinary( break; } +#ifdef FEATURE_SIMD case GT_NE: { var_types simdType = Compiler::getSIMDTypeForSize(simdSize); @@ -8567,6 +8583,7 @@ ValueNum ValueNumStore::EvalHWIntrinsicFunBinary( } break; } +#endif // FEATURE_SIMD case GT_OR: { @@ -8610,6 +8627,7 @@ ValueNum ValueNumStore::EvalHWIntrinsicFunBinary( case GT_SUB: { +#ifdef FEATURE_SIMD if (varTypeIsFloating(baseType)) { // Handle `x - NaN == NaN` and `NaN - x == NaN` @@ -8622,6 +8640,7 @@ ValueNum ValueNumStore::EvalHWIntrinsicFunBinary( // We cannot handle `x - -0 == x` since `-0 - -0 == 0` } +#endif // FEATURE_SIMD // Handle `x - 0 == x` ValueNum zeroVN = VNZeroForType(type); @@ -8734,7 +8753,7 @@ ValueNum ValueNumStore::EvalHWIntrinsicFunBinary( break; } #endif - +#ifdef FEATURE_SIMD case NI_Vector128_op_Equality: #if defined(TARGET_ARM64) case NI_Vector64_op_Equality: @@ -8776,7 +8795,7 @@ ValueNum ValueNumStore::EvalHWIntrinsicFunBinary( } break; } - +#endif // FEATURE_SIMD default: { break; @@ -8858,7 +8877,7 @@ ValueNum ValueNumStore::EvalHWIntrinsicFunBinary( default: break; } - +#ifdef FEATURE_SIMD switch (ni) { case NI_Vector128_op_Equality: @@ -8900,11 +8919,13 @@ ValueNum ValueNumStore::EvalHWIntrinsicFunBinary( break; } } +#endif // FEATURE_SIMD } return VNForFunc(type, func, arg0VN, arg1VN, resultTypeVN); } +#ifdef FEATURE_SIMD ValueNum EvaluateSimdWithElementFloating( ValueNumStore* vns, var_types simdType, var_types baseType, ValueNum arg0VN, int32_t arg1, double arg2) { @@ -9012,11 +9033,13 @@ ValueNum EvaluateSimdWithElementIntegral( } } } +#endif // FEATURE_SIMD ValueNum ValueNumStore::EvalHWIntrinsicFunTernary( GenTreeHWIntrinsic* tree, VNFunc func, ValueNum arg0VN, ValueNum arg1VN, ValueNum arg2VN, ValueNum resultTypeVN) { - var_types type = tree->TypeGet(); + var_types type = tree->TypeGet(); +#ifdef FEATURE_SIMD var_types baseType = tree->GetSimdBaseType(); NamedIntrinsic ni = tree->GetHWIntrinsicId(); @@ -9124,6 +9147,7 @@ ValueNum ValueNumStore::EvalHWIntrinsicFunTernary( break; } } +#endif // FEATURE_SIMD return VNForFunc(type, func, arg0VN, arg1VN, arg2VN, resultTypeVN); } @@ -12933,7 +12957,6 @@ void Compiler::fgValueNumberHWIntrinsic(GenTreeHWIntrinsic* tree) ValueNumPair normalPair = ValueNumPair(); const size_t opCount = tree->GetOperandCount(); - if ((opCount > 3) || ((JitConfig.JitDisableSimdVN() & 2) == 2) || HWIntrinsicInfo::HasSpecialSideEffect(intrinsicId)) { diff --git a/src/coreclr/jit/valuenum.h b/src/coreclr/jit/valuenum.h index a643789a2843ee..e61a367839a397 100644 --- a/src/coreclr/jit/valuenum.h +++ b/src/coreclr/jit/valuenum.h @@ -692,7 +692,6 @@ class ValueNumStore // It has an unreached() for a "typ" that has no all bits set value, such as TYP_VOID. ValueNum VNAllBitsForType(var_types typ); -#ifdef FEATURE_SIMD // Returns the value number broadcast of the given "simdType" and "simdBaseType". ValueNum VNBroadcastForSimdType(var_types simdType, var_types simdBaseType, ValueNum valVN); @@ -702,6 +701,7 @@ class ValueNumStore // A helper function for constructing VNF_SimdType VNs. ValueNum VNForSimdType(unsigned simdSize, CorInfoType simdBaseJitType); +#ifdef FEATURE_SIMD // Returns if a value number represents NaN in all elements bool VNIsVectorNaN(var_types simdType, var_types simdBaseType, ValueNum valVN); diff --git a/src/coreclr/jit/valuenumfuncs.h b/src/coreclr/jit/valuenumfuncs.h index 227f4f26b11e9b..6a5dac01e6d3a6 100644 --- a/src/coreclr/jit/valuenumfuncs.h +++ b/src/coreclr/jit/valuenumfuncs.h @@ -176,7 +176,7 @@ ValueNumFuncDef(ADD_UN_OVF, 2, true, false, false) // unsigned overflow checkin ValueNumFuncDef(SUB_UN_OVF, 2, false, false, false) ValueNumFuncDef(MUL_UN_OVF, 2, true, false, false) -#ifdef FEATURE_SIMD +#ifdef FEATURE_HW_INTRINSICS ValueNumFuncDef(SimdType, 2, false, false, false) // A value number function to compose a SIMD type #endif @@ -201,8 +201,10 @@ ValueNumFuncDef(HWI_##isa##_##name, ((argCount == -1) ? -1 : (argCount + 1)), (( //TODO-LOONGARCH64-CQ: add LoongArch64's Hardware Intrinsics Instructions if supported. #elif defined (TARGET_RISCV64) - //TODO-RISCV64-CQ: add RISCV64's Hardware Intrinsics Instructions if supported. - +#define HARDWARE_INTRINSIC(isa, name, size, argCount, t1, t2, t3, t4, t5, t6, t7, t8, t9, t10, category, flag) \ +ValueNumFuncDef(HWI_##isa##_##name, ((argCount == -1) ? -1 : (argCount + 1)), ((flag) & HW_Flag_Commutative) >> 0, false, false) // All of the HARDWARE_INTRINSICS for riscv64 +#include "hwintrinsiclistriscv64.h" +#define VNF_HWI_FIRST FIRST_NI_RiscV64Base #else #error Unsupported platform #endif diff --git a/src/tests/readytorun/JittedMethodsCountingTest/JittedMethodsCountingTest.cs b/src/tests/readytorun/JittedMethodsCountingTest/JittedMethodsCountingTest.cs index f4ac4837c00a6c..33bdddb142dde6 100644 --- a/src/tests/readytorun/JittedMethodsCountingTest/JittedMethodsCountingTest.cs +++ b/src/tests/readytorun/JittedMethodsCountingTest/JittedMethodsCountingTest.cs @@ -6,25 +6,17 @@ using InteropServices = System.Runtime.InteropServices; using JitInfo = System.Runtime.JitInfo; +using X86 = System.Runtime.Intrinsics.X86; public class JittedMethodsCountingTest { private const int MAX_JITTED_METHODS_ACCEPTED = 70; - [Fact] + public static bool IsEnabled => IsReadyToRunEnabled() && IsHardwareIntrinsicsEnabled() && IsSSEEnabled(); + + [ConditionalFact(typeof(JittedMethodsCountingTest), nameof(IsEnabled))] public static int TestEntryPoint() { - // If either of DOTNET_ReadyToRun, DOTNET_EnableHWIntrinsics, or - // DOTNET_EnableSSE(2) are disabled (i.e. set to "0"), then this test - // ought to be skipped. - if (!IsReadyToRunEnabled() || !IsHardwareIntrinsicsEnabled() || !IsSSEEnabled()) - { - Console.WriteLine("\nThis test is only supported in ReadyToRun scenarios" - + " with Hardware Intrinsics and SSE(2) enabled." - + " Skipping...\n"); - return 100; - } - Console.WriteLine("\nHello World from Jitted Methods Counting Test!"); long jits = JitInfo.GetCompiledMethodCount(currentThread: false); @@ -51,12 +43,5 @@ private static bool IsHardwareIntrinsicsEnabled() || dotnetEnableHWIntrinsics != "0"); } - private static bool IsSSEEnabled() - { - string? dotnetSSE = Environment.GetEnvironmentVariable("DOTNET_EnableSSE"); - string? dotnetSSE2 = Environment.GetEnvironmentVariable("DOTNET_EnableSSE2"); - - return ((string.IsNullOrEmpty(dotnetSSE) || dotnetSSE != "0") - && (string.IsNullOrEmpty(dotnetSSE2) || dotnetSSE2 != "0")); - } + private static bool IsSSEEnabled() => X86.Sse.IsSupported || X86.Sse2.IsSupported; }