diff --git a/clang/lib/CIR/CodeGen/CIRGenBuilder.h b/clang/lib/CIR/CodeGen/CIRGenBuilder.h index 601f58382a4a..15209b64b2fd 100644 --- a/clang/lib/CIR/CodeGen/CIRGenBuilder.h +++ b/clang/lib/CIR/CodeGen/CIRGenBuilder.h @@ -796,6 +796,34 @@ class CIRGenBuilderTy : public CIRBaseBuilderTy { return CIRBaseBuilderTy::createStore(loc, flag, dst); } + mlir::cir::VecShuffleOp + createVecShuffle(mlir::Location loc, mlir::Value vec1, mlir::Value vec2, + llvm::ArrayRef maskAttrs) { + auto vecType = mlir::cast(vec1.getType()); + auto resultTy = mlir::cir::VectorType::get( + getContext(), vecType.getEltType(), maskAttrs.size()); + return CIRBaseBuilderTy::create( + loc, resultTy, vec1, vec2, getArrayAttr(maskAttrs)); + } + + mlir::cir::VecShuffleOp createVecShuffle(mlir::Location loc, mlir::Value vec1, + mlir::Value vec2, + llvm::ArrayRef mask) { + llvm::SmallVector maskAttrs; + for (int32_t idx : mask) { + maskAttrs.push_back(mlir::cir::IntAttr::get(getSInt32Ty(), idx)); + } + + return createVecShuffle(loc, vec1, vec2, maskAttrs); + } + + mlir::cir::VecShuffleOp createVecShuffle(mlir::Location loc, mlir::Value vec1, + llvm::ArrayRef mask) { + // FIXME(cir): Support use cir.vec.shuffle with single vec + // Workaround: pass Vec as both vec1 and vec2 + return createVecShuffle(loc, vec1, vec1, mask); + } + mlir::cir::StoreOp createAlignedStore(mlir::Location loc, mlir::Value val, mlir::Value dst, clang::CharUnits align = clang::CharUnits::One(), diff --git a/clang/lib/CIR/CodeGen/CIRGenExpr.cpp b/clang/lib/CIR/CodeGen/CIRGenExpr.cpp index bb89ba0e404e..739745bbbfff 100644 --- a/clang/lib/CIR/CodeGen/CIRGenExpr.cpp +++ b/clang/lib/CIR/CodeGen/CIRGenExpr.cpp @@ -653,9 +653,55 @@ RValue CIRGenFunction::buildLoadOfLValue(LValue LV, SourceLocation Loc) { getLoc(Loc), load, LV.getVectorIdx())); } + if (LV.isExtVectorElt()) { + return buildLoadOfExtVectorElementLValue(LV); + } + llvm_unreachable("NYI"); } +int64_t CIRGenFunction::getAccessedFieldNo(unsigned int idx, + const mlir::ArrayAttr elts) { + auto elt = mlir::dyn_cast(elts[idx]); + assert(elt && "The indices should be integer attributes"); + return elt.getInt(); +} + +// If this is a reference to a subset of the elements of a vector, create an +// appropriate shufflevector. +RValue CIRGenFunction::buildLoadOfExtVectorElementLValue(LValue LV) { + mlir::Location loc = LV.getExtVectorPointer().getLoc(); + mlir::Value Vec = builder.createLoad(loc, LV.getExtVectorAddress()); + + // HLSL allows treating scalars as one-element vectors. Converting the scalar + // IR value to a vector here allows the rest of codegen to behave as normal. + if (getLangOpts().HLSL && !mlir::isa(Vec.getType())) { + llvm_unreachable("HLSL NYI"); + } + + const mlir::ArrayAttr Elts = LV.getExtVectorElts(); + + // If the result of the expression is a non-vector type, we must be extracting + // a single element. Just codegen as an extractelement. + const auto *ExprVT = LV.getType()->getAs(); + if (!ExprVT) { + int64_t InIdx = getAccessedFieldNo(0, Elts); + mlir::cir::ConstantOp Elt = + builder.getConstInt(loc, builder.getSInt64Ty(), InIdx); + return RValue::get(builder.create(loc, Vec, Elt)); + } + + // Always use shuffle vector to try to retain the original program structure + unsigned NumResultElts = ExprVT->getNumElements(); + + SmallVector Mask; + for (unsigned i = 0; i != NumResultElts; ++i) + Mask.push_back(getAccessedFieldNo(i, Elts)); + + Vec = builder.createVecShuffle(loc, Vec, Mask); + return RValue::get(Vec); +} + RValue CIRGenFunction::buildLoadOfBitfieldLValue(LValue LV, SourceLocation Loc) { const CIRGenBitFieldInfo &info = LV.getBitFieldInfo(); @@ -674,6 +720,80 @@ RValue CIRGenFunction::buildLoadOfBitfieldLValue(LValue LV, return RValue::get(field); } +void CIRGenFunction::buildStoreThroughExtVectorComponentLValue(RValue Src, + LValue Dst) { + mlir::Location loc = Dst.getExtVectorPointer().getLoc(); + + // HLSL allows storing to scalar values through ExtVector component LValues. + // To support this we need to handle the case where the destination address is + // a scalar. + Address DstAddr = Dst.getExtVectorAddress(); + if (!mlir::isa(DstAddr.getElementType())) { + llvm_unreachable("HLSL NYI"); + } + + // This access turns into a read/modify/write of the vector. Load the input + // value now. + mlir::Value Vec = builder.createLoad(loc, DstAddr); + const mlir::ArrayAttr Elts = Dst.getExtVectorElts(); + + mlir::Value SrcVal = Src.getScalarVal(); + + if (const clang::VectorType *VTy = + Dst.getType()->getAs()) { + unsigned NumSrcElts = VTy->getNumElements(); + unsigned NumDstElts = cast(Vec.getType()).getSize(); + if (NumDstElts == NumSrcElts) { + // Use shuffle vector is the src and destination are the same number of + // elements and restore the vector mask since it is on the side it will be + // stored. + SmallVector Mask(NumDstElts); + for (unsigned i = 0; i != NumSrcElts; ++i) + Mask[getAccessedFieldNo(i, Elts)] = i; + + Vec = builder.createVecShuffle(loc, SrcVal, Mask); + } else if (NumDstElts > NumSrcElts) { + // Extended the source vector to the same length and then shuffle it + // into the destination. + // FIXME: since we're shuffling with undef, can we just use the indices + // into that? This could be simpler. + SmallVector ExtMask; + for (unsigned i = 0; i != NumSrcElts; ++i) + ExtMask.push_back(i); + ExtMask.resize(NumDstElts, -1); + mlir::Value ExtSrcVal = builder.createVecShuffle(loc, SrcVal, ExtMask); + // build identity + SmallVector Mask; + for (unsigned i = 0; i != NumDstElts; ++i) + Mask.push_back(i); + + // When the vector size is odd and .odd or .hi is used, the last element + // of the Elts constant array will be one past the size of the vector. + // Ignore the last element here, if it is greater than the mask size. + if (getAccessedFieldNo(NumSrcElts - 1, Elts) == Mask.size()) + llvm_unreachable("NYI"); + + // modify when what gets shuffled in + for (unsigned i = 0; i != NumSrcElts; ++i) + Mask[getAccessedFieldNo(i, Elts)] = i + NumDstElts; + Vec = builder.createVecShuffle(loc, Vec, ExtSrcVal, Mask); + } else { + // We should never shorten the vector + llvm_unreachable("unexpected shorten vector length"); + } + } else { + // If the Src is a scalar (not a vector), and the target is a vector it must + // be updating one element. + unsigned InIdx = getAccessedFieldNo(0, Elts); + auto Elt = builder.getSInt64(InIdx, loc); + + Vec = builder.create(loc, Vec, SrcVal, Elt); + } + + builder.createStore(loc, Vec, Dst.getExtVectorAddress(), + Dst.isVolatileQualified()); +} + void CIRGenFunction::buildStoreThroughLValue(RValue Src, LValue Dst, bool isInit) { if (!Dst.isSimple()) { @@ -686,6 +806,10 @@ void CIRGenFunction::buildStoreThroughLValue(RValue Src, LValue Dst, builder.createStore(loc, Vector, Dst.getVectorAddress()); return; } + + if (Dst.isExtVectorElt()) + return buildStoreThroughExtVectorComponentLValue(Src, Dst); + assert(Dst.isBitField() && "NIY LValue type"); mlir::Value result; return buildStoreThroughBitfieldLValue(Src, Dst, result); @@ -979,6 +1103,71 @@ CIRGenFunction::buildPointerToDataMemberBinaryExpr(const BinaryOperator *E) { return makeAddrLValue(memberAddr, memberPtrTy->getPointeeType(), baseInfo); } +LValue +CIRGenFunction::buildExtVectorElementExpr(const ExtVectorElementExpr *E) { + // Emit the base vector as an l-value. + LValue Base; + + // ExtVectorElementExpr's base can either be a vector or pointer to vector. + if (E->isArrow()) { + // If it is a pointer to a vector, emit the address and form an lvalue with + // it. + LValueBaseInfo BaseInfo; + // TODO(cir): Support TBAA + assert(!MissingFeatures::tbaa()); + Address Ptr = buildPointerWithAlignment(E->getBase(), &BaseInfo); + const auto *PT = E->getBase()->getType()->castAs(); + Base = makeAddrLValue(Ptr, PT->getPointeeType(), BaseInfo); + Base.getQuals().removeObjCGCAttr(); + } else if (E->getBase()->isGLValue()) { + // Otherwise, if the base is an lvalue ( as in the case of foo.x.x), + // emit the base as an lvalue. + assert(E->getBase()->getType()->isVectorType()); + Base = buildLValue(E->getBase()); + } else { + // Otherwise, the base is a normal rvalue (as in (V+V).x), emit it as such. + assert(E->getBase()->getType()->isVectorType() && + "Result must be a vector"); + mlir::Value Vec = buildScalarExpr(E->getBase()); + + // Store the vector to memory (because LValue wants an address). + QualType BaseTy = E->getBase()->getType(); + Address VecMem = CreateMemTemp(BaseTy, Vec.getLoc(), "tmp"); + builder.createStore(Vec.getLoc(), Vec, VecMem); + Base = makeAddrLValue(VecMem, BaseTy, AlignmentSource::Decl); + } + + QualType type = + E->getType().withCVRQualifiers(Base.getQuals().getCVRQualifiers()); + + // Encode the element access list into a vector of unsigned indices. + SmallVector indices; + E->getEncodedElementAccess(indices); + + if (Base.isSimple()) { + SmallVector attrElts; + for (uint32_t i : indices) { + attrElts.push_back(static_cast(i)); + } + auto elts = builder.getI64ArrayAttr(attrElts); + return LValue::MakeExtVectorElt(Base.getAddress(), elts, type, + Base.getBaseInfo()); + } + assert(Base.isExtVectorElt() && "Can only subscript lvalue vec elts here!"); + + mlir::ArrayAttr baseElts = Base.getExtVectorElts(); + + // Composite the two indices + SmallVector attrElts; + for (uint32_t i : indices) { + attrElts.push_back(getAccessedFieldNo(i, baseElts)); + } + auto elts = builder.getI64ArrayAttr(attrElts); + + return LValue::MakeExtVectorElt(Base.getExtVectorAddress(), elts, type, + Base.getBaseInfo()); +} + LValue CIRGenFunction::buildBinaryOperatorLValue(const BinaryOperator *E) { // Comma expressions just emit their LHS then their RHS as an l-value. if (E->getOpcode() == BO_Comma) { @@ -2263,6 +2452,8 @@ LValue CIRGenFunction::buildLValue(const Expr *E) { return buildConditionalOperatorLValue(cast(E)); case Expr::ArraySubscriptExprClass: return buildArraySubscriptExpr(cast(E)); + case Expr::ExtVectorElementExprClass: + return buildExtVectorElementExpr(cast(E)); case Expr::BinaryOperatorClass: return buildBinaryOperatorLValue(cast(E)); case Expr::CompoundAssignOperatorClass: { diff --git a/clang/lib/CIR/CodeGen/CIRGenExprScalar.cpp b/clang/lib/CIR/CodeGen/CIRGenExprScalar.cpp index 8e57367a35bb..d15a6a6a272a 100644 --- a/clang/lib/CIR/CodeGen/CIRGenExprScalar.cpp +++ b/clang/lib/CIR/CodeGen/CIRGenExprScalar.cpp @@ -328,8 +328,12 @@ class ScalarExprEmitter : public StmtVisitor { E->getSrcExpr()->getType(), E->getType(), E->getSourceRange().getBegin()); } + + mlir::Value VisitExtVectorElementExpr(Expr *E) { + return buildLoadOfLValue(E); + } + mlir::Value VisitMemberExpr(MemberExpr *E); - mlir::Value VisitExtVectorelementExpr(Expr *E) { llvm_unreachable("NYI"); } mlir::Value VisitCompoundLiteralEpxr(CompoundLiteralExpr *E) { llvm_unreachable("NYI"); } diff --git a/clang/lib/CIR/CodeGen/CIRGenFunction.h b/clang/lib/CIR/CodeGen/CIRGenFunction.h index 696df531dfb7..0f0c8f44fa8e 100644 --- a/clang/lib/CIR/CodeGen/CIRGenFunction.h +++ b/clang/lib/CIR/CodeGen/CIRGenFunction.h @@ -763,6 +763,12 @@ class CIRGenFunction : public CIRGenTypeCache { mlir::Location Loc, LValueBaseInfo BaseInfo, bool isNontemporal = false); + int64_t getAccessedFieldNo(unsigned idx, const mlir::ArrayAttr elts); + + RValue buildLoadOfExtVectorElementLValue(LValue LV); + + void buildStoreThroughExtVectorComponentLValue(RValue Src, LValue Dst); + RValue buildLoadOfBitfieldLValue(LValue LV, SourceLocation Loc); /// Load a scalar value from an address, taking care to appropriately convert @@ -1223,6 +1229,7 @@ class CIRGenFunction : public CIRGenTypeCache { LValue lvalue, bool capturedByInit = false); LValue buildDeclRefLValue(const clang::DeclRefExpr *E); + LValue buildExtVectorElementExpr(const ExtVectorElementExpr *E); LValue buildBinaryOperatorLValue(const clang::BinaryOperator *E); LValue buildCompoundAssignmentLValue(const clang::CompoundAssignOperator *E); LValue buildUnaryOpLValue(const clang::UnaryOperator *E); diff --git a/clang/lib/CIR/CodeGen/CIRGenValue.h b/clang/lib/CIR/CodeGen/CIRGenValue.h index 4862d32df245..408dcdcd605d 100644 --- a/clang/lib/CIR/CodeGen/CIRGenValue.h +++ b/clang/lib/CIR/CodeGen/CIRGenValue.h @@ -207,7 +207,8 @@ class LValue { unsigned Alignment; mlir::Value V; mlir::Type ElementType; - mlir::Value VectorIdx; // Index for vector subscript + mlir::Value VectorIdx; // Index for vector subscript + mlir::Attribute VectorElts; // ExtVector element subset: V.xyx LValueBaseInfo BaseInfo; const CIRGenBitFieldInfo *BitFieldInfo{0}; @@ -316,6 +317,20 @@ class LValue { return VectorIdx; } + // extended vector elements. + Address getExtVectorAddress() const { + assert(isExtVectorElt()); + return Address(getExtVectorPointer(), ElementType, getAlignment()); + } + mlir::Value getExtVectorPointer() const { + assert(isExtVectorElt()); + return V; + } + mlir::ArrayAttr getExtVectorElts() const { + assert(isExtVectorElt()); + return mlir::cast(VectorElts); + } + static LValue MakeVectorElt(Address vecAddress, mlir::Value Index, clang::QualType type, LValueBaseInfo BaseInfo) { LValue R; @@ -328,6 +343,19 @@ class LValue { return R; } + static LValue MakeExtVectorElt(Address vecAddress, mlir::ArrayAttr elts, + clang::QualType type, + LValueBaseInfo baseInfo) { + LValue R; + R.LVType = ExtVectorElt; + R.V = vecAddress.getPointer(); + R.ElementType = vecAddress.getElementType(); + R.VectorElts = elts; + R.Initialize(type, type.getQualifiers(), vecAddress.getAlignment(), + baseInfo); + return R; + } + // bitfield lvalue Address getBitFieldAddress() const { return Address(getBitFieldPointer(), ElementType, getAlignment()); diff --git a/clang/test/CIR/CodeGen/vectype-ext.cpp b/clang/test/CIR/CodeGen/vectype-ext.cpp new file mode 100644 index 000000000000..a2dd42dc201d --- /dev/null +++ b/clang/test/CIR/CodeGen/vectype-ext.cpp @@ -0,0 +1,454 @@ +// RUN: %clang_cc1 -std=c++17 -fclangir -emit-cir -triple x86_64-unknown-linux-gnu %s -o %t.cir +// RUN: FileCheck --input-file=%t.cir %s -check-prefix=CIR +// RUN: %clang_cc1 -std=c++17 -fclangir -S -emit-llvm -triple x86_64-unknown-linux-gnu %s -o %t.ll +// RUN: FileCheck --input-file=%t.ll %s -check-prefix=LLVM + +typedef int vi4 __attribute__((ext_vector_type(4))); +typedef int vi2 __attribute__((ext_vector_type(2))); +typedef double vd2 __attribute__((ext_vector_type(2))); +typedef long vl2 __attribute__((ext_vector_type(2))); +typedef unsigned short vus2 __attribute__((ext_vector_type(2))); + +// CIR: cir.func {{@.*vector_int_test.*}} +// LLVM: define void {{@.*vector_int_test.*}} +void vector_int_test(int x) { + + // Vector constant. Not yet implemented. Expected results will change from + // cir.vec.create to cir.const. + vi4 a = { 1, 2, 3, 4 }; + // CIR: %{{[0-9]+}} = cir.vec.create(%{{[0-9]+}}, %{{[0-9]+}}, %{{[0-9]+}}, %{{[0-9]+}} : !s32i, !s32i, !s32i, !s32i) : !cir.vector + // LLVM: store <4 x i32> , ptr %{{[0-9]+}}, align 16 + + // Non-const vector initialization. + vi4 b = { x, 5, 6, x + 1 }; + // CIR: %{{[0-9]+}} = cir.vec.create(%{{[0-9]+}}, %{{[0-9]+}}, %{{[0-9]+}}, %{{[0-9]+}} : !s32i, !s32i, !s32i, !s32i) : !cir.vector + // LLVM: %[[#X1:]] = load i32, ptr %{{[0-9]+}}, align 4 + // LLVM-NEXT: %[[#X2:]] = load i32, ptr %{{[0-9]+}}, align 4 + // LLVM-NEXT: %[[#SUM:]] = add i32 %[[#X2]], 1 + // LLVM-NEXT: %[[#VEC1:]] = insertelement <4 x i32> undef, i32 %[[#X1]], i64 0 + // LLVM-NEXT: %[[#VEC2:]] = insertelement <4 x i32> %[[#VEC1]], i32 5, i64 1 + // LLVM-NEXT: %[[#VEC3:]] = insertelement <4 x i32> %[[#VEC2]], i32 6, i64 2 + // LLVM-NEXT: %[[#VEC4:]] = insertelement <4 x i32> %[[#VEC3]], i32 %[[#SUM]], i64 3 + // LLVM-NEXT: store <4 x i32> %[[#VEC4]], ptr %{{[0-9]+}}, align 16 + + // Incomplete vector initialization. + vi4 bb = { x, x + 1 }; + // CIR: %[[#zero:]] = cir.const #cir.int<0> : !s32i + // CIR: %{{[0-9]+}} = cir.vec.create(%{{[0-9]+}}, %{{[0-9]+}}, %[[#zero]], %[[#zero]] : !s32i, !s32i, !s32i, !s32i) : !cir.vector + // LLVM: %[[#X1:]] = load i32, ptr %{{[0-9]+}}, align 4 + // LLVM-NEXT: %[[#X2:]] = load i32, ptr %{{[0-9]+}}, align 4 + // LLVM-NEXT: %[[#SUM:]] = add i32 %[[#X2]], 1 + // LLVM-NEXT: %[[#VEC1:]] = insertelement <4 x i32> undef, i32 %[[#X1]], i64 0 + // LLVM-NEXT: %[[#VEC2:]] = insertelement <4 x i32> %[[#VEC1]], i32 %[[#SUM]], i64 1 + // LLVM-NEXT: %[[#VEC3:]] = insertelement <4 x i32> %[[#VEC2]], i32 0, i64 2 + // LLVM-NEXT: %[[#VEC4:]] = insertelement <4 x i32> %[[#VEC3]], i32 0, i64 3 + // LLVM-NEXT: store <4 x i32> %[[#VEC4]], ptr %{{[0-9]+}}, align 16 + + + // Scalar to vector conversion, a.k.a. vector splat. Only valid as an + // operand of a binary operator, not as a regular conversion. + bb = a + 7; + // CIR: %[[#seven:]] = cir.const #cir.int<7> : !s32i + // CIR: %{{[0-9]+}} = cir.vec.splat %[[#seven]] : !s32i, !cir.vector + // LLVM: %[[#A:]] = load <4 x i32>, ptr %{{[0-9]+}}, align 16 + // LLVM-NEXT: %[[#BB:]] = add <4 x i32> %[[#A]], + // LLVM-NEXT: store <4 x i32> %[[#BB]], ptr %{{[0-9]+}}, align 16 + + // Vector to vector conversion + vd2 bbb = { }; + bb = (vi4)bbb; + // CIR: %{{[0-9]+}} = cir.cast(bitcast, %{{[0-9]+}} : !cir.vector), !cir.vector + // LLVM: %{{[0-9]+}} = bitcast <2 x double> %{{[0-9]+}} to <4 x i32> + + // Extract element + int c = a[x]; + // CIR: %{{[0-9]+}} = cir.vec.extract %{{[0-9]+}}[%{{[0-9]+}} : !s32i] : !cir.vector + // LLVM: %[[#A:]] = load <4 x i32>, ptr %{{[0-9]+}}, align 16 + // LLVM-NEXT: %[[#X:]] = load i32, ptr %{{[0-9]+}}, align 4 + // LLVM-NEXT: %[[#EXT:]] = extractelement <4 x i32> %[[#A]], i32 %[[#X]] + // LLVM-NEXT: store i32 %[[#EXT]], ptr %{{[0-9]+}}, align 4 + + // Insert element + a[x] = x; + // CIR: %[[#LOADEDVI:]] = cir.load %[[#STORAGEVI:]] : !cir.ptr>, !cir.vector + // CIR: %[[#UPDATEDVI:]] = cir.vec.insert %{{[0-9]+}}, %[[#LOADEDVI]][%{{[0-9]+}} : !s32i] : !cir.vector + // CIR: cir.store %[[#UPDATEDVI]], %[[#STORAGEVI]] : !cir.vector, !cir.ptr> + // LLVM: %[[#X1:]] = load i32, ptr %{{[0-9]+}}, align 4 + // LLVM-NEXT: %[[#X2:]] = load i32, ptr %{{[0-9]+}}, align 4 + // LLVM-NEXT: %[[#A:]] = load <4 x i32>, ptr %{{[0-9]+}}, align 16 + // LLVM-NEXT: %[[#INS:]] = insertelement <4 x i32> %[[#A]], i32 %[[#X1]], i32 %[[#X2]] + // LLVM-NEXT: store <4 x i32> %[[#INS]], ptr %{{[0-9]+}}, align 16 + + // Compound assignment + a[x] += a[0]; + // CIR: %[[#RHSCA:]] = cir.vec.extract %{{[0-9]+}}[%{{[0-9]+}} : !s32i] : !cir.vector + // CIR: %[[#LHSCA:]] = cir.vec.extract %{{[0-9]+}}[%{{[0-9]+}} : !s32i] : !cir.vector + // CIR: %[[#SUMCA:]] = cir.binop(add, %[[#LHSCA]], %[[#RHSCA]]) : !s32i + // CIR: cir.vec.insert %[[#SUMCA]], %{{[0-9]+}}[%{{[0-9]+}} : !s32i] : !cir.vector + // LLVM: %[[#A1:]] = load <4 x i32>, ptr %{{[0-9]+}}, align 16 + // LLVM-NEXT: %[[#RHSCA:]] = extractelement <4 x i32> %[[#A1]], i32 0 + // LLVM-NEXT: %[[#X:]] = load i32, ptr %{{[0-9]+}}, align 4 + // LLVM-NEXT: %[[#A2:]] = load <4 x i32>, ptr %{{[0-9]+}}, align 16 + // LLVM-NEXT: %[[#LHSCA:]] = extractelement <4 x i32> %[[#A2]], i32 %[[#X]] + // LLVM-NEXT: %[[#SUMCA:]] = add i32 %[[#LHSCA]], %[[#RHSCA]] + // LLVM-NEXT: %[[#A3:]] = load <4 x i32>, ptr %{{[0-9]+}}, align 16 + // LLVM-NEXT: %[[#RES:]] = insertelement <4 x i32> %[[#A3]], i32 %[[#SUMCA]], i32 %[[#X]] + // LLVM-NEXT: store <4 x i32> %[[#RES]], ptr %{{[0-9]+}}, align 16 + + // Binary arithmetic operations + vi4 d = a + b; + // CIR: %{{[0-9]+}} = cir.binop(add, %{{[0-9]+}}, %{{[0-9]+}}) : !cir.vector + // LLVM: %{{[0-9]+}} = add <4 x i32> %{{[0-9]+}}, %{{[0-9]+}} + vi4 e = a - b; + // CIR: %{{[0-9]+}} = cir.binop(sub, %{{[0-9]+}}, %{{[0-9]+}}) : !cir.vector + // LLVM: %{{[0-9]+}} = sub <4 x i32> %{{[0-9]+}}, %{{[0-9]+}} + vi4 f = a * b; + // CIR: %{{[0-9]+}} = cir.binop(mul, %{{[0-9]+}}, %{{[0-9]+}}) : !cir.vector + // LLVM: %{{[0-9]+}} = mul <4 x i32> %{{[0-9]+}}, %{{[0-9]+}} + vi4 g = a / b; + // CIR: %{{[0-9]+}} = cir.binop(div, %{{[0-9]+}}, %{{[0-9]+}}) : !cir.vector + // LLVM: %{{[0-9]+}} = sdiv <4 x i32> %{{[0-9]+}}, %{{[0-9]+}} + vi4 h = a % b; + // CIR: %{{[0-9]+}} = cir.binop(rem, %{{[0-9]+}}, %{{[0-9]+}}) : !cir.vector + // LLVM: %{{[0-9]+}} = srem <4 x i32> %{{[0-9]+}}, %{{[0-9]+}} + vi4 i = a & b; + // CIR: %{{[0-9]+}} = cir.binop(and, %{{[0-9]+}}, %{{[0-9]+}}) : !cir.vector + // LLVM: %{{[0-9]+}} = and <4 x i32> %{{[0-9]+}}, %{{[0-9]+}} + vi4 j = a | b; + // CIR: %{{[0-9]+}} = cir.binop(or, %{{[0-9]+}}, %{{[0-9]+}}) : !cir.vector + // LLVM: %{{[0-9]+}} = or <4 x i32> %{{[0-9]+}}, %{{[0-9]+}} + vi4 k = a ^ b; + // CIR: %{{[0-9]+}} = cir.binop(xor, %{{[0-9]+}}, %{{[0-9]+}}) : !cir.vector + // LLVM: %{{[0-9]+}} = xor <4 x i32> %{{[0-9]+}}, %{{[0-9]+}} + + // Unary arithmetic operations + vi4 l = +a; + // CIR: %{{[0-9]+}} = cir.unary(plus, %{{[0-9]+}}) : !cir.vector, !cir.vector + // LLVM: %[[#VAL:]] = load <4 x i32>, ptr %{{[0-9]+}}, align 16 + // LLVM-NEXT: store <4 x i32> %[[#VAL]], ptr %{{[0-9]+}}, align 16 + vi4 m = -a; + // CIR: %{{[0-9]+}} = cir.unary(minus, %{{[0-9]+}}) : !cir.vector, !cir.vector + // LLVM: %[[#VAL:]] = load <4 x i32>, ptr %{{[0-9]+}}, align 16 + // LLVM-NEXT: %[[#RES:]] = sub <4 x i32> zeroinitializer, %[[#VAL]] + // LLVM-NEXT: store <4 x i32> %[[#RES]], ptr %{{[0-9]+}}, align 16 + vi4 n = ~a; + // CIR: %{{[0-9]+}} = cir.unary(not, %{{[0-9]+}}) : !cir.vector, !cir.vector + // LLVM: %[[#VAL:]] = load <4 x i32>, ptr %{{[0-9]+}}, align 16 + // LLVM-NEXT: %[[#RES:]] = xor <4 x i32> , %[[#VAL]] + // LLVM-NEXT: store <4 x i32> %[[#RES]], ptr %{{[0-9]+}}, align 16 + + // TODO: Ternary conditional operator + + // Comparisons + vi4 o = a == b; + // CIR: %{{[0-9]+}} = cir.vec.cmp(eq, %{{[0-9]+}}, %{{[0-9]+}}) : !cir.vector, !cir.vector + // LLVM: %[[#RES:]] = icmp eq <4 x i32> %{{[0-9]+}}, %{{[0-9]+}} + // LLVM-NEXT: %[[#EXT:]] = sext <4 x i1> %[[#RES]] to <4 x i32> + vi4 p = a != b; + // CIR: %{{[0-9]+}} = cir.vec.cmp(ne, %{{[0-9]+}}, %{{[0-9]+}}) : !cir.vector, !cir.vector + // LLVM: %[[#RES:]] = icmp ne <4 x i32> %{{[0-9]+}}, %{{[0-9]+}} + // LLVM-NEXT: %[[#EXT:]] = sext <4 x i1> %[[#RES]] to <4 x i32> + vi4 q = a < b; + // CIR: %{{[0-9]+}} = cir.vec.cmp(lt, %{{[0-9]+}}, %{{[0-9]+}}) : !cir.vector, !cir.vector + // LLVM: %[[#RES:]] = icmp slt <4 x i32> %{{[0-9]+}}, %{{[0-9]+}} + // LLVM-NEXT: %[[#EXT:]] = sext <4 x i1> %[[#RES]] to <4 x i32> + vi4 r = a > b; + // CIR: %{{[0-9]+}} = cir.vec.cmp(gt, %{{[0-9]+}}, %{{[0-9]+}}) : !cir.vector, !cir.vector + // LLVM: %[[#RES:]] = icmp sgt <4 x i32> %{{[0-9]+}}, %{{[0-9]+}} + // LLVM-NEXT: %[[#EXT:]] = sext <4 x i1> %[[#RES]] to <4 x i32> + vi4 s = a <= b; + // CIR: %{{[0-9]+}} = cir.vec.cmp(le, %{{[0-9]+}}, %{{[0-9]+}}) : !cir.vector, !cir.vector + // LLVM: %[[#RES:]] = icmp sle <4 x i32> %{{[0-9]+}}, %{{[0-9]+}} + // LLVM-NEXT: %[[#EXT:]] = sext <4 x i1> %[[#RES]] to <4 x i32> + vi4 t = a >= b; + // CIR: %{{[0-9]+}} = cir.vec.cmp(ge, %{{[0-9]+}}, %{{[0-9]+}}) : !cir.vector, !cir.vector + // LLVM: %[[#RES:]] = icmp sge <4 x i32> %{{[0-9]+}}, %{{[0-9]+}} + // LLVM-NEXT: %[[#EXT:]] = sext <4 x i1> %[[#RES]] to <4 x i32> + + // __builtin_shufflevector + vi4 u = __builtin_shufflevector(a, b, 7, 5, 3, 1); + // CIR: %{{[0-9]+}} = cir.vec.shuffle(%{{[0-9]+}}, %{{[0-9]+}} : !cir.vector) [#cir.int<7> : !s64i, #cir.int<5> : !s64i, #cir.int<3> : !s64i, #cir.int<1> : !s64i] : !cir.vector + + // LLVM: %[[#A:]] = load <4 x i32>, ptr %{{[0-9]+}}, align 16 + // LLVM-NEXT: %[[#B:]] = load <4 x i32>, ptr %{{[0-9]+}}, align 16 + // LLVM-NEXT: %[[#SHFL:]] = shufflevector <4 x i32> %[[#A]], <4 x i32> %[[#B]], <4 x i32> + // LLVM-NEXT: store <4 x i32> %[[#SHFL]], ptr %{{[0-9]+}}, align 16 + + vi4 v = __builtin_shufflevector(a, b); + // CIR: %{{[0-9]+}} = cir.vec.shuffle.dynamic %{{[0-9]+}} : !cir.vector, %{{[0-9]+}} : !cir.vector + + // LLVM: %[[#A:]] = load <4 x i32>, ptr %{{[0-9]+}}, align 16 + // LLVM-NEXT: %[[#B:]] = load <4 x i32>, ptr %{{[0-9]+}}, align 16 + // LLVM-NEXT: %[[#IDXMOD:]] = and <4 x i32> %[[#B]], + // LLVM-NEXT: %[[#IDX0:]] = extractelement <4 x i32> %[[#IDXMOD]], i64 0 + // LLVM-NEXT: %[[#EXT1:]] = extractelement <4 x i32> %[[#A]], i32 %[[#IDX0]] + // LLVM-NEXT: %[[#INS1:]] = insertelement <4 x i32> undef, i32 %[[#EXT1]], i64 0 + // LLVM-NEXT: %[[#IDX1:]] = extractelement <4 x i32> %[[#IDXMOD]], i64 1 + // LLVM-NEXT: %[[#EXT2:]] = extractelement <4 x i32> %[[#A]], i32 %[[#IDX1]] + // LLVM-NEXT: %[[#INS2:]] = insertelement <4 x i32> %[[#INS1]], i32 %[[#EXT2]], i64 1 + // LLVM-NEXT: %[[#IDX2:]] = extractelement <4 x i32> %[[#IDXMOD]], i64 2 + // LLVM-NEXT: %[[#EXT3:]] = extractelement <4 x i32> %[[#A]], i32 %[[#IDX2]] + // LLVM-NEXT: %[[#INS3:]] = insertelement <4 x i32> %[[#INS2]], i32 %[[#EXT3]], i64 2 + // LLVM-NEXT: %[[#IDX3:]] = extractelement <4 x i32> %[[#IDXMOD]], i64 3 + // LLVM-NEXT: %[[#EXT4:]] = extractelement <4 x i32> %[[#A]], i32 %[[#IDX3]] + // LLVM-NEXT: %[[#INS4:]] = insertelement <4 x i32> %[[#INS3]], i32 %[[#EXT4]], i64 3 + // LLVM-NEXT: store <4 x i32> %[[#INS4]], ptr %{{[0-9]+}}, align 16 +} + +// CIR: cir.func {{@.*vector_double_test.*}} +// LLVM: define void {{@.*vector_double_test.*}} +void vector_double_test(int x, double y) { + // Vector constant. Not yet implemented. Expected results will change from + // cir.vec.create to cir.const. + vd2 a = { 1.5, 2.5 }; + // CIR: %{{[0-9]+}} = cir.vec.create(%{{[0-9]+}}, %{{[0-9]+}} : !cir.double, !cir.double) : !cir.vector + + // LLVM: store <2 x double> , ptr %{{[0-9]+}}, align 16 + + // Non-const vector initialization. + vd2 b = { y, y + 1.0 }; + // CIR: %{{[0-9]+}} = cir.vec.create(%{{[0-9]+}}, %{{[0-9]+}} : !cir.double, !cir.double) : !cir.vector + + // LLVM: %[[#Y1:]] = load double, ptr %{{[0-9]+}}, align 8 + // LLVM-NEXT: %[[#Y2:]] = load double, ptr %{{[0-9]+}}, align 8 + // LLVM-NEXT: %[[#SUM:]] = fadd double %[[#Y2]], 1.000000e+00 + // LLVM-NEXT: %[[#VEC1:]] = insertelement <2 x double> undef, double %[[#Y1]], i64 0 + // LLVM-NEXT: %[[#VEC2:]] = insertelement <2 x double> %[[#VEC1]], double %[[#SUM]], i64 1 + // LLVM-NEXT: store <2 x double> %[[#VEC2]], ptr %{{[0-9]+}}, align 16 + + // Incomplete vector initialization + vd2 bb = { y }; + // CIR: [[#dzero:]] = cir.const #cir.fp<0.000000e+00> : !cir.double + // CIR: %{{[0-9]+}} = cir.vec.create(%{{[0-9]+}}, %[[#dzero]] : !cir.double, !cir.double) : !cir.vector + + // LLVM: %[[#Y1:]] = load double, ptr %{{[0-9]+}}, align 8 + // LLVM-NEXT: %[[#VEC1:]] = insertelement <2 x double> undef, double %[[#Y1]], i64 0 + // LLVM-NEXT: %[[#VEC2:]] = insertelement <2 x double> %[[#VEC1]], double 0.000000e+00, i64 1 + // LLVM-NEXT: store <2 x double> %[[#VEC2]], ptr %{{[0-9]+}}, align 16 + + // Scalar to vector conversion, a.k.a. vector splat. Only valid as an + // operand of a binary operator, not as a regular conversion. + bb = a + 2.5; + // CIR: %[[#twohalf:]] = cir.const #cir.fp<2.500000e+00> : !cir.double + // CIR: %{{[0-9]+}} = cir.vec.splat %[[#twohalf]] : !cir.double, !cir.vector + + // LLVM: %[[#A:]] = load <2 x double>, ptr %{{[0-9]+}}, align 16 + // LLVM-NEXT: %[[#BB:]] = fadd <2 x double> %[[#A]], + // LLVM-NEXT: store <2 x double> %[[#BB]], ptr %{{[0-9]+}}, align 16 + + // Extract element + double c = a[x]; + // CIR: %{{[0-9]+}} = cir.vec.extract %{{[0-9]+}}[%{{[0-9]+}} : !s32i] : !cir.vector + // LLVM: %{{[0-9]+}} = extractelement <2 x double> %{{[0-9]+}}, i32 %{{[0-9]+}} + + // Insert element + a[x] = y; + // CIR: %[[#LOADEDVF:]] = cir.load %[[#STORAGEVF:]] : !cir.ptr>, !cir.vector + // CIR: %[[#UPDATEDVF:]] = cir.vec.insert %{{[0-9]+}}, %[[#LOADEDVF]][%{{[0-9]+}} : !s32i] : !cir.vector + // CIR: cir.store %[[#UPDATEDVF]], %[[#STORAGEVF]] : !cir.vector, !cir.ptr> + + // LLVM: %[[#Y:]] = load double, ptr %{{[0-9]+}}, align 8 + // LLVM-NEXT: %[[#X:]] = load i32, ptr %{{[0-9]+}}, align 4 + // LLVM-NEXT: %[[#A:]] = load <2 x double>, ptr %{{[0-9]+}}, align 16 + // LLVM-NEXT: %[[#INS:]] = insertelement <2 x double> %[[#A]], double %[[#Y]], i32 %[[#X]] + // LLVM-NEXT: store <2 x double> %[[#INS]], ptr %{{[0-9]+}}, align 16 + + // Binary arithmetic operations + vd2 d = a + b; + // CIR: %{{[0-9]+}} = cir.binop(add, %{{[0-9]+}}, %{{[0-9]+}}) : !cir.vector + // LLVM: %{{[0-9]+}} = fadd <2 x double> %{{[0-9]+}}, %{{[0-9]+}} + vd2 e = a - b; + // CIR: %{{[0-9]+}} = cir.binop(sub, %{{[0-9]+}}, %{{[0-9]+}}) : !cir.vector + // LLVM: %{{[0-9]+}} = fsub <2 x double> %{{[0-9]+}}, %{{[0-9]+}} + vd2 f = a * b; + // CIR: %{{[0-9]+}} = cir.binop(mul, %{{[0-9]+}}, %{{[0-9]+}}) : !cir.vector + // LLVM: %{{[0-9]+}} = fmul <2 x double> %{{[0-9]+}}, %{{[0-9]+}} + vd2 g = a / b; + // CIR: %{{[0-9]+}} = cir.binop(div, %{{[0-9]+}}, %{{[0-9]+}}) : !cir.vector + // LLVM: %{{[0-9]+}} = fdiv <2 x double> %{{[0-9]+}}, %{{[0-9]+}} + + // Unary arithmetic operations + vd2 l = +a; + // CIR: %{{[0-9]+}} = cir.unary(plus, %{{[0-9]+}}) : !cir.vector, !cir.vector + // LLVM: %[[#VAL:]] = load <2 x double>, ptr %{{[0-9]+}}, align 16 + // LLVM-NEXT: store <2 x double> %[[#VAL]], ptr %{{[0-9]+}}, align 16 + vd2 m = -a; + // CIR: %{{[0-9]+}} = cir.unary(minus, %{{[0-9]+}}) : !cir.vector, !cir.vector + // LLVM: %[[#VAL:]] = load <2 x double>, ptr %{{[0-9]+}}, align 16 + // LLVM-NEXT: %[[#RES:]] = fneg <2 x double> %[[#VAL]] + // LLVM-NEXT: store <2 x double> %[[#RES]], ptr %{{[0-9]+}}, align 16 + + // Comparisons + vl2 o = a == b; + // CIR: %{{[0-9]+}} = cir.vec.cmp(eq, %{{[0-9]+}}, %{{[0-9]+}}) : !cir.vector, !cir.vector + // LLVM: %[[#RES:]] = fcmp oeq <2 x double> %{{[0-9]+}}, %{{[0-9]+}} + // LLVM-NEXT: sext <2 x i1> %[[#RES:]] to <2 x i64> + vl2 p = a != b; + // CIR: %{{[0-9]+}} = cir.vec.cmp(ne, %{{[0-9]+}}, %{{[0-9]+}}) : !cir.vector, !cir.vector + // LLVM: %[[#RES:]] = fcmp une <2 x double> %{{[0-9]+}}, %{{[0-9]+}} + // LLVM-NEXT: sext <2 x i1> %[[#RES:]] to <2 x i64> + vl2 q = a < b; + // CIR: %{{[0-9]+}} = cir.vec.cmp(lt, %{{[0-9]+}}, %{{[0-9]+}}) : !cir.vector, !cir.vector + // LLVM: %[[#RES:]] = fcmp olt <2 x double> %{{[0-9]+}}, %{{[0-9]+}} + // LLVM-NEXT: sext <2 x i1> %[[#RES:]] to <2 x i64> + vl2 r = a > b; + // CIR: %{{[0-9]+}} = cir.vec.cmp(gt, %{{[0-9]+}}, %{{[0-9]+}}) : !cir.vector, !cir.vector + // LLVM: %[[#RES:]] = fcmp ogt <2 x double> %{{[0-9]+}}, %{{[0-9]+}} + // LLVM-NEXT: sext <2 x i1> %[[#RES:]] to <2 x i64> + vl2 s = a <= b; + // CIR: %{{[0-9]+}} = cir.vec.cmp(le, %{{[0-9]+}}, %{{[0-9]+}}) : !cir.vector, !cir.vector + // LLVM: %[[#RES:]] = fcmp ole <2 x double> %{{[0-9]+}}, %{{[0-9]+}} + // LLVM-NEXT: sext <2 x i1> %[[#RES:]] to <2 x i64> + vl2 t = a >= b; + // CIR: %{{[0-9]+}} = cir.vec.cmp(ge, %{{[0-9]+}}, %{{[0-9]+}}) : !cir.vector, !cir.vector + // LLVM: %[[#RES:]] = fcmp oge <2 x double> %{{[0-9]+}}, %{{[0-9]+}} + // LLVM-NEXT: sext <2 x i1> %[[#RES:]] to <2 x i64> + + // __builtin_convertvector + vus2 w = __builtin_convertvector(a, vus2); + // CIR: %{{[0-9]+}} = cir.cast(float_to_int, %{{[0-9]+}} : !cir.vector), !cir.vector + // LLVM: %{{[0-9]+}} = fptoui <2 x double> %{{[0-9]+}} to <2 x i16> +} + +// CIR: cir.func {{@.*test_load.*}} +// LLVM: define void {{@.*test_load.*}} +void test_load() { + vi4 a = { 1, 2, 3, 4 }; + + vi2 b; + + b = a.wz; + // CIR: %[[#LOAD1:]] = cir.load %{{[0-9]+}} : !cir.ptr>, !cir.vector + // CIR-NEXT: %[[#SHUFFLE1:]] = cir.vec.shuffle(%[[#LOAD1]], %[[#LOAD1]] : !cir.vector) [#cir.int<3> : !s32i, #cir.int<2> : !s32i] : !cir.vector + // CIR-NEXT: cir.store %[[#SHUFFLE1]], %{{[0-9]+}} : !cir.vector, !cir.ptr> + + // LLVM: %[[#LOAD1:]] = load <4 x i32>, ptr %{{[0-9]+}}, align 16 + // LLVM-NEXT: %[[#SHUFFLE1:]] = shufflevector <4 x i32> %[[#LOAD1]], <4 x i32> %[[#LOAD1]], <2 x i32> + // LLVM-NEXT: store <2 x i32> %[[#SHUFFLE1]], ptr %{{[0-9]+}}, align 8 + + int one_elem_load = a.s2; + // CIR-NEXT: %[[#LOAD8:]] = cir.load %{{[0-9]+}} : !cir.ptr>, !cir.vector + // CIR-NEXT: %[[#EXTRACT_INDEX:]] = cir.const #cir.int<2> : !s64i + // CIR-NEXT: %[[#EXTRACT1:]] = cir.vec.extract %[[#LOAD8]][%[[#EXTRACT_INDEX]] : !s64i] : !cir.vector + // CIR-NEXT: cir.store %[[#EXTRACT1]], %{{[0-9]+}} : !s32i, !cir.ptr + + // LLVM-NEXT: %[[#LOAD8:]] = load <4 x i32>, ptr %{{[0-9]+}}, align 16 + // LLVM-NEXT: %[[#EXTRACT1:]] = extractelement <4 x i32> %[[#LOAD8]], i64 2 + // LLVM-NEXT: store i32 %[[#EXTRACT1]], ptr %{{[0-9]+}}, align 4 + +} + +// CIR: cir.func {{@.*test_store.*}} +// LLVM: define void {{@.*test_store.*}} +void test_store() { + vi4 a; + // CIR: %[[#PVECA:]] = cir.alloca !cir.vector + // LLVM: %[[#PVECA:]] = alloca <4 x i32> + + vi2 b = {1, 2}; + // CIR-NEXT: %[[#PVECB:]] = cir.alloca !cir.vector + // LLVM-NEXT: %[[#PVECB:]] = alloca <2 x i32> + + a.xy = b; + // CIR: %[[#LOAD4RHS:]] = cir.load %{{[0-9]+}} : !cir.ptr>, !cir.vector + // CIR-NEXT: %[[#LOAD5LHS:]] = cir.load %{{[0-9]+}} : !cir.ptr>, !cir.vector + // CIR-NEXT: %[[#SHUFFLE5:]] = cir.vec.shuffle(%[[#LOAD4RHS]], %[[#LOAD4RHS]] : !cir.vector) [#cir.int<0> : !s32i, #cir.int<1> : !s32i, #cir.int<-1> : !s32i, #cir.int<-1> : !s32i] : !cir.vector + // CIR-NEXT: %[[#SHUFFLE6:]] = cir.vec.shuffle(%[[#LOAD5LHS]], %[[#SHUFFLE5]] : !cir.vector) [#cir.int<4> : !s32i, #cir.int<5> : !s32i, #cir.int<2> : !s32i, #cir.int<3> : !s32i] : !cir.vector + // CIR-NEXT: cir.store %[[#SHUFFLE6]], %{{[0-9]+}} : !cir.vector, !cir.ptr> + + // LLVM: %[[#LOAD4RHS:]] = load <2 x i32>, ptr %{{[0-9]+}}, align 8 + // LLVM-NEXT: %[[#LOAD5LHS:]] = load <4 x i32>, ptr %{{[0-9]+}}, align 16 + // LLVM-NEXT: %[[#SHUFFLE5:]] = shufflevector <2 x i32> %[[#LOAD4RHS]], <2 x i32> %[[#LOAD4RHS]], <4 x i32> + // LLVM-NEXT: %[[#SHUFFLE6:]] = shufflevector <4 x i32> %[[#LOAD5LHS]], <4 x i32> %[[#SHUFFLE5]], <4 x i32> + // LLVM-NEXT: store <4 x i32> %[[#SHUFFLE6]], ptr %{{[0-9]+}}, align 16 + + // load single element + a.s0 = 1; + // CIR-NEXT: cir.const #cir.int<1> + // CIR-NEXT: %[[#LOAD7:]] = cir.load %{{[0-9]+}} : !cir.ptr>, !cir.vector + // CIR-NEXT: %[[#INSERT_INDEX:]] = cir.const #cir.int<0> : !s64i + // CIR-NEXT: %[[#INSERT1:]] = cir.vec.insert %{{[0-9]+}}, %[[#LOAD7]][%[[#INSERT_INDEX]] : !s64i] : !cir.vector + // CIR-NEXT: cir.store %[[#INSERT1]], %{{[0-9]+}} : !cir.vector, !cir.ptr> + + // LLVM-NEXT: %[[#LOAD7:]] = load <4 x i32>, ptr %{{[0-9]+}}, align 16 + // LLVM-NEXT: %[[#INSERT1:]] = insertelement <4 x i32> %[[#LOAD7]], i32 1, i64 0 + // LLVM-NEXT: store <4 x i32> %[[#INSERT1]], ptr %{{[0-9]+}}, align 16 + + // extend length from 2 to 4, then merge two vectors + a.lo = b; + // CIR: %[[#VECB:]] = cir.load %[[#PVECB]] : !cir.ptr>, !cir.vector + // CIR-NEXT: %[[#VECA:]] = cir.load %[[#PVECA]] : !cir.ptr>, !cir.vector + // CIR-NEXT: %[[#EXTVECB:]] = cir.vec.shuffle(%[[#VECB]], %[[#VECB]] : !cir.vector) [#cir.int<0> : !s32i, #cir.int<1> : !s32i, #cir.int<-1> : !s32i, #cir.int<-1> : !s32i] : !cir.vector + // CIR-NEXT: %[[#RESULT:]] = cir.vec.shuffle(%[[#VECA]], %[[#EXTVECB]] : !cir.vector) [#cir.int<4> : !s32i, #cir.int<5> : !s32i, #cir.int<2> : !s32i, #cir.int<3> : !s32i] : !cir.vector + // CIR-NEXT: cir.store %[[#RESULT]], %[[#PVECA]] : !cir.vector, !cir.ptr> + + // LLVM: %[[#VECB:]] = load <2 x i32>, ptr %[[#PVECB]], align 8 + // LLVM-NEXT: %[[#VECA:]] = load <4 x i32>, ptr %[[#PVECA]], align 16 + // LLVM-NEXT: %[[#EXTVECB:]] = shufflevector <2 x i32> %[[#VECB]], <2 x i32> %[[#VECB]], <4 x i32> + // LLVM-NEXT: %[[#RESULT:]] = shufflevector <4 x i32> %[[#VECA]], <4 x i32> %[[#EXTVECB]], <4 x i32> + // LLVM-NEXT: store <4 x i32> %[[#RESULT]], ptr %[[#PVECA]], align 16 + +} + +// CIR: cir.func {{@.*test_build_lvalue.*}} +// LLVM: define void {{@.*test_build_lvalue.*}} +void test_build_lvalue() { + // special cases only + + vi4 *pv, v; + + // CIR-NEXT: %[[#ALLOCAPV:]] = cir.alloca !cir.ptr>, !cir.ptr>>, ["pv"] {alignment = 8 : i64} + // CIR-NEXT: %[[#ALLOCAV:]] = cir.alloca !cir.vector, !cir.ptr>, ["v"] {alignment = 16 : i64} + // CIR-NEXT: %[[#ALLOCAS:]] = cir.alloca !s32i, !cir.ptr, ["s", init] {alignment = 4 : i64} + // CIR-NEXT: %[[#ALLOCATMP:]] = cir.alloca !cir.vector, !cir.ptr>, ["tmp"] {alignment = 16 : i64} + // CIR-NEXT: %[[#ALLOCAR:]] = cir.alloca !s32i, !cir.ptr, ["r", init] {alignment = 4 : i64} + + // LLVM-NEXT: %[[#ALLOCAPV:]] = alloca ptr, i64 1, align 8 + // LLVM-NEXT: %[[#ALLOCAV:]] = alloca <4 x i32>, i64 1, align 16 + // LLVM-NEXT: %[[#ALLOCAS:]] = alloca i32, i64 1, align 4 + // LLVM-NEXT: %[[#ALLOCATMP:]] = alloca <4 x i32>, i64 1, align 16 + // LLVM-NEXT: %[[#ALLOCAR:]] = alloca i32, i64 1, align 4 + + pv->x = 99; + // CIR-NEXT: %[[#VAL:]] = cir.const #cir.int<99> : !s32i + // CIR-NEXT: %[[#PV:]] = cir.load %[[#ALLOCAPV]] : !cir.ptr>>, !cir.ptr> + // CIR-NEXT: %[[#V:]] = cir.load %[[#PV]] : !cir.ptr>, !cir.vector + // CIR-NEXT: %[[#IDX:]] = cir.const #cir.int<0> : !s64i + // CIR-NEXT: %[[#RESULT:]] = cir.vec.insert %[[#VAL]], %[[#V]][%[[#IDX]] : !s64i] : !cir.vector + // CIR-NEXT: cir.store %[[#RESULT]], %[[#PV]] : !cir.vector, !cir.ptr> + + // LLVM-NEXT: %[[#PV:]] = load ptr, ptr %[[#ALLOCAPV]], align 8 + // LLVM-NEXT: %[[#V:]] = load <4 x i32>, ptr %[[#PV]], align 16 + // LLVM-NEXT: %[[#RESULT:]] = insertelement <4 x i32> %[[#V]], i32 99, i64 0 + // LLVM-NEXT: store <4 x i32> %[[#RESULT]], ptr %[[#PV]], align 16 + + int s = (v+v).x; + + // CIR-NEXT: %[[#LOAD1:]] = cir.load %[[#ALLOCAV]] : !cir.ptr>, !cir.vector + // CIR-NEXT: %[[#LOAD2:]] = cir.load %[[#ALLOCAV]] : !cir.ptr>, !cir.vector + // CIR-NEXT: %[[#SUM:]] = cir.binop(add, %[[#LOAD1]], %[[#LOAD2]]) : !cir.vector + // CIR-NEXT: cir.store %[[#SUM]], %[[#ALLOCATMP]] : !cir.vector, !cir.ptr> + // CIR-NEXT: %[[#TMP:]] = cir.load %[[#ALLOCATMP]] : !cir.ptr>, !cir.vector + // CIR-NEXT: %[[#IDX:]] = cir.const #cir.int<0> : !s64i + // CIR-NEXT: %[[#RESULT:]] = cir.vec.extract %[[#TMP]][%[[#IDX]] : !s64i] : !cir.vector + // CIR-NEXT: cir.store %[[#RESULT]], %[[#ALLOCAS]] : !s32i, !cir.ptr + + // LLVM-NEXT: %[[#LOAD1:]] = load <4 x i32>, ptr %{{[0-9]+}}, align 16 + // LLVM-NEXT: %[[#LOAD2:]] = load <4 x i32>, ptr %{{[0-9]+}}, align 16 + // LLVM-NEXT: %[[#SUM:]] = add <4 x i32> %[[#LOAD1]], %[[#LOAD2]] + // LLVM-NEXT: store <4 x i32> %[[#SUM]], ptr %[[#ALLOCATMP]], align 16 + // LLVM-NEXT: %[[#TMP:]] = load <4 x i32>, ptr %[[#ALLOCATMP]], align 16 + // LLVM-NEXT: %[[#RESULT:]] = extractelement <4 x i32> %[[#TMP]], i64 0 + // LLVM-NEXT: store i32 %[[#RESULT]], ptr %[[#ALLOCAS]], align 4 + + int r = v.xy.x; + // CIR-NEXT: %[[#V:]] = cir.load %[[#ALLOCAV]] : !cir.ptr>, !cir.vector + // CIR-NEXT: %[[#IDX:]] = cir.const #cir.int<0> : !s64i + // CIR-NEXT: %[[#RESULT:]] = cir.vec.extract %[[#V]][%[[#IDX]] : !s64i] : !cir.vector + // CIR-NEXT: cir.store %[[#RESULT]], %[[#ALLOCAR]] : !s32i, !cir.ptr + + // LLVM-NEXT: %[[#V:]] = load <4 x i32>, ptr %[[#ALLOCAV]], align 16 + // LLVM-NEXT: %[[#RESULT:]] = extractelement <4 x i32> %[[#V]], i64 0 + // LLVM-NEXT: store i32 %[[#RESULT]], ptr %[[#ALLOCAR]], align 4 + +}