@@ -10814,11 +10814,11 @@ static SDValue lowerShuffleToEXPAND(const SDLoc &DL, MVT VT,
10814
10814
return DAG.getNode(X86ISD::EXPAND, DL, VT, ExpandedVector, ZeroVector, VMask);
10815
10815
}
10816
10816
10817
- static bool matchVectorShuffleWithUNPCK (MVT VT, SDValue &V1, SDValue &V2,
10818
- unsigned &UnpackOpcode, bool IsUnary,
10819
- ArrayRef<int> TargetMask,
10820
- const SDLoc &DL, SelectionDAG &DAG,
10821
- const X86Subtarget &Subtarget) {
10817
+ static bool matchShuffleWithUNPCK (MVT VT, SDValue &V1, SDValue &V2,
10818
+ unsigned &UnpackOpcode, bool IsUnary,
10819
+ ArrayRef<int> TargetMask, const SDLoc &DL ,
10820
+ SelectionDAG &DAG,
10821
+ const X86Subtarget &Subtarget) {
10822
10822
int NumElts = VT.getVectorNumElements();
10823
10823
10824
10824
bool Undef1 = true, Undef2 = true, Zero1 = true, Zero2 = true;
@@ -10926,8 +10926,8 @@ static SDValue lowerShuffleWithUNPCK(const SDLoc &DL, MVT VT,
10926
10926
return SDValue();
10927
10927
}
10928
10928
10929
- static bool matchVectorShuffleAsVPMOV (ArrayRef<int> Mask, bool SwappedOps,
10930
- int Delta) {
10929
+ static bool matchShuffleAsVPMOV (ArrayRef<int> Mask, bool SwappedOps,
10930
+ int Delta) {
10931
10931
int Size = (int)Mask.size();
10932
10932
int Split = Size / Delta;
10933
10933
int TruncatedVectorStart = SwappedOps ? Size : 0;
@@ -11012,20 +11012,19 @@ static SDValue lowerShuffleWithVPMOV(const SDLoc &DL, ArrayRef<int> Mask,
11012
11012
11013
11013
// The first half/quarter of the mask should refer to every second/fourth
11014
11014
// element of the vector truncated and bitcasted.
11015
- if (!matchVectorShuffleAsVPMOV (Mask, SwappedOps, 2) &&
11016
- !matchVectorShuffleAsVPMOV (Mask, SwappedOps, 4))
11015
+ if (!matchShuffleAsVPMOV (Mask, SwappedOps, 2) &&
11016
+ !matchShuffleAsVPMOV (Mask, SwappedOps, 4))
11017
11017
return SDValue();
11018
11018
11019
11019
return DAG.getNode(X86ISD::VTRUNC, DL, VT, Src);
11020
11020
}
11021
11021
11022
11022
// X86 has dedicated pack instructions that can handle specific truncation
11023
11023
// operations: PACKSS and PACKUS.
11024
- static bool matchVectorShuffleWithPACK(MVT VT, MVT &SrcVT, SDValue &V1,
11025
- SDValue &V2, unsigned &PackOpcode,
11026
- ArrayRef<int> TargetMask,
11027
- SelectionDAG &DAG,
11028
- const X86Subtarget &Subtarget) {
11024
+ static bool matchShuffleWithPACK(MVT VT, MVT &SrcVT, SDValue &V1, SDValue &V2,
11025
+ unsigned &PackOpcode, ArrayRef<int> TargetMask,
11026
+ SelectionDAG &DAG,
11027
+ const X86Subtarget &Subtarget) {
11029
11028
unsigned NumElts = VT.getVectorNumElements();
11030
11029
unsigned BitSize = VT.getScalarSizeInBits();
11031
11030
MVT PackSVT = MVT::getIntegerVT(BitSize * 2);
@@ -11078,8 +11077,8 @@ static SDValue lowerShuffleWithPACK(const SDLoc &DL, MVT VT, ArrayRef<int> Mask,
11078
11077
const X86Subtarget &Subtarget) {
11079
11078
MVT PackVT;
11080
11079
unsigned PackOpcode;
11081
- if (matchVectorShuffleWithPACK (VT, PackVT, V1, V2, PackOpcode, Mask, DAG,
11082
- Subtarget))
11080
+ if (matchShuffleWithPACK (VT, PackVT, V1, V2, PackOpcode, Mask, DAG,
11081
+ Subtarget))
11083
11082
return DAG.getNode(PackOpcode, DL, VT, DAG.getBitcast(PackVT, V1),
11084
11083
DAG.getBitcast(PackVT, V2));
11085
11084
@@ -11170,10 +11169,10 @@ static SDValue getVectorMaskingNode(SDValue Op, SDValue Mask,
11170
11169
const X86Subtarget &Subtarget,
11171
11170
SelectionDAG &DAG);
11172
11171
11173
- static bool matchVectorShuffleAsBlend (SDValue V1, SDValue V2,
11174
- MutableArrayRef<int> Mask,
11175
- const APInt &Zeroable, bool &ForceV1Zero,
11176
- bool &ForceV2Zero, uint64_t &BlendMask) {
11172
+ static bool matchShuffleAsBlend (SDValue V1, SDValue V2,
11173
+ MutableArrayRef<int> Mask,
11174
+ const APInt &Zeroable, bool &ForceV1Zero,
11175
+ bool &ForceV2Zero, uint64_t &BlendMask) {
11177
11176
bool V1IsZeroOrUndef =
11178
11177
V1.isUndef() || ISD::isBuildVectorAllZeros(V1.getNode());
11179
11178
bool V2IsZeroOrUndef =
@@ -11236,8 +11235,8 @@ static SDValue lowerShuffleAsBlend(const SDLoc &DL, MVT VT, SDValue V1,
11236
11235
uint64_t BlendMask = 0;
11237
11236
bool ForceV1Zero = false, ForceV2Zero = false;
11238
11237
SmallVector<int, 64> Mask(Original.begin(), Original.end());
11239
- if (!matchVectorShuffleAsBlend (V1, V2, Mask, Zeroable, ForceV1Zero, ForceV2Zero,
11240
- BlendMask))
11238
+ if (!matchShuffleAsBlend (V1, V2, Mask, Zeroable, ForceV1Zero, ForceV2Zero,
11239
+ BlendMask))
11241
11240
return SDValue();
11242
11241
11243
11242
// Create a REAL zero vector - ISD::isBuildVectorAllZeros allows UNDEFs.
@@ -11807,9 +11806,11 @@ static SDValue lowerShuffleAsRotate(const SDLoc &DL, MVT VT, SDValue V1,
11807
11806
}
11808
11807
11809
11808
/// Try to lower a vector shuffle as a byte shift sequence.
11810
- static SDValue lowerVectorShuffleAsByteShiftMask(
11811
- const SDLoc &DL, MVT VT, SDValue V1, SDValue V2, ArrayRef<int> Mask,
11812
- const APInt &Zeroable, const X86Subtarget &Subtarget, SelectionDAG &DAG) {
11809
+ static SDValue lowerShuffleAsByteShiftMask(const SDLoc &DL, MVT VT, SDValue V1,
11810
+ SDValue V2, ArrayRef<int> Mask,
11811
+ const APInt &Zeroable,
11812
+ const X86Subtarget &Subtarget,
11813
+ SelectionDAG &DAG) {
11813
11814
assert(!isNoopShuffleMask(Mask) && "We shouldn't lower no-op shuffles!");
11814
11815
assert(VT.is128BitVector() && "Only 128-bit vectors supported");
11815
11816
@@ -14254,8 +14255,8 @@ static SDValue lowerV8I16Shuffle(const SDLoc &DL, ArrayRef<int> Mask,
14254
14255
return BitBlend;
14255
14256
14256
14257
// Try to use byte shift instructions to mask.
14257
- if (SDValue V = lowerVectorShuffleAsByteShiftMask(
14258
- DL, MVT::v8i16, V1, V2, Mask, Zeroable, Subtarget, DAG))
14258
+ if (SDValue V = lowerShuffleAsByteShiftMask(DL, MVT::v8i16, V1, V2, Mask,
14259
+ Zeroable, Subtarget, DAG))
14259
14260
return V;
14260
14261
14261
14262
// Try to lower by permuting the inputs into an unpack instruction.
@@ -14516,8 +14517,8 @@ static SDValue lowerV16I8Shuffle(const SDLoc &DL, ArrayRef<int> Mask,
14516
14517
return V;
14517
14518
14518
14519
// Try to use byte shift instructions to mask.
14519
- if (SDValue V = lowerVectorShuffleAsByteShiftMask(
14520
- DL, MVT::v16i8, V1, V2, Mask, Zeroable, Subtarget, DAG))
14520
+ if (SDValue V = lowerShuffleAsByteShiftMask(DL, MVT::v16i8, V1, V2, Mask,
14521
+ Zeroable, Subtarget, DAG))
14521
14522
return V;
14522
14523
14523
14524
// Check for SSSE3 which lets us lower all v16i8 shuffles much more directly
@@ -17243,8 +17244,8 @@ static bool canonicalizeShuffleMaskWithCommute(ArrayRef<int> Mask) {
17243
17244
/// above in helper routines. The canonicalization attempts to widen shuffles
17244
17245
/// to involve fewer lanes of wider elements, consolidate symmetric patterns
17245
17246
/// s.t. only one of the two inputs needs to be tested, etc.
17246
- static SDValue lowerVectorShuffle (SDValue Op, const X86Subtarget &Subtarget,
17247
- SelectionDAG &DAG) {
17247
+ static SDValue lowerVECTOR_SHUFFLE (SDValue Op, const X86Subtarget &Subtarget,
17248
+ SelectionDAG &DAG) {
17248
17249
ShuffleVectorSDNode *SVOp = cast<ShuffleVectorSDNode>(Op);
17249
17250
ArrayRef<int> OrigMask = SVOp->getMask();
17250
17251
SDValue V1 = Op.getOperand(0);
@@ -28457,7 +28458,7 @@ SDValue X86TargetLowering::LowerOperation(SDValue Op, SelectionDAG &DAG) const {
28457
28458
case ISD::BITREVERSE: return LowerBITREVERSE(Op, Subtarget, DAG);
28458
28459
case ISD::BUILD_VECTOR: return LowerBUILD_VECTOR(Op, DAG);
28459
28460
case ISD::CONCAT_VECTORS: return LowerCONCAT_VECTORS(Op, Subtarget, DAG);
28460
- case ISD::VECTOR_SHUFFLE: return lowerVectorShuffle (Op, Subtarget, DAG);
28461
+ case ISD::VECTOR_SHUFFLE: return lowerVECTOR_SHUFFLE (Op, Subtarget, DAG);
28461
28462
case ISD::VSELECT: return LowerVSELECT(Op, DAG);
28462
28463
case ISD::EXTRACT_VECTOR_ELT: return LowerEXTRACT_VECTOR_ELT(Op, DAG);
28463
28464
case ISD::INSERT_VECTOR_ELT: return LowerINSERT_VECTOR_ELT(Op, DAG);
@@ -33075,8 +33076,8 @@ static bool matchBinaryShuffle(MVT MaskVT, ArrayRef<int> Mask,
33075
33076
if (((MaskVT == MVT::v8i16 || MaskVT == MVT::v16i8) && Subtarget.hasSSE2()) ||
33076
33077
((MaskVT == MVT::v16i16 || MaskVT == MVT::v32i8) && Subtarget.hasInt256()) ||
33077
33078
((MaskVT == MVT::v32i16 || MaskVT == MVT::v64i8) && Subtarget.hasBWI())) {
33078
- if (matchVectorShuffleWithPACK (MaskVT, SrcVT, V1, V2, Shuffle, Mask, DAG,
33079
- Subtarget)) {
33079
+ if (matchShuffleWithPACK (MaskVT, SrcVT, V1, V2, Shuffle, Mask, DAG,
33080
+ Subtarget)) {
33080
33081
DstVT = MaskVT;
33081
33082
return true;
33082
33083
}
@@ -33088,8 +33089,8 @@ static bool matchBinaryShuffle(MVT MaskVT, ArrayRef<int> Mask,
33088
33089
(MaskVT.is256BitVector() && 32 <= EltSizeInBits && Subtarget.hasAVX()) ||
33089
33090
(MaskVT.is256BitVector() && Subtarget.hasAVX2()) ||
33090
33091
(MaskVT.is512BitVector() && Subtarget.hasAVX512())) {
33091
- if (matchVectorShuffleWithUNPCK (MaskVT, V1, V2, Shuffle, IsUnary, Mask, DL,
33092
- DAG, Subtarget)) {
33092
+ if (matchShuffleWithUNPCK (MaskVT, V1, V2, Shuffle, IsUnary, Mask, DL, DAG ,
33093
+ Subtarget)) {
33093
33094
SrcVT = DstVT = MaskVT;
33094
33095
if (MaskVT.is256BitVector() && !Subtarget.hasAVX2())
33095
33096
SrcVT = DstVT = (32 == EltSizeInBits ? MVT::v8f32 : MVT::v4f64);
@@ -33127,8 +33128,8 @@ static bool matchBinaryPermuteShuffle(
33127
33128
uint64_t BlendMask = 0;
33128
33129
bool ForceV1Zero = false, ForceV2Zero = false;
33129
33130
SmallVector<int, 8> TargetMask(Mask.begin(), Mask.end());
33130
- if (matchVectorShuffleAsBlend (V1, V2, TargetMask, Zeroable, ForceV1Zero,
33131
- ForceV2Zero, BlendMask)) {
33131
+ if (matchShuffleAsBlend (V1, V2, TargetMask, Zeroable, ForceV1Zero,
33132
+ ForceV2Zero, BlendMask)) {
33132
33133
if (MaskVT == MVT::v16i16) {
33133
33134
// We can only use v16i16 PBLENDW if the lanes are repeated.
33134
33135
SmallVector<int, 8> RepeatedMask;
0 commit comments