@@ -28846,7 +28846,7 @@ SDValue X86TargetLowering::LowerWin64_INT128_TO_FP(SDValue Op,
28846
28846
28847
28847
// Return true if the required (according to Opcode) shift-imm form is natively
28848
28848
// supported by the Subtarget
28849
- static bool SupportedVectorShiftWithImm (MVT VT, const X86Subtarget &Subtarget,
28849
+ static bool supportedVectorShiftWithImm (MVT VT, const X86Subtarget &Subtarget,
28850
28850
unsigned Opcode) {
28851
28851
if (VT.getScalarSizeInBits() < 16)
28852
28852
return false;
@@ -28866,14 +28866,14 @@ static bool SupportedVectorShiftWithImm(MVT VT, const X86Subtarget &Subtarget,
28866
28866
// The shift amount is a variable, but it is the same for all vector lanes.
28867
28867
// These instructions are defined together with shift-immediate.
28868
28868
static
28869
- bool SupportedVectorShiftWithBaseAmnt (MVT VT, const X86Subtarget &Subtarget,
28869
+ bool supportedVectorShiftWithBaseAmnt (MVT VT, const X86Subtarget &Subtarget,
28870
28870
unsigned Opcode) {
28871
- return SupportedVectorShiftWithImm (VT, Subtarget, Opcode);
28871
+ return supportedVectorShiftWithImm (VT, Subtarget, Opcode);
28872
28872
}
28873
28873
28874
28874
// Return true if the required (according to Opcode) variable-shift form is
28875
28875
// natively supported by the Subtarget
28876
- static bool SupportedVectorVarShift (MVT VT, const X86Subtarget &Subtarget,
28876
+ static bool supportedVectorVarShift (MVT VT, const X86Subtarget &Subtarget,
28877
28877
unsigned Opcode) {
28878
28878
28879
28879
if (!Subtarget.hasInt256() || VT.getScalarSizeInBits() < 16)
@@ -28949,7 +28949,7 @@ static SDValue LowerScalarImmediateShift(SDValue Op, SelectionDAG &DAG,
28949
28949
28950
28950
uint64_t ShiftAmt = APIntShiftAmt.getZExtValue();
28951
28951
28952
- if (SupportedVectorShiftWithImm (VT, Subtarget, Op.getOpcode()))
28952
+ if (supportedVectorShiftWithImm (VT, Subtarget, Op.getOpcode()))
28953
28953
return getTargetVShiftByConstNode(X86Opc, dl, VT, R, ShiftAmt, DAG);
28954
28954
28955
28955
// i64 SRA needs to be performed as partial shifts.
@@ -29033,7 +29033,7 @@ static SDValue LowerScalarVariableShift(SDValue Op, SelectionDAG &DAG,
29033
29033
unsigned X86OpcV = getTargetVShiftUniformOpcode(Opcode, true);
29034
29034
29035
29035
if (SDValue BaseShAmt = DAG.getSplatValue(Amt)) {
29036
- if (SupportedVectorShiftWithBaseAmnt (VT, Subtarget, Opcode)) {
29036
+ if (supportedVectorShiftWithBaseAmnt (VT, Subtarget, Opcode)) {
29037
29037
MVT EltVT = VT.getVectorElementType();
29038
29038
assert(EltVT.bitsLE(MVT::i64) && "Unexpected element type!");
29039
29039
if (EltVT != MVT::i64 && EltVT.bitsGT(MVT::i32))
@@ -29051,7 +29051,7 @@ static SDValue LowerScalarVariableShift(SDValue Op, SelectionDAG &DAG,
29051
29051
!Subtarget.hasXOP()) {
29052
29052
unsigned NumElts = VT.getVectorNumElements();
29053
29053
MVT ExtVT = MVT::getVectorVT(MVT::i16, NumElts / 2);
29054
- if (SupportedVectorShiftWithBaseAmnt (ExtVT, Subtarget, Opcode)) {
29054
+ if (supportedVectorShiftWithBaseAmnt (ExtVT, Subtarget, Opcode)) {
29055
29055
unsigned LogicalOp = (Opcode == ISD::SHL ? ISD::SHL : ISD::SRL);
29056
29056
unsigned LogicalX86Op = getTargetVShiftUniformOpcode(LogicalOp, false);
29057
29057
BaseShAmt = DAG.getNode(ISD::ZERO_EXTEND, dl, MVT::i32, BaseShAmt);
@@ -29103,7 +29103,7 @@ static SDValue LowerScalarVariableShift(SDValue Op, SelectionDAG &DAG,
29103
29103
return SDValue();
29104
29104
}
29105
29105
29106
- if (SupportedVectorShiftWithBaseAmnt (VT, Subtarget, Op.getOpcode()))
29106
+ if (supportedVectorShiftWithBaseAmnt (VT, Subtarget, Op.getOpcode()))
29107
29107
return DAG.getNode(X86OpcV, dl, VT, R, Op.getOperand(1));
29108
29108
}
29109
29109
return SDValue();
@@ -29196,7 +29196,7 @@ static SDValue LowerShift(SDValue Op, const X86Subtarget &Subtarget,
29196
29196
if (SDValue V = LowerScalarVariableShift(Op, DAG, Subtarget))
29197
29197
return V;
29198
29198
29199
- if (SupportedVectorVarShift (VT, Subtarget, Opc))
29199
+ if (supportedVectorVarShift (VT, Subtarget, Opc))
29200
29200
return Op;
29201
29201
29202
29202
// XOP has 128-bit variable logical/arithmetic shifts.
@@ -29815,8 +29815,8 @@ static SDValue LowerRotate(SDValue Op, const X86Subtarget &Subtarget,
29815
29815
}
29816
29816
29817
29817
bool ConstantAmt = ISD::isBuildVectorOfConstantSDNodes(Amt.getNode());
29818
- bool LegalVarShifts = SupportedVectorVarShift (VT, Subtarget, ISD::SHL) &&
29819
- SupportedVectorVarShift (VT, Subtarget, ISD::SRL);
29818
+ bool LegalVarShifts = supportedVectorVarShift (VT, Subtarget, ISD::SHL) &&
29819
+ supportedVectorVarShift (VT, Subtarget, ISD::SRL);
29820
29820
29821
29821
// Fallback for splats + all supported variable shifts.
29822
29822
// Fallback for non-constants AVX2 vXi16 as well.
@@ -45816,7 +45816,7 @@ static SDValue combineAndMaskToShift(SDNode *N, SelectionDAG &DAG,
45816
45816
if (isBitwiseNot(Op0))
45817
45817
return SDValue();
45818
45818
45819
- if (!SupportedVectorShiftWithImm (VT0.getSimpleVT(), Subtarget, ISD::SRL))
45819
+ if (!supportedVectorShiftWithImm (VT0.getSimpleVT(), Subtarget, ISD::SRL))
45820
45820
return SDValue();
45821
45821
45822
45822
unsigned EltBitWidth = VT0.getScalarSizeInBits();
0 commit comments