-
Notifications
You must be signed in to change notification settings - Fork 14.3k
[InstCombine] Add folds for (fp_binop ({s|u}itofp x), ({s|u}itofp y))
#82555
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
@llvm/pr-subscribers-llvm-transforms Author: None (goldsteinn) Changes
Patch is 49.29 KiB, truncated to 20.00 KiB below, full version: https://github.com/llvm/llvm-project/pull/82555.diff 7 Files Affected:
diff --git a/llvm/lib/Transforms/InstCombine/InstCombineAddSub.cpp b/llvm/lib/Transforms/InstCombine/InstCombineAddSub.cpp
index cfff5df9ff5005..f0ca83bb225f1f 100644
--- a/llvm/lib/Transforms/InstCombine/InstCombineAddSub.cpp
+++ b/llvm/lib/Transforms/InstCombine/InstCombineAddSub.cpp
@@ -1867,64 +1867,10 @@ Instruction *InstCombinerImpl::visitFAdd(BinaryOperator &I) {
// Check for (fadd double (sitofp x), y), see if we can merge this into an
// integer add followed by a promotion.
- Value *LHS = I.getOperand(0), *RHS = I.getOperand(1);
- if (SIToFPInst *LHSConv = dyn_cast<SIToFPInst>(LHS)) {
- Value *LHSIntVal = LHSConv->getOperand(0);
- Type *FPType = LHSConv->getType();
-
- // TODO: This check is overly conservative. In many cases known bits
- // analysis can tell us that the result of the addition has less significant
- // bits than the integer type can hold.
- auto IsValidPromotion = [](Type *FTy, Type *ITy) {
- Type *FScalarTy = FTy->getScalarType();
- Type *IScalarTy = ITy->getScalarType();
-
- // Do we have enough bits in the significand to represent the result of
- // the integer addition?
- unsigned MaxRepresentableBits =
- APFloat::semanticsPrecision(FScalarTy->getFltSemantics());
- return IScalarTy->getIntegerBitWidth() <= MaxRepresentableBits;
- };
-
- // (fadd double (sitofp x), fpcst) --> (sitofp (add int x, intcst))
- // ... if the constant fits in the integer value. This is useful for things
- // like (double)(x & 1234) + 4.0 -> (double)((X & 1234)+4) which no longer
- // requires a constant pool load, and generally allows the add to be better
- // instcombined.
- if (ConstantFP *CFP = dyn_cast<ConstantFP>(RHS))
- if (IsValidPromotion(FPType, LHSIntVal->getType())) {
- Constant *CI = ConstantFoldCastOperand(Instruction::FPToSI, CFP,
- LHSIntVal->getType(), DL);
- if (LHSConv->hasOneUse() &&
- ConstantFoldCastOperand(Instruction::SIToFP, CI, I.getType(), DL) ==
- CFP &&
- willNotOverflowSignedAdd(LHSIntVal, CI, I)) {
- // Insert the new integer add.
- Value *NewAdd = Builder.CreateNSWAdd(LHSIntVal, CI, "addconv");
- return new SIToFPInst(NewAdd, I.getType());
- }
- }
-
- // (fadd double (sitofp x), (sitofp y)) --> (sitofp (add int x, y))
- if (SIToFPInst *RHSConv = dyn_cast<SIToFPInst>(RHS)) {
- Value *RHSIntVal = RHSConv->getOperand(0);
- // It's enough to check LHS types only because we require int types to
- // be the same for this transform.
- if (IsValidPromotion(FPType, LHSIntVal->getType())) {
- // Only do this if x/y have the same type, if at least one of them has a
- // single use (so we don't increase the number of int->fp conversions),
- // and if the integer add will not overflow.
- if (LHSIntVal->getType() == RHSIntVal->getType() &&
- (LHSConv->hasOneUse() || RHSConv->hasOneUse()) &&
- willNotOverflowSignedAdd(LHSIntVal, RHSIntVal, I)) {
- // Insert the new integer add.
- Value *NewAdd = Builder.CreateNSWAdd(LHSIntVal, RHSIntVal, "addconv");
- return new SIToFPInst(NewAdd, I.getType());
- }
- }
- }
- }
+ if (Instruction *R = foldFBinOpOfIntCasts(I))
+ return R;
+ Value *LHS = I.getOperand(0), *RHS = I.getOperand(1);
// Handle specials cases for FAdd with selects feeding the operation
if (Value *V = SimplifySelectsFeedingBinaryOp(I, LHS, RHS))
return replaceInstUsesWith(I, V);
@@ -2832,6 +2778,9 @@ Instruction *InstCombinerImpl::visitFSub(BinaryOperator &I) {
if (Instruction *X = foldFNegIntoConstant(I, DL))
return X;
+ if (Instruction *R = foldFBinOpOfIntCasts(I))
+ return R;
+
Value *X, *Y;
Constant *C;
diff --git a/llvm/lib/Transforms/InstCombine/InstCombineInternal.h b/llvm/lib/Transforms/InstCombine/InstCombineInternal.h
index 0b4283bc37650a..57148d719d9b61 100644
--- a/llvm/lib/Transforms/InstCombine/InstCombineInternal.h
+++ b/llvm/lib/Transforms/InstCombine/InstCombineInternal.h
@@ -379,6 +379,7 @@ class LLVM_LIBRARY_VISIBILITY InstCombinerImpl final
Instruction *scalarizePHI(ExtractElementInst &EI, PHINode *PN);
Instruction *foldBitcastExtElt(ExtractElementInst &ExtElt);
Instruction *foldCastedBitwiseLogic(BinaryOperator &I);
+ Instruction *foldFBinOpOfIntCasts(BinaryOperator &I);
Instruction *foldBinopOfSextBoolToSelect(BinaryOperator &I);
Instruction *narrowBinOp(TruncInst &Trunc);
Instruction *narrowMaskedBinOp(BinaryOperator &And);
diff --git a/llvm/lib/Transforms/InstCombine/InstCombineMulDivRem.cpp b/llvm/lib/Transforms/InstCombine/InstCombineMulDivRem.cpp
index 0bd4b6d1a835af..3ebf6b3d9bf7f7 100644
--- a/llvm/lib/Transforms/InstCombine/InstCombineMulDivRem.cpp
+++ b/llvm/lib/Transforms/InstCombine/InstCombineMulDivRem.cpp
@@ -769,6 +769,9 @@ Instruction *InstCombinerImpl::visitFMul(BinaryOperator &I) {
if (Instruction *R = foldFPSignBitOps(I))
return R;
+ if (Instruction *R = foldFBinOpOfIntCasts(I))
+ return R;
+
// X * -1.0 --> -X
Value *Op0 = I.getOperand(0), *Op1 = I.getOperand(1);
if (match(Op1, m_SpecificFP(-1.0)))
diff --git a/llvm/lib/Transforms/InstCombine/InstructionCombining.cpp b/llvm/lib/Transforms/InstCombine/InstructionCombining.cpp
index 4af455c37c788c..e5bbf59151bf9a 100644
--- a/llvm/lib/Transforms/InstCombine/InstructionCombining.cpp
+++ b/llvm/lib/Transforms/InstCombine/InstructionCombining.cpp
@@ -1401,6 +1401,171 @@ Value *InstCombinerImpl::dyn_castNegVal(Value *V) const {
return nullptr;
}
+// Try to fold:
+// 1) (fp_binop ({s|u}itofp x), ({s|u}itofp y))
+// -> ({s|u}itofp (int_binop x, y))
+// 2) (fp_binop ({s|u}itofp x), FpC)
+// -> ({s|u}itofp (int_binop x, (fpto{s|u}i FpC)))
+Instruction *InstCombinerImpl::foldFBinOpOfIntCasts(BinaryOperator &BO) {
+ Value *IntOps[2];
+ Constant *Op1FpC = nullptr;
+
+ // Check for:
+ // 1) (binop ({s|u}itofp x), ({s|u}itofp y))
+ // 2) (binop ({s|u}itofp x), FpC)
+ if (!match(BO.getOperand(0), m_SIToFP(m_Value(IntOps[0]))) &&
+ !match(BO.getOperand(0), m_UIToFP(m_Value(IntOps[0]))))
+ return nullptr;
+
+ if (!match(BO.getOperand(1), m_Constant(Op1FpC)) &&
+ !match(BO.getOperand(1), m_SIToFP(m_Value(IntOps[1]))) &&
+ !match(BO.getOperand(1), m_UIToFP(m_Value(IntOps[1]))))
+ return nullptr;
+
+
+ Type *FPTy = BO.getType();
+ Type *IntTy = IntOps[0]->getType();
+
+ // Do we have signed casts?
+ bool OpsFromSigned = isa<SIToFPInst>(BO.getOperand(0));
+
+
+ unsigned IntSz = IntTy->getScalarSizeInBits();
+ // This is the maximum number of inuse bits by the integer where the int -> fp
+ // casts are exact.
+ unsigned MaxRepresentableBits =
+ APFloat::semanticsPrecision(FPTy->getScalarType()->getFltSemantics());
+
+ // Cache KnownBits a bit to potentially save some analysis.
+ std::optional<KnownBits> OpsKnown[2];
+
+ // Preserve known number of leading bits. This can allow us to trivial nsw/nuw
+ // checks later on.
+ unsigned NumUsedLeadingBits[2] = {IntSz, IntSz};
+
+ auto IsNonZero = [&](unsigned OpNo) -> bool {
+ if (OpsKnown[OpNo].has_value() && OpsKnown[OpNo]->isNonZero())
+ return true;
+ return isKnownNonZero(IntOps[OpNo], SQ.DL);
+ };
+
+ auto IsNonNeg = [&](unsigned OpNo) -> bool {
+ if (OpsKnown[OpNo].has_value() && OpsKnown[OpNo]->isNonNegative())
+ return true;
+ return isKnownNonNegative(IntOps[OpNo], SQ);
+ };
+
+ // Check if we know for certain that ({s|u}itofp op) is exact.
+ auto IsValidPromotion = [&](unsigned OpNo) -> bool {
+ // If fp precision >= bitwidth(op) then its exact.
+ if (MaxRepresentableBits >= IntSz)
+ ;
+ // Otherwise if its signed cast check that fp precisions >= bitwidth(op) -
+ // numSignBits(op).
+ else if (OpsFromSigned)
+ NumUsedLeadingBits[OpNo] = IntSz - ComputeNumSignBits(IntOps[OpNo]);
+ // Finally for unsigned check that fp precision >= bitwidth(op) -
+ // numLeadingZeros(op).
+ else {
+ if (!OpsKnown[OpNo].has_value())
+ OpsKnown[OpNo] = computeKnownBits(IntOps[OpNo], /*Depth*/ 0, &BO);
+ NumUsedLeadingBits[OpNo] = IntSz - OpsKnown[OpNo]->countMinLeadingZeros();
+ }
+ // NB: We could also check if op is known to be a power of 2 or zero (which
+ // will always be representable). Its unlikely, however, that is we are
+ // unable to bound op in any way we will be able to pass the overflow checks
+ // later on.
+
+ if (MaxRepresentableBits < NumUsedLeadingBits[OpNo])
+ return false;
+ // Signed + Mul also requires that op is non-zero to avoid -0 cases.
+ return (OpsFromSigned && BO.getOpcode() == Instruction::FMul)
+ ? IsNonZero(OpNo)
+ : true;
+
+ };
+
+ // If we have a constant rhs, see if we can losslessly convert it to an int.
+ if (Op1FpC != nullptr) {
+ Constant *Op1IntC = ConstantFoldCastOperand(
+ OpsFromSigned ? Instruction::FPToSI : Instruction::FPToUI, Op1FpC,
+ IntTy, DL);
+ if (Op1IntC == nullptr)
+ return nullptr;
+ if (ConstantFoldCastOperand(OpsFromSigned ? Instruction::SIToFP
+ : Instruction::UIToFP,
+ Op1IntC, FPTy, DL) != Op1FpC)
+ return nullptr;
+
+ // First try to keep sign of cast the same.
+ IntOps[1] = Op1IntC;
+ }
+
+ // Ensure lhs/rhs integer types match.
+ if (IntTy != IntOps[1]->getType())
+ return nullptr;
+
+
+ if (Op1FpC == nullptr) {
+ if (OpsFromSigned != isa<SIToFPInst>(BO.getOperand(1))) {
+ // If we have a signed + unsigned, see if we can treat both as signed
+ // (uitofp nneg x) == (sitofp nneg x).
+ if (OpsFromSigned ? !IsNonNeg(1) : !IsNonNeg(0))
+ return nullptr;
+ OpsFromSigned = true;
+ }
+ if (!IsValidPromotion(1))
+ return nullptr;
+ }
+ if (!IsValidPromotion(0))
+ return nullptr;
+
+ // Final we check if the integer version of the binop will not overflow.
+ BinaryOperator::BinaryOps IntOpc;
+ // Because of the precision check, we can often rule out overflows.
+ bool NeedsOverflowCheck = true;
+ // Try to conservatively rule out overflow based on the already done precision
+ // checks.
+ unsigned OverflowMaxOutputBits = OpsFromSigned ? 2 : 1;
+ unsigned OverflowMaxCurBits =
+ std::max(NumUsedLeadingBits[0], NumUsedLeadingBits[1]);
+ bool OutputSigned = OpsFromSigned;
+ switch (BO.getOpcode()) {
+ case Instruction::FAdd:
+ IntOpc = Instruction::Add;
+ OverflowMaxOutputBits += OverflowMaxCurBits;
+ break;
+ case Instruction::FSub:
+ IntOpc = Instruction::Sub;
+ OverflowMaxOutputBits += OverflowMaxCurBits;
+ break;
+ case Instruction::FMul:
+ IntOpc = Instruction::Mul;
+ OverflowMaxOutputBits += OverflowMaxCurBits * 2;
+ break;
+ default:
+ llvm_unreachable("Unsupported binop");
+ }
+ // The precision check may have already ruled out overflow.
+ if (OverflowMaxOutputBits < IntSz) {
+ NeedsOverflowCheck = false;
+ // We can bound unsigned overflow from sub to in range signed value (this is
+ // what allows us to avoid the overflow check for sub).
+ if (IntOpc == Instruction::Sub)
+ OutputSigned = true;
+ }
+
+ // Precision check did not rule out overflow, so need to check.
+ if (NeedsOverflowCheck &&
+ !willNotOverflow(IntOpc, IntOps[0], IntOps[1], BO, OutputSigned))
+ return nullptr;
+
+ Value *IntBinOp = Builder.CreateBinOp(IntOpc, IntOps[0], IntOps[1]);
+ if (OutputSigned)
+ return new SIToFPInst(IntBinOp, FPTy);
+ return new UIToFPInst(IntBinOp, FPTy);
+}
+
/// A binop with a constant operand and a sign-extended boolean operand may be
/// converted into a select of constants by applying the binary operation to
/// the constant with the two possible values of the extended boolean (0 or -1).
diff --git a/llvm/test/Transforms/InstCombine/add-sitofp.ll b/llvm/test/Transforms/InstCombine/add-sitofp.ll
index db44b806593b64..bb97e9002c29b8 100644
--- a/llvm/test/Transforms/InstCombine/add-sitofp.ll
+++ b/llvm/test/Transforms/InstCombine/add-sitofp.ll
@@ -5,8 +5,8 @@ define double @x(i32 %a, i32 %b) {
; CHECK-LABEL: @x(
; CHECK-NEXT: [[M:%.*]] = lshr i32 [[A:%.*]], 24
; CHECK-NEXT: [[N:%.*]] = and i32 [[M]], [[B:%.*]]
-; CHECK-NEXT: [[ADDCONV:%.*]] = add nuw nsw i32 [[N]], 1
-; CHECK-NEXT: [[P:%.*]] = sitofp i32 [[ADDCONV]] to double
+; CHECK-NEXT: [[TMP1:%.*]] = add nuw nsw i32 [[N]], 1
+; CHECK-NEXT: [[P:%.*]] = sitofp i32 [[TMP1]] to double
; CHECK-NEXT: ret double [[P]]
;
%m = lshr i32 %a, 24
@@ -19,8 +19,8 @@ define double @x(i32 %a, i32 %b) {
define double @test(i32 %a) {
; CHECK-LABEL: @test(
; CHECK-NEXT: [[A_AND:%.*]] = and i32 [[A:%.*]], 1073741823
-; CHECK-NEXT: [[ADDCONV:%.*]] = add nuw nsw i32 [[A_AND]], 1
-; CHECK-NEXT: [[RES:%.*]] = sitofp i32 [[ADDCONV]] to double
+; CHECK-NEXT: [[TMP1:%.*]] = add nuw nsw i32 [[A_AND]], 1
+; CHECK-NEXT: [[RES:%.*]] = sitofp i32 [[TMP1]] to double
; CHECK-NEXT: ret double [[RES]]
;
; Drop two highest bits to guarantee that %a + 1 doesn't overflow
@@ -48,8 +48,8 @@ define double @test_2(i32 %a, i32 %b) {
; CHECK-LABEL: @test_2(
; CHECK-NEXT: [[A_AND:%.*]] = and i32 [[A:%.*]], 1073741823
; CHECK-NEXT: [[B_AND:%.*]] = and i32 [[B:%.*]], 1073741823
-; CHECK-NEXT: [[ADDCONV:%.*]] = add nuw nsw i32 [[A_AND]], [[B_AND]]
-; CHECK-NEXT: [[RES:%.*]] = sitofp i32 [[ADDCONV]] to double
+; CHECK-NEXT: [[TMP1:%.*]] = add nuw nsw i32 [[A_AND]], [[B_AND]]
+; CHECK-NEXT: [[RES:%.*]] = sitofp i32 [[TMP1]] to double
; CHECK-NEXT: ret double [[RES]]
;
; Drop two highest bits to guarantee that %a + %b doesn't overflow
@@ -90,8 +90,8 @@ define float @test_3(i32 %a, i32 %b) {
; CHECK-LABEL: @test_3(
; CHECK-NEXT: [[M:%.*]] = lshr i32 [[A:%.*]], 24
; CHECK-NEXT: [[N:%.*]] = and i32 [[M]], [[B:%.*]]
-; CHECK-NEXT: [[O:%.*]] = sitofp i32 [[N]] to float
-; CHECK-NEXT: [[P:%.*]] = fadd float [[O]], 1.000000e+00
+; CHECK-NEXT: [[TMP1:%.*]] = add nuw nsw i32 [[N]], 1
+; CHECK-NEXT: [[P:%.*]] = sitofp i32 [[TMP1]] to float
; CHECK-NEXT: ret float [[P]]
;
%m = lshr i32 %a, 24
@@ -105,8 +105,8 @@ define <4 x double> @test_4(<4 x i32> %a, <4 x i32> %b) {
; CHECK-LABEL: @test_4(
; CHECK-NEXT: [[A_AND:%.*]] = and <4 x i32> [[A:%.*]], <i32 1073741823, i32 1073741823, i32 1073741823, i32 1073741823>
; CHECK-NEXT: [[B_AND:%.*]] = and <4 x i32> [[B:%.*]], <i32 1073741823, i32 1073741823, i32 1073741823, i32 1073741823>
-; CHECK-NEXT: [[ADDCONV:%.*]] = add nuw nsw <4 x i32> [[A_AND]], [[B_AND]]
-; CHECK-NEXT: [[RES:%.*]] = sitofp <4 x i32> [[ADDCONV]] to <4 x double>
+; CHECK-NEXT: [[TMP1:%.*]] = add nuw nsw <4 x i32> [[A_AND]], [[B_AND]]
+; CHECK-NEXT: [[RES:%.*]] = sitofp <4 x i32> [[TMP1]] to <4 x double>
; CHECK-NEXT: ret <4 x double> [[RES]]
;
; Drop two highest bits to guarantee that %a + %b doesn't overflow
diff --git a/llvm/test/Transforms/InstCombine/binop-itofp.ll b/llvm/test/Transforms/InstCombine/binop-itofp.ll
new file mode 100644
index 00000000000000..ffa89374579145
--- /dev/null
+++ b/llvm/test/Transforms/InstCombine/binop-itofp.ll
@@ -0,0 +1,998 @@
+; NOTE: Assertions have been autogenerated by utils/update_test_checks.py
+; RUN: opt < %s -passes=instcombine -S | FileCheck %s
+
+define half @test_ui_ui_i8_add(i8 noundef %x_in, i8 noundef %y_in) {
+; CHECK-LABEL: @test_ui_ui_i8_add(
+; CHECK-NEXT: [[X:%.*]] = and i8 [[X_IN:%.*]], 127
+; CHECK-NEXT: [[Y:%.*]] = and i8 [[Y_IN:%.*]], 127
+; CHECK-NEXT: [[TMP1:%.*]] = add nuw i8 [[X]], [[Y]]
+; CHECK-NEXT: [[R:%.*]] = uitofp i8 [[TMP1]] to half
+; CHECK-NEXT: ret half [[R]]
+;
+ %x = and i8 %x_in, 127
+ %y = and i8 %y_in, 127
+ %xf = uitofp i8 %x to half
+ %yf = uitofp i8 %y to half
+ %r = fadd half %xf, %yf
+ ret half %r
+}
+
+define half @test_ui_ui_i8_add_fail_overflow(i8 noundef %x_in, i8 noundef %y_in) {
+; CHECK-LABEL: @test_ui_ui_i8_add_fail_overflow(
+; CHECK-NEXT: [[X:%.*]] = and i8 [[X_IN:%.*]], 127
+; CHECK-NEXT: [[Y:%.*]] = and i8 [[Y_IN:%.*]], -127
+; CHECK-NEXT: [[XF:%.*]] = uitofp i8 [[X]] to half
+; CHECK-NEXT: [[YF:%.*]] = uitofp i8 [[Y]] to half
+; CHECK-NEXT: [[R:%.*]] = fadd half [[XF]], [[YF]]
+; CHECK-NEXT: ret half [[R]]
+;
+ %x = and i8 %x_in, 127
+ %y = and i8 %y_in, 129
+ %xf = uitofp i8 %x to half
+ %yf = uitofp i8 %y to half
+ %r = fadd half %xf, %yf
+ ret half %r
+}
+
+define half @test_ui_ui_i8_add_C(i8 noundef %x_in) {
+; CHECK-LABEL: @test_ui_ui_i8_add_C(
+; CHECK-NEXT: [[TMP1:%.*]] = or i8 [[X_IN:%.*]], -128
+; CHECK-NEXT: [[R:%.*]] = uitofp i8 [[TMP1]] to half
+; CHECK-NEXT: ret half [[R]]
+;
+ %x = and i8 %x_in, 127
+ %xf = uitofp i8 %x to half
+ %r = fadd half %xf, 128.0
+ ret half %r
+}
+
+define half @test_ui_ui_i8_add_C_fail_no_repr(i8 noundef %x_in) {
+; CHECK-LABEL: @test_ui_ui_i8_add_C_fail_no_repr(
+; CHECK-NEXT: [[X:%.*]] = and i8 [[X_IN:%.*]], 127
+; CHECK-NEXT: [[XF:%.*]] = uitofp i8 [[X]] to half
+; CHECK-NEXT: [[R:%.*]] = fadd half [[XF]], 0xH57F8
+; CHECK-NEXT: ret half [[R]]
+;
+ %x = and i8 %x_in, 127
+ %xf = uitofp i8 %x to half
+ %r = fadd half %xf, 127.5
+ ret half %r
+}
+
+define half @test_ui_ui_i8_add_C_fail_overflow(i8 noundef %x_in) {
+; CHECK-LABEL: @test_ui_ui_i8_add_C_fail_overflow(
+; CHECK-NEXT: [[X:%.*]] = and i8 [[X_IN:%.*]], 127
+; CHECK-NEXT: [[XF:%.*]] = uitofp i8 [[X]] to half
+; CHECK-NEXT: [[R:%.*]] = fadd half [[XF]], 0xH5808
+; CHECK-NEXT: ret half [[R]]
+;
+ %x = and i8 %x_in, 127
+ %xf = uitofp i8 %x to half
+ %r = fadd half %xf, 129.0
+ ret half %r
+}
+
+define half @test_si_si_i8_add(i8 noundef %x_in, i8 noundef %y_in) {
+; CHECK-LABEL: @test_si_si_i8_add(
+; CHECK-NEXT: [[X:%.*]] = or i8 [[X_IN:%.*]], -64
+; CHECK-NEXT: [[Y:%.*]] = or i8 [[Y_IN:%.*]], -64
+; CHECK-NEXT: [[TMP1:%.*]] = add nsw i8 [[X]], [[Y]]
+; CHECK-NEXT: [[R:%.*]] = sitofp i8 [[TMP1]] to half
+; CHECK-NEXT: ret half [[R]]
+;
+ %x = or i8 %x_in, -64
+ %y = or i8 %y_in, -64
+ %xf = sitofp i8 %x to half
+ %yf = sitofp i8 %y to half
+ %r = fadd half %xf, %yf
+ ret half %r
+}
+
+define half @test_si_si_i8_add_fail_overflow(i8 noundef %x_in, i8 noundef %y_in) {
+; CHECK-LABEL: @test_si_si_i8_add_fail_overflow(
+; CHECK-NEXT: [[X:%.*]] = or i8 [[X_IN:%.*]], -64
+; CHECK-NEXT: [[Y:%.*]] = or i8 [[Y_IN:%.*]], -65
+; CHECK-NEXT: [[XF:%.*]] = sitofp i8 [[X]] to half
+; CHECK-NEXT: [[YF:%.*]] = sitofp i8 [[Y]] to half
+; CHECK-NEXT: [[R:%.*]] = fadd half [[XF]], [[YF]]
+; CHECK-NEXT: ret half [[R]]
+;
+ %x = or i8 %x_in, -64
+ %y = or i8 %y_in, -65
+ %xf = sitofp i8 %x to half
+ %yf = sitofp i8 %y to half
+ %r = fadd half %xf, %yf
+ ret half %r
+}
+
+define half @test_ui_si_i8_add(i8 noundef %x_in, i8 noundef %y_in) {
+; CHECK-LABEL: @test_ui_si_i8_add(
+; CHECK-NEXT: [[X:%.*]] = and i8 [[X_IN:%.*]], 63
+; CHECK-NEXT: [[Y:%.*]] = and i8 [[Y_IN:%.*]], 63
+; CHECK-NEXT: [[TMP1:%.*]] = add nuw nsw i8 [[X]], [[Y]]
+; CHECK-NEXT: [[R:%.*]] = sitofp i8 [[TMP1]] to half
+; CHECK-NEXT: ret half [[R]]
+;
+ %x = and i8 %x_in, 63
+ %y = and i8 %y_in, 63
+ %xf = sitofp i8 %x to half
+ %yf = uitofp i8 %y to half
+ %r = fadd half %xf, %yf
+ ret half %r
+}
+
+define half @test_ui_si_i8_add_overflow(i8 noundef %x_in, i8 noundef %y_in) {
+; CHECK-LABEL: @test_ui_si_i8_add_overflow(
+; CHECK-NEXT: [[X:%.*]] = and i8 [[X_IN:%.*]], 63
+; CHECK-NEXT: [[Y:%.*]] = and i8 [[Y_IN:%.*]], 65
+; CHECK-NEXT: [[XF:%.*]] = sitofp i8 [[X]] to half
+; CHECK-NEXT: [[YF:%.*]] = uitofp i8 [[Y]] to half
+; CHECK-NEXT: [[R:%.*]] = fadd half [[XF]], [[YF]]
+; CHECK-NEXT: ret half [[R]]
+;
+ %x = and i8 %x_in, 63
+ %y = and i8 %y_in, 65
+ %xf = sitofp i8 %x to half
+ %yf = uitofp i8 %y to half
+ %r = fadd half %xf, %yf
+ ret half %r
+}
+
+define half @test_ui_ui_i8_sub_C(i8 noundef %x_in) {
+; CHECK-LABEL: @test_ui_ui_i8_sub_C(
+; CHECK-NEXT: [[TMP1:%.*]] = and i8 [[X_IN:%.*]], 127
+; CHECK-NEXT: [[R:%.*]] = uitofp i8 [[TMP1]] to half
+; CHECK-NEXT: ret half [[R]]
+;
+ %x = or i8 %x_in, 128
+ %xf = uitofp i8 %x to half
+ %r = fsub half %xf, 128.0
+ ret half %r
+}
+
+define half @test_ui_ui_i8_sub_C_fail_overflow(i8 noundef %x_in)...
[truncated]
|
(fp_binop ({s|u}itofp x), ({s|u}itofp y))
Result of all the proofs run locally:
|
✅ With the latest revision this PR passed the C/C++ code formatter. |
50df541
to
d04e70f
Compare
unsigned NumUsedLeadingBits[2] = {IntSz, IntSz}; | ||
|
||
auto IsNonZero = [&](unsigned OpNo) -> bool { | ||
if (OpsKnown[OpNo].hasKnownBits() && |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
if (OpsKnown[OpNo].hasKnownBits() && | |
if ( |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Don't think we want to do this. We may unnecessarily compute knownbits. The idea isKnownNonZero
is what we want to spend our compute on, we only check knownbits if we have it already as an early out. Same applies to nonneg.
}; | ||
|
||
auto IsNonNeg = [&](unsigned OpNo) -> bool { | ||
if (OpsKnown[OpNo].hasKnownBits() && |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
if (OpsKnown[OpNo].hasKnownBits() && | |
if ( |
// Otherwise if its signed cast check that fp precisions >= bitwidth(op) - | ||
// numSignBits(op). | ||
else if (OpsFromSigned) | ||
NumUsedLeadingBits[OpNo] = IntSz - ComputeNumSignBits(IntOps[OpNo]); |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Can we pass WithCache
to ComputeNumSignBits/ComputeMaxSignificantBits
to reuse the KnownBits
result?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I started working on a patch for that, but it ends up spanning across alot of value tracking. Would rather leave as TODO for now. I can add comment that it should be changed if/as we expand usage of the WithCache in ValueTracking. (overflow checks also have same issue).
That okay?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
The big issue I was running into, is we need to start caching Depth
and adding logic for if we want to override what we have cached (or potentially introduce regressions).
It was brought up in the original commit: #66611 (comment)
but nikic was opposed. Maybe we can revisit in the future if there is a proper compile time advantage.
d04e70f
to
b974bff
Compare
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
can you also test/prove the binop (sitofp), (uitofp) cases
return (OpsFromSigned && BO.getOpcode() == Instruction::FMul) | ||
? IsNonZero(OpNo) | ||
: true; |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
OpsFromSigned && Opcode == fmul && IsNonZero(OpNo)
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Hmm, thats not the same.
It could be: !OpsFromSigned || Opcode != fmul || isNonZero(OpNo)
. Would you prefer that?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Basically we only need the IsNonZero for sitofp
+ fmul
. Otherwise we are good.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yes, select of bool to bool value is a weird pattern
The |
unsigned MaxRepresentableBits = | ||
APFloat::semanticsPrecision(FPTy->getScalarType()->getFltSemantics()); |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Off by 1 depending on signed or unsigned?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Yes, we are overly conservative if IntSz == MaxRepresentableBits + 1
and its signed (not we can't expand the bound for say sitofp i16 to half
b.c it won't sign extend). So for things like i12
, i24
, ...
I'll add comment to the affect, although don't think its really worth increase code complexity to handle a special case that will almost never actually apply (weird integer widths are rare).
Proofs:
; $> /home/noah/programs/opensource/llvm-dev/src/alive2/build/alive-tv (-smt-to=200000000)
----------------------------------------
define half @src_sisi_add_i12(i12 noundef %x, i12 noundef %y) {
#0:
%overflow_info = sadd_overflow i12 noundef %x, noundef %y
%does_overflow = extractvalue {i12, i1, i8} %overflow_info, 1
%doesnot_overflow = xor i1 %does_overflow, 1
assume i1 %doesnot_overflow
%xf = sitofp i12 noundef %x to half
%yf = sitofp i12 noundef %y to half
%r = fadd half %xf, %yf
ret half %r
}
=>
define half @tgt_sisi_add_i12(i12 noundef %x, i12 noundef %y) {
#0:
%overflow_info = sadd_overflow i12 noundef %x, noundef %y
%does_overflow = extractvalue {i12, i1, i8} %overflow_info, 1
%doesnot_overflow = xor i1 %does_overflow, 1
assume i1 %doesnot_overflow
%xy = add i12 noundef %x, noundef %y
%r = sitofp i12 %xy to half
ret half %r
}
Transformation seems to be correct!
----------------------------------------
define half @src_uiui_add_i12(i12 noundef %x, i12 noundef %y) {
#0:
%overflow_info = uadd_overflow i12 noundef %x, noundef %y
%does_overflow = extractvalue {i12, i1, i8} %overflow_info, 1
%doesnot_overflow = xor i1 %does_overflow, 1
assume i1 %doesnot_overflow
%xf = uitofp i12 noundef %x to half
%yf = uitofp i12 noundef %y to half
%r = fadd half %xf, %yf
ret half %r
}
=>
define half @tgt_uiui_add_i12(i12 noundef %x, i12 noundef %y) {
#0:
%overflow_info = uadd_overflow i12 noundef %x, noundef %y
%does_overflow = extractvalue {i12, i1, i8} %overflow_info, 1
%doesnot_overflow = xor i1 %does_overflow, 1
assume i1 %doesnot_overflow
%xy = add i12 noundef %x, noundef %y
%r = uitofp i12 %xy to half
ret half %r
}
Transformation doesn't verify!
ERROR: Value mismatch
Example:
i12 noundef %x = #x80d (2061, -2035)
i12 noundef %y = #x002 (2)
Source:
{i12, i1, i8} %overflow_info = { #x80f (2063, -2033), #x0 (0), poison }
i1 %does_overflow = #x0 (0)
i1 %doesnot_overflow = #x1 (1)
half %xf = #x6806 (2060)
half %yf = #x4000 (2)
half %r = #x6807 (2062)
Target:
{i12, i1, i8} %overflow_info = { #x80f (2063, -2033), #x0 (0), poison }
i1 %does_overflow = #x0 (0)
i1 %doesnot_overflow = #x1 (1)
i12 %xy = #x80f (2063, -2033)
half %r = #x6808 (2064)
Source value: #x6807 (2062)
Target value: #x6808 (2064)
Summary:
1 correct transformations
1 incorrect transformations
0 failed-to-prove transformations
0 Alive2 errors
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Didn't add the comment here?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Added it at the check vs IntSz
in isValidPromotion
.
b974bff
to
498f112
Compare
ping |
// handled specially. We can't, however, increase the bound arbitrarily for | ||
// `sitofp` as for larger sizes, it won't sign extend. | ||
if (MaxRepresentableBits >= IntSz) | ||
; |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Empty statement weird here; invert conditions?
…tructionCombiner; NFC
The full fold is one of the following: 1) `(fp_binop ({s|u}itofp x), ({s|u}itofp y))` -> `({s|u}itofp (int_binop x, y))` 2) `(fp_binop ({s|u}itofp x), FpC)` -> `({s|u}itofp (int_binop x, (fpto{s|u}i FpC)))` And support the following binops: `fmul` -> `mul` `fadd` -> `add` `fsub` -> `sub` Proofs: https://alive2.llvm.org/ce/z/zuacA8 The proofs timeout, so they must be reproduced locally.
498f112
to
0e4d930
Compare
unsigned MaxRepresentableBits = | ||
APFloat::semanticsPrecision(FPTy->getScalarType()->getFltSemantics()); |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Didn't add the comment here?
This commit #include <stdio.h>
static int g_12 = 0xD010L;
static float g_842 = 1.0;
static volatile int g_2345 = 1;
static float(safe_mul_func_float_f_f)(float sf1, float sf2) {
return (((0x1.0p-100f * sf1) * (0x1.0p-28f * sf2)) >
(0x1.0p-100f * (0x1.0p-28f * 3.40282347e+38F)))
? (sf1)
: (sf1 * sf2);
}
static void func_10(int p_11) {
g_842 = safe_mul_func_float_f_f(0.0, (short)p_11);
g_12 = 65529UL;
g_2345 = p_11;
}
int main(int argc, char *argv[]) {
func_10(g_12);
printf("%f\n", g_842);
return 0;
} Before: This is my compilation option:
|
Reproduce, I think the issue is we don't check constants are positive for |
Yeah adding a non-zero check for constant fixes. Will have patch up soon. |
Bug was introduced in llvm#82555 We where missing check that the constant was non-zero for signed + mul transform.
Fix is up at: #85298 |
Bug was introduced in llvm#82555 We where missing check that the constant was non-zero for signed + mul transform.
Bug was introduced in llvm#82555 We where missing check that the constant was non-zero for signed + mul transform.
Fix pushed. |
We already handle the `+x` case, and noticed it was missing in the bug affecting llvm#82555 Proofs: https://alive2.llvm.org/ce/z/WUSvmV
We already handle the `+x` case, and noticed it was missing in the bug affecting llvm#82555 Proofs: https://alive2.llvm.org/ce/z/WUSvmV
We already handle the `+x` case, and noticed it was missing in the bug affecting #82555 Proofs: https://alive2.llvm.org/ce/z/WUSvmV Closes #85345
We already handle the `+x` case, and noticed it was missing in the bug affecting llvm#82555 Proofs: https://alive2.llvm.org/ce/z/WUSvmV Closes llvm#85345
The full fold is one of the following:
(fp_binop ({s|u}itofp x), ({s|u}itofp y))
->
({s|u}itofp (int_binop x, y))
(fp_binop ({s|u}itofp x), FpC)
->
({s|u}itofp (int_binop x, (fpto{s|u}i FpC)))
And support the following binops:
fmul
->mul
fadd
->add
fsub
->sub
Proofs: https://alive2.llvm.org/ce/z/zuacA8
The proofs timeout, so they must be reproduced locally.