static Value *canonicalizeSaturatedAdd(ICmpInst *Cmp, Value *TVal, Value *FVal,
InstCombiner::BuilderTy &Builder) {
- // Match an unsigned saturated add with constant.
- Value *X = Cmp->getOperand(0);
- const APInt *CmpC, *AddC;
- if (!Cmp->hasOneUse() || Cmp->getPredicate() != ICmpInst::ICMP_ULT ||
- !match(Cmp->getOperand(1), m_APInt(CmpC)) || !match(FVal, m_AllOnes()) ||
- !match(TVal, m_Add(m_Specific(X), m_APInt(AddC))) || ~(*AddC) != *CmpC)
+ if (!Cmp->hasOneUse() || Cmp->getPredicate() != ICmpInst::ICMP_ULT)
return nullptr;
- // Commute compare and select operands:
- // select (icmp ult X, C), (add X, ~C), -1 -->
- // select (icmp ugt X, C), -1, (add X, ~C)
- Value *NewCmp = Builder.CreateICmp(ICmpInst::ICMP_UGT, X, Cmp->getOperand(1));
- return Builder.CreateSelect(NewCmp, FVal, TVal);
+ // Match unsigned saturated add of 2 variables with an unnecessary 'not'.
+ // TODO: There are more variations of this pattern.
+ Value *Cmp0 = Cmp->getOperand(0);
+ Value *Cmp1 = Cmp->getOperand(1);
+ Value *X, *Y;
+ if (match(TVal, m_AllOnes()) && match(Cmp0, m_Not(m_Value(X))) &&
+ match(FVal, m_c_Add(m_Specific(X), m_Value(Y))) && Y == Cmp1) {
+ // Change the comparison to use the sum (false value of the select). That is
+ // the canonical pattern match form for uadd.with.overflow and eliminates a
+ // use of the 'not' op:
+ // (~X u< Y) ? -1 : (X + Y) --> ((X + Y) u< Y) ? -1 : (X + Y)
+ // (~X u< Y) ? -1 : (Y + X) --> ((Y + X) u< Y) ? -1 : (Y + X)
+ Value *NewCmp = Builder.CreateICmp(ICmpInst::ICMP_ULT, FVal, Y);
+ return Builder.CreateSelect(NewCmp, TVal, FVal);
+ }
+
+ // Match unsigned saturated add with constant.
+ const APInt *C, *CmpC;
+ if (match(TVal, m_Add(m_Value(X), m_APInt(C))) && X == Cmp0 &&
+ match(FVal, m_AllOnes()) && match(Cmp1, m_APInt(CmpC)) && *CmpC == ~*C) {
+ // Commute compare predicate and select operands:
+ // (X u< ~C) ? (X + C) : -1 --> (X u> ~C) ? -1 : (X + C)
+ Value *NewCmp = Builder.CreateICmp(ICmpInst::ICMP_UGT, X, Cmp1);
+ return Builder.CreateSelect(NewCmp, FVal, TVal);
+ }
+
+ return nullptr;
}
/// Attempt to fold a cttz/ctlz followed by a icmp plus select into a single
define i32 @uadd_sat(i32 %x, i32 %y) {
; CHECK-LABEL: @uadd_sat(
-; CHECK-NEXT: [[NOTX:%.*]] = xor i32 [[X:%.*]], -1
-; CHECK-NEXT: [[A:%.*]] = add i32 [[Y:%.*]], [[X]]
-; CHECK-NEXT: [[C:%.*]] = icmp ult i32 [[NOTX]], [[Y]]
-; CHECK-NEXT: [[R:%.*]] = select i1 [[C]], i32 -1, i32 [[A]]
-; CHECK-NEXT: ret i32 [[R]]
+; CHECK-NEXT: [[A:%.*]] = add i32 [[Y:%.*]], [[X:%.*]]
+; CHECK-NEXT: [[TMP1:%.*]] = icmp ult i32 [[A]], [[Y]]
+; CHECK-NEXT: [[TMP2:%.*]] = select i1 [[TMP1]], i32 -1, i32 [[A]]
+; CHECK-NEXT: ret i32 [[TMP2]]
;
%notx = xor i32 %x, -1
%a = add i32 %y, %x
define i32 @uadd_sat_commute1(i32 %xp, i32 %y) {
; CHECK-LABEL: @uadd_sat_commute1(
; CHECK-NEXT: [[X:%.*]] = urem i32 42, [[XP:%.*]]
-; CHECK-NEXT: [[NOTX:%.*]] = xor i32 [[X]], -1
; CHECK-NEXT: [[A:%.*]] = add i32 [[X]], [[Y:%.*]]
-; CHECK-NEXT: [[C:%.*]] = icmp ult i32 [[NOTX]], [[Y]]
-; CHECK-NEXT: [[R:%.*]] = select i1 [[C]], i32 -1, i32 [[A]]
-; CHECK-NEXT: ret i32 [[R]]
+; CHECK-NEXT: [[TMP1:%.*]] = icmp ult i32 [[A]], [[Y]]
+; CHECK-NEXT: [[TMP2:%.*]] = select i1 [[TMP1]], i32 -1, i32 [[A]]
+; CHECK-NEXT: ret i32 [[TMP2]]
;
%x = urem i32 42, %xp ; thwart complexity-based-canonicalization
%notx = xor i32 %x, -1