-
Notifications
You must be signed in to change notification settings - Fork 15.1k
Backport "[Clang][CodeGen] Fix type for atomic float incdec operators (#107075)" #107184
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
|
@llvm/pr-subscribers-clang @llvm/pr-subscribers-backend-x86 Author: Yingwei Zheng (dtcxzyw) ChangesBackport 9fef09f. Patch is 31.58 KiB, truncated to 20.00 KiB below, full version: https://github.com/llvm/llvm-project/pull/107184.diff 3 Files Affected:
diff --git a/clang/lib/CodeGen/CGExprScalar.cpp b/clang/lib/CodeGen/CGExprScalar.cpp
index a17d68424bbce5..6e212e74676e8d 100644
--- a/clang/lib/CodeGen/CGExprScalar.cpp
+++ b/clang/lib/CodeGen/CGExprScalar.cpp
@@ -2833,18 +2833,22 @@ ScalarExprEmitter::EmitScalarPrePostIncDec(const UnaryOperator *E, LValue LV,
llvm::AtomicOrdering::SequentiallyConsistent);
return isPre ? Builder.CreateBinOp(op, old, amt) : old;
}
- // Special case for atomic increment/decrement on floats
+ // Special case for atomic increment/decrement on floats.
+ // Bail out non-power-of-2-sized floating point types (e.g., x86_fp80).
if (type->isFloatingType()) {
- llvm::AtomicRMWInst::BinOp aop =
- isInc ? llvm::AtomicRMWInst::FAdd : llvm::AtomicRMWInst::FSub;
- llvm::Instruction::BinaryOps op =
- isInc ? llvm::Instruction::FAdd : llvm::Instruction::FSub;
- llvm::Value *amt = llvm::ConstantFP::get(
- VMContext, llvm::APFloat(static_cast<float>(1.0)));
- llvm::Value *old =
- Builder.CreateAtomicRMW(aop, LV.getAddress(), amt,
- llvm::AtomicOrdering::SequentiallyConsistent);
- return isPre ? Builder.CreateBinOp(op, old, amt) : old;
+ llvm::Type *Ty = ConvertType(type);
+ if (llvm::has_single_bit(Ty->getScalarSizeInBits())) {
+ llvm::AtomicRMWInst::BinOp aop =
+ isInc ? llvm::AtomicRMWInst::FAdd : llvm::AtomicRMWInst::FSub;
+ llvm::Instruction::BinaryOps op =
+ isInc ? llvm::Instruction::FAdd : llvm::Instruction::FSub;
+ llvm::Value *amt = llvm::ConstantFP::get(Ty, 1.0);
+ llvm::AtomicRMWInst *old = Builder.CreateAtomicRMW(
+ aop, LV.getAddress(), amt,
+ llvm::AtomicOrdering::SequentiallyConsistent);
+
+ return isPre ? Builder.CreateBinOp(op, old, amt) : old;
+ }
}
value = EmitLoadOfLValue(LV, E->getExprLoc());
input = value;
diff --git a/clang/test/CodeGen/X86/x86-atomic-double.c b/clang/test/CodeGen/X86/x86-atomic-double.c
index 2354c89cc2b170..09c8f70c3db854 100644
--- a/clang/test/CodeGen/X86/x86-atomic-double.c
+++ b/clang/test/CodeGen/X86/x86-atomic-double.c
@@ -6,20 +6,14 @@
// X64-LABEL: define dso_local double @test_double_post_inc(
// X64-SAME: ) #[[ATTR0:[0-9]+]] {
// X64-NEXT: entry:
-// X64-NEXT: [[RETVAL:%.*]] = alloca double, align 8
-// X64-NEXT: [[TMP0:%.*]] = atomicrmw fadd ptr @test_double_post_inc.n, float 1.000000e+00 seq_cst, align 8
-// X64-NEXT: store float [[TMP0]], ptr [[RETVAL]], align 8
-// X64-NEXT: [[TMP1:%.*]] = load double, ptr [[RETVAL]], align 8
-// X64-NEXT: ret double [[TMP1]]
+// X64-NEXT: [[TMP0:%.*]] = atomicrmw fadd ptr @test_double_post_inc.n, double 1.000000e+00 seq_cst, align 8
+// X64-NEXT: ret double [[TMP0]]
//
// X86-LABEL: define dso_local double @test_double_post_inc(
// X86-SAME: ) #[[ATTR0:[0-9]+]] {
// X86-NEXT: entry:
-// X86-NEXT: [[RETVAL:%.*]] = alloca double, align 4
-// X86-NEXT: [[TMP0:%.*]] = atomicrmw fadd ptr @test_double_post_inc.n, float 1.000000e+00 seq_cst, align 8
-// X86-NEXT: store float [[TMP0]], ptr [[RETVAL]], align 4
-// X86-NEXT: [[TMP1:%.*]] = load double, ptr [[RETVAL]], align 4
-// X86-NEXT: ret double [[TMP1]]
+// X86-NEXT: [[TMP0:%.*]] = atomicrmw fadd ptr @test_double_post_inc.n, double 1.000000e+00 seq_cst, align 8
+// X86-NEXT: ret double [[TMP0]]
//
double test_double_post_inc()
{
@@ -30,20 +24,14 @@ double test_double_post_inc()
// X64-LABEL: define dso_local double @test_double_post_dc(
// X64-SAME: ) #[[ATTR0]] {
// X64-NEXT: entry:
-// X64-NEXT: [[RETVAL:%.*]] = alloca double, align 8
-// X64-NEXT: [[TMP0:%.*]] = atomicrmw fsub ptr @test_double_post_dc.n, float 1.000000e+00 seq_cst, align 8
-// X64-NEXT: store float [[TMP0]], ptr [[RETVAL]], align 8
-// X64-NEXT: [[TMP1:%.*]] = load double, ptr [[RETVAL]], align 8
-// X64-NEXT: ret double [[TMP1]]
+// X64-NEXT: [[TMP0:%.*]] = atomicrmw fsub ptr @test_double_post_dc.n, double 1.000000e+00 seq_cst, align 8
+// X64-NEXT: ret double [[TMP0]]
//
// X86-LABEL: define dso_local double @test_double_post_dc(
// X86-SAME: ) #[[ATTR0]] {
// X86-NEXT: entry:
-// X86-NEXT: [[RETVAL:%.*]] = alloca double, align 4
-// X86-NEXT: [[TMP0:%.*]] = atomicrmw fsub ptr @test_double_post_dc.n, float 1.000000e+00 seq_cst, align 8
-// X86-NEXT: store float [[TMP0]], ptr [[RETVAL]], align 4
-// X86-NEXT: [[TMP1:%.*]] = load double, ptr [[RETVAL]], align 4
-// X86-NEXT: ret double [[TMP1]]
+// X86-NEXT: [[TMP0:%.*]] = atomicrmw fsub ptr @test_double_post_dc.n, double 1.000000e+00 seq_cst, align 8
+// X86-NEXT: ret double [[TMP0]]
//
double test_double_post_dc()
{
@@ -54,22 +42,16 @@ double test_double_post_dc()
// X64-LABEL: define dso_local double @test_double_pre_dc(
// X64-SAME: ) #[[ATTR0]] {
// X64-NEXT: entry:
-// X64-NEXT: [[RETVAL:%.*]] = alloca double, align 8
-// X64-NEXT: [[TMP0:%.*]] = atomicrmw fsub ptr @test_double_pre_dc.n, float 1.000000e+00 seq_cst, align 8
-// X64-NEXT: [[TMP1:%.*]] = fsub float [[TMP0]], 1.000000e+00
-// X64-NEXT: store float [[TMP1]], ptr [[RETVAL]], align 8
-// X64-NEXT: [[TMP2:%.*]] = load double, ptr [[RETVAL]], align 8
-// X64-NEXT: ret double [[TMP2]]
+// X64-NEXT: [[TMP0:%.*]] = atomicrmw fsub ptr @test_double_pre_dc.n, double 1.000000e+00 seq_cst, align 8
+// X64-NEXT: [[TMP1:%.*]] = fsub double [[TMP0]], 1.000000e+00
+// X64-NEXT: ret double [[TMP1]]
//
// X86-LABEL: define dso_local double @test_double_pre_dc(
// X86-SAME: ) #[[ATTR0]] {
// X86-NEXT: entry:
-// X86-NEXT: [[RETVAL:%.*]] = alloca double, align 4
-// X86-NEXT: [[TMP0:%.*]] = atomicrmw fsub ptr @test_double_pre_dc.n, float 1.000000e+00 seq_cst, align 8
-// X86-NEXT: [[TMP1:%.*]] = fsub float [[TMP0]], 1.000000e+00
-// X86-NEXT: store float [[TMP1]], ptr [[RETVAL]], align 4
-// X86-NEXT: [[TMP2:%.*]] = load double, ptr [[RETVAL]], align 4
-// X86-NEXT: ret double [[TMP2]]
+// X86-NEXT: [[TMP0:%.*]] = atomicrmw fsub ptr @test_double_pre_dc.n, double 1.000000e+00 seq_cst, align 8
+// X86-NEXT: [[TMP1:%.*]] = fsub double [[TMP0]], 1.000000e+00
+// X86-NEXT: ret double [[TMP1]]
//
double test_double_pre_dc()
{
@@ -80,25 +62,43 @@ double test_double_pre_dc()
// X64-LABEL: define dso_local double @test_double_pre_inc(
// X64-SAME: ) #[[ATTR0]] {
// X64-NEXT: entry:
-// X64-NEXT: [[RETVAL:%.*]] = alloca double, align 8
-// X64-NEXT: [[TMP0:%.*]] = atomicrmw fadd ptr @test_double_pre_inc.n, float 1.000000e+00 seq_cst, align 8
-// X64-NEXT: [[TMP1:%.*]] = fadd float [[TMP0]], 1.000000e+00
-// X64-NEXT: store float [[TMP1]], ptr [[RETVAL]], align 8
-// X64-NEXT: [[TMP2:%.*]] = load double, ptr [[RETVAL]], align 8
-// X64-NEXT: ret double [[TMP2]]
+// X64-NEXT: [[TMP0:%.*]] = atomicrmw fadd ptr @test_double_pre_inc.n, double 1.000000e+00 seq_cst, align 8
+// X64-NEXT: [[TMP1:%.*]] = fadd double [[TMP0]], 1.000000e+00
+// X64-NEXT: ret double [[TMP1]]
//
// X86-LABEL: define dso_local double @test_double_pre_inc(
// X86-SAME: ) #[[ATTR0]] {
// X86-NEXT: entry:
-// X86-NEXT: [[RETVAL:%.*]] = alloca double, align 4
-// X86-NEXT: [[TMP0:%.*]] = atomicrmw fadd ptr @test_double_pre_inc.n, float 1.000000e+00 seq_cst, align 8
-// X86-NEXT: [[TMP1:%.*]] = fadd float [[TMP0]], 1.000000e+00
-// X86-NEXT: store float [[TMP1]], ptr [[RETVAL]], align 4
-// X86-NEXT: [[TMP2:%.*]] = load double, ptr [[RETVAL]], align 4
-// X86-NEXT: ret double [[TMP2]]
+// X86-NEXT: [[TMP0:%.*]] = atomicrmw fadd ptr @test_double_pre_inc.n, double 1.000000e+00 seq_cst, align 8
+// X86-NEXT: [[TMP1:%.*]] = fadd double [[TMP0]], 1.000000e+00
+// X86-NEXT: ret double [[TMP1]]
//
double test_double_pre_inc()
{
static _Atomic double n;
return ++n;
}
+
+// X64-LABEL: define dso_local i32 @pr107054(
+// X64-SAME: ) #[[ATTR0]] {
+// X64-NEXT: entry:
+// X64-NEXT: [[TMP0:%.*]] = atomicrmw fadd ptr @pr107054.n, double 1.000000e+00 seq_cst, align 8
+// X64-NEXT: [[TMP1:%.*]] = fadd double [[TMP0]], 1.000000e+00
+// X64-NEXT: [[CMP:%.*]] = fcmp oeq double [[TMP1]], 1.000000e+00
+// X64-NEXT: [[CONV:%.*]] = zext i1 [[CMP]] to i32
+// X64-NEXT: ret i32 [[CONV]]
+//
+// X86-LABEL: define dso_local i32 @pr107054(
+// X86-SAME: ) #[[ATTR0]] {
+// X86-NEXT: entry:
+// X86-NEXT: [[TMP0:%.*]] = atomicrmw fadd ptr @pr107054.n, double 1.000000e+00 seq_cst, align 8
+// X86-NEXT: [[TMP1:%.*]] = fadd double [[TMP0]], 1.000000e+00
+// X86-NEXT: [[CMP:%.*]] = fcmp oeq double [[TMP1]], 1.000000e+00
+// X86-NEXT: [[CONV:%.*]] = zext i1 [[CMP]] to i32
+// X86-NEXT: ret i32 [[CONV]]
+//
+int pr107054()
+{
+ static _Atomic double n;
+ return (++n) == 1;
+}
diff --git a/clang/test/CodeGen/X86/x86-atomic-long_double.c b/clang/test/CodeGen/X86/x86-atomic-long_double.c
index 2c3f381f13511e..9c82784807daca 100644
--- a/clang/test/CodeGen/X86/x86-atomic-long_double.c
+++ b/clang/test/CodeGen/X86/x86-atomic-long_double.c
@@ -4,29 +4,60 @@
// X64-LABEL: define dso_local x86_fp80 @testinc(
// X64-SAME: ptr noundef [[ADDR:%.*]]) #[[ATTR0:[0-9]+]] {
-// X64-NEXT: [[ENTRY:.*:]]
-// X64-NEXT: [[RETVAL:%.*]] = alloca x86_fp80, align 16
+// X64-NEXT: [[ENTRY:.*]]:
// X64-NEXT: [[ADDR_ADDR:%.*]] = alloca ptr, align 8
+// X64-NEXT: [[ATOMIC_TEMP:%.*]] = alloca x86_fp80, align 16
+// X64-NEXT: [[ATOMIC_TEMP1:%.*]] = alloca x86_fp80, align 16
+// X64-NEXT: [[ATOMIC_TEMP2:%.*]] = alloca x86_fp80, align 16
+// X64-NEXT: [[ATOMIC_TEMP3:%.*]] = alloca x86_fp80, align 16
// X64-NEXT: store ptr [[ADDR]], ptr [[ADDR_ADDR]], align 8
// X64-NEXT: [[TMP0:%.*]] = load ptr, ptr [[ADDR_ADDR]], align 8
-// X64-NEXT: [[TMP1:%.*]] = atomicrmw fadd ptr [[TMP0]], float 1.000000e+00 seq_cst, align 16
-// X64-NEXT: [[TMP2:%.*]] = fadd float [[TMP1]], 1.000000e+00
-// X64-NEXT: store float [[TMP2]], ptr [[RETVAL]], align 16
-// X64-NEXT: [[TMP3:%.*]] = load x86_fp80, ptr [[RETVAL]], align 16
-// X64-NEXT: ret x86_fp80 [[TMP3]]
+// X64-NEXT: [[ATOMIC_LOAD:%.*]] = load atomic i128, ptr [[TMP0]] seq_cst, align 16
+// X64-NEXT: store i128 [[ATOMIC_LOAD]], ptr [[ATOMIC_TEMP]], align 16
+// X64-NEXT: [[TMP1:%.*]] = load x86_fp80, ptr [[ATOMIC_TEMP]], align 16
+// X64-NEXT: br label %[[ATOMIC_OP:.*]]
+// X64: [[ATOMIC_OP]]:
+// X64-NEXT: [[TMP2:%.*]] = phi x86_fp80 [ [[TMP1]], %[[ENTRY]] ], [ [[TMP8:%.*]], %[[ATOMIC_OP]] ]
+// X64-NEXT: [[INC:%.*]] = fadd x86_fp80 [[TMP2]], 0xK3FFF8000000000000000
+// X64-NEXT: call void @llvm.memset.p0.i64(ptr align 16 [[ATOMIC_TEMP1]], i8 0, i64 16, i1 false)
+// X64-NEXT: store x86_fp80 [[TMP2]], ptr [[ATOMIC_TEMP1]], align 16
+// X64-NEXT: [[TMP3:%.*]] = load i128, ptr [[ATOMIC_TEMP1]], align 16
+// X64-NEXT: call void @llvm.memset.p0.i64(ptr align 16 [[ATOMIC_TEMP2]], i8 0, i64 16, i1 false)
+// X64-NEXT: store x86_fp80 [[INC]], ptr [[ATOMIC_TEMP2]], align 16
+// X64-NEXT: [[TMP4:%.*]] = load i128, ptr [[ATOMIC_TEMP2]], align 16
+// X64-NEXT: [[TMP5:%.*]] = cmpxchg ptr [[TMP0]], i128 [[TMP3]], i128 [[TMP4]] seq_cst seq_cst, align 16
+// X64-NEXT: [[TMP6:%.*]] = extractvalue { i128, i1 } [[TMP5]], 0
+// X64-NEXT: [[TMP7:%.*]] = extractvalue { i128, i1 } [[TMP5]], 1
+// X64-NEXT: store i128 [[TMP6]], ptr [[ATOMIC_TEMP3]], align 16
+// X64-NEXT: [[TMP8]] = load x86_fp80, ptr [[ATOMIC_TEMP3]], align 16
+// X64-NEXT: br i1 [[TMP7]], label %[[ATOMIC_CONT:.*]], label %[[ATOMIC_OP]]
+// X64: [[ATOMIC_CONT]]:
+// X64-NEXT: ret x86_fp80 [[INC]]
//
// X86-LABEL: define dso_local x86_fp80 @testinc(
// X86-SAME: ptr noundef [[ADDR:%.*]]) #[[ATTR0:[0-9]+]] {
-// X86-NEXT: [[ENTRY:.*:]]
-// X86-NEXT: [[RETVAL:%.*]] = alloca x86_fp80, align 4
+// X86-NEXT: [[ENTRY:.*]]:
// X86-NEXT: [[ADDR_ADDR:%.*]] = alloca ptr, align 4
+// X86-NEXT: [[ATOMIC_TEMP:%.*]] = alloca x86_fp80, align 4
+// X86-NEXT: [[ATOMIC_TEMP1:%.*]] = alloca x86_fp80, align 4
+// X86-NEXT: [[ATOMIC_TEMP2:%.*]] = alloca x86_fp80, align 4
// X86-NEXT: store ptr [[ADDR]], ptr [[ADDR_ADDR]], align 4
// X86-NEXT: [[TMP0:%.*]] = load ptr, ptr [[ADDR_ADDR]], align 4
-// X86-NEXT: [[TMP1:%.*]] = atomicrmw fadd ptr [[TMP0]], float 1.000000e+00 seq_cst, align 4
-// X86-NEXT: [[TMP2:%.*]] = fadd float [[TMP1]], 1.000000e+00
-// X86-NEXT: store float [[TMP2]], ptr [[RETVAL]], align 4
-// X86-NEXT: [[TMP3:%.*]] = load x86_fp80, ptr [[RETVAL]], align 4
-// X86-NEXT: ret x86_fp80 [[TMP3]]
+// X86-NEXT: call void @__atomic_load(i32 noundef 12, ptr noundef [[TMP0]], ptr noundef [[ATOMIC_TEMP]], i32 noundef 5)
+// X86-NEXT: [[TMP1:%.*]] = load x86_fp80, ptr [[ATOMIC_TEMP]], align 4
+// X86-NEXT: br label %[[ATOMIC_OP:.*]]
+// X86: [[ATOMIC_OP]]:
+// X86-NEXT: [[TMP2:%.*]] = phi x86_fp80 [ [[TMP1]], %[[ENTRY]] ], [ [[TMP3:%.*]], %[[ATOMIC_OP]] ]
+// X86-NEXT: [[INC:%.*]] = fadd x86_fp80 [[TMP2]], 0xK3FFF8000000000000000
+// X86-NEXT: call void @llvm.memset.p0.i64(ptr align 4 [[ATOMIC_TEMP1]], i8 0, i64 12, i1 false)
+// X86-NEXT: store x86_fp80 [[TMP2]], ptr [[ATOMIC_TEMP1]], align 4
+// X86-NEXT: call void @llvm.memset.p0.i64(ptr align 4 [[ATOMIC_TEMP2]], i8 0, i64 12, i1 false)
+// X86-NEXT: store x86_fp80 [[INC]], ptr [[ATOMIC_TEMP2]], align 4
+// X86-NEXT: [[CALL:%.*]] = call zeroext i1 @__atomic_compare_exchange(i32 noundef 12, ptr noundef [[TMP0]], ptr noundef [[ATOMIC_TEMP1]], ptr noundef [[ATOMIC_TEMP2]], i32 noundef 5, i32 noundef 5)
+// X86-NEXT: [[TMP3]] = load x86_fp80, ptr [[ATOMIC_TEMP1]], align 4
+// X86-NEXT: br i1 [[CALL]], label %[[ATOMIC_CONT:.*]], label %[[ATOMIC_OP]]
+// X86: [[ATOMIC_CONT]]:
+// X86-NEXT: ret x86_fp80 [[INC]]
//
long double testinc(_Atomic long double *addr) {
@@ -35,27 +66,60 @@ long double testinc(_Atomic long double *addr) {
// X64-LABEL: define dso_local x86_fp80 @testdec(
// X64-SAME: ptr noundef [[ADDR:%.*]]) #[[ATTR0]] {
-// X64-NEXT: [[ENTRY:.*:]]
-// X64-NEXT: [[RETVAL:%.*]] = alloca x86_fp80, align 16
+// X64-NEXT: [[ENTRY:.*]]:
// X64-NEXT: [[ADDR_ADDR:%.*]] = alloca ptr, align 8
+// X64-NEXT: [[ATOMIC_TEMP:%.*]] = alloca x86_fp80, align 16
+// X64-NEXT: [[ATOMIC_TEMP1:%.*]] = alloca x86_fp80, align 16
+// X64-NEXT: [[ATOMIC_TEMP2:%.*]] = alloca x86_fp80, align 16
+// X64-NEXT: [[ATOMIC_TEMP3:%.*]] = alloca x86_fp80, align 16
// X64-NEXT: store ptr [[ADDR]], ptr [[ADDR_ADDR]], align 8
// X64-NEXT: [[TMP0:%.*]] = load ptr, ptr [[ADDR_ADDR]], align 8
-// X64-NEXT: [[TMP1:%.*]] = atomicrmw fsub ptr [[TMP0]], float 1.000000e+00 seq_cst, align 16
-// X64-NEXT: store float [[TMP1]], ptr [[RETVAL]], align 16
-// X64-NEXT: [[TMP2:%.*]] = load x86_fp80, ptr [[RETVAL]], align 16
-// X64-NEXT: ret x86_fp80 [[TMP2]]
+// X64-NEXT: [[ATOMIC_LOAD:%.*]] = load atomic i128, ptr [[TMP0]] seq_cst, align 16
+// X64-NEXT: store i128 [[ATOMIC_LOAD]], ptr [[ATOMIC_TEMP]], align 16
+// X64-NEXT: [[TMP1:%.*]] = load x86_fp80, ptr [[ATOMIC_TEMP]], align 16
+// X64-NEXT: br label %[[ATOMIC_OP:.*]]
+// X64: [[ATOMIC_OP]]:
+// X64-NEXT: [[TMP2:%.*]] = phi x86_fp80 [ [[TMP1]], %[[ENTRY]] ], [ [[TMP8:%.*]], %[[ATOMIC_OP]] ]
+// X64-NEXT: [[DEC:%.*]] = fadd x86_fp80 [[TMP2]], 0xKBFFF8000000000000000
+// X64-NEXT: call void @llvm.memset.p0.i64(ptr align 16 [[ATOMIC_TEMP1]], i8 0, i64 16, i1 false)
+// X64-NEXT: store x86_fp80 [[TMP2]], ptr [[ATOMIC_TEMP1]], align 16
+// X64-NEXT: [[TMP3:%.*]] = load i128, ptr [[ATOMIC_TEMP1]], align 16
+// X64-NEXT: call void @llvm.memset.p0.i64(ptr align 16 [[ATOMIC_TEMP2]], i8 0, i64 16, i1 false)
+// X64-NEXT: store x86_fp80 [[DEC]], ptr [[ATOMIC_TEMP2]], align 16
+// X64-NEXT: [[TMP4:%.*]] = load i128, ptr [[ATOMIC_TEMP2]], align 16
+// X64-NEXT: [[TMP5:%.*]] = cmpxchg ptr [[TMP0]], i128 [[TMP3]], i128 [[TMP4]] seq_cst seq_cst, align 16
+// X64-NEXT: [[TMP6:%.*]] = extractvalue { i128, i1 } [[TMP5]], 0
+// X64-NEXT: [[TMP7:%.*]] = extractvalue { i128, i1 } [[TMP5]], 1
+// X64-NEXT: store i128 [[TMP6]], ptr [[ATOMIC_TEMP3]], align 16
+// X64-NEXT: [[TMP8]] = load x86_fp80, ptr [[ATOMIC_TEMP3]], align 16
+// X64-NEXT: br i1 [[TMP7]], label %[[ATOMIC_CONT:.*]], label %[[ATOMIC_OP]]
+// X64: [[ATOMIC_CONT]]:
+// X64-NEXT: ret x86_fp80 [[TMP1]]
//
// X86-LABEL: define dso_local x86_fp80 @testdec(
// X86-SAME: ptr noundef [[ADDR:%.*]]) #[[ATTR0]] {
-// X86-NEXT: [[ENTRY:.*:]]
-// X86-NEXT: [[RETVAL:%.*]] = alloca x86_fp80, align 4
+// X86-NEXT: [[ENTRY:.*]]:
// X86-NEXT: [[ADDR_ADDR:%.*]] = alloca ptr, align 4
+// X86-NEXT: [[ATOMIC_TEMP:%.*]] = alloca x86_fp80, align 4
+// X86-NEXT: [[ATOMIC_TEMP1:%.*]] = alloca x86_fp80, align 4
+// X86-NEXT: [[ATOMIC_TEMP2:%.*]] = alloca x86_fp80, align 4
// X86-NEXT: store ptr [[ADDR]], ptr [[ADDR_ADDR]], align 4
// X86-NEXT: [[TMP0:%.*]] = load ptr, ptr [[ADDR_ADDR]], align 4
-// X86-NEXT: [[TMP1:%.*]] = atomicrmw fsub ptr [[TMP0]], float 1.000000e+00 seq_cst, align 4
-// X86-NEXT: store float [[TMP1]], ptr [[RETVAL]], align 4
-// X86-NEXT: [[TMP2:%.*]] = load x86_fp80, ptr [[RETVAL]], align 4
-// X86-NEXT: ret x86_fp80 [[TMP2]]
+// X86-NEXT: call void @__atomic_load(i32 noundef 12, ptr noundef [[TMP0]], ptr noundef [[ATOMIC_TEMP]], i32 noundef 5)
+// X86-NEXT: [[TMP1:%.*]] = load x86_fp80, ptr [[ATOMIC_TEMP]], align 4
+// X86-NEXT: br label %[[ATOMIC_OP:.*]]
+// X86: [[ATOMIC_OP]]:
+// X86-NEXT: [[TMP2:%.*]] = phi x86_fp80 [ [[TMP1]], %[[ENTRY]] ], [ [[TMP3:%.*]], %[[ATOMIC_OP]] ]
+// X86-NEXT: [[DEC:%.*]] = fadd x86_fp80 [[TMP2]], 0xKBFFF8000000000000000
+// X86-NEXT: call void @llvm.memset.p0.i64(ptr align 4 [[ATOMIC_TEMP1]], i8 0, i64 12, i1 false)
+// X86-NEXT: store x86_fp80 [[TMP2]], ptr [[ATOMIC_TEMP1]], align 4
+// X86-NEXT: call void @llvm.memset.p0.i64(ptr align 4 [[ATOMIC_TEMP2]], i8 0, i64 12, i1 false)
+// X86-NEXT: store x86_fp80 [[DEC]], ptr [[ATOMIC_TEMP2]], align 4
+// X86-NEXT: [[CALL:%.*]] = call zeroext i1 @__atomic_compare_exchange(i32 noundef 12, ptr noundef [[TMP0]], ptr noundef [[ATOMIC_TEMP1]], ptr noundef [[ATOMIC_TEMP2]], i32 noundef 5, i32 noundef 5)
+// X86-NEXT: [[TMP3]] = load x86_fp80, ptr [[ATOMIC_TEMP1]], align 4
+// X86-NEXT: br i1 [[CALL]], label %[[ATOMIC_CONT:.*]], label %[[ATOMIC_OP]]
+// X86: [[ATOMIC_CONT]]:
+// X86-NEXT: ret x86_fp80 [[TMP1]]
//
long double testdec(_Atomic long double *addr) {
@@ -175,29 +239,60 @@ long double testassign(_Atomic long double *addr) {
// X64-LABEL: define dso_local x86_fp80 @test_volatile_inc(
// X64-SAME: ptr noundef [[ADDR:%.*]]) #[[ATTR0]] {
-// X64-NEXT: [[ENTRY:.*:]]
-// X64-NEXT: [[RETVAL:%.*]] = alloca x86_fp80, align 16
+// X64-NEXT: [[ENTRY:.*]]:
// X64-NEXT: [[ADDR_ADDR:%.*]] = alloca ptr, align 8
+// X64-NEXT: [[ATOMIC_TEMP:%.*]] = alloca x86_fp80, align 16
+// X64-NEXT: [[ATOMIC_TEMP1:%.*]] = alloca x86_fp80, align 16
+// X64-NEXT: [[ATOMIC_TEMP2:%.*]] = alloca x86_fp80, align 16
+// X64-NEXT: [[ATOMIC_TEMP3:%.*]] = alloca x86_fp80, align 16
// X64-NEXT: store ptr [[ADDR]], ptr [[ADDR_ADDR]], align 8
// X64-NEXT: [[TMP0:%.*]] = load ptr, ptr [[ADDR_ADDR]], align 8
-// X64-NEXT: [[TMP1:%.*]] = atomicrmw fadd ptr [[TMP0]], float 1.000000e+00 seq_cst, align 16
-// X64-NEXT: [[TMP2:%.*]] = fadd float [[TMP1]], 1.000000e+00
-// X64-NEXT: store float [[TMP2]], ptr [[RETVAL]], align 16
-// X64-NEXT: [[TMP3:%.*]] = load x86_fp80, ptr [[RETVAL]], align 16
-// X64-NEXT: ret x86_fp80 [[TMP3]]
+// X64-NEXT: [[ATOMIC_LOAD:%.*]] = load atomic volatile i128, ptr [[TMP0]] seq_cst, align 16
+// X64-NEXT: store i128 [[ATOMIC_LOAD]], ptr [[ATOMIC_TEMP]], align 16
+// X64-NEXT: [[TMP1:%.*]] = load x86_fp80, ptr [[ATOMIC_TEMP]], align 16
+// X64-NEXT: br label %[[ATOMIC_OP:.*]]
+// X64: [[ATOMIC_OP]]:
+// X64-NEXT: [[TMP2:%.*]] = phi x86_fp80 [ [[TMP1]], %[[ENTRY]] ], [ [[TMP8:%.*]], %[[ATOMIC_OP]] ]
+// X64-NEXT: [[INC:%.*]] = fadd x86_fp80 [[TMP2]], 0xK3FFF8000000000000000
+// X64-NEXT: call void @llvm.memset.p0.i64(ptr align 16 [[ATOMIC_TEMP1]], i8 0, i64 16, i1 false)
+// X64-NEXT: store x86_fp80 [[TMP2]], ptr [[ATOMIC_TEMP1]], align 16
+// X64-NEXT: [[TMP3:%...
[truncated]
|
|
Hi, since we are wrapping up LLVM 19.1.0 we are very strict with the fixes we pick at this point. Can you please respond to the following questions to help me understand if this has to be included in the final release or not. Is this PR a fix for a regression or a critical issue? What is the risk of accepting this into the release branch? What is the risk of NOT accepting this into the release branch? |
|
@RKSimon ? |
) `llvm::ConstantFP::get(llvm::LLVMContext&, APFloat(float))` always returns a f32 constant. Fix llvm#107054.
|
@dtcxzyw (or anyone else). If you would like to add a note about this fix in the release notes (completely optional). Please reply to this comment with a one or two sentence description of the fix. When you are done, please add the release:note label to this PR. |
Backport 9fef09f.