From 65de13b975f0b05335481d779d893defedf41b53 Mon Sep 17 00:00:00 2001 From: Matt Arsenault Date: Tue, 11 Jun 2019 01:35:07 +0000 Subject: [PATCH] AtomicExpand: Don't crash on non-0 alloca This now produces garbage on AMDGPU with a call to an nonexistent, anonymous libcall but won't assert. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@363022 91177308-0d34-0410-b5e6-96231b3b80d8 --- lib/CodeGen/AtomicExpandPass.cpp | 8 +++- lib/Target/AMDGPU/AMDGPUISelLowering.cpp | 1 + .../AtomicExpand/AMDGPU/unaligned-atomic.ll | 37 +++++++++++++++++++ 3 files changed, 44 insertions(+), 2 deletions(-) create mode 100644 test/Transforms/AtomicExpand/AMDGPU/unaligned-atomic.ll diff --git a/lib/CodeGen/AtomicExpandPass.cpp b/lib/CodeGen/AtomicExpandPass.cpp index 7aca67a327b..dc7eaf6a5fe 100644 --- a/lib/CodeGen/AtomicExpandPass.cpp +++ b/lib/CodeGen/AtomicExpandPass.cpp @@ -1712,8 +1712,11 @@ bool AtomicExpand::expandAtomicOpToLibcall( if (CASExpected) { AllocaCASExpected = AllocaBuilder.CreateAlloca(CASExpected->getType()); AllocaCASExpected->setAlignment(AllocaAlignment); + unsigned AllocaAS = AllocaCASExpected->getType()->getPointerAddressSpace(); + AllocaCASExpected_i8 = - Builder.CreateBitCast(AllocaCASExpected, Type::getInt8PtrTy(Ctx)); + Builder.CreateBitCast(AllocaCASExpected, + Type::getInt8PtrTy(Ctx, AllocaAS)); Builder.CreateLifetimeStart(AllocaCASExpected_i8, SizeVal64); Builder.CreateAlignedStore(CASExpected, AllocaCASExpected, AllocaAlignment); Args.push_back(AllocaCASExpected_i8); @@ -1740,8 +1743,9 @@ bool AtomicExpand::expandAtomicOpToLibcall( if (!CASExpected && HasResult && !UseSizedLibcall) { AllocaResult = AllocaBuilder.CreateAlloca(I->getType()); AllocaResult->setAlignment(AllocaAlignment); + unsigned AllocaAS = AllocaResult->getType()->getPointerAddressSpace(); AllocaResult_i8 = - Builder.CreateBitCast(AllocaResult, Type::getInt8PtrTy(Ctx)); + Builder.CreateBitCast(AllocaResult, Type::getInt8PtrTy(Ctx, AllocaAS)); Builder.CreateLifetimeStart(AllocaResult_i8, SizeVal64); Args.push_back(AllocaResult_i8); } diff --git a/lib/Target/AMDGPU/AMDGPUISelLowering.cpp b/lib/Target/AMDGPU/AMDGPUISelLowering.cpp index 64e710fba7f..cdb406d27ad 100644 --- a/lib/Target/AMDGPU/AMDGPUISelLowering.cpp +++ b/lib/Target/AMDGPU/AMDGPUISelLowering.cpp @@ -524,6 +524,7 @@ AMDGPUTargetLowering::AMDGPUTargetLowering(const TargetMachine &TM, setHasMultipleConditionRegisters(true); setMinCmpXchgSizeInBits(32); + setSupportsUnalignedAtomics(false); PredictableSelectIsExpensive = false; diff --git a/test/Transforms/AtomicExpand/AMDGPU/unaligned-atomic.ll b/test/Transforms/AtomicExpand/AMDGPU/unaligned-atomic.ll new file mode 100644 index 00000000000..3d21e151001 --- /dev/null +++ b/test/Transforms/AtomicExpand/AMDGPU/unaligned-atomic.ll @@ -0,0 +1,37 @@ +; NOTE: Assertions have been autogenerated by utils/update_test_checks.py +; RUN: opt -S -mtriple=amdgcn-amd-amdhsa -mcpu=gfx900 -atomic-expand %s | FileCheck -check-prefix=GCN %s + +; FIXME: This should not introduce a libcall, much less one to an +; anonymous function. + +define i32 @atomic_load_global_align1(i32 addrspace(1)* %ptr) { +; GCN-LABEL: @atomic_load_global_align1( +; GCN-NEXT: [[TMP1:%.*]] = bitcast i32 addrspace(1)* [[PTR:%.*]] to i8 addrspace(1)* +; GCN-NEXT: [[TMP2:%.*]] = addrspacecast i8 addrspace(1)* [[TMP1]] to i8* +; GCN-NEXT: [[TMP3:%.*]] = alloca i32, align 4 +; GCN-NEXT: [[TMP4:%.*]] = bitcast i32* [[TMP3]] to i8* +; GCN-NEXT: call void @llvm.lifetime.start.p0i8(i64 4, i8* [[TMP4]]) +; GCN-NEXT: call void @0(i64 4, i8* [[TMP2]], i8* [[TMP4]], i32 5) +; GCN-NEXT: [[TMP5:%.*]] = load i32, i32* [[TMP3]], align 4 +; GCN-NEXT: call void @llvm.lifetime.end.p0i8(i64 4, i8* [[TMP4]]) +; GCN-NEXT: ret i32 [[TMP5]] +; + %val = load atomic i32, i32 addrspace(1)* %ptr seq_cst, align 1 + ret i32 %val +} + +define void @atomic_store_global_align1(i32 addrspace(1)* %ptr, i32 %val) { +; GCN-LABEL: @atomic_store_global_align1( +; GCN-NEXT: [[TMP1:%.*]] = bitcast i32 addrspace(1)* [[PTR:%.*]] to i8 addrspace(1)* +; GCN-NEXT: [[TMP2:%.*]] = addrspacecast i8 addrspace(1)* [[TMP1]] to i8* +; GCN-NEXT: [[TMP3:%.*]] = alloca i32, align 4 +; GCN-NEXT: [[TMP4:%.*]] = bitcast i32* [[TMP3]] to i8* +; GCN-NEXT: call void @llvm.lifetime.start.p0i8(i64 4, i8* [[TMP4]]) +; GCN-NEXT: store i32 [[VAL:%.*]], i32* [[TMP3]], align 4 +; GCN-NEXT: call void @1(i64 4, i8* [[TMP2]], i8* [[TMP4]], i32 0) +; GCN-NEXT: call void @llvm.lifetime.end.p0i8(i64 4, i8* [[TMP4]]) +; GCN-NEXT: ret void +; + store atomic i32 %val, i32 addrspace(1)* %ptr monotonic, align 1 + ret void +} -- 2.50.1