From d7c4504964b63d6d7b04d8d4b1dd0c8b39533f8d Mon Sep 17 00:00:00 2001 From: Simon Pilgrim Date: Mon, 18 Sep 2017 17:58:31 +0000 Subject: [PATCH] [X86][AVX] Improve (i8 bitcast (v8i1 x)) handling for 256-bit vector compare results. As commented on D37849, AVX1 targets were missing a chance to use vmovmskps for v8f32/v8i32 results for bool vector bitcasts git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@313547 91177308-0d34-0410-b5e6-96231b3b80d8 --- lib/Target/X86/X86ISelLowering.cpp | 2 +- test/CodeGen/X86/bitcast-setcc-256.ll | 30 ++++++++------------------- 2 files changed, 10 insertions(+), 22 deletions(-) diff --git a/lib/Target/X86/X86ISelLowering.cpp b/lib/Target/X86/X86ISelLowering.cpp index 8a442d56385..ea9a62434a8 100644 --- a/lib/Target/X86/X86ISelLowering.cpp +++ b/lib/Target/X86/X86ISelLowering.cpp @@ -29392,7 +29392,7 @@ static SDValue combineBitcastvxi1(SelectionDAG &DAG, SDValue BitCast, // the compare. if (N0->getOpcode() == ISD::SETCC && N0->getOperand(0)->getValueType(0).is256BitVector() && - Subtarget.hasInt256()) { + Subtarget.hasAVX()) { SExtVT = MVT::v8i32; FPCastVT = MVT::v8f32; } diff --git a/test/CodeGen/X86/bitcast-setcc-256.ll b/test/CodeGen/X86/bitcast-setcc-256.ll index 6da258114a5..92fba2674a9 100644 --- a/test/CodeGen/X86/bitcast-setcc-256.ll +++ b/test/CodeGen/X86/bitcast-setcc-256.ll @@ -77,9 +77,8 @@ define i8 @v8i32(<8 x i32> %a, <8 x i32> %b) { ; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm3 ; AVX1-NEXT: vpcmpgtd %xmm2, %xmm3, %xmm2 ; AVX1-NEXT: vpcmpgtd %xmm1, %xmm0, %xmm0 -; AVX1-NEXT: vpacksswb %xmm2, %xmm0, %xmm0 -; AVX1-NEXT: vpshufb {{.*#+}} xmm0 = xmm0[0,2,4,6,8,10,12,14,u,u,u,u,u,u,u,u] -; AVX1-NEXT: vpmovmskb %xmm0, %eax +; AVX1-NEXT: vinsertf128 $1, %xmm2, %ymm0, %ymm0 +; AVX1-NEXT: vmovmskps %ymm0, %eax ; AVX1-NEXT: # kill: %AL %AL %EAX ; AVX1-NEXT: vzeroupper ; AVX1-NEXT: retq @@ -126,24 +125,13 @@ define i8 @v8f32(<8 x float> %a, <8 x float> %b) { ; SSSE3-NEXT: # kill: %AL %AL %EAX ; SSSE3-NEXT: retq ; -; AVX1-LABEL: v8f32: -; AVX1: # BB#0: -; AVX1-NEXT: vcmpltps %ymm0, %ymm1, %ymm0 -; AVX1-NEXT: vextractf128 $1, %ymm0, %xmm1 -; AVX1-NEXT: vpacksswb %xmm1, %xmm0, %xmm0 -; AVX1-NEXT: vpshufb {{.*#+}} xmm0 = xmm0[0,2,4,6,8,10,12,14,u,u,u,u,u,u,u,u] -; AVX1-NEXT: vpmovmskb %xmm0, %eax -; AVX1-NEXT: # kill: %AL %AL %EAX -; AVX1-NEXT: vzeroupper -; AVX1-NEXT: retq -; -; AVX2-LABEL: v8f32: -; AVX2: # BB#0: -; AVX2-NEXT: vcmpltps %ymm0, %ymm1, %ymm0 -; AVX2-NEXT: vmovmskps %ymm0, %eax -; AVX2-NEXT: # kill: %AL %AL %EAX -; AVX2-NEXT: vzeroupper -; AVX2-NEXT: retq +; AVX12-LABEL: v8f32: +; AVX12: # BB#0: +; AVX12-NEXT: vcmpltps %ymm0, %ymm1, %ymm0 +; AVX12-NEXT: vmovmskps %ymm0, %eax +; AVX12-NEXT: # kill: %AL %AL %EAX +; AVX12-NEXT: vzeroupper +; AVX12-NEXT: retq ; ; AVX512-LABEL: v8f32: ; AVX512: # BB#0: -- 2.40.0