; SSE2-NEXT: andps %xmm3, %xmm2
; SSE2-NEXT: movaps %xmm0, %xmm4
; SSE2-NEXT: shufps {{.*#+}} xmm4 = xmm4[3,1,2,3]
-; SSE2-NEXT: movaps {{.*#+}} xmm5 = [nan,nan,nan,nan]
+; SSE2-NEXT: movaps {{.*#+}} xmm5
; SSE2-NEXT: andps %xmm5, %xmm4
; SSE2-NEXT: orps %xmm2, %xmm4
; SSE2-NEXT: movaps %xmm1, %xmm2
; AVX: # BB#0:
; AVX-NEXT: vmovaps {{.*#+}} xmm2 = [-0.000000e+00,-0.000000e+00,-0.000000e+00,-0.000000e+00]
; AVX-NEXT: vandps %xmm2, %xmm1, %xmm3
-; AVX-NEXT: vmovaps {{.*#+}} xmm4 = [nan,nan,nan,nan]
+; AVX-NEXT: vmovaps {{.*#+}} xmm4
; AVX-NEXT: vandps %xmm4, %xmm0, %xmm5
; AVX-NEXT: vorps %xmm3, %xmm5, %xmm3
; AVX-NEXT: vmovshdup {{.*#+}} xmm5 = xmm1[1,1,3,3]
; SSE2-NEXT: andps %xmm8, %xmm0
; SSE2-NEXT: movaps %xmm5, %xmm7
; SSE2-NEXT: shufps {{.*#+}} xmm7 = xmm7[3,1,2,3]
-; SSE2-NEXT: movaps {{.*#+}} xmm6 = [nan,nan,nan,nan]
+; SSE2-NEXT: movaps {{.*#+}} xmm6
; SSE2-NEXT: andps %xmm6, %xmm7
; SSE2-NEXT: orps %xmm0, %xmm7
; SSE2-NEXT: movaps %xmm2, %xmm0
; AVX-NEXT: vmovaps {{.*#+}} xmm2 = [-0.000000e+00,-0.000000e+00,-0.000000e+00,-0.000000e+00]
; AVX-NEXT: vandps %xmm2, %xmm4, %xmm5
; AVX-NEXT: vextractf128 $1, %ymm0, %xmm6
-; AVX-NEXT: vmovaps {{.*#+}} xmm3 = [nan,nan,nan,nan]
+; AVX-NEXT: vmovaps {{.*#+}} xmm3
; AVX-NEXT: vandps %xmm3, %xmm6, %xmm7
; AVX-NEXT: vorps %xmm5, %xmm7, %xmm8
; AVX-NEXT: vmovshdup {{.*#+}} xmm7 = xmm4[1,1,3,3]
; SSE2-NEXT: movaps {{.*#+}} xmm3 = [-0.000000e+00,-0.000000e+00]
; SSE2-NEXT: movaps %xmm1, %xmm4
; SSE2-NEXT: andps %xmm3, %xmm4
-; SSE2-NEXT: movaps {{.*#+}} xmm5 = [nan,nan]
+; SSE2-NEXT: movaps {{.*#+}} xmm5
; SSE2-NEXT: movaps %xmm0, %xmm2
; SSE2-NEXT: andps %xmm5, %xmm2
; SSE2-NEXT: orps %xmm4, %xmm2
; AVX: # BB#0:
; AVX-NEXT: vmovapd {{.*#+}} xmm2 = [-0.000000e+00,-0.000000e+00]
; AVX-NEXT: vandpd %xmm2, %xmm1, %xmm3
-; AVX-NEXT: vmovapd {{.*#+}} xmm4 = [nan,nan]
+; AVX-NEXT: vmovapd {{.*#+}} xmm4
; AVX-NEXT: vandpd %xmm4, %xmm0, %xmm5
; AVX-NEXT: vorpd %xmm3, %xmm5, %xmm3
; AVX-NEXT: vpermilpd {{.*#+}} xmm1 = xmm1[1,0]
; SSE2-NEXT: movaps {{.*#+}} xmm5 = [-0.000000e+00,-0.000000e+00]
; SSE2-NEXT: movaps %xmm2, %xmm6
; SSE2-NEXT: andps %xmm5, %xmm6
-; SSE2-NEXT: movaps {{.*#+}} xmm7 = [nan,nan]
+; SSE2-NEXT: movaps {{.*#+}} xmm7
; SSE2-NEXT: andps %xmm7, %xmm0
; SSE2-NEXT: orps %xmm6, %xmm0
; SSE2-NEXT: movhlps {{.*#+}} xmm2 = xmm2[1,1]
; AVX-NEXT: vmovapd {{.*#+}} xmm3 = [-0.000000e+00,-0.000000e+00]
; AVX-NEXT: vandpd %xmm3, %xmm2, %xmm4
; AVX-NEXT: vextractf128 $1, %ymm0, %xmm5
-; AVX-NEXT: vmovapd {{.*#+}} xmm6 = [nan,nan]
+; AVX-NEXT: vmovapd {{.*#+}} xmm6
; AVX-NEXT: vandpd %xmm6, %xmm5, %xmm7
; AVX-NEXT: vorpd %xmm4, %xmm7, %xmm4
; AVX-NEXT: vpermilpd {{.*#+}} xmm2 = xmm2[1,0]