X86vfpround, "{1to2}", "{x}">, EVEX_V128;
defm Z256 : avx512_vcvt_fp<opc, OpcodeStr, v4f32x_info, v4f64x_info, fpround,
"{1to4}", "{y}">, EVEX_V256;
+
+ def : InstAlias<OpcodeStr##"x\t{$src, $dst|$dst, $src}",
+ (!cast<Instruction>(NAME # "Z128rr") VR128X:$dst, VR128X:$src), 0>;
+ def : InstAlias<OpcodeStr##"x\t{$src, $dst|$dst, $src}",
+ (!cast<Instruction>(NAME # "Z128rm") VR128X:$dst, f128mem:$src), 0>;
+ def : InstAlias<OpcodeStr##"y\t{$src, $dst|$dst, $src}",
+ (!cast<Instruction>(NAME # "Z256rr") VR128X:$dst, VR256X:$src), 0>;
+ def : InstAlias<OpcodeStr##"y\t{$src, $dst|$dst, $src}",
+ (!cast<Instruction>(NAME # "Z256rm") VR128X:$dst, f256mem:$src), 0>;
}
}
OpNode128, "{1to2}", "{x}">, EVEX_V128;
defm Z256 : avx512_vcvt_fp<opc, OpcodeStr, v4i32x_info, v4f64x_info, OpNode,
"{1to4}", "{y}">, EVEX_V256;
+
+ def : InstAlias<OpcodeStr##"x\t{$src, $dst|$dst, $src}",
+ (!cast<Instruction>(NAME # "Z128rr") VR128X:$dst, VR128X:$src), 0>;
+ def : InstAlias<OpcodeStr##"x\t{$src, $dst|$dst, $src}",
+ (!cast<Instruction>(NAME # "Z128rm") VR128X:$dst, i128mem:$src), 0>;
+ def : InstAlias<OpcodeStr##"y\t{$src, $dst|$dst, $src}",
+ (!cast<Instruction>(NAME # "Z256rr") VR128X:$dst, VR256X:$src), 0>;
+ def : InstAlias<OpcodeStr##"y\t{$src, $dst|$dst, $src}",
+ (!cast<Instruction>(NAME # "Z256rm") VR128X:$dst, i256mem:$src), 0>;
}
}
"{1to2}", "{x}">, EVEX_V128;
defm Z256 : avx512_vcvt_fp<opc, OpcodeStr, v4i32x_info, v4f64x_info, OpNode,
"{1to4}", "{y}">, EVEX_V256;
+
+ def : InstAlias<OpcodeStr##"x\t{$src, $dst|$dst, $src}",
+ (!cast<Instruction>(NAME # "Z128rr") VR128X:$dst, VR128X:$src), 0>;
+ def : InstAlias<OpcodeStr##"x\t{$src, $dst|$dst, $src}",
+ (!cast<Instruction>(NAME # "Z128rm") VR128X:$dst, f128mem:$src), 0>;
+ def : InstAlias<OpcodeStr##"y\t{$src, $dst|$dst, $src}",
+ (!cast<Instruction>(NAME # "Z256rr") VR128X:$dst, VR256X:$src), 0>;
+ def : InstAlias<OpcodeStr##"y\t{$src, $dst|$dst, $src}",
+ (!cast<Instruction>(NAME # "Z256rm") VR128X:$dst, f256mem:$src), 0>;
}
}
"{1to2}", "{x}">, EVEX_V128;
defm Z256 : avx512_vcvt_fp<opc, OpcodeStr, v4f32x_info, v4i64x_info, OpNode,
"{1to4}", "{y}">, EVEX_V256;
+
+ def : InstAlias<OpcodeStr##"x\t{$src, $dst|$dst, $src}",
+ (!cast<Instruction>(NAME # "Z128rr") VR128X:$dst, VR128X:$src), 0>;
+ def : InstAlias<OpcodeStr##"x\t{$src, $dst|$dst, $src}",
+ (!cast<Instruction>(NAME # "Z128rm") VR128X:$dst, i128mem:$src), 0>;
+ def : InstAlias<OpcodeStr##"y\t{$src, $dst|$dst, $src}",
+ (!cast<Instruction>(NAME # "Z256rr") VR128X:$dst, VR256X:$src), 0>;
+ def : InstAlias<OpcodeStr##"y\t{$src, $dst|$dst, $src}",
+ (!cast<Instruction>(NAME # "Z256rm") VR128X:$dst, i256mem:$src), 0>;
}
}
{ X86::VCVTSS2SIrr, X86::VCVTSS2SIrm, 0 },
{ X86::VCVTDQ2PDrr, X86::VCVTDQ2PDrm, 0 },
{ X86::VCVTDQ2PSrr, X86::VCVTDQ2PSrm, 0 },
- { X86::VCVTPD2DQrr, X86::VCVTPD2DQXrm, 0 },
- { X86::VCVTPD2PSrr, X86::VCVTPD2PSXrm, 0 },
+ { X86::VCVTPD2DQrr, X86::VCVTPD2DQrm, 0 },
+ { X86::VCVTPD2PSrr, X86::VCVTPD2PSrm, 0 },
{ X86::VCVTPS2DQrr, X86::VCVTPS2DQrm, 0 },
{ X86::VCVTPS2PDrr, X86::VCVTPS2PDrm, 0 },
- { X86::VCVTTPD2DQrr, X86::VCVTTPD2DQXrm, 0 },
+ { X86::VCVTTPD2DQrr, X86::VCVTTPD2DQrm, 0 },
{ X86::VCVTTPS2DQrr, X86::VCVTTPS2DQrm, 0 },
{ X86::VMOV64toPQIrr, X86::VMOVQI2PQIrm, 0 },
{ X86::VMOV64toSDrr, X86::VMOV64toSDrm, 0 },
// XMM only
def : InstAlias<"vcvtpd2dqx\t{$src, $dst|$dst, $src}",
(VCVTPD2DQrr VR128:$dst, VR128:$src), 0>;
-def VCVTPD2DQXrm : SDI<0xE6, MRMSrcMem, (outs VR128:$dst), (ins f128mem:$src),
- "vcvtpd2dqx\t{$src, $dst|$dst, $src}",
- [(set VR128:$dst,
- (v4i32 (X86cvtp2Int (loadv2f64 addr:$src))))]>, VEX,
- Sched<[WriteCvtF2ILd]>;
+def VCVTPD2DQrm : SDI<0xE6, MRMSrcMem, (outs VR128:$dst), (ins f128mem:$src),
+ "vcvtpd2dq{x}\t{$src, $dst|$dst, $src}",
+ [(set VR128:$dst,
+ (v4i32 (X86cvtp2Int (loadv2f64 addr:$src))))]>, VEX,
+ Sched<[WriteCvtF2ILd]>;
+def : InstAlias<"vcvtpd2dqx\t{$src, $dst|$dst, $src}",
+ (VCVTPD2DQrm VR128:$dst, f128mem:$src), 0>;
// YMM only
def VCVTPD2DQYrr : SDI<0xE6, MRMSrcReg, (outs VR128:$dst), (ins VR256:$src),
- "vcvtpd2dq{y}\t{$src, $dst|$dst, $src}",
+ "vcvtpd2dq\t{$src, $dst|$dst, $src}",
[(set VR128:$dst,
(v4i32 (X86cvtp2Int (v4f64 VR256:$src))))]>,
VEX, VEX_L, Sched<[WriteCvtF2I]>;
[(set VR128:$dst,
(v4i32 (X86cvtp2Int (loadv4f64 addr:$src))))]>,
VEX, VEX_L, Sched<[WriteCvtF2ILd]>;
-def : InstAlias<"vcvtpd2dq\t{$src, $dst|$dst, $src}",
+def : InstAlias<"vcvtpd2dqy\t{$src, $dst|$dst, $src}",
(VCVTPD2DQYrr VR128:$dst, VR256:$src), 0>;
+def : InstAlias<"vcvtpd2dqy\t{$src, $dst|$dst, $src}",
+ (VCVTPD2DQYrm VR128:$dst, f256mem:$src), 0>;
}
def CVTPD2DQrm : SDI<0xE6, MRMSrcMem, (outs VR128:$dst), (ins f128mem:$src),
def : InstAlias<"vcvttpd2dqx\t{$src, $dst|$dst, $src}",
(VCVTTPD2DQrr VR128:$dst, VR128:$src), 0>;
let Predicates = [HasAVX, NoVLX] in
-def VCVTTPD2DQXrm : VPDI<0xE6, MRMSrcMem, (outs VR128:$dst), (ins f128mem:$src),
- "cvttpd2dqx\t{$src, $dst|$dst, $src}",
- [(set VR128:$dst,
- (v4i32 (X86cvttpd2dq (loadv2f64 addr:$src))))],
- IIC_SSE_CVT_PD_RM>, VEX, Sched<[WriteCvtF2ILd]>;
+def VCVTTPD2DQrm : VPDI<0xE6, MRMSrcMem, (outs VR128:$dst), (ins f128mem:$src),
+ "cvttpd2dq{x}\t{$src, $dst|$dst, $src}",
+ [(set VR128:$dst,
+ (v4i32 (X86cvttpd2dq (loadv2f64 addr:$src))))],
+ IIC_SSE_CVT_PD_RM>, VEX, Sched<[WriteCvtF2ILd]>;
+def : InstAlias<"vcvttpd2dqx\t{$src, $dst|$dst, $src}",
+ (VCVTTPD2DQrm VR128:$dst, f128mem:$src), 0>;
// YMM only
let Predicates = [HasAVX, NoVLX] in {
def VCVTTPD2DQYrr : VPDI<0xE6, MRMSrcReg, (outs VR128:$dst), (ins VR256:$src),
- "cvttpd2dq{y}\t{$src, $dst|$dst, $src}",
+ "cvttpd2dq\t{$src, $dst|$dst, $src}",
[(set VR128:$dst,
(v4i32 (fp_to_sint (v4f64 VR256:$src))))],
IIC_SSE_CVT_PD_RR>, VEX, VEX_L, Sched<[WriteCvtF2I]>;
(v4i32 (fp_to_sint (loadv4f64 addr:$src))))],
IIC_SSE_CVT_PD_RM>, VEX, VEX_L, Sched<[WriteCvtF2ILd]>;
}
-def : InstAlias<"vcvttpd2dq\t{$src, $dst|$dst, $src}",
+def : InstAlias<"vcvttpd2dqy\t{$src, $dst|$dst, $src}",
(VCVTTPD2DQYrr VR128:$dst, VR256:$src), 0>;
+def : InstAlias<"vcvttpd2dqy\t{$src, $dst|$dst, $src}",
+ (VCVTTPD2DQYrm VR128:$dst, f256mem:$src), 0>;
let Predicates = [HasAVX, NoVLX] in {
let AddedComplexity = 15 in
def : InstAlias<"vcvtpd2psx\t{$src, $dst|$dst, $src}",
(VCVTPD2PSrr VR128:$dst, VR128:$src), 0>;
let Predicates = [HasAVX, NoVLX] in
-def VCVTPD2PSXrm : VPDI<0x5A, MRMSrcMem, (outs VR128:$dst), (ins f128mem:$src),
- "cvtpd2psx\t{$src, $dst|$dst, $src}",
- [(set VR128:$dst, (X86vfpround (loadv2f64 addr:$src)))],
- IIC_SSE_CVT_PD_RM>, VEX, Sched<[WriteCvtF2FLd]>;
+def VCVTPD2PSrm : VPDI<0x5A, MRMSrcMem, (outs VR128:$dst), (ins f128mem:$src),
+ "cvtpd2ps{x}\t{$src, $dst|$dst, $src}",
+ [(set VR128:$dst, (X86vfpround (loadv2f64 addr:$src)))],
+ IIC_SSE_CVT_PD_RM>, VEX, Sched<[WriteCvtF2FLd]>;
+def : InstAlias<"vcvtpd2psx\t{$src, $dst|$dst, $src}",
+ (VCVTPD2PSrm VR128:$dst, f128mem:$src), 0>;
// YMM only
let Predicates = [HasAVX, NoVLX] in {
def VCVTPD2PSYrr : VPDI<0x5A, MRMSrcReg, (outs VR128:$dst), (ins VR256:$src),
- "cvtpd2ps{y}\t{$src, $dst|$dst, $src}",
+ "cvtpd2ps\t{$src, $dst|$dst, $src}",
[(set VR128:$dst, (fpround VR256:$src))],
IIC_SSE_CVT_PD_RR>, VEX, VEX_L, Sched<[WriteCvtF2F]>;
def VCVTPD2PSYrm : VPDI<0x5A, MRMSrcMem, (outs VR128:$dst), (ins f256mem:$src),
[(set VR128:$dst, (fpround (loadv4f64 addr:$src)))],
IIC_SSE_CVT_PD_RM>, VEX, VEX_L, Sched<[WriteCvtF2FLd]>;
}
-def : InstAlias<"vcvtpd2ps\t{$src, $dst|$dst, $src}",
+def : InstAlias<"vcvtpd2psy\t{$src, $dst|$dst, $src}",
(VCVTPD2PSYrr VR128:$dst, VR256:$src), 0>;
+def : InstAlias<"vcvtpd2psy\t{$src, $dst|$dst, $src}",
+ (VCVTPD2PSYrm VR128:$dst, f256mem:$src), 0>;
def CVTPD2PSrr : PDI<0x5A, MRMSrcReg, (outs VR128:$dst), (ins VR128:$src),
"cvtpd2ps\t{$src, $dst|$dst, $src}",
define <4 x i32> @fptosi01(<4 x double> %a) {
; CHECK-LABEL: fptosi01:
; CHECK: # BB#0:
-; CHECK-NEXT: vcvttpd2dqy %ymm0, %xmm0
+; CHECK-NEXT: vcvttpd2dq %ymm0, %xmm0
; CHECK-NEXT: vzeroupper
; CHECK-NEXT: retq
%b = fptosi <4 x double> %a to <4 x i32>
define <8 x float> @fptrunc00(<8 x double> %b) nounwind {
; CHECK-LABEL: fptrunc00:
; CHECK: # BB#0:
-; CHECK-NEXT: vcvtpd2psy %ymm0, %xmm0
-; CHECK-NEXT: vcvtpd2psy %ymm1, %xmm1
+; CHECK-NEXT: vcvtpd2ps %ymm0, %xmm0
+; CHECK-NEXT: vcvtpd2ps %ymm1, %xmm1
; CHECK-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
; CHECK-NEXT: retq
%a = fptrunc <8 x double> %b to <8 x float>
;; Check that FP_TO_SINT and FP_TO_UINT generate convert with truncate
; CHECK-LABEL: test1:
-; CHECK: vcvttpd2dqy
+; CHECK: vcvttpd2dq
; CHECK: ret
; CHECK-LABEL: test2:
-; CHECK: vcvttpd2dqy
+; CHECK: vcvttpd2dq
; CHECK: ret
define <4 x i8> @test1(<4 x double> %d) {
define <2 x i64> @test_mm256_cvtpd_epi32(<4 x double> %a0) nounwind {
; X32-LABEL: test_mm256_cvtpd_epi32:
; X32: # BB#0:
-; X32-NEXT: vcvtpd2dqy %ymm0, %xmm0
+; X32-NEXT: vcvtpd2dq %ymm0, %xmm0
; X32-NEXT: vzeroupper
; X32-NEXT: retl
;
; X64-LABEL: test_mm256_cvtpd_epi32:
; X64: # BB#0:
-; X64-NEXT: vcvtpd2dqy %ymm0, %xmm0
+; X64-NEXT: vcvtpd2dq %ymm0, %xmm0
; X64-NEXT: vzeroupper
; X64-NEXT: retq
%cvt = call <4 x i32> @llvm.x86.avx.cvt.pd2dq.256(<4 x double> %a0)
define <4 x float> @test_mm256_cvtpd_ps(<4 x double> %a0) nounwind {
; X32-LABEL: test_mm256_cvtpd_ps:
; X32: # BB#0:
-; X32-NEXT: vcvtpd2psy %ymm0, %xmm0
+; X32-NEXT: vcvtpd2ps %ymm0, %xmm0
; X32-NEXT: vzeroupper
; X32-NEXT: retl
;
; X64-LABEL: test_mm256_cvtpd_ps:
; X64: # BB#0:
-; X64-NEXT: vcvtpd2psy %ymm0, %xmm0
+; X64-NEXT: vcvtpd2ps %ymm0, %xmm0
; X64-NEXT: vzeroupper
; X64-NEXT: retq
%res = call <4 x float> @llvm.x86.avx.cvt.pd2.ps.256(<4 x double> %a0)
define <2 x i64> @test_mm256_cvttpd_epi32(<4 x double> %a0) nounwind {
; X32-LABEL: test_mm256_cvttpd_epi32:
; X32: # BB#0:
-; X32-NEXT: vcvttpd2dqy %ymm0, %xmm0
+; X32-NEXT: vcvttpd2dq %ymm0, %xmm0
; X32-NEXT: vzeroupper
; X32-NEXT: retl
;
; X64-LABEL: test_mm256_cvttpd_epi32:
; X64: # BB#0:
-; X64-NEXT: vcvttpd2dqy %ymm0, %xmm0
+; X64-NEXT: vcvttpd2dq %ymm0, %xmm0
; X64-NEXT: vzeroupper
; X64-NEXT: retq
%cvt = call <4 x i32> @llvm.x86.avx.cvtt.pd2dq.256(<4 x double> %a0)
define <4 x float> @test_x86_avx_cvt_pd2_ps_256(<4 x double> %a0) {
; AVX-LABEL: test_x86_avx_cvt_pd2_ps_256:
; AVX: ## BB#0:
-; AVX-NEXT: vcvtpd2psy %ymm0, %xmm0 ## encoding: [0xc5,0xfd,0x5a,0xc0]
+; AVX-NEXT: vcvtpd2ps %ymm0, %xmm0 ## encoding: [0xc5,0xfd,0x5a,0xc0]
; AVX-NEXT: vzeroupper ## encoding: [0xc5,0xf8,0x77]
; AVX-NEXT: retl ## encoding: [0xc3]
;
define <4 x i32> @test_x86_avx_cvt_pd2dq_256(<4 x double> %a0) {
; AVX-LABEL: test_x86_avx_cvt_pd2dq_256:
; AVX: ## BB#0:
-; AVX-NEXT: vcvtpd2dqy %ymm0, %xmm0 ## encoding: [0xc5,0xff,0xe6,0xc0]
+; AVX-NEXT: vcvtpd2dq %ymm0, %xmm0 ## encoding: [0xc5,0xff,0xe6,0xc0]
; AVX-NEXT: vzeroupper ## encoding: [0xc5,0xf8,0x77]
; AVX-NEXT: retl ## encoding: [0xc3]
;
define <4 x i32> @test_x86_avx_cvtt_pd2dq_256(<4 x double> %a0) {
; AVX-LABEL: test_x86_avx_cvtt_pd2dq_256:
; AVX: ## BB#0:
-; AVX-NEXT: vcvttpd2dqy %ymm0, %xmm0 ## encoding: [0xc5,0xfd,0xe6,0xc0]
+; AVX-NEXT: vcvttpd2dq %ymm0, %xmm0 ## encoding: [0xc5,0xfd,0xe6,0xc0]
; AVX-NEXT: vzeroupper ## encoding: [0xc5,0xf8,0x77]
; AVX-NEXT: retl ## encoding: [0xc3]
;
}
define <4 x i32> @fptosi03(<4 x double> %a) {
-; KNL-LABEL: fptosi03:
-; KNL: ## BB#0:
-; KNL-NEXT: vcvttpd2dqy %ymm0, %xmm0
-; KNL-NEXT: retq
-;
-; SKX-LABEL: fptosi03:
-; SKX: ## BB#0:
-; SKX-NEXT: vcvttpd2dq %ymm0, %xmm0
-; SKX-NEXT: retq
+; ALL-LABEL: fptosi03:
+; ALL: ## BB#0:
+; ALL-NEXT: vcvttpd2dq %ymm0, %xmm0
+; ALL-NEXT: retq
%b = fptosi <4 x double> %a to <4 x i32>
ret <4 x i32> %b
}
}
define <4 x float> @fptrunc01(<4 x double> %b) {
-; KNL-LABEL: fptrunc01:
-; KNL: ## BB#0:
-; KNL-NEXT: vcvtpd2psy %ymm0, %xmm0
-; KNL-NEXT: retq
-;
-; SKX-LABEL: fptrunc01:
-; SKX: ## BB#0:
-; SKX-NEXT: vcvtpd2ps %ymm0, %xmm0
-; SKX-NEXT: retq
+; ALL-LABEL: fptrunc01:
+; ALL: ## BB#0:
+; ALL-NEXT: vcvtpd2ps %ymm0, %xmm0
+; ALL-NEXT: retq
%a = fptrunc <4 x double> %b to <4 x float>
ret <4 x float> %a
}
; KNL: ## BB#0:
; KNL-NEXT: vpslld $31, %xmm1, %xmm1
; KNL-NEXT: vpsrad $31, %xmm1, %xmm1
-; KNL-NEXT: vcvtpd2psy %ymm0, %xmm0
+; KNL-NEXT: vcvtpd2ps %ymm0, %xmm0
; KNL-NEXT: vpand %xmm0, %xmm1, %xmm0
; KNL-NEXT: retq
;
; AVX: # BB#0:
; AVX-NEXT: vbroadcastss {{.*}}(%rip), %xmm2
; AVX-NEXT: vandpd %xmm2, %xmm0, %xmm0
-; AVX-NEXT: vcvtpd2psy %ymm1, %xmm1
+; AVX-NEXT: vcvtpd2ps %ymm1, %xmm1
; AVX-NEXT: vbroadcastss {{.*}}(%rip), %xmm2
; AVX-NEXT: vandpd %xmm2, %xmm1, %xmm1
; AVX-NEXT: vorpd %xmm1, %xmm0, %xmm0
;
; AVX-LABEL: test3:
; AVX: # BB#0:
-; AVX-NEXT: vcvtpd2psy %ymm0, %xmm0
+; AVX-NEXT: vcvtpd2ps %ymm0, %xmm0
; AVX-NEXT: vzeroupper
; AVX-NEXT: retl
%y = fptrunc <4 x double> %x to <4 x float>
;
; AVX-LABEL: test4:
; AVX: # BB#0:
-; AVX-NEXT: vcvtpd2psy %ymm0, %xmm0
-; AVX-NEXT: vcvtpd2psy %ymm1, %xmm1
+; AVX-NEXT: vcvtpd2ps %ymm0, %xmm0
+; AVX-NEXT: vcvtpd2ps %ymm1, %xmm1
; AVX-NEXT: vinsertf128 $1, %xmm1, %ymm0, %ymm0
; AVX-NEXT: retl
%y = fptrunc <8 x double> %x to <8 x float>
; AVX-LABEL: fptosi_4f64_to_2i32:
; AVX: # BB#0:
; AVX-NEXT: # kill: %XMM0<def> %XMM0<kill> %YMM0<def>
-; AVX-NEXT: vcvttpd2dqy %ymm0, %xmm0
+; AVX-NEXT: vcvttpd2dq %ymm0, %xmm0
; AVX-NEXT: vzeroupper
; AVX-NEXT: retq
;
-; AVX512F-LABEL: fptosi_4f64_to_2i32:
-; AVX512F: # BB#0:
-; AVX512F-NEXT: # kill: %XMM0<def> %XMM0<kill> %YMM0<def>
-; AVX512F-NEXT: vcvttpd2dqy %ymm0, %xmm0
-; AVX512F-NEXT: retq
-;
-; AVX512VL-LABEL: fptosi_4f64_to_2i32:
-; AVX512VL: # BB#0:
-; AVX512VL-NEXT: # kill: %XMM0<def> %XMM0<kill> %YMM0<def>
-; AVX512VL-NEXT: vcvttpd2dq %ymm0, %xmm0
-; AVX512VL-NEXT: retq
-;
-; AVX512VLDQ-LABEL: fptosi_4f64_to_2i32:
-; AVX512VLDQ: # BB#0:
-; AVX512VLDQ-NEXT: # kill: %XMM0<def> %XMM0<kill> %YMM0<def>
-; AVX512VLDQ-NEXT: vcvttpd2dq %ymm0, %xmm0
-; AVX512VLDQ-NEXT: retq
+; AVX512-LABEL: fptosi_4f64_to_2i32:
+; AVX512: # BB#0:
+; AVX512-NEXT: # kill: %XMM0<def> %XMM0<kill> %YMM0<def>
+; AVX512-NEXT: vcvttpd2dq %ymm0, %xmm0
+; AVX512-NEXT: retq
%ext = shufflevector <2 x double> %a, <2 x double> undef, <4 x i32> <i32 0, i32 1, i32 undef, i32 undef>
%cvt = fptosi <4 x double> %ext to <4 x i32>
ret <4 x i32> %cvt
;
; AVX-LABEL: fptosi_4f64_to_4i32:
; AVX: # BB#0:
-; AVX-NEXT: vcvttpd2dqy %ymm0, %xmm0
+; AVX-NEXT: vcvttpd2dq %ymm0, %xmm0
; AVX-NEXT: vzeroupper
; AVX-NEXT: retq
;
-; AVX512F-LABEL: fptosi_4f64_to_4i32:
-; AVX512F: # BB#0:
-; AVX512F-NEXT: vcvttpd2dqy %ymm0, %xmm0
-; AVX512F-NEXT: retq
-;
-; AVX512VL-LABEL: fptosi_4f64_to_4i32:
-; AVX512VL: # BB#0:
-; AVX512VL-NEXT: vcvttpd2dq %ymm0, %xmm0
-; AVX512VL-NEXT: retq
-;
-; AVX512VLDQ-LABEL: fptosi_4f64_to_4i32:
-; AVX512VLDQ: # BB#0:
-; AVX512VLDQ-NEXT: vcvttpd2dq %ymm0, %xmm0
-; AVX512VLDQ-NEXT: retq
+; AVX512-LABEL: fptosi_4f64_to_4i32:
+; AVX512: # BB#0:
+; AVX512-NEXT: vcvttpd2dq %ymm0, %xmm0
+; AVX512-NEXT: retq
%cvt = fptosi <4 x double> %a to <4 x i32>
ret <4 x i32> %cvt
}
// CHECK: vpgatherqd xmm10, xmmword ptr [r15 + 2*ymm9], xmm8
// CHECK: encoding: [0xc4,0x02,0x3d,0x91,0x14,0x4f]
vpgatherqd xmm10, xmmword ptr [r15 + 2*ymm9], xmm8
+
+// CHECK: vcvtpd2ps xmm0, xmm15
+// CHECK: encoding: [0xc4,0xc1,0x79,0x5a,0xc7]
+ vcvtpd2ps xmm0, xmm15
+
+// CHECK: vcvtpd2ps xmm0, xmm15
+// CHECK: encoding: [0xc4,0xc1,0x79,0x5a,0xc7]
+ vcvtpd2psx xmm0, xmm15
+
+// CHECK: vcvtpd2ps xmm0, xmmword ptr [rax]
+// CHECK: encoding: [0xc5,0xf9,0x5a,0x00]
+ vcvtpd2ps xmm0, xmmword ptr [rax]
+
+// CHECK: vcvtpd2ps xmm0, xmmword ptr [rax]
+// CHECK: encoding: [0xc5,0xf9,0x5a,0x00]
+ vcvtpd2psx xmm0, xmmword ptr [rax]
+
+// CHECK: vcvtpd2ps xmm0, ymm15
+// CHECK: encoding: [0xc4,0xc1,0x7d,0x5a,0xc7]
+ vcvtpd2ps xmm0, ymm15
+
+// CHECK: vcvtpd2ps xmm0, ymm15
+// CHECK: encoding: [0xc4,0xc1,0x7d,0x5a,0xc7]
+ vcvtpd2psy xmm0, ymm15
+
+// CHECK: vcvtpd2ps xmm0, ymmword ptr [rax]
+// CHECK: encoding: [0xc5,0xfd,0x5a,0x00]
+ vcvtpd2ps xmm0, ymmword ptr [rax]
+
+// CHECK: vcvtpd2ps xmm0, ymmword ptr [rax]
+// CHECK: encoding: [0xc5,0xfd,0x5a,0x00]
+ vcvtpd2psy xmm0, ymmword ptr [rax]
+
+// CHECK: vcvtpd2dq xmm0, xmm15
+// CHECK: encoding: [0xc4,0xc1,0x7b,0xe6,0xc7]
+ vcvtpd2dq xmm0, xmm15
+
+// CHECK: vcvtpd2dq xmm0, xmm15
+// CHECK: encoding: [0xc4,0xc1,0x7b,0xe6,0xc7]
+ vcvtpd2dqx xmm0, xmm15
+
+// CHECK: vcvtpd2dq xmm0, xmmword ptr [rax]
+// CHECK: encoding: [0xc5,0xfb,0xe6,0x00]
+ vcvtpd2dq xmm0, xmmword ptr [rax]
+
+// CHECK: vcvtpd2dq xmm0, xmmword ptr [rax]
+// CHECK: encoding: [0xc5,0xfb,0xe6,0x00]
+ vcvtpd2dqx xmm0, xmmword ptr [rax]
+
+// CHECK: vcvtpd2dq xmm0, ymm15
+// CHECK: encoding: [0xc4,0xc1,0x7f,0xe6,0xc7]
+ vcvtpd2dq xmm0, ymm15
+
+// CHECK: vcvtpd2dq xmm0, ymm15
+// CHECK: encoding: [0xc4,0xc1,0x7f,0xe6,0xc7]
+ vcvtpd2dqy xmm0, ymm15
+
+// CHECK: vcvtpd2dq xmm0, ymmword ptr [rax]
+// CHECK: encoding: [0xc5,0xff,0xe6,0x00]
+ vcvtpd2dq xmm0, ymmword ptr [rax]
+
+// CHECK: vcvtpd2dq xmm0, ymmword ptr [rax]
+// CHECK: encoding: [0xc5,0xff,0xe6,0x00]
+ vcvtpd2dqy xmm0, ymmword ptr [rax]
+
+// CHECK: vcvttpd2dq xmm0, xmm15
+// CHECK: encoding: [0xc4,0xc1,0x79,0xe6,0xc7]
+ vcvttpd2dq xmm0, xmm15
+
+// CHECK: vcvttpd2dq xmm0, xmm15
+// CHECK: encoding: [0xc4,0xc1,0x79,0xe6,0xc7]
+ vcvttpd2dqx xmm0, xmm15
+
+// CHECK: vcvttpd2dq xmm0, xmmword ptr [rax]
+// CHECK: encoding: [0xc5,0xf9,0xe6,0x00]
+ vcvttpd2dq xmm0, xmmword ptr [rax]
+
+// CHECK: vcvttpd2dq xmm0, xmmword ptr [rax]
+// CHECK: encoding: [0xc5,0xf9,0xe6,0x00]
+ vcvttpd2dqx xmm0, xmmword ptr [rax]
+
+// CHECK: vcvttpd2dq xmm0, ymm15
+// CHECK: encoding: [0xc4,0xc1,0x7d,0xe6,0xc7]
+ vcvttpd2dq xmm0, ymm15
+
+// CHECK: vcvttpd2dq xmm0, ymm15
+// CHECK: encoding: [0xc4,0xc1,0x7d,0xe6,0xc7]
+ vcvttpd2dqy xmm0, ymm15
+
+// CHECK: vcvttpd2dq xmm0, ymmword ptr [rax]
+// CHECK: encoding: [0xc5,0xfd,0xe6,0x00]
+ vcvttpd2dq xmm0, ymmword ptr [rax]
+
+// CHECK: vcvttpd2dq xmm0, ymmword ptr [rax]
+// CHECK: encoding: [0xc5,0xfd,0xe6,0x00]
+ vcvttpd2dqy xmm0, ymmword ptr [rax]
-// RUN: llvm-mc -triple x86_64-unknown-unknown -mcpu=knl -mattr=+avx512vl -x86-asm-syntax=intel -output-asm-variant=1 --show-encoding %s | FileCheck %s
+// RUN: llvm-mc -triple x86_64-unknown-unknown -mcpu=knl -mattr=+avx512vl,+avx512dq -x86-asm-syntax=intel -output-asm-variant=1 --show-encoding %s | FileCheck %s
// CHECK: vcmppd k3, xmm27, xmm23, 171
// CHECK: encoding: [0x62,0xb1,0xa5,0x00,0xc2,0xdf,0xab]
// CHECK: encoding: [0x62,0xa2,0x7d,0x21,0xa3,0xbc,0xb9,0x00,0x04,0x00,0x00]
vscatterqps xmmword ptr [rcx + 4*ymm31 + 1024] {k1}, xmm23
+// CHECK: vcvtpd2ps xmm0, xmm23
+// CHECK: encoding: [0x62,0xb1,0xfd,0x08,0x5a,0xc7]
+ vcvtpd2ps xmm0, xmm23
+
+// CHECK: vcvtpd2ps xmm0, xmm23
+// CHECK: encoding: [0x62,0xb1,0xfd,0x08,0x5a,0xc7]
+ vcvtpd2psx xmm0, xmm23
+
+// CHECK: vcvtpd2ps xmm16, xmmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xfd,0x08,0x5a,0x00]
+ vcvtpd2ps xmm16, xmmword ptr [rax]
+
+// CHECK: vcvtpd2ps xmm16, xmmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xfd,0x08,0x5a,0x00]
+ vcvtpd2psx xmm16, xmmword ptr [rax]
+
+// CHECK: vcvtpd2ps xmm0, ymm23
+// CHECK: encoding: [0x62,0xb1,0xfd,0x28,0x5a,0xc7]
+ vcvtpd2ps xmm0, ymm23
+
+// CHECK: vcvtpd2ps xmm0, ymm23
+// CHECK: encoding: [0x62,0xb1,0xfd,0x28,0x5a,0xc7]
+ vcvtpd2psy xmm0, ymm23
+
+// CHECK: vcvtpd2ps xmm16, ymmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xfd,0x28,0x5a,0x00]
+ vcvtpd2ps xmm16, ymmword ptr [rax]
+
+// CHECK: vcvtpd2ps xmm16, ymmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xfd,0x28,0x5a,0x00]
+ vcvtpd2psy xmm16, ymmword ptr [rax]
+
+// CHECK: vcvtpd2dq xmm0, xmm23
+// CHECK: encoding: [0x62,0xb1,0xff,0x08,0xe6,0xc7]
+ vcvtpd2dq xmm0, xmm23
+
+// CHECK: vcvtpd2dq xmm0, xmm23
+// CHECK: encoding: [0x62,0xb1,0xff,0x08,0xe6,0xc7]
+ vcvtpd2dqx xmm0, xmm23
+
+// CHECK: vcvtpd2dq xmm16, xmmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xff,0x08,0xe6,0x00]
+ vcvtpd2dq xmm16, xmmword ptr [rax]
+
+// CHECK: vcvtpd2dq xmm16, xmmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xff,0x08,0xe6,0x00]
+ vcvtpd2dqx xmm16, xmmword ptr [rax]
+
+// CHECK: vcvtpd2dq xmm0, ymm23
+// CHECK: encoding: [0x62,0xb1,0xff,0x28,0xe6,0xc7]
+ vcvtpd2dq xmm0, ymm23
+
+// CHECK: vcvtpd2dq xmm0, ymm23
+// CHECK: encoding: [0x62,0xb1,0xff,0x28,0xe6,0xc7]
+ vcvtpd2dqy xmm0, ymm23
+
+// CHECK: vcvtpd2dq xmm16, ymmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xff,0x28,0xe6,0x00]
+ vcvtpd2dq xmm16, ymmword ptr [rax]
+
+// CHECK: vcvtpd2dq xmm16, ymmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xff,0x28,0xe6,0x00]
+ vcvtpd2dqy xmm16, ymmword ptr [rax]
+
+// CHECK: vcvtpd2udq xmm0, xmm23
+// CHECK: encoding: [0x62,0xb1,0xfc,0x08,0x79,0xc7]
+ vcvtpd2udq xmm0, xmm23
+
+// CHECK: vcvtpd2udq xmm0, xmm23
+// CHECK: encoding: [0x62,0xb1,0xfc,0x08,0x79,0xc7]
+ vcvtpd2udqx xmm0, xmm23
+
+// CHECK: vcvtpd2udq xmm16, xmmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xfc,0x08,0x79,0x00]
+ vcvtpd2udq xmm16, xmmword ptr [rax]
+
+// CHECK: vcvtpd2udq xmm16, xmmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xfc,0x08,0x79,0x00]
+ vcvtpd2udqx xmm16, xmmword ptr [rax]
+
+// CHECK: vcvtpd2udq xmm0, ymm23
+// CHECK: encoding: [0x62,0xb1,0xfc,0x28,0x79,0xc7]
+ vcvtpd2udq xmm0, ymm23
+
+// CHECK: vcvtpd2udq xmm0, ymm23
+// CHECK: encoding: [0x62,0xb1,0xfc,0x28,0x79,0xc7]
+ vcvtpd2udqy xmm0, ymm23
+
+// CHECK: vcvtpd2udq xmm16, ymmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xfc,0x28,0x79,0x00]
+ vcvtpd2udq xmm16, ymmword ptr [rax]
+
+// CHECK: vcvtpd2udq xmm16, ymmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xfc,0x28,0x79,0x00]
+ vcvtpd2udqy xmm16, ymmword ptr [rax]
+
+// CHECK: vcvttpd2dq xmm0, xmm23
+// CHECK: encoding: [0x62,0xb1,0xfd,0x08,0xe6,0xc7]
+ vcvttpd2dq xmm0, xmm23
+
+// CHECK: vcvttpd2dq xmm0, xmm23
+// CHECK: encoding: [0x62,0xb1,0xfd,0x08,0xe6,0xc7]
+ vcvttpd2dqx xmm0, xmm23
+
+// CHECK: vcvttpd2dq xmm16, xmmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xfd,0x08,0xe6,0x00]
+ vcvttpd2dq xmm16, xmmword ptr [rax]
+
+// CHECK: vcvttpd2dq xmm16, xmmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xfd,0x08,0xe6,0x00]
+ vcvttpd2dqx xmm16, xmmword ptr [rax]
+
+// CHECK: vcvttpd2dq xmm0, ymm23
+// CHECK: encoding: [0x62,0xb1,0xfd,0x28,0xe6,0xc7]
+ vcvttpd2dq xmm0, ymm23
+
+// CHECK: vcvttpd2dq xmm0, ymm23
+// CHECK: encoding: [0x62,0xb1,0xfd,0x28,0xe6,0xc7]
+ vcvttpd2dqy xmm0, ymm23
+
+// CHECK: vcvttpd2dq xmm16, ymmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xfd,0x28,0xe6,0x00]
+ vcvttpd2dq xmm16, ymmword ptr [rax]
+
+// CHECK: vcvttpd2dq xmm16, ymmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xfd,0x28,0xe6,0x00]
+ vcvttpd2dqy xmm16, ymmword ptr [rax]
+
+// CHECK: vcvttpd2udq xmm0, xmm23
+// CHECK: encoding: [0x62,0xb1,0xfc,0x08,0x78,0xc7]
+ vcvttpd2udq xmm0, xmm23
+
+// CHECK: vcvttpd2udq xmm0, xmm23
+// CHECK: encoding: [0x62,0xb1,0xfc,0x08,0x78,0xc7]
+ vcvttpd2udqx xmm0, xmm23
+
+// CHECK: vcvttpd2udq xmm16, xmmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xfc,0x08,0x78,0x00]
+ vcvttpd2udq xmm16, xmmword ptr [rax]
+
+// CHECK: vcvttpd2udq xmm16, xmmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xfc,0x08,0x78,0x00]
+ vcvttpd2udqx xmm16, xmmword ptr [rax]
+
+// CHECK: vcvttpd2udq xmm0, ymm23
+// CHECK: encoding: [0x62,0xb1,0xfc,0x28,0x78,0xc7]
+ vcvttpd2udq xmm0, ymm23
+
+// CHECK: vcvttpd2udq xmm0, ymm23
+// CHECK: encoding: [0x62,0xb1,0xfc,0x28,0x78,0xc7]
+ vcvttpd2udqy xmm0, ymm23
+
+// CHECK: vcvttpd2udq xmm16, ymmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xfc,0x28,0x78,0x00]
+ vcvttpd2udq xmm16, ymmword ptr [rax]
+
+// CHECK: vcvttpd2udq xmm16, ymmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xfc,0x28,0x78,0x00]
+ vcvttpd2udqy xmm16, ymmword ptr [rax]
+
+// CHECK: vcvtqq2ps xmm0, xmm23
+// CHECK: encoding: [0x62,0xb1,0xfc,0x08,0x5b,0xc7]
+ vcvtqq2ps xmm0, xmm23
+
+// CHECK: vcvtqq2ps xmm0, xmm23
+// CHECK: encoding: [0x62,0xb1,0xfc,0x08,0x5b,0xc7]
+ vcvtqq2psx xmm0, xmm23
+
+// CHECK: vcvtqq2ps xmm16, xmmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xfc,0x08,0x5b,0x00]
+ vcvtqq2ps xmm16, xmmword ptr [rax]
+
+// CHECK: vcvtqq2ps xmm16, xmmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xfc,0x08,0x5b,0x00]
+ vcvtqq2psx xmm16, xmmword ptr [rax]
+
+// CHECK: vcvtqq2ps xmm0, ymm23
+// CHECK: encoding: [0x62,0xb1,0xfc,0x28,0x5b,0xc7]
+ vcvtqq2ps xmm0, ymm23
+
+// CHECK: vcvtqq2ps xmm0, ymm23
+// CHECK: encoding: [0x62,0xb1,0xfc,0x28,0x5b,0xc7]
+ vcvtqq2psy xmm0, ymm23
+
+// CHECK: vcvtqq2ps xmm16, ymmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xfc,0x28,0x5b,0x00]
+ vcvtqq2ps xmm16, ymmword ptr [rax]
+
+// CHECK: vcvtqq2ps xmm16, ymmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xfc,0x28,0x5b,0x00]
+ vcvtqq2psy xmm16, ymmword ptr [rax]
+
+// CHECK: vcvtuqq2ps xmm0, xmm23
+// CHECK: encoding: [0x62,0xb1,0xff,0x08,0x7a,0xc7]
+ vcvtuqq2ps xmm0, xmm23
+
+// CHECK: vcvtuqq2ps xmm0, xmm23
+// CHECK: encoding: [0x62,0xb1,0xff,0x08,0x7a,0xc7]
+ vcvtuqq2psx xmm0, xmm23
+
+// CHECK: vcvtuqq2ps xmm16, xmmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xff,0x08,0x7a,0x00]
+ vcvtuqq2ps xmm16, xmmword ptr [rax]
+
+// CHECK: vcvtuqq2ps xmm16, xmmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xff,0x08,0x7a,0x00]
+ vcvtuqq2psx xmm16, xmmword ptr [rax]
+
+// CHECK: vcvtuqq2ps xmm0, ymm23
+// CHECK: encoding: [0x62,0xb1,0xff,0x28,0x7a,0xc7]
+ vcvtuqq2ps xmm0, ymm23
+
+// CHECK: vcvtuqq2ps xmm0, ymm23
+// CHECK: encoding: [0x62,0xb1,0xff,0x28,0x7a,0xc7]
+ vcvtuqq2psy xmm0, ymm23
+
+// CHECK: vcvtuqq2ps xmm16, ymmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xff,0x28,0x7a,0x00]
+ vcvtuqq2ps xmm16, ymmword ptr [rax]
+
+// CHECK: vcvtuqq2ps xmm16, ymmword ptr [rax]
+// CHECK: encoding: [0x62,0xe1,0xff,0x28,0x7a,0x00]
+ vcvtuqq2psy xmm16, ymmword ptr [rax]
// CHECK: encoding: [0xc5,0xf9,0xe6,0xe9]
vcvttpd2dq %xmm1, %xmm5
-// CHECK: vcvttpd2dqy %ymm2, %xmm5
+// CHECK: vcvttpd2dq %ymm2, %xmm5
// CHECK: encoding: [0xc5,0xfd,0xe6,0xea]
vcvttpd2dq %ymm2, %xmm5
// CHECK: vcvttpd2dq %xmm1, %xmm5
// CHECK: encoding: [0xc5,0xf9,0xe6,0xe9]
- vcvttpd2dqx %xmm1, %xmm5
+ vcvttpd2dq %xmm1, %xmm5
// CHECK: vcvttpd2dqx (%eax), %xmm1
// CHECK: encoding: [0xc5,0xf9,0xe6,0x08]
vcvttpd2dqx (%eax), %xmm1
-// CHECK: vcvttpd2dqy %ymm2, %xmm1
+// CHECK: vcvttpd2dq %ymm2, %xmm1
// CHECK: encoding: [0xc5,0xfd,0xe6,0xca]
vcvttpd2dqy %ymm2, %xmm1
// CHECK: encoding: [0xc5,0xfd,0xe6,0x08]
vcvttpd2dqy (%eax), %xmm1
-// CHECK: vcvtpd2psy %ymm2, %xmm5
+// CHECK: vcvtpd2ps %ymm2, %xmm5
// CHECK: encoding: [0xc5,0xfd,0x5a,0xea]
vcvtpd2ps %ymm2, %xmm5
// CHECK: encoding: [0xc5,0xf9,0x5a,0x08]
vcvtpd2psx (%eax), %xmm1
-// CHECK: vcvtpd2psy %ymm2, %xmm1
+// CHECK: vcvtpd2ps %ymm2, %xmm1
// CHECK: encoding: [0xc5,0xfd,0x5a,0xca]
vcvtpd2psy %ymm2, %xmm1
// CHECK: encoding: [0xc5,0xfd,0x5a,0x08]
vcvtpd2psy (%eax), %xmm1
-// CHECK: vcvtpd2dqy %ymm2, %xmm5
+// CHECK: vcvtpd2dq %ymm2, %xmm5
// CHECK: encoding: [0xc5,0xff,0xe6,0xea]
vcvtpd2dq %ymm2, %xmm5
-// CHECK: vcvtpd2dqy %ymm2, %xmm1
+// CHECK: vcvtpd2dq %ymm2, %xmm1
// CHECK: encoding: [0xc5,0xff,0xe6,0xca]
vcvtpd2dqy %ymm2, %xmm1
// CHECK: encoding: [0xc4,0x41,0x79,0xe6,0xd3]
vcvttpd2dq %xmm11, %xmm10
-// CHECK: vcvttpd2dqy %ymm12, %xmm10
+// CHECK: vcvttpd2dq %ymm12, %xmm10
// CHECK: encoding: [0xc4,0x41,0x7d,0xe6,0xd4]
vcvttpd2dq %ymm12, %xmm10
// CHECK: encoding: [0xc5,0x79,0xe6,0x18]
vcvttpd2dqx (%rax), %xmm11
-// CHECK: vcvttpd2dqy %ymm12, %xmm11
+// CHECK: vcvttpd2dq %ymm12, %xmm11
// CHECK: encoding: [0xc4,0x41,0x7d,0xe6,0xdc]
vcvttpd2dqy %ymm12, %xmm11
// CHECK: encoding: [0xc5,0x7d,0xe6,0x18]
vcvttpd2dqy (%rax), %xmm11
-// CHECK: vcvtpd2psy %ymm12, %xmm10
+// CHECK: vcvtpd2ps %ymm12, %xmm10
// CHECK: encoding: [0xc4,0x41,0x7d,0x5a,0xd4]
vcvtpd2ps %ymm12, %xmm10
// CHECK: encoding: [0xc5,0x79,0x5a,0x18]
vcvtpd2psx (%rax), %xmm11
-// CHECK: vcvtpd2psy %ymm12, %xmm11
+// CHECK: vcvtpd2ps %ymm12, %xmm11
// CHECK: encoding: [0xc4,0x41,0x7d,0x5a,0xdc]
vcvtpd2psy %ymm12, %xmm11
// CHECK: encoding: [0xc5,0x7d,0x5a,0x18]
vcvtpd2psy (%rax), %xmm11
-// CHECK: vcvtpd2dqy %ymm12, %xmm10
+// CHECK: vcvtpd2dq %ymm12, %xmm10
// CHECK: encoding: [0xc4,0x41,0x7f,0xe6,0xd4]
vcvtpd2dq %ymm12, %xmm10
-// CHECK: vcvtpd2dqy %ymm12, %xmm11
+// CHECK: vcvtpd2dq %ymm12, %xmm11
// CHECK: encoding: [0xc4,0x41,0x7f,0xe6,0xdc]
vcvtpd2dqy %ymm12, %xmm11