diff options
Diffstat (limited to 'test/CodeGen/X86/logical-load-fold.ll')
-rw-r--r-- | test/CodeGen/X86/logical-load-fold.ll | 53 |
1 files changed, 53 insertions, 0 deletions
diff --git a/test/CodeGen/X86/logical-load-fold.ll b/test/CodeGen/X86/logical-load-fold.ll new file mode 100644 index 0000000..5aac2d7 --- /dev/null +++ b/test/CodeGen/X86/logical-load-fold.ll @@ -0,0 +1,53 @@ +; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=sse2,sse-unaligned-mem | FileCheck %s --check-prefix=SSE2 +; RUN: llc < %s -mtriple=x86_64-unknown-unknown -mattr=avx | FileCheck %s --check-prefix=AVX + +; Although we have the ability to fold an unaligned load with AVX +; and under special conditions with some SSE implementations, we +; can not fold the load under any circumstances in these test +; cases because they are not 16-byte loads. The load must be +; executed as a scalar ('movs*') with a zero extension to +; 128-bits and then used in the packed logical ('andp*') op. +; PR22371 - http://llvm.org/bugs/show_bug.cgi?id=22371 + +define double @load_double_no_fold(double %x, double %y) { +; SSE2-LABEL: load_double_no_fold: +; SSE2: BB#0: +; SSE2-NEXT: cmplesd %xmm0, %xmm1 +; SSE2-NEXT: movsd {{.*#+}} xmm0 = mem[0],zero +; SSE2-NEXT: andpd %xmm1, %xmm0 +; SSE2-NEXT: retq +; +; AVX-LABEL: load_double_no_fold: +; AVX: BB#0: +; AVX-NEXT: vcmplesd %xmm0, %xmm1, %xmm0 +; AVX-NEXT: vmovsd {{.*#+}} xmm1 = mem[0],zero +; AVX-NEXT: vandpd %xmm1, %xmm0, %xmm0 +; AVX-NEXT: retq + + %cmp = fcmp oge double %x, %y + %zext = zext i1 %cmp to i32 + %conv = sitofp i32 %zext to double + ret double %conv +} + +define float @load_float_no_fold(float %x, float %y) { +; SSE2-LABEL: load_float_no_fold: +; SSE2: BB#0: +; SSE2-NEXT: cmpless %xmm0, %xmm1 +; SSE2-NEXT: movss {{.*#+}} xmm0 = mem[0],zero,zero,zero +; SSE2-NEXT: andps %xmm1, %xmm0 +; SSE2-NEXT: retq +; +; AVX-LABEL: load_float_no_fold: +; AVX: BB#0: +; AVX-NEXT: vcmpless %xmm0, %xmm1, %xmm0 +; AVX-NEXT: vmovss {{.*#+}} xmm1 = mem[0],zero,zero,zero +; AVX-NEXT: vandps %xmm1, %xmm0, %xmm0 +; AVX-NEXT: retq + + %cmp = fcmp oge float %x, %y + %zext = zext i1 %cmp to i32 + %conv = sitofp i32 %zext to float + ret float %conv +} + |