diff options
author | Stephen Hines <srhines@google.com> | 2015-04-01 18:49:24 +0000 |
---|---|---|
committer | Gerrit Code Review <noreply-gerritcodereview@google.com> | 2015-04-01 18:49:26 +0000 |
commit | 3fa16bd6062e23bcdb82ed4dd965674792e6b761 (patch) | |
tree | 9348fc507292f7e8715d22d64ce5a32131b4f875 /test/CodeGen/X86/lower-vec-shift-2.ll | |
parent | beed47390a60f6f0c77532b3d3f76bb47ef49423 (diff) | |
parent | ebe69fe11e48d322045d5949c83283927a0d790b (diff) | |
download | external_llvm-3fa16bd6062e23bcdb82ed4dd965674792e6b761.zip external_llvm-3fa16bd6062e23bcdb82ed4dd965674792e6b761.tar.gz external_llvm-3fa16bd6062e23bcdb82ed4dd965674792e6b761.tar.bz2 |
Merge "Update aosp/master LLVM for rebase to r230699."
Diffstat (limited to 'test/CodeGen/X86/lower-vec-shift-2.ll')
-rw-r--r-- | test/CodeGen/X86/lower-vec-shift-2.ll | 157 |
1 files changed, 157 insertions, 0 deletions
diff --git a/test/CodeGen/X86/lower-vec-shift-2.ll b/test/CodeGen/X86/lower-vec-shift-2.ll new file mode 100644 index 0000000..fb8fbba --- /dev/null +++ b/test/CodeGen/X86/lower-vec-shift-2.ll @@ -0,0 +1,157 @@ +; RUN: llc -mtriple=x86_64-unknown-unknown -mattr=+sse2 < %s | FileCheck %s --check-prefix=SSE2 +; RUN: llc -mtriple=x86_64-unknown-unknown -mattr=+avx < %s | FileCheck %s --check-prefix=AVX + +define <8 x i16> @test1(<8 x i16> %A, <8 x i16> %B) { +; SSE2-LABEL: test1: +; SSE2: # BB#0: # %entry +; SSE2-NEXT: movd %xmm1, %eax +; SSE2-NEXT: movzwl %ax, %eax +; SSE2-NEXT: movd %eax, %xmm1 +; SSE2-NEXT: psllw %xmm1, %xmm0 +; SSE2-NEXT: retq +; +; AVX-LABEL: test1: +; AVX: # BB#0: # %entry +; AVX-NEXT: vpxor %xmm2, %xmm2, %xmm2 +; AVX-NEXT: vpblendw {{.*#+}} xmm1 = xmm1[0],xmm2[1,2,3,4,5,6,7] +; AVX-NEXT: vpsllw %xmm1, %xmm0, %xmm0 +; AVX-NEXT: retq +entry: + %vecinit14 = shufflevector <8 x i16> %B, <8 x i16> undef, <8 x i32> zeroinitializer + %shl = shl <8 x i16> %A, %vecinit14 + ret <8 x i16> %shl +} + +define <4 x i32> @test2(<4 x i32> %A, <4 x i32> %B) { +; SSE2-LABEL: test2: +; SSE2: # BB#0: # %entry +; SSE2-NEXT: xorps %xmm2, %xmm2 +; SSE2-NEXT: movss {{.*#+}} xmm2 = xmm1[0],xmm2[1,2,3] +; SSE2-NEXT: pslld %xmm2, %xmm0 +; SSE2-NEXT: retq +; +; AVX-LABEL: test2: +; AVX: # BB#0: # %entry +; AVX-NEXT: vpxor %xmm2, %xmm2, %xmm2 +; AVX-NEXT: vpblendw {{.*#+}} xmm1 = xmm1[0,1],xmm2[2,3,4,5,6,7] +; AVX-NEXT: vpslld %xmm1, %xmm0, %xmm0 +; AVX-NEXT: retq +entry: + %vecinit6 = shufflevector <4 x i32> %B, <4 x i32> undef, <4 x i32> zeroinitializer + %shl = shl <4 x i32> %A, %vecinit6 + ret <4 x i32> %shl +} + +define <2 x i64> @test3(<2 x i64> %A, <2 x i64> %B) { +; SSE2-LABEL: test3: +; SSE2: # BB#0: # %entry +; SSE2-NEXT: psllq %xmm1, %xmm0 +; SSE2-NEXT: retq +; +; AVX-LABEL: test3: +; AVX: # BB#0: # %entry +; AVX-NEXT: vpsllq %xmm1, %xmm0, %xmm0 +; AVX-NEXT: retq +entry: + %vecinit2 = shufflevector <2 x i64> %B, <2 x i64> undef, <2 x i32> zeroinitializer + %shl = shl <2 x i64> %A, %vecinit2 + ret <2 x i64> %shl +} + +define <8 x i16> @test4(<8 x i16> %A, <8 x i16> %B) { +; SSE2-LABEL: test4: +; SSE2: # BB#0: # %entry +; SSE2-NEXT: movd %xmm1, %eax +; SSE2-NEXT: movzwl %ax, %eax +; SSE2-NEXT: movd %eax, %xmm1 +; SSE2-NEXT: psrlw %xmm1, %xmm0 +; SSE2-NEXT: retq +; +; AVX-LABEL: test4: +; AVX: # BB#0: # %entry +; AVX-NEXT: vpxor %xmm2, %xmm2, %xmm2 +; AVX-NEXT: vpblendw {{.*#+}} xmm1 = xmm1[0],xmm2[1,2,3,4,5,6,7] +; AVX-NEXT: vpsrlw %xmm1, %xmm0, %xmm0 +; AVX-NEXT: retq +entry: + %vecinit14 = shufflevector <8 x i16> %B, <8 x i16> undef, <8 x i32> zeroinitializer + %shr = lshr <8 x i16> %A, %vecinit14 + ret <8 x i16> %shr +} + +define <4 x i32> @test5(<4 x i32> %A, <4 x i32> %B) { +; SSE2-LABEL: test5: +; SSE2: # BB#0: # %entry +; SSE2-NEXT: xorps %xmm2, %xmm2 +; SSE2-NEXT: movss {{.*#+}} xmm2 = xmm1[0],xmm2[1,2,3] +; SSE2-NEXT: psrld %xmm2, %xmm0 +; SSE2-NEXT: retq +; +; AVX-LABEL: test5: +; AVX: # BB#0: # %entry +; AVX-NEXT: vpxor %xmm2, %xmm2, %xmm2 +; AVX-NEXT: vpblendw {{.*#+}} xmm1 = xmm1[0,1],xmm2[2,3,4,5,6,7] +; AVX-NEXT: vpsrld %xmm1, %xmm0, %xmm0 +; AVX-NEXT: retq +entry: + %vecinit6 = shufflevector <4 x i32> %B, <4 x i32> undef, <4 x i32> zeroinitializer + %shr = lshr <4 x i32> %A, %vecinit6 + ret <4 x i32> %shr +} + +define <2 x i64> @test6(<2 x i64> %A, <2 x i64> %B) { +; SSE2-LABEL: test6: +; SSE2: # BB#0: # %entry +; SSE2-NEXT: psrlq %xmm1, %xmm0 +; SSE2-NEXT: retq +; +; AVX-LABEL: test6: +; AVX: # BB#0: # %entry +; AVX-NEXT: vpsrlq %xmm1, %xmm0, %xmm0 +; AVX-NEXT: retq +entry: + %vecinit2 = shufflevector <2 x i64> %B, <2 x i64> undef, <2 x i32> zeroinitializer + %shr = lshr <2 x i64> %A, %vecinit2 + ret <2 x i64> %shr +} + +define <8 x i16> @test7(<8 x i16> %A, <8 x i16> %B) { +; SSE2-LABEL: test7: +; SSE2: # BB#0: # %entry +; SSE2-NEXT: movd %xmm1, %eax +; SSE2-NEXT: movzwl %ax, %eax +; SSE2-NEXT: movd %eax, %xmm1 +; SSE2-NEXT: psraw %xmm1, %xmm0 +; SSE2-NEXT: retq +; +; AVX-LABEL: test7: +; AVX: # BB#0: # %entry +; AVX-NEXT: vpxor %xmm2, %xmm2, %xmm2 +; AVX-NEXT: vpblendw {{.*#+}} xmm1 = xmm1[0],xmm2[1,2,3,4,5,6,7] +; AVX-NEXT: vpsraw %xmm1, %xmm0, %xmm0 +; AVX-NEXT: retq +entry: + %vecinit14 = shufflevector <8 x i16> %B, <8 x i16> undef, <8 x i32> zeroinitializer + %shr = ashr <8 x i16> %A, %vecinit14 + ret <8 x i16> %shr +} + +define <4 x i32> @test8(<4 x i32> %A, <4 x i32> %B) { +; SSE2-LABEL: test8: +; SSE2: # BB#0: # %entry +; SSE2-NEXT: xorps %xmm2, %xmm2 +; SSE2-NEXT: movss {{.*#+}} xmm2 = xmm1[0],xmm2[1,2,3] +; SSE2-NEXT: psrad %xmm2, %xmm0 +; SSE2-NEXT: retq +; +; AVX-LABEL: test8: +; AVX: # BB#0: # %entry +; AVX-NEXT: vpxor %xmm2, %xmm2, %xmm2 +; AVX-NEXT: vpblendw {{.*#+}} xmm1 = xmm1[0,1],xmm2[2,3,4,5,6,7] +; AVX-NEXT: vpsrad %xmm1, %xmm0, %xmm0 +; AVX-NEXT: retq +entry: + %vecinit6 = shufflevector <4 x i32> %B, <4 x i32> undef, <4 x i32> zeroinitializer + %shr = ashr <4 x i32> %A, %vecinit6 + ret <4 x i32> %shr +} |