diff options
author | Daniel Sanders <daniel.sanders@imgtec.com> | 2013-09-23 12:57:42 +0000 |
---|---|---|
committer | Daniel Sanders <daniel.sanders@imgtec.com> | 2013-09-23 12:57:42 +0000 |
commit | 4e812c1f4a723f0fa0e8714610e08be593c759b8 (patch) | |
tree | 2f13b4639f8479ec83bdb4f34897a1c377ea52a2 /test/CodeGen/Mips/msa | |
parent | 57ebcb28a63d8646fd8fd69cfd9e6766066e342f (diff) | |
download | external_llvm-4e812c1f4a723f0fa0e8714610e08be593c759b8.zip external_llvm-4e812c1f4a723f0fa0e8714610e08be593c759b8.tar.gz external_llvm-4e812c1f4a723f0fa0e8714610e08be593c759b8.tar.bz2 |
[mips][msa] Added support for matching and, or, and xor from normal IR (i.e. not intrinsics)
git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@191194 91177308-0d34-0410-b5e6-96231b3b80d8
Diffstat (limited to 'test/CodeGen/Mips/msa')
-rw-r--r-- | test/CodeGen/Mips/msa/bitwise.ll | 192 | ||||
-rw-r--r-- | test/CodeGen/Mips/msa/vec.ll | 198 |
2 files changed, 390 insertions, 0 deletions
diff --git a/test/CodeGen/Mips/msa/bitwise.ll b/test/CodeGen/Mips/msa/bitwise.ll index 61d2ccb..90b9589 100644 --- a/test/CodeGen/Mips/msa/bitwise.ll +++ b/test/CodeGen/Mips/msa/bitwise.ll @@ -1,5 +1,197 @@ ; RUN: llc -march=mips -mattr=+msa < %s | FileCheck %s +define void @and_v16i8(<16 x i8>* %c, <16 x i8>* %a, <16 x i8>* %b) nounwind { + ; CHECK: and_v16i8: + + %1 = load <16 x i8>* %a + ; CHECK-DAG: ld.b [[R1:\$w[0-9]+]], 0($5) + %2 = load <16 x i8>* %b + ; CHECK-DAG: ld.b [[R2:\$w[0-9]+]], 0($6) + %3 = and <16 x i8> %1, %2 + ; CHECK-DAG: and.v [[R3:\$w[0-9]+]], [[R1]], [[R2]] + store <16 x i8> %3, <16 x i8>* %c + ; CHECK-DAG: st.b [[R3]], 0($4) + + ret void + ; CHECK: .size and_v16i8 +} + +define void @and_v8i16(<8 x i16>* %c, <8 x i16>* %a, <8 x i16>* %b) nounwind { + ; CHECK: and_v8i16: + + %1 = load <8 x i16>* %a + ; CHECK-DAG: ld.h [[R1:\$w[0-9]+]], 0($5) + %2 = load <8 x i16>* %b + ; CHECK-DAG: ld.h [[R2:\$w[0-9]+]], 0($6) + %3 = and <8 x i16> %1, %2 + ; CHECK-DAG: and.v [[R3:\$w[0-9]+]], [[R1]], [[R2]] + store <8 x i16> %3, <8 x i16>* %c + ; CHECK-DAG: st.h [[R3]], 0($4) + + ret void + ; CHECK: .size and_v8i16 +} + +define void @and_v4i32(<4 x i32>* %c, <4 x i32>* %a, <4 x i32>* %b) nounwind { + ; CHECK: and_v4i32: + + %1 = load <4 x i32>* %a + ; CHECK-DAG: ld.w [[R1:\$w[0-9]+]], 0($5) + %2 = load <4 x i32>* %b + ; CHECK-DAG: ld.w [[R2:\$w[0-9]+]], 0($6) + %3 = and <4 x i32> %1, %2 + ; CHECK-DAG: and.v [[R3:\$w[0-9]+]], [[R1]], [[R2]] + store <4 x i32> %3, <4 x i32>* %c + ; CHECK-DAG: st.w [[R3]], 0($4) + + ret void + ; CHECK: .size and_v4i32 +} + +define void @and_v2i64(<2 x i64>* %c, <2 x i64>* %a, <2 x i64>* %b) nounwind { + ; CHECK: and_v2i64: + + %1 = load <2 x i64>* %a + ; CHECK-DAG: ld.d [[R1:\$w[0-9]+]], 0($5) + %2 = load <2 x i64>* %b + ; CHECK-DAG: ld.d [[R2:\$w[0-9]+]], 0($6) + %3 = and <2 x i64> %1, %2 + ; CHECK-DAG: and.v [[R3:\$w[0-9]+]], [[R1]], [[R2]] + store <2 x i64> %3, <2 x i64>* %c + ; CHECK-DAG: st.d [[R3]], 0($4) + + ret void + ; CHECK: .size and_v2i64 +} + +define void @or_v16i8(<16 x i8>* %c, <16 x i8>* %a, <16 x i8>* %b) nounwind { + ; CHECK: or_v16i8: + + %1 = load <16 x i8>* %a + ; CHECK-DAG: ld.b [[R1:\$w[0-9]+]], 0($5) + %2 = load <16 x i8>* %b + ; CHECK-DAG: ld.b [[R2:\$w[0-9]+]], 0($6) + %3 = or <16 x i8> %1, %2 + ; CHECK-DAG: or.v [[R3:\$w[0-9]+]], [[R1]], [[R2]] + store <16 x i8> %3, <16 x i8>* %c + ; CHECK-DAG: st.b [[R3]], 0($4) + + ret void + ; CHECK: .size or_v16i8 +} + +define void @or_v8i16(<8 x i16>* %c, <8 x i16>* %a, <8 x i16>* %b) nounwind { + ; CHECK: or_v8i16: + + %1 = load <8 x i16>* %a + ; CHECK-DAG: ld.h [[R1:\$w[0-9]+]], 0($5) + %2 = load <8 x i16>* %b + ; CHECK-DAG: ld.h [[R2:\$w[0-9]+]], 0($6) + %3 = or <8 x i16> %1, %2 + ; CHECK-DAG: or.v [[R3:\$w[0-9]+]], [[R1]], [[R2]] + store <8 x i16> %3, <8 x i16>* %c + ; CHECK-DAG: st.h [[R3]], 0($4) + + ret void + ; CHECK: .size or_v8i16 +} + +define void @or_v4i32(<4 x i32>* %c, <4 x i32>* %a, <4 x i32>* %b) nounwind { + ; CHECK: or_v4i32: + + %1 = load <4 x i32>* %a + ; CHECK-DAG: ld.w [[R1:\$w[0-9]+]], 0($5) + %2 = load <4 x i32>* %b + ; CHECK-DAG: ld.w [[R2:\$w[0-9]+]], 0($6) + %3 = or <4 x i32> %1, %2 + ; CHECK-DAG: or.v [[R3:\$w[0-9]+]], [[R1]], [[R2]] + store <4 x i32> %3, <4 x i32>* %c + ; CHECK-DAG: st.w [[R3]], 0($4) + + ret void + ; CHECK: .size or_v4i32 +} + +define void @or_v2i64(<2 x i64>* %c, <2 x i64>* %a, <2 x i64>* %b) nounwind { + ; CHECK: or_v2i64: + + %1 = load <2 x i64>* %a + ; CHECK-DAG: ld.d [[R1:\$w[0-9]+]], 0($5) + %2 = load <2 x i64>* %b + ; CHECK-DAG: ld.d [[R2:\$w[0-9]+]], 0($6) + %3 = or <2 x i64> %1, %2 + ; CHECK-DAG: or.v [[R3:\$w[0-9]+]], [[R1]], [[R2]] + store <2 x i64> %3, <2 x i64>* %c + ; CHECK-DAG: st.d [[R3]], 0($4) + + ret void + ; CHECK: .size or_v2i64 +} + +define void @xor_v16i8(<16 x i8>* %c, <16 x i8>* %a, <16 x i8>* %b) nounwind { + ; CHECK: xor_v16i8: + + %1 = load <16 x i8>* %a + ; CHECK-DAG: ld.b [[R1:\$w[0-9]+]], 0($5) + %2 = load <16 x i8>* %b + ; CHECK-DAG: ld.b [[R2:\$w[0-9]+]], 0($6) + %3 = xor <16 x i8> %1, %2 + ; CHECK-DAG: xor.v [[R3:\$w[0-9]+]], [[R1]], [[R2]] + store <16 x i8> %3, <16 x i8>* %c + ; CHECK-DAG: st.b [[R3]], 0($4) + + ret void + ; CHECK: .size xor_v16i8 +} + +define void @xor_v8i16(<8 x i16>* %c, <8 x i16>* %a, <8 x i16>* %b) nounwind { + ; CHECK: xor_v8i16: + + %1 = load <8 x i16>* %a + ; CHECK-DAG: ld.h [[R1:\$w[0-9]+]], 0($5) + %2 = load <8 x i16>* %b + ; CHECK-DAG: ld.h [[R2:\$w[0-9]+]], 0($6) + %3 = xor <8 x i16> %1, %2 + ; CHECK-DAG: xor.v [[R3:\$w[0-9]+]], [[R1]], [[R2]] + store <8 x i16> %3, <8 x i16>* %c + ; CHECK-DAG: st.h [[R3]], 0($4) + + ret void + ; CHECK: .size xor_v8i16 +} + +define void @xor_v4i32(<4 x i32>* %c, <4 x i32>* %a, <4 x i32>* %b) nounwind { + ; CHECK: xor_v4i32: + + %1 = load <4 x i32>* %a + ; CHECK-DAG: ld.w [[R1:\$w[0-9]+]], 0($5) + %2 = load <4 x i32>* %b + ; CHECK-DAG: ld.w [[R2:\$w[0-9]+]], 0($6) + %3 = xor <4 x i32> %1, %2 + ; CHECK-DAG: xor.v [[R3:\$w[0-9]+]], [[R1]], [[R2]] + store <4 x i32> %3, <4 x i32>* %c + ; CHECK-DAG: st.w [[R3]], 0($4) + + ret void + ; CHECK: .size xor_v4i32 +} + +define void @xor_v2i64(<2 x i64>* %c, <2 x i64>* %a, <2 x i64>* %b) nounwind { + ; CHECK: xor_v2i64: + + %1 = load <2 x i64>* %a + ; CHECK-DAG: ld.d [[R1:\$w[0-9]+]], 0($5) + %2 = load <2 x i64>* %b + ; CHECK-DAG: ld.d [[R2:\$w[0-9]+]], 0($6) + %3 = xor <2 x i64> %1, %2 + ; CHECK-DAG: xor.v [[R3:\$w[0-9]+]], [[R1]], [[R2]] + store <2 x i64> %3, <2 x i64>* %c + ; CHECK-DAG: st.d [[R3]], 0($4) + + ret void + ; CHECK: .size xor_v2i64 +} + define void @sll_v16i8(<16 x i8>* %c, <16 x i8>* %a, <16 x i8>* %b) nounwind { ; CHECK: sll_v16i8: diff --git a/test/CodeGen/Mips/msa/vec.ll b/test/CodeGen/Mips/msa/vec.ll index 13a23fa..7ad640b 100644 --- a/test/CodeGen/Mips/msa/vec.ll +++ b/test/CodeGen/Mips/msa/vec.ll @@ -95,6 +95,72 @@ entry: ; ANYENDIAN: st.b ; ANYENDIAN: .size llvm_mips_and_v_d_test ; +define void @and_v_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_and_v_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_and_v_b_ARG2 + %2 = and <16 x i8> %0, %1 + store <16 x i8> %2, <16 x i8>* @llvm_mips_and_v_b_RES + ret void +} + +; CHECK: and_v_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: and.v +; CHECK: st.b +; CHECK: .size and_v_b_test +; +define void @and_v_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_and_v_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_and_v_h_ARG2 + %2 = and <8 x i16> %0, %1 + store <8 x i16> %2, <8 x i16>* @llvm_mips_and_v_h_RES + ret void +} + +; CHECK: and_v_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: and.v +; CHECK: st.h +; CHECK: .size and_v_h_test +; + +define void @and_v_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_and_v_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_and_v_w_ARG2 + %2 = and <4 x i32> %0, %1 + store <4 x i32> %2, <4 x i32>* @llvm_mips_and_v_w_RES + ret void +} + +; CHECK: and_v_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: and.v +; CHECK: st.w +; CHECK: .size and_v_w_test +; + +define void @and_v_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_and_v_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_and_v_d_ARG2 + %2 = and <2 x i64> %0, %1 + store <2 x i64> %2, <2 x i64>* @llvm_mips_and_v_d_RES + ret void +} + +; CHECK: and_v_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: and.v +; CHECK: st.d +; CHECK: .size and_v_d_test +; @llvm_mips_bmnz_v_b_ARG1 = global <16 x i8> <i8 0, i8 1, i8 2, i8 3, i8 4, i8 5, i8 6, i8 7, i8 8, i8 9, i8 10, i8 11, i8 12, i8 13, i8 14, i8 15>, align 16 @llvm_mips_bmnz_v_b_ARG2 = global <16 x i8> <i8 16, i8 17, i8 18, i8 19, i8 20, i8 21, i8 22, i8 23, i8 24, i8 25, i8 26, i8 27, i8 28, i8 29, i8 30, i8 31>, align 16 @llvm_mips_bmnz_v_b_RES = global <16 x i8> <i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0>, align 16 @@ -555,6 +621,72 @@ entry: ; ANYENDIAN: st.b ; ANYENDIAN: .size llvm_mips_or_v_d_test ; +define void @or_v_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_or_v_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_or_v_b_ARG2 + %2 = or <16 x i8> %0, %1 + store <16 x i8> %2, <16 x i8>* @llvm_mips_or_v_b_RES + ret void +} + +; CHECK: or_v_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: or.v +; CHECK: st.b +; CHECK: .size or_v_b_test +; +define void @or_v_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_or_v_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_or_v_h_ARG2 + %2 = or <8 x i16> %0, %1 + store <8 x i16> %2, <8 x i16>* @llvm_mips_or_v_h_RES + ret void +} + +; CHECK: or_v_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: or.v +; CHECK: st.h +; CHECK: .size or_v_h_test +; + +define void @or_v_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_or_v_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_or_v_w_ARG2 + %2 = or <4 x i32> %0, %1 + store <4 x i32> %2, <4 x i32>* @llvm_mips_or_v_w_RES + ret void +} + +; CHECK: or_v_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: or.v +; CHECK: st.w +; CHECK: .size or_v_w_test +; + +define void @or_v_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_or_v_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_or_v_d_ARG2 + %2 = or <2 x i64> %0, %1 + store <2 x i64> %2, <2 x i64>* @llvm_mips_or_v_d_RES + ret void +} + +; CHECK: or_v_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: or.v +; CHECK: st.d +; CHECK: .size or_v_d_test +; @llvm_mips_xor_v_b_ARG1 = global <16 x i8> <i8 0, i8 1, i8 2, i8 3, i8 4, i8 5, i8 6, i8 7, i8 8, i8 9, i8 10, i8 11, i8 12, i8 13, i8 14, i8 15>, align 16 @llvm_mips_xor_v_b_ARG2 = global <16 x i8> <i8 16, i8 17, i8 18, i8 19, i8 20, i8 21, i8 22, i8 23, i8 24, i8 25, i8 26, i8 27, i8 28, i8 29, i8 30, i8 31>, align 16 @llvm_mips_xor_v_b_RES = global <16 x i8> <i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0, i8 0>, align 16 @@ -647,6 +779,72 @@ entry: ; ANYENDIAN: st.b ; ANYENDIAN: .size llvm_mips_xor_v_d_test ; +define void @xor_v_b_test() nounwind { +entry: + %0 = load <16 x i8>* @llvm_mips_xor_v_b_ARG1 + %1 = load <16 x i8>* @llvm_mips_xor_v_b_ARG2 + %2 = xor <16 x i8> %0, %1 + store <16 x i8> %2, <16 x i8>* @llvm_mips_xor_v_b_RES + ret void +} + +; CHECK: xor_v_b_test: +; CHECK: ld.b +; CHECK: ld.b +; CHECK: xor.v +; CHECK: st.b +; CHECK: .size xor_v_b_test +; +define void @xor_v_h_test() nounwind { +entry: + %0 = load <8 x i16>* @llvm_mips_xor_v_h_ARG1 + %1 = load <8 x i16>* @llvm_mips_xor_v_h_ARG2 + %2 = xor <8 x i16> %0, %1 + store <8 x i16> %2, <8 x i16>* @llvm_mips_xor_v_h_RES + ret void +} + +; CHECK: xor_v_h_test: +; CHECK: ld.h +; CHECK: ld.h +; CHECK: xor.v +; CHECK: st.h +; CHECK: .size xor_v_h_test +; + +define void @xor_v_w_test() nounwind { +entry: + %0 = load <4 x i32>* @llvm_mips_xor_v_w_ARG1 + %1 = load <4 x i32>* @llvm_mips_xor_v_w_ARG2 + %2 = xor <4 x i32> %0, %1 + store <4 x i32> %2, <4 x i32>* @llvm_mips_xor_v_w_RES + ret void +} + +; CHECK: xor_v_w_test: +; CHECK: ld.w +; CHECK: ld.w +; CHECK: xor.v +; CHECK: st.w +; CHECK: .size xor_v_w_test +; + +define void @xor_v_d_test() nounwind { +entry: + %0 = load <2 x i64>* @llvm_mips_xor_v_d_ARG1 + %1 = load <2 x i64>* @llvm_mips_xor_v_d_ARG2 + %2 = xor <2 x i64> %0, %1 + store <2 x i64> %2, <2 x i64>* @llvm_mips_xor_v_d_RES + ret void +} + +; CHECK: xor_v_d_test: +; CHECK: ld.d +; CHECK: ld.d +; CHECK: xor.v +; CHECK: st.d +; CHECK: .size xor_v_d_test +; declare <16 x i8> @llvm.mips.and.v(<16 x i8>, <16 x i8>) nounwind declare <16 x i8> @llvm.mips.bmnz.v(<16 x i8>, <16 x i8>) nounwind declare <16 x i8> @llvm.mips.bmz.v(<16 x i8>, <16 x i8>) nounwind |