diff options
Diffstat (limited to 'test/CodeGen/AArch64/arm64-ldp.ll')
-rw-r--r-- | test/CodeGen/AArch64/arm64-ldp.ll | 149 |
1 files changed, 149 insertions, 0 deletions
diff --git a/test/CodeGen/AArch64/arm64-ldp.ll b/test/CodeGen/AArch64/arm64-ldp.ll new file mode 100644 index 0000000..5a98626 --- /dev/null +++ b/test/CodeGen/AArch64/arm64-ldp.ll @@ -0,0 +1,149 @@ +; RUN: llc < %s -march=arm64 -verify-machineinstrs | FileCheck %s +; RUN: llc < %s -march=arm64 -aarch64-unscaled-mem-op=true\ +; RUN: -verify-machineinstrs | FileCheck -check-prefix=LDUR_CHK %s + +; CHECK: ldp_int +; CHECK: ldp +define i32 @ldp_int(i32* %p) nounwind { + %tmp = load i32* %p, align 4 + %add.ptr = getelementptr inbounds i32* %p, i64 1 + %tmp1 = load i32* %add.ptr, align 4 + %add = add nsw i32 %tmp1, %tmp + ret i32 %add +} + +; CHECK: ldp_long +; CHECK: ldp +define i64 @ldp_long(i64* %p) nounwind { + %tmp = load i64* %p, align 8 + %add.ptr = getelementptr inbounds i64* %p, i64 1 + %tmp1 = load i64* %add.ptr, align 8 + %add = add nsw i64 %tmp1, %tmp + ret i64 %add +} + +; CHECK: ldp_float +; CHECK: ldp +define float @ldp_float(float* %p) nounwind { + %tmp = load float* %p, align 4 + %add.ptr = getelementptr inbounds float* %p, i64 1 + %tmp1 = load float* %add.ptr, align 4 + %add = fadd float %tmp, %tmp1 + ret float %add +} + +; CHECK: ldp_double +; CHECK: ldp +define double @ldp_double(double* %p) nounwind { + %tmp = load double* %p, align 8 + %add.ptr = getelementptr inbounds double* %p, i64 1 + %tmp1 = load double* %add.ptr, align 8 + %add = fadd double %tmp, %tmp1 + ret double %add +} + +; Test the load/store optimizer---combine ldurs into a ldp, if appropriate +define i32 @ldur_int(i32* %a) nounwind { +; LDUR_CHK: ldur_int +; LDUR_CHK: ldp [[DST1:w[0-9]+]], [[DST2:w[0-9]+]], [x0, #-8] +; LDUR_CHK-NEXT: add w{{[0-9]+}}, [[DST2]], [[DST1]] +; LDUR_CHK-NEXT: ret + %p1 = getelementptr inbounds i32* %a, i32 -1 + %tmp1 = load i32* %p1, align 2 + %p2 = getelementptr inbounds i32* %a, i32 -2 + %tmp2 = load i32* %p2, align 2 + %tmp3 = add i32 %tmp1, %tmp2 + ret i32 %tmp3 +} + +define i64 @ldur_long(i64* %a) nounwind ssp { +; LDUR_CHK: ldur_long +; LDUR_CHK: ldp [[DST1:x[0-9]+]], [[DST2:x[0-9]+]], [x0, #-16] +; LDUR_CHK-NEXT: add x{{[0-9]+}}, [[DST2]], [[DST1]] +; LDUR_CHK-NEXT: ret + %p1 = getelementptr inbounds i64* %a, i64 -1 + %tmp1 = load i64* %p1, align 2 + %p2 = getelementptr inbounds i64* %a, i64 -2 + %tmp2 = load i64* %p2, align 2 + %tmp3 = add i64 %tmp1, %tmp2 + ret i64 %tmp3 +} + +define float @ldur_float(float* %a) { +; LDUR_CHK: ldur_float +; LDUR_CHK: ldp [[DST1:s[0-9]+]], [[DST2:s[0-9]+]], [x0, #-8] +; LDUR_CHK-NEXT: add s{{[0-9]+}}, [[DST2]], [[DST1]] +; LDUR_CHK-NEXT: ret + %p1 = getelementptr inbounds float* %a, i64 -1 + %tmp1 = load float* %p1, align 2 + %p2 = getelementptr inbounds float* %a, i64 -2 + %tmp2 = load float* %p2, align 2 + %tmp3 = fadd float %tmp1, %tmp2 + ret float %tmp3 +} + +define double @ldur_double(double* %a) { +; LDUR_CHK: ldur_double +; LDUR_CHK: ldp [[DST1:d[0-9]+]], [[DST2:d[0-9]+]], [x0, #-16] +; LDUR_CHK-NEXT: add d{{[0-9]+}}, [[DST2]], [[DST1]] +; LDUR_CHK-NEXT: ret + %p1 = getelementptr inbounds double* %a, i64 -1 + %tmp1 = load double* %p1, align 2 + %p2 = getelementptr inbounds double* %a, i64 -2 + %tmp2 = load double* %p2, align 2 + %tmp3 = fadd double %tmp1, %tmp2 + ret double %tmp3 +} + +; Now check some boundary conditions +define i64 @pairUpBarelyIn(i64* %a) nounwind ssp { +; LDUR_CHK: pairUpBarelyIn +; LDUR_CHK-NOT: ldur +; LDUR_CHK: ldp [[DST1:x[0-9]+]], [[DST2:x[0-9]+]], [x0, #-256] +; LDUR_CHK-NEXT: add x{{[0-9]+}}, [[DST2]], [[DST1]] +; LDUR_CHK-NEXT: ret + %p1 = getelementptr inbounds i64* %a, i64 -31 + %tmp1 = load i64* %p1, align 2 + %p2 = getelementptr inbounds i64* %a, i64 -32 + %tmp2 = load i64* %p2, align 2 + %tmp3 = add i64 %tmp1, %tmp2 + ret i64 %tmp3 +} + +define i64 @pairUpBarelyOut(i64* %a) nounwind ssp { +; LDUR_CHK: pairUpBarelyOut +; LDUR_CHK-NOT: ldp +; Don't be fragile about which loads or manipulations of the base register +; are used---just check that there isn't an ldp before the add +; LDUR_CHK: add +; LDUR_CHK-NEXT: ret + %p1 = getelementptr inbounds i64* %a, i64 -32 + %tmp1 = load i64* %p1, align 2 + %p2 = getelementptr inbounds i64* %a, i64 -33 + %tmp2 = load i64* %p2, align 2 + %tmp3 = add i64 %tmp1, %tmp2 + ret i64 %tmp3 +} + +define i64 @pairUpNotAligned(i64* %a) nounwind ssp { +; LDUR_CHK: pairUpNotAligned +; LDUR_CHK-NOT: ldp +; LDUR_CHK: ldur +; LDUR_CHK-NEXT: ldur +; LDUR_CHK-NEXT: add +; LDUR_CHK-NEXT: ret + %p1 = getelementptr inbounds i64* %a, i64 -18 + %bp1 = bitcast i64* %p1 to i8* + %bp1p1 = getelementptr inbounds i8* %bp1, i64 1 + %dp1 = bitcast i8* %bp1p1 to i64* + %tmp1 = load i64* %dp1, align 1 + + %p2 = getelementptr inbounds i64* %a, i64 -17 + %bp2 = bitcast i64* %p2 to i8* + %bp2p1 = getelementptr inbounds i8* %bp2, i64 1 + %dp2 = bitcast i8* %bp2p1 to i64* + %tmp2 = load i64* %dp2, align 1 + + %tmp3 = add i64 %tmp1, %tmp2 + ret i64 %tmp3 +} |