diff options
Diffstat (limited to 'test/CodeGen/PowerPC/vsx-fma-m.ll')
-rw-r--r-- | test/CodeGen/PowerPC/vsx-fma-m.ll | 36 |
1 files changed, 18 insertions, 18 deletions
diff --git a/test/CodeGen/PowerPC/vsx-fma-m.ll b/test/CodeGen/PowerPC/vsx-fma-m.ll index ab36072..64185a4 100644 --- a/test/CodeGen/PowerPC/vsx-fma-m.ll +++ b/test/CodeGen/PowerPC/vsx-fma-m.ll @@ -12,7 +12,7 @@ entry: %0 = tail call double @llvm.fma.f64(double %b, double %c, double %a) store double %0, double* %d, align 8 %1 = tail call double @llvm.fma.f64(double %b, double %e, double %a) - %arrayidx1 = getelementptr inbounds double* %d, i64 1 + %arrayidx1 = getelementptr inbounds double, double* %d, i64 1 store double %1, double* %arrayidx1, align 8 ret void @@ -39,10 +39,10 @@ entry: %0 = tail call double @llvm.fma.f64(double %b, double %c, double %a) store double %0, double* %d, align 8 %1 = tail call double @llvm.fma.f64(double %b, double %e, double %a) - %arrayidx1 = getelementptr inbounds double* %d, i64 1 + %arrayidx1 = getelementptr inbounds double, double* %d, i64 1 store double %1, double* %arrayidx1, align 8 %2 = tail call double @llvm.fma.f64(double %b, double %f, double %a) - %arrayidx2 = getelementptr inbounds double* %d, i64 2 + %arrayidx2 = getelementptr inbounds double, double* %d, i64 2 store double %2, double* %arrayidx2, align 8 ret void @@ -77,12 +77,12 @@ entry: store double %0, double* %d, align 8 %1 = tail call double @llvm.fma.f64(double %b, double %e, double %a) %2 = tail call double @llvm.fma.f64(double %b, double %c, double %1) - %arrayidx1 = getelementptr inbounds double* %d, i64 3 + %arrayidx1 = getelementptr inbounds double, double* %d, i64 3 store double %2, double* %arrayidx1, align 8 %3 = tail call double @llvm.fma.f64(double %b, double %f, double %a) - %arrayidx2 = getelementptr inbounds double* %d, i64 2 + %arrayidx2 = getelementptr inbounds double, double* %d, i64 2 store double %3, double* %arrayidx2, align 8 - %arrayidx3 = getelementptr inbounds double* %d, i64 1 + %arrayidx3 = getelementptr inbounds double, double* %d, i64 1 store double %1, double* %arrayidx3, align 8 ret void @@ -125,13 +125,13 @@ entry: %0 = tail call double @llvm.fma.f64(double %b, double %c, double %a) store double %0, double* %d, align 8 %1 = tail call double @llvm.fma.f64(double %b, double %e, double %a) - %arrayidx1 = getelementptr inbounds double* %d, i64 1 + %arrayidx1 = getelementptr inbounds double, double* %d, i64 1 store double %1, double* %arrayidx1, align 8 %2 = tail call double @llvm.fma.f64(double %b, double %c, double %1) - %arrayidx3 = getelementptr inbounds double* %d, i64 3 + %arrayidx3 = getelementptr inbounds double, double* %d, i64 3 store double %2, double* %arrayidx3, align 8 %3 = tail call double @llvm.fma.f64(double %b, double %f, double %a) - %arrayidx4 = getelementptr inbounds double* %d, i64 2 + %arrayidx4 = getelementptr inbounds double, double* %d, i64 2 store double %3, double* %arrayidx4, align 8 ret void @@ -178,7 +178,7 @@ entry: %0 = tail call <2 x double> @llvm.fma.v2f64(<2 x double> %b, <2 x double> %c, <2 x double> %a) store <2 x double> %0, <2 x double>* %d, align 8 %1 = tail call <2 x double> @llvm.fma.v2f64(<2 x double> %b, <2 x double> %e, <2 x double> %a) - %arrayidx1 = getelementptr inbounds <2 x double>* %d, i64 1 + %arrayidx1 = getelementptr inbounds <2 x double>, <2 x double>* %d, i64 1 store <2 x double> %1, <2 x double>* %arrayidx1, align 8 ret void @@ -205,10 +205,10 @@ entry: %0 = tail call <2 x double> @llvm.fma.v2f64(<2 x double> %b, <2 x double> %c, <2 x double> %a) store <2 x double> %0, <2 x double>* %d, align 8 %1 = tail call <2 x double> @llvm.fma.v2f64(<2 x double> %b, <2 x double> %e, <2 x double> %a) - %arrayidx1 = getelementptr inbounds <2 x double>* %d, i64 1 + %arrayidx1 = getelementptr inbounds <2 x double>, <2 x double>* %d, i64 1 store <2 x double> %1, <2 x double>* %arrayidx1, align 8 %2 = tail call <2 x double> @llvm.fma.v2f64(<2 x double> %b, <2 x double> %f, <2 x double> %a) - %arrayidx2 = getelementptr inbounds <2 x double>* %d, i64 2 + %arrayidx2 = getelementptr inbounds <2 x double>, <2 x double>* %d, i64 2 store <2 x double> %2, <2 x double>* %arrayidx2, align 8 ret void @@ -243,12 +243,12 @@ entry: store <2 x double> %0, <2 x double>* %d, align 8 %1 = tail call <2 x double> @llvm.fma.v2f64(<2 x double> %b, <2 x double> %e, <2 x double> %a) %2 = tail call <2 x double> @llvm.fma.v2f64(<2 x double> %b, <2 x double> %c, <2 x double> %1) - %arrayidx1 = getelementptr inbounds <2 x double>* %d, i64 3 + %arrayidx1 = getelementptr inbounds <2 x double>, <2 x double>* %d, i64 3 store <2 x double> %2, <2 x double>* %arrayidx1, align 8 %3 = tail call <2 x double> @llvm.fma.v2f64(<2 x double> %b, <2 x double> %f, <2 x double> %a) - %arrayidx2 = getelementptr inbounds <2 x double>* %d, i64 2 + %arrayidx2 = getelementptr inbounds <2 x double>, <2 x double>* %d, i64 2 store <2 x double> %3, <2 x double>* %arrayidx2, align 8 - %arrayidx3 = getelementptr inbounds <2 x double>* %d, i64 1 + %arrayidx3 = getelementptr inbounds <2 x double>, <2 x double>* %d, i64 1 store <2 x double> %1, <2 x double>* %arrayidx3, align 8 ret void @@ -300,13 +300,13 @@ entry: %0 = tail call <2 x double> @llvm.fma.v2f64(<2 x double> %b, <2 x double> %c, <2 x double> %a) store <2 x double> %0, <2 x double>* %d, align 8 %1 = tail call <2 x double> @llvm.fma.v2f64(<2 x double> %b, <2 x double> %e, <2 x double> %a) - %arrayidx1 = getelementptr inbounds <2 x double>* %d, i64 1 + %arrayidx1 = getelementptr inbounds <2 x double>, <2 x double>* %d, i64 1 store <2 x double> %1, <2 x double>* %arrayidx1, align 8 %2 = tail call <2 x double> @llvm.fma.v2f64(<2 x double> %b, <2 x double> %c, <2 x double> %1) - %arrayidx3 = getelementptr inbounds <2 x double>* %d, i64 3 + %arrayidx3 = getelementptr inbounds <2 x double>, <2 x double>* %d, i64 3 store <2 x double> %2, <2 x double>* %arrayidx3, align 8 %3 = tail call <2 x double> @llvm.fma.v2f64(<2 x double> %b, <2 x double> %f, <2 x double> %a) - %arrayidx4 = getelementptr inbounds <2 x double>* %d, i64 2 + %arrayidx4 = getelementptr inbounds <2 x double>, <2 x double>* %d, i64 2 store <2 x double> %3, <2 x double>* %arrayidx4, align 8 ret void |