aboutsummaryrefslogtreecommitdiffstats
path: root/test/CodeGen/X86/gather-addresses.ll
blob: 5f48b1e32b1613eb1d7f37cb0e4ee7f86c0f863c (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
; RUN: llc -mtriple=x86_64-linux -mcpu=nehalem < %s | FileCheck %s --check-prefix=LIN
; RUN: llc -mtriple=x86_64-win32 -mcpu=nehalem < %s | FileCheck %s --check-prefix=WIN
; rdar://7398554

; When doing vector gather-scatter index calculation with 32-bit indices,
; bounce the vector off of cache rather than shuffling each individual
; element out of the index vector.

; CHECK: foo:
; LIN: movaps	(%rsi), %xmm0
; LIN: andps	(%rdx), %xmm0
; LIN: movaps	%xmm0, -24(%rsp)
; LIN: movslq	-24(%rsp), %[[REG1:r.+]]
; LIN: movslq	-20(%rsp), %[[REG2:r.+]]
; LIN: movslq	-16(%rsp), %[[REG3:r.+]]
; LIN: movslq	-12(%rsp), %[[REG4:r.+]]
; LIN: movsd	(%rdi,%[[REG1]],8), %xmm0
; LIN: movhpd	(%rdi,%[[REG2]],8), %xmm0
; LIN: movsd	(%rdi,%[[REG3]],8), %xmm1
; LIN: movhpd	(%rdi,%[[REG4]],8), %xmm1

; WIN: movaps	(%rdx), %xmm0
; WIN: andps	(%r8), %xmm0
; WIN: movaps	%xmm0, (%rsp)
; WIN: movslq	(%rsp), %[[REG1:r.+]]
; WIN: movslq	4(%rsp), %[[REG2:r.+]]
; WIN: movslq	8(%rsp), %[[REG3:r.+]]
; WIN: movslq	12(%rsp), %[[REG4:r.+]]
; WIN: movsd	(%rcx,%[[REG1]],8), %xmm0
; WIN: movhpd	(%rcx,%[[REG2]],8), %xmm0
; WIN: movsd	(%rcx,%[[REG3]],8), %xmm1
; WIN: movhpd	(%rcx,%[[REG4]],8), %xmm1

define <4 x double> @foo(double* %p, <4 x i32>* %i, <4 x i32>* %h) nounwind {
  %a = load <4 x i32>* %i
  %b = load <4 x i32>* %h
  %j = and <4 x i32> %a, %b
  %d0 = extractelement <4 x i32> %j, i32 0
  %d1 = extractelement <4 x i32> %j, i32 1
  %d2 = extractelement <4 x i32> %j, i32 2
  %d3 = extractelement <4 x i32> %j, i32 3
  %q0 = getelementptr double* %p, i32 %d0
  %q1 = getelementptr double* %p, i32 %d1
  %q2 = getelementptr double* %p, i32 %d2
  %q3 = getelementptr double* %p, i32 %d3
  %r0 = load double* %q0
  %r1 = load double* %q1
  %r2 = load double* %q2
  %r3 = load double* %q3
  %v0 = insertelement <4 x double> undef, double %r0, i32 0
  %v1 = insertelement <4 x double> %v0, double %r1, i32 1
  %v2 = insertelement <4 x double> %v1, double %r2, i32 2
  %v3 = insertelement <4 x double> %v2, double %r3, i32 3
  ret <4 x double> %v3
}