1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
|
; Test extensions of f32 to f128.
;
; RUN: llc < %s -mtriple=s390x-linux-gnu | FileCheck %s
; Check register extension.
define void @f1(fp128 *%dst, float %val) {
; CHECK-LABEL: f1:
; CHECK: lxebr %f0, %f0
; CHECK: std %f0, 0(%r2)
; CHECK: std %f2, 8(%r2)
; CHECK: br %r14
%res = fpext float %val to fp128
store fp128 %res, fp128 *%dst
ret void
}
; Check the low end of the LXEB range.
define void @f2(fp128 *%dst, float *%ptr) {
; CHECK-LABEL: f2:
; CHECK: lxeb %f0, 0(%r3)
; CHECK: std %f0, 0(%r2)
; CHECK: std %f2, 8(%r2)
; CHECK: br %r14
%val = load float , float *%ptr
%res = fpext float %val to fp128
store fp128 %res, fp128 *%dst
ret void
}
; Check the high end of the aligned LXEB range.
define void @f3(fp128 *%dst, float *%base) {
; CHECK-LABEL: f3:
; CHECK: lxeb %f0, 4092(%r3)
; CHECK: std %f0, 0(%r2)
; CHECK: std %f2, 8(%r2)
; CHECK: br %r14
%ptr = getelementptr float, float *%base, i64 1023
%val = load float , float *%ptr
%res = fpext float %val to fp128
store fp128 %res, fp128 *%dst
ret void
}
; Check the next word up, which needs separate address logic.
; Other sequences besides this one would be OK.
define void @f4(fp128 *%dst, float *%base) {
; CHECK-LABEL: f4:
; CHECK: aghi %r3, 4096
; CHECK: lxeb %f0, 0(%r3)
; CHECK: std %f0, 0(%r2)
; CHECK: std %f2, 8(%r2)
; CHECK: br %r14
%ptr = getelementptr float, float *%base, i64 1024
%val = load float , float *%ptr
%res = fpext float %val to fp128
store fp128 %res, fp128 *%dst
ret void
}
; Check negative displacements, which also need separate address logic.
define void @f5(fp128 *%dst, float *%base) {
; CHECK-LABEL: f5:
; CHECK: aghi %r3, -4
; CHECK: lxeb %f0, 0(%r3)
; CHECK: std %f0, 0(%r2)
; CHECK: std %f2, 8(%r2)
; CHECK: br %r14
%ptr = getelementptr float, float *%base, i64 -1
%val = load float , float *%ptr
%res = fpext float %val to fp128
store fp128 %res, fp128 *%dst
ret void
}
; Check that LXEB allows indices.
define void @f6(fp128 *%dst, float *%base, i64 %index) {
; CHECK-LABEL: f6:
; CHECK: sllg %r1, %r4, 2
; CHECK: lxeb %f0, 400(%r1,%r3)
; CHECK: std %f0, 0(%r2)
; CHECK: std %f2, 8(%r2)
; CHECK: br %r14
%ptr1 = getelementptr float, float *%base, i64 %index
%ptr2 = getelementptr float, float *%ptr1, i64 100
%val = load float , float *%ptr2
%res = fpext float %val to fp128
store fp128 %res, fp128 *%dst
ret void
}
; Test a case where we spill the source of at least one LXEBR. We want
; to use LXEB if possible.
define void @f7(fp128 *%ptr1, float *%ptr2) {
; CHECK-LABEL: f7:
; CHECK: lxeb {{%f[0-9]+}}, 16{{[04]}}(%r15)
; CHECK: br %r14
%val0 = load volatile float , float *%ptr2
%val1 = load volatile float , float *%ptr2
%val2 = load volatile float , float *%ptr2
%val3 = load volatile float , float *%ptr2
%val4 = load volatile float , float *%ptr2
%val5 = load volatile float , float *%ptr2
%val6 = load volatile float , float *%ptr2
%val7 = load volatile float , float *%ptr2
%val8 = load volatile float , float *%ptr2
%val9 = load volatile float , float *%ptr2
%val10 = load volatile float , float *%ptr2
%val11 = load volatile float , float *%ptr2
%val12 = load volatile float , float *%ptr2
%val13 = load volatile float , float *%ptr2
%val14 = load volatile float , float *%ptr2
%val15 = load volatile float , float *%ptr2
%val16 = load volatile float , float *%ptr2
%ext0 = fpext float %val0 to fp128
%ext1 = fpext float %val1 to fp128
%ext2 = fpext float %val2 to fp128
%ext3 = fpext float %val3 to fp128
%ext4 = fpext float %val4 to fp128
%ext5 = fpext float %val5 to fp128
%ext6 = fpext float %val6 to fp128
%ext7 = fpext float %val7 to fp128
%ext8 = fpext float %val8 to fp128
%ext9 = fpext float %val9 to fp128
%ext10 = fpext float %val10 to fp128
%ext11 = fpext float %val11 to fp128
%ext12 = fpext float %val12 to fp128
%ext13 = fpext float %val13 to fp128
%ext14 = fpext float %val14 to fp128
%ext15 = fpext float %val15 to fp128
%ext16 = fpext float %val16 to fp128
store volatile float %val0, float *%ptr2
store volatile float %val1, float *%ptr2
store volatile float %val2, float *%ptr2
store volatile float %val3, float *%ptr2
store volatile float %val4, float *%ptr2
store volatile float %val5, float *%ptr2
store volatile float %val6, float *%ptr2
store volatile float %val7, float *%ptr2
store volatile float %val8, float *%ptr2
store volatile float %val9, float *%ptr2
store volatile float %val10, float *%ptr2
store volatile float %val11, float *%ptr2
store volatile float %val12, float *%ptr2
store volatile float %val13, float *%ptr2
store volatile float %val14, float *%ptr2
store volatile float %val15, float *%ptr2
store volatile float %val16, float *%ptr2
store volatile fp128 %ext0, fp128 *%ptr1
store volatile fp128 %ext1, fp128 *%ptr1
store volatile fp128 %ext2, fp128 *%ptr1
store volatile fp128 %ext3, fp128 *%ptr1
store volatile fp128 %ext4, fp128 *%ptr1
store volatile fp128 %ext5, fp128 *%ptr1
store volatile fp128 %ext6, fp128 *%ptr1
store volatile fp128 %ext7, fp128 *%ptr1
store volatile fp128 %ext8, fp128 *%ptr1
store volatile fp128 %ext9, fp128 *%ptr1
store volatile fp128 %ext10, fp128 *%ptr1
store volatile fp128 %ext11, fp128 *%ptr1
store volatile fp128 %ext12, fp128 *%ptr1
store volatile fp128 %ext13, fp128 *%ptr1
store volatile fp128 %ext14, fp128 *%ptr1
store volatile fp128 %ext15, fp128 *%ptr1
store volatile fp128 %ext16, fp128 *%ptr1
ret void
}
|