1
1
//@ assembly-output: emit-asm
2
- //@ only- x86
3
- // FIXME(#114479): LLVM miscompiles loading and storing `f32` and `f64` when SSE is disabled.
4
- // There's no compiletest directive to ignore a test on i586 only, so just always explicitly enable
5
- // SSE2 .
6
- // Use the same target CPU as `i686` so that LLVM orders the instructions in the same order.
7
- //@ compile-flags: -Ctarget-feature=+sse2 -Ctarget-cpu=pentium4
2
+ //@ needs-llvm-components: x86
3
+ //@ revisions: sse nosse
4
+ //@[sse] compile-flags: --target i686-unknown-linux-gnu
5
+ // We make SSE available but don't use it for the ABI .
6
+ //@[nosse] compile-flags: -- target i586-unknown-linux-gnu -Ctarget-feature=+sse2 -Ctarget-cpu=pentium4
7
+
8
8
// Force frame pointers to make ASM more consistent between targets
9
9
//@ compile-flags: -O -C force-frame-pointers
10
10
//@ filecheck-flags: --implicit-check-not fld --implicit-check-not fst
11
- //@ revisions: normal win
12
- //@[normal] ignore-windows
13
- //@[win] only-windows
14
11
15
- #![ crate_type = "lib" ]
16
12
#![ feature( f16, f128) ]
13
+ #![ feature( no_core, lang_items, rustc_attrs, repr_simd) ]
14
+ #![ no_core]
15
+ #![ crate_type = "lib" ]
16
+
17
+ #[ lang = "sized" ]
18
+ trait Sized { }
19
+
20
+ #[ lang = "copy" ]
21
+ trait Copy { }
22
+
23
+ impl Copy for f16 { }
24
+ impl Copy for f32 { }
25
+ impl Copy for f64 { }
26
+ impl Copy for f128 { }
27
+ impl Copy for usize { }
17
28
18
29
// Tests that returning `f32` and `f64` with the "Rust" ABI on 32-bit x86 doesn't use the x87
19
30
// floating point stack, as loading and storing `f32`s and `f64`s to and from the x87 stack quietens
24
35
// CHECK-LABEL: return_f32:
25
36
#[ no_mangle]
26
37
pub fn return_f32 ( x : f32 ) -> f32 {
27
- // CHECK: movl {{.*}}(%ebp), %eax
38
+ // sse: movss {{.*}}(%ebp), %xmm0
39
+ // nosse: movl {{.*}}(%ebp), %eax
28
40
// CHECK-NOT: ax
29
41
// CHECK: retl
30
42
x
@@ -33,9 +45,11 @@ pub fn return_f32(x: f32) -> f32 {
33
45
// CHECK-LABEL: return_f64:
34
46
#[ no_mangle]
35
47
pub fn return_f64 ( x : f64 ) -> f64 {
36
- // CHECK: movl [[#%d,OFFSET:]](%ebp), %[[PTR:.*]]
37
- // CHECK-NEXT: movsd [[#%d,OFFSET+4]](%ebp), %[[VAL:.*]]
38
- // CHECK-NEXT: movsd %[[VAL]], (%[[PTR]])
48
+ // nosse: movl [[#%d,OFFSET:]](%ebp), %[[PTR:.*]]
49
+ // nosse-NEXT: movsd [[#%d,OFFSET+4]](%ebp), %[[VAL:.*]]
50
+ // nosse-NEXT: movsd %[[VAL]], (%[[PTR]])
51
+ // sse: movsd {{.*}}(%ebp), %xmm0
52
+ // sse-NOT: ax
39
53
// CHECK: retl
40
54
x
41
55
}
@@ -148,7 +162,8 @@ pub unsafe fn call_f32(x: &mut f32) {
148
162
}
149
163
// CHECK: movl {{.*}}(%ebp), %[[PTR:.*]]
150
164
// CHECK: calll {{()|_}}get_f32
151
- // CHECK-NEXT: movl %eax, (%[[PTR]])
165
+ // sse-NEXT: movss %xmm0, (%[[PTR]])
166
+ // nosse-NEXT: movl %eax, (%[[PTR]])
152
167
* x = get_f32 ( ) ;
153
168
}
154
169
@@ -160,8 +175,9 @@ pub unsafe fn call_f64(x: &mut f64) {
160
175
}
161
176
// CHECK: movl {{.*}}(%ebp), %[[PTR:.*]]
162
177
// CHECK: calll {{()|_}}get_f64
163
- // CHECK: movsd {{.*}}(%{{ebp|esp}}), %[[VAL:.*]]
164
- // CHECK-NEXT: movsd %[[VAL:.*]], (%[[PTR]])
178
+ // sse: movlps %xmm0, (%[[PTR]])
179
+ // nosse: movsd {{.*}}(%{{ebp|esp}}), %[[VAL:.*]]
180
+ // nosse-NEXT: movsd %[[VAL:.*]], (%[[PTR]])
165
181
* x = get_f64 ( ) ;
166
182
}
167
183
@@ -190,10 +206,8 @@ pub unsafe fn call_f64_f64(x: &mut (f64, f64)) {
190
206
}
191
207
// CHECK: movl {{.*}}(%ebp), %[[PTR:.*]]
192
208
// CHECK: calll {{()|_}}get_f64_f64
193
- // normal: movsd [[#%d,OFFSET:]](%ebp), %[[VAL1:.*]]
194
- // normal-NEXT: movsd [[#%d,OFFSET+8]](%ebp), %[[VAL2:.*]]
195
- // win: movsd (%esp), %[[VAL1:.*]]
196
- // win-NEXT: movsd 8(%esp), %[[VAL2:.*]]
209
+ // CHECK: movsd [[#%d,OFFSET:]](%ebp), %[[VAL1:.*]]
210
+ // CHECK-NEXT: movsd [[#%d,OFFSET+8]](%ebp), %[[VAL2:.*]]
197
211
// CHECK-NEXT: movsd %[[VAL1]], (%[[PTR]])
198
212
// CHECK-NEXT: movsd %[[VAL2]], 8(%[[PTR]])
199
213
* x = get_f64_f64 ( ) ;
@@ -207,13 +221,10 @@ pub unsafe fn call_f32_f64(x: &mut (f32, f64)) {
207
221
}
208
222
// CHECK: movl {{.*}}(%ebp), %[[PTR:.*]]
209
223
// CHECK: calll {{()|_}}get_f32_f64
210
- // normal: movss [[#%d,OFFSET:]](%ebp), %[[VAL1:.*]]
211
- // normal-NEXT: movsd [[#%d,OFFSET+4]](%ebp), %[[VAL2:.*]]
212
- // win: movss (%esp), %[[VAL1:.*]]
213
- // win-NEXT: movsd 8(%esp), %[[VAL2:.*]]
224
+ // CHECK: movss [[#%d,OFFSET:]](%ebp), %[[VAL1:.*]]
225
+ // CHECK-NEXT: movsd [[#%d,OFFSET+4]](%ebp), %[[VAL2:.*]]
214
226
// CHECK-NEXT: movss %[[VAL1]], (%[[PTR]])
215
- // normal-NEXT: movsd %[[VAL2]], 4(%[[PTR]])
216
- // win-NEXT: movsd %[[VAL2]], 8(%[[PTR]])
227
+ // CHECK-NEXT: movsd %[[VAL2]], 4(%[[PTR]])
217
228
* x = get_f32_f64 ( ) ;
218
229
}
219
230
@@ -225,10 +236,8 @@ pub unsafe fn call_f64_f32(x: &mut (f64, f32)) {
225
236
}
226
237
// CHECK: movl {{.*}}(%ebp), %[[PTR:.*]]
227
238
// CHECK: calll {{()|_}}get_f64_f32
228
- // normal: movsd [[#%d,OFFSET:]](%ebp), %[[VAL1:.*]]
229
- // normal-NEXT: movss [[#%d,OFFSET+8]](%ebp), %[[VAL2:.*]]
230
- // win: movsd (%esp), %[[VAL1:.*]]
231
- // win-NEXT: movss 8(%esp), %[[VAL2:.*]]
239
+ // CHECK: movsd [[#%d,OFFSET:]](%ebp), %[[VAL1:.*]]
240
+ // CHECK-NEXT: movss [[#%d,OFFSET+8]](%ebp), %[[VAL2:.*]]
232
241
// CHECK-NEXT: movsd %[[VAL1]], (%[[PTR]])
233
242
// CHECK-NEXT: movss %[[VAL2]], 8(%[[PTR]])
234
243
* x = get_f64_f32 ( ) ;
@@ -257,10 +266,8 @@ pub unsafe fn call_f64_other(x: &mut (f64, usize)) {
257
266
}
258
267
// CHECK: movl {{.*}}(%ebp), %[[PTR:.*]]
259
268
// CHECK: calll {{()|_}}get_f64_other
260
- // normal: movsd [[#%d,OFFSET:]](%ebp), %[[VAL1:.*]]
261
- // normal-NEXT: movl [[#%d,OFFSET+8]](%ebp), %[[VAL2:.*]]
262
- // win: movsd (%esp), %[[VAL1:.*]]
263
- // win-NEXT: movl 8(%esp), %[[VAL2:.*]]
269
+ // CHECK: movsd [[#%d,OFFSET:]](%ebp), %[[VAL1:.*]]
270
+ // CHECK-NEXT: movl [[#%d,OFFSET+8]](%ebp), %[[VAL2:.*]]
264
271
// CHECK-NEXT: movsd %[[VAL1]], (%[[PTR]])
265
272
// CHECK-NEXT: movl %[[VAL2]], 8(%[[PTR]])
266
273
* x = get_f64_other ( ) ;
@@ -289,13 +296,10 @@ pub unsafe fn call_other_f64(x: &mut (usize, f64)) {
289
296
}
290
297
// CHECK: movl {{.*}}(%ebp), %[[PTR:.*]]
291
298
// CHECK: calll {{()|_}}get_other_f64
292
- // normal: movl [[#%d,OFFSET:]](%ebp), %[[VAL1:.*]]
293
- // normal-NEXT: movsd [[#%d,OFFSET+4]](%ebp), %[[VAL2:.*]]
294
- // win: movl (%esp), %[[VAL1:.*]]
295
- // win-NEXT: movsd 8(%esp), %[[VAL2:.*]]
299
+ // CHECK: movl [[#%d,OFFSET:]](%ebp), %[[VAL1:.*]]
300
+ // CHECK-NEXT: movsd [[#%d,OFFSET+4]](%ebp), %[[VAL2:.*]]
296
301
// CHECK-NEXT: movl %[[VAL1]], (%[[PTR]])
297
- // normal-NEXT: movsd %[[VAL2]], 4(%[[PTR]])
298
- // win-NEXT: movsd %[[VAL2]], 8(%[[PTR]])
302
+ // CHECK-NEXT: movsd %[[VAL2]], 4(%[[PTR]])
299
303
* x = get_other_f64 ( ) ;
300
304
}
301
305
@@ -307,7 +311,8 @@ pub unsafe fn call_other_f64(x: &mut (usize, f64)) {
307
311
pub fn return_f16 ( x : f16 ) -> f16 {
308
312
// CHECK: pushl %ebp
309
313
// CHECK: movl %esp, %ebp
310
- // CHECK: movzwl 8(%ebp), %eax
314
+ // nosse: movzwl 8(%ebp), %eax
315
+ // sse: pinsrw $0, 8(%ebp), %xmm0
311
316
// CHECK: popl %ebp
312
317
// CHECK: retl
313
318
x
@@ -316,15 +321,18 @@ pub fn return_f16(x: f16) -> f16 {
316
321
// CHECK-LABEL: return_f128:
317
322
#[ no_mangle]
318
323
pub fn return_f128 ( x : f128 ) -> f128 {
319
- // CHECK: movl [[#%d,OFFSET:]](%ebp), %[[PTR:.*]]
320
- // CHECK-NEXT: movl [[#%d,OFFSET+4]](%ebp), %[[VAL1:.*]]
321
- // CHECK-NEXT: movl [[#%d,OFFSET+8]](%ebp), %[[VAL2:.*]]
322
- // CHECK-NEXT: movl [[#%d,OFFSET+12]](%ebp), %[[VAL3:.*]]
323
- // CHECK-NEXT: movl [[#%d,OFFSET+16]](%ebp), %[[VAL4:.*]]
324
- // CHECK-NEXT: movl %[[VAL4:.*]] 12(%[[PTR]])
325
- // CHECK-NEXT: movl %[[VAL3:.*]] 8(%[[PTR]])
326
- // CHECK-NEXT: movl %[[VAL2:.*]] 4(%[[PTR]])
327
- // CHECK-NEXT: movl %[[VAL1:.*]] (%[[PTR]])
324
+ // CHECK: pushl %ebp
325
+ // sse: movaps [[#%d,OFFSET:]](%ebp), %xmm0
326
+ // nosse: movl [[#%d,OFFSET:]](%ebp), %[[PTR:.*]]
327
+ // nosse-NEXT: movl [[#%d,OFFSET+4]](%ebp), %[[VAL1:.*]]
328
+ // nosse-NEXT: movl [[#%d,OFFSET+8]](%ebp), %[[VAL2:.*]]
329
+ // nosse-NEXT: movl [[#%d,OFFSET+12]](%ebp), %[[VAL3:.*]]
330
+ // nosse-NEXT: movl [[#%d,OFFSET+16]](%ebp), %[[VAL4:.*]]
331
+ // nosse-NEXT: movl %[[VAL4:.*]] 12(%[[PTR]])
332
+ // nosse-NEXT: movl %[[VAL3:.*]] 8(%[[PTR]])
333
+ // nosse-NEXT: movl %[[VAL2:.*]] 4(%[[PTR]])
334
+ // nosse-NEXT: movl %[[VAL1:.*]] (%[[PTR]])
335
+ // CHECK: popl %ebp
328
336
// CHECK: retl
329
337
x
330
338
}
0 commit comments