|
1 | 1 | ; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
2 |
| -; RUN: llc -mtriple=x86_64-pc-linux < %s | FileCheck %s --check-prefix=X86-64 |
| 2 | +; RUN: llc -mtriple=x86_64-pc-linux < %s | FileCheck %s --check-prefix=X64 |
3 | 3 | ; RUN: llc -mtriple=i686-pc-linux < %s | FileCheck %s --check-prefix=X86
|
4 | 4 |
|
5 | 5 | define double @foo(double %0) #0 {
|
6 |
| -; X86-64-LABEL: foo: |
7 |
| -; X86-64: # %bb.0: |
8 |
| -; X86-64-NEXT: subq $24, %rsp |
9 |
| -; X86-64-NEXT: movsd %xmm0, (%rsp) # 8-byte Spill |
10 |
| -; X86-64-NEXT: movl $1024, %edi # imm = 0x400 |
11 |
| -; X86-64-NEXT: callq fesetround@PLT |
12 |
| -; X86-64-NEXT: movsd {{.*#+}} xmm1 = [1.0E+0,0.0E+0] |
13 |
| -; X86-64-NEXT: divsd (%rsp), %xmm1 # 8-byte Folded Reload |
14 |
| -; X86-64-NEXT: movsd %xmm1, {{[-0-9]+}}(%r{{[sb]}}p) # 8-byte Spill |
15 |
| -; X86-64-NEXT: movl $1024, %edi # imm = 0x400 |
16 |
| -; X86-64-NEXT: callq fesetround@PLT |
17 |
| -; X86-64-NEXT: movsd {{.*#+}} xmm0 = [1.0E+0,0.0E+0] |
18 |
| -; X86-64-NEXT: divsd (%rsp), %xmm0 # 8-byte Folded Reload |
19 |
| -; X86-64-NEXT: movsd %xmm0, {{[-0-9]+}}(%r{{[sb]}}p) # 8-byte Spill |
20 |
| -; X86-64-NEXT: movl $1024, %edi # imm = 0x400 |
21 |
| -; X86-64-NEXT: callq fesetround@PLT |
22 |
| -; X86-64-NEXT: movsd {{.*#+}} xmm2 = [1.0E+0,0.0E+0] |
23 |
| -; X86-64-NEXT: divsd (%rsp), %xmm2 # 8-byte Folded Reload |
24 |
| -; X86-64-NEXT: movsd {{[-0-9]+}}(%r{{[sb]}}p), %xmm0 # 8-byte Reload |
25 |
| -; X86-64-NEXT: # xmm0 = mem[0],zero |
26 |
| -; X86-64-NEXT: movsd {{[-0-9]+}}(%r{{[sb]}}p), %xmm1 # 8-byte Reload |
27 |
| -; X86-64-NEXT: # xmm1 = mem[0],zero |
28 |
| -; X86-64-NEXT: callq fma@PLT |
29 |
| -; X86-64-NEXT: addq $24, %rsp |
30 |
| -; X86-64-NEXT: retq |
| 6 | +; X64-LABEL: foo: |
| 7 | +; X64: # %bb.0: |
| 8 | +; X64-NEXT: subq $24, %rsp |
| 9 | +; X64-NEXT: movsd %xmm0, (%rsp) # 8-byte Spill |
| 10 | +; X64-NEXT: movl $1024, %edi # imm = 0x400 |
| 11 | +; X64-NEXT: callq fesetround@PLT |
| 12 | +; X64-NEXT: movsd {{.*#+}} xmm1 = [1.0E+0,0.0E+0] |
| 13 | +; X64-NEXT: divsd (%rsp), %xmm1 # 8-byte Folded Reload |
| 14 | +; X64-NEXT: movsd %xmm1, {{[-0-9]+}}(%r{{[sb]}}p) # 8-byte Spill |
| 15 | +; X64-NEXT: movl $1024, %edi # imm = 0x400 |
| 16 | +; X64-NEXT: callq fesetround@PLT |
| 17 | +; X64-NEXT: movsd {{.*#+}} xmm0 = [1.0E+0,0.0E+0] |
| 18 | +; X64-NEXT: divsd (%rsp), %xmm0 # 8-byte Folded Reload |
| 19 | +; X64-NEXT: movsd %xmm0, {{[-0-9]+}}(%r{{[sb]}}p) # 8-byte Spill |
| 20 | +; X64-NEXT: movl $1024, %edi # imm = 0x400 |
| 21 | +; X64-NEXT: callq fesetround@PLT |
| 22 | +; X64-NEXT: movsd {{.*#+}} xmm2 = [1.0E+0,0.0E+0] |
| 23 | +; X64-NEXT: divsd (%rsp), %xmm2 # 8-byte Folded Reload |
| 24 | +; X64-NEXT: movsd {{[-0-9]+}}(%r{{[sb]}}p), %xmm0 # 8-byte Reload |
| 25 | +; X64-NEXT: # xmm0 = mem[0],zero |
| 26 | +; X64-NEXT: movsd {{[-0-9]+}}(%r{{[sb]}}p), %xmm1 # 8-byte Reload |
| 27 | +; X64-NEXT: # xmm1 = mem[0],zero |
| 28 | +; X64-NEXT: callq fma@PLT |
| 29 | +; X64-NEXT: addq $24, %rsp |
| 30 | +; X64-NEXT: retq |
31 | 31 | ;
|
32 | 32 | ; X86-LABEL: foo:
|
33 | 33 | ; X86: # %bb.0:
|
|
0 commit comments