@@ -1013,14 +1013,13 @@ define <2 x i64> @arg_i64_v2i64(<2 x i64> %v, i64 %x, i32 %y) nounwind {
1013
1013
; X86AVX2-NEXT: movl 12(%ebp), %ecx
1014
1014
; X86AVX2-NEXT: movl 16(%ebp), %edx
1015
1015
; X86AVX2-NEXT: vmovaps %xmm0, (%esp)
1016
- ; X86AVX2-NEXT: leal (%edx,%edx), %esi
1017
- ; X86AVX2-NEXT: andl $3 , %esi
1018
- ; X86AVX2-NEXT: movl %eax, (%esp,%esi,4 )
1016
+ ; X86AVX2-NEXT: leal 1 (%edx,%edx), %esi
1017
+ ; X86AVX2-NEXT: andl $1 , %edx
1018
+ ; X86AVX2-NEXT: movl %eax, (%esp,%edx,8 )
1019
1019
; X86AVX2-NEXT: vmovaps (%esp), %xmm0
1020
1020
; X86AVX2-NEXT: vmovaps %xmm0, {{[0-9]+}}(%esp)
1021
- ; X86AVX2-NEXT: leal 1(%edx,%edx), %eax
1022
- ; X86AVX2-NEXT: andl $3, %eax
1023
- ; X86AVX2-NEXT: movl %ecx, 16(%esp,%eax,4)
1021
+ ; X86AVX2-NEXT: andl $3, %esi
1022
+ ; X86AVX2-NEXT: movl %ecx, 16(%esp,%esi,4)
1024
1023
; X86AVX2-NEXT: vmovaps {{[0-9]+}}(%esp), %xmm0
1025
1024
; X86AVX2-NEXT: leal -4(%ebp), %esp
1026
1025
; X86AVX2-NEXT: popl %esi
@@ -1362,14 +1361,13 @@ define <2 x i64> @load_i64_v2i64(<2 x i64> %v, ptr %p, i32 %y) nounwind {
1362
1361
; X86AVX2-NEXT: movl (%ecx), %edx
1363
1362
; X86AVX2-NEXT: movl 4(%ecx), %ecx
1364
1363
; X86AVX2-NEXT: vmovaps %xmm0, (%esp)
1365
- ; X86AVX2-NEXT: leal (%eax,%eax), %esi
1366
- ; X86AVX2-NEXT: andl $3 , %esi
1367
- ; X86AVX2-NEXT: movl %edx, (%esp,%esi,4 )
1364
+ ; X86AVX2-NEXT: leal 1 (%eax,%eax), %esi
1365
+ ; X86AVX2-NEXT: andl $1 , %eax
1366
+ ; X86AVX2-NEXT: movl %edx, (%esp,%eax,8 )
1368
1367
; X86AVX2-NEXT: vmovaps (%esp), %xmm0
1369
1368
; X86AVX2-NEXT: vmovaps %xmm0, {{[0-9]+}}(%esp)
1370
- ; X86AVX2-NEXT: leal 1(%eax,%eax), %eax
1371
- ; X86AVX2-NEXT: andl $3, %eax
1372
- ; X86AVX2-NEXT: movl %ecx, 16(%esp,%eax,4)
1369
+ ; X86AVX2-NEXT: andl $3, %esi
1370
+ ; X86AVX2-NEXT: movl %ecx, 16(%esp,%esi,4)
1373
1371
; X86AVX2-NEXT: vmovaps {{[0-9]+}}(%esp), %xmm0
1374
1372
; X86AVX2-NEXT: leal -4(%ebp), %esp
1375
1373
; X86AVX2-NEXT: popl %esi
@@ -1746,14 +1744,13 @@ define <4 x i64> @arg_i64_v4i64(<4 x i64> %v, i64 %x, i32 %y) nounwind {
1746
1744
; X86AVX2-NEXT: movl 12(%ebp), %ecx
1747
1745
; X86AVX2-NEXT: movl 16(%ebp), %edx
1748
1746
; X86AVX2-NEXT: vmovaps %ymm0, (%esp)
1749
- ; X86AVX2-NEXT: leal (%edx,%edx), %esi
1750
- ; X86AVX2-NEXT: andl $7 , %esi
1751
- ; X86AVX2-NEXT: movl %eax, (%esp,%esi,4 )
1747
+ ; X86AVX2-NEXT: leal 1 (%edx,%edx), %esi
1748
+ ; X86AVX2-NEXT: andl $3 , %edx
1749
+ ; X86AVX2-NEXT: movl %eax, (%esp,%edx,8 )
1752
1750
; X86AVX2-NEXT: vmovaps (%esp), %ymm0
1753
1751
; X86AVX2-NEXT: vmovaps %ymm0, {{[0-9]+}}(%esp)
1754
- ; X86AVX2-NEXT: leal 1(%edx,%edx), %eax
1755
- ; X86AVX2-NEXT: andl $7, %eax
1756
- ; X86AVX2-NEXT: movl %ecx, 32(%esp,%eax,4)
1752
+ ; X86AVX2-NEXT: andl $7, %esi
1753
+ ; X86AVX2-NEXT: movl %ecx, 32(%esp,%esi,4)
1757
1754
; X86AVX2-NEXT: vmovaps {{[0-9]+}}(%esp), %ymm0
1758
1755
; X86AVX2-NEXT: leal -4(%ebp), %esp
1759
1756
; X86AVX2-NEXT: popl %esi
@@ -2128,14 +2125,13 @@ define <4 x i64> @load_i64_v4i64(<4 x i64> %v, ptr %p, i32 %y) nounwind {
2128
2125
; X86AVX2-NEXT: movl (%ecx), %edx
2129
2126
; X86AVX2-NEXT: movl 4(%ecx), %ecx
2130
2127
; X86AVX2-NEXT: vmovaps %ymm0, (%esp)
2131
- ; X86AVX2-NEXT: leal (%eax,%eax), %esi
2132
- ; X86AVX2-NEXT: andl $7 , %esi
2133
- ; X86AVX2-NEXT: movl %edx, (%esp,%esi,4 )
2128
+ ; X86AVX2-NEXT: leal 1 (%eax,%eax), %esi
2129
+ ; X86AVX2-NEXT: andl $3 , %eax
2130
+ ; X86AVX2-NEXT: movl %edx, (%esp,%eax,8 )
2134
2131
; X86AVX2-NEXT: vmovaps (%esp), %ymm0
2135
2132
; X86AVX2-NEXT: vmovaps %ymm0, {{[0-9]+}}(%esp)
2136
- ; X86AVX2-NEXT: leal 1(%eax,%eax), %eax
2137
- ; X86AVX2-NEXT: andl $7, %eax
2138
- ; X86AVX2-NEXT: movl %ecx, 32(%esp,%eax,4)
2133
+ ; X86AVX2-NEXT: andl $7, %esi
2134
+ ; X86AVX2-NEXT: movl %ecx, 32(%esp,%esi,4)
2139
2135
; X86AVX2-NEXT: vmovaps {{[0-9]+}}(%esp), %ymm0
2140
2136
; X86AVX2-NEXT: leal -4(%ebp), %esp
2141
2137
; X86AVX2-NEXT: popl %esi
0 commit comments