631 lines
18 KiB
LLVM
631 lines
18 KiB
LLVM
; NOTE: Assertions have been autogenerated by utils/update_llc_test_checks.py
|
|
; RUN: llc < %s -mtriple=i686-unknown | FileCheck %s --check-prefixes=X86
|
|
; RUN: llc < %s -mtriple=x86_64-unknown | FileCheck %s --check-prefixes=X64
|
|
; RUN: llc < %s -mtriple=x86_64-unknown -mattr=+tbm | FileCheck %s --check-prefixes=X64
|
|
|
|
; PR35908 - Fold ADD/SUB and bit extracts into ADC/SBB+BT
|
|
;
|
|
; int test_add_add(int x, int y, int z) { return ((x + y) + bool(z & (1 << 30))); }
|
|
; int test_add_sub(int x, int y, int z) { return ((x - y) + bool(z & (1 << 30))); }
|
|
; int test_sub_add(int x, int y, int z) { return ((x + y) - bool(z & (1 << 30))); }
|
|
; int test_sub_sub(int x, int y, int z) { return (x - (y - bool(z & (1 << 30)))); }
|
|
|
|
;
|
|
; Constant Bit Indices
|
|
;
|
|
|
|
define i32 @test_i32_add_add_idx(i32 %x, i32 %y, i32 %z) nounwind {
|
|
; X86-LABEL: test_i32_add_add_idx:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: btl $30, {{[0-9]+}}(%esp)
|
|
; X86-NEXT: adcl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_i32_add_add_idx:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: movl %edi, %eax
|
|
; X64-NEXT: btl $30, %edx
|
|
; X64-NEXT: adcl %esi, %eax
|
|
; X64-NEXT: retq
|
|
%add = add i32 %y, %x
|
|
%shift = lshr i32 %z, 30
|
|
%mask = and i32 %shift, 1
|
|
%add1 = add i32 %add, %mask
|
|
ret i32 %add1
|
|
}
|
|
|
|
define i32 @test_i32_add_add_commute_idx(i32 %x, i32 %y, i32 %z) nounwind {
|
|
; X86-LABEL: test_i32_add_add_commute_idx:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: btl $2, {{[0-9]+}}(%esp)
|
|
; X86-NEXT: adcl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_i32_add_add_commute_idx:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: movl %edi, %eax
|
|
; X64-NEXT: btl $2, %edx
|
|
; X64-NEXT: adcl %esi, %eax
|
|
; X64-NEXT: retq
|
|
%add = add i32 %y, %x
|
|
%shift = lshr i32 %z, 2
|
|
%mask = and i32 %shift, 1
|
|
%add1 = add i32 %mask, %add
|
|
ret i32 %add1
|
|
}
|
|
|
|
define i32 @test_i32_add_add_idx0(i32 %x, i32 %y, i32 %z) nounwind {
|
|
; X86-LABEL: test_i32_add_add_idx0:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NEXT: addl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NEXT: andl $1, %eax
|
|
; X86-NEXT: addl %ecx, %eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_i32_add_add_idx0:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: # kill: def $edx killed $edx def $rdx
|
|
; X64-NEXT: # kill: def $edi killed $edi def $rdi
|
|
; X64-NEXT: andl $1, %edx
|
|
; X64-NEXT: leal (%rdx,%rdi), %eax
|
|
; X64-NEXT: addl %esi, %eax
|
|
; X64-NEXT: retq
|
|
%add = add i32 %y, %x
|
|
%mask = and i32 %z, 1
|
|
%add1 = add i32 %mask, %add
|
|
ret i32 %add1
|
|
}
|
|
|
|
define i24 @test_i24_add_add_idx(i24 %x, i24 %y, i24 %z) nounwind {
|
|
; X86-LABEL: test_i24_add_add_idx:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: btl $15, {{[0-9]+}}(%esp)
|
|
; X86-NEXT: adcl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_i24_add_add_idx:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: movl %edi, %eax
|
|
; X64-NEXT: btl $15, %edx
|
|
; X64-NEXT: adcl %esi, %eax
|
|
; X64-NEXT: retq
|
|
%add = add i24 %y, %x
|
|
%shift = lshr i24 %z, 15
|
|
%mask = and i24 %shift, 1
|
|
%add1 = add i24 %add, %mask
|
|
ret i24 %add1
|
|
}
|
|
|
|
define i128 @test_i128_add_add_idx(i128 %x, i128 %y, i128 %z) nounwind {
|
|
; X86-LABEL: test_i128_add_add_idx:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: pushl %edi
|
|
; X86-NEXT: pushl %esi
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %esi
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %edi
|
|
; X86-NEXT: addl {{[0-9]+}}(%esp), %esi
|
|
; X86-NEXT: adcl {{[0-9]+}}(%esp), %edi
|
|
; X86-NEXT: adcl {{[0-9]+}}(%esp), %edx
|
|
; X86-NEXT: adcl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NEXT: btl $5, {{[0-9]+}}(%esp)
|
|
; X86-NEXT: adcl $0, %esi
|
|
; X86-NEXT: adcl $0, %edi
|
|
; X86-NEXT: adcl $0, %edx
|
|
; X86-NEXT: adcl $0, %ecx
|
|
; X86-NEXT: movl %edi, 4(%eax)
|
|
; X86-NEXT: movl %esi, (%eax)
|
|
; X86-NEXT: movl %edx, 8(%eax)
|
|
; X86-NEXT: movl %ecx, 12(%eax)
|
|
; X86-NEXT: popl %esi
|
|
; X86-NEXT: popl %edi
|
|
; X86-NEXT: retl $4
|
|
;
|
|
; X64-LABEL: test_i128_add_add_idx:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: movq %rdi, %rax
|
|
; X64-NEXT: addq %rdx, %rax
|
|
; X64-NEXT: adcq %rcx, %rsi
|
|
; X64-NEXT: btl $5, %r9d
|
|
; X64-NEXT: adcq $0, %rax
|
|
; X64-NEXT: adcq $0, %rsi
|
|
; X64-NEXT: movq %rsi, %rdx
|
|
; X64-NEXT: retq
|
|
%add = add i128 %y, %x
|
|
%shift = lshr i128 %z, 69
|
|
%mask = and i128 %shift, 1
|
|
%add1 = add i128 %add, %mask
|
|
ret i128 %add1
|
|
}
|
|
|
|
define i32 @test_i32_add_sub_idx(i32 %x, i32 %y, i32 %z) nounwind {
|
|
; X86-LABEL: test_i32_add_sub_idx:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NEXT: subl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NEXT: shrl $31, %eax
|
|
; X86-NEXT: addl %ecx, %eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_i32_add_sub_idx:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: # kill: def $edx killed $edx def $rdx
|
|
; X64-NEXT: # kill: def $edi killed $edi def $rdi
|
|
; X64-NEXT: subl %esi, %edi
|
|
; X64-NEXT: shrl $31, %edx
|
|
; X64-NEXT: leal (%rdx,%rdi), %eax
|
|
; X64-NEXT: retq
|
|
%sub = sub i32 %x, %y
|
|
%shift = lshr i32 %z, 31
|
|
%mask = and i32 %shift, 1
|
|
%add = add i32 %sub, %mask
|
|
ret i32 %add
|
|
}
|
|
|
|
define i32 @test_i32_add_sub_commute_idx(i32 %x, i32 %y, i32 %z) nounwind {
|
|
; X86-LABEL: test_i32_add_sub_commute_idx:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: subl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: btl $8, {{[0-9]+}}(%esp)
|
|
; X86-NEXT: adcl $0, %eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_i32_add_sub_commute_idx:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: movl %edi, %eax
|
|
; X64-NEXT: subl %esi, %eax
|
|
; X64-NEXT: btl $8, %edx
|
|
; X64-NEXT: adcl $0, %eax
|
|
; X64-NEXT: retq
|
|
%sub = sub i32 %x, %y
|
|
%shift = lshr i32 %z, 8
|
|
%mask = and i32 %shift, 1
|
|
%add = add i32 %mask, %sub
|
|
ret i32 %add
|
|
}
|
|
|
|
define i32 @test_i32_sub_add_idx(i32 %x, i32 %y, i32 %z) nounwind {
|
|
; X86-LABEL: test_i32_sub_add_idx:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: addl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: btl $1, {{[0-9]+}}(%esp)
|
|
; X86-NEXT: sbbl $0, %eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_i32_sub_add_idx:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: # kill: def $esi killed $esi def $rsi
|
|
; X64-NEXT: # kill: def $edi killed $edi def $rdi
|
|
; X64-NEXT: leal (%rdi,%rsi), %eax
|
|
; X64-NEXT: btl $1, %edx
|
|
; X64-NEXT: sbbl $0, %eax
|
|
; X64-NEXT: retq
|
|
%add = add i32 %y, %x
|
|
%shift = lshr i32 %z, 1
|
|
%mask = and i32 %shift, 1
|
|
%sub = sub i32 %add, %mask
|
|
ret i32 %sub
|
|
}
|
|
|
|
define i32 @test_i32_sub_add_commute_idx(i32 %x, i32 %y, i32 %z) nounwind {
|
|
; X86-LABEL: test_i32_sub_add_commute_idx:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: btl $1, {{[0-9]+}}(%esp)
|
|
; X86-NEXT: sbbl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_i32_sub_add_commute_idx:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: movl %edi, %eax
|
|
; X64-NEXT: btl $1, %edx
|
|
; X64-NEXT: sbbl %esi, %eax
|
|
; X64-NEXT: retq
|
|
%shift = lshr i32 %z, 1
|
|
%mask = and i32 %shift, 1
|
|
%add = add i32 %y, %mask
|
|
%sub = sub i32 %x, %add
|
|
ret i32 %sub
|
|
}
|
|
|
|
define i32 @test_i32_sub_add_sext_idx(i32 %x, i32 %y, i32 %z) nounwind {
|
|
; X86-LABEL: test_i32_sub_add_sext_idx:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NEXT: shll $25, %ecx
|
|
; X86-NEXT: sarl $31, %ecx
|
|
; X86-NEXT: addl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NEXT: subl %ecx, %eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_i32_sub_add_sext_idx:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: movl %edi, %eax
|
|
; X64-NEXT: shll $25, %edx
|
|
; X64-NEXT: sarl $31, %edx
|
|
; X64-NEXT: addl %esi, %edx
|
|
; X64-NEXT: subl %edx, %eax
|
|
; X64-NEXT: retq
|
|
%shl = shl i32 %z, 25
|
|
%sext = ashr i32 %shl, 31
|
|
%add = add i32 %y, %sext
|
|
%sub = sub i32 %x, %add
|
|
ret i32 %sub
|
|
}
|
|
|
|
define i32 @test_i32_sub_sub_idx(i32 %x, i32 %y, i32 %z) nounwind {
|
|
; X86-LABEL: test_i32_sub_sub_idx:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: btl $16, {{[0-9]+}}(%esp)
|
|
; X86-NEXT: adcl $0, %eax
|
|
; X86-NEXT: subl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_i32_sub_sub_idx:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: movl %edi, %eax
|
|
; X64-NEXT: btl $16, %edx
|
|
; X64-NEXT: adcl $0, %eax
|
|
; X64-NEXT: subl %esi, %eax
|
|
; X64-NEXT: retq
|
|
%shift = lshr i32 %z, 16
|
|
%mask = and i32 %shift, 1
|
|
%sub0 = sub i32 %y, %mask
|
|
%sub1 = sub i32 %x, %sub0
|
|
ret i32 %sub1
|
|
}
|
|
|
|
define i32 @test_i32_sub_sub_commute_idx(i32 %x, i32 %y, i32 %z) nounwind {
|
|
; X86-LABEL: test_i32_sub_sub_commute_idx:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: btl $15, {{[0-9]+}}(%esp)
|
|
; X86-NEXT: sbbl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_i32_sub_sub_commute_idx:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: movl %edi, %eax
|
|
; X64-NEXT: btl $15, %edx
|
|
; X64-NEXT: sbbl %esi, %eax
|
|
; X64-NEXT: retq
|
|
%shift = lshr i32 %z, 15
|
|
%mask = and i32 %shift, 1
|
|
%sub0 = sub i32 %x, %y
|
|
%sub1 = sub i32 %sub0, %mask
|
|
ret i32 %sub1
|
|
}
|
|
|
|
define i32 @test_i32_sub_sum_idx(i32 %x, i32 %y, i32 %z) nounwind {
|
|
; X86-LABEL: test_i32_sub_sum_idx:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: addl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: btl $30, {{[0-9]+}}(%esp)
|
|
; X86-NEXT: sbbl $0, %eax
|
|
; X86-NEXT: negl %eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_i32_sub_sum_idx:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: # kill: def $esi killed $esi def $rsi
|
|
; X64-NEXT: # kill: def $edi killed $edi def $rdi
|
|
; X64-NEXT: leal (%rdi,%rsi), %eax
|
|
; X64-NEXT: btl $30, %edx
|
|
; X64-NEXT: sbbl $0, %eax
|
|
; X64-NEXT: negl %eax
|
|
; X64-NEXT: retq
|
|
%shift = lshr i32 %z, 30
|
|
%mask = and i32 %shift, 1
|
|
%add = add i32 %y, %x
|
|
%sub = sub i32 %mask, %add
|
|
ret i32 %sub
|
|
}
|
|
|
|
;
|
|
; Variable Bit Indices
|
|
;
|
|
|
|
define i32 @test_i32_add_add_var(i32 %x, i32 %y, i32 %z, i32 %w) nounwind {
|
|
; X86-LABEL: test_i32_add_add_var:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
|
|
; X86-NEXT: btl %ecx, %edx
|
|
; X86-NEXT: adcl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_i32_add_add_var:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: movl %edi, %eax
|
|
; X64-NEXT: btl %ecx, %edx
|
|
; X64-NEXT: adcl %esi, %eax
|
|
; X64-NEXT: retq
|
|
%add = add i32 %y, %x
|
|
%shift = lshr i32 %z, %w
|
|
%mask = and i32 %shift, 1
|
|
%add1 = add i32 %add, %mask
|
|
ret i32 %add1
|
|
}
|
|
|
|
define i32 @test_i32_add_add_commute_var(i32 %x, i32 %y, i32 %z, i32 %w) nounwind {
|
|
; X86-LABEL: test_i32_add_add_commute_var:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
|
|
; X86-NEXT: btl %ecx, %edx
|
|
; X86-NEXT: adcl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_i32_add_add_commute_var:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: movl %edi, %eax
|
|
; X64-NEXT: btl %ecx, %edx
|
|
; X64-NEXT: adcl %esi, %eax
|
|
; X64-NEXT: retq
|
|
%add = add i32 %y, %x
|
|
%shift = lshr i32 %z, %w
|
|
%mask = and i32 %shift, 1
|
|
%add1 = add i32 %mask, %add
|
|
ret i32 %add1
|
|
}
|
|
|
|
define i64 @test_i64_add_add_var(i64 %x, i64 %y, i64 %z, i64 %w) nounwind {
|
|
; X86-LABEL: test_i64_add_add_var:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: pushl %ebx
|
|
; X86-NEXT: pushl %edi
|
|
; X86-NEXT: pushl %esi
|
|
; X86-NEXT: movzbl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %esi
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %ebx
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
|
|
; X86-NEXT: addl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: adcl {{[0-9]+}}(%esp), %edx
|
|
; X86-NEXT: movl %ebx, %edi
|
|
; X86-NEXT: shrl %cl, %edi
|
|
; X86-NEXT: shrdl %cl, %ebx, %esi
|
|
; X86-NEXT: testb $32, %cl
|
|
; X86-NEXT: jne .LBB15_2
|
|
; X86-NEXT: # %bb.1:
|
|
; X86-NEXT: movl %esi, %edi
|
|
; X86-NEXT: .LBB15_2:
|
|
; X86-NEXT: andl $1, %edi
|
|
; X86-NEXT: addl %edi, %eax
|
|
; X86-NEXT: adcl $0, %edx
|
|
; X86-NEXT: popl %esi
|
|
; X86-NEXT: popl %edi
|
|
; X86-NEXT: popl %ebx
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_i64_add_add_var:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: movq %rdi, %rax
|
|
; X64-NEXT: btq %rcx, %rdx
|
|
; X64-NEXT: adcq %rsi, %rax
|
|
; X64-NEXT: retq
|
|
%add = add i64 %y, %x
|
|
%shift = lshr i64 %z, %w
|
|
%mask = and i64 %shift, 1
|
|
%add1 = add i64 %add, %mask
|
|
ret i64 %add1
|
|
}
|
|
|
|
define i32 @test_i32_add_sub_var(i32 %x, i32 %y, i32 %z, i32 %w) nounwind {
|
|
; X86-LABEL: test_i32_add_sub_var:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: subl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: btl %ecx, %edx
|
|
; X86-NEXT: adcl $0, %eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_i32_add_sub_var:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: movl %edi, %eax
|
|
; X64-NEXT: subl %esi, %eax
|
|
; X64-NEXT: btl %ecx, %edx
|
|
; X64-NEXT: adcl $0, %eax
|
|
; X64-NEXT: retq
|
|
%sub = sub i32 %x, %y
|
|
%shift = lshr i32 %z, %w
|
|
%mask = and i32 %shift, 1
|
|
%add = add i32 %sub, %mask
|
|
ret i32 %add
|
|
}
|
|
|
|
define i32 @test_i32_add_sub_commute_var(i32 %x, i32 %y, i32 %z, i32 %w) nounwind {
|
|
; X86-LABEL: test_i32_add_sub_commute_var:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: subl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: btl %ecx, %edx
|
|
; X86-NEXT: adcl $0, %eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_i32_add_sub_commute_var:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: movl %edi, %eax
|
|
; X64-NEXT: subl %esi, %eax
|
|
; X64-NEXT: btl %ecx, %edx
|
|
; X64-NEXT: adcl $0, %eax
|
|
; X64-NEXT: retq
|
|
%sub = sub i32 %x, %y
|
|
%shift = lshr i32 %z, %w
|
|
%mask = and i32 %shift, 1
|
|
%add = add i32 %mask, %sub
|
|
ret i32 %add
|
|
}
|
|
|
|
define i32 @test_i32_sub_add_var(i32 %x, i32 %y, i32 %z, i32 %w) nounwind {
|
|
; X86-LABEL: test_i32_sub_add_var:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: addl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: btl %ecx, %edx
|
|
; X86-NEXT: sbbl $0, %eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_i32_sub_add_var:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: # kill: def $esi killed $esi def $rsi
|
|
; X64-NEXT: # kill: def $edi killed $edi def $rdi
|
|
; X64-NEXT: leal (%rdi,%rsi), %eax
|
|
; X64-NEXT: btl %ecx, %edx
|
|
; X64-NEXT: sbbl $0, %eax
|
|
; X64-NEXT: retq
|
|
%add = add i32 %y, %x
|
|
%shift = lshr i32 %z, %w
|
|
%mask = and i32 %shift, 1
|
|
%sub = sub i32 %add, %mask
|
|
ret i32 %sub
|
|
}
|
|
|
|
define i32 @test_i32_sub_add_commute_var(i32 %x, i32 %y, i32 %z, i32 %w) nounwind {
|
|
; X86-LABEL: test_i32_sub_add_commute_var:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
|
|
; X86-NEXT: btl %ecx, %edx
|
|
; X86-NEXT: sbbl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_i32_sub_add_commute_var:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: movl %edi, %eax
|
|
; X64-NEXT: btl %ecx, %edx
|
|
; X64-NEXT: sbbl %esi, %eax
|
|
; X64-NEXT: retq
|
|
%shift = lshr i32 %z, %w
|
|
%mask = and i32 %shift, 1
|
|
%add = add i32 %y, %mask
|
|
%sub = sub i32 %x, %add
|
|
ret i32 %sub
|
|
}
|
|
|
|
define i32 @test_i32_sub_add_sext_var(i32 %x, i32 %y, i32 %z, i32 %w) nounwind {
|
|
; X86-LABEL: test_i32_sub_add_sext_var:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movzbl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
|
|
; X86-NEXT: shll %cl, %edx
|
|
; X86-NEXT: sarl $31, %edx
|
|
; X86-NEXT: addl {{[0-9]+}}(%esp), %edx
|
|
; X86-NEXT: subl %edx, %eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_i32_sub_add_sext_var:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: movl %edi, %eax
|
|
; X64-NEXT: # kill: def $cl killed $cl killed $ecx
|
|
; X64-NEXT: shll %cl, %edx
|
|
; X64-NEXT: sarl $31, %edx
|
|
; X64-NEXT: addl %esi, %edx
|
|
; X64-NEXT: subl %edx, %eax
|
|
; X64-NEXT: retq
|
|
%shl = shl i32 %z, %w
|
|
%sext = ashr i32 %shl, 31
|
|
%add = add i32 %y, %sext
|
|
%sub = sub i32 %x, %add
|
|
ret i32 %sub
|
|
}
|
|
|
|
define i32 @test_i32_sub_sub_var(i32 %x, i32 %y, i32 %z, i32 %w) nounwind {
|
|
; X86-LABEL: test_i32_sub_sub_var:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
|
|
; X86-NEXT: btl %ecx, %edx
|
|
; X86-NEXT: adcl $0, %eax
|
|
; X86-NEXT: subl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_i32_sub_sub_var:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: movl %edi, %eax
|
|
; X64-NEXT: btl %ecx, %edx
|
|
; X64-NEXT: adcl $0, %eax
|
|
; X64-NEXT: subl %esi, %eax
|
|
; X64-NEXT: retq
|
|
%shift = lshr i32 %z, %w
|
|
%mask = and i32 %shift, 1
|
|
%sub0 = sub i32 %y, %mask
|
|
%sub1 = sub i32 %x, %sub0
|
|
ret i32 %sub1
|
|
}
|
|
|
|
define i32 @test_i32_sub_sub_commute_var(i32 %x, i32 %y, i32 %z, i32 %w) nounwind {
|
|
; X86-LABEL: test_i32_sub_sub_commute_var:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
|
|
; X86-NEXT: btl %ecx, %edx
|
|
; X86-NEXT: sbbl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_i32_sub_sub_commute_var:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: movl %edi, %eax
|
|
; X64-NEXT: btl %ecx, %edx
|
|
; X64-NEXT: sbbl %esi, %eax
|
|
; X64-NEXT: retq
|
|
%shift = lshr i32 %z, %w
|
|
%mask = and i32 %shift, 1
|
|
%sub0 = sub i32 %x, %y
|
|
%sub1 = sub i32 %sub0, %mask
|
|
ret i32 %sub1
|
|
}
|
|
|
|
define i32 @test_i32_sub_sum_var(i32 %x, i32 %y, i32 %z, i32 %w) nounwind {
|
|
; X86-LABEL: test_i32_sub_sum_var:
|
|
; X86: # %bb.0:
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %ecx
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %edx
|
|
; X86-NEXT: movl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: addl {{[0-9]+}}(%esp), %eax
|
|
; X86-NEXT: btl %ecx, %edx
|
|
; X86-NEXT: sbbl $0, %eax
|
|
; X86-NEXT: negl %eax
|
|
; X86-NEXT: retl
|
|
;
|
|
; X64-LABEL: test_i32_sub_sum_var:
|
|
; X64: # %bb.0:
|
|
; X64-NEXT: # kill: def $esi killed $esi def $rsi
|
|
; X64-NEXT: # kill: def $edi killed $edi def $rdi
|
|
; X64-NEXT: leal (%rdi,%rsi), %eax
|
|
; X64-NEXT: btl %ecx, %edx
|
|
; X64-NEXT: sbbl $0, %eax
|
|
; X64-NEXT: negl %eax
|
|
; X64-NEXT: retq
|
|
%shift = lshr i32 %z, %w
|
|
%mask = and i32 %shift, 1
|
|
%add = add i32 %y, %x
|
|
%sub = sub i32 %mask, %add
|
|
ret i32 %sub
|
|
}
|