From 1e362133d5a39cd0708461cc1c591ed095e8a1fe Mon Sep 17 00:00:00 2001 From: Chandler Carruth Date: Wed, 6 Sep 2017 06:28:08 +0000 Subject: [PATCH] [x86] Fix PR34377 by disabling cmov conversion when we relied on it performing a zext of a register. On the PR there is discussion of how to more effectively handle this, but this patch prevents us from miscompiling code. Differential Revision: https://reviews.llvm.org/D37504 llvm-svn: 312620 --- lib/Target/X86/X86CmovConversion.cpp | 10 ++++++++++ test/CodeGen/X86/cmov-into-branch.ll | 26 ++++++++++++++++++++++---- 2 files changed, 32 insertions(+), 4 deletions(-) diff --git a/lib/Target/X86/X86CmovConversion.cpp b/lib/Target/X86/X86CmovConversion.cpp index f7967004e20..e31a7949f0b 100644 --- a/lib/Target/X86/X86CmovConversion.cpp +++ b/lib/Target/X86/X86CmovConversion.cpp @@ -289,6 +289,16 @@ bool X86CmovConverterPass::collectCmovCandidates( // Can't handle mixed conditions with memory operands. SkipGroup = true; } + // Check if we were relying on zero-extending behavior of the CMOV. + if (!SkipGroup && + llvm::any_of( + MRI->use_nodbg_instructions(I.defs().begin()->getReg()), + [&](MachineInstr &UseI) { + return UseI.getOpcode() == X86::SUBREG_TO_REG; + })) + // FIXME: We should model the cost of using an explicit MOV to handle + // the zero-extension rather than just refusing to handle this. + SkipGroup = true; continue; } // If Group is empty, keep looking for first CMOV in the range. diff --git a/test/CodeGen/X86/cmov-into-branch.ll b/test/CodeGen/X86/cmov-into-branch.ll index e3803950164..0a2246700e6 100644 --- a/test/CodeGen/X86/cmov-into-branch.ll +++ b/test/CodeGen/X86/cmov-into-branch.ll @@ -61,6 +61,24 @@ define i32 @test5(i32 %a, i32* nocapture %b, i32 %x, i32 %y) { ret i32 %cond5 } +; Zero-extended select. +define void @test6(i32 %a, i32 %x, i32* %y.ptr, i64* %z.ptr) { +; CHECK-LABEL: test6: +; CHECK: # BB#0: # %entry +; CHECK-NEXT: # kill: %ESI %ESI %RSI +; CHECK-NEXT: testl %edi, %edi +; CHECK-NEXT: cmovnsl (%rdx), %esi +; CHECK-NEXT: movq %rsi, (%rcx) +; CHECK-NEXT: retq +entry: + %y = load i32, i32* %y.ptr + %cmp = icmp slt i32 %a, 0 + %z = select i1 %cmp, i32 %x, i32 %y + %z.ext = zext i32 %z to i64 + store i64 %z.ext, i64* %z.ptr + ret void +} + ; If a select is not obviously predictable, don't turn it into a branch. define i32 @weighted_select1(i32 %a, i32 %b) { ; CHECK-LABEL: weighted_select1: @@ -79,10 +97,10 @@ define i32 @weighted_select2(i32 %a, i32 %b) { ; CHECK-LABEL: weighted_select2: ; CHECK: # BB#0: ; CHECK-NEXT: testl %edi, %edi -; CHECK-NEXT: jne .LBB5_2 +; CHECK-NEXT: jne .LBB6_2 ; CHECK-NEXT: # BB#1: # %select.false ; CHECK-NEXT: movl %esi, %edi -; CHECK-NEXT: .LBB5_2: # %select.end +; CHECK-NEXT: .LBB6_2: # %select.end ; CHECK-NEXT: movl %edi, %eax ; CHECK-NEXT: retq %cmp = icmp ne i32 %a, 0 @@ -98,11 +116,11 @@ define i32 @weighted_select3(i32 %a, i32 %b) { ; CHECK-LABEL: weighted_select3: ; CHECK: # BB#0: ; CHECK-NEXT: testl %edi, %edi -; CHECK-NEXT: je .LBB6_1 +; CHECK-NEXT: je .LBB7_1 ; CHECK-NEXT: # BB#2: # %select.end ; CHECK-NEXT: movl %edi, %eax ; CHECK-NEXT: retq -; CHECK-NEXT: .LBB6_1: # %select.false +; CHECK-NEXT: .LBB7_1: # %select.false ; CHECK-NEXT: movl %esi, %edi ; CHECK-NEXT: movl %edi, %eax ; CHECK-NEXT: retq