From fc0b80d9746e5fd4b45057ab814c67371fb0f9ea Mon Sep 17 00:00:00 2001 From: Evan Cheng Date: Fri, 13 Mar 2009 22:59:14 +0000 Subject: [PATCH] Fix PR3784: If the source of a phi comes from a bb ended with an invoke, make sure the copy is inserted before the try range (unless it's used as an input to the invoke, then insert it after the last use), not at the end of the bb. Also re-apply r66140 which was disabled as a workaround. git-svn-id: https://llvm.org/svn/llvm-project/llvm/trunk@66976 91177308-0d34-0410-b5e6-96231b3b80d8 --- lib/CodeGen/PHIElimination.cpp | 40 +++++++++++++++++++++++ lib/Transforms/Scalar/CodeGenPrepare.cpp | 6 ++++ test/CodeGen/X86/2009-03-13-PHIElimBug.ll | 37 +++++++++++++++++++++ test/CodeGen/X86/split-eh-lpad-edges.ll | 34 +++++++++++++++++++ 4 files changed, 117 insertions(+) create mode 100644 test/CodeGen/X86/2009-03-13-PHIElimBug.ll diff --git a/lib/CodeGen/PHIElimination.cpp b/lib/CodeGen/PHIElimination.cpp index b32db2e062f..6f3c82c7081 100644 --- a/lib/CodeGen/PHIElimination.cpp +++ b/lib/CodeGen/PHIElimination.cpp @@ -31,6 +31,7 @@ using namespace llvm; STATISTIC(NumAtomic, "Number of atomic phis lowered"); +STATISTIC(NumEH, "Number of EH try blocks skipped"); namespace { class VISIBILITY_HIDDEN PNE : public MachineFunctionPass { @@ -65,6 +66,9 @@ namespace { /// void analyzePHINodes(const MachineFunction& Fn); + void WalkPassEHTryRange(MachineBasicBlock &MBB, + MachineBasicBlock::iterator &I, unsigned SrcReg); + typedef std::pair BBVRegPair; typedef std::map VRegPHIUse; @@ -140,6 +144,39 @@ static bool isSourceDefinedByImplicitDef(const MachineInstr *MPhi, return true; } +void PNE::WalkPassEHTryRange(MachineBasicBlock &MBB, + MachineBasicBlock::iterator &I, unsigned SrcReg) { + if (I == MBB.begin()) + return; + MachineBasicBlock::iterator PI = prior(I); + if (PI->getOpcode() != TargetInstrInfo::EH_LABEL) + return; + + // Trying to walk pass the EH try range. If we run into a use instruction, + // we want to insert the copy there. + SmallPtrSet UsesInMBB; + for (MachineRegisterInfo::use_iterator UI = MRI->use_begin(SrcReg), + UE = MRI->use_end(); UI != UE; ++UI) { + MachineInstr *UseMI = &*UI; + if (UseMI->getParent() == &MBB) + UsesInMBB.insert(UseMI); + } + + while (PI != MBB.begin()) { + --PI; + if (PI->getOpcode() == TargetInstrInfo::EH_LABEL) { + ++NumEH; + I = PI; + return; + } else if (UsesInMBB.count(&*PI)) { + ++NumEH; + I = next(PI); + return; + } + } + return; +} + /// LowerAtomicPHINode - Lower the PHI node at the top of the specified block, /// under the assuption that it needs to be lowered in a way that supports /// atomic execution of PHIs. This lowering method is always correct all of the @@ -238,6 +275,9 @@ void PNE::LowerAtomicPHINode(MachineBasicBlock &MBB, // in the block (or end()). MachineBasicBlock::iterator InsertPos = opBlock.getFirstTerminator(); + // Walk pass EH try range if needed. + WalkPassEHTryRange(opBlock, InsertPos, SrcReg); + // Insert the copy. TII->copyRegToReg(opBlock, InsertPos, IncomingReg, SrcReg, RC, RC); diff --git a/lib/Transforms/Scalar/CodeGenPrepare.cpp b/lib/Transforms/Scalar/CodeGenPrepare.cpp index 1f3a13cc9ce..85e57661c15 100644 --- a/lib/Transforms/Scalar/CodeGenPrepare.cpp +++ b/lib/Transforms/Scalar/CodeGenPrepare.cpp @@ -327,6 +327,12 @@ static void SplitEdgeNicely(TerminatorInst *TI, unsigned SuccNum, assert(isa(Dest->begin()) && "This should only be called if Dest has a PHI!"); + // Do not split edges to EH landing pads. + if (InvokeInst *Invoke = dyn_cast(TI)) { + if (Invoke->getSuccessor(1) == Dest) + return; + } + // As a hack, never split backedges of loops. Even though the copy for any // PHIs inserted on the backedge would be dead for exits from the loop, we // assume that the cost of *splitting* the backedge would be too high. diff --git a/test/CodeGen/X86/2009-03-13-PHIElimBug.ll b/test/CodeGen/X86/2009-03-13-PHIElimBug.ll new file mode 100644 index 00000000000..ea4150c22c2 --- /dev/null +++ b/test/CodeGen/X86/2009-03-13-PHIElimBug.ll @@ -0,0 +1,37 @@ +; RUN: llvm-as < %s | llc -mtriple=i386-pc-linux-gnu -stats |& grep phielim | grep {Number of EH try blocks skipped} | grep 4 +; PR3784 + + %struct.c38002a__arr___XUB = type { i32, i32 } + %struct.c38002a__arr_name = type { [0 x i32]*, %struct.c38002a__arr___XUB* } + %struct.c38002a__rec = type { i32, %struct.c38002a__arr_name } + +define void @_ada_c38002a() { +entry: + %0 = invoke i8* @__gnat_malloc(i32 12) + to label %invcont unwind label %lpad ; [#uses=0] + +invcont: ; preds = %entry + %1 = invoke i8* @__gnat_malloc(i32 20) + to label %invcont1 unwind label %lpad ; [#uses=0] + +invcont1: ; preds = %invcont + %2 = invoke i32 @report__ident_int(i32 2) + to label %.noexc unwind label %lpad ; [#uses=0] + +.noexc: ; preds = %invcont1 + %3 = invoke i32 @report__ident_int(i32 3) + to label %.noexc88 unwind label %lpad ; [#uses=0] + +.noexc88: ; preds = %.noexc + unreachable + +lpad: ; preds = %.noexc, %invcont1, %invcont, %entry + %r.0 = phi %struct.c38002a__rec* [ null, %entry ], [ null, %invcont ], [ null, %invcont1 ], [ null, %.noexc ] ; <%struct.c38002a__rec*> [#uses=1] + %4 = getelementptr %struct.c38002a__rec* %r.0, i32 0, i32 0 ; [#uses=1] + %5 = load i32* %4, align 4 ; [#uses=0] + ret void +} + +declare i32 @report__ident_int(i32) + +declare i8* @__gnat_malloc(i32) diff --git a/test/CodeGen/X86/split-eh-lpad-edges.ll b/test/CodeGen/X86/split-eh-lpad-edges.ll index 2cc8e54672b..7808ad63168 100644 --- a/test/CodeGen/X86/split-eh-lpad-edges.ll +++ b/test/CodeGen/X86/split-eh-lpad-edges.ll @@ -1,4 +1,38 @@ ; RUN: llvm-as < %s | llc -mtriple=x86_64-apple-darwin | not grep jmp +; rdar://6647639 + + %struct.FetchPlanHeader = type { i8*, i8*, i32, i8*, i8*, i8*, i8*, i8*, %struct.NSObject* (%struct.NSObject*, %struct.objc_selector*, ...)*, %struct.__attributeDescriptionFlags } + %struct.NSArray = type { %struct.NSObject } + %struct.NSAutoreleasePool = type { %struct.NSObject, i8*, i8*, i8*, i8* } + %struct.NSObject = type { %struct.NSObject* } + %struct.__attributeDescriptionFlags = type <{ i32 }> + %struct._message_ref_t = type { %struct.NSObject* (%struct.NSObject*, %struct._message_ref_t*, ...)*, %struct.objc_selector* } + %struct.objc_selector = type opaque +@"\01l_objc_msgSend_fixup_alloc" = external global %struct._message_ref_t, align 16 ; <%struct._message_ref_t*> [#uses=2] + +define %struct.NSArray* @newFetchedRowsForFetchPlan_MT(%struct.FetchPlanHeader* %fetchPlan, %struct.objc_selector* %selectionMethod, %struct.NSObject* %selectionParameter) ssp { +entry: + %0 = invoke %struct.NSObject* null(%struct.NSObject* null, %struct._message_ref_t* @"\01l_objc_msgSend_fixup_alloc") + to label %invcont unwind label %lpad ; <%struct.NSObject*> [#uses=1] + +invcont: ; preds = %entry + %1 = invoke %struct.NSObject* (%struct.NSObject*, %struct.objc_selector*, ...)* @objc_msgSend(%struct.NSObject* %0, %struct.objc_selector* null) + to label %invcont26 unwind label %lpad ; <%struct.NSObject*> [#uses=0] + +invcont26: ; preds = %invcont + %2 = invoke %struct.NSObject* null(%struct.NSObject* null, %struct._message_ref_t* @"\01l_objc_msgSend_fixup_alloc") + to label %invcont27 unwind label %lpad ; <%struct.NSObject*> [#uses=0] + +invcont27: ; preds = %invcont26 + unreachable + +lpad: ; preds = %invcont26, %invcont, %entry + %pool.1 = phi %struct.NSAutoreleasePool* [ null, %entry ], [ null, %invcont ], [ null, %invcont26 ] ; <%struct.NSAutoreleasePool*> [#uses=0] + unreachable +} + +declare %struct.NSObject* @objc_msgSend(%struct.NSObject*, %struct.objc_selector*, ...) +; RUN: llvm-as < %s | llc -mtriple=x86_64-apple-darwin | not grep jmp ; rdar://6647639 ; XFAIL: * -- 2.34.1