Skip to content

Commit

Permalink
[AAPointerInfo] handle multiple offsets in PHI
Browse files Browse the repository at this point in the history
The arguments to a PHI may represent a recurrence by eventually using the output
of the PHI itself. This is now handled by checking for cycles in the control
flow. If a PHI is not in a recurrence, it is now able to report multiple offsets
instead of conservatively reporting unknown.

Reviewed By: jdoerfert

Differential Revision: https://reviews.llvm.org/D138991
  • Loading branch information
ssahasra committed Dec 15, 2022
1 parent b05b897 commit 88db516
Show file tree
Hide file tree
Showing 7 changed files with 314 additions and 37 deletions.
44 changes: 23 additions & 21 deletions llvm/include/llvm/Analysis/CycleAnalysis.h
Expand Up @@ -27,6 +27,27 @@ extern template class GenericCycle<SSAContext>;
using CycleInfo = GenericCycleInfo<SSAContext>;
using Cycle = CycleInfo::CycleT;

/// Legacy analysis pass which computes a \ref CycleInfo.
class CycleInfoWrapperPass : public FunctionPass {
Function *F = nullptr;
CycleInfo CI;

public:
static char ID;

CycleInfoWrapperPass();

CycleInfo &getResult() { return CI; }
const CycleInfo &getResult() const { return CI; }

bool runOnFunction(Function &F) override;
void getAnalysisUsage(AnalysisUsage &AU) const override;
void releaseMemory() override;
void print(raw_ostream &OS, const Module *M = nullptr) const override;

// TODO: verify analysis?
};

/// Analysis pass which computes a \ref CycleInfo.
class CycleAnalysis : public AnalysisInfoMixin<CycleAnalysis> {
friend AnalysisInfoMixin<CycleAnalysis>;
Expand All @@ -36,6 +57,8 @@ class CycleAnalysis : public AnalysisInfoMixin<CycleAnalysis> {
/// Provide the result typedef for this analysis pass.
using Result = CycleInfo;

using LegacyWrapper = CycleInfoWrapperPass;

/// Run the analysis pass over a function and produce a dominator tree.
CycleInfo run(Function &F, FunctionAnalysisManager &);

Expand All @@ -52,27 +75,6 @@ class CycleInfoPrinterPass : public PassInfoMixin<CycleInfoPrinterPass> {
PreservedAnalyses run(Function &F, FunctionAnalysisManager &AM);
};

/// Legacy analysis pass which computes a \ref CycleInfo.
class CycleInfoWrapperPass : public FunctionPass {
Function *F = nullptr;
CycleInfo CI;

public:
static char ID;

CycleInfoWrapperPass();

CycleInfo &getCycleInfo() { return CI; }
const CycleInfo &getCycleInfo() const { return CI; }

bool runOnFunction(Function &F) override;
void getAnalysisUsage(AnalysisUsage &AU) const override;
void releaseMemory() override;
void print(raw_ostream &OS, const Module *M = nullptr) const override;

// TODO: verify analysis?
};

} // end namespace llvm

#endif // LLVM_ANALYSIS_CYCLEANALYSIS_H
32 changes: 28 additions & 4 deletions llvm/include/llvm/Transforms/IPO/Attributor.h
Expand Up @@ -1090,20 +1090,44 @@ class SubsumingPositionIterator {
iterator end() { return IRPositions.end(); }
};

/// Wrapper for FunctoinAnalysisManager.
/// Wrapper for FunctionAnalysisManager.
struct AnalysisGetter {
// The client may be running the old pass manager, in which case, we need to
// map the requested Analysis to its equivalent wrapper in the old pass
// manager. The scheme implemented here does not require every Analysis to be
// updated. Only those new analyses that the client cares about in the old
// pass manager need to expose a LegacyWrapper type, and that wrapper should
// support a getResult() method that matches the new Analysis.
//
// We need SFINAE to check for the LegacyWrapper, but function templates don't
// allow partial specialization, which is needed in this case. So instead, we
// use a constexpr bool to perform the SFINAE, and then use this information
// inside the function template.
template <typename, typename = void> static constexpr bool HasLegacyWrapper{};
template <typename Analysis>
static constexpr bool HasLegacyWrapper<
Analysis, std::void_t<typename Analysis::LegacyWrapper>> = true;

template <typename Analysis>
typename Analysis::Result *getAnalysis(const Function &F) {
if (!FAM || !F.getParent())
return nullptr;
return &FAM->getResult<Analysis>(const_cast<Function &>(F));
if (FAM)
return &FAM->getResult<Analysis>(const_cast<Function &>(F));
if constexpr (HasLegacyWrapper<Analysis>)
if (LegacyPass)
return &LegacyPass
->getAnalysis<typename Analysis::LegacyWrapper>(
const_cast<Function &>(F))
.getResult();
return nullptr;
}

AnalysisGetter(FunctionAnalysisManager &FAM) : FAM(&FAM) {}
AnalysisGetter(Pass *P) : LegacyPass(P) {}
AnalysisGetter() = default;

private:
FunctionAnalysisManager *FAM = nullptr;
Pass *LegacyPass = nullptr;
};

/// Data structure to hold cached (LLVM-IR) information.
Expand Down
17 changes: 15 additions & 2 deletions llvm/lib/Target/AMDGPU/AMDGPUAttributor.cpp
Expand Up @@ -13,6 +13,7 @@
#include "AMDGPU.h"
#include "GCNSubtarget.h"
#include "Utils/AMDGPUBaseInfo.h"
#include "llvm/Analysis/CycleAnalysis.h"
#include "llvm/CodeGen/TargetPassConfig.h"
#include "llvm/IR/IntrinsicsAMDGPU.h"
#include "llvm/IR/IntrinsicsR600.h"
Expand All @@ -21,6 +22,10 @@

#define DEBUG_TYPE "amdgpu-attributor"

namespace llvm {
void initializeCycleInfoWrapperPassPass(PassRegistry &);
}

using namespace llvm;

#define AMDGPU_ATTRIBUTE(Name, Str) Name##_POS,
Expand Down Expand Up @@ -747,7 +752,7 @@ class AMDGPUAttributor : public ModulePass {

bool runOnModule(Module &M) override {
SetVector<Function *> Functions;
AnalysisGetter AG;
AnalysisGetter AG(this);
for (Function &F : M) {
if (!F.isIntrinsic())
Functions.insert(&F);
Expand Down Expand Up @@ -782,6 +787,10 @@ class AMDGPUAttributor : public ModulePass {
return Change == ChangeStatus::CHANGED;
}

void getAnalysisUsage(AnalysisUsage &AU) const override {
AU.addRequired<CycleInfoWrapperPass>();
}

StringRef getPassName() const override { return "AMDGPU Attributor"; }
TargetMachine *TM;
static char ID;
Expand All @@ -791,4 +800,8 @@ class AMDGPUAttributor : public ModulePass {
char AMDGPUAttributor::ID = 0;

Pass *llvm::createAMDGPUAttributorPass() { return new AMDGPUAttributor(); }
INITIALIZE_PASS(AMDGPUAttributor, DEBUG_TYPE, "AMDGPU Attributor", false, false)
INITIALIZE_PASS_BEGIN(AMDGPUAttributor, DEBUG_TYPE, "AMDGPU Attributor", false,
false)
INITIALIZE_PASS_DEPENDENCY(CycleInfoWrapperPass);
INITIALIZE_PASS_END(AMDGPUAttributor, DEBUG_TYPE, "AMDGPU Attributor", false,
false)
49 changes: 39 additions & 10 deletions llvm/lib/Transforms/IPO/AttributorAttributes.cpp
Expand Up @@ -28,6 +28,7 @@
#include "llvm/Analysis/AssumeBundleQueries.h"
#include "llvm/Analysis/AssumptionCache.h"
#include "llvm/Analysis/CaptureTracking.h"
#include "llvm/Analysis/CycleAnalysis.h"
#include "llvm/Analysis/InstructionSimplify.h"
#include "llvm/Analysis/LazyValueInfo.h"
#include "llvm/Analysis/MemoryBuiltins.h"
Expand Down Expand Up @@ -1442,10 +1443,13 @@ ChangeStatus AAPointerInfoFloating::updateImpl(Attributor &A) {
return true;
};

const auto *F = getAnchorScope();
const auto *CI =
F ? A.getInfoCache().getAnalysisResultForFunction<CycleAnalysis>(*F)
: nullptr;
const auto *TLI =
getAnchorScope()
? A.getInfoCache().getTargetLibraryInfoForFunction(*getAnchorScope())
: nullptr;
F ? A.getInfoCache().getTargetLibraryInfoForFunction(*F) : nullptr;

auto UsePred = [&](const Use &U, bool &Follow) -> bool {
Value *CurPtr = U.get();
User *Usr = U.getUser();
Expand Down Expand Up @@ -1517,31 +1521,56 @@ ChangeStatus AAPointerInfoFloating::updateImpl(Attributor &A) {
return true;
}

// Check if the PHI operand is not dependent on the PHI itself.
// Check if the PHI operand can be traced back to AssociatedValue.
APInt Offset(
DL.getIndexSizeInBits(CurPtr->getType()->getPointerAddressSpace()),
0);
Value *CurPtrBase = CurPtr->stripAndAccumulateConstantOffsets(
DL, Offset, /* AllowNonInbounds */ true);
auto It = OffsetInfoMap.find(CurPtrBase);
if (It != OffsetInfoMap.end()) {
if (It == OffsetInfoMap.end()) {
LLVM_DEBUG(dbgs() << "[AAPointerInfo] PHI operand is too complex "
<< *CurPtr << " in " << *Usr << "\n");
UsrOI.setUnknown();
Follow = true;
return true;
}

auto mayBeInCycleHeader = [](const CycleInfo *CI, const Instruction *I) {
if (!CI)
return true;
auto *BB = I->getParent();
auto *C = CI->getCycle(BB);
if (!C)
return false;
return BB == C->getHeader();
};

// Check if the PHI operand is not dependent on the PHI itself. Every
// recurrence is a cyclic net of PHIs in the data flow, and has an
// equivalent Cycle in the control flow. One of those PHIs must be in the
// header of that control flow Cycle. This is independent of the choice of
// Cycles reported by CycleInfo. It is sufficient to check the PHIs in
// every Cycle header; if such a node is marked unknown, this will
// eventually propagate through the whole net of PHIs in the recurrence.
if (mayBeInCycleHeader(CI, cast<Instruction>(Usr))) {
auto BaseOI = It->getSecond();
BaseOI.addToAll(Offset.getZExtValue());
if (IsFirstPHIUser || BaseOI == UsrOI) {
LLVM_DEBUG(dbgs() << "[AAPointerInfo] PHI is invariant " << *CurPtr
<< " in " << *Usr << "\n");
return HandlePassthroughUser(Usr, PtrOI, Follow);
}

LLVM_DEBUG(
dbgs() << "[AAPointerInfo] PHI operand pointer offset mismatch "
<< *CurPtr << " in " << *Usr << "\n");
} else {
LLVM_DEBUG(dbgs() << "[AAPointerInfo] PHI operand is too complex "
<< *CurPtr << " in " << *Usr << "\n");
UsrOI.setUnknown();
Follow = true;
return true;
}

// TODO: Approximate in case we know the direction of the recurrence.
UsrOI.setUnknown();
UsrOI.merge(PtrOI);
Follow = true;
return true;
}
Expand Down
30 changes: 30 additions & 0 deletions llvm/test/CodeGen/AMDGPU/implicitarg-attributes.ll
Expand Up @@ -63,6 +63,36 @@ entry:
ret void
}

; CHECK-NOT: hidden_hostcall_buffer
; CHECK-NOT: hidden_multigrid_sync_arg
; CHECK-LABEL: .name: kernel_3

define amdgpu_kernel void @kernel_3(i32 addrspace(1)* %a, i1 %cond) {
entry:
%tmp7 = tail call i8 addrspace(4)* @llvm.amdgcn.implicitarg.ptr()
br i1 %cond, label %old, label %new

old: ; preds = %entry
%tmp4 = getelementptr i8, i8 addrspace(4)* %tmp7, i64 12
br label %join

new: ; preds = %entry
%tmp12 = getelementptr inbounds i8, i8 addrspace(4)* %tmp7, i64 18
br label %join

join: ; preds = %new, %old
%.in.in.in = phi i8 addrspace(4)* [ %tmp12, %new ], [ %tmp4, %old ]
%.in.in = bitcast i8 addrspace(4)* %.in.in.in to i16 addrspace(4)*

;;; THIS USE of implicitarg_ptr should not produce hostcall metadata
%.in = load i16, i16 addrspace(4)* %.in.in, align 2

%idx.ext = sext i16 %.in to i64
%add.ptr3 = getelementptr inbounds i32, i32 addrspace(1)* %a, i64 %idx.ext
%tmp16 = atomicrmw add i32 addrspace(1)* %add.ptr3, i32 15 syncscope("agent-one-as") monotonic, align 4
ret void
}

declare i32 @llvm.amdgcn.workitem.id.x()

declare align 4 i8 addrspace(4)* @llvm.amdgcn.implicitarg.ptr()
Expand Down
10 changes: 10 additions & 0 deletions llvm/test/CodeGen/AMDGPU/llc-pipeline.ll
Expand Up @@ -51,6 +51,8 @@
; GCN-O0-NEXT: Scalarize Masked Memory Intrinsics
; GCN-O0-NEXT: Expand reduction intrinsics
; GCN-O0-NEXT: AMDGPU Attributor
; GCN-O0-NEXT: FunctionPass Manager
; GCN-O0-NEXT: Cycle Info Analysis
; GCN-O0-NEXT: CallGraph Construction
; GCN-O0-NEXT: Call Graph SCC Pass Manager
; GCN-O0-NEXT: AMDGPU Annotate Kernel Features
Expand Down Expand Up @@ -225,6 +227,8 @@
; GCN-O1-NEXT: Natural Loop Information
; GCN-O1-NEXT: TLS Variable Hoist
; GCN-O1-NEXT: AMDGPU Attributor
; GCN-O1-NEXT: FunctionPass Manager
; GCN-O1-NEXT: Cycle Info Analysis
; GCN-O1-NEXT: CallGraph Construction
; GCN-O1-NEXT: Call Graph SCC Pass Manager
; GCN-O1-NEXT: AMDGPU Annotate Kernel Features
Expand Down Expand Up @@ -509,6 +513,8 @@
; GCN-O1-OPTS-NEXT: TLS Variable Hoist
; GCN-O1-OPTS-NEXT: Early CSE
; GCN-O1-OPTS-NEXT: AMDGPU Attributor
; GCN-O1-OPTS-NEXT: FunctionPass Manager
; GCN-O1-OPTS-NEXT: Cycle Info Analysis
; GCN-O1-OPTS-NEXT: CallGraph Construction
; GCN-O1-OPTS-NEXT: Call Graph SCC Pass Manager
; GCN-O1-OPTS-NEXT: AMDGPU Annotate Kernel Features
Expand Down Expand Up @@ -807,6 +813,8 @@
; GCN-O2-NEXT: TLS Variable Hoist
; GCN-O2-NEXT: Early CSE
; GCN-O2-NEXT: AMDGPU Attributor
; GCN-O2-NEXT: FunctionPass Manager
; GCN-O2-NEXT: Cycle Info Analysis
; GCN-O2-NEXT: CallGraph Construction
; GCN-O2-NEXT: Call Graph SCC Pass Manager
; GCN-O2-NEXT: AMDGPU Annotate Kernel Features
Expand Down Expand Up @@ -1118,6 +1126,8 @@
; GCN-O3-NEXT: Optimization Remark Emitter
; GCN-O3-NEXT: Global Value Numbering
; GCN-O3-NEXT: AMDGPU Attributor
; GCN-O3-NEXT: FunctionPass Manager
; GCN-O3-NEXT: Cycle Info Analysis
; GCN-O3-NEXT: CallGraph Construction
; GCN-O3-NEXT: Call Graph SCC Pass Manager
; GCN-O3-NEXT: AMDGPU Annotate Kernel Features
Expand Down

0 comments on commit 88db516

Please sign in to comment.