diff --git a/llvm/lib/Target/AMDGPU/AMDGPUSubtarget.cpp b/llvm/lib/Target/AMDGPU/AMDGPUSubtarget.cpp index 519f05fd402f4b..8015eb5734e30a 100644 --- a/llvm/lib/Target/AMDGPU/AMDGPUSubtarget.cpp +++ b/llvm/lib/Target/AMDGPU/AMDGPUSubtarget.cpp @@ -823,7 +823,7 @@ struct FillMFMAShadowMutation : ScheduleDAGMutation { for (unsigned I = 0; I < Succs.size(); ++I) { for (const SDep &SI : Succs[I]->Succs) { const SUnit *SU = SI.getSUnit(); - if (SU != Succs[I] && llvm::find(Succs, SU) == Succs.end()) + if (SU != Succs[I] && !llvm::is_contained(Succs, SU)) Succs.push_back(SU); } } @@ -831,7 +831,7 @@ struct FillMFMAShadowMutation : ScheduleDAGMutation { SmallPtrSet Visited; while (!Preds.empty()) { const SUnit *SU = Preds.pop_back_val(); - if (llvm::find(Succs, SU) != Succs.end()) + if (llvm::is_contained(Succs, SU)) return false; Visited.insert(SU); for (const SDep &SI : SU->Preds) diff --git a/llvm/lib/Target/AMDGPU/GCNNSAReassign.cpp b/llvm/lib/Target/AMDGPU/GCNNSAReassign.cpp index 200b2d36848da8..89887b6665703e 100644 --- a/llvm/lib/Target/AMDGPU/GCNNSAReassign.cpp +++ b/llvm/lib/Target/AMDGPU/GCNNSAReassign.cpp @@ -279,7 +279,7 @@ bool GCNNSAReassign::runOnMachineFunction(MachineFunction &MF) { const MachineOperand &Op = MI->getOperand(VAddr0Idx + I); Register Reg = Op.getReg(); LiveInterval *LI = &LIS->getInterval(Reg); - if (llvm::find(Intervals, LI) != Intervals.end()) { + if (llvm::is_contained(Intervals, LI)) { // Same register used, unable to make sequential Intervals.clear(); break; diff --git a/llvm/lib/Target/AMDGPU/SIInsertSkips.cpp b/llvm/lib/Target/AMDGPU/SIInsertSkips.cpp index 07e88478722017..7601d8e48def4f 100644 --- a/llvm/lib/Target/AMDGPU/SIInsertSkips.cpp +++ b/llvm/lib/Target/AMDGPU/SIInsertSkips.cpp @@ -240,8 +240,8 @@ void SIInsertSkips::skipIfDead(MachineBasicBlock &MBB, // In this case, we write the "null_export; s_endpgm" skip code in the // already-existing basic block. auto NextBBI = std::next(MBB.getIterator()); - bool NoSuccessor = I == MBB.end() && - llvm::find(MBB.successors(), &*NextBBI) == MBB.succ_end(); + bool NoSuccessor = + I == MBB.end() && !llvm::is_contained(MBB.successors(), &*NextBBI); if (NoSuccessor) { generatePsEndPgm(MBB, I, DL, TII); diff --git a/llvm/lib/Target/AMDGPU/SILowerI1Copies.cpp b/llvm/lib/Target/AMDGPU/SILowerI1Copies.cpp index cb8eab674b4011..b8590f6128f6c7 100644 --- a/llvm/lib/Target/AMDGPU/SILowerI1Copies.cpp +++ b/llvm/lib/Target/AMDGPU/SILowerI1Copies.cpp @@ -213,7 +213,7 @@ class PhiIncomingAnalysis { ReachableMap[MBB] = true; if (HaveReachablePred) { for (MachineBasicBlock *UnreachablePred : Stack) { - if (llvm::find(Predecessors, UnreachablePred) == Predecessors.end()) + if (!llvm::is_contained(Predecessors, UnreachablePred)) Predecessors.push_back(UnreachablePred); } } @@ -347,7 +347,7 @@ class LoopFinder { if (DomIt != Visited.end() && DomIt->second <= LoopLevel) return true; - if (llvm::find(Blocks, &MBB) != Blocks.end()) + if (llvm::is_contained(Blocks, &MBB)) return true; return false;