summaryrefslogtreecommitdiff
path: root/contrib/llvm/lib/Analysis/CodeMetrics.cpp
diff options
context:
space:
mode:
authorDimitry Andric <dim@FreeBSD.org>2019-12-20 19:53:05 +0000
committerDimitry Andric <dim@FreeBSD.org>2019-12-20 19:53:05 +0000
commit0b57cec536236d46e3dba9bd041533462f33dbb7 (patch)
tree56229dbdbbf76d18580f72f789003db17246c8d9 /contrib/llvm/lib/Analysis/CodeMetrics.cpp
parent718ef55ec7785aae63f98f8ca05dc07ed399c16d (diff)
Notes
Diffstat (limited to 'contrib/llvm/lib/Analysis/CodeMetrics.cpp')
-rw-r--r--contrib/llvm/lib/Analysis/CodeMetrics.cpp195
1 files changed, 0 insertions, 195 deletions
diff --git a/contrib/llvm/lib/Analysis/CodeMetrics.cpp b/contrib/llvm/lib/Analysis/CodeMetrics.cpp
deleted file mode 100644
index 627d955c865f..000000000000
--- a/contrib/llvm/lib/Analysis/CodeMetrics.cpp
+++ /dev/null
@@ -1,195 +0,0 @@
-//===- CodeMetrics.cpp - Code cost measurements ---------------------------===//
-//
-// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
-// See https://llvm.org/LICENSE.txt for license information.
-// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
-//
-//===----------------------------------------------------------------------===//
-//
-// This file implements code cost measurement utilities.
-//
-//===----------------------------------------------------------------------===//
-
-#include "llvm/Analysis/CodeMetrics.h"
-#include "llvm/Analysis/AssumptionCache.h"
-#include "llvm/Analysis/LoopInfo.h"
-#include "llvm/Analysis/TargetTransformInfo.h"
-#include "llvm/Analysis/ValueTracking.h"
-#include "llvm/IR/DataLayout.h"
-#include "llvm/IR/Function.h"
-#include "llvm/Support/Debug.h"
-#include "llvm/Support/raw_ostream.h"
-
-#define DEBUG_TYPE "code-metrics"
-
-using namespace llvm;
-
-static void
-appendSpeculatableOperands(const Value *V,
- SmallPtrSetImpl<const Value *> &Visited,
- SmallVectorImpl<const Value *> &Worklist) {
- const User *U = dyn_cast<User>(V);
- if (!U)
- return;
-
- for (const Value *Operand : U->operands())
- if (Visited.insert(Operand).second)
- if (isSafeToSpeculativelyExecute(Operand))
- Worklist.push_back(Operand);
-}
-
-static void completeEphemeralValues(SmallPtrSetImpl<const Value *> &Visited,
- SmallVectorImpl<const Value *> &Worklist,
- SmallPtrSetImpl<const Value *> &EphValues) {
- // Note: We don't speculate PHIs here, so we'll miss instruction chains kept
- // alive only by ephemeral values.
-
- // Walk the worklist using an index but without caching the size so we can
- // append more entries as we process the worklist. This forms a queue without
- // quadratic behavior by just leaving processed nodes at the head of the
- // worklist forever.
- for (int i = 0; i < (int)Worklist.size(); ++i) {
- const Value *V = Worklist[i];
-
- assert(Visited.count(V) &&
- "Failed to add a worklist entry to our visited set!");
-
- // If all uses of this value are ephemeral, then so is this value.
- if (!all_of(V->users(), [&](const User *U) { return EphValues.count(U); }))
- continue;
-
- EphValues.insert(V);
- LLVM_DEBUG(dbgs() << "Ephemeral Value: " << *V << "\n");
-
- // Append any more operands to consider.
- appendSpeculatableOperands(V, Visited, Worklist);
- }
-}
-
-// Find all ephemeral values.
-void CodeMetrics::collectEphemeralValues(
- const Loop *L, AssumptionCache *AC,
- SmallPtrSetImpl<const Value *> &EphValues) {
- SmallPtrSet<const Value *, 32> Visited;
- SmallVector<const Value *, 16> Worklist;
-
- for (auto &AssumeVH : AC->assumptions()) {
- if (!AssumeVH)
- continue;
- Instruction *I = cast<Instruction>(AssumeVH);
-
- // Filter out call sites outside of the loop so we don't do a function's
- // worth of work for each of its loops (and, in the common case, ephemeral
- // values in the loop are likely due to @llvm.assume calls in the loop).
- if (!L->contains(I->getParent()))
- continue;
-
- if (EphValues.insert(I).second)
- appendSpeculatableOperands(I, Visited, Worklist);
- }
-
- completeEphemeralValues(Visited, Worklist, EphValues);
-}
-
-void CodeMetrics::collectEphemeralValues(
- const Function *F, AssumptionCache *AC,
- SmallPtrSetImpl<const Value *> &EphValues) {
- SmallPtrSet<const Value *, 32> Visited;
- SmallVector<const Value *, 16> Worklist;
-
- for (auto &AssumeVH : AC->assumptions()) {
- if (!AssumeVH)
- continue;
- Instruction *I = cast<Instruction>(AssumeVH);
- assert(I->getParent()->getParent() == F &&
- "Found assumption for the wrong function!");
-
- if (EphValues.insert(I).second)
- appendSpeculatableOperands(I, Visited, Worklist);
- }
-
- completeEphemeralValues(Visited, Worklist, EphValues);
-}
-
-/// Fill in the current structure with information gleaned from the specified
-/// block.
-void CodeMetrics::analyzeBasicBlock(const BasicBlock *BB,
- const TargetTransformInfo &TTI,
- const SmallPtrSetImpl<const Value*> &EphValues) {
- ++NumBlocks;
- unsigned NumInstsBeforeThisBB = NumInsts;
- for (const Instruction &I : *BB) {
- // Skip ephemeral values.
- if (EphValues.count(&I))
- continue;
-
- // Special handling for calls.
- if (const auto *Call = dyn_cast<CallBase>(&I)) {
- if (const Function *F = Call->getCalledFunction()) {
- // If a function is both internal and has a single use, then it is
- // extremely likely to get inlined in the future (it was probably
- // exposed by an interleaved devirtualization pass).
- if (!Call->isNoInline() && F->hasInternalLinkage() && F->hasOneUse())
- ++NumInlineCandidates;
-
- // If this call is to function itself, then the function is recursive.
- // Inlining it into other functions is a bad idea, because this is
- // basically just a form of loop peeling, and our metrics aren't useful
- // for that case.
- if (F == BB->getParent())
- isRecursive = true;
-
- if (TTI.isLoweredToCall(F))
- ++NumCalls;
- } else {
- // We don't want inline asm to count as a call - that would prevent loop
- // unrolling. The argument setup cost is still real, though.
- if (!Call->isInlineAsm())
- ++NumCalls;
- }
- }
-
- if (const AllocaInst *AI = dyn_cast<AllocaInst>(&I)) {
- if (!AI->isStaticAlloca())
- this->usesDynamicAlloca = true;
- }
-
- if (isa<ExtractElementInst>(I) || I.getType()->isVectorTy())
- ++NumVectorInsts;
-
- if (I.getType()->isTokenTy() && I.isUsedOutsideOfBlock(BB))
- notDuplicatable = true;
-
- if (const CallInst *CI = dyn_cast<CallInst>(&I)) {
- if (CI->cannotDuplicate())
- notDuplicatable = true;
- if (CI->isConvergent())
- convergent = true;
- }
-
- if (const InvokeInst *InvI = dyn_cast<InvokeInst>(&I))
- if (InvI->cannotDuplicate())
- notDuplicatable = true;
-
- NumInsts += TTI.getUserCost(&I);
- }
-
- if (isa<ReturnInst>(BB->getTerminator()))
- ++NumRets;
-
- // We never want to inline functions that contain an indirectbr. This is
- // incorrect because all the blockaddress's (in static global initializers
- // for example) would be referring to the original function, and this indirect
- // jump would jump from the inlined copy of the function into the original
- // function which is extremely undefined behavior.
- // FIXME: This logic isn't really right; we can safely inline functions
- // with indirectbr's as long as no other function or global references the
- // blockaddress of a block within the current function. And as a QOI issue,
- // if someone is using a blockaddress without an indirectbr, and that
- // reference somehow ends up in another function or global, we probably
- // don't want to inline this function.
- notDuplicatable |= isa<IndirectBrInst>(BB->getTerminator());
-
- // Remember NumInsts for this BB.
- NumBBInsts[BB] = NumInsts - NumInstsBeforeThisBB;
-}