1//===- Attributor.h --- Module-wide attribute deduction ---------*- C++ -*-===//
2//
3// Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions.
4// See https://llvm.org/LICENSE.txt for license information.
5// SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception
6//
7//===----------------------------------------------------------------------===//
8//
9// Attributor: An inter procedural (abstract) "attribute" deduction framework.
10//
11// The Attributor framework is an inter procedural abstract analysis (fixpoint
12// iteration analysis). The goal is to allow easy deduction of new attributes as
13// well as information exchange between abstract attributes in-flight.
14//
15// The Attributor class is the driver and the link between the various abstract
16// attributes. The Attributor will iterate until a fixpoint state is reached by
17// all abstract attributes in-flight, or until it will enforce a pessimistic fix
18// point because an iteration limit is reached.
19//
20// Abstract attributes, derived from the AbstractAttribute class, actually
21// describe properties of the code. They can correspond to actual LLVM-IR
22// attributes, or they can be more general, ultimately unrelated to LLVM-IR
23// attributes. The latter is useful when an abstract attributes provides
24// information to other abstract attributes in-flight but we might not want to
25// manifest the information. The Attributor allows to query in-flight abstract
26// attributes through the `Attributor::getAAFor` method (see the method
27// description for an example). If the method is used by an abstract attribute
28// P, and it results in an abstract attribute Q, the Attributor will
29// automatically capture a potential dependence from Q to P. This dependence
30// will cause P to be reevaluated whenever Q changes in the future.
31//
32// The Attributor will only reevaluate abstract attributes that might have
33// changed since the last iteration. That means that the Attribute will not
34// revisit all instructions/blocks/functions in the module but only query
35// an update from a subset of the abstract attributes.
36//
37// The update method `AbstractAttribute::updateImpl` is implemented by the
38// specific "abstract attribute" subclasses. The method is invoked whenever the
39// currently assumed state (see the AbstractState class) might not be valid
40// anymore. This can, for example, happen if the state was dependent on another
41// abstract attribute that changed. In every invocation, the update method has
42// to adjust the internal state of an abstract attribute to a point that is
43// justifiable by the underlying IR and the current state of abstract attributes
44// in-flight. Since the IR is given and assumed to be valid, the information
45// derived from it can be assumed to hold. However, information derived from
46// other abstract attributes is conditional on various things. If the justifying
47// state changed, the `updateImpl` has to revisit the situation and potentially
48// find another justification or limit the optimistic assumes made.
49//
50// Change is the key in this framework. Until a state of no-change, thus a
51// fixpoint, is reached, the Attributor will query the abstract attributes
52// in-flight to re-evaluate their state. If the (current) state is too
53// optimistic, hence it cannot be justified anymore through other abstract
54// attributes or the state of the IR, the state of the abstract attribute will
55// have to change. Generally, we assume abstract attribute state to be a finite
56// height lattice and the update function to be monotone. However, these
57// conditions are not enforced because the iteration limit will guarantee
58// termination. If an optimistic fixpoint is reached, or a pessimistic fix
59// point is enforced after a timeout, the abstract attributes are tasked to
60// manifest their result in the IR for passes to come.
61//
62// Attribute manifestation is not mandatory. If desired, there is support to
63// generate a single or multiple LLVM-IR attributes already in the helper struct
64// IRAttribute. In the simplest case, a subclass inherits from IRAttribute with
65// a proper Attribute::AttrKind as template parameter. The Attributor
66// manifestation framework will then create and place a new attribute if it is
67// allowed to do so (based on the abstract state). Other use cases can be
68// achieved by overloading AbstractAttribute or IRAttribute methods.
69//
70//
71// The "mechanics" of adding a new "abstract attribute":
72// - Define a class (transitively) inheriting from AbstractAttribute and one
73//   (which could be the same) that (transitively) inherits from AbstractState.
74//   For the latter, consider the already available BooleanState and
75//   {Inc,Dec,Bit}IntegerState if they fit your needs, e.g., you require only a
76//   number tracking or bit-encoding.
77// - Implement all pure methods. Also use overloading if the attribute is not
78//   conforming with the "default" behavior: A (set of) LLVM-IR attribute(s) for
79//   an argument, call site argument, function return value, or function. See
80//   the class and method descriptions for more information on the two
81//   "Abstract" classes and their respective methods.
82// - Register opportunities for the new abstract attribute in the
83//   `Attributor::identifyDefaultAbstractAttributes` method if it should be
84//   counted as a 'default' attribute.
85// - Add sufficient tests.
86// - Add a Statistics object for bookkeeping. If it is a simple (set of)
87//   attribute(s) manifested through the Attributor manifestation framework, see
88//   the bookkeeping function in Attributor.cpp.
89// - If instructions with a certain opcode are interesting to the attribute, add
90//   that opcode to the switch in `Attributor::identifyAbstractAttributes`. This
91//   will make it possible to query all those instructions through the
92//   `InformationCache::getOpcodeInstMapForFunction` interface and eliminate the
93//   need to traverse the IR repeatedly.
94//
95//===----------------------------------------------------------------------===//
96
97#ifndef LLVM_TRANSFORMS_IPO_ATTRIBUTOR_H
98#define LLVM_TRANSFORMS_IPO_ATTRIBUTOR_H
99
100#include "llvm/ADT/DenseSet.h"
101#include "llvm/ADT/GraphTraits.h"
102#include "llvm/ADT/MapVector.h"
103#include "llvm/ADT/STLExtras.h"
104#include "llvm/ADT/SetOperations.h"
105#include "llvm/ADT/SetVector.h"
106#include "llvm/ADT/SmallSet.h"
107#include "llvm/ADT/iterator.h"
108#include "llvm/Analysis/AssumeBundleQueries.h"
109#include "llvm/Analysis/CFG.h"
110#include "llvm/Analysis/CGSCCPassManager.h"
111#include "llvm/Analysis/LazyCallGraph.h"
112#include "llvm/Analysis/LoopInfo.h"
113#include "llvm/Analysis/MemoryLocation.h"
114#include "llvm/Analysis/MustExecute.h"
115#include "llvm/Analysis/OptimizationRemarkEmitter.h"
116#include "llvm/Analysis/PostDominators.h"
117#include "llvm/Analysis/TargetLibraryInfo.h"
118#include "llvm/IR/AbstractCallSite.h"
119#include "llvm/IR/Attributes.h"
120#include "llvm/IR/ConstantRange.h"
121#include "llvm/IR/Constants.h"
122#include "llvm/IR/GlobalValue.h"
123#include "llvm/IR/InstIterator.h"
124#include "llvm/IR/Instruction.h"
125#include "llvm/IR/Instructions.h"
126#include "llvm/IR/PassManager.h"
127#include "llvm/IR/Value.h"
128#include "llvm/Support/Alignment.h"
129#include "llvm/Support/Allocator.h"
130#include "llvm/Support/Casting.h"
131#include "llvm/Support/DOTGraphTraits.h"
132#include "llvm/Support/DebugCounter.h"
133#include "llvm/Support/ErrorHandling.h"
134#include "llvm/Support/ModRef.h"
135#include "llvm/Support/TimeProfiler.h"
136#include "llvm/Support/TypeSize.h"
137#include "llvm/TargetParser/Triple.h"
138#include "llvm/Transforms/Utils/CallGraphUpdater.h"
139
140#include <limits>
141#include <map>
142#include <optional>
143
144namespace llvm {
145
146class DataLayout;
147class LLVMContext;
148class Pass;
149template <typename Fn> class function_ref;
150struct AADepGraphNode;
151struct AADepGraph;
152struct Attributor;
153struct AbstractAttribute;
154struct InformationCache;
155struct AAIsDead;
156struct AttributorCallGraph;
157struct IRPosition;
158
159class Function;
160
161/// Abstract Attribute helper functions.
162namespace AA {
163using InstExclusionSetTy = SmallPtrSet<Instruction *, 4>;
164
165enum class GPUAddressSpace : unsigned {
166  Generic = 0,
167  Global = 1,
168  Shared = 3,
169  Constant = 4,
170  Local = 5,
171};
172
173/// Return true iff \p M target a GPU (and we can use GPU AS reasoning).
174bool isGPU(const Module &M);
175
176/// Flags to distinguish intra-procedural queries from *potentially*
177/// inter-procedural queries. Not that information can be valid for both and
178/// therefore both bits might be set.
179enum ValueScope : uint8_t {
180  Intraprocedural = 1,
181  Interprocedural = 2,
182  AnyScope = Intraprocedural | Interprocedural,
183};
184
185struct ValueAndContext : public std::pair<Value *, const Instruction *> {
186  using Base = std::pair<Value *, const Instruction *>;
187  ValueAndContext(const Base &B) : Base(B) {}
188  ValueAndContext(Value &V, const Instruction *CtxI) : Base(&V, CtxI) {}
189  ValueAndContext(Value &V, const Instruction &CtxI) : Base(&V, &CtxI) {}
190
191  Value *getValue() const { return this->first; }
192  const Instruction *getCtxI() const { return this->second; }
193};
194
195/// Return true if \p I is a `nosync` instruction. Use generic reasoning and
196/// potentially the corresponding AANoSync.
197bool isNoSyncInst(Attributor &A, const Instruction &I,
198                  const AbstractAttribute &QueryingAA);
199
200/// Return true if \p V is dynamically unique, that is, there are no two
201/// "instances" of \p V at runtime with different values.
202/// Note: If \p ForAnalysisOnly is set we only check that the Attributor will
203/// never use \p V to represent two "instances" not that \p V could not
204/// technically represent them.
205bool isDynamicallyUnique(Attributor &A, const AbstractAttribute &QueryingAA,
206                         const Value &V, bool ForAnalysisOnly = true);
207
208/// Return true if \p V is a valid value in \p Scope, that is a constant or an
209/// instruction/argument of \p Scope.
210bool isValidInScope(const Value &V, const Function *Scope);
211
212/// Return true if the value of \p VAC is a valid at the position of \p VAC,
213/// that is a constant, an argument of the same function, or an instruction in
214/// that function that dominates the position.
215bool isValidAtPosition(const ValueAndContext &VAC, InformationCache &InfoCache);
216
217/// Try to convert \p V to type \p Ty without introducing new instructions. If
218/// this is not possible return `nullptr`. Note: this function basically knows
219/// how to cast various constants.
220Value *getWithType(Value &V, Type &Ty);
221
222/// Return the combination of \p A and \p B such that the result is a possible
223/// value of both. \p B is potentially casted to match the type \p Ty or the
224/// type of \p A if \p Ty is null.
225///
226/// Examples:
227///        X + none  => X
228/// not_none + undef => not_none
229///          V1 + V2 => nullptr
230std::optional<Value *>
231combineOptionalValuesInAAValueLatice(const std::optional<Value *> &A,
232                                     const std::optional<Value *> &B, Type *Ty);
233
234/// Helper to represent an access offset and size, with logic to deal with
235/// uncertainty and check for overlapping accesses.
236struct RangeTy {
237  int64_t Offset = Unassigned;
238  int64_t Size = Unassigned;
239
240  RangeTy(int64_t Offset, int64_t Size) : Offset(Offset), Size(Size) {}
241  RangeTy() = default;
242  static RangeTy getUnknown() { return RangeTy{Unknown, Unknown}; }
243
244  /// Return true if offset or size are unknown.
245  bool offsetOrSizeAreUnknown() const {
246    return Offset == RangeTy::Unknown || Size == RangeTy::Unknown;
247  }
248
249  /// Return true if offset and size are unknown, thus this is the default
250  /// unknown object.
251  bool offsetAndSizeAreUnknown() const {
252    return Offset == RangeTy::Unknown && Size == RangeTy::Unknown;
253  }
254
255  /// Return true if the offset and size are unassigned.
256  bool isUnassigned() const {
257    assert((Offset == RangeTy::Unassigned) == (Size == RangeTy::Unassigned) &&
258           "Inconsistent state!");
259    return Offset == RangeTy::Unassigned;
260  }
261
262  /// Return true if this offset and size pair might describe an address that
263  /// overlaps with \p Range.
264  bool mayOverlap(const RangeTy &Range) const {
265    // Any unknown value and we are giving up -> overlap.
266    if (offsetOrSizeAreUnknown() || Range.offsetOrSizeAreUnknown())
267      return true;
268
269    // Check if one offset point is in the other interval [offset,
270    // offset+size].
271    return Range.Offset + Range.Size > Offset && Range.Offset < Offset + Size;
272  }
273
274  RangeTy &operator&=(const RangeTy &R) {
275    if (R.isUnassigned())
276      return *this;
277    if (isUnassigned())
278      return *this = R;
279    if (Offset == Unknown || R.Offset == Unknown)
280      Offset = Unknown;
281    if (Size == Unknown || R.Size == Unknown)
282      Size = Unknown;
283    if (offsetAndSizeAreUnknown())
284      return *this;
285    if (Offset == Unknown) {
286      Size = std::max(Size, R.Size);
287    } else if (Size == Unknown) {
288      Offset = std::min(Offset, R.Offset);
289    } else {
290      Offset = std::min(Offset, R.Offset);
291      Size = std::max(Offset + Size, R.Offset + R.Size) - Offset;
292    }
293    return *this;
294  }
295
296  /// Comparison for sorting ranges by offset.
297  ///
298  /// Returns true if the offset \p L is less than that of \p R.
299  inline static bool OffsetLessThan(const RangeTy &L, const RangeTy &R) {
300    return L.Offset < R.Offset;
301  }
302
303  /// Constants used to represent special offsets or sizes.
304  /// - We cannot assume that Offsets and Size are non-negative.
305  /// - The constants should not clash with DenseMapInfo, such as EmptyKey
306  ///   (INT64_MAX) and TombstoneKey (INT64_MIN).
307  /// We use values "in the middle" of the 64 bit range to represent these
308  /// special cases.
309  static constexpr int64_t Unassigned = std::numeric_limits<int32_t>::min();
310  static constexpr int64_t Unknown = std::numeric_limits<int32_t>::max();
311};
312
313inline raw_ostream &operator<<(raw_ostream &OS, const RangeTy &R) {
314  OS << "[" << R.Offset << ", " << R.Size << "]";
315  return OS;
316}
317
318inline bool operator==(const RangeTy &A, const RangeTy &B) {
319  return A.Offset == B.Offset && A.Size == B.Size;
320}
321
322inline bool operator!=(const RangeTy &A, const RangeTy &B) { return !(A == B); }
323
324/// Return the initial value of \p Obj with type \p Ty if that is a constant.
325Constant *getInitialValueForObj(Attributor &A,
326                                const AbstractAttribute &QueryingAA, Value &Obj,
327                                Type &Ty, const TargetLibraryInfo *TLI,
328                                const DataLayout &DL,
329                                RangeTy *RangePtr = nullptr);
330
331/// Collect all potential values \p LI could read into \p PotentialValues. That
332/// is, the only values read by \p LI are assumed to be known and all are in
333/// \p PotentialValues. \p PotentialValueOrigins will contain all the
334/// instructions that might have put a potential value into \p PotentialValues.
335/// Dependences onto \p QueryingAA are properly tracked, \p
336/// UsedAssumedInformation will inform the caller if assumed information was
337/// used.
338///
339/// \returns True if the assumed potential copies are all in \p PotentialValues,
340///          false if something went wrong and the copies could not be
341///          determined.
342bool getPotentiallyLoadedValues(
343    Attributor &A, LoadInst &LI, SmallSetVector<Value *, 4> &PotentialValues,
344    SmallSetVector<Instruction *, 4> &PotentialValueOrigins,
345    const AbstractAttribute &QueryingAA, bool &UsedAssumedInformation,
346    bool OnlyExact = false);
347
348/// Collect all potential values of the one stored by \p SI into
349/// \p PotentialCopies. That is, the only copies that were made via the
350/// store are assumed to be known and all are in \p PotentialCopies. Dependences
351/// onto \p QueryingAA are properly tracked, \p UsedAssumedInformation will
352/// inform the caller if assumed information was used.
353///
354/// \returns True if the assumed potential copies are all in \p PotentialCopies,
355///          false if something went wrong and the copies could not be
356///          determined.
357bool getPotentialCopiesOfStoredValue(
358    Attributor &A, StoreInst &SI, SmallSetVector<Value *, 4> &PotentialCopies,
359    const AbstractAttribute &QueryingAA, bool &UsedAssumedInformation,
360    bool OnlyExact = false);
361
362/// Return true if \p IRP is readonly. This will query respective AAs that
363/// deduce the information and introduce dependences for \p QueryingAA.
364bool isAssumedReadOnly(Attributor &A, const IRPosition &IRP,
365                       const AbstractAttribute &QueryingAA, bool &IsKnown);
366
367/// Return true if \p IRP is readnone. This will query respective AAs that
368/// deduce the information and introduce dependences for \p QueryingAA.
369bool isAssumedReadNone(Attributor &A, const IRPosition &IRP,
370                       const AbstractAttribute &QueryingAA, bool &IsKnown);
371
372/// Return true if \p ToI is potentially reachable from \p FromI without running
373/// into any instruction in \p ExclusionSet The two instructions do not need to
374/// be in the same function. \p GoBackwardsCB can be provided to convey domain
375/// knowledge about the "lifespan" the user is interested in. By default, the
376/// callers of \p FromI are checked as well to determine if \p ToI can be
377/// reached. If the query is not interested in callers beyond a certain point,
378/// e.g., a GPU kernel entry or the function containing an alloca, the
379/// \p GoBackwardsCB should return false.
380bool isPotentiallyReachable(
381    Attributor &A, const Instruction &FromI, const Instruction &ToI,
382    const AbstractAttribute &QueryingAA,
383    const AA::InstExclusionSetTy *ExclusionSet = nullptr,
384    std::function<bool(const Function &F)> GoBackwardsCB = nullptr);
385
386/// Same as above but it is sufficient to reach any instruction in \p ToFn.
387bool isPotentiallyReachable(
388    Attributor &A, const Instruction &FromI, const Function &ToFn,
389    const AbstractAttribute &QueryingAA,
390    const AA::InstExclusionSetTy *ExclusionSet = nullptr,
391    std::function<bool(const Function &F)> GoBackwardsCB = nullptr);
392
393/// Return true if \p Obj is assumed to be a thread local object.
394bool isAssumedThreadLocalObject(Attributor &A, Value &Obj,
395                                const AbstractAttribute &QueryingAA);
396
397/// Return true if \p I is potentially affected by a barrier.
398bool isPotentiallyAffectedByBarrier(Attributor &A, const Instruction &I,
399                                    const AbstractAttribute &QueryingAA);
400bool isPotentiallyAffectedByBarrier(Attributor &A, ArrayRef<const Value *> Ptrs,
401                                    const AbstractAttribute &QueryingAA,
402                                    const Instruction *CtxI);
403} // namespace AA
404
405template <>
406struct DenseMapInfo<AA::ValueAndContext>
407    : public DenseMapInfo<AA::ValueAndContext::Base> {
408  using Base = DenseMapInfo<AA::ValueAndContext::Base>;
409  static inline AA::ValueAndContext getEmptyKey() {
410    return Base::getEmptyKey();
411  }
412  static inline AA::ValueAndContext getTombstoneKey() {
413    return Base::getTombstoneKey();
414  }
415  static unsigned getHashValue(const AA::ValueAndContext &VAC) {
416    return Base::getHashValue(VAC);
417  }
418
419  static bool isEqual(const AA::ValueAndContext &LHS,
420                      const AA::ValueAndContext &RHS) {
421    return Base::isEqual(LHS, RHS);
422  }
423};
424
425template <>
426struct DenseMapInfo<AA::ValueScope> : public DenseMapInfo<unsigned char> {
427  using Base = DenseMapInfo<unsigned char>;
428  static inline AA::ValueScope getEmptyKey() {
429    return AA::ValueScope(Base::getEmptyKey());
430  }
431  static inline AA::ValueScope getTombstoneKey() {
432    return AA::ValueScope(Base::getTombstoneKey());
433  }
434  static unsigned getHashValue(const AA::ValueScope &S) {
435    return Base::getHashValue(S);
436  }
437
438  static bool isEqual(const AA::ValueScope &LHS, const AA::ValueScope &RHS) {
439    return Base::isEqual(LHS, RHS);
440  }
441};
442
443template <>
444struct DenseMapInfo<const AA::InstExclusionSetTy *>
445    : public DenseMapInfo<void *> {
446  using super = DenseMapInfo<void *>;
447  static inline const AA::InstExclusionSetTy *getEmptyKey() {
448    return static_cast<const AA::InstExclusionSetTy *>(super::getEmptyKey());
449  }
450  static inline const AA::InstExclusionSetTy *getTombstoneKey() {
451    return static_cast<const AA::InstExclusionSetTy *>(
452        super::getTombstoneKey());
453  }
454  static unsigned getHashValue(const AA::InstExclusionSetTy *BES) {
455    unsigned H = 0;
456    if (BES)
457      for (const auto *II : *BES)
458        H += DenseMapInfo<const Instruction *>::getHashValue(II);
459    return H;
460  }
461  static bool isEqual(const AA::InstExclusionSetTy *LHS,
462                      const AA::InstExclusionSetTy *RHS) {
463    if (LHS == RHS)
464      return true;
465    if (LHS == getEmptyKey() || RHS == getEmptyKey() ||
466        LHS == getTombstoneKey() || RHS == getTombstoneKey())
467      return false;
468    auto SizeLHS = LHS ? LHS->size() : 0;
469    auto SizeRHS = RHS ? RHS->size() : 0;
470    if (SizeLHS != SizeRHS)
471      return false;
472    if (SizeRHS == 0)
473      return true;
474    return llvm::set_is_subset(*LHS, *RHS);
475  }
476};
477
478/// The value passed to the line option that defines the maximal initialization
479/// chain length.
480extern unsigned MaxInitializationChainLength;
481
482///{
483enum class ChangeStatus {
484  CHANGED,
485  UNCHANGED,
486};
487
488ChangeStatus operator|(ChangeStatus l, ChangeStatus r);
489ChangeStatus &operator|=(ChangeStatus &l, ChangeStatus r);
490ChangeStatus operator&(ChangeStatus l, ChangeStatus r);
491ChangeStatus &operator&=(ChangeStatus &l, ChangeStatus r);
492
493enum class DepClassTy {
494  REQUIRED, ///< The target cannot be valid if the source is not.
495  OPTIONAL, ///< The target may be valid if the source is not.
496  NONE,     ///< Do not track a dependence between source and target.
497};
498///}
499
500/// The data structure for the nodes of a dependency graph
501struct AADepGraphNode {
502public:
503  virtual ~AADepGraphNode() = default;
504  using DepTy = PointerIntPair<AADepGraphNode *, 1>;
505  using DepSetTy = SmallSetVector<DepTy, 2>;
506
507protected:
508  /// Set of dependency graph nodes which should be updated if this one
509  /// is updated. The bit encodes if it is optional.
510  DepSetTy Deps;
511
512  static AADepGraphNode *DepGetVal(const DepTy &DT) { return DT.getPointer(); }
513  static AbstractAttribute *DepGetValAA(const DepTy &DT) {
514    return cast<AbstractAttribute>(DT.getPointer());
515  }
516
517  operator AbstractAttribute *() { return cast<AbstractAttribute>(this); }
518
519public:
520  using iterator = mapped_iterator<DepSetTy::iterator, decltype(&DepGetVal)>;
521  using aaiterator =
522      mapped_iterator<DepSetTy::iterator, decltype(&DepGetValAA)>;
523
524  aaiterator begin() { return aaiterator(Deps.begin(), &DepGetValAA); }
525  aaiterator end() { return aaiterator(Deps.end(), &DepGetValAA); }
526  iterator child_begin() { return iterator(Deps.begin(), &DepGetVal); }
527  iterator child_end() { return iterator(Deps.end(), &DepGetVal); }
528
529  void print(raw_ostream &OS) const { print(nullptr, OS); }
530  virtual void print(Attributor *, raw_ostream &OS) const {
531    OS << "AADepNode Impl\n";
532  }
533  DepSetTy &getDeps() { return Deps; }
534
535  friend struct Attributor;
536  friend struct AADepGraph;
537};
538
539/// The data structure for the dependency graph
540///
541/// Note that in this graph if there is an edge from A to B (A -> B),
542/// then it means that B depends on A, and when the state of A is
543/// updated, node B should also be updated
544struct AADepGraph {
545  AADepGraph() = default;
546  ~AADepGraph() = default;
547
548  using DepTy = AADepGraphNode::DepTy;
549  static AADepGraphNode *DepGetVal(const DepTy &DT) { return DT.getPointer(); }
550  using iterator =
551      mapped_iterator<AADepGraphNode::DepSetTy::iterator, decltype(&DepGetVal)>;
552
553  /// There is no root node for the dependency graph. But the SCCIterator
554  /// requires a single entry point, so we maintain a fake("synthetic") root
555  /// node that depends on every node.
556  AADepGraphNode SyntheticRoot;
557  AADepGraphNode *GetEntryNode() { return &SyntheticRoot; }
558
559  iterator begin() { return SyntheticRoot.child_begin(); }
560  iterator end() { return SyntheticRoot.child_end(); }
561
562  void viewGraph();
563
564  /// Dump graph to file
565  void dumpGraph();
566
567  /// Print dependency graph
568  void print();
569};
570
571/// Helper to describe and deal with positions in the LLVM-IR.
572///
573/// A position in the IR is described by an anchor value and an "offset" that
574/// could be the argument number, for call sites and arguments, or an indicator
575/// of the "position kind". The kinds, specified in the Kind enum below, include
576/// the locations in the attribute list, i.a., function scope and return value,
577/// as well as a distinction between call sites and functions. Finally, there
578/// are floating values that do not have a corresponding attribute list
579/// position.
580struct IRPosition {
581  // NOTE: In the future this definition can be changed to support recursive
582  // functions.
583  using CallBaseContext = CallBase;
584
585  /// The positions we distinguish in the IR.
586  enum Kind : char {
587    IRP_INVALID,  ///< An invalid position.
588    IRP_FLOAT,    ///< A position that is not associated with a spot suitable
589                  ///< for attributes. This could be any value or instruction.
590    IRP_RETURNED, ///< An attribute for the function return value.
591    IRP_CALL_SITE_RETURNED, ///< An attribute for a call site return value.
592    IRP_FUNCTION,           ///< An attribute for a function (scope).
593    IRP_CALL_SITE,          ///< An attribute for a call site (function scope).
594    IRP_ARGUMENT,           ///< An attribute for a function argument.
595    IRP_CALL_SITE_ARGUMENT, ///< An attribute for a call site argument.
596  };
597
598  /// Default constructor available to create invalid positions implicitly. All
599  /// other positions need to be created explicitly through the appropriate
600  /// static member function.
601  IRPosition() : Enc(nullptr, ENC_VALUE) { verify(); }
602
603  /// Create a position describing the value of \p V.
604  static const IRPosition value(const Value &V,
605                                const CallBaseContext *CBContext = nullptr) {
606    if (auto *Arg = dyn_cast<Argument>(&V))
607      return IRPosition::argument(*Arg, CBContext);
608    if (auto *CB = dyn_cast<CallBase>(&V))
609      return IRPosition::callsite_returned(*CB);
610    return IRPosition(const_cast<Value &>(V), IRP_FLOAT, CBContext);
611  }
612
613  /// Create a position describing the instruction \p I. This is different from
614  /// the value version because call sites are treated as intrusctions rather
615  /// than their return value in this function.
616  static const IRPosition inst(const Instruction &I,
617                               const CallBaseContext *CBContext = nullptr) {
618    return IRPosition(const_cast<Instruction &>(I), IRP_FLOAT, CBContext);
619  }
620
621  /// Create a position describing the function scope of \p F.
622  /// \p CBContext is used for call base specific analysis.
623  static const IRPosition function(const Function &F,
624                                   const CallBaseContext *CBContext = nullptr) {
625    return IRPosition(const_cast<Function &>(F), IRP_FUNCTION, CBContext);
626  }
627
628  /// Create a position describing the returned value of \p F.
629  /// \p CBContext is used for call base specific analysis.
630  static const IRPosition returned(const Function &F,
631                                   const CallBaseContext *CBContext = nullptr) {
632    return IRPosition(const_cast<Function &>(F), IRP_RETURNED, CBContext);
633  }
634
635  /// Create a position describing the argument \p Arg.
636  /// \p CBContext is used for call base specific analysis.
637  static const IRPosition argument(const Argument &Arg,
638                                   const CallBaseContext *CBContext = nullptr) {
639    return IRPosition(const_cast<Argument &>(Arg), IRP_ARGUMENT, CBContext);
640  }
641
642  /// Create a position describing the function scope of \p CB.
643  static const IRPosition callsite_function(const CallBase &CB) {
644    return IRPosition(const_cast<CallBase &>(CB), IRP_CALL_SITE);
645  }
646
647  /// Create a position describing the returned value of \p CB.
648  static const IRPosition callsite_returned(const CallBase &CB) {
649    return IRPosition(const_cast<CallBase &>(CB), IRP_CALL_SITE_RETURNED);
650  }
651
652  /// Create a position describing the argument of \p CB at position \p ArgNo.
653  static const IRPosition callsite_argument(const CallBase &CB,
654                                            unsigned ArgNo) {
655    return IRPosition(const_cast<Use &>(CB.getArgOperandUse(ArgNo)),
656                      IRP_CALL_SITE_ARGUMENT);
657  }
658
659  /// Create a position describing the argument of \p ACS at position \p ArgNo.
660  static const IRPosition callsite_argument(AbstractCallSite ACS,
661                                            unsigned ArgNo) {
662    if (ACS.getNumArgOperands() <= ArgNo)
663      return IRPosition();
664    int CSArgNo = ACS.getCallArgOperandNo(ArgNo);
665    if (CSArgNo >= 0)
666      return IRPosition::callsite_argument(
667          cast<CallBase>(*ACS.getInstruction()), CSArgNo);
668    return IRPosition();
669  }
670
671  /// Create a position with function scope matching the "context" of \p IRP.
672  /// If \p IRP is a call site (see isAnyCallSitePosition()) then the result
673  /// will be a call site position, otherwise the function position of the
674  /// associated function.
675  static const IRPosition
676  function_scope(const IRPosition &IRP,
677                 const CallBaseContext *CBContext = nullptr) {
678    if (IRP.isAnyCallSitePosition()) {
679      return IRPosition::callsite_function(
680          cast<CallBase>(IRP.getAnchorValue()));
681    }
682    assert(IRP.getAssociatedFunction());
683    return IRPosition::function(*IRP.getAssociatedFunction(), CBContext);
684  }
685
686  bool operator==(const IRPosition &RHS) const {
687    return Enc == RHS.Enc && RHS.CBContext == CBContext;
688  }
689  bool operator!=(const IRPosition &RHS) const { return !(*this == RHS); }
690
691  /// Return the value this abstract attribute is anchored with.
692  ///
693  /// The anchor value might not be the associated value if the latter is not
694  /// sufficient to determine where arguments will be manifested. This is, so
695  /// far, only the case for call site arguments as the value is not sufficient
696  /// to pinpoint them. Instead, we can use the call site as an anchor.
697  Value &getAnchorValue() const {
698    switch (getEncodingBits()) {
699    case ENC_VALUE:
700    case ENC_RETURNED_VALUE:
701    case ENC_FLOATING_FUNCTION:
702      return *getAsValuePtr();
703    case ENC_CALL_SITE_ARGUMENT_USE:
704      return *(getAsUsePtr()->getUser());
705    default:
706      llvm_unreachable("Unkown encoding!");
707    };
708  }
709
710  /// Return the associated function, if any.
711  Function *getAssociatedFunction() const {
712    if (auto *CB = dyn_cast<CallBase>(&getAnchorValue())) {
713      // We reuse the logic that associates callback calles to arguments of a
714      // call site here to identify the callback callee as the associated
715      // function.
716      if (Argument *Arg = getAssociatedArgument())
717        return Arg->getParent();
718      return dyn_cast_if_present<Function>(
719          CB->getCalledOperand()->stripPointerCasts());
720    }
721    return getAnchorScope();
722  }
723
724  /// Return the associated argument, if any.
725  Argument *getAssociatedArgument() const;
726
727  /// Return true if the position refers to a function interface, that is the
728  /// function scope, the function return, or an argument.
729  bool isFnInterfaceKind() const {
730    switch (getPositionKind()) {
731    case IRPosition::IRP_FUNCTION:
732    case IRPosition::IRP_RETURNED:
733    case IRPosition::IRP_ARGUMENT:
734      return true;
735    default:
736      return false;
737    }
738  }
739
740  /// Return true if this is a function or call site position.
741  bool isFunctionScope() const {
742    switch (getPositionKind()) {
743    case IRPosition::IRP_CALL_SITE:
744    case IRPosition::IRP_FUNCTION:
745      return true;
746    default:
747      return false;
748    };
749  }
750
751  /// Return the Function surrounding the anchor value.
752  Function *getAnchorScope() const {
753    Value &V = getAnchorValue();
754    if (isa<Function>(V))
755      return &cast<Function>(V);
756    if (isa<Argument>(V))
757      return cast<Argument>(V).getParent();
758    if (isa<Instruction>(V))
759      return cast<Instruction>(V).getFunction();
760    return nullptr;
761  }
762
763  /// Return the context instruction, if any.
764  Instruction *getCtxI() const {
765    Value &V = getAnchorValue();
766    if (auto *I = dyn_cast<Instruction>(&V))
767      return I;
768    if (auto *Arg = dyn_cast<Argument>(&V))
769      if (!Arg->getParent()->isDeclaration())
770        return &Arg->getParent()->getEntryBlock().front();
771    if (auto *F = dyn_cast<Function>(&V))
772      if (!F->isDeclaration())
773        return &(F->getEntryBlock().front());
774    return nullptr;
775  }
776
777  /// Return the value this abstract attribute is associated with.
778  Value &getAssociatedValue() const {
779    if (getCallSiteArgNo() < 0 || isa<Argument>(&getAnchorValue()))
780      return getAnchorValue();
781    assert(isa<CallBase>(&getAnchorValue()) && "Expected a call base!");
782    return *cast<CallBase>(&getAnchorValue())
783                ->getArgOperand(getCallSiteArgNo());
784  }
785
786  /// Return the type this abstract attribute is associated with.
787  Type *getAssociatedType() const {
788    if (getPositionKind() == IRPosition::IRP_RETURNED)
789      return getAssociatedFunction()->getReturnType();
790    return getAssociatedValue().getType();
791  }
792
793  /// Return the callee argument number of the associated value if it is an
794  /// argument or call site argument, otherwise a negative value. In contrast to
795  /// `getCallSiteArgNo` this method will always return the "argument number"
796  /// from the perspective of the callee. This may not the same as the call site
797  /// if this is a callback call.
798  int getCalleeArgNo() const {
799    return getArgNo(/* CallbackCalleeArgIfApplicable */ true);
800  }
801
802  /// Return the call site argument number of the associated value if it is an
803  /// argument or call site argument, otherwise a negative value. In contrast to
804  /// `getCalleArgNo` this method will always return the "operand number" from
805  /// the perspective of the call site. This may not the same as the callee
806  /// perspective if this is a callback call.
807  int getCallSiteArgNo() const {
808    return getArgNo(/* CallbackCalleeArgIfApplicable */ false);
809  }
810
811  /// Return the index in the attribute list for this position.
812  unsigned getAttrIdx() const {
813    switch (getPositionKind()) {
814    case IRPosition::IRP_INVALID:
815    case IRPosition::IRP_FLOAT:
816      break;
817    case IRPosition::IRP_FUNCTION:
818    case IRPosition::IRP_CALL_SITE:
819      return AttributeList::FunctionIndex;
820    case IRPosition::IRP_RETURNED:
821    case IRPosition::IRP_CALL_SITE_RETURNED:
822      return AttributeList::ReturnIndex;
823    case IRPosition::IRP_ARGUMENT:
824      return getCalleeArgNo() + AttributeList::FirstArgIndex;
825    case IRPosition::IRP_CALL_SITE_ARGUMENT:
826      return getCallSiteArgNo() + AttributeList::FirstArgIndex;
827    }
828    llvm_unreachable(
829        "There is no attribute index for a floating or invalid position!");
830  }
831
832  /// Return the value attributes are attached to.
833  Value *getAttrListAnchor() const {
834    if (auto *CB = dyn_cast<CallBase>(&getAnchorValue()))
835      return CB;
836    return getAssociatedFunction();
837  }
838
839  /// Return the attributes associated with this function or call site scope.
840  AttributeList getAttrList() const {
841    if (auto *CB = dyn_cast<CallBase>(&getAnchorValue()))
842      return CB->getAttributes();
843    return getAssociatedFunction()->getAttributes();
844  }
845
846  /// Update the attributes associated with this function or call site scope.
847  void setAttrList(const AttributeList &AttrList) const {
848    if (auto *CB = dyn_cast<CallBase>(&getAnchorValue()))
849      return CB->setAttributes(AttrList);
850    return getAssociatedFunction()->setAttributes(AttrList);
851  }
852
853  /// Return the number of arguments associated with this function or call site
854  /// scope.
855  unsigned getNumArgs() const {
856    assert((getPositionKind() == IRP_CALL_SITE ||
857            getPositionKind() == IRP_FUNCTION) &&
858           "Only valid for function/call site positions!");
859    if (auto *CB = dyn_cast<CallBase>(&getAnchorValue()))
860      return CB->arg_size();
861    return getAssociatedFunction()->arg_size();
862  }
863
864  /// Return theargument \p ArgNo associated with this function or call site
865  /// scope.
866  Value *getArg(unsigned ArgNo) const {
867    assert((getPositionKind() == IRP_CALL_SITE ||
868            getPositionKind() == IRP_FUNCTION) &&
869           "Only valid for function/call site positions!");
870    if (auto *CB = dyn_cast<CallBase>(&getAnchorValue()))
871      return CB->getArgOperand(ArgNo);
872    return getAssociatedFunction()->getArg(ArgNo);
873  }
874
875  /// Return the associated position kind.
876  Kind getPositionKind() const {
877    char EncodingBits = getEncodingBits();
878    if (EncodingBits == ENC_CALL_SITE_ARGUMENT_USE)
879      return IRP_CALL_SITE_ARGUMENT;
880    if (EncodingBits == ENC_FLOATING_FUNCTION)
881      return IRP_FLOAT;
882
883    Value *V = getAsValuePtr();
884    if (!V)
885      return IRP_INVALID;
886    if (isa<Argument>(V))
887      return IRP_ARGUMENT;
888    if (isa<Function>(V))
889      return isReturnPosition(EncodingBits) ? IRP_RETURNED : IRP_FUNCTION;
890    if (isa<CallBase>(V))
891      return isReturnPosition(EncodingBits) ? IRP_CALL_SITE_RETURNED
892                                            : IRP_CALL_SITE;
893    return IRP_FLOAT;
894  }
895
896  bool isAnyCallSitePosition() const {
897    switch (getPositionKind()) {
898    case IRPosition::IRP_CALL_SITE:
899    case IRPosition::IRP_CALL_SITE_RETURNED:
900    case IRPosition::IRP_CALL_SITE_ARGUMENT:
901      return true;
902    default:
903      return false;
904    }
905  }
906
907  /// Return true if the position is an argument or call site argument.
908  bool isArgumentPosition() const {
909    switch (getPositionKind()) {
910    case IRPosition::IRP_ARGUMENT:
911    case IRPosition::IRP_CALL_SITE_ARGUMENT:
912      return true;
913    default:
914      return false;
915    }
916  }
917
918  /// Return the same position without the call base context.
919  IRPosition stripCallBaseContext() const {
920    IRPosition Result = *this;
921    Result.CBContext = nullptr;
922    return Result;
923  }
924
925  /// Get the call base context from the position.
926  const CallBaseContext *getCallBaseContext() const { return CBContext; }
927
928  /// Check if the position has any call base context.
929  bool hasCallBaseContext() const { return CBContext != nullptr; }
930
931  /// Special DenseMap key values.
932  ///
933  ///{
934  static const IRPosition EmptyKey;
935  static const IRPosition TombstoneKey;
936  ///}
937
938  /// Conversion into a void * to allow reuse of pointer hashing.
939  operator void *() const { return Enc.getOpaqueValue(); }
940
941private:
942  /// Private constructor for special values only!
943  explicit IRPosition(void *Ptr, const CallBaseContext *CBContext = nullptr)
944      : CBContext(CBContext) {
945    Enc.setFromOpaqueValue(Ptr);
946  }
947
948  /// IRPosition anchored at \p AnchorVal with kind/argument numbet \p PK.
949  explicit IRPosition(Value &AnchorVal, Kind PK,
950                      const CallBaseContext *CBContext = nullptr)
951      : CBContext(CBContext) {
952    switch (PK) {
953    case IRPosition::IRP_INVALID:
954      llvm_unreachable("Cannot create invalid IRP with an anchor value!");
955      break;
956    case IRPosition::IRP_FLOAT:
957      // Special case for floating functions.
958      if (isa<Function>(AnchorVal) || isa<CallBase>(AnchorVal))
959        Enc = {&AnchorVal, ENC_FLOATING_FUNCTION};
960      else
961        Enc = {&AnchorVal, ENC_VALUE};
962      break;
963    case IRPosition::IRP_FUNCTION:
964    case IRPosition::IRP_CALL_SITE:
965      Enc = {&AnchorVal, ENC_VALUE};
966      break;
967    case IRPosition::IRP_RETURNED:
968    case IRPosition::IRP_CALL_SITE_RETURNED:
969      Enc = {&AnchorVal, ENC_RETURNED_VALUE};
970      break;
971    case IRPosition::IRP_ARGUMENT:
972      Enc = {&AnchorVal, ENC_VALUE};
973      break;
974    case IRPosition::IRP_CALL_SITE_ARGUMENT:
975      llvm_unreachable(
976          "Cannot create call site argument IRP with an anchor value!");
977      break;
978    }
979    verify();
980  }
981
982  /// Return the callee argument number of the associated value if it is an
983  /// argument or call site argument. See also `getCalleeArgNo` and
984  /// `getCallSiteArgNo`.
985  int getArgNo(bool CallbackCalleeArgIfApplicable) const {
986    if (CallbackCalleeArgIfApplicable)
987      if (Argument *Arg = getAssociatedArgument())
988        return Arg->getArgNo();
989    switch (getPositionKind()) {
990    case IRPosition::IRP_ARGUMENT:
991      return cast<Argument>(getAsValuePtr())->getArgNo();
992    case IRPosition::IRP_CALL_SITE_ARGUMENT: {
993      Use &U = *getAsUsePtr();
994      return cast<CallBase>(U.getUser())->getArgOperandNo(&U);
995    }
996    default:
997      return -1;
998    }
999  }
1000
1001  /// IRPosition for the use \p U. The position kind \p PK needs to be
1002  /// IRP_CALL_SITE_ARGUMENT, the anchor value is the user, the associated value
1003  /// the used value.
1004  explicit IRPosition(Use &U, Kind PK) {
1005    assert(PK == IRP_CALL_SITE_ARGUMENT &&
1006           "Use constructor is for call site arguments only!");
1007    Enc = {&U, ENC_CALL_SITE_ARGUMENT_USE};
1008    verify();
1009  }
1010
1011  /// Verify internal invariants.
1012  void verify();
1013
1014  /// Return the underlying pointer as Value *, valid for all positions but
1015  /// IRP_CALL_SITE_ARGUMENT.
1016  Value *getAsValuePtr() const {
1017    assert(getEncodingBits() != ENC_CALL_SITE_ARGUMENT_USE &&
1018           "Not a value pointer!");
1019    return reinterpret_cast<Value *>(Enc.getPointer());
1020  }
1021
1022  /// Return the underlying pointer as Use *, valid only for
1023  /// IRP_CALL_SITE_ARGUMENT positions.
1024  Use *getAsUsePtr() const {
1025    assert(getEncodingBits() == ENC_CALL_SITE_ARGUMENT_USE &&
1026           "Not a value pointer!");
1027    return reinterpret_cast<Use *>(Enc.getPointer());
1028  }
1029
1030  /// Return true if \p EncodingBits describe a returned or call site returned
1031  /// position.
1032  static bool isReturnPosition(char EncodingBits) {
1033    return EncodingBits == ENC_RETURNED_VALUE;
1034  }
1035
1036  /// Return true if the encoding bits describe a returned or call site returned
1037  /// position.
1038  bool isReturnPosition() const { return isReturnPosition(getEncodingBits()); }
1039
1040  /// The encoding of the IRPosition is a combination of a pointer and two
1041  /// encoding bits. The values of the encoding bits are defined in the enum
1042  /// below. The pointer is either a Value* (for the first three encoding bit
1043  /// combinations) or Use* (for ENC_CALL_SITE_ARGUMENT_USE).
1044  ///
1045  ///{
1046  enum {
1047    ENC_VALUE = 0b00,
1048    ENC_RETURNED_VALUE = 0b01,
1049    ENC_FLOATING_FUNCTION = 0b10,
1050    ENC_CALL_SITE_ARGUMENT_USE = 0b11,
1051  };
1052
1053  // Reserve the maximal amount of bits so there is no need to mask out the
1054  // remaining ones. We will not encode anything else in the pointer anyway.
1055  static constexpr int NumEncodingBits =
1056      PointerLikeTypeTraits<void *>::NumLowBitsAvailable;
1057  static_assert(NumEncodingBits >= 2, "At least two bits are required!");
1058
1059  /// The pointer with the encoding bits.
1060  PointerIntPair<void *, NumEncodingBits, char> Enc;
1061  ///}
1062
1063  /// Call base context. Used for callsite specific analysis.
1064  const CallBaseContext *CBContext = nullptr;
1065
1066  /// Return the encoding bits.
1067  char getEncodingBits() const { return Enc.getInt(); }
1068};
1069
1070/// Helper that allows IRPosition as a key in a DenseMap.
1071template <> struct DenseMapInfo<IRPosition> {
1072  static inline IRPosition getEmptyKey() { return IRPosition::EmptyKey; }
1073  static inline IRPosition getTombstoneKey() {
1074    return IRPosition::TombstoneKey;
1075  }
1076  static unsigned getHashValue(const IRPosition &IRP) {
1077    return (DenseMapInfo<void *>::getHashValue(IRP) << 4) ^
1078           (DenseMapInfo<Value *>::getHashValue(IRP.getCallBaseContext()));
1079  }
1080
1081  static bool isEqual(const IRPosition &a, const IRPosition &b) {
1082    return a == b;
1083  }
1084};
1085
1086/// A visitor class for IR positions.
1087///
1088/// Given a position P, the SubsumingPositionIterator allows to visit "subsuming
1089/// positions" wrt. attributes/information. Thus, if a piece of information
1090/// holds for a subsuming position, it also holds for the position P.
1091///
1092/// The subsuming positions always include the initial position and then,
1093/// depending on the position kind, additionally the following ones:
1094/// - for IRP_RETURNED:
1095///   - the function (IRP_FUNCTION)
1096/// - for IRP_ARGUMENT:
1097///   - the function (IRP_FUNCTION)
1098/// - for IRP_CALL_SITE:
1099///   - the callee (IRP_FUNCTION), if known
1100/// - for IRP_CALL_SITE_RETURNED:
1101///   - the callee (IRP_RETURNED), if known
1102///   - the call site (IRP_FUNCTION)
1103///   - the callee (IRP_FUNCTION), if known
1104/// - for IRP_CALL_SITE_ARGUMENT:
1105///   - the argument of the callee (IRP_ARGUMENT), if known
1106///   - the callee (IRP_FUNCTION), if known
1107///   - the position the call site argument is associated with if it is not
1108///     anchored to the call site, e.g., if it is an argument then the argument
1109///     (IRP_ARGUMENT)
1110class SubsumingPositionIterator {
1111  SmallVector<IRPosition, 4> IRPositions;
1112  using iterator = decltype(IRPositions)::iterator;
1113
1114public:
1115  SubsumingPositionIterator(const IRPosition &IRP);
1116  iterator begin() { return IRPositions.begin(); }
1117  iterator end() { return IRPositions.end(); }
1118};
1119
1120/// Wrapper for FunctionAnalysisManager.
1121struct AnalysisGetter {
1122  // The client may be running the old pass manager, in which case, we need to
1123  // map the requested Analysis to its equivalent wrapper in the old pass
1124  // manager. The scheme implemented here does not require every Analysis to be
1125  // updated. Only those new analyses that the client cares about in the old
1126  // pass manager need to expose a LegacyWrapper type, and that wrapper should
1127  // support a getResult() method that matches the new Analysis.
1128  //
1129  // We need SFINAE to check for the LegacyWrapper, but function templates don't
1130  // allow partial specialization, which is needed in this case. So instead, we
1131  // use a constexpr bool to perform the SFINAE, and then use this information
1132  // inside the function template.
1133  template <typename, typename = void>
1134  static constexpr bool HasLegacyWrapper = false;
1135
1136  template <typename Analysis>
1137  typename Analysis::Result *getAnalysis(const Function &F,
1138                                         bool RequestCachedOnly = false) {
1139    if (!LegacyPass && !FAM)
1140      return nullptr;
1141    if (FAM) {
1142      if (CachedOnly || RequestCachedOnly)
1143        return FAM->getCachedResult<Analysis>(const_cast<Function &>(F));
1144      return &FAM->getResult<Analysis>(const_cast<Function &>(F));
1145    }
1146    if constexpr (HasLegacyWrapper<Analysis>) {
1147      if (!CachedOnly && !RequestCachedOnly)
1148        return &LegacyPass
1149                    ->getAnalysis<typename Analysis::LegacyWrapper>(
1150                        const_cast<Function &>(F))
1151                    .getResult();
1152      if (auto *P =
1153              LegacyPass
1154                  ->getAnalysisIfAvailable<typename Analysis::LegacyWrapper>())
1155        return &P->getResult();
1156    }
1157    return nullptr;
1158  }
1159
1160  /// Invalidates the analyses. Valid only when using the new pass manager.
1161  void invalidateAnalyses() {
1162    assert(FAM && "Can only be used from the new PM!");
1163    FAM->clear();
1164  }
1165
1166  AnalysisGetter(FunctionAnalysisManager &FAM, bool CachedOnly = false)
1167      : FAM(&FAM), CachedOnly(CachedOnly) {}
1168  AnalysisGetter(Pass *P, bool CachedOnly = false)
1169      : LegacyPass(P), CachedOnly(CachedOnly) {}
1170  AnalysisGetter() = default;
1171
1172private:
1173  FunctionAnalysisManager *FAM = nullptr;
1174  Pass *LegacyPass = nullptr;
1175
1176  /// If \p CachedOnly is true, no pass is created, just existing results are
1177  /// used. Also available per request.
1178  bool CachedOnly = false;
1179};
1180
1181template <typename Analysis>
1182constexpr bool AnalysisGetter::HasLegacyWrapper<
1183    Analysis, std::void_t<typename Analysis::LegacyWrapper>> = true;
1184
1185/// Data structure to hold cached (LLVM-IR) information.
1186///
1187/// All attributes are given an InformationCache object at creation time to
1188/// avoid inspection of the IR by all of them individually. This default
1189/// InformationCache will hold information required by 'default' attributes,
1190/// thus the ones deduced when Attributor::identifyDefaultAbstractAttributes(..)
1191/// is called.
1192///
1193/// If custom abstract attributes, registered manually through
1194/// Attributor::registerAA(...), need more information, especially if it is not
1195/// reusable, it is advised to inherit from the InformationCache and cast the
1196/// instance down in the abstract attributes.
1197struct InformationCache {
1198  InformationCache(const Module &M, AnalysisGetter &AG,
1199                   BumpPtrAllocator &Allocator, SetVector<Function *> *CGSCC,
1200                   bool UseExplorer = true)
1201      : CGSCC(CGSCC), DL(M.getDataLayout()), Allocator(Allocator), AG(AG),
1202        TargetTriple(M.getTargetTriple()) {
1203    if (UseExplorer)
1204      Explorer = new (Allocator) MustBeExecutedContextExplorer(
1205          /* ExploreInterBlock */ true, /* ExploreCFGForward */ true,
1206          /* ExploreCFGBackward */ true,
1207          /* LIGetter */
1208          [&](const Function &F) { return AG.getAnalysis<LoopAnalysis>(F); },
1209          /* DTGetter */
1210          [&](const Function &F) {
1211            return AG.getAnalysis<DominatorTreeAnalysis>(F);
1212          },
1213          /* PDTGetter */
1214          [&](const Function &F) {
1215            return AG.getAnalysis<PostDominatorTreeAnalysis>(F);
1216          });
1217  }
1218
1219  ~InformationCache() {
1220    // The FunctionInfo objects are allocated via a BumpPtrAllocator, we call
1221    // the destructor manually.
1222    for (auto &It : FuncInfoMap)
1223      It.getSecond()->~FunctionInfo();
1224    // Same is true for the instruction exclusions sets.
1225    using AA::InstExclusionSetTy;
1226    for (auto *BES : BESets)
1227      BES->~InstExclusionSetTy();
1228    if (Explorer)
1229      Explorer->~MustBeExecutedContextExplorer();
1230  }
1231
1232  /// Apply \p CB to all uses of \p F. If \p LookThroughConstantExprUses is
1233  /// true, constant expression users are not given to \p CB but their uses are
1234  /// traversed transitively.
1235  template <typename CBTy>
1236  static void foreachUse(Function &F, CBTy CB,
1237                         bool LookThroughConstantExprUses = true) {
1238    SmallVector<Use *, 8> Worklist(make_pointer_range(F.uses()));
1239
1240    for (unsigned Idx = 0; Idx < Worklist.size(); ++Idx) {
1241      Use &U = *Worklist[Idx];
1242
1243      // Allow use in constant bitcasts and simply look through them.
1244      if (LookThroughConstantExprUses && isa<ConstantExpr>(U.getUser())) {
1245        for (Use &CEU : cast<ConstantExpr>(U.getUser())->uses())
1246          Worklist.push_back(&CEU);
1247        continue;
1248      }
1249
1250      CB(U);
1251    }
1252  }
1253
1254  /// The CG-SCC the pass is run on, or nullptr if it is a module pass.
1255  const SetVector<Function *> *const CGSCC = nullptr;
1256
1257  /// A vector type to hold instructions.
1258  using InstructionVectorTy = SmallVector<Instruction *, 8>;
1259
1260  /// A map type from opcodes to instructions with this opcode.
1261  using OpcodeInstMapTy = DenseMap<unsigned, InstructionVectorTy *>;
1262
1263  /// Return the map that relates "interesting" opcodes with all instructions
1264  /// with that opcode in \p F.
1265  OpcodeInstMapTy &getOpcodeInstMapForFunction(const Function &F) {
1266    return getFunctionInfo(F).OpcodeInstMap;
1267  }
1268
1269  /// Return the instructions in \p F that may read or write memory.
1270  InstructionVectorTy &getReadOrWriteInstsForFunction(const Function &F) {
1271    return getFunctionInfo(F).RWInsts;
1272  }
1273
1274  /// Return MustBeExecutedContextExplorer
1275  MustBeExecutedContextExplorer *getMustBeExecutedContextExplorer() {
1276    return Explorer;
1277  }
1278
1279  /// Return TargetLibraryInfo for function \p F.
1280  TargetLibraryInfo *getTargetLibraryInfoForFunction(const Function &F) {
1281    return AG.getAnalysis<TargetLibraryAnalysis>(F);
1282  }
1283
1284  /// Return true if \p Arg is involved in a must-tail call, thus the argument
1285  /// of the caller or callee.
1286  bool isInvolvedInMustTailCall(const Argument &Arg) {
1287    FunctionInfo &FI = getFunctionInfo(*Arg.getParent());
1288    return FI.CalledViaMustTail || FI.ContainsMustTailCall;
1289  }
1290
1291  bool isOnlyUsedByAssume(const Instruction &I) const {
1292    return AssumeOnlyValues.contains(&I);
1293  }
1294
1295  /// Invalidates the cached analyses. Valid only when using the new pass
1296  /// manager.
1297  void invalidateAnalyses() { AG.invalidateAnalyses(); }
1298
1299  /// Return the analysis result from a pass \p AP for function \p F.
1300  template <typename AP>
1301  typename AP::Result *getAnalysisResultForFunction(const Function &F,
1302                                                    bool CachedOnly = false) {
1303    return AG.getAnalysis<AP>(F, CachedOnly);
1304  }
1305
1306  /// Return datalayout used in the module.
1307  const DataLayout &getDL() { return DL; }
1308
1309  /// Return the map conaining all the knowledge we have from `llvm.assume`s.
1310  const RetainedKnowledgeMap &getKnowledgeMap() const { return KnowledgeMap; }
1311
1312  /// Given \p BES, return a uniqued version.
1313  const AA::InstExclusionSetTy *
1314  getOrCreateUniqueBlockExecutionSet(const AA::InstExclusionSetTy *BES) {
1315    auto It = BESets.find(BES);
1316    if (It != BESets.end())
1317      return *It;
1318    auto *UniqueBES = new (Allocator) AA::InstExclusionSetTy(*BES);
1319    bool Success = BESets.insert(UniqueBES).second;
1320    (void)Success;
1321    assert(Success && "Expected only new entries to be added");
1322    return UniqueBES;
1323  }
1324
1325  /// Return true if the stack (llvm::Alloca) can be accessed by other threads.
1326  bool stackIsAccessibleByOtherThreads() { return !targetIsGPU(); }
1327
1328  /// Return true if the target is a GPU.
1329  bool targetIsGPU() {
1330    return TargetTriple.isAMDGPU() || TargetTriple.isNVPTX();
1331  }
1332
1333  /// Return all functions that might be called indirectly, only valid for
1334  /// closed world modules (see isClosedWorldModule).
1335  const ArrayRef<Function *>
1336  getIndirectlyCallableFunctions(Attributor &A) const;
1337
1338private:
1339  struct FunctionInfo {
1340    ~FunctionInfo();
1341
1342    /// A nested map that remembers all instructions in a function with a
1343    /// certain instruction opcode (Instruction::getOpcode()).
1344    OpcodeInstMapTy OpcodeInstMap;
1345
1346    /// A map from functions to their instructions that may read or write
1347    /// memory.
1348    InstructionVectorTy RWInsts;
1349
1350    /// Function is called by a `musttail` call.
1351    bool CalledViaMustTail;
1352
1353    /// Function contains a `musttail` call.
1354    bool ContainsMustTailCall;
1355  };
1356
1357  /// A map type from functions to informatio about it.
1358  DenseMap<const Function *, FunctionInfo *> FuncInfoMap;
1359
1360  /// Return information about the function \p F, potentially by creating it.
1361  FunctionInfo &getFunctionInfo(const Function &F) {
1362    FunctionInfo *&FI = FuncInfoMap[&F];
1363    if (!FI) {
1364      FI = new (Allocator) FunctionInfo();
1365      initializeInformationCache(F, *FI);
1366    }
1367    return *FI;
1368  }
1369
1370  /// Vector of functions that might be callable indirectly, i.a., via a
1371  /// function pointer.
1372  SmallVector<Function *> IndirectlyCallableFunctions;
1373
1374  /// Initialize the function information cache \p FI for the function \p F.
1375  ///
1376  /// This method needs to be called for all function that might be looked at
1377  /// through the information cache interface *prior* to looking at them.
1378  void initializeInformationCache(const Function &F, FunctionInfo &FI);
1379
1380  /// The datalayout used in the module.
1381  const DataLayout &DL;
1382
1383  /// The allocator used to allocate memory, e.g. for `FunctionInfo`s.
1384  BumpPtrAllocator &Allocator;
1385
1386  /// MustBeExecutedContextExplorer
1387  MustBeExecutedContextExplorer *Explorer = nullptr;
1388
1389  /// A map with knowledge retained in `llvm.assume` instructions.
1390  RetainedKnowledgeMap KnowledgeMap;
1391
1392  /// A container for all instructions that are only used by `llvm.assume`.
1393  SetVector<const Instruction *> AssumeOnlyValues;
1394
1395  /// Cache for block sets to allow reuse.
1396  DenseSet<const AA::InstExclusionSetTy *> BESets;
1397
1398  /// Getters for analysis.
1399  AnalysisGetter &AG;
1400
1401  /// Set of inlineable functions
1402  SmallPtrSet<const Function *, 8> InlineableFunctions;
1403
1404  /// The triple describing the target machine.
1405  Triple TargetTriple;
1406
1407  /// Give the Attributor access to the members so
1408  /// Attributor::identifyDefaultAbstractAttributes(...) can initialize them.
1409  friend struct Attributor;
1410};
1411
1412/// Configuration for the Attributor.
1413struct AttributorConfig {
1414
1415  AttributorConfig(CallGraphUpdater &CGUpdater) : CGUpdater(CGUpdater) {}
1416
1417  /// Is the user of the Attributor a module pass or not. This determines what
1418  /// IR we can look at and modify. If it is a module pass we might deduce facts
1419  /// outside the initial function set and modify functions outside that set,
1420  /// but only as part of the optimization of the functions in the initial
1421  /// function set. For CGSCC passes we can look at the IR of the module slice
1422  /// but never run any deduction, or perform any modification, outside the
1423  /// initial function set (which we assume is the SCC).
1424  bool IsModulePass = true;
1425
1426  /// Flag to determine if we can delete functions or keep dead ones around.
1427  bool DeleteFns = true;
1428
1429  /// Flag to determine if we rewrite function signatures.
1430  bool RewriteSignatures = true;
1431
1432  /// Flag to determine if we want to initialize all default AAs for an internal
1433  /// function marked live. See also: InitializationCallback>
1434  bool DefaultInitializeLiveInternals = true;
1435
1436  /// Flag to determine if we should skip all liveness checks early on.
1437  bool UseLiveness = true;
1438
1439  /// Flag to indicate if the entire world is contained in this module, that
1440  /// is, no outside functions exist.
1441  bool IsClosedWorldModule = false;
1442
1443  /// Callback function to be invoked on internal functions marked live.
1444  std::function<void(Attributor &A, const Function &F)> InitializationCallback =
1445      nullptr;
1446
1447  /// Callback function to determine if an indirect call targets should be made
1448  /// direct call targets (with an if-cascade).
1449  std::function<bool(Attributor &A, const AbstractAttribute &AA, CallBase &CB,
1450                     Function &AssummedCallee)>
1451      IndirectCalleeSpecializationCallback = nullptr;
1452
1453  /// Helper to update an underlying call graph and to delete functions.
1454  CallGraphUpdater &CGUpdater;
1455
1456  /// If not null, a set limiting the attribute opportunities.
1457  DenseSet<const char *> *Allowed = nullptr;
1458
1459  /// Maximum number of iterations to run until fixpoint.
1460  std::optional<unsigned> MaxFixpointIterations;
1461
1462  /// A callback function that returns an ORE object from a Function pointer.
1463  ///{
1464  using OptimizationRemarkGetter =
1465      function_ref<OptimizationRemarkEmitter &(Function *)>;
1466  OptimizationRemarkGetter OREGetter = nullptr;
1467  ///}
1468
1469  /// The name of the pass running the attributor, used to emit remarks.
1470  const char *PassName = nullptr;
1471
1472  using IPOAmendableCBTy = function_ref<bool(const Function &F)>;
1473  IPOAmendableCBTy IPOAmendableCB;
1474};
1475
1476/// A debug counter to limit the number of AAs created.
1477DEBUG_COUNTER(NumAbstractAttributes, "num-abstract-attributes",
1478              "How many AAs should be initialized");
1479
1480/// The fixpoint analysis framework that orchestrates the attribute deduction.
1481///
1482/// The Attributor provides a general abstract analysis framework (guided
1483/// fixpoint iteration) as well as helper functions for the deduction of
1484/// (LLVM-IR) attributes. However, also other code properties can be deduced,
1485/// propagated, and ultimately manifested through the Attributor framework. This
1486/// is particularly useful if these properties interact with attributes and a
1487/// co-scheduled deduction allows to improve the solution. Even if not, thus if
1488/// attributes/properties are completely isolated, they should use the
1489/// Attributor framework to reduce the number of fixpoint iteration frameworks
1490/// in the code base. Note that the Attributor design makes sure that isolated
1491/// attributes are not impacted, in any way, by others derived at the same time
1492/// if there is no cross-reasoning performed.
1493///
1494/// The public facing interface of the Attributor is kept simple and basically
1495/// allows abstract attributes to one thing, query abstract attributes
1496/// in-flight. There are two reasons to do this:
1497///    a) The optimistic state of one abstract attribute can justify an
1498///       optimistic state of another, allowing to framework to end up with an
1499///       optimistic (=best possible) fixpoint instead of one based solely on
1500///       information in the IR.
1501///    b) This avoids reimplementing various kinds of lookups, e.g., to check
1502///       for existing IR attributes, in favor of a single lookups interface
1503///       provided by an abstract attribute subclass.
1504///
1505/// NOTE: The mechanics of adding a new "concrete" abstract attribute are
1506///       described in the file comment.
1507struct Attributor {
1508
1509  /// Constructor
1510  ///
1511  /// \param Functions The set of functions we are deriving attributes for.
1512  /// \param InfoCache Cache to hold various information accessible for
1513  ///                  the abstract attributes.
1514  /// \param Configuration The Attributor configuration which determines what
1515  ///                      generic features to use.
1516  Attributor(SetVector<Function *> &Functions, InformationCache &InfoCache,
1517             AttributorConfig Configuration);
1518
1519  ~Attributor();
1520
1521  /// Run the analyses until a fixpoint is reached or enforced (timeout).
1522  ///
1523  /// The attributes registered with this Attributor can be used after as long
1524  /// as the Attributor is not destroyed (it owns the attributes now).
1525  ///
1526  /// \Returns CHANGED if the IR was changed, otherwise UNCHANGED.
1527  ChangeStatus run();
1528
1529  /// Lookup an abstract attribute of type \p AAType at position \p IRP. While
1530  /// no abstract attribute is found equivalent positions are checked, see
1531  /// SubsumingPositionIterator. Thus, the returned abstract attribute
1532  /// might be anchored at a different position, e.g., the callee if \p IRP is a
1533  /// call base.
1534  ///
1535  /// This method is the only (supported) way an abstract attribute can retrieve
1536  /// information from another abstract attribute. As an example, take an
1537  /// abstract attribute that determines the memory access behavior for a
1538  /// argument (readnone, readonly, ...). It should use `getAAFor` to get the
1539  /// most optimistic information for other abstract attributes in-flight, e.g.
1540  /// the one reasoning about the "captured" state for the argument or the one
1541  /// reasoning on the memory access behavior of the function as a whole.
1542  ///
1543  /// If the DepClass enum is set to `DepClassTy::None` the dependence from
1544  /// \p QueryingAA to the return abstract attribute is not automatically
1545  /// recorded. This should only be used if the caller will record the
1546  /// dependence explicitly if necessary, thus if it the returned abstract
1547  /// attribute is used for reasoning. To record the dependences explicitly use
1548  /// the `Attributor::recordDependence` method.
1549  template <typename AAType>
1550  const AAType *getAAFor(const AbstractAttribute &QueryingAA,
1551                         const IRPosition &IRP, DepClassTy DepClass) {
1552    return getOrCreateAAFor<AAType>(IRP, &QueryingAA, DepClass,
1553                                    /* ForceUpdate */ false);
1554  }
1555
1556  /// The version of getAAFor that allows to omit a querying abstract
1557  /// attribute. Using this after Attributor started running is restricted to
1558  /// only the Attributor itself. Initial seeding of AAs can be done via this
1559  /// function.
1560  /// NOTE: ForceUpdate is ignored in any stage other than the update stage.
1561  template <typename AAType>
1562  const AAType *getOrCreateAAFor(IRPosition IRP,
1563                                 const AbstractAttribute *QueryingAA,
1564                                 DepClassTy DepClass, bool ForceUpdate = false,
1565                                 bool UpdateAfterInit = true) {
1566    if (!shouldPropagateCallBaseContext(IRP))
1567      IRP = IRP.stripCallBaseContext();
1568
1569    if (AAType *AAPtr = lookupAAFor<AAType>(IRP, QueryingAA, DepClass,
1570                                            /* AllowInvalidState */ true)) {
1571      if (ForceUpdate && Phase == AttributorPhase::UPDATE)
1572        updateAA(*AAPtr);
1573      return AAPtr;
1574    }
1575
1576    bool ShouldUpdateAA;
1577    if (!shouldInitialize<AAType>(IRP, ShouldUpdateAA))
1578      return nullptr;
1579
1580    if (!DebugCounter::shouldExecute(NumAbstractAttributes))
1581      return nullptr;
1582
1583    // No matching attribute found, create one.
1584    // Use the static create method.
1585    auto &AA = AAType::createForPosition(IRP, *this);
1586
1587    // Always register a new attribute to make sure we clean up the allocated
1588    // memory properly.
1589    registerAA(AA);
1590
1591    // If we are currenty seeding attributes, enforce seeding rules.
1592    if (Phase == AttributorPhase::SEEDING && !shouldSeedAttribute(AA)) {
1593      AA.getState().indicatePessimisticFixpoint();
1594      return &AA;
1595    }
1596
1597    // Bootstrap the new attribute with an initial update to propagate
1598    // information, e.g., function -> call site.
1599    {
1600      TimeTraceScope TimeScope("initialize", [&]() {
1601        return AA.getName() +
1602               std::to_string(AA.getIRPosition().getPositionKind());
1603      });
1604      ++InitializationChainLength;
1605      AA.initialize(*this);
1606      --InitializationChainLength;
1607    }
1608
1609    if (!ShouldUpdateAA) {
1610      AA.getState().indicatePessimisticFixpoint();
1611      return &AA;
1612    }
1613
1614    // Allow seeded attributes to declare dependencies.
1615    // Remember the seeding state.
1616    if (UpdateAfterInit) {
1617      AttributorPhase OldPhase = Phase;
1618      Phase = AttributorPhase::UPDATE;
1619
1620      updateAA(AA);
1621
1622      Phase = OldPhase;
1623    }
1624
1625    if (QueryingAA && AA.getState().isValidState())
1626      recordDependence(AA, const_cast<AbstractAttribute &>(*QueryingAA),
1627                       DepClass);
1628    return &AA;
1629  }
1630
1631  template <typename AAType>
1632  const AAType *getOrCreateAAFor(const IRPosition &IRP) {
1633    return getOrCreateAAFor<AAType>(IRP, /* QueryingAA */ nullptr,
1634                                    DepClassTy::NONE);
1635  }
1636
1637  /// Return the attribute of \p AAType for \p IRP if existing and valid. This
1638  /// also allows non-AA users lookup.
1639  template <typename AAType>
1640  AAType *lookupAAFor(const IRPosition &IRP,
1641                      const AbstractAttribute *QueryingAA = nullptr,
1642                      DepClassTy DepClass = DepClassTy::OPTIONAL,
1643                      bool AllowInvalidState = false) {
1644    static_assert(std::is_base_of<AbstractAttribute, AAType>::value,
1645                  "Cannot query an attribute with a type not derived from "
1646                  "'AbstractAttribute'!");
1647    // Lookup the abstract attribute of type AAType. If found, return it after
1648    // registering a dependence of QueryingAA on the one returned attribute.
1649    AbstractAttribute *AAPtr = AAMap.lookup({&AAType::ID, IRP});
1650    if (!AAPtr)
1651      return nullptr;
1652
1653    AAType *AA = static_cast<AAType *>(AAPtr);
1654
1655    // Do not register a dependence on an attribute with an invalid state.
1656    if (DepClass != DepClassTy::NONE && QueryingAA &&
1657        AA->getState().isValidState())
1658      recordDependence(*AA, const_cast<AbstractAttribute &>(*QueryingAA),
1659                       DepClass);
1660
1661    // Return nullptr if this attribute has an invalid state.
1662    if (!AllowInvalidState && !AA->getState().isValidState())
1663      return nullptr;
1664    return AA;
1665  }
1666
1667  /// Allows a query AA to request an update if a new query was received.
1668  void registerForUpdate(AbstractAttribute &AA);
1669
1670  /// Explicitly record a dependence from \p FromAA to \p ToAA, that is if
1671  /// \p FromAA changes \p ToAA should be updated as well.
1672  ///
1673  /// This method should be used in conjunction with the `getAAFor` method and
1674  /// with the DepClass enum passed to the method set to None. This can
1675  /// be beneficial to avoid false dependences but it requires the users of
1676  /// `getAAFor` to explicitly record true dependences through this method.
1677  /// The \p DepClass flag indicates if the dependence is striclty necessary.
1678  /// That means for required dependences, if \p FromAA changes to an invalid
1679  /// state, \p ToAA can be moved to a pessimistic fixpoint because it required
1680  /// information from \p FromAA but none are available anymore.
1681  void recordDependence(const AbstractAttribute &FromAA,
1682                        const AbstractAttribute &ToAA, DepClassTy DepClass);
1683
1684  /// Introduce a new abstract attribute into the fixpoint analysis.
1685  ///
1686  /// Note that ownership of the attribute is given to the Attributor. It will
1687  /// invoke delete for the Attributor on destruction of the Attributor.
1688  ///
1689  /// Attributes are identified by their IR position (AAType::getIRPosition())
1690  /// and the address of their static member (see AAType::ID).
1691  template <typename AAType> AAType &registerAA(AAType &AA) {
1692    static_assert(std::is_base_of<AbstractAttribute, AAType>::value,
1693                  "Cannot register an attribute with a type not derived from "
1694                  "'AbstractAttribute'!");
1695    // Put the attribute in the lookup map structure and the container we use to
1696    // keep track of all attributes.
1697    const IRPosition &IRP = AA.getIRPosition();
1698    AbstractAttribute *&AAPtr = AAMap[{&AAType::ID, IRP}];
1699
1700    assert(!AAPtr && "Attribute already in map!");
1701    AAPtr = &AA;
1702
1703    // Register AA with the synthetic root only before the manifest stage.
1704    if (Phase == AttributorPhase::SEEDING || Phase == AttributorPhase::UPDATE)
1705      DG.SyntheticRoot.Deps.insert(
1706          AADepGraphNode::DepTy(&AA, unsigned(DepClassTy::REQUIRED)));
1707
1708    return AA;
1709  }
1710
1711  /// Return the internal information cache.
1712  InformationCache &getInfoCache() { return InfoCache; }
1713
1714  /// Return true if this is a module pass, false otherwise.
1715  bool isModulePass() const { return Configuration.IsModulePass; }
1716
1717  /// Return true if we should specialize the call site \b CB for the potential
1718  /// callee \p Fn.
1719  bool shouldSpecializeCallSiteForCallee(const AbstractAttribute &AA,
1720                                         CallBase &CB, Function &Callee) {
1721    return Configuration.IndirectCalleeSpecializationCallback
1722               ? Configuration.IndirectCalleeSpecializationCallback(*this, AA,
1723                                                                    CB, Callee)
1724               : true;
1725  }
1726
1727  /// Return true if the module contains the whole world, thus, no outside
1728  /// functions exist.
1729  bool isClosedWorldModule() const;
1730
1731  /// Return true if we derive attributes for \p Fn
1732  bool isRunOn(Function &Fn) const { return isRunOn(&Fn); }
1733  bool isRunOn(Function *Fn) const {
1734    return Functions.empty() || Functions.count(Fn);
1735  }
1736
1737  template <typename AAType> bool shouldUpdateAA(const IRPosition &IRP) {
1738    // If this is queried in the manifest stage, we force the AA to indicate
1739    // pessimistic fixpoint immediately.
1740    if (Phase == AttributorPhase::MANIFEST || Phase == AttributorPhase::CLEANUP)
1741      return false;
1742
1743    Function *AssociatedFn = IRP.getAssociatedFunction();
1744
1745    if (IRP.isAnyCallSitePosition()) {
1746      // Check if we require a callee but there is none.
1747      if (!AssociatedFn && AAType::requiresCalleeForCallBase())
1748        return false;
1749
1750      // Check if we require non-asm but it is inline asm.
1751      if (AAType::requiresNonAsmForCallBase() &&
1752          cast<CallBase>(IRP.getAnchorValue()).isInlineAsm())
1753        return false;
1754    }
1755
1756    // Check if we require a calles but we can't see all.
1757    if (AAType::requiresCallersForArgOrFunction())
1758      if (IRP.getPositionKind() == IRPosition::IRP_FUNCTION ||
1759          IRP.getPositionKind() == IRPosition::IRP_ARGUMENT)
1760        if (!AssociatedFn->hasLocalLinkage())
1761          return false;
1762
1763    if (!AAType::isValidIRPositionForUpdate(*this, IRP))
1764      return false;
1765
1766    // We update only AAs associated with functions in the Functions set or
1767    // call sites of them.
1768    return (!AssociatedFn || isModulePass() || isRunOn(AssociatedFn) ||
1769            isRunOn(IRP.getAnchorScope()));
1770  }
1771
1772  template <typename AAType>
1773  bool shouldInitialize(const IRPosition &IRP, bool &ShouldUpdateAA) {
1774    if (!AAType::isValidIRPositionForInit(*this, IRP))
1775      return false;
1776
1777    if (Configuration.Allowed && !Configuration.Allowed->count(&AAType::ID))
1778      return false;
1779
1780    // For now we skip anything in naked and optnone functions.
1781    const Function *AnchorFn = IRP.getAnchorScope();
1782    if (AnchorFn && (AnchorFn->hasFnAttribute(Attribute::Naked) ||
1783                     AnchorFn->hasFnAttribute(Attribute::OptimizeNone)))
1784      return false;
1785
1786    // Avoid too many nested initializations to prevent a stack overflow.
1787    if (InitializationChainLength > MaxInitializationChainLength)
1788      return false;
1789
1790    ShouldUpdateAA = shouldUpdateAA<AAType>(IRP);
1791
1792    return !AAType::hasTrivialInitializer() || ShouldUpdateAA;
1793  }
1794
1795  /// Determine opportunities to derive 'default' attributes in \p F and create
1796  /// abstract attribute objects for them.
1797  ///
1798  /// \param F The function that is checked for attribute opportunities.
1799  ///
1800  /// Note that abstract attribute instances are generally created even if the
1801  /// IR already contains the information they would deduce. The most important
1802  /// reason for this is the single interface, the one of the abstract attribute
1803  /// instance, which can be queried without the need to look at the IR in
1804  /// various places.
1805  void identifyDefaultAbstractAttributes(Function &F);
1806
1807  /// Determine whether the function \p F is IPO amendable
1808  ///
1809  /// If a function is exactly defined or it has alwaysinline attribute
1810  /// and is viable to be inlined, we say it is IPO amendable
1811  bool isFunctionIPOAmendable(const Function &F) {
1812    return F.hasExactDefinition() || InfoCache.InlineableFunctions.count(&F) ||
1813           (Configuration.IPOAmendableCB && Configuration.IPOAmendableCB(F));
1814  }
1815
1816  /// Mark the internal function \p F as live.
1817  ///
1818  /// This will trigger the identification and initialization of attributes for
1819  /// \p F.
1820  void markLiveInternalFunction(const Function &F) {
1821    assert(F.hasLocalLinkage() &&
1822           "Only local linkage is assumed dead initially.");
1823
1824    if (Configuration.DefaultInitializeLiveInternals)
1825      identifyDefaultAbstractAttributes(const_cast<Function &>(F));
1826    if (Configuration.InitializationCallback)
1827      Configuration.InitializationCallback(*this, F);
1828  }
1829
1830  /// Helper function to remove callsite.
1831  void removeCallSite(CallInst *CI) {
1832    if (!CI)
1833      return;
1834
1835    Configuration.CGUpdater.removeCallSite(*CI);
1836  }
1837
1838  /// Record that \p U is to be replaces with \p NV after information was
1839  /// manifested. This also triggers deletion of trivially dead istructions.
1840  bool changeUseAfterManifest(Use &U, Value &NV) {
1841    Value *&V = ToBeChangedUses[&U];
1842    if (V && (V->stripPointerCasts() == NV.stripPointerCasts() ||
1843              isa_and_nonnull<UndefValue>(V)))
1844      return false;
1845    assert((!V || V == &NV || isa<UndefValue>(NV)) &&
1846           "Use was registered twice for replacement with different values!");
1847    V = &NV;
1848    return true;
1849  }
1850
1851  /// Helper function to replace all uses associated with \p IRP with \p NV.
1852  /// Return true if there is any change. The flag \p ChangeDroppable indicates
1853  /// if dropppable uses should be changed too.
1854  bool changeAfterManifest(const IRPosition IRP, Value &NV,
1855                           bool ChangeDroppable = true) {
1856    if (IRP.getPositionKind() == IRPosition::IRP_CALL_SITE_ARGUMENT) {
1857      auto *CB = cast<CallBase>(IRP.getCtxI());
1858      return changeUseAfterManifest(
1859          CB->getArgOperandUse(IRP.getCallSiteArgNo()), NV);
1860    }
1861    Value &V = IRP.getAssociatedValue();
1862    auto &Entry = ToBeChangedValues[&V];
1863    Value *CurNV = get<0>(Entry);
1864    if (CurNV && (CurNV->stripPointerCasts() == NV.stripPointerCasts() ||
1865                  isa<UndefValue>(CurNV)))
1866      return false;
1867    assert((!CurNV || CurNV == &NV || isa<UndefValue>(NV)) &&
1868           "Value replacement was registered twice with different values!");
1869    Entry = {&NV, ChangeDroppable};
1870    return true;
1871  }
1872
1873  /// Record that \p I is to be replaced with `unreachable` after information
1874  /// was manifested.
1875  void changeToUnreachableAfterManifest(Instruction *I) {
1876    ToBeChangedToUnreachableInsts.insert(I);
1877  }
1878
1879  /// Record that \p II has at least one dead successor block. This information
1880  /// is used, e.g., to replace \p II with a call, after information was
1881  /// manifested.
1882  void registerInvokeWithDeadSuccessor(InvokeInst &II) {
1883    InvokeWithDeadSuccessor.insert(&II);
1884  }
1885
1886  /// Record that \p I is deleted after information was manifested. This also
1887  /// triggers deletion of trivially dead istructions.
1888  void deleteAfterManifest(Instruction &I) { ToBeDeletedInsts.insert(&I); }
1889
1890  /// Record that \p BB is deleted after information was manifested. This also
1891  /// triggers deletion of trivially dead istructions.
1892  void deleteAfterManifest(BasicBlock &BB) { ToBeDeletedBlocks.insert(&BB); }
1893
1894  // Record that \p BB is added during the manifest of an AA. Added basic blocks
1895  // are preserved in the IR.
1896  void registerManifestAddedBasicBlock(BasicBlock &BB) {
1897    ManifestAddedBlocks.insert(&BB);
1898  }
1899
1900  /// Record that \p F is deleted after information was manifested.
1901  void deleteAfterManifest(Function &F) {
1902    if (Configuration.DeleteFns)
1903      ToBeDeletedFunctions.insert(&F);
1904  }
1905
1906  /// Return the attributes of kind \p AK existing in the IR as operand bundles
1907  /// of an llvm.assume.
1908  bool getAttrsFromAssumes(const IRPosition &IRP, Attribute::AttrKind AK,
1909                           SmallVectorImpl<Attribute> &Attrs);
1910
1911  /// Return true if any kind in \p AKs existing in the IR at a position that
1912  /// will affect this one. See also getAttrs(...).
1913  /// \param IgnoreSubsumingPositions Flag to determine if subsuming positions,
1914  ///                                 e.g., the function position if this is an
1915  ///                                 argument position, should be ignored.
1916  bool hasAttr(const IRPosition &IRP, ArrayRef<Attribute::AttrKind> AKs,
1917               bool IgnoreSubsumingPositions = false,
1918               Attribute::AttrKind ImpliedAttributeKind = Attribute::None);
1919
1920  /// Return the attributes of any kind in \p AKs existing in the IR at a
1921  /// position that will affect this one. While each position can only have a
1922  /// single attribute of any kind in \p AKs, there are "subsuming" positions
1923  /// that could have an attribute as well. This method returns all attributes
1924  /// found in \p Attrs.
1925  /// \param IgnoreSubsumingPositions Flag to determine if subsuming positions,
1926  ///                                 e.g., the function position if this is an
1927  ///                                 argument position, should be ignored.
1928  void getAttrs(const IRPosition &IRP, ArrayRef<Attribute::AttrKind> AKs,
1929                SmallVectorImpl<Attribute> &Attrs,
1930                bool IgnoreSubsumingPositions = false);
1931
1932  /// Remove all \p AttrKinds attached to \p IRP.
1933  ChangeStatus removeAttrs(const IRPosition &IRP,
1934                           ArrayRef<Attribute::AttrKind> AttrKinds);
1935  ChangeStatus removeAttrs(const IRPosition &IRP, ArrayRef<StringRef> Attrs);
1936
1937  /// Attach \p DeducedAttrs to \p IRP, if \p ForceReplace is set we do this
1938  /// even if the same attribute kind was already present.
1939  ChangeStatus manifestAttrs(const IRPosition &IRP,
1940                             ArrayRef<Attribute> DeducedAttrs,
1941                             bool ForceReplace = false);
1942
1943private:
1944  /// Helper to check \p Attrs for \p AK, if not found, check if \p
1945  /// AAType::isImpliedByIR is true, and if not, create AAType for \p IRP.
1946  template <Attribute::AttrKind AK, typename AAType>
1947  void checkAndQueryIRAttr(const IRPosition &IRP, AttributeSet Attrs);
1948
1949  /// Helper to apply \p CB on all attributes of type \p AttrDescs of \p IRP.
1950  template <typename DescTy>
1951  ChangeStatus updateAttrMap(const IRPosition &IRP, ArrayRef<DescTy> AttrDescs,
1952                             function_ref<bool(const DescTy &, AttributeSet,
1953                                               AttributeMask &, AttrBuilder &)>
1954                                 CB);
1955
1956  /// Mapping from functions/call sites to their attributes.
1957  DenseMap<Value *, AttributeList> AttrsMap;
1958
1959public:
1960  /// If \p IRP is assumed to be a constant, return it, if it is unclear yet,
1961  /// return std::nullopt, otherwise return `nullptr`.
1962  std::optional<Constant *> getAssumedConstant(const IRPosition &IRP,
1963                                               const AbstractAttribute &AA,
1964                                               bool &UsedAssumedInformation);
1965  std::optional<Constant *> getAssumedConstant(const Value &V,
1966                                               const AbstractAttribute &AA,
1967                                               bool &UsedAssumedInformation) {
1968    return getAssumedConstant(IRPosition::value(V), AA, UsedAssumedInformation);
1969  }
1970
1971  /// If \p V is assumed simplified, return it, if it is unclear yet,
1972  /// return std::nullopt, otherwise return `nullptr`.
1973  std::optional<Value *> getAssumedSimplified(const IRPosition &IRP,
1974                                              const AbstractAttribute &AA,
1975                                              bool &UsedAssumedInformation,
1976                                              AA::ValueScope S) {
1977    return getAssumedSimplified(IRP, &AA, UsedAssumedInformation, S);
1978  }
1979  std::optional<Value *> getAssumedSimplified(const Value &V,
1980                                              const AbstractAttribute &AA,
1981                                              bool &UsedAssumedInformation,
1982                                              AA::ValueScope S) {
1983    return getAssumedSimplified(IRPosition::value(V), AA,
1984                                UsedAssumedInformation, S);
1985  }
1986
1987  /// If \p V is assumed simplified, return it, if it is unclear yet,
1988  /// return std::nullopt, otherwise return `nullptr`. Same as the public
1989  /// version except that it can be used without recording dependences on any \p
1990  /// AA.
1991  std::optional<Value *> getAssumedSimplified(const IRPosition &V,
1992                                              const AbstractAttribute *AA,
1993                                              bool &UsedAssumedInformation,
1994                                              AA::ValueScope S);
1995
1996  /// Try to simplify \p IRP and in the scope \p S. If successful, true is
1997  /// returned and all potential values \p IRP can take are put into \p Values.
1998  /// If the result in \p Values contains select or PHI instructions it means
1999  /// those could not be simplified to a single value. Recursive calls with
2000  /// these instructions will yield their respective potential values. If false
2001  /// is returned no other information is valid.
2002  bool getAssumedSimplifiedValues(const IRPosition &IRP,
2003                                  const AbstractAttribute *AA,
2004                                  SmallVectorImpl<AA::ValueAndContext> &Values,
2005                                  AA::ValueScope S,
2006                                  bool &UsedAssumedInformation,
2007                                  bool RecurseForSelectAndPHI = true);
2008
2009  /// Register \p CB as a simplification callback.
2010  /// `Attributor::getAssumedSimplified` will use these callbacks before
2011  /// we it will ask `AAValueSimplify`. It is important to ensure this
2012  /// is called before `identifyDefaultAbstractAttributes`, assuming the
2013  /// latter is called at all.
2014  using SimplifictionCallbackTy = std::function<std::optional<Value *>(
2015      const IRPosition &, const AbstractAttribute *, bool &)>;
2016  void registerSimplificationCallback(const IRPosition &IRP,
2017                                      const SimplifictionCallbackTy &CB) {
2018    SimplificationCallbacks[IRP].emplace_back(CB);
2019  }
2020
2021  /// Return true if there is a simplification callback for \p IRP.
2022  bool hasSimplificationCallback(const IRPosition &IRP) {
2023    return SimplificationCallbacks.count(IRP);
2024  }
2025
2026  /// Register \p CB as a simplification callback.
2027  /// Similar to \p registerSimplificationCallback, the call back will be called
2028  /// first when we simplify a global variable \p GV.
2029  using GlobalVariableSimplifictionCallbackTy =
2030      std::function<std::optional<Constant *>(
2031          const GlobalVariable &, const AbstractAttribute *, bool &)>;
2032  void registerGlobalVariableSimplificationCallback(
2033      const GlobalVariable &GV,
2034      const GlobalVariableSimplifictionCallbackTy &CB) {
2035    GlobalVariableSimplificationCallbacks[&GV].emplace_back(CB);
2036  }
2037
2038  /// Return true if there is a simplification callback for \p GV.
2039  bool hasGlobalVariableSimplificationCallback(const GlobalVariable &GV) {
2040    return GlobalVariableSimplificationCallbacks.count(&GV);
2041  }
2042
2043  /// Return \p std::nullopt if there is no call back registered for \p GV or
2044  /// the call back is still not sure if \p GV can be simplified. Return \p
2045  /// nullptr if \p GV can't be simplified.
2046  std::optional<Constant *>
2047  getAssumedInitializerFromCallBack(const GlobalVariable &GV,
2048                                    const AbstractAttribute *AA,
2049                                    bool &UsedAssumedInformation) {
2050    assert(GlobalVariableSimplificationCallbacks.contains(&GV));
2051    for (auto &CB : GlobalVariableSimplificationCallbacks.lookup(&GV)) {
2052      auto SimplifiedGV = CB(GV, AA, UsedAssumedInformation);
2053      // For now we assume the call back will not return a std::nullopt.
2054      assert(SimplifiedGV.has_value() && "SimplifiedGV has not value");
2055      return *SimplifiedGV;
2056    }
2057    llvm_unreachable("there must be a callback registered");
2058  }
2059
2060  using VirtualUseCallbackTy =
2061      std::function<bool(Attributor &, const AbstractAttribute *)>;
2062  void registerVirtualUseCallback(const Value &V,
2063                                  const VirtualUseCallbackTy &CB) {
2064    VirtualUseCallbacks[&V].emplace_back(CB);
2065  }
2066
2067private:
2068  /// The vector with all simplification callbacks registered by outside AAs.
2069  DenseMap<IRPosition, SmallVector<SimplifictionCallbackTy, 1>>
2070      SimplificationCallbacks;
2071
2072  /// The vector with all simplification callbacks for global variables
2073  /// registered by outside AAs.
2074  DenseMap<const GlobalVariable *,
2075           SmallVector<GlobalVariableSimplifictionCallbackTy, 1>>
2076      GlobalVariableSimplificationCallbacks;
2077
2078  DenseMap<const Value *, SmallVector<VirtualUseCallbackTy, 1>>
2079      VirtualUseCallbacks;
2080
2081public:
2082  /// Translate \p V from the callee context into the call site context.
2083  std::optional<Value *>
2084  translateArgumentToCallSiteContent(std::optional<Value *> V, CallBase &CB,
2085                                     const AbstractAttribute &AA,
2086                                     bool &UsedAssumedInformation);
2087
2088  /// Return true if \p AA (or its context instruction) is assumed dead.
2089  ///
2090  /// If \p LivenessAA is not provided it is queried.
2091  bool isAssumedDead(const AbstractAttribute &AA, const AAIsDead *LivenessAA,
2092                     bool &UsedAssumedInformation,
2093                     bool CheckBBLivenessOnly = false,
2094                     DepClassTy DepClass = DepClassTy::OPTIONAL);
2095
2096  /// Return true if \p I is assumed dead.
2097  ///
2098  /// If \p LivenessAA is not provided it is queried.
2099  bool isAssumedDead(const Instruction &I, const AbstractAttribute *QueryingAA,
2100                     const AAIsDead *LivenessAA, bool &UsedAssumedInformation,
2101                     bool CheckBBLivenessOnly = false,
2102                     DepClassTy DepClass = DepClassTy::OPTIONAL,
2103                     bool CheckForDeadStore = false);
2104
2105  /// Return true if \p U is assumed dead.
2106  ///
2107  /// If \p FnLivenessAA is not provided it is queried.
2108  bool isAssumedDead(const Use &U, const AbstractAttribute *QueryingAA,
2109                     const AAIsDead *FnLivenessAA, bool &UsedAssumedInformation,
2110                     bool CheckBBLivenessOnly = false,
2111                     DepClassTy DepClass = DepClassTy::OPTIONAL);
2112
2113  /// Return true if \p IRP is assumed dead.
2114  ///
2115  /// If \p FnLivenessAA is not provided it is queried.
2116  bool isAssumedDead(const IRPosition &IRP, const AbstractAttribute *QueryingAA,
2117                     const AAIsDead *FnLivenessAA, bool &UsedAssumedInformation,
2118                     bool CheckBBLivenessOnly = false,
2119                     DepClassTy DepClass = DepClassTy::OPTIONAL);
2120
2121  /// Return true if \p BB is assumed dead.
2122  ///
2123  /// If \p LivenessAA is not provided it is queried.
2124  bool isAssumedDead(const BasicBlock &BB, const AbstractAttribute *QueryingAA,
2125                     const AAIsDead *FnLivenessAA,
2126                     DepClassTy DepClass = DepClassTy::OPTIONAL);
2127
2128  /// Check \p Pred on all potential Callees of \p CB.
2129  ///
2130  /// This method will evaluate \p Pred with all potential callees of \p CB as
2131  /// input and return true if \p Pred does. If some callees might be unknown
2132  /// this function will return false.
2133  bool checkForAllCallees(
2134      function_ref<bool(ArrayRef<const Function *> Callees)> Pred,
2135      const AbstractAttribute &QueryingAA, const CallBase &CB);
2136
2137  /// Check \p Pred on all (transitive) uses of \p V.
2138  ///
2139  /// This method will evaluate \p Pred on all (transitive) uses of the
2140  /// associated value and return true if \p Pred holds every time.
2141  /// If uses are skipped in favor of equivalent ones, e.g., if we look through
2142  /// memory, the \p EquivalentUseCB will be used to give the caller an idea
2143  /// what original used was replaced by a new one (or new ones). The visit is
2144  /// cut short if \p EquivalentUseCB returns false and the function will return
2145  /// false as well.
2146  bool checkForAllUses(function_ref<bool(const Use &, bool &)> Pred,
2147                       const AbstractAttribute &QueryingAA, const Value &V,
2148                       bool CheckBBLivenessOnly = false,
2149                       DepClassTy LivenessDepClass = DepClassTy::OPTIONAL,
2150                       bool IgnoreDroppableUses = true,
2151                       function_ref<bool(const Use &OldU, const Use &NewU)>
2152                           EquivalentUseCB = nullptr);
2153
2154  /// Emit a remark generically.
2155  ///
2156  /// This template function can be used to generically emit a remark. The
2157  /// RemarkKind should be one of the following:
2158  ///   - OptimizationRemark to indicate a successful optimization attempt
2159  ///   - OptimizationRemarkMissed to report a failed optimization attempt
2160  ///   - OptimizationRemarkAnalysis to provide additional information about an
2161  ///     optimization attempt
2162  ///
2163  /// The remark is built using a callback function \p RemarkCB that takes a
2164  /// RemarkKind as input and returns a RemarkKind.
2165  template <typename RemarkKind, typename RemarkCallBack>
2166  void emitRemark(Instruction *I, StringRef RemarkName,
2167                  RemarkCallBack &&RemarkCB) const {
2168    if (!Configuration.OREGetter)
2169      return;
2170
2171    Function *F = I->getFunction();
2172    auto &ORE = Configuration.OREGetter(F);
2173
2174    if (RemarkName.starts_with("OMP"))
2175      ORE.emit([&]() {
2176        return RemarkCB(RemarkKind(Configuration.PassName, RemarkName, I))
2177               << " [" << RemarkName << "]";
2178      });
2179    else
2180      ORE.emit([&]() {
2181        return RemarkCB(RemarkKind(Configuration.PassName, RemarkName, I));
2182      });
2183  }
2184
2185  /// Emit a remark on a function.
2186  template <typename RemarkKind, typename RemarkCallBack>
2187  void emitRemark(Function *F, StringRef RemarkName,
2188                  RemarkCallBack &&RemarkCB) const {
2189    if (!Configuration.OREGetter)
2190      return;
2191
2192    auto &ORE = Configuration.OREGetter(F);
2193
2194    if (RemarkName.starts_with("OMP"))
2195      ORE.emit([&]() {
2196        return RemarkCB(RemarkKind(Configuration.PassName, RemarkName, F))
2197               << " [" << RemarkName << "]";
2198      });
2199    else
2200      ORE.emit([&]() {
2201        return RemarkCB(RemarkKind(Configuration.PassName, RemarkName, F));
2202      });
2203  }
2204
2205  /// Helper struct used in the communication between an abstract attribute (AA)
2206  /// that wants to change the signature of a function and the Attributor which
2207  /// applies the changes. The struct is partially initialized with the
2208  /// information from the AA (see the constructor). All other members are
2209  /// provided by the Attributor prior to invoking any callbacks.
2210  struct ArgumentReplacementInfo {
2211    /// Callee repair callback type
2212    ///
2213    /// The function repair callback is invoked once to rewire the replacement
2214    /// arguments in the body of the new function. The argument replacement info
2215    /// is passed, as build from the registerFunctionSignatureRewrite call, as
2216    /// well as the replacement function and an iteratore to the first
2217    /// replacement argument.
2218    using CalleeRepairCBTy = std::function<void(
2219        const ArgumentReplacementInfo &, Function &, Function::arg_iterator)>;
2220
2221    /// Abstract call site (ACS) repair callback type
2222    ///
2223    /// The abstract call site repair callback is invoked once on every abstract
2224    /// call site of the replaced function (\see ReplacedFn). The callback needs
2225    /// to provide the operands for the call to the new replacement function.
2226    /// The number and type of the operands appended to the provided vector
2227    /// (second argument) is defined by the number and types determined through
2228    /// the replacement type vector (\see ReplacementTypes). The first argument
2229    /// is the ArgumentReplacementInfo object registered with the Attributor
2230    /// through the registerFunctionSignatureRewrite call.
2231    using ACSRepairCBTy =
2232        std::function<void(const ArgumentReplacementInfo &, AbstractCallSite,
2233                           SmallVectorImpl<Value *> &)>;
2234
2235    /// Simple getters, see the corresponding members for details.
2236    ///{
2237
2238    Attributor &getAttributor() const { return A; }
2239    const Function &getReplacedFn() const { return ReplacedFn; }
2240    const Argument &getReplacedArg() const { return ReplacedArg; }
2241    unsigned getNumReplacementArgs() const { return ReplacementTypes.size(); }
2242    const SmallVectorImpl<Type *> &getReplacementTypes() const {
2243      return ReplacementTypes;
2244    }
2245
2246    ///}
2247
2248  private:
2249    /// Constructor that takes the argument to be replaced, the types of
2250    /// the replacement arguments, as well as callbacks to repair the call sites
2251    /// and new function after the replacement happened.
2252    ArgumentReplacementInfo(Attributor &A, Argument &Arg,
2253                            ArrayRef<Type *> ReplacementTypes,
2254                            CalleeRepairCBTy &&CalleeRepairCB,
2255                            ACSRepairCBTy &&ACSRepairCB)
2256        : A(A), ReplacedFn(*Arg.getParent()), ReplacedArg(Arg),
2257          ReplacementTypes(ReplacementTypes.begin(), ReplacementTypes.end()),
2258          CalleeRepairCB(std::move(CalleeRepairCB)),
2259          ACSRepairCB(std::move(ACSRepairCB)) {}
2260
2261    /// Reference to the attributor to allow access from the callbacks.
2262    Attributor &A;
2263
2264    /// The "old" function replaced by ReplacementFn.
2265    const Function &ReplacedFn;
2266
2267    /// The "old" argument replaced by new ones defined via ReplacementTypes.
2268    const Argument &ReplacedArg;
2269
2270    /// The types of the arguments replacing ReplacedArg.
2271    const SmallVector<Type *, 8> ReplacementTypes;
2272
2273    /// Callee repair callback, see CalleeRepairCBTy.
2274    const CalleeRepairCBTy CalleeRepairCB;
2275
2276    /// Abstract call site (ACS) repair callback, see ACSRepairCBTy.
2277    const ACSRepairCBTy ACSRepairCB;
2278
2279    /// Allow access to the private members from the Attributor.
2280    friend struct Attributor;
2281  };
2282
2283  /// Check if we can rewrite a function signature.
2284  ///
2285  /// The argument \p Arg is replaced with new ones defined by the number,
2286  /// order, and types in \p ReplacementTypes.
2287  ///
2288  /// \returns True, if the replacement can be registered, via
2289  /// registerFunctionSignatureRewrite, false otherwise.
2290  bool isValidFunctionSignatureRewrite(Argument &Arg,
2291                                       ArrayRef<Type *> ReplacementTypes);
2292
2293  /// Register a rewrite for a function signature.
2294  ///
2295  /// The argument \p Arg is replaced with new ones defined by the number,
2296  /// order, and types in \p ReplacementTypes. The rewiring at the call sites is
2297  /// done through \p ACSRepairCB and at the callee site through
2298  /// \p CalleeRepairCB.
2299  ///
2300  /// \returns True, if the replacement was registered, false otherwise.
2301  bool registerFunctionSignatureRewrite(
2302      Argument &Arg, ArrayRef<Type *> ReplacementTypes,
2303      ArgumentReplacementInfo::CalleeRepairCBTy &&CalleeRepairCB,
2304      ArgumentReplacementInfo::ACSRepairCBTy &&ACSRepairCB);
2305
2306  /// Check \p Pred on all function call sites.
2307  ///
2308  /// This method will evaluate \p Pred on call sites and return
2309  /// true if \p Pred holds in every call sites. However, this is only possible
2310  /// all call sites are known, hence the function has internal linkage.
2311  /// If true is returned, \p UsedAssumedInformation is set if assumed
2312  /// information was used to skip or simplify potential call sites.
2313  bool checkForAllCallSites(function_ref<bool(AbstractCallSite)> Pred,
2314                            const AbstractAttribute &QueryingAA,
2315                            bool RequireAllCallSites,
2316                            bool &UsedAssumedInformation);
2317
2318  /// Check \p Pred on all call sites of \p Fn.
2319  ///
2320  /// This method will evaluate \p Pred on call sites and return
2321  /// true if \p Pred holds in every call sites. However, this is only possible
2322  /// all call sites are known, hence the function has internal linkage.
2323  /// If true is returned, \p UsedAssumedInformation is set if assumed
2324  /// information was used to skip or simplify potential call sites.
2325  bool checkForAllCallSites(function_ref<bool(AbstractCallSite)> Pred,
2326                            const Function &Fn, bool RequireAllCallSites,
2327                            const AbstractAttribute *QueryingAA,
2328                            bool &UsedAssumedInformation,
2329                            bool CheckPotentiallyDead = false);
2330
2331  /// Check \p Pred on all values potentially returned by the function
2332  /// associated with \p QueryingAA.
2333  ///
2334  /// This is the context insensitive version of the method above.
2335  bool
2336  checkForAllReturnedValues(function_ref<bool(Value &)> Pred,
2337                            const AbstractAttribute &QueryingAA,
2338                            AA::ValueScope S = AA::ValueScope::Intraprocedural,
2339                            bool RecurseForSelectAndPHI = true);
2340
2341  /// Check \p Pred on all instructions in \p Fn with an opcode present in
2342  /// \p Opcodes.
2343  ///
2344  /// This method will evaluate \p Pred on all instructions with an opcode
2345  /// present in \p Opcode and return true if \p Pred holds on all of them.
2346  bool checkForAllInstructions(function_ref<bool(Instruction &)> Pred,
2347                               const Function *Fn,
2348                               const AbstractAttribute *QueryingAA,
2349                               ArrayRef<unsigned> Opcodes,
2350                               bool &UsedAssumedInformation,
2351                               bool CheckBBLivenessOnly = false,
2352                               bool CheckPotentiallyDead = false);
2353
2354  /// Check \p Pred on all instructions with an opcode present in \p Opcodes.
2355  ///
2356  /// This method will evaluate \p Pred on all instructions with an opcode
2357  /// present in \p Opcode and return true if \p Pred holds on all of them.
2358  bool checkForAllInstructions(function_ref<bool(Instruction &)> Pred,
2359                               const AbstractAttribute &QueryingAA,
2360                               ArrayRef<unsigned> Opcodes,
2361                               bool &UsedAssumedInformation,
2362                               bool CheckBBLivenessOnly = false,
2363                               bool CheckPotentiallyDead = false);
2364
2365  /// Check \p Pred on all call-like instructions (=CallBased derived).
2366  ///
2367  /// See checkForAllCallLikeInstructions(...) for more information.
2368  bool checkForAllCallLikeInstructions(function_ref<bool(Instruction &)> Pred,
2369                                       const AbstractAttribute &QueryingAA,
2370                                       bool &UsedAssumedInformation,
2371                                       bool CheckBBLivenessOnly = false,
2372                                       bool CheckPotentiallyDead = false) {
2373    return checkForAllInstructions(
2374        Pred, QueryingAA,
2375        {(unsigned)Instruction::Invoke, (unsigned)Instruction::CallBr,
2376         (unsigned)Instruction::Call},
2377        UsedAssumedInformation, CheckBBLivenessOnly, CheckPotentiallyDead);
2378  }
2379
2380  /// Check \p Pred on all Read/Write instructions.
2381  ///
2382  /// This method will evaluate \p Pred on all instructions that read or write
2383  /// to memory present in the information cache and return true if \p Pred
2384  /// holds on all of them.
2385  bool checkForAllReadWriteInstructions(function_ref<bool(Instruction &)> Pred,
2386                                        AbstractAttribute &QueryingAA,
2387                                        bool &UsedAssumedInformation);
2388
2389  /// Create a shallow wrapper for \p F such that \p F has internal linkage
2390  /// afterwards. It also sets the original \p F 's name to anonymous
2391  ///
2392  /// A wrapper is a function with the same type (and attributes) as \p F
2393  /// that will only call \p F and return the result, if any.
2394  ///
2395  /// Assuming the declaration of looks like:
2396  ///   rty F(aty0 arg0, ..., atyN argN);
2397  ///
2398  /// The wrapper will then look as follows:
2399  ///   rty wrapper(aty0 arg0, ..., atyN argN) {
2400  ///     return F(arg0, ..., argN);
2401  ///   }
2402  ///
2403  static void createShallowWrapper(Function &F);
2404
2405  /// Returns true if the function \p F can be internalized. i.e. it has a
2406  /// compatible linkage.
2407  static bool isInternalizable(Function &F);
2408
2409  /// Make another copy of the function \p F such that the copied version has
2410  /// internal linkage afterwards and can be analysed. Then we replace all uses
2411  /// of the original function to the copied one
2412  ///
2413  /// Only non-locally linked functions that have `linkonce_odr` or `weak_odr`
2414  /// linkage can be internalized because these linkages guarantee that other
2415  /// definitions with the same name have the same semantics as this one.
2416  ///
2417  /// This will only be run if the `attributor-allow-deep-wrappers` option is
2418  /// set, or if the function is called with \p Force set to true.
2419  ///
2420  /// If the function \p F failed to be internalized the return value will be a
2421  /// null pointer.
2422  static Function *internalizeFunction(Function &F, bool Force = false);
2423
2424  /// Make copies of each function in the set \p FnSet such that the copied
2425  /// version has internal linkage afterwards and can be analysed. Then we
2426  /// replace all uses of the original function to the copied one. The map
2427  /// \p FnMap contains a mapping of functions to their internalized versions.
2428  ///
2429  /// Only non-locally linked functions that have `linkonce_odr` or `weak_odr`
2430  /// linkage can be internalized because these linkages guarantee that other
2431  /// definitions with the same name have the same semantics as this one.
2432  ///
2433  /// This version will internalize all the functions in the set \p FnSet at
2434  /// once and then replace the uses. This prevents internalized functions being
2435  /// called by external functions when there is an internalized version in the
2436  /// module.
2437  static bool internalizeFunctions(SmallPtrSetImpl<Function *> &FnSet,
2438                                   DenseMap<Function *, Function *> &FnMap);
2439
2440  /// Return the data layout associated with the anchor scope.
2441  const DataLayout &getDataLayout() const { return InfoCache.DL; }
2442
2443  /// The allocator used to allocate memory, e.g. for `AbstractAttribute`s.
2444  BumpPtrAllocator &Allocator;
2445
2446  const SmallSetVector<Function *, 8> &getModifiedFunctions() {
2447    return CGModifiedFunctions;
2448  }
2449
2450private:
2451  /// This method will do fixpoint iteration until fixpoint or the
2452  /// maximum iteration count is reached.
2453  ///
2454  /// If the maximum iteration count is reached, This method will
2455  /// indicate pessimistic fixpoint on attributes that transitively depend
2456  /// on attributes that were scheduled for an update.
2457  void runTillFixpoint();
2458
2459  /// Gets called after scheduling, manifests attributes to the LLVM IR.
2460  ChangeStatus manifestAttributes();
2461
2462  /// Gets called after attributes have been manifested, cleans up the IR.
2463  /// Deletes dead functions, blocks and instructions.
2464  /// Rewrites function signitures and updates the call graph.
2465  ChangeStatus cleanupIR();
2466
2467  /// Identify internal functions that are effectively dead, thus not reachable
2468  /// from a live entry point. The functions are added to ToBeDeletedFunctions.
2469  void identifyDeadInternalFunctions();
2470
2471  /// Run `::update` on \p AA and track the dependences queried while doing so.
2472  /// Also adjust the state if we know further updates are not necessary.
2473  ChangeStatus updateAA(AbstractAttribute &AA);
2474
2475  /// Remember the dependences on the top of the dependence stack such that they
2476  /// may trigger further updates. (\see DependenceStack)
2477  void rememberDependences();
2478
2479  /// Determine if CallBase context in \p IRP should be propagated.
2480  bool shouldPropagateCallBaseContext(const IRPosition &IRP);
2481
2482  /// Apply all requested function signature rewrites
2483  /// (\see registerFunctionSignatureRewrite) and return Changed if the module
2484  /// was altered.
2485  ChangeStatus
2486  rewriteFunctionSignatures(SmallSetVector<Function *, 8> &ModifiedFns);
2487
2488  /// Check if the Attribute \p AA should be seeded.
2489  /// See getOrCreateAAFor.
2490  bool shouldSeedAttribute(AbstractAttribute &AA);
2491
2492  /// A nested map to lookup abstract attributes based on the argument position
2493  /// on the outer level, and the addresses of the static member (AAType::ID) on
2494  /// the inner level.
2495  ///{
2496  using AAMapKeyTy = std::pair<const char *, IRPosition>;
2497  DenseMap<AAMapKeyTy, AbstractAttribute *> AAMap;
2498  ///}
2499
2500  /// Map to remember all requested signature changes (= argument replacements).
2501  DenseMap<Function *, SmallVector<std::unique_ptr<ArgumentReplacementInfo>, 8>>
2502      ArgumentReplacementMap;
2503
2504  /// The set of functions we are deriving attributes for.
2505  SetVector<Function *> &Functions;
2506
2507  /// The information cache that holds pre-processed (LLVM-IR) information.
2508  InformationCache &InfoCache;
2509
2510  /// Abstract Attribute dependency graph
2511  AADepGraph DG;
2512
2513  /// Set of functions for which we modified the content such that it might
2514  /// impact the call graph.
2515  SmallSetVector<Function *, 8> CGModifiedFunctions;
2516
2517  /// Information about a dependence. If FromAA is changed ToAA needs to be
2518  /// updated as well.
2519  struct DepInfo {
2520    const AbstractAttribute *FromAA;
2521    const AbstractAttribute *ToAA;
2522    DepClassTy DepClass;
2523  };
2524
2525  /// The dependence stack is used to track dependences during an
2526  /// `AbstractAttribute::update` call. As `AbstractAttribute::update` can be
2527  /// recursive we might have multiple vectors of dependences in here. The stack
2528  /// size, should be adjusted according to the expected recursion depth and the
2529  /// inner dependence vector size to the expected number of dependences per
2530  /// abstract attribute. Since the inner vectors are actually allocated on the
2531  /// stack we can be generous with their size.
2532  using DependenceVector = SmallVector<DepInfo, 8>;
2533  SmallVector<DependenceVector *, 16> DependenceStack;
2534
2535  /// A set to remember the functions we already assume to be live and visited.
2536  DenseSet<const Function *> VisitedFunctions;
2537
2538  /// Uses we replace with a new value after manifest is done. We will remove
2539  /// then trivially dead instructions as well.
2540  SmallMapVector<Use *, Value *, 32> ToBeChangedUses;
2541
2542  /// Values we replace with a new value after manifest is done. We will remove
2543  /// then trivially dead instructions as well.
2544  SmallMapVector<Value *, PointerIntPair<Value *, 1, bool>, 32>
2545      ToBeChangedValues;
2546
2547  /// Instructions we replace with `unreachable` insts after manifest is done.
2548  SmallSetVector<WeakVH, 16> ToBeChangedToUnreachableInsts;
2549
2550  /// Invoke instructions with at least a single dead successor block.
2551  SmallSetVector<WeakVH, 16> InvokeWithDeadSuccessor;
2552
2553  /// A flag that indicates which stage of the process we are in. Initially, the
2554  /// phase is SEEDING. Phase is changed in `Attributor::run()`
2555  enum class AttributorPhase {
2556    SEEDING,
2557    UPDATE,
2558    MANIFEST,
2559    CLEANUP,
2560  } Phase = AttributorPhase::SEEDING;
2561
2562  /// The current initialization chain length. Tracked to avoid stack overflows.
2563  unsigned InitializationChainLength = 0;
2564
2565  /// Functions, blocks, and instructions we delete after manifest is done.
2566  ///
2567  ///{
2568  SmallPtrSet<BasicBlock *, 8> ManifestAddedBlocks;
2569  SmallSetVector<Function *, 8> ToBeDeletedFunctions;
2570  SmallSetVector<BasicBlock *, 8> ToBeDeletedBlocks;
2571  SmallSetVector<WeakVH, 8> ToBeDeletedInsts;
2572  ///}
2573
2574  /// Container with all the query AAs that requested an update via
2575  /// registerForUpdate.
2576  SmallSetVector<AbstractAttribute *, 16> QueryAAsAwaitingUpdate;
2577
2578  /// User provided configuration for this Attributor instance.
2579  const AttributorConfig Configuration;
2580
2581  friend AADepGraph;
2582  friend AttributorCallGraph;
2583};
2584
2585/// An interface to query the internal state of an abstract attribute.
2586///
2587/// The abstract state is a minimal interface that allows the Attributor to
2588/// communicate with the abstract attributes about their internal state without
2589/// enforcing or exposing implementation details, e.g., the (existence of an)
2590/// underlying lattice.
2591///
2592/// It is sufficient to be able to query if a state is (1) valid or invalid, (2)
2593/// at a fixpoint, and to indicate to the state that (3) an optimistic fixpoint
2594/// was reached or (4) a pessimistic fixpoint was enforced.
2595///
2596/// All methods need to be implemented by the subclass. For the common use case,
2597/// a single boolean state or a bit-encoded state, the BooleanState and
2598/// {Inc,Dec,Bit}IntegerState classes are already provided. An abstract
2599/// attribute can inherit from them to get the abstract state interface and
2600/// additional methods to directly modify the state based if needed. See the
2601/// class comments for help.
2602struct AbstractState {
2603  virtual ~AbstractState() = default;
2604
2605  /// Return if this abstract state is in a valid state. If false, no
2606  /// information provided should be used.
2607  virtual bool isValidState() const = 0;
2608
2609  /// Return if this abstract state is fixed, thus does not need to be updated
2610  /// if information changes as it cannot change itself.
2611  virtual bool isAtFixpoint() const = 0;
2612
2613  /// Indicate that the abstract state should converge to the optimistic state.
2614  ///
2615  /// This will usually make the optimistically assumed state the known to be
2616  /// true state.
2617  ///
2618  /// \returns ChangeStatus::UNCHANGED as the assumed value should not change.
2619  virtual ChangeStatus indicateOptimisticFixpoint() = 0;
2620
2621  /// Indicate that the abstract state should converge to the pessimistic state.
2622  ///
2623  /// This will usually revert the optimistically assumed state to the known to
2624  /// be true state.
2625  ///
2626  /// \returns ChangeStatus::CHANGED as the assumed value may change.
2627  virtual ChangeStatus indicatePessimisticFixpoint() = 0;
2628};
2629
2630/// Simple state with integers encoding.
2631///
2632/// The interface ensures that the assumed bits are always a subset of the known
2633/// bits. Users can only add known bits and, except through adding known bits,
2634/// they can only remove assumed bits. This should guarantee monotonicity and
2635/// thereby the existence of a fixpoint (if used correctly). The fixpoint is
2636/// reached when the assumed and known state/bits are equal. Users can
2637/// force/inidicate a fixpoint. If an optimistic one is indicated, the known
2638/// state will catch up with the assumed one, for a pessimistic fixpoint it is
2639/// the other way around.
2640template <typename base_ty, base_ty BestState, base_ty WorstState>
2641struct IntegerStateBase : public AbstractState {
2642  using base_t = base_ty;
2643
2644  IntegerStateBase() = default;
2645  IntegerStateBase(base_t Assumed) : Assumed(Assumed) {}
2646
2647  /// Return the best possible representable state.
2648  static constexpr base_t getBestState() { return BestState; }
2649  static constexpr base_t getBestState(const IntegerStateBase &) {
2650    return getBestState();
2651  }
2652
2653  /// Return the worst possible representable state.
2654  static constexpr base_t getWorstState() { return WorstState; }
2655  static constexpr base_t getWorstState(const IntegerStateBase &) {
2656    return getWorstState();
2657  }
2658
2659  /// See AbstractState::isValidState()
2660  /// NOTE: For now we simply pretend that the worst possible state is invalid.
2661  bool isValidState() const override { return Assumed != getWorstState(); }
2662
2663  /// See AbstractState::isAtFixpoint()
2664  bool isAtFixpoint() const override { return Assumed == Known; }
2665
2666  /// See AbstractState::indicateOptimisticFixpoint(...)
2667  ChangeStatus indicateOptimisticFixpoint() override {
2668    Known = Assumed;
2669    return ChangeStatus::UNCHANGED;
2670  }
2671
2672  /// See AbstractState::indicatePessimisticFixpoint(...)
2673  ChangeStatus indicatePessimisticFixpoint() override {
2674    Assumed = Known;
2675    return ChangeStatus::CHANGED;
2676  }
2677
2678  /// Return the known state encoding
2679  base_t getKnown() const { return Known; }
2680
2681  /// Return the assumed state encoding.
2682  base_t getAssumed() const { return Assumed; }
2683
2684  /// Equality for IntegerStateBase.
2685  bool
2686  operator==(const IntegerStateBase<base_t, BestState, WorstState> &R) const {
2687    return this->getAssumed() == R.getAssumed() &&
2688           this->getKnown() == R.getKnown();
2689  }
2690
2691  /// Inequality for IntegerStateBase.
2692  bool
2693  operator!=(const IntegerStateBase<base_t, BestState, WorstState> &R) const {
2694    return !(*this == R);
2695  }
2696
2697  /// "Clamp" this state with \p R. The result is subtype dependent but it is
2698  /// intended that only information assumed in both states will be assumed in
2699  /// this one afterwards.
2700  void operator^=(const IntegerStateBase<base_t, BestState, WorstState> &R) {
2701    handleNewAssumedValue(R.getAssumed());
2702  }
2703
2704  /// "Clamp" this state with \p R. The result is subtype dependent but it is
2705  /// intended that information known in either state will be known in
2706  /// this one afterwards.
2707  void operator+=(const IntegerStateBase<base_t, BestState, WorstState> &R) {
2708    handleNewKnownValue(R.getKnown());
2709  }
2710
2711  void operator|=(const IntegerStateBase<base_t, BestState, WorstState> &R) {
2712    joinOR(R.getAssumed(), R.getKnown());
2713  }
2714
2715  void operator&=(const IntegerStateBase<base_t, BestState, WorstState> &R) {
2716    joinAND(R.getAssumed(), R.getKnown());
2717  }
2718
2719protected:
2720  /// Handle a new assumed value \p Value. Subtype dependent.
2721  virtual void handleNewAssumedValue(base_t Value) = 0;
2722
2723  /// Handle a new known value \p Value. Subtype dependent.
2724  virtual void handleNewKnownValue(base_t Value) = 0;
2725
2726  /// Handle a  value \p Value. Subtype dependent.
2727  virtual void joinOR(base_t AssumedValue, base_t KnownValue) = 0;
2728
2729  /// Handle a new assumed value \p Value. Subtype dependent.
2730  virtual void joinAND(base_t AssumedValue, base_t KnownValue) = 0;
2731
2732  /// The known state encoding in an integer of type base_t.
2733  base_t Known = getWorstState();
2734
2735  /// The assumed state encoding in an integer of type base_t.
2736  base_t Assumed = getBestState();
2737};
2738
2739/// Specialization of the integer state for a bit-wise encoding.
2740template <typename base_ty = uint32_t, base_ty BestState = ~base_ty(0),
2741          base_ty WorstState = 0>
2742struct BitIntegerState
2743    : public IntegerStateBase<base_ty, BestState, WorstState> {
2744  using super = IntegerStateBase<base_ty, BestState, WorstState>;
2745  using base_t = base_ty;
2746  BitIntegerState() = default;
2747  BitIntegerState(base_t Assumed) : super(Assumed) {}
2748
2749  /// Return true if the bits set in \p BitsEncoding are "known bits".
2750  bool isKnown(base_t BitsEncoding = BestState) const {
2751    return (this->Known & BitsEncoding) == BitsEncoding;
2752  }
2753
2754  /// Return true if the bits set in \p BitsEncoding are "assumed bits".
2755  bool isAssumed(base_t BitsEncoding = BestState) const {
2756    return (this->Assumed & BitsEncoding) == BitsEncoding;
2757  }
2758
2759  /// Add the bits in \p BitsEncoding to the "known bits".
2760  BitIntegerState &addKnownBits(base_t Bits) {
2761    // Make sure we never miss any "known bits".
2762    this->Assumed |= Bits;
2763    this->Known |= Bits;
2764    return *this;
2765  }
2766
2767  /// Remove the bits in \p BitsEncoding from the "assumed bits" if not known.
2768  BitIntegerState &removeAssumedBits(base_t BitsEncoding) {
2769    return intersectAssumedBits(~BitsEncoding);
2770  }
2771
2772  /// Remove the bits in \p BitsEncoding from the "known bits".
2773  BitIntegerState &removeKnownBits(base_t BitsEncoding) {
2774    this->Known = (this->Known & ~BitsEncoding);
2775    return *this;
2776  }
2777
2778  /// Keep only "assumed bits" also set in \p BitsEncoding but all known ones.
2779  BitIntegerState &intersectAssumedBits(base_t BitsEncoding) {
2780    // Make sure we never lose any "known bits".
2781    this->Assumed = (this->Assumed & BitsEncoding) | this->Known;
2782    return *this;
2783  }
2784
2785private:
2786  void handleNewAssumedValue(base_t Value) override {
2787    intersectAssumedBits(Value);
2788  }
2789  void handleNewKnownValue(base_t Value) override { addKnownBits(Value); }
2790  void joinOR(base_t AssumedValue, base_t KnownValue) override {
2791    this->Known |= KnownValue;
2792    this->Assumed |= AssumedValue;
2793  }
2794  void joinAND(base_t AssumedValue, base_t KnownValue) override {
2795    this->Known &= KnownValue;
2796    this->Assumed &= AssumedValue;
2797  }
2798};
2799
2800/// Specialization of the integer state for an increasing value, hence ~0u is
2801/// the best state and 0 the worst.
2802template <typename base_ty = uint32_t, base_ty BestState = ~base_ty(0),
2803          base_ty WorstState = 0>
2804struct IncIntegerState
2805    : public IntegerStateBase<base_ty, BestState, WorstState> {
2806  using super = IntegerStateBase<base_ty, BestState, WorstState>;
2807  using base_t = base_ty;
2808
2809  IncIntegerState() : super() {}
2810  IncIntegerState(base_t Assumed) : super(Assumed) {}
2811
2812  /// Return the best possible representable state.
2813  static constexpr base_t getBestState() { return BestState; }
2814  static constexpr base_t
2815  getBestState(const IncIntegerState<base_ty, BestState, WorstState> &) {
2816    return getBestState();
2817  }
2818
2819  /// Take minimum of assumed and \p Value.
2820  IncIntegerState &takeAssumedMinimum(base_t Value) {
2821    // Make sure we never lose "known value".
2822    this->Assumed = std::max(std::min(this->Assumed, Value), this->Known);
2823    return *this;
2824  }
2825
2826  /// Take maximum of known and \p Value.
2827  IncIntegerState &takeKnownMaximum(base_t Value) {
2828    // Make sure we never lose "known value".
2829    this->Assumed = std::max(Value, this->Assumed);
2830    this->Known = std::max(Value, this->Known);
2831    return *this;
2832  }
2833
2834private:
2835  void handleNewAssumedValue(base_t Value) override {
2836    takeAssumedMinimum(Value);
2837  }
2838  void handleNewKnownValue(base_t Value) override { takeKnownMaximum(Value); }
2839  void joinOR(base_t AssumedValue, base_t KnownValue) override {
2840    this->Known = std::max(this->Known, KnownValue);
2841    this->Assumed = std::max(this->Assumed, AssumedValue);
2842  }
2843  void joinAND(base_t AssumedValue, base_t KnownValue) override {
2844    this->Known = std::min(this->Known, KnownValue);
2845    this->Assumed = std::min(this->Assumed, AssumedValue);
2846  }
2847};
2848
2849/// Specialization of the integer state for a decreasing value, hence 0 is the
2850/// best state and ~0u the worst.
2851template <typename base_ty = uint32_t>
2852struct DecIntegerState : public IntegerStateBase<base_ty, 0, ~base_ty(0)> {
2853  using base_t = base_ty;
2854
2855  /// Take maximum of assumed and \p Value.
2856  DecIntegerState &takeAssumedMaximum(base_t Value) {
2857    // Make sure we never lose "known value".
2858    this->Assumed = std::min(std::max(this->Assumed, Value), this->Known);
2859    return *this;
2860  }
2861
2862  /// Take minimum of known and \p Value.
2863  DecIntegerState &takeKnownMinimum(base_t Value) {
2864    // Make sure we never lose "known value".
2865    this->Assumed = std::min(Value, this->Assumed);
2866    this->Known = std::min(Value, this->Known);
2867    return *this;
2868  }
2869
2870private:
2871  void handleNewAssumedValue(base_t Value) override {
2872    takeAssumedMaximum(Value);
2873  }
2874  void handleNewKnownValue(base_t Value) override { takeKnownMinimum(Value); }
2875  void joinOR(base_t AssumedValue, base_t KnownValue) override {
2876    this->Assumed = std::min(this->Assumed, KnownValue);
2877    this->Assumed = std::min(this->Assumed, AssumedValue);
2878  }
2879  void joinAND(base_t AssumedValue, base_t KnownValue) override {
2880    this->Assumed = std::max(this->Assumed, KnownValue);
2881    this->Assumed = std::max(this->Assumed, AssumedValue);
2882  }
2883};
2884
2885/// Simple wrapper for a single bit (boolean) state.
2886struct BooleanState : public IntegerStateBase<bool, true, false> {
2887  using super = IntegerStateBase<bool, true, false>;
2888  using base_t = IntegerStateBase::base_t;
2889
2890  BooleanState() = default;
2891  BooleanState(base_t Assumed) : super(Assumed) {}
2892
2893  /// Set the assumed value to \p Value but never below the known one.
2894  void setAssumed(bool Value) { Assumed &= (Known | Value); }
2895
2896  /// Set the known and asssumed value to \p Value.
2897  void setKnown(bool Value) {
2898    Known |= Value;
2899    Assumed |= Value;
2900  }
2901
2902  /// Return true if the state is assumed to hold.
2903  bool isAssumed() const { return getAssumed(); }
2904
2905  /// Return true if the state is known to hold.
2906  bool isKnown() const { return getKnown(); }
2907
2908private:
2909  void handleNewAssumedValue(base_t Value) override {
2910    if (!Value)
2911      Assumed = Known;
2912  }
2913  void handleNewKnownValue(base_t Value) override {
2914    if (Value)
2915      Known = (Assumed = Value);
2916  }
2917  void joinOR(base_t AssumedValue, base_t KnownValue) override {
2918    Known |= KnownValue;
2919    Assumed |= AssumedValue;
2920  }
2921  void joinAND(base_t AssumedValue, base_t KnownValue) override {
2922    Known &= KnownValue;
2923    Assumed &= AssumedValue;
2924  }
2925};
2926
2927/// State for an integer range.
2928struct IntegerRangeState : public AbstractState {
2929
2930  /// Bitwidth of the associated value.
2931  uint32_t BitWidth;
2932
2933  /// State representing assumed range, initially set to empty.
2934  ConstantRange Assumed;
2935
2936  /// State representing known range, initially set to [-inf, inf].
2937  ConstantRange Known;
2938
2939  IntegerRangeState(uint32_t BitWidth)
2940      : BitWidth(BitWidth), Assumed(ConstantRange::getEmpty(BitWidth)),
2941        Known(ConstantRange::getFull(BitWidth)) {}
2942
2943  IntegerRangeState(const ConstantRange &CR)
2944      : BitWidth(CR.getBitWidth()), Assumed(CR),
2945        Known(getWorstState(CR.getBitWidth())) {}
2946
2947  /// Return the worst possible representable state.
2948  static ConstantRange getWorstState(uint32_t BitWidth) {
2949    return ConstantRange::getFull(BitWidth);
2950  }
2951
2952  /// Return the best possible representable state.
2953  static ConstantRange getBestState(uint32_t BitWidth) {
2954    return ConstantRange::getEmpty(BitWidth);
2955  }
2956  static ConstantRange getBestState(const IntegerRangeState &IRS) {
2957    return getBestState(IRS.getBitWidth());
2958  }
2959
2960  /// Return associated values' bit width.
2961  uint32_t getBitWidth() const { return BitWidth; }
2962
2963  /// See AbstractState::isValidState()
2964  bool isValidState() const override {
2965    return BitWidth > 0 && !Assumed.isFullSet();
2966  }
2967
2968  /// See AbstractState::isAtFixpoint()
2969  bool isAtFixpoint() const override { return Assumed == Known; }
2970
2971  /// See AbstractState::indicateOptimisticFixpoint(...)
2972  ChangeStatus indicateOptimisticFixpoint() override {
2973    Known = Assumed;
2974    return ChangeStatus::CHANGED;
2975  }
2976
2977  /// See AbstractState::indicatePessimisticFixpoint(...)
2978  ChangeStatus indicatePessimisticFixpoint() override {
2979    Assumed = Known;
2980    return ChangeStatus::CHANGED;
2981  }
2982
2983  /// Return the known state encoding
2984  ConstantRange getKnown() const { return Known; }
2985
2986  /// Return the assumed state encoding.
2987  ConstantRange getAssumed() const { return Assumed; }
2988
2989  /// Unite assumed range with the passed state.
2990  void unionAssumed(const ConstantRange &R) {
2991    // Don't lose a known range.
2992    Assumed = Assumed.unionWith(R).intersectWith(Known);
2993  }
2994
2995  /// See IntegerRangeState::unionAssumed(..).
2996  void unionAssumed(const IntegerRangeState &R) {
2997    unionAssumed(R.getAssumed());
2998  }
2999
3000  /// Intersect known range with the passed state.
3001  void intersectKnown(const ConstantRange &R) {
3002    Assumed = Assumed.intersectWith(R);
3003    Known = Known.intersectWith(R);
3004  }
3005
3006  /// See IntegerRangeState::intersectKnown(..).
3007  void intersectKnown(const IntegerRangeState &R) {
3008    intersectKnown(R.getKnown());
3009  }
3010
3011  /// Equality for IntegerRangeState.
3012  bool operator==(const IntegerRangeState &R) const {
3013    return getAssumed() == R.getAssumed() && getKnown() == R.getKnown();
3014  }
3015
3016  /// "Clamp" this state with \p R. The result is subtype dependent but it is
3017  /// intended that only information assumed in both states will be assumed in
3018  /// this one afterwards.
3019  IntegerRangeState operator^=(const IntegerRangeState &R) {
3020    // NOTE: `^=` operator seems like `intersect` but in this case, we need to
3021    // take `union`.
3022    unionAssumed(R);
3023    return *this;
3024  }
3025
3026  IntegerRangeState operator&=(const IntegerRangeState &R) {
3027    // NOTE: `&=` operator seems like `intersect` but in this case, we need to
3028    // take `union`.
3029    Known = Known.unionWith(R.getKnown());
3030    Assumed = Assumed.unionWith(R.getAssumed());
3031    return *this;
3032  }
3033};
3034
3035/// Simple state for a set.
3036///
3037/// This represents a state containing a set of values. The interface supports
3038/// modelling sets that contain all possible elements. The state's internal
3039/// value is modified using union or intersection operations.
3040template <typename BaseTy> struct SetState : public AbstractState {
3041  /// A wrapper around a set that has semantics for handling unions and
3042  /// intersections with a "universal" set that contains all elements.
3043  struct SetContents {
3044    /// Creates a universal set with no concrete elements or an empty set.
3045    SetContents(bool Universal) : Universal(Universal) {}
3046
3047    /// Creates a non-universal set with concrete values.
3048    SetContents(const DenseSet<BaseTy> &Assumptions)
3049        : Universal(false), Set(Assumptions) {}
3050
3051    SetContents(bool Universal, const DenseSet<BaseTy> &Assumptions)
3052        : Universal(Universal), Set(Assumptions) {}
3053
3054    const DenseSet<BaseTy> &getSet() const { return Set; }
3055
3056    bool isUniversal() const { return Universal; }
3057
3058    bool empty() const { return Set.empty() && !Universal; }
3059
3060    /// Finds A := A ^ B where A or B could be the "Universal" set which
3061    /// contains every possible attribute. Returns true if changes were made.
3062    bool getIntersection(const SetContents &RHS) {
3063      bool IsUniversal = Universal;
3064      unsigned Size = Set.size();
3065
3066      // A := A ^ U = A
3067      if (RHS.isUniversal())
3068        return false;
3069
3070      // A := U ^ B = B
3071      if (Universal)
3072        Set = RHS.getSet();
3073      else
3074        set_intersect(Set, RHS.getSet());
3075
3076      Universal &= RHS.isUniversal();
3077      return IsUniversal != Universal || Size != Set.size();
3078    }
3079
3080    /// Finds A := A u B where A or B could be the "Universal" set which
3081    /// contains every possible attribute. returns true if changes were made.
3082    bool getUnion(const SetContents &RHS) {
3083      bool IsUniversal = Universal;
3084      unsigned Size = Set.size();
3085
3086      // A := A u U = U = U u B
3087      if (!RHS.isUniversal() && !Universal)
3088        set_union(Set, RHS.getSet());
3089
3090      Universal |= RHS.isUniversal();
3091      return IsUniversal != Universal || Size != Set.size();
3092    }
3093
3094  private:
3095    /// Indicates if this set is "universal", containing every possible element.
3096    bool Universal;
3097
3098    /// The set of currently active assumptions.
3099    DenseSet<BaseTy> Set;
3100  };
3101
3102  SetState() : Known(false), Assumed(true), IsAtFixedpoint(false) {}
3103
3104  /// Initializes the known state with an initial set and initializes the
3105  /// assumed state as universal.
3106  SetState(const DenseSet<BaseTy> &Known)
3107      : Known(Known), Assumed(true), IsAtFixedpoint(false) {}
3108
3109  /// See AbstractState::isValidState()
3110  bool isValidState() const override { return !Assumed.empty(); }
3111
3112  /// See AbstractState::isAtFixpoint()
3113  bool isAtFixpoint() const override { return IsAtFixedpoint; }
3114
3115  /// See AbstractState::indicateOptimisticFixpoint(...)
3116  ChangeStatus indicateOptimisticFixpoint() override {
3117    IsAtFixedpoint = true;
3118    Known = Assumed;
3119    return ChangeStatus::UNCHANGED;
3120  }
3121
3122  /// See AbstractState::indicatePessimisticFixpoint(...)
3123  ChangeStatus indicatePessimisticFixpoint() override {
3124    IsAtFixedpoint = true;
3125    Assumed = Known;
3126    return ChangeStatus::CHANGED;
3127  }
3128
3129  /// Return the known state encoding.
3130  const SetContents &getKnown() const { return Known; }
3131
3132  /// Return the assumed state encoding.
3133  const SetContents &getAssumed() const { return Assumed; }
3134
3135  /// Returns if the set state contains the element.
3136  bool setContains(const BaseTy &Elem) const {
3137    return Assumed.getSet().contains(Elem) || Known.getSet().contains(Elem);
3138  }
3139
3140  /// Performs the set intersection between this set and \p RHS. Returns true if
3141  /// changes were made.
3142  bool getIntersection(const SetContents &RHS) {
3143    bool IsUniversal = Assumed.isUniversal();
3144    unsigned SizeBefore = Assumed.getSet().size();
3145
3146    // Get intersection and make sure that the known set is still a proper
3147    // subset of the assumed set. A := K u (A ^ R).
3148    Assumed.getIntersection(RHS);
3149    Assumed.getUnion(Known);
3150
3151    return SizeBefore != Assumed.getSet().size() ||
3152           IsUniversal != Assumed.isUniversal();
3153  }
3154
3155  /// Performs the set union between this set and \p RHS. Returns true if
3156  /// changes were made.
3157  bool getUnion(const SetContents &RHS) { return Assumed.getUnion(RHS); }
3158
3159private:
3160  /// The set of values known for this state.
3161  SetContents Known;
3162
3163  /// The set of assumed values for this state.
3164  SetContents Assumed;
3165
3166  bool IsAtFixedpoint;
3167};
3168
3169/// Helper to tie a abstract state implementation to an abstract attribute.
3170template <typename StateTy, typename BaseType, class... Ts>
3171struct StateWrapper : public BaseType, public StateTy {
3172  /// Provide static access to the type of the state.
3173  using StateType = StateTy;
3174
3175  StateWrapper(const IRPosition &IRP, Ts... Args)
3176      : BaseType(IRP), StateTy(Args...) {}
3177
3178  /// See AbstractAttribute::getState(...).
3179  StateType &getState() override { return *this; }
3180
3181  /// See AbstractAttribute::getState(...).
3182  const StateType &getState() const override { return *this; }
3183};
3184
3185/// Helper class that provides common functionality to manifest IR attributes.
3186template <Attribute::AttrKind AK, typename BaseType, typename AAType>
3187struct IRAttribute : public BaseType {
3188  IRAttribute(const IRPosition &IRP) : BaseType(IRP) {}
3189
3190  /// Most boolean IRAttribute AAs don't do anything non-trivial
3191  /// in their initializers while non-boolean ones often do. Subclasses can
3192  /// change this.
3193  static bool hasTrivialInitializer() { return Attribute::isEnumAttrKind(AK); }
3194
3195  /// Compile time access to the IR attribute kind.
3196  static constexpr Attribute::AttrKind IRAttributeKind = AK;
3197
3198  /// Return true if the IR attribute(s) associated with this AA are implied for
3199  /// an undef value.
3200  static bool isImpliedByUndef() { return true; }
3201
3202  /// Return true if the IR attribute(s) associated with this AA are implied for
3203  /// an poison value.
3204  static bool isImpliedByPoison() { return true; }
3205
3206  static bool isImpliedByIR(Attributor &A, const IRPosition &IRP,
3207                            Attribute::AttrKind ImpliedAttributeKind = AK,
3208                            bool IgnoreSubsumingPositions = false) {
3209    if (AAType::isImpliedByUndef() && isa<UndefValue>(IRP.getAssociatedValue()))
3210      return true;
3211    if (AAType::isImpliedByPoison() &&
3212        isa<PoisonValue>(IRP.getAssociatedValue()))
3213      return true;
3214    return A.hasAttr(IRP, {ImpliedAttributeKind}, IgnoreSubsumingPositions,
3215                     ImpliedAttributeKind);
3216  }
3217
3218  /// See AbstractAttribute::manifest(...).
3219  ChangeStatus manifest(Attributor &A) override {
3220    if (isa<UndefValue>(this->getIRPosition().getAssociatedValue()))
3221      return ChangeStatus::UNCHANGED;
3222    SmallVector<Attribute, 4> DeducedAttrs;
3223    getDeducedAttributes(A, this->getAnchorValue().getContext(), DeducedAttrs);
3224    if (DeducedAttrs.empty())
3225      return ChangeStatus::UNCHANGED;
3226    return A.manifestAttrs(this->getIRPosition(), DeducedAttrs);
3227  }
3228
3229  /// Return the kind that identifies the abstract attribute implementation.
3230  Attribute::AttrKind getAttrKind() const { return AK; }
3231
3232  /// Return the deduced attributes in \p Attrs.
3233  virtual void getDeducedAttributes(Attributor &A, LLVMContext &Ctx,
3234                                    SmallVectorImpl<Attribute> &Attrs) const {
3235    Attrs.emplace_back(Attribute::get(Ctx, getAttrKind()));
3236  }
3237};
3238
3239/// Base struct for all "concrete attribute" deductions.
3240///
3241/// The abstract attribute is a minimal interface that allows the Attributor to
3242/// orchestrate the abstract/fixpoint analysis. The design allows to hide away
3243/// implementation choices made for the subclasses but also to structure their
3244/// implementation and simplify the use of other abstract attributes in-flight.
3245///
3246/// To allow easy creation of new attributes, most methods have default
3247/// implementations. The ones that do not are generally straight forward, except
3248/// `AbstractAttribute::updateImpl` which is the location of most reasoning
3249/// associated with the abstract attribute. The update is invoked by the
3250/// Attributor in case the situation used to justify the current optimistic
3251/// state might have changed. The Attributor determines this automatically
3252/// by monitoring the `Attributor::getAAFor` calls made by abstract attributes.
3253///
3254/// The `updateImpl` method should inspect the IR and other abstract attributes
3255/// in-flight to justify the best possible (=optimistic) state. The actual
3256/// implementation is, similar to the underlying abstract state encoding, not
3257/// exposed. In the most common case, the `updateImpl` will go through a list of
3258/// reasons why its optimistic state is valid given the current information. If
3259/// any combination of them holds and is sufficient to justify the current
3260/// optimistic state, the method shall return UNCHAGED. If not, the optimistic
3261/// state is adjusted to the situation and the method shall return CHANGED.
3262///
3263/// If the manifestation of the "concrete attribute" deduced by the subclass
3264/// differs from the "default" behavior, which is a (set of) LLVM-IR
3265/// attribute(s) for an argument, call site argument, function return value, or
3266/// function, the `AbstractAttribute::manifest` method should be overloaded.
3267///
3268/// NOTE: If the state obtained via getState() is INVALID, thus if
3269///       AbstractAttribute::getState().isValidState() returns false, no
3270///       information provided by the methods of this class should be used.
3271/// NOTE: The Attributor currently has certain limitations to what we can do.
3272///       As a general rule of thumb, "concrete" abstract attributes should *for
3273///       now* only perform "backward" information propagation. That means
3274///       optimistic information obtained through abstract attributes should
3275///       only be used at positions that precede the origin of the information
3276///       with regards to the program flow. More practically, information can
3277///       *now* be propagated from instructions to their enclosing function, but
3278///       *not* from call sites to the called function. The mechanisms to allow
3279///       both directions will be added in the future.
3280/// NOTE: The mechanics of adding a new "concrete" abstract attribute are
3281///       described in the file comment.
3282struct AbstractAttribute : public IRPosition, public AADepGraphNode {
3283  using StateType = AbstractState;
3284
3285  AbstractAttribute(const IRPosition &IRP) : IRPosition(IRP) {}
3286
3287  /// Virtual destructor.
3288  virtual ~AbstractAttribute() = default;
3289
3290  /// Compile time access to the IR attribute kind.
3291  static constexpr Attribute::AttrKind IRAttributeKind = Attribute::None;
3292
3293  /// This function is used to identify if an \p DGN is of type
3294  /// AbstractAttribute so that the dyn_cast and cast can use such information
3295  /// to cast an AADepGraphNode to an AbstractAttribute.
3296  ///
3297  /// We eagerly return true here because all AADepGraphNodes except for the
3298  /// Synthethis Node are of type AbstractAttribute
3299  static bool classof(const AADepGraphNode *DGN) { return true; }
3300
3301  /// Return false if this AA does anything non-trivial (hence not done by
3302  /// default) in its initializer.
3303  static bool hasTrivialInitializer() { return false; }
3304
3305  /// Return true if this AA requires a "callee" (or an associted function) for
3306  /// a call site positon. Default is optimistic to minimize AAs.
3307  static bool requiresCalleeForCallBase() { return false; }
3308
3309  /// Return true if this AA requires non-asm "callee" for a call site positon.
3310  static bool requiresNonAsmForCallBase() { return true; }
3311
3312  /// Return true if this AA requires all callees for an argument or function
3313  /// positon.
3314  static bool requiresCallersForArgOrFunction() { return false; }
3315
3316  /// Return false if an AA should not be created for \p IRP.
3317  static bool isValidIRPositionForInit(Attributor &A, const IRPosition &IRP) {
3318    return true;
3319  }
3320
3321  /// Return false if an AA should not be updated for \p IRP.
3322  static bool isValidIRPositionForUpdate(Attributor &A, const IRPosition &IRP) {
3323    Function *AssociatedFn = IRP.getAssociatedFunction();
3324    bool IsFnInterface = IRP.isFnInterfaceKind();
3325    assert((!IsFnInterface || AssociatedFn) &&
3326           "Function interface without a function?");
3327
3328    // TODO: Not all attributes require an exact definition. Find a way to
3329    //       enable deduction for some but not all attributes in case the
3330    //       definition might be changed at runtime, see also
3331    //       http://lists.llvm.org/pipermail/llvm-dev/2018-February/121275.html.
3332    // TODO: We could always determine abstract attributes and if sufficient
3333    //       information was found we could duplicate the functions that do not
3334    //       have an exact definition.
3335    return !IsFnInterface || A.isFunctionIPOAmendable(*AssociatedFn);
3336  }
3337
3338  /// Initialize the state with the information in the Attributor \p A.
3339  ///
3340  /// This function is called by the Attributor once all abstract attributes
3341  /// have been identified. It can and shall be used for task like:
3342  ///  - identify existing knowledge in the IR and use it for the "known state"
3343  ///  - perform any work that is not going to change over time, e.g., determine
3344  ///    a subset of the IR, or attributes in-flight, that have to be looked at
3345  ///    in the `updateImpl` method.
3346  virtual void initialize(Attributor &A) {}
3347
3348  /// A query AA is always scheduled as long as we do updates because it does
3349  /// lazy computation that cannot be determined to be done from the outside.
3350  /// However, while query AAs will not be fixed if they do not have outstanding
3351  /// dependences, we will only schedule them like other AAs. If a query AA that
3352  /// received a new query it needs to request an update via
3353  /// `Attributor::requestUpdateForAA`.
3354  virtual bool isQueryAA() const { return false; }
3355
3356  /// Return the internal abstract state for inspection.
3357  virtual StateType &getState() = 0;
3358  virtual const StateType &getState() const = 0;
3359
3360  /// Return an IR position, see struct IRPosition.
3361  const IRPosition &getIRPosition() const { return *this; };
3362  IRPosition &getIRPosition() { return *this; };
3363
3364  /// Helper functions, for debug purposes only.
3365  ///{
3366  void print(raw_ostream &OS) const { print(nullptr, OS); }
3367  void print(Attributor *, raw_ostream &OS) const override;
3368  virtual void printWithDeps(raw_ostream &OS) const;
3369  void dump() const { this->print(dbgs()); }
3370
3371  /// This function should return the "summarized" assumed state as string.
3372  virtual const std::string getAsStr(Attributor *A) const = 0;
3373
3374  /// This function should return the name of the AbstractAttribute
3375  virtual const std::string getName() const = 0;
3376
3377  /// This function should return the address of the ID of the AbstractAttribute
3378  virtual const char *getIdAddr() const = 0;
3379  ///}
3380
3381  /// Allow the Attributor access to the protected methods.
3382  friend struct Attributor;
3383
3384protected:
3385  /// Hook for the Attributor to trigger an update of the internal state.
3386  ///
3387  /// If this attribute is already fixed, this method will return UNCHANGED,
3388  /// otherwise it delegates to `AbstractAttribute::updateImpl`.
3389  ///
3390  /// \Return CHANGED if the internal state changed, otherwise UNCHANGED.
3391  ChangeStatus update(Attributor &A);
3392
3393  /// Hook for the Attributor to trigger the manifestation of the information
3394  /// represented by the abstract attribute in the LLVM-IR.
3395  ///
3396  /// \Return CHANGED if the IR was altered, otherwise UNCHANGED.
3397  virtual ChangeStatus manifest(Attributor &A) {
3398    return ChangeStatus::UNCHANGED;
3399  }
3400
3401  /// Hook to enable custom statistic tracking, called after manifest that
3402  /// resulted in a change if statistics are enabled.
3403  ///
3404  /// We require subclasses to provide an implementation so we remember to
3405  /// add statistics for them.
3406  virtual void trackStatistics() const = 0;
3407
3408  /// The actual update/transfer function which has to be implemented by the
3409  /// derived classes.
3410  ///
3411  /// If it is called, the environment has changed and we have to determine if
3412  /// the current information is still valid or adjust it otherwise.
3413  ///
3414  /// \Return CHANGED if the internal state changed, otherwise UNCHANGED.
3415  virtual ChangeStatus updateImpl(Attributor &A) = 0;
3416};
3417
3418/// Forward declarations of output streams for debug purposes.
3419///
3420///{
3421raw_ostream &operator<<(raw_ostream &OS, const AbstractAttribute &AA);
3422raw_ostream &operator<<(raw_ostream &OS, ChangeStatus S);
3423raw_ostream &operator<<(raw_ostream &OS, IRPosition::Kind);
3424raw_ostream &operator<<(raw_ostream &OS, const IRPosition &);
3425raw_ostream &operator<<(raw_ostream &OS, const AbstractState &State);
3426template <typename base_ty, base_ty BestState, base_ty WorstState>
3427raw_ostream &
3428operator<<(raw_ostream &OS,
3429           const IntegerStateBase<base_ty, BestState, WorstState> &S) {
3430  return OS << "(" << S.getKnown() << "-" << S.getAssumed() << ")"
3431            << static_cast<const AbstractState &>(S);
3432}
3433raw_ostream &operator<<(raw_ostream &OS, const IntegerRangeState &State);
3434///}
3435
3436struct AttributorPass : public PassInfoMixin<AttributorPass> {
3437  PreservedAnalyses run(Module &M, ModuleAnalysisManager &AM);
3438};
3439struct AttributorCGSCCPass : public PassInfoMixin<AttributorCGSCCPass> {
3440  PreservedAnalyses run(LazyCallGraph::SCC &C, CGSCCAnalysisManager &AM,
3441                        LazyCallGraph &CG, CGSCCUpdateResult &UR);
3442};
3443
3444/// A more lightweight version of the Attributor which only runs attribute
3445/// inference but no simplifications.
3446struct AttributorLightPass : public PassInfoMixin<AttributorLightPass> {
3447  PreservedAnalyses run(Module &M, ModuleAnalysisManager &AM);
3448};
3449
3450/// A more lightweight version of the Attributor which only runs attribute
3451/// inference but no simplifications.
3452struct AttributorLightCGSCCPass
3453    : public PassInfoMixin<AttributorLightCGSCCPass> {
3454  PreservedAnalyses run(LazyCallGraph::SCC &C, CGSCCAnalysisManager &AM,
3455                        LazyCallGraph &CG, CGSCCUpdateResult &UR);
3456};
3457
3458/// Helper function to clamp a state \p S of type \p StateType with the
3459/// information in \p R and indicate/return if \p S did change (as-in update is
3460/// required to be run again).
3461template <typename StateType>
3462ChangeStatus clampStateAndIndicateChange(StateType &S, const StateType &R) {
3463  auto Assumed = S.getAssumed();
3464  S ^= R;
3465  return Assumed == S.getAssumed() ? ChangeStatus::UNCHANGED
3466                                   : ChangeStatus::CHANGED;
3467}
3468
3469/// ----------------------------------------------------------------------------
3470///                       Abstract Attribute Classes
3471/// ----------------------------------------------------------------------------
3472
3473struct AANoUnwind
3474    : public IRAttribute<Attribute::NoUnwind,
3475                         StateWrapper<BooleanState, AbstractAttribute>,
3476                         AANoUnwind> {
3477  AANoUnwind(const IRPosition &IRP, Attributor &A) : IRAttribute(IRP) {}
3478
3479  /// Returns true if nounwind is assumed.
3480  bool isAssumedNoUnwind() const { return getAssumed(); }
3481
3482  /// Returns true if nounwind is known.
3483  bool isKnownNoUnwind() const { return getKnown(); }
3484
3485  /// Create an abstract attribute view for the position \p IRP.
3486  static AANoUnwind &createForPosition(const IRPosition &IRP, Attributor &A);
3487
3488  /// See AbstractAttribute::getName()
3489  const std::string getName() const override { return "AANoUnwind"; }
3490
3491  /// See AbstractAttribute::getIdAddr()
3492  const char *getIdAddr() const override { return &ID; }
3493
3494  /// This function should return true if the type of the \p AA is AANoUnwind
3495  static bool classof(const AbstractAttribute *AA) {
3496    return (AA->getIdAddr() == &ID);
3497  }
3498
3499  /// Unique ID (due to the unique address)
3500  static const char ID;
3501};
3502
3503struct AANoSync
3504    : public IRAttribute<Attribute::NoSync,
3505                         StateWrapper<BooleanState, AbstractAttribute>,
3506                         AANoSync> {
3507  AANoSync(const IRPosition &IRP, Attributor &A) : IRAttribute(IRP) {}
3508
3509  static bool isImpliedByIR(Attributor &A, const IRPosition &IRP,
3510                            Attribute::AttrKind ImpliedAttributeKind,
3511                            bool IgnoreSubsumingPositions = false) {
3512    // Note: This is also run for non-IPO amendable functions.
3513    assert(ImpliedAttributeKind == Attribute::NoSync);
3514    if (A.hasAttr(IRP, {Attribute::NoSync}, IgnoreSubsumingPositions,
3515                  Attribute::NoSync))
3516      return true;
3517
3518    // Check for readonly + non-convergent.
3519    // TODO: We should be able to use hasAttr for Attributes, not only
3520    // AttrKinds.
3521    Function *F = IRP.getAssociatedFunction();
3522    if (!F || F->isConvergent())
3523      return false;
3524
3525    SmallVector<Attribute, 2> Attrs;
3526    A.getAttrs(IRP, {Attribute::Memory}, Attrs, IgnoreSubsumingPositions);
3527
3528    MemoryEffects ME = MemoryEffects::unknown();
3529    for (const Attribute &Attr : Attrs)
3530      ME &= Attr.getMemoryEffects();
3531
3532    if (!ME.onlyReadsMemory())
3533      return false;
3534
3535    A.manifestAttrs(IRP, Attribute::get(F->getContext(), Attribute::NoSync));
3536    return true;
3537  }
3538
3539  /// See AbstractAttribute::isValidIRPositionForInit
3540  static bool isValidIRPositionForInit(Attributor &A, const IRPosition &IRP) {
3541    if (!IRP.isFunctionScope() &&
3542        !IRP.getAssociatedType()->isPtrOrPtrVectorTy())
3543      return false;
3544    return IRAttribute::isValidIRPositionForInit(A, IRP);
3545  }
3546
3547  /// Returns true if "nosync" is assumed.
3548  bool isAssumedNoSync() const { return getAssumed(); }
3549
3550  /// Returns true if "nosync" is known.
3551  bool isKnownNoSync() const { return getKnown(); }
3552
3553  /// Helper function used to determine whether an instruction is non-relaxed
3554  /// atomic. In other words, if an atomic instruction does not have unordered
3555  /// or monotonic ordering
3556  static bool isNonRelaxedAtomic(const Instruction *I);
3557
3558  /// Helper function specific for intrinsics which are potentially volatile.
3559  static bool isNoSyncIntrinsic(const Instruction *I);
3560
3561  /// Helper function to determine if \p CB is an aligned (GPU) barrier. Aligned
3562  /// barriers have to be executed by all threads. The flag \p ExecutedAligned
3563  /// indicates if the call is executed by all threads in a (thread) block in an
3564  /// aligned way. If that is the case, non-aligned barriers are effectively
3565  /// aligned barriers.
3566  static bool isAlignedBarrier(const CallBase &CB, bool ExecutedAligned);
3567
3568  /// Create an abstract attribute view for the position \p IRP.
3569  static AANoSync &createForPosition(const IRPosition &IRP, Attributor &A);
3570
3571  /// See AbstractAttribute::getName()
3572  const std::string getName() const override { return "AANoSync"; }
3573
3574  /// See AbstractAttribute::getIdAddr()
3575  const char *getIdAddr() const override { return &ID; }
3576
3577  /// This function should return true if the type of the \p AA is AANoSync
3578  static bool classof(const AbstractAttribute *AA) {
3579    return (AA->getIdAddr() == &ID);
3580  }
3581
3582  /// Unique ID (due to the unique address)
3583  static const char ID;
3584};
3585
3586/// An abstract interface for all nonnull attributes.
3587struct AAMustProgress
3588    : public IRAttribute<Attribute::MustProgress,
3589                         StateWrapper<BooleanState, AbstractAttribute>,
3590                         AAMustProgress> {
3591  AAMustProgress(const IRPosition &IRP, Attributor &A) : IRAttribute(IRP) {}
3592
3593  static bool isImpliedByIR(Attributor &A, const IRPosition &IRP,
3594                            Attribute::AttrKind ImpliedAttributeKind,
3595                            bool IgnoreSubsumingPositions = false) {
3596    // Note: This is also run for non-IPO amendable functions.
3597    assert(ImpliedAttributeKind == Attribute::MustProgress);
3598    return A.hasAttr(IRP, {Attribute::MustProgress, Attribute::WillReturn},
3599                     IgnoreSubsumingPositions, Attribute::MustProgress);
3600  }
3601
3602  /// Return true if we assume that the underlying value is nonnull.
3603  bool isAssumedMustProgress() const { return getAssumed(); }
3604
3605  /// Return true if we know that underlying value is nonnull.
3606  bool isKnownMustProgress() const { return getKnown(); }
3607
3608  /// Create an abstract attribute view for the position \p IRP.
3609  static AAMustProgress &createForPosition(const IRPosition &IRP,
3610                                           Attributor &A);
3611
3612  /// See AbstractAttribute::getName()
3613  const std::string getName() const override { return "AAMustProgress"; }
3614
3615  /// See AbstractAttribute::getIdAddr()
3616  const char *getIdAddr() const override { return &ID; }
3617
3618  /// This function should return true if the type of the \p AA is
3619  /// AAMustProgress
3620  static bool classof(const AbstractAttribute *AA) {
3621    return (AA->getIdAddr() == &ID);
3622  }
3623
3624  /// Unique ID (due to the unique address)
3625  static const char ID;
3626};
3627
3628/// An abstract interface for all nonnull attributes.
3629struct AANonNull
3630    : public IRAttribute<Attribute::NonNull,
3631                         StateWrapper<BooleanState, AbstractAttribute>,
3632                         AANonNull> {
3633  AANonNull(const IRPosition &IRP, Attributor &A) : IRAttribute(IRP) {}
3634
3635  /// See AbstractAttribute::hasTrivialInitializer.
3636  static bool hasTrivialInitializer() { return false; }
3637
3638  /// See IRAttribute::isImpliedByUndef.
3639  /// Undef is not necessarily nonnull as nonnull + noundef would cause poison.
3640  /// Poison implies nonnull though.
3641  static bool isImpliedByUndef() { return false; }
3642
3643  /// See AbstractAttribute::isValidIRPositionForInit
3644  static bool isValidIRPositionForInit(Attributor &A, const IRPosition &IRP) {
3645    if (!IRP.getAssociatedType()->isPtrOrPtrVectorTy())
3646      return false;
3647    return IRAttribute::isValidIRPositionForInit(A, IRP);
3648  }
3649
3650  /// See AbstractAttribute::isImpliedByIR(...).
3651  static bool isImpliedByIR(Attributor &A, const IRPosition &IRP,
3652                            Attribute::AttrKind ImpliedAttributeKind,
3653                            bool IgnoreSubsumingPositions = false);
3654
3655  /// Return true if we assume that the underlying value is nonnull.
3656  bool isAssumedNonNull() const { return getAssumed(); }
3657
3658  /// Return true if we know that underlying value is nonnull.
3659  bool isKnownNonNull() const { return getKnown(); }
3660
3661  /// Create an abstract attribute view for the position \p IRP.
3662  static AANonNull &createForPosition(const IRPosition &IRP, Attributor &A);
3663
3664  /// See AbstractAttribute::getName()
3665  const std::string getName() const override { return "AANonNull"; }
3666
3667  /// See AbstractAttribute::getIdAddr()
3668  const char *getIdAddr() const override { return &ID; }
3669
3670  /// This function should return true if the type of the \p AA is AANonNull
3671  static bool classof(const AbstractAttribute *AA) {
3672    return (AA->getIdAddr() == &ID);
3673  }
3674
3675  /// Unique ID (due to the unique address)
3676  static const char ID;
3677};
3678
3679/// An abstract attribute for norecurse.
3680struct AANoRecurse
3681    : public IRAttribute<Attribute::NoRecurse,
3682                         StateWrapper<BooleanState, AbstractAttribute>,
3683                         AANoRecurse> {
3684  AANoRecurse(const IRPosition &IRP, Attributor &A) : IRAttribute(IRP) {}
3685
3686  /// Return true if "norecurse" is assumed.
3687  bool isAssumedNoRecurse() const { return getAssumed(); }
3688
3689  /// Return true if "norecurse" is known.
3690  bool isKnownNoRecurse() const { return getKnown(); }
3691
3692  /// Create an abstract attribute view for the position \p IRP.
3693  static AANoRecurse &createForPosition(const IRPosition &IRP, Attributor &A);
3694
3695  /// See AbstractAttribute::getName()
3696  const std::string getName() const override { return "AANoRecurse"; }
3697
3698  /// See AbstractAttribute::getIdAddr()
3699  const char *getIdAddr() const override { return &ID; }
3700
3701  /// This function should return true if the type of the \p AA is AANoRecurse
3702  static bool classof(const AbstractAttribute *AA) {
3703    return (AA->getIdAddr() == &ID);
3704  }
3705
3706  /// Unique ID (due to the unique address)
3707  static const char ID;
3708};
3709
3710/// An abstract attribute for willreturn.
3711struct AAWillReturn
3712    : public IRAttribute<Attribute::WillReturn,
3713                         StateWrapper<BooleanState, AbstractAttribute>,
3714                         AAWillReturn> {
3715  AAWillReturn(const IRPosition &IRP, Attributor &A) : IRAttribute(IRP) {}
3716
3717  static bool isImpliedByIR(Attributor &A, const IRPosition &IRP,
3718                            Attribute::AttrKind ImpliedAttributeKind,
3719                            bool IgnoreSubsumingPositions = false) {
3720    // Note: This is also run for non-IPO amendable functions.
3721    assert(ImpliedAttributeKind == Attribute::WillReturn);
3722    if (IRAttribute::isImpliedByIR(A, IRP, ImpliedAttributeKind,
3723                                   IgnoreSubsumingPositions))
3724      return true;
3725    if (!isImpliedByMustprogressAndReadonly(A, IRP))
3726      return false;
3727    A.manifestAttrs(IRP, Attribute::get(IRP.getAnchorValue().getContext(),
3728                                        Attribute::WillReturn));
3729    return true;
3730  }
3731
3732  /// Check for `mustprogress` and `readonly` as they imply `willreturn`.
3733  static bool isImpliedByMustprogressAndReadonly(Attributor &A,
3734                                                 const IRPosition &IRP) {
3735    // Check for `mustprogress` in the scope and the associated function which
3736    // might be different if this is a call site.
3737    if (!A.hasAttr(IRP, {Attribute::MustProgress}))
3738      return false;
3739
3740    SmallVector<Attribute, 2> Attrs;
3741    A.getAttrs(IRP, {Attribute::Memory}, Attrs,
3742               /* IgnoreSubsumingPositions */ false);
3743
3744    MemoryEffects ME = MemoryEffects::unknown();
3745    for (const Attribute &Attr : Attrs)
3746      ME &= Attr.getMemoryEffects();
3747    return ME.onlyReadsMemory();
3748  }
3749
3750  /// Return true if "willreturn" is assumed.
3751  bool isAssumedWillReturn() const { return getAssumed(); }
3752
3753  /// Return true if "willreturn" is known.
3754  bool isKnownWillReturn() const { return getKnown(); }
3755
3756  /// Create an abstract attribute view for the position \p IRP.
3757  static AAWillReturn &createForPosition(const IRPosition &IRP, Attributor &A);
3758
3759  /// See AbstractAttribute::getName()
3760  const std::string getName() const override { return "AAWillReturn"; }
3761
3762  /// See AbstractAttribute::getIdAddr()
3763  const char *getIdAddr() const override { return &ID; }
3764
3765  /// This function should return true if the type of the \p AA is AAWillReturn
3766  static bool classof(const AbstractAttribute *AA) {
3767    return (AA->getIdAddr() == &ID);
3768  }
3769
3770  /// Unique ID (due to the unique address)
3771  static const char ID;
3772};
3773
3774/// An abstract attribute for undefined behavior.
3775struct AAUndefinedBehavior
3776    : public StateWrapper<BooleanState, AbstractAttribute> {
3777  using Base = StateWrapper<BooleanState, AbstractAttribute>;
3778  AAUndefinedBehavior(const IRPosition &IRP, Attributor &A) : Base(IRP) {}
3779
3780  /// Return true if "undefined behavior" is assumed.
3781  bool isAssumedToCauseUB() const { return getAssumed(); }
3782
3783  /// Return true if "undefined behavior" is assumed for a specific instruction.
3784  virtual bool isAssumedToCauseUB(Instruction *I) const = 0;
3785
3786  /// Return true if "undefined behavior" is known.
3787  bool isKnownToCauseUB() const { return getKnown(); }
3788
3789  /// Return true if "undefined behavior" is known for a specific instruction.
3790  virtual bool isKnownToCauseUB(Instruction *I) const = 0;
3791
3792  /// Create an abstract attribute view for the position \p IRP.
3793  static AAUndefinedBehavior &createForPosition(const IRPosition &IRP,
3794                                                Attributor &A);
3795
3796  /// See AbstractAttribute::getName()
3797  const std::string getName() const override { return "AAUndefinedBehavior"; }
3798
3799  /// See AbstractAttribute::getIdAddr()
3800  const char *getIdAddr() const override { return &ID; }
3801
3802  /// This function should return true if the type of the \p AA is
3803  /// AAUndefineBehavior
3804  static bool classof(const AbstractAttribute *AA) {
3805    return (AA->getIdAddr() == &ID);
3806  }
3807
3808  /// Unique ID (due to the unique address)
3809  static const char ID;
3810};
3811
3812/// An abstract interface to determine reachability of point A to B.
3813struct AAIntraFnReachability
3814    : public StateWrapper<BooleanState, AbstractAttribute> {
3815  using Base = StateWrapper<BooleanState, AbstractAttribute>;
3816  AAIntraFnReachability(const IRPosition &IRP, Attributor &A) : Base(IRP) {}
3817
3818  /// Returns true if 'From' instruction is assumed to reach, 'To' instruction.
3819  /// Users should provide two positions they are interested in, and the class
3820  /// determines (and caches) reachability.
3821  virtual bool isAssumedReachable(
3822      Attributor &A, const Instruction &From, const Instruction &To,
3823      const AA::InstExclusionSetTy *ExclusionSet = nullptr) const = 0;
3824
3825  /// Create an abstract attribute view for the position \p IRP.
3826  static AAIntraFnReachability &createForPosition(const IRPosition &IRP,
3827                                                  Attributor &A);
3828
3829  /// See AbstractAttribute::getName()
3830  const std::string getName() const override { return "AAIntraFnReachability"; }
3831
3832  /// See AbstractAttribute::getIdAddr()
3833  const char *getIdAddr() const override { return &ID; }
3834
3835  /// This function should return true if the type of the \p AA is
3836  /// AAIntraFnReachability
3837  static bool classof(const AbstractAttribute *AA) {
3838    return (AA->getIdAddr() == &ID);
3839  }
3840
3841  /// Unique ID (due to the unique address)
3842  static const char ID;
3843};
3844
3845/// An abstract interface for all noalias attributes.
3846struct AANoAlias
3847    : public IRAttribute<Attribute::NoAlias,
3848                         StateWrapper<BooleanState, AbstractAttribute>,
3849                         AANoAlias> {
3850  AANoAlias(const IRPosition &IRP, Attributor &A) : IRAttribute(IRP) {}
3851
3852  /// See AbstractAttribute::isValidIRPositionForInit
3853  static bool isValidIRPositionForInit(Attributor &A, const IRPosition &IRP) {
3854    if (!IRP.getAssociatedType()->isPtrOrPtrVectorTy())
3855      return false;
3856    return IRAttribute::isValidIRPositionForInit(A, IRP);
3857  }
3858
3859  /// See IRAttribute::isImpliedByIR
3860  static bool isImpliedByIR(Attributor &A, const IRPosition &IRP,
3861                            Attribute::AttrKind ImpliedAttributeKind,
3862                            bool IgnoreSubsumingPositions = false);
3863
3864  /// See AbstractAttribute::requiresCallersForArgOrFunction
3865  static bool requiresCallersForArgOrFunction() { return true; }
3866
3867  /// Return true if we assume that the underlying value is alias.
3868  bool isAssumedNoAlias() const { return getAssumed(); }
3869
3870  /// Return true if we know that underlying value is noalias.
3871  bool isKnownNoAlias() const { return getKnown(); }
3872
3873  /// Create an abstract attribute view for the position \p IRP.
3874  static AANoAlias &createForPosition(const IRPosition &IRP, Attributor &A);
3875
3876  /// See AbstractAttribute::getName()
3877  const std::string getName() const override { return "AANoAlias"; }
3878
3879  /// See AbstractAttribute::getIdAddr()
3880  const char *getIdAddr() const override { return &ID; }
3881
3882  /// This function should return true if the type of the \p AA is AANoAlias
3883  static bool classof(const AbstractAttribute *AA) {
3884    return (AA->getIdAddr() == &ID);
3885  }
3886
3887  /// Unique ID (due to the unique address)
3888  static const char ID;
3889};
3890
3891/// An AbstractAttribute for nofree.
3892struct AANoFree
3893    : public IRAttribute<Attribute::NoFree,
3894                         StateWrapper<BooleanState, AbstractAttribute>,
3895                         AANoFree> {
3896  AANoFree(const IRPosition &IRP, Attributor &A) : IRAttribute(IRP) {}
3897
3898  /// See IRAttribute::isImpliedByIR
3899  static bool isImpliedByIR(Attributor &A, const IRPosition &IRP,
3900                            Attribute::AttrKind ImpliedAttributeKind,
3901                            bool IgnoreSubsumingPositions = false) {
3902    // Note: This is also run for non-IPO amendable functions.
3903    assert(ImpliedAttributeKind == Attribute::NoFree);
3904    return A.hasAttr(
3905        IRP, {Attribute::ReadNone, Attribute::ReadOnly, Attribute::NoFree},
3906        IgnoreSubsumingPositions, Attribute::NoFree);
3907  }
3908
3909  /// See AbstractAttribute::isValidIRPositionForInit
3910  static bool isValidIRPositionForInit(Attributor &A, const IRPosition &IRP) {
3911    if (!IRP.isFunctionScope() &&
3912        !IRP.getAssociatedType()->isPtrOrPtrVectorTy())
3913      return false;
3914    return IRAttribute::isValidIRPositionForInit(A, IRP);
3915  }
3916
3917  /// Return true if "nofree" is assumed.
3918  bool isAssumedNoFree() const { return getAssumed(); }
3919
3920  /// Return true if "nofree" is known.
3921  bool isKnownNoFree() const { return getKnown(); }
3922
3923  /// Create an abstract attribute view for the position \p IRP.
3924  static AANoFree &createForPosition(const IRPosition &IRP, Attributor &A);
3925
3926  /// See AbstractAttribute::getName()
3927  const std::string getName() const override { return "AANoFree"; }
3928
3929  /// See AbstractAttribute::getIdAddr()
3930  const char *getIdAddr() const override { return &ID; }
3931
3932  /// This function should return true if the type of the \p AA is AANoFree
3933  static bool classof(const AbstractAttribute *AA) {
3934    return (AA->getIdAddr() == &ID);
3935  }
3936
3937  /// Unique ID (due to the unique address)
3938  static const char ID;
3939};
3940
3941/// An AbstractAttribute for noreturn.
3942struct AANoReturn
3943    : public IRAttribute<Attribute::NoReturn,
3944                         StateWrapper<BooleanState, AbstractAttribute>,
3945                         AANoReturn> {
3946  AANoReturn(const IRPosition &IRP, Attributor &A) : IRAttribute(IRP) {}
3947
3948  /// Return true if the underlying object is assumed to never return.
3949  bool isAssumedNoReturn() const { return getAssumed(); }
3950
3951  /// Return true if the underlying object is known to never return.
3952  bool isKnownNoReturn() const { return getKnown(); }
3953
3954  /// Create an abstract attribute view for the position \p IRP.
3955  static AANoReturn &createForPosition(const IRPosition &IRP, Attributor &A);
3956
3957  /// See AbstractAttribute::getName()
3958  const std::string getName() const override { return "AANoReturn"; }
3959
3960  /// See AbstractAttribute::getIdAddr()
3961  const char *getIdAddr() const override { return &ID; }
3962
3963  /// This function should return true if the type of the \p AA is AANoReturn
3964  static bool classof(const AbstractAttribute *AA) {
3965    return (AA->getIdAddr() == &ID);
3966  }
3967
3968  /// Unique ID (due to the unique address)
3969  static const char ID;
3970};
3971
3972/// An abstract interface for liveness abstract attribute.
3973struct AAIsDead
3974    : public StateWrapper<BitIntegerState<uint8_t, 3, 0>, AbstractAttribute> {
3975  using Base = StateWrapper<BitIntegerState<uint8_t, 3, 0>, AbstractAttribute>;
3976  AAIsDead(const IRPosition &IRP, Attributor &A) : Base(IRP) {}
3977
3978  /// See AbstractAttribute::isValidIRPositionForInit
3979  static bool isValidIRPositionForInit(Attributor &A, const IRPosition &IRP) {
3980    if (IRP.getPositionKind() == IRPosition::IRP_FUNCTION)
3981      return isa<Function>(IRP.getAnchorValue()) &&
3982             !cast<Function>(IRP.getAnchorValue()).isDeclaration();
3983    return true;
3984  }
3985
3986  /// State encoding bits. A set bit in the state means the property holds.
3987  enum {
3988    HAS_NO_EFFECT = 1 << 0,
3989    IS_REMOVABLE = 1 << 1,
3990
3991    IS_DEAD = HAS_NO_EFFECT | IS_REMOVABLE,
3992  };
3993  static_assert(IS_DEAD == getBestState(), "Unexpected BEST_STATE value");
3994
3995protected:
3996  /// The query functions are protected such that other attributes need to go
3997  /// through the Attributor interfaces: `Attributor::isAssumedDead(...)`
3998
3999  /// Returns true if the underlying value is assumed dead.
4000  virtual bool isAssumedDead() const = 0;
4001
4002  /// Returns true if the underlying value is known dead.
4003  virtual bool isKnownDead() const = 0;
4004
4005  /// Returns true if \p BB is known dead.
4006  virtual bool isKnownDead(const BasicBlock *BB) const = 0;
4007
4008  /// Returns true if \p I is assumed dead.
4009  virtual bool isAssumedDead(const Instruction *I) const = 0;
4010
4011  /// Returns true if \p I is known dead.
4012  virtual bool isKnownDead(const Instruction *I) const = 0;
4013
4014  /// Return true if the underlying value is a store that is known to be
4015  /// removable. This is different from dead stores as the removable store
4016  /// can have an effect on live values, especially loads, but that effect
4017  /// is propagated which allows us to remove the store in turn.
4018  virtual bool isRemovableStore() const { return false; }
4019
4020  /// This method is used to check if at least one instruction in a collection
4021  /// of instructions is live.
4022  template <typename T> bool isLiveInstSet(T begin, T end) const {
4023    for (const auto &I : llvm::make_range(begin, end)) {
4024      assert(I->getFunction() == getIRPosition().getAssociatedFunction() &&
4025             "Instruction must be in the same anchor scope function.");
4026
4027      if (!isAssumedDead(I))
4028        return true;
4029    }
4030
4031    return false;
4032  }
4033
4034public:
4035  /// Create an abstract attribute view for the position \p IRP.
4036  static AAIsDead &createForPosition(const IRPosition &IRP, Attributor &A);
4037
4038  /// Determine if \p F might catch asynchronous exceptions.
4039  static bool mayCatchAsynchronousExceptions(const Function &F) {
4040    return F.hasPersonalityFn() && !canSimplifyInvokeNoUnwind(&F);
4041  }
4042
4043  /// Returns true if \p BB is assumed dead.
4044  virtual bool isAssumedDead(const BasicBlock *BB) const = 0;
4045
4046  /// Return if the edge from \p From BB to \p To BB is assumed dead.
4047  /// This is specifically useful in AAReachability.
4048  virtual bool isEdgeDead(const BasicBlock *From, const BasicBlock *To) const {
4049    return false;
4050  }
4051
4052  /// See AbstractAttribute::getName()
4053  const std::string getName() const override { return "AAIsDead"; }
4054
4055  /// See AbstractAttribute::getIdAddr()
4056  const char *getIdAddr() const override { return &ID; }
4057
4058  /// This function should return true if the type of the \p AA is AAIsDead
4059  static bool classof(const AbstractAttribute *AA) {
4060    return (AA->getIdAddr() == &ID);
4061  }
4062
4063  /// Unique ID (due to the unique address)
4064  static const char ID;
4065
4066  friend struct Attributor;
4067};
4068
4069/// State for dereferenceable attribute
4070struct DerefState : AbstractState {
4071
4072  static DerefState getBestState() { return DerefState(); }
4073  static DerefState getBestState(const DerefState &) { return getBestState(); }
4074
4075  /// Return the worst possible representable state.
4076  static DerefState getWorstState() {
4077    DerefState DS;
4078    DS.indicatePessimisticFixpoint();
4079    return DS;
4080  }
4081  static DerefState getWorstState(const DerefState &) {
4082    return getWorstState();
4083  }
4084
4085  /// State representing for dereferenceable bytes.
4086  IncIntegerState<> DerefBytesState;
4087
4088  /// Map representing for accessed memory offsets and sizes.
4089  /// A key is Offset and a value is size.
4090  /// If there is a load/store instruction something like,
4091  ///   p[offset] = v;
4092  /// (offset, sizeof(v)) will be inserted to this map.
4093  /// std::map is used because we want to iterate keys in ascending order.
4094  std::map<int64_t, uint64_t> AccessedBytesMap;
4095
4096  /// Helper function to calculate dereferenceable bytes from current known
4097  /// bytes and accessed bytes.
4098  ///
4099  /// int f(int *A){
4100  ///    *A = 0;
4101  ///    *(A+2) = 2;
4102  ///    *(A+1) = 1;
4103  ///    *(A+10) = 10;
4104  /// }
4105  /// ```
4106  /// In that case, AccessedBytesMap is `{0:4, 4:4, 8:4, 40:4}`.
4107  /// AccessedBytesMap is std::map so it is iterated in accending order on
4108  /// key(Offset). So KnownBytes will be updated like this:
4109  ///
4110  /// |Access | KnownBytes
4111  /// |(0, 4)| 0 -> 4
4112  /// |(4, 4)| 4 -> 8
4113  /// |(8, 4)| 8 -> 12
4114  /// |(40, 4) | 12 (break)
4115  void computeKnownDerefBytesFromAccessedMap() {
4116    int64_t KnownBytes = DerefBytesState.getKnown();
4117    for (auto &Access : AccessedBytesMap) {
4118      if (KnownBytes < Access.first)
4119        break;
4120      KnownBytes = std::max(KnownBytes, Access.first + (int64_t)Access.second);
4121    }
4122
4123    DerefBytesState.takeKnownMaximum(KnownBytes);
4124  }
4125
4126  /// State representing that whether the value is globaly dereferenceable.
4127  BooleanState GlobalState;
4128
4129  /// See AbstractState::isValidState()
4130  bool isValidState() const override { return DerefBytesState.isValidState(); }
4131
4132  /// See AbstractState::isAtFixpoint()
4133  bool isAtFixpoint() const override {
4134    return !isValidState() ||
4135           (DerefBytesState.isAtFixpoint() && GlobalState.isAtFixpoint());
4136  }
4137
4138  /// See AbstractState::indicateOptimisticFixpoint(...)
4139  ChangeStatus indicateOptimisticFixpoint() override {
4140    DerefBytesState.indicateOptimisticFixpoint();
4141    GlobalState.indicateOptimisticFixpoint();
4142    return ChangeStatus::UNCHANGED;
4143  }
4144
4145  /// See AbstractState::indicatePessimisticFixpoint(...)
4146  ChangeStatus indicatePessimisticFixpoint() override {
4147    DerefBytesState.indicatePessimisticFixpoint();
4148    GlobalState.indicatePessimisticFixpoint();
4149    return ChangeStatus::CHANGED;
4150  }
4151
4152  /// Update known dereferenceable bytes.
4153  void takeKnownDerefBytesMaximum(uint64_t Bytes) {
4154    DerefBytesState.takeKnownMaximum(Bytes);
4155
4156    // Known bytes might increase.
4157    computeKnownDerefBytesFromAccessedMap();
4158  }
4159
4160  /// Update assumed dereferenceable bytes.
4161  void takeAssumedDerefBytesMinimum(uint64_t Bytes) {
4162    DerefBytesState.takeAssumedMinimum(Bytes);
4163  }
4164
4165  /// Add accessed bytes to the map.
4166  void addAccessedBytes(int64_t Offset, uint64_t Size) {
4167    uint64_t &AccessedBytes = AccessedBytesMap[Offset];
4168    AccessedBytes = std::max(AccessedBytes, Size);
4169
4170    // Known bytes might increase.
4171    computeKnownDerefBytesFromAccessedMap();
4172  }
4173
4174  /// Equality for DerefState.
4175  bool operator==(const DerefState &R) const {
4176    return this->DerefBytesState == R.DerefBytesState &&
4177           this->GlobalState == R.GlobalState;
4178  }
4179
4180  /// Inequality for DerefState.
4181  bool operator!=(const DerefState &R) const { return !(*this == R); }
4182
4183  /// See IntegerStateBase::operator^=
4184  DerefState operator^=(const DerefState &R) {
4185    DerefBytesState ^= R.DerefBytesState;
4186    GlobalState ^= R.GlobalState;
4187    return *this;
4188  }
4189
4190  /// See IntegerStateBase::operator+=
4191  DerefState operator+=(const DerefState &R) {
4192    DerefBytesState += R.DerefBytesState;
4193    GlobalState += R.GlobalState;
4194    return *this;
4195  }
4196
4197  /// See IntegerStateBase::operator&=
4198  DerefState operator&=(const DerefState &R) {
4199    DerefBytesState &= R.DerefBytesState;
4200    GlobalState &= R.GlobalState;
4201    return *this;
4202  }
4203
4204  /// See IntegerStateBase::operator|=
4205  DerefState operator|=(const DerefState &R) {
4206    DerefBytesState |= R.DerefBytesState;
4207    GlobalState |= R.GlobalState;
4208    return *this;
4209  }
4210};
4211
4212/// An abstract interface for all dereferenceable attribute.
4213struct AADereferenceable
4214    : public IRAttribute<Attribute::Dereferenceable,
4215                         StateWrapper<DerefState, AbstractAttribute>,
4216                         AADereferenceable> {
4217  AADereferenceable(const IRPosition &IRP, Attributor &A) : IRAttribute(IRP) {}
4218
4219  /// See AbstractAttribute::isValidIRPositionForInit
4220  static bool isValidIRPositionForInit(Attributor &A, const IRPosition &IRP) {
4221    if (!IRP.getAssociatedType()->isPtrOrPtrVectorTy())
4222      return false;
4223    return IRAttribute::isValidIRPositionForInit(A, IRP);
4224  }
4225
4226  /// Return true if we assume that underlying value is
4227  /// dereferenceable(_or_null) globally.
4228  bool isAssumedGlobal() const { return GlobalState.getAssumed(); }
4229
4230  /// Return true if we know that underlying value is
4231  /// dereferenceable(_or_null) globally.
4232  bool isKnownGlobal() const { return GlobalState.getKnown(); }
4233
4234  /// Return assumed dereferenceable bytes.
4235  uint32_t getAssumedDereferenceableBytes() const {
4236    return DerefBytesState.getAssumed();
4237  }
4238
4239  /// Return known dereferenceable bytes.
4240  uint32_t getKnownDereferenceableBytes() const {
4241    return DerefBytesState.getKnown();
4242  }
4243
4244  /// Create an abstract attribute view for the position \p IRP.
4245  static AADereferenceable &createForPosition(const IRPosition &IRP,
4246                                              Attributor &A);
4247
4248  /// See AbstractAttribute::getName()
4249  const std::string getName() const override { return "AADereferenceable"; }
4250
4251  /// See AbstractAttribute::getIdAddr()
4252  const char *getIdAddr() const override { return &ID; }
4253
4254  /// This function should return true if the type of the \p AA is
4255  /// AADereferenceable
4256  static bool classof(const AbstractAttribute *AA) {
4257    return (AA->getIdAddr() == &ID);
4258  }
4259
4260  /// Unique ID (due to the unique address)
4261  static const char ID;
4262};
4263
4264using AAAlignmentStateType =
4265    IncIntegerState<uint64_t, Value::MaximumAlignment, 1>;
4266/// An abstract interface for all align attributes.
4267struct AAAlign
4268    : public IRAttribute<Attribute::Alignment,
4269                         StateWrapper<AAAlignmentStateType, AbstractAttribute>,
4270                         AAAlign> {
4271  AAAlign(const IRPosition &IRP, Attributor &A) : IRAttribute(IRP) {}
4272
4273  /// See AbstractAttribute::isValidIRPositionForInit
4274  static bool isValidIRPositionForInit(Attributor &A, const IRPosition &IRP) {
4275    if (!IRP.getAssociatedType()->isPtrOrPtrVectorTy())
4276      return false;
4277    return IRAttribute::isValidIRPositionForInit(A, IRP);
4278  }
4279
4280  /// Return assumed alignment.
4281  Align getAssumedAlign() const { return Align(getAssumed()); }
4282
4283  /// Return known alignment.
4284  Align getKnownAlign() const { return Align(getKnown()); }
4285
4286  /// See AbstractAttribute::getName()
4287  const std::string getName() const override { return "AAAlign"; }
4288
4289  /// See AbstractAttribute::getIdAddr()
4290  const char *getIdAddr() const override { return &ID; }
4291
4292  /// This function should return true if the type of the \p AA is AAAlign
4293  static bool classof(const AbstractAttribute *AA) {
4294    return (AA->getIdAddr() == &ID);
4295  }
4296
4297  /// Create an abstract attribute view for the position \p IRP.
4298  static AAAlign &createForPosition(const IRPosition &IRP, Attributor &A);
4299
4300  /// Unique ID (due to the unique address)
4301  static const char ID;
4302};
4303
4304/// An abstract interface to track if a value leaves it's defining function
4305/// instance.
4306/// TODO: We should make it a ternary AA tracking uniqueness, and uniqueness
4307/// wrt. the Attributor analysis separately.
4308struct AAInstanceInfo : public StateWrapper<BooleanState, AbstractAttribute> {
4309  AAInstanceInfo(const IRPosition &IRP, Attributor &A)
4310      : StateWrapper<BooleanState, AbstractAttribute>(IRP) {}
4311
4312  /// Return true if we know that the underlying value is unique in its scope
4313  /// wrt. the Attributor analysis. That means it might not be unique but we can
4314  /// still use pointer equality without risking to represent two instances with
4315  /// one `llvm::Value`.
4316  bool isKnownUniqueForAnalysis() const { return isKnown(); }
4317
4318  /// Return true if we assume that the underlying value is unique in its scope
4319  /// wrt. the Attributor analysis. That means it might not be unique but we can
4320  /// still use pointer equality without risking to represent two instances with
4321  /// one `llvm::Value`.
4322  bool isAssumedUniqueForAnalysis() const { return isAssumed(); }
4323
4324  /// Create an abstract attribute view for the position \p IRP.
4325  static AAInstanceInfo &createForPosition(const IRPosition &IRP,
4326                                           Attributor &A);
4327
4328  /// See AbstractAttribute::getName()
4329  const std::string getName() const override { return "AAInstanceInfo"; }
4330
4331  /// See AbstractAttribute::getIdAddr()
4332  const char *getIdAddr() const override { return &ID; }
4333
4334  /// This function should return true if the type of the \p AA is
4335  /// AAInstanceInfo
4336  static bool classof(const AbstractAttribute *AA) {
4337    return (AA->getIdAddr() == &ID);
4338  }
4339
4340  /// Unique ID (due to the unique address)
4341  static const char ID;
4342};
4343
4344/// An abstract interface for all nocapture attributes.
4345struct AANoCapture
4346    : public IRAttribute<
4347          Attribute::NoCapture,
4348          StateWrapper<BitIntegerState<uint16_t, 7, 0>, AbstractAttribute>,
4349          AANoCapture> {
4350  AANoCapture(const IRPosition &IRP, Attributor &A) : IRAttribute(IRP) {}
4351
4352  /// See IRAttribute::isImpliedByIR
4353  static bool isImpliedByIR(Attributor &A, const IRPosition &IRP,
4354                            Attribute::AttrKind ImpliedAttributeKind,
4355                            bool IgnoreSubsumingPositions = false);
4356
4357  /// Update \p State according to the capture capabilities of \p F for position
4358  /// \p IRP.
4359  static void determineFunctionCaptureCapabilities(const IRPosition &IRP,
4360                                                   const Function &F,
4361                                                   BitIntegerState &State);
4362
4363  /// See AbstractAttribute::isValidIRPositionForInit
4364  static bool isValidIRPositionForInit(Attributor &A, const IRPosition &IRP) {
4365    if (!IRP.getAssociatedType()->isPtrOrPtrVectorTy())
4366      return false;
4367    return IRAttribute::isValidIRPositionForInit(A, IRP);
4368  }
4369
4370  /// State encoding bits. A set bit in the state means the property holds.
4371  /// NO_CAPTURE is the best possible state, 0 the worst possible state.
4372  enum {
4373    NOT_CAPTURED_IN_MEM = 1 << 0,
4374    NOT_CAPTURED_IN_INT = 1 << 1,
4375    NOT_CAPTURED_IN_RET = 1 << 2,
4376
4377    /// If we do not capture the value in memory or through integers we can only
4378    /// communicate it back as a derived pointer.
4379    NO_CAPTURE_MAYBE_RETURNED = NOT_CAPTURED_IN_MEM | NOT_CAPTURED_IN_INT,
4380
4381    /// If we do not capture the value in memory, through integers, or as a
4382    /// derived pointer we know it is not captured.
4383    NO_CAPTURE =
4384        NOT_CAPTURED_IN_MEM | NOT_CAPTURED_IN_INT | NOT_CAPTURED_IN_RET,
4385  };
4386
4387  /// Return true if we know that the underlying value is not captured in its
4388  /// respective scope.
4389  bool isKnownNoCapture() const { return isKnown(NO_CAPTURE); }
4390
4391  /// Return true if we assume that the underlying value is not captured in its
4392  /// respective scope.
4393  bool isAssumedNoCapture() const { return isAssumed(NO_CAPTURE); }
4394
4395  /// Return true if we know that the underlying value is not captured in its
4396  /// respective scope but we allow it to escape through a "return".
4397  bool isKnownNoCaptureMaybeReturned() const {
4398    return isKnown(NO_CAPTURE_MAYBE_RETURNED);
4399  }
4400
4401  /// Return true if we assume that the underlying value is not captured in its
4402  /// respective scope but we allow it to escape through a "return".
4403  bool isAssumedNoCaptureMaybeReturned() const {
4404    return isAssumed(NO_CAPTURE_MAYBE_RETURNED);
4405  }
4406
4407  /// Create an abstract attribute view for the position \p IRP.
4408  static AANoCapture &createForPosition(const IRPosition &IRP, Attributor &A);
4409
4410  /// See AbstractAttribute::getName()
4411  const std::string getName() const override { return "AANoCapture"; }
4412
4413  /// See AbstractAttribute::getIdAddr()
4414  const char *getIdAddr() const override { return &ID; }
4415
4416  /// This function should return true if the type of the \p AA is AANoCapture
4417  static bool classof(const AbstractAttribute *AA) {
4418    return (AA->getIdAddr() == &ID);
4419  }
4420
4421  /// Unique ID (due to the unique address)
4422  static const char ID;
4423};
4424
4425struct ValueSimplifyStateType : public AbstractState {
4426
4427  ValueSimplifyStateType(Type *Ty) : Ty(Ty) {}
4428
4429  static ValueSimplifyStateType getBestState(Type *Ty) {
4430    return ValueSimplifyStateType(Ty);
4431  }
4432  static ValueSimplifyStateType getBestState(const ValueSimplifyStateType &VS) {
4433    return getBestState(VS.Ty);
4434  }
4435
4436  /// Return the worst possible representable state.
4437  static ValueSimplifyStateType getWorstState(Type *Ty) {
4438    ValueSimplifyStateType DS(Ty);
4439    DS.indicatePessimisticFixpoint();
4440    return DS;
4441  }
4442  static ValueSimplifyStateType
4443  getWorstState(const ValueSimplifyStateType &VS) {
4444    return getWorstState(VS.Ty);
4445  }
4446
4447  /// See AbstractState::isValidState(...)
4448  bool isValidState() const override { return BS.isValidState(); }
4449
4450  /// See AbstractState::isAtFixpoint(...)
4451  bool isAtFixpoint() const override { return BS.isAtFixpoint(); }
4452
4453  /// Return the assumed state encoding.
4454  ValueSimplifyStateType getAssumed() { return *this; }
4455  const ValueSimplifyStateType &getAssumed() const { return *this; }
4456
4457  /// See AbstractState::indicatePessimisticFixpoint(...)
4458  ChangeStatus indicatePessimisticFixpoint() override {
4459    return BS.indicatePessimisticFixpoint();
4460  }
4461
4462  /// See AbstractState::indicateOptimisticFixpoint(...)
4463  ChangeStatus indicateOptimisticFixpoint() override {
4464    return BS.indicateOptimisticFixpoint();
4465  }
4466
4467  /// "Clamp" this state with \p PVS.
4468  ValueSimplifyStateType operator^=(const ValueSimplifyStateType &VS) {
4469    BS ^= VS.BS;
4470    unionAssumed(VS.SimplifiedAssociatedValue);
4471    return *this;
4472  }
4473
4474  bool operator==(const ValueSimplifyStateType &RHS) const {
4475    if (isValidState() != RHS.isValidState())
4476      return false;
4477    if (!isValidState() && !RHS.isValidState())
4478      return true;
4479    return SimplifiedAssociatedValue == RHS.SimplifiedAssociatedValue;
4480  }
4481
4482protected:
4483  /// The type of the original value.
4484  Type *Ty;
4485
4486  /// Merge \p Other into the currently assumed simplified value
4487  bool unionAssumed(std::optional<Value *> Other);
4488
4489  /// Helper to track validity and fixpoint
4490  BooleanState BS;
4491
4492  /// An assumed simplified value. Initially, it is set to std::nullopt, which
4493  /// means that the value is not clear under current assumption. If in the
4494  /// pessimistic state, getAssumedSimplifiedValue doesn't return this value but
4495  /// returns orignal associated value.
4496  std::optional<Value *> SimplifiedAssociatedValue;
4497};
4498
4499/// An abstract interface for value simplify abstract attribute.
4500struct AAValueSimplify
4501    : public StateWrapper<ValueSimplifyStateType, AbstractAttribute, Type *> {
4502  using Base = StateWrapper<ValueSimplifyStateType, AbstractAttribute, Type *>;
4503  AAValueSimplify(const IRPosition &IRP, Attributor &A)
4504      : Base(IRP, IRP.getAssociatedType()) {}
4505
4506  /// Create an abstract attribute view for the position \p IRP.
4507  static AAValueSimplify &createForPosition(const IRPosition &IRP,
4508                                            Attributor &A);
4509
4510  /// See AbstractAttribute::getName()
4511  const std::string getName() const override { return "AAValueSimplify"; }
4512
4513  /// See AbstractAttribute::getIdAddr()
4514  const char *getIdAddr() const override { return &ID; }
4515
4516  /// This function should return true if the type of the \p AA is
4517  /// AAValueSimplify
4518  static bool classof(const AbstractAttribute *AA) {
4519    return (AA->getIdAddr() == &ID);
4520  }
4521
4522  /// Unique ID (due to the unique address)
4523  static const char ID;
4524
4525private:
4526  /// Return an assumed simplified value if a single candidate is found. If
4527  /// there cannot be one, return original value. If it is not clear yet, return
4528  /// std::nullopt.
4529  ///
4530  /// Use `Attributor::getAssumedSimplified` for value simplification.
4531  virtual std::optional<Value *>
4532  getAssumedSimplifiedValue(Attributor &A) const = 0;
4533
4534  friend struct Attributor;
4535};
4536
4537struct AAHeapToStack : public StateWrapper<BooleanState, AbstractAttribute> {
4538  using Base = StateWrapper<BooleanState, AbstractAttribute>;
4539  AAHeapToStack(const IRPosition &IRP, Attributor &A) : Base(IRP) {}
4540
4541  /// Returns true if HeapToStack conversion is assumed to be possible.
4542  virtual bool isAssumedHeapToStack(const CallBase &CB) const = 0;
4543
4544  /// Returns true if HeapToStack conversion is assumed and the CB is a
4545  /// callsite to a free operation to be removed.
4546  virtual bool isAssumedHeapToStackRemovedFree(CallBase &CB) const = 0;
4547
4548  /// Create an abstract attribute view for the position \p IRP.
4549  static AAHeapToStack &createForPosition(const IRPosition &IRP, Attributor &A);
4550
4551  /// See AbstractAttribute::getName()
4552  const std::string getName() const override { return "AAHeapToStack"; }
4553
4554  /// See AbstractAttribute::getIdAddr()
4555  const char *getIdAddr() const override { return &ID; }
4556
4557  /// This function should return true if the type of the \p AA is AAHeapToStack
4558  static bool classof(const AbstractAttribute *AA) {
4559    return (AA->getIdAddr() == &ID);
4560  }
4561
4562  /// Unique ID (due to the unique address)
4563  static const char ID;
4564};
4565
4566/// An abstract interface for privatizability.
4567///
4568/// A pointer is privatizable if it can be replaced by a new, private one.
4569/// Privatizing pointer reduces the use count, interaction between unrelated
4570/// code parts.
4571///
4572/// In order for a pointer to be privatizable its value cannot be observed
4573/// (=nocapture), it is (for now) not written (=readonly & noalias), we know
4574/// what values are necessary to make the private copy look like the original
4575/// one, and the values we need can be loaded (=dereferenceable).
4576struct AAPrivatizablePtr
4577    : public StateWrapper<BooleanState, AbstractAttribute> {
4578  using Base = StateWrapper<BooleanState, AbstractAttribute>;
4579  AAPrivatizablePtr(const IRPosition &IRP, Attributor &A) : Base(IRP) {}
4580
4581  /// See AbstractAttribute::isValidIRPositionForInit
4582  static bool isValidIRPositionForInit(Attributor &A, const IRPosition &IRP) {
4583    if (!IRP.getAssociatedType()->isPtrOrPtrVectorTy())
4584      return false;
4585    return AbstractAttribute::isValidIRPositionForInit(A, IRP);
4586  }
4587
4588  /// Returns true if pointer privatization is assumed to be possible.
4589  bool isAssumedPrivatizablePtr() const { return getAssumed(); }
4590
4591  /// Returns true if pointer privatization is known to be possible.
4592  bool isKnownPrivatizablePtr() const { return getKnown(); }
4593
4594  /// See AbstractAttribute::requiresCallersForArgOrFunction
4595  static bool requiresCallersForArgOrFunction() { return true; }
4596
4597  /// Return the type we can choose for a private copy of the underlying
4598  /// value. std::nullopt means it is not clear yet, nullptr means there is
4599  /// none.
4600  virtual std::optional<Type *> getPrivatizableType() const = 0;
4601
4602  /// Create an abstract attribute view for the position \p IRP.
4603  static AAPrivatizablePtr &createForPosition(const IRPosition &IRP,
4604                                              Attributor &A);
4605
4606  /// See AbstractAttribute::getName()
4607  const std::string getName() const override { return "AAPrivatizablePtr"; }
4608
4609  /// See AbstractAttribute::getIdAddr()
4610  const char *getIdAddr() const override { return &ID; }
4611
4612  /// This function should return true if the type of the \p AA is
4613  /// AAPricatizablePtr
4614  static bool classof(const AbstractAttribute *AA) {
4615    return (AA->getIdAddr() == &ID);
4616  }
4617
4618  /// Unique ID (due to the unique address)
4619  static const char ID;
4620};
4621
4622/// An abstract interface for memory access kind related attributes
4623/// (readnone/readonly/writeonly).
4624struct AAMemoryBehavior
4625    : public IRAttribute<
4626          Attribute::None,
4627          StateWrapper<BitIntegerState<uint8_t, 3>, AbstractAttribute>,
4628          AAMemoryBehavior> {
4629  AAMemoryBehavior(const IRPosition &IRP, Attributor &A) : IRAttribute(IRP) {}
4630
4631  /// See AbstractAttribute::hasTrivialInitializer.
4632  static bool hasTrivialInitializer() { return false; }
4633
4634  /// See AbstractAttribute::isValidIRPositionForInit
4635  static bool isValidIRPositionForInit(Attributor &A, const IRPosition &IRP) {
4636    if (!IRP.isFunctionScope() &&
4637        !IRP.getAssociatedType()->isPtrOrPtrVectorTy())
4638      return false;
4639    return IRAttribute::isValidIRPositionForInit(A, IRP);
4640  }
4641
4642  /// State encoding bits. A set bit in the state means the property holds.
4643  /// BEST_STATE is the best possible state, 0 the worst possible state.
4644  enum {
4645    NO_READS = 1 << 0,
4646    NO_WRITES = 1 << 1,
4647    NO_ACCESSES = NO_READS | NO_WRITES,
4648
4649    BEST_STATE = NO_ACCESSES,
4650  };
4651  static_assert(BEST_STATE == getBestState(), "Unexpected BEST_STATE value");
4652
4653  /// Return true if we know that the underlying value is not read or accessed
4654  /// in its respective scope.
4655  bool isKnownReadNone() const { return isKnown(NO_ACCESSES); }
4656
4657  /// Return true if we assume that the underlying value is not read or accessed
4658  /// in its respective scope.
4659  bool isAssumedReadNone() const { return isAssumed(NO_ACCESSES); }
4660
4661  /// Return true if we know that the underlying value is not accessed
4662  /// (=written) in its respective scope.
4663  bool isKnownReadOnly() const { return isKnown(NO_WRITES); }
4664
4665  /// Return true if we assume that the underlying value is not accessed
4666  /// (=written) in its respective scope.
4667  bool isAssumedReadOnly() const { return isAssumed(NO_WRITES); }
4668
4669  /// Return true if we know that the underlying value is not read in its
4670  /// respective scope.
4671  bool isKnownWriteOnly() const { return isKnown(NO_READS); }
4672
4673  /// Return true if we assume that the underlying value is not read in its
4674  /// respective scope.
4675  bool isAssumedWriteOnly() const { return isAssumed(NO_READS); }
4676
4677  /// Create an abstract attribute view for the position \p IRP.
4678  static AAMemoryBehavior &createForPosition(const IRPosition &IRP,
4679                                             Attributor &A);
4680
4681  /// See AbstractAttribute::getName()
4682  const std::string getName() const override { return "AAMemoryBehavior"; }
4683
4684  /// See AbstractAttribute::getIdAddr()
4685  const char *getIdAddr() const override { return &ID; }
4686
4687  /// This function should return true if the type of the \p AA is
4688  /// AAMemoryBehavior
4689  static bool classof(const AbstractAttribute *AA) {
4690    return (AA->getIdAddr() == &ID);
4691  }
4692
4693  /// Unique ID (due to the unique address)
4694  static const char ID;
4695};
4696
4697/// An abstract interface for all memory location attributes
4698/// (readnone/argmemonly/inaccessiblememonly/inaccessibleorargmemonly).
4699struct AAMemoryLocation
4700    : public IRAttribute<
4701          Attribute::None,
4702          StateWrapper<BitIntegerState<uint32_t, 511>, AbstractAttribute>,
4703          AAMemoryLocation> {
4704  using MemoryLocationsKind = StateType::base_t;
4705
4706  AAMemoryLocation(const IRPosition &IRP, Attributor &A) : IRAttribute(IRP) {}
4707
4708  /// See AbstractAttribute::requiresCalleeForCallBase.
4709  static bool requiresCalleeForCallBase() { return true; }
4710
4711  /// See AbstractAttribute::hasTrivialInitializer.
4712  static bool hasTrivialInitializer() { return false; }
4713
4714  /// See AbstractAttribute::isValidIRPositionForInit
4715  static bool isValidIRPositionForInit(Attributor &A, const IRPosition &IRP) {
4716    if (!IRP.isFunctionScope() &&
4717        !IRP.getAssociatedType()->isPtrOrPtrVectorTy())
4718      return false;
4719    return IRAttribute::isValidIRPositionForInit(A, IRP);
4720  }
4721
4722  /// Encoding of different locations that could be accessed by a memory
4723  /// access.
4724  enum {
4725    ALL_LOCATIONS = 0,
4726    NO_LOCAL_MEM = 1 << 0,
4727    NO_CONST_MEM = 1 << 1,
4728    NO_GLOBAL_INTERNAL_MEM = 1 << 2,
4729    NO_GLOBAL_EXTERNAL_MEM = 1 << 3,
4730    NO_GLOBAL_MEM = NO_GLOBAL_INTERNAL_MEM | NO_GLOBAL_EXTERNAL_MEM,
4731    NO_ARGUMENT_MEM = 1 << 4,
4732    NO_INACCESSIBLE_MEM = 1 << 5,
4733    NO_MALLOCED_MEM = 1 << 6,
4734    NO_UNKOWN_MEM = 1 << 7,
4735    NO_LOCATIONS = NO_LOCAL_MEM | NO_CONST_MEM | NO_GLOBAL_INTERNAL_MEM |
4736                   NO_GLOBAL_EXTERNAL_MEM | NO_ARGUMENT_MEM |
4737                   NO_INACCESSIBLE_MEM | NO_MALLOCED_MEM | NO_UNKOWN_MEM,
4738
4739    // Helper bit to track if we gave up or not.
4740    VALID_STATE = NO_LOCATIONS + 1,
4741
4742    BEST_STATE = NO_LOCATIONS | VALID_STATE,
4743  };
4744  static_assert(BEST_STATE == getBestState(), "Unexpected BEST_STATE value");
4745
4746  /// Return true if we know that the associated functions has no observable
4747  /// accesses.
4748  bool isKnownReadNone() const { return isKnown(NO_LOCATIONS); }
4749
4750  /// Return true if we assume that the associated functions has no observable
4751  /// accesses.
4752  bool isAssumedReadNone() const {
4753    return isAssumed(NO_LOCATIONS) || isAssumedStackOnly();
4754  }
4755
4756  /// Return true if we know that the associated functions has at most
4757  /// local/stack accesses.
4758  bool isKnowStackOnly() const {
4759    return isKnown(inverseLocation(NO_LOCAL_MEM, true, true));
4760  }
4761
4762  /// Return true if we assume that the associated functions has at most
4763  /// local/stack accesses.
4764  bool isAssumedStackOnly() const {
4765    return isAssumed(inverseLocation(NO_LOCAL_MEM, true, true));
4766  }
4767
4768  /// Return true if we know that the underlying value will only access
4769  /// inaccesible memory only (see Attribute::InaccessibleMemOnly).
4770  bool isKnownInaccessibleMemOnly() const {
4771    return isKnown(inverseLocation(NO_INACCESSIBLE_MEM, true, true));
4772  }
4773
4774  /// Return true if we assume that the underlying value will only access
4775  /// inaccesible memory only (see Attribute::InaccessibleMemOnly).
4776  bool isAssumedInaccessibleMemOnly() const {
4777    return isAssumed(inverseLocation(NO_INACCESSIBLE_MEM, true, true));
4778  }
4779
4780  /// Return true if we know that the underlying value will only access
4781  /// argument pointees (see Attribute::ArgMemOnly).
4782  bool isKnownArgMemOnly() const {
4783    return isKnown(inverseLocation(NO_ARGUMENT_MEM, true, true));
4784  }
4785
4786  /// Return true if we assume that the underlying value will only access
4787  /// argument pointees (see Attribute::ArgMemOnly).
4788  bool isAssumedArgMemOnly() const {
4789    return isAssumed(inverseLocation(NO_ARGUMENT_MEM, true, true));
4790  }
4791
4792  /// Return true if we know that the underlying value will only access
4793  /// inaccesible memory or argument pointees (see
4794  /// Attribute::InaccessibleOrArgMemOnly).
4795  bool isKnownInaccessibleOrArgMemOnly() const {
4796    return isKnown(
4797        inverseLocation(NO_INACCESSIBLE_MEM | NO_ARGUMENT_MEM, true, true));
4798  }
4799
4800  /// Return true if we assume that the underlying value will only access
4801  /// inaccesible memory or argument pointees (see
4802  /// Attribute::InaccessibleOrArgMemOnly).
4803  bool isAssumedInaccessibleOrArgMemOnly() const {
4804    return isAssumed(
4805        inverseLocation(NO_INACCESSIBLE_MEM | NO_ARGUMENT_MEM, true, true));
4806  }
4807
4808  /// Return true if the underlying value may access memory through arguement
4809  /// pointers of the associated function, if any.
4810  bool mayAccessArgMem() const { return !isAssumed(NO_ARGUMENT_MEM); }
4811
4812  /// Return true if only the memory locations specififed by \p MLK are assumed
4813  /// to be accessed by the associated function.
4814  bool isAssumedSpecifiedMemOnly(MemoryLocationsKind MLK) const {
4815    return isAssumed(MLK);
4816  }
4817
4818  /// Return the locations that are assumed to be not accessed by the associated
4819  /// function, if any.
4820  MemoryLocationsKind getAssumedNotAccessedLocation() const {
4821    return getAssumed();
4822  }
4823
4824  /// Return the inverse of location \p Loc, thus for NO_XXX the return
4825  /// describes ONLY_XXX. The flags \p AndLocalMem and \p AndConstMem determine
4826  /// if local (=stack) and constant memory are allowed as well. Most of the
4827  /// time we do want them to be included, e.g., argmemonly allows accesses via
4828  /// argument pointers or local or constant memory accesses.
4829  static MemoryLocationsKind
4830  inverseLocation(MemoryLocationsKind Loc, bool AndLocalMem, bool AndConstMem) {
4831    return NO_LOCATIONS & ~(Loc | (AndLocalMem ? NO_LOCAL_MEM : 0) |
4832                            (AndConstMem ? NO_CONST_MEM : 0));
4833  };
4834
4835  /// Return the locations encoded by \p MLK as a readable string.
4836  static std::string getMemoryLocationsAsStr(MemoryLocationsKind MLK);
4837
4838  /// Simple enum to distinguish read/write/read-write accesses.
4839  enum AccessKind {
4840    NONE = 0,
4841    READ = 1 << 0,
4842    WRITE = 1 << 1,
4843    READ_WRITE = READ | WRITE,
4844  };
4845
4846  /// Check \p Pred on all accesses to the memory kinds specified by \p MLK.
4847  ///
4848  /// This method will evaluate \p Pred on all accesses (access instruction +
4849  /// underlying accessed memory pointer) and it will return true if \p Pred
4850  /// holds every time.
4851  virtual bool checkForAllAccessesToMemoryKind(
4852      function_ref<bool(const Instruction *, const Value *, AccessKind,
4853                        MemoryLocationsKind)>
4854          Pred,
4855      MemoryLocationsKind MLK) const = 0;
4856
4857  /// Create an abstract attribute view for the position \p IRP.
4858  static AAMemoryLocation &createForPosition(const IRPosition &IRP,
4859                                             Attributor &A);
4860
4861  /// See AbstractState::getAsStr(Attributor).
4862  const std::string getAsStr(Attributor *A) const override {
4863    return getMemoryLocationsAsStr(getAssumedNotAccessedLocation());
4864  }
4865
4866  /// See AbstractAttribute::getName()
4867  const std::string getName() const override { return "AAMemoryLocation"; }
4868
4869  /// See AbstractAttribute::getIdAddr()
4870  const char *getIdAddr() const override { return &ID; }
4871
4872  /// This function should return true if the type of the \p AA is
4873  /// AAMemoryLocation
4874  static bool classof(const AbstractAttribute *AA) {
4875    return (AA->getIdAddr() == &ID);
4876  }
4877
4878  /// Unique ID (due to the unique address)
4879  static const char ID;
4880};
4881
4882/// An abstract interface for range value analysis.
4883struct AAValueConstantRange
4884    : public StateWrapper<IntegerRangeState, AbstractAttribute, uint32_t> {
4885  using Base = StateWrapper<IntegerRangeState, AbstractAttribute, uint32_t>;
4886  AAValueConstantRange(const IRPosition &IRP, Attributor &A)
4887      : Base(IRP, IRP.getAssociatedType()->getIntegerBitWidth()) {}
4888
4889  /// See AbstractAttribute::isValidIRPositionForInit
4890  static bool isValidIRPositionForInit(Attributor &A, const IRPosition &IRP) {
4891    if (!IRP.getAssociatedType()->isIntegerTy())
4892      return false;
4893    return AbstractAttribute::isValidIRPositionForInit(A, IRP);
4894  }
4895
4896  /// See AbstractAttribute::requiresCallersForArgOrFunction
4897  static bool requiresCallersForArgOrFunction() { return true; }
4898
4899  /// See AbstractAttribute::getState(...).
4900  IntegerRangeState &getState() override { return *this; }
4901  const IntegerRangeState &getState() const override { return *this; }
4902
4903  /// Create an abstract attribute view for the position \p IRP.
4904  static AAValueConstantRange &createForPosition(const IRPosition &IRP,
4905                                                 Attributor &A);
4906
4907  /// Return an assumed range for the associated value a program point \p CtxI.
4908  /// If \p I is nullptr, simply return an assumed range.
4909  virtual ConstantRange
4910  getAssumedConstantRange(Attributor &A,
4911                          const Instruction *CtxI = nullptr) const = 0;
4912
4913  /// Return a known range for the associated value at a program point \p CtxI.
4914  /// If \p I is nullptr, simply return a known range.
4915  virtual ConstantRange
4916  getKnownConstantRange(Attributor &A,
4917                        const Instruction *CtxI = nullptr) const = 0;
4918
4919  /// Return an assumed constant for the associated value a program point \p
4920  /// CtxI.
4921  std::optional<Constant *>
4922  getAssumedConstant(Attributor &A, const Instruction *CtxI = nullptr) const {
4923    ConstantRange RangeV = getAssumedConstantRange(A, CtxI);
4924    if (auto *C = RangeV.getSingleElement()) {
4925      Type *Ty = getAssociatedValue().getType();
4926      return cast_or_null<Constant>(
4927          AA::getWithType(*ConstantInt::get(Ty->getContext(), *C), *Ty));
4928    }
4929    if (RangeV.isEmptySet())
4930      return std::nullopt;
4931    return nullptr;
4932  }
4933
4934  /// See AbstractAttribute::getName()
4935  const std::string getName() const override { return "AAValueConstantRange"; }
4936
4937  /// See AbstractAttribute::getIdAddr()
4938  const char *getIdAddr() const override { return &ID; }
4939
4940  /// This function should return true if the type of the \p AA is
4941  /// AAValueConstantRange
4942  static bool classof(const AbstractAttribute *AA) {
4943    return (AA->getIdAddr() == &ID);
4944  }
4945
4946  /// Unique ID (due to the unique address)
4947  static const char ID;
4948};
4949
4950/// A class for a set state.
4951/// The assumed boolean state indicates whether the corresponding set is full
4952/// set or not. If the assumed state is false, this is the worst state. The
4953/// worst state (invalid state) of set of potential values is when the set
4954/// contains every possible value (i.e. we cannot in any way limit the value
4955/// that the target position can take). That never happens naturally, we only
4956/// force it. As for the conditions under which we force it, see
4957/// AAPotentialConstantValues.
4958template <typename MemberTy> struct PotentialValuesState : AbstractState {
4959  using SetTy = SmallSetVector<MemberTy, 8>;
4960
4961  PotentialValuesState() : IsValidState(true), UndefIsContained(false) {}
4962
4963  PotentialValuesState(bool IsValid)
4964      : IsValidState(IsValid), UndefIsContained(false) {}
4965
4966  /// See AbstractState::isValidState(...)
4967  bool isValidState() const override { return IsValidState.isValidState(); }
4968
4969  /// See AbstractState::isAtFixpoint(...)
4970  bool isAtFixpoint() const override { return IsValidState.isAtFixpoint(); }
4971
4972  /// See AbstractState::indicatePessimisticFixpoint(...)
4973  ChangeStatus indicatePessimisticFixpoint() override {
4974    return IsValidState.indicatePessimisticFixpoint();
4975  }
4976
4977  /// See AbstractState::indicateOptimisticFixpoint(...)
4978  ChangeStatus indicateOptimisticFixpoint() override {
4979    return IsValidState.indicateOptimisticFixpoint();
4980  }
4981
4982  /// Return the assumed state
4983  PotentialValuesState &getAssumed() { return *this; }
4984  const PotentialValuesState &getAssumed() const { return *this; }
4985
4986  /// Return this set. We should check whether this set is valid or not by
4987  /// isValidState() before calling this function.
4988  const SetTy &getAssumedSet() const {
4989    assert(isValidState() && "This set shoud not be used when it is invalid!");
4990    return Set;
4991  }
4992
4993  /// Returns whether this state contains an undef value or not.
4994  bool undefIsContained() const {
4995    assert(isValidState() && "This flag shoud not be used when it is invalid!");
4996    return UndefIsContained;
4997  }
4998
4999  bool operator==(const PotentialValuesState &RHS) const {
5000    if (isValidState() != RHS.isValidState())
5001      return false;
5002    if (!isValidState() && !RHS.isValidState())
5003      return true;
5004    if (undefIsContained() != RHS.undefIsContained())
5005      return false;
5006    return Set == RHS.getAssumedSet();
5007  }
5008
5009  /// Maximum number of potential values to be tracked.
5010  /// This is set by -attributor-max-potential-values command line option
5011  static unsigned MaxPotentialValues;
5012
5013  /// Return empty set as the best state of potential values.
5014  static PotentialValuesState getBestState() {
5015    return PotentialValuesState(true);
5016  }
5017
5018  static PotentialValuesState getBestState(const PotentialValuesState &PVS) {
5019    return getBestState();
5020  }
5021
5022  /// Return full set as the worst state of potential values.
5023  static PotentialValuesState getWorstState() {
5024    return PotentialValuesState(false);
5025  }
5026
5027  /// Union assumed set with the passed value.
5028  void unionAssumed(const MemberTy &C) { insert(C); }
5029
5030  /// Union assumed set with assumed set of the passed state \p PVS.
5031  void unionAssumed(const PotentialValuesState &PVS) { unionWith(PVS); }
5032
5033  /// Union assumed set with an undef value.
5034  void unionAssumedWithUndef() { unionWithUndef(); }
5035
5036  /// "Clamp" this state with \p PVS.
5037  PotentialValuesState operator^=(const PotentialValuesState &PVS) {
5038    IsValidState ^= PVS.IsValidState;
5039    unionAssumed(PVS);
5040    return *this;
5041  }
5042
5043  PotentialValuesState operator&=(const PotentialValuesState &PVS) {
5044    IsValidState &= PVS.IsValidState;
5045    unionAssumed(PVS);
5046    return *this;
5047  }
5048
5049  bool contains(const MemberTy &V) const {
5050    return !isValidState() ? true : Set.contains(V);
5051  }
5052
5053protected:
5054  SetTy &getAssumedSet() {
5055    assert(isValidState() && "This set shoud not be used when it is invalid!");
5056    return Set;
5057  }
5058
5059private:
5060  /// Check the size of this set, and invalidate when the size is no
5061  /// less than \p MaxPotentialValues threshold.
5062  void checkAndInvalidate() {
5063    if (Set.size() >= MaxPotentialValues)
5064      indicatePessimisticFixpoint();
5065    else
5066      reduceUndefValue();
5067  }
5068
5069  /// If this state contains both undef and not undef, we can reduce
5070  /// undef to the not undef value.
5071  void reduceUndefValue() { UndefIsContained = UndefIsContained & Set.empty(); }
5072
5073  /// Insert an element into this set.
5074  void insert(const MemberTy &C) {
5075    if (!isValidState())
5076      return;
5077    Set.insert(C);
5078    checkAndInvalidate();
5079  }
5080
5081  /// Take union with R.
5082  void unionWith(const PotentialValuesState &R) {
5083    /// If this is a full set, do nothing.
5084    if (!isValidState())
5085      return;
5086    /// If R is full set, change L to a full set.
5087    if (!R.isValidState()) {
5088      indicatePessimisticFixpoint();
5089      return;
5090    }
5091    for (const MemberTy &C : R.Set)
5092      Set.insert(C);
5093    UndefIsContained |= R.undefIsContained();
5094    checkAndInvalidate();
5095  }
5096
5097  /// Take union with an undef value.
5098  void unionWithUndef() {
5099    UndefIsContained = true;
5100    reduceUndefValue();
5101  }
5102
5103  /// Take intersection with R.
5104  void intersectWith(const PotentialValuesState &R) {
5105    /// If R is a full set, do nothing.
5106    if (!R.isValidState())
5107      return;
5108    /// If this is a full set, change this to R.
5109    if (!isValidState()) {
5110      *this = R;
5111      return;
5112    }
5113    SetTy IntersectSet;
5114    for (const MemberTy &C : Set) {
5115      if (R.Set.count(C))
5116        IntersectSet.insert(C);
5117    }
5118    Set = IntersectSet;
5119    UndefIsContained &= R.undefIsContained();
5120    reduceUndefValue();
5121  }
5122
5123  /// A helper state which indicate whether this state is valid or not.
5124  BooleanState IsValidState;
5125
5126  /// Container for potential values
5127  SetTy Set;
5128
5129  /// Flag for undef value
5130  bool UndefIsContained;
5131};
5132
5133struct DenormalFPMathState : public AbstractState {
5134  struct DenormalState {
5135    DenormalMode Mode = DenormalMode::getInvalid();
5136    DenormalMode ModeF32 = DenormalMode::getInvalid();
5137
5138    bool operator==(const DenormalState Other) const {
5139      return Mode == Other.Mode && ModeF32 == Other.ModeF32;
5140    }
5141
5142    bool operator!=(const DenormalState Other) const {
5143      return Mode != Other.Mode || ModeF32 != Other.ModeF32;
5144    }
5145
5146    bool isValid() const {
5147      return Mode.isValid() && ModeF32.isValid();
5148    }
5149
5150    static DenormalMode::DenormalModeKind
5151    unionDenormalKind(DenormalMode::DenormalModeKind Callee,
5152                      DenormalMode::DenormalModeKind Caller) {
5153      if (Caller == Callee)
5154        return Caller;
5155      if (Callee == DenormalMode::Dynamic)
5156        return Caller;
5157      if (Caller == DenormalMode::Dynamic)
5158        return Callee;
5159      return DenormalMode::Invalid;
5160    }
5161
5162    static DenormalMode unionAssumed(DenormalMode Callee, DenormalMode Caller) {
5163      return DenormalMode{unionDenormalKind(Callee.Output, Caller.Output),
5164                          unionDenormalKind(Callee.Input, Caller.Input)};
5165    }
5166
5167    DenormalState unionWith(DenormalState Caller) const {
5168      DenormalState Callee(*this);
5169      Callee.Mode = unionAssumed(Callee.Mode, Caller.Mode);
5170      Callee.ModeF32 = unionAssumed(Callee.ModeF32, Caller.ModeF32);
5171      return Callee;
5172    }
5173  };
5174
5175  DenormalState Known;
5176
5177  /// Explicitly track whether we've hit a fixed point.
5178  bool IsAtFixedpoint = false;
5179
5180  DenormalFPMathState() = default;
5181
5182  DenormalState getKnown() const { return Known; }
5183
5184  // There's only really known or unknown, there's no speculatively assumable
5185  // state.
5186  DenormalState getAssumed() const { return Known; }
5187
5188  bool isValidState() const override {
5189    return Known.isValid();
5190  }
5191
5192  /// Return true if there are no dynamic components to the denormal mode worth
5193  /// specializing.
5194  bool isModeFixed() const {
5195    return Known.Mode.Input != DenormalMode::Dynamic &&
5196           Known.Mode.Output != DenormalMode::Dynamic &&
5197           Known.ModeF32.Input != DenormalMode::Dynamic &&
5198           Known.ModeF32.Output != DenormalMode::Dynamic;
5199  }
5200
5201  bool isAtFixpoint() const override {
5202    return IsAtFixedpoint;
5203  }
5204
5205  ChangeStatus indicateFixpoint() {
5206    bool Changed = !IsAtFixedpoint;
5207    IsAtFixedpoint = true;
5208    return Changed ? ChangeStatus::CHANGED : ChangeStatus::UNCHANGED;
5209  }
5210
5211  ChangeStatus indicateOptimisticFixpoint() override {
5212    return indicateFixpoint();
5213  }
5214
5215  ChangeStatus indicatePessimisticFixpoint() override {
5216    return indicateFixpoint();
5217  }
5218
5219  DenormalFPMathState operator^=(const DenormalFPMathState &Caller) {
5220    Known = Known.unionWith(Caller.getKnown());
5221    return *this;
5222  }
5223};
5224
5225using PotentialConstantIntValuesState = PotentialValuesState<APInt>;
5226using PotentialLLVMValuesState =
5227    PotentialValuesState<std::pair<AA::ValueAndContext, AA::ValueScope>>;
5228
5229raw_ostream &operator<<(raw_ostream &OS,
5230                        const PotentialConstantIntValuesState &R);
5231raw_ostream &operator<<(raw_ostream &OS, const PotentialLLVMValuesState &R);
5232
5233/// An abstract interface for potential values analysis.
5234///
5235/// This AA collects potential values for each IR position.
5236/// An assumed set of potential values is initialized with the empty set (the
5237/// best state) and it will grow monotonically as we find more potential values
5238/// for this position.
5239/// The set might be forced to the worst state, that is, to contain every
5240/// possible value for this position in 2 cases.
5241///   1. We surpassed the \p MaxPotentialValues threshold. This includes the
5242///      case that this position is affected (e.g. because of an operation) by a
5243///      Value that is in the worst state.
5244///   2. We tried to initialize on a Value that we cannot handle (e.g. an
5245///      operator we do not currently handle).
5246///
5247/// For non constant integers see AAPotentialValues.
5248struct AAPotentialConstantValues
5249    : public StateWrapper<PotentialConstantIntValuesState, AbstractAttribute> {
5250  using Base = StateWrapper<PotentialConstantIntValuesState, AbstractAttribute>;
5251  AAPotentialConstantValues(const IRPosition &IRP, Attributor &A) : Base(IRP) {}
5252
5253  /// See AbstractAttribute::isValidIRPositionForInit
5254  static bool isValidIRPositionForInit(Attributor &A, const IRPosition &IRP) {
5255    if (!IRP.getAssociatedType()->isIntegerTy())
5256      return false;
5257    return AbstractAttribute::isValidIRPositionForInit(A, IRP);
5258  }
5259
5260  /// See AbstractAttribute::requiresCallersForArgOrFunction
5261  static bool requiresCallersForArgOrFunction() { return true; }
5262
5263  /// See AbstractAttribute::getState(...).
5264  PotentialConstantIntValuesState &getState() override { return *this; }
5265  const PotentialConstantIntValuesState &getState() const override {
5266    return *this;
5267  }
5268
5269  /// Create an abstract attribute view for the position \p IRP.
5270  static AAPotentialConstantValues &createForPosition(const IRPosition &IRP,
5271                                                      Attributor &A);
5272
5273  /// Return assumed constant for the associated value
5274  std::optional<Constant *>
5275  getAssumedConstant(Attributor &A, const Instruction *CtxI = nullptr) const {
5276    if (!isValidState())
5277      return nullptr;
5278    if (getAssumedSet().size() == 1) {
5279      Type *Ty = getAssociatedValue().getType();
5280      return cast_or_null<Constant>(AA::getWithType(
5281          *ConstantInt::get(Ty->getContext(), *(getAssumedSet().begin())),
5282          *Ty));
5283    }
5284    if (getAssumedSet().size() == 0) {
5285      if (undefIsContained())
5286        return UndefValue::get(getAssociatedValue().getType());
5287      return std::nullopt;
5288    }
5289
5290    return nullptr;
5291  }
5292
5293  /// See AbstractAttribute::getName()
5294  const std::string getName() const override {
5295    return "AAPotentialConstantValues";
5296  }
5297
5298  /// See AbstractAttribute::getIdAddr()
5299  const char *getIdAddr() const override { return &ID; }
5300
5301  /// This function should return true if the type of the \p AA is
5302  /// AAPotentialConstantValues
5303  static bool classof(const AbstractAttribute *AA) {
5304    return (AA->getIdAddr() == &ID);
5305  }
5306
5307  /// Unique ID (due to the unique address)
5308  static const char ID;
5309};
5310
5311struct AAPotentialValues
5312    : public StateWrapper<PotentialLLVMValuesState, AbstractAttribute> {
5313  using Base = StateWrapper<PotentialLLVMValuesState, AbstractAttribute>;
5314  AAPotentialValues(const IRPosition &IRP, Attributor &A) : Base(IRP) {}
5315
5316  /// See AbstractAttribute::requiresCallersForArgOrFunction
5317  static bool requiresCallersForArgOrFunction() { return true; }
5318
5319  /// See AbstractAttribute::getState(...).
5320  PotentialLLVMValuesState &getState() override { return *this; }
5321  const PotentialLLVMValuesState &getState() const override { return *this; }
5322
5323  /// Create an abstract attribute view for the position \p IRP.
5324  static AAPotentialValues &createForPosition(const IRPosition &IRP,
5325                                              Attributor &A);
5326
5327  /// Extract the single value in \p Values if any.
5328  static Value *getSingleValue(Attributor &A, const AbstractAttribute &AA,
5329                               const IRPosition &IRP,
5330                               SmallVectorImpl<AA::ValueAndContext> &Values);
5331
5332  /// See AbstractAttribute::getName()
5333  const std::string getName() const override { return "AAPotentialValues"; }
5334
5335  /// See AbstractAttribute::getIdAddr()
5336  const char *getIdAddr() const override { return &ID; }
5337
5338  /// This function should return true if the type of the \p AA is
5339  /// AAPotentialValues
5340  static bool classof(const AbstractAttribute *AA) {
5341    return (AA->getIdAddr() == &ID);
5342  }
5343
5344  /// Unique ID (due to the unique address)
5345  static const char ID;
5346
5347private:
5348  virtual bool getAssumedSimplifiedValues(
5349      Attributor &A, SmallVectorImpl<AA::ValueAndContext> &Values,
5350      AA::ValueScope, bool RecurseForSelectAndPHI = false) const = 0;
5351
5352  friend struct Attributor;
5353};
5354
5355/// An abstract interface for all noundef attributes.
5356struct AANoUndef
5357    : public IRAttribute<Attribute::NoUndef,
5358                         StateWrapper<BooleanState, AbstractAttribute>,
5359                         AANoUndef> {
5360  AANoUndef(const IRPosition &IRP, Attributor &A) : IRAttribute(IRP) {}
5361
5362  /// See IRAttribute::isImpliedByUndef
5363  static bool isImpliedByUndef() { return false; }
5364
5365  /// See IRAttribute::isImpliedByPoison
5366  static bool isImpliedByPoison() { return false; }
5367
5368  /// See IRAttribute::isImpliedByIR
5369  static bool isImpliedByIR(Attributor &A, const IRPosition &IRP,
5370                            Attribute::AttrKind ImpliedAttributeKind,
5371                            bool IgnoreSubsumingPositions = false);
5372
5373  /// Return true if we assume that the underlying value is noundef.
5374  bool isAssumedNoUndef() const { return getAssumed(); }
5375
5376  /// Return true if we know that underlying value is noundef.
5377  bool isKnownNoUndef() const { return getKnown(); }
5378
5379  /// Create an abstract attribute view for the position \p IRP.
5380  static AANoUndef &createForPosition(const IRPosition &IRP, Attributor &A);
5381
5382  /// See AbstractAttribute::getName()
5383  const std::string getName() const override { return "AANoUndef"; }
5384
5385  /// See AbstractAttribute::getIdAddr()
5386  const char *getIdAddr() const override { return &ID; }
5387
5388  /// This function should return true if the type of the \p AA is AANoUndef
5389  static bool classof(const AbstractAttribute *AA) {
5390    return (AA->getIdAddr() == &ID);
5391  }
5392
5393  /// Unique ID (due to the unique address)
5394  static const char ID;
5395};
5396
5397struct AANoFPClass
5398    : public IRAttribute<
5399          Attribute::NoFPClass,
5400          StateWrapper<BitIntegerState<uint32_t, fcAllFlags, fcNone>,
5401                       AbstractAttribute>,
5402          AANoFPClass> {
5403  using Base = StateWrapper<BitIntegerState<uint32_t, fcAllFlags, fcNone>,
5404                            AbstractAttribute>;
5405
5406  AANoFPClass(const IRPosition &IRP, Attributor &A) : IRAttribute(IRP) {}
5407
5408  /// See AbstractAttribute::isValidIRPositionForInit
5409  static bool isValidIRPositionForInit(Attributor &A, const IRPosition &IRP) {
5410    Type *Ty = IRP.getAssociatedType();
5411    do {
5412      if (Ty->isFPOrFPVectorTy())
5413        return IRAttribute::isValidIRPositionForInit(A, IRP);
5414      if (!Ty->isArrayTy())
5415        break;
5416      Ty = Ty->getArrayElementType();
5417    } while (true);
5418    return false;
5419  }
5420
5421  /// Return true if we assume that the underlying value is nofpclass.
5422  FPClassTest getAssumedNoFPClass() const {
5423    return static_cast<FPClassTest>(getAssumed());
5424  }
5425
5426  /// Create an abstract attribute view for the position \p IRP.
5427  static AANoFPClass &createForPosition(const IRPosition &IRP, Attributor &A);
5428
5429  /// See AbstractAttribute::getName()
5430  const std::string getName() const override { return "AANoFPClass"; }
5431
5432  /// See AbstractAttribute::getIdAddr()
5433  const char *getIdAddr() const override { return &ID; }
5434
5435  /// This function should return true if the type of the \p AA is AANoFPClass
5436  static bool classof(const AbstractAttribute *AA) {
5437    return (AA->getIdAddr() == &ID);
5438  }
5439
5440  /// Unique ID (due to the unique address)
5441  static const char ID;
5442};
5443
5444struct AACallGraphNode;
5445struct AACallEdges;
5446
5447/// An Iterator for call edges, creates AACallEdges attributes in a lazy way.
5448/// This iterator becomes invalid if the underlying edge list changes.
5449/// So This shouldn't outlive a iteration of Attributor.
5450class AACallEdgeIterator
5451    : public iterator_adaptor_base<AACallEdgeIterator,
5452                                   SetVector<Function *>::iterator> {
5453  AACallEdgeIterator(Attributor &A, SetVector<Function *>::iterator Begin)
5454      : iterator_adaptor_base(Begin), A(A) {}
5455
5456public:
5457  AACallGraphNode *operator*() const;
5458
5459private:
5460  Attributor &A;
5461  friend AACallEdges;
5462  friend AttributorCallGraph;
5463};
5464
5465struct AACallGraphNode {
5466  AACallGraphNode(Attributor &A) : A(A) {}
5467  virtual ~AACallGraphNode() = default;
5468
5469  virtual AACallEdgeIterator optimisticEdgesBegin() const = 0;
5470  virtual AACallEdgeIterator optimisticEdgesEnd() const = 0;
5471
5472  /// Iterator range for exploring the call graph.
5473  iterator_range<AACallEdgeIterator> optimisticEdgesRange() const {
5474    return iterator_range<AACallEdgeIterator>(optimisticEdgesBegin(),
5475                                              optimisticEdgesEnd());
5476  }
5477
5478protected:
5479  /// Reference to Attributor needed for GraphTraits implementation.
5480  Attributor &A;
5481};
5482
5483/// An abstract state for querying live call edges.
5484/// This interface uses the Attributor's optimistic liveness
5485/// information to compute the edges that are alive.
5486struct AACallEdges : public StateWrapper<BooleanState, AbstractAttribute>,
5487                     AACallGraphNode {
5488  using Base = StateWrapper<BooleanState, AbstractAttribute>;
5489
5490  AACallEdges(const IRPosition &IRP, Attributor &A)
5491      : Base(IRP), AACallGraphNode(A) {}
5492
5493  /// See AbstractAttribute::requiresNonAsmForCallBase.
5494  static bool requiresNonAsmForCallBase() { return false; }
5495
5496  /// Get the optimistic edges.
5497  virtual const SetVector<Function *> &getOptimisticEdges() const = 0;
5498
5499  /// Is there any call with a unknown callee.
5500  virtual bool hasUnknownCallee() const = 0;
5501
5502  /// Is there any call with a unknown callee, excluding any inline asm.
5503  virtual bool hasNonAsmUnknownCallee() const = 0;
5504
5505  /// Iterator for exploring the call graph.
5506  AACallEdgeIterator optimisticEdgesBegin() const override {
5507    return AACallEdgeIterator(A, getOptimisticEdges().begin());
5508  }
5509
5510  /// Iterator for exploring the call graph.
5511  AACallEdgeIterator optimisticEdgesEnd() const override {
5512    return AACallEdgeIterator(A, getOptimisticEdges().end());
5513  }
5514
5515  /// Create an abstract attribute view for the position \p IRP.
5516  static AACallEdges &createForPosition(const IRPosition &IRP, Attributor &A);
5517
5518  /// See AbstractAttribute::getName()
5519  const std::string getName() const override { return "AACallEdges"; }
5520
5521  /// See AbstractAttribute::getIdAddr()
5522  const char *getIdAddr() const override { return &ID; }
5523
5524  /// This function should return true if the type of the \p AA is AACallEdges.
5525  static bool classof(const AbstractAttribute *AA) {
5526    return (AA->getIdAddr() == &ID);
5527  }
5528
5529  /// Unique ID (due to the unique address)
5530  static const char ID;
5531};
5532
5533// Synthetic root node for the Attributor's internal call graph.
5534struct AttributorCallGraph : public AACallGraphNode {
5535  AttributorCallGraph(Attributor &A) : AACallGraphNode(A) {}
5536  virtual ~AttributorCallGraph() = default;
5537
5538  AACallEdgeIterator optimisticEdgesBegin() const override {
5539    return AACallEdgeIterator(A, A.Functions.begin());
5540  }
5541
5542  AACallEdgeIterator optimisticEdgesEnd() const override {
5543    return AACallEdgeIterator(A, A.Functions.end());
5544  }
5545
5546  /// Force populate the entire call graph.
5547  void populateAll() const {
5548    for (const AACallGraphNode *AA : optimisticEdgesRange()) {
5549      // Nothing else to do here.
5550      (void)AA;
5551    }
5552  }
5553
5554  void print();
5555};
5556
5557template <> struct GraphTraits<AACallGraphNode *> {
5558  using NodeRef = AACallGraphNode *;
5559  using ChildIteratorType = AACallEdgeIterator;
5560
5561  static AACallEdgeIterator child_begin(AACallGraphNode *Node) {
5562    return Node->optimisticEdgesBegin();
5563  }
5564
5565  static AACallEdgeIterator child_end(AACallGraphNode *Node) {
5566    return Node->optimisticEdgesEnd();
5567  }
5568};
5569
5570template <>
5571struct GraphTraits<AttributorCallGraph *>
5572    : public GraphTraits<AACallGraphNode *> {
5573  using nodes_iterator = AACallEdgeIterator;
5574
5575  static AACallGraphNode *getEntryNode(AttributorCallGraph *G) {
5576    return static_cast<AACallGraphNode *>(G);
5577  }
5578
5579  static AACallEdgeIterator nodes_begin(const AttributorCallGraph *G) {
5580    return G->optimisticEdgesBegin();
5581  }
5582
5583  static AACallEdgeIterator nodes_end(const AttributorCallGraph *G) {
5584    return G->optimisticEdgesEnd();
5585  }
5586};
5587
5588template <>
5589struct DOTGraphTraits<AttributorCallGraph *> : public DefaultDOTGraphTraits {
5590  DOTGraphTraits(bool Simple = false) : DefaultDOTGraphTraits(Simple) {}
5591
5592  std::string getNodeLabel(const AACallGraphNode *Node,
5593                           const AttributorCallGraph *Graph) {
5594    const AACallEdges *AACE = static_cast<const AACallEdges *>(Node);
5595    return AACE->getAssociatedFunction()->getName().str();
5596  }
5597
5598  static bool isNodeHidden(const AACallGraphNode *Node,
5599                           const AttributorCallGraph *Graph) {
5600    // Hide the synth root.
5601    return static_cast<const AACallGraphNode *>(Graph) == Node;
5602  }
5603};
5604
5605struct AAExecutionDomain
5606    : public StateWrapper<BooleanState, AbstractAttribute> {
5607  using Base = StateWrapper<BooleanState, AbstractAttribute>;
5608  AAExecutionDomain(const IRPosition &IRP, Attributor &A) : Base(IRP) {}
5609
5610  /// Summary about the execution domain of a block or instruction.
5611  struct ExecutionDomainTy {
5612    using BarriersSetTy = SmallPtrSet<CallBase *, 2>;
5613    using AssumesSetTy = SmallPtrSet<AssumeInst *, 4>;
5614
5615    void addAssumeInst(Attributor &A, AssumeInst &AI) {
5616      EncounteredAssumes.insert(&AI);
5617    }
5618
5619    void addAlignedBarrier(Attributor &A, CallBase &CB) {
5620      AlignedBarriers.insert(&CB);
5621    }
5622
5623    void clearAssumeInstAndAlignedBarriers() {
5624      EncounteredAssumes.clear();
5625      AlignedBarriers.clear();
5626    }
5627
5628    bool IsExecutedByInitialThreadOnly = true;
5629    bool IsReachedFromAlignedBarrierOnly = true;
5630    bool IsReachingAlignedBarrierOnly = true;
5631    bool EncounteredNonLocalSideEffect = false;
5632    BarriersSetTy AlignedBarriers;
5633    AssumesSetTy EncounteredAssumes;
5634  };
5635
5636  /// Create an abstract attribute view for the position \p IRP.
5637  static AAExecutionDomain &createForPosition(const IRPosition &IRP,
5638                                              Attributor &A);
5639
5640  /// See AbstractAttribute::getName().
5641  const std::string getName() const override { return "AAExecutionDomain"; }
5642
5643  /// See AbstractAttribute::getIdAddr().
5644  const char *getIdAddr() const override { return &ID; }
5645
5646  /// Check if an instruction is executed only by the initial thread.
5647  bool isExecutedByInitialThreadOnly(const Instruction &I) const {
5648    return isExecutedByInitialThreadOnly(*I.getParent());
5649  }
5650
5651  /// Check if a basic block is executed only by the initial thread.
5652  virtual bool isExecutedByInitialThreadOnly(const BasicBlock &) const = 0;
5653
5654  /// Check if the instruction \p I is executed in an aligned region, that is,
5655  /// the synchronizing effects before and after \p I are both aligned barriers.
5656  /// This effectively means all threads execute \p I together.
5657  virtual bool isExecutedInAlignedRegion(Attributor &A,
5658                                         const Instruction &I) const = 0;
5659
5660  virtual ExecutionDomainTy getExecutionDomain(const BasicBlock &) const = 0;
5661  /// Return the execution domain with which the call \p CB is entered and the
5662  /// one with which it is left.
5663  virtual std::pair<ExecutionDomainTy, ExecutionDomainTy>
5664  getExecutionDomain(const CallBase &CB) const = 0;
5665  virtual ExecutionDomainTy getFunctionExecutionDomain() const = 0;
5666
5667  /// Helper function to determine if \p FI is a no-op given the information
5668  /// about its execution from \p ExecDomainAA.
5669  virtual bool isNoOpFence(const FenceInst &FI) const = 0;
5670
5671  /// This function should return true if the type of the \p AA is
5672  /// AAExecutionDomain.
5673  static bool classof(const AbstractAttribute *AA) {
5674    return (AA->getIdAddr() == &ID);
5675  }
5676
5677  /// Unique ID (due to the unique address)
5678  static const char ID;
5679};
5680
5681/// An abstract Attribute for computing reachability between functions.
5682struct AAInterFnReachability
5683    : public StateWrapper<BooleanState, AbstractAttribute> {
5684  using Base = StateWrapper<BooleanState, AbstractAttribute>;
5685
5686  AAInterFnReachability(const IRPosition &IRP, Attributor &A) : Base(IRP) {}
5687
5688  /// If the function represented by this possition can reach \p Fn.
5689  bool canReach(Attributor &A, const Function &Fn) const {
5690    Function *Scope = getAnchorScope();
5691    if (!Scope || Scope->isDeclaration())
5692      return true;
5693    return instructionCanReach(A, Scope->getEntryBlock().front(), Fn);
5694  }
5695
5696  /// Can  \p Inst reach \p Fn.
5697  /// See also AA::isPotentiallyReachable.
5698  virtual bool instructionCanReach(
5699      Attributor &A, const Instruction &Inst, const Function &Fn,
5700      const AA::InstExclusionSetTy *ExclusionSet = nullptr) const = 0;
5701
5702  /// Create an abstract attribute view for the position \p IRP.
5703  static AAInterFnReachability &createForPosition(const IRPosition &IRP,
5704                                                  Attributor &A);
5705
5706  /// See AbstractAttribute::getName()
5707  const std::string getName() const override { return "AAInterFnReachability"; }
5708
5709  /// See AbstractAttribute::getIdAddr()
5710  const char *getIdAddr() const override { return &ID; }
5711
5712  /// This function should return true if the type of the \p AA is AACallEdges.
5713  static bool classof(const AbstractAttribute *AA) {
5714    return (AA->getIdAddr() == &ID);
5715  }
5716
5717  /// Unique ID (due to the unique address)
5718  static const char ID;
5719};
5720
5721/// An abstract Attribute for determining the necessity of the convergent
5722/// attribute.
5723struct AANonConvergent : public StateWrapper<BooleanState, AbstractAttribute> {
5724  using Base = StateWrapper<BooleanState, AbstractAttribute>;
5725
5726  AANonConvergent(const IRPosition &IRP, Attributor &A) : Base(IRP) {}
5727
5728  /// Create an abstract attribute view for the position \p IRP.
5729  static AANonConvergent &createForPosition(const IRPosition &IRP,
5730                                            Attributor &A);
5731
5732  /// Return true if "non-convergent" is assumed.
5733  bool isAssumedNotConvergent() const { return getAssumed(); }
5734
5735  /// Return true if "non-convergent" is known.
5736  bool isKnownNotConvergent() const { return getKnown(); }
5737
5738  /// See AbstractAttribute::getName()
5739  const std::string getName() const override { return "AANonConvergent"; }
5740
5741  /// See AbstractAttribute::getIdAddr()
5742  const char *getIdAddr() const override { return &ID; }
5743
5744  /// This function should return true if the type of the \p AA is
5745  /// AANonConvergent.
5746  static bool classof(const AbstractAttribute *AA) {
5747    return (AA->getIdAddr() == &ID);
5748  }
5749
5750  /// Unique ID (due to the unique address)
5751  static const char ID;
5752};
5753
5754/// An abstract interface for struct information.
5755struct AAPointerInfo : public AbstractAttribute {
5756  AAPointerInfo(const IRPosition &IRP) : AbstractAttribute(IRP) {}
5757
5758  /// See AbstractAttribute::isValidIRPositionForInit
5759  static bool isValidIRPositionForInit(Attributor &A, const IRPosition &IRP) {
5760    if (!IRP.getAssociatedType()->isPtrOrPtrVectorTy())
5761      return false;
5762    return AbstractAttribute::isValidIRPositionForInit(A, IRP);
5763  }
5764
5765  enum AccessKind {
5766    // First two bits to distinguish may and must accesses.
5767    AK_MUST = 1 << 0,
5768    AK_MAY = 1 << 1,
5769
5770    // Then two bits for read and write. These are not exclusive.
5771    AK_R = 1 << 2,
5772    AK_W = 1 << 3,
5773    AK_RW = AK_R | AK_W,
5774
5775    // One special case for assumptions about memory content. These
5776    // are neither reads nor writes. They are however always modeled
5777    // as read to avoid using them for write removal.
5778    AK_ASSUMPTION = (1 << 4) | AK_MUST,
5779
5780    // Helper for easy access.
5781    AK_MAY_READ = AK_MAY | AK_R,
5782    AK_MAY_WRITE = AK_MAY | AK_W,
5783    AK_MAY_READ_WRITE = AK_MAY | AK_R | AK_W,
5784    AK_MUST_READ = AK_MUST | AK_R,
5785    AK_MUST_WRITE = AK_MUST | AK_W,
5786    AK_MUST_READ_WRITE = AK_MUST | AK_R | AK_W,
5787  };
5788
5789  /// A container for a list of ranges.
5790  struct RangeList {
5791    // The set of ranges rarely contains more than one element, and is unlikely
5792    // to contain more than say four elements. So we find the middle-ground with
5793    // a sorted vector. This avoids hard-coding a rarely used number like "four"
5794    // into every instance of a SmallSet.
5795    using RangeTy = AA::RangeTy;
5796    using VecTy = SmallVector<RangeTy>;
5797    using iterator = VecTy::iterator;
5798    using const_iterator = VecTy::const_iterator;
5799    VecTy Ranges;
5800
5801    RangeList(const RangeTy &R) { Ranges.push_back(R); }
5802    RangeList(ArrayRef<int64_t> Offsets, int64_t Size) {
5803      Ranges.reserve(Offsets.size());
5804      for (unsigned i = 0, e = Offsets.size(); i != e; ++i) {
5805        assert(((i + 1 == e) || Offsets[i] < Offsets[i + 1]) &&
5806               "Expected strictly ascending offsets.");
5807        Ranges.emplace_back(Offsets[i], Size);
5808      }
5809    }
5810    RangeList() = default;
5811
5812    iterator begin() { return Ranges.begin(); }
5813    iterator end() { return Ranges.end(); }
5814    const_iterator begin() const { return Ranges.begin(); }
5815    const_iterator end() const { return Ranges.end(); }
5816
5817    // Helpers required for std::set_difference
5818    using value_type = RangeTy;
5819    void push_back(const RangeTy &R) {
5820      assert((Ranges.empty() || RangeTy::OffsetLessThan(Ranges.back(), R)) &&
5821             "Ensure the last element is the greatest.");
5822      Ranges.push_back(R);
5823    }
5824
5825    /// Copy ranges from \p L that are not in \p R, into \p D.
5826    static void set_difference(const RangeList &L, const RangeList &R,
5827                               RangeList &D) {
5828      std::set_difference(L.begin(), L.end(), R.begin(), R.end(),
5829                          std::back_inserter(D), RangeTy::OffsetLessThan);
5830    }
5831
5832    unsigned size() const { return Ranges.size(); }
5833
5834    bool operator==(const RangeList &OI) const { return Ranges == OI.Ranges; }
5835
5836    /// Merge the ranges in \p RHS into the current ranges.
5837    /// - Merging a list of  unknown ranges makes the current list unknown.
5838    /// - Ranges with the same offset are merged according to RangeTy::operator&
5839    /// \return true if the current RangeList changed.
5840    bool merge(const RangeList &RHS) {
5841      if (isUnknown())
5842        return false;
5843      if (RHS.isUnknown()) {
5844        setUnknown();
5845        return true;
5846      }
5847
5848      if (Ranges.empty()) {
5849        Ranges = RHS.Ranges;
5850        return true;
5851      }
5852
5853      bool Changed = false;
5854      auto LPos = Ranges.begin();
5855      for (auto &R : RHS.Ranges) {
5856        auto Result = insert(LPos, R);
5857        if (isUnknown())
5858          return true;
5859        LPos = Result.first;
5860        Changed |= Result.second;
5861      }
5862      return Changed;
5863    }
5864
5865    /// Insert \p R at the given iterator \p Pos, and merge if necessary.
5866    ///
5867    /// This assumes that all ranges before \p Pos are OffsetLessThan \p R, and
5868    /// then maintains the sorted order for the suffix list.
5869    ///
5870    /// \return The place of insertion and true iff anything changed.
5871    std::pair<iterator, bool> insert(iterator Pos, const RangeTy &R) {
5872      if (isUnknown())
5873        return std::make_pair(Ranges.begin(), false);
5874      if (R.offsetOrSizeAreUnknown()) {
5875        return std::make_pair(setUnknown(), true);
5876      }
5877
5878      // Maintain this as a sorted vector of unique entries.
5879      auto LB = std::lower_bound(Pos, Ranges.end(), R, RangeTy::OffsetLessThan);
5880      if (LB == Ranges.end() || LB->Offset != R.Offset)
5881        return std::make_pair(Ranges.insert(LB, R), true);
5882      bool Changed = *LB != R;
5883      *LB &= R;
5884      if (LB->offsetOrSizeAreUnknown())
5885        return std::make_pair(setUnknown(), true);
5886      return std::make_pair(LB, Changed);
5887    }
5888
5889    /// Insert the given range \p R, maintaining sorted order.
5890    ///
5891    /// \return The place of insertion and true iff anything changed.
5892    std::pair<iterator, bool> insert(const RangeTy &R) {
5893      return insert(Ranges.begin(), R);
5894    }
5895
5896    /// Add the increment \p Inc to the offset of every range.
5897    void addToAllOffsets(int64_t Inc) {
5898      assert(!isUnassigned() &&
5899             "Cannot increment if the offset is not yet computed!");
5900      if (isUnknown())
5901        return;
5902      for (auto &R : Ranges) {
5903        R.Offset += Inc;
5904      }
5905    }
5906
5907    /// Return true iff there is exactly one range and it is known.
5908    bool isUnique() const {
5909      return Ranges.size() == 1 && !Ranges.front().offsetOrSizeAreUnknown();
5910    }
5911
5912    /// Return the unique range, assuming it exists.
5913    const RangeTy &getUnique() const {
5914      assert(isUnique() && "No unique range to return!");
5915      return Ranges.front();
5916    }
5917
5918    /// Return true iff the list contains an unknown range.
5919    bool isUnknown() const {
5920      if (isUnassigned())
5921        return false;
5922      if (Ranges.front().offsetOrSizeAreUnknown()) {
5923        assert(Ranges.size() == 1 && "Unknown is a singleton range.");
5924        return true;
5925      }
5926      return false;
5927    }
5928
5929    /// Discard all ranges and insert a single unknown range.
5930    iterator setUnknown() {
5931      Ranges.clear();
5932      Ranges.push_back(RangeTy::getUnknown());
5933      return Ranges.begin();
5934    }
5935
5936    /// Return true if no ranges have been inserted.
5937    bool isUnassigned() const { return Ranges.size() == 0; }
5938  };
5939
5940  /// An access description.
5941  struct Access {
5942    Access(Instruction *I, int64_t Offset, int64_t Size,
5943           std::optional<Value *> Content, AccessKind Kind, Type *Ty)
5944        : LocalI(I), RemoteI(I), Content(Content), Ranges(Offset, Size),
5945          Kind(Kind), Ty(Ty) {
5946      verify();
5947    }
5948    Access(Instruction *LocalI, Instruction *RemoteI, const RangeList &Ranges,
5949           std::optional<Value *> Content, AccessKind K, Type *Ty)
5950        : LocalI(LocalI), RemoteI(RemoteI), Content(Content), Ranges(Ranges),
5951          Kind(K), Ty(Ty) {
5952      if (Ranges.size() > 1) {
5953        Kind = AccessKind(Kind | AK_MAY);
5954        Kind = AccessKind(Kind & ~AK_MUST);
5955      }
5956      verify();
5957    }
5958    Access(Instruction *LocalI, Instruction *RemoteI, int64_t Offset,
5959           int64_t Size, std::optional<Value *> Content, AccessKind Kind,
5960           Type *Ty)
5961        : LocalI(LocalI), RemoteI(RemoteI), Content(Content),
5962          Ranges(Offset, Size), Kind(Kind), Ty(Ty) {
5963      verify();
5964    }
5965    Access(const Access &Other) = default;
5966
5967    Access &operator=(const Access &Other) = default;
5968    bool operator==(const Access &R) const {
5969      return LocalI == R.LocalI && RemoteI == R.RemoteI && Ranges == R.Ranges &&
5970             Content == R.Content && Kind == R.Kind;
5971    }
5972    bool operator!=(const Access &R) const { return !(*this == R); }
5973
5974    Access &operator&=(const Access &R) {
5975      assert(RemoteI == R.RemoteI && "Expected same instruction!");
5976      assert(LocalI == R.LocalI && "Expected same instruction!");
5977
5978      // Note that every Access object corresponds to a unique Value, and only
5979      // accesses to the same Value are merged. Hence we assume that all ranges
5980      // are the same size. If ranges can be different size, then the contents
5981      // must be dropped.
5982      Ranges.merge(R.Ranges);
5983      Content =
5984          AA::combineOptionalValuesInAAValueLatice(Content, R.Content, Ty);
5985
5986      // Combine the access kind, which results in a bitwise union.
5987      // If there is more than one range, then this must be a MAY.
5988      // If we combine a may and a must access we clear the must bit.
5989      Kind = AccessKind(Kind | R.Kind);
5990      if ((Kind & AK_MAY) || Ranges.size() > 1) {
5991        Kind = AccessKind(Kind | AK_MAY);
5992        Kind = AccessKind(Kind & ~AK_MUST);
5993      }
5994      verify();
5995      return *this;
5996    }
5997
5998    void verify() {
5999      assert(isMustAccess() + isMayAccess() == 1 &&
6000             "Expect must or may access, not both.");
6001      assert(isAssumption() + isWrite() <= 1 &&
6002             "Expect assumption access or write access, never both.");
6003      assert((isMayAccess() || Ranges.size() == 1) &&
6004             "Cannot be a must access if there are multiple ranges.");
6005    }
6006
6007    /// Return the access kind.
6008    AccessKind getKind() const { return Kind; }
6009
6010    /// Return true if this is a read access.
6011    bool isRead() const { return Kind & AK_R; }
6012
6013    /// Return true if this is a write access.
6014    bool isWrite() const { return Kind & AK_W; }
6015
6016    /// Return true if this is a write access.
6017    bool isWriteOrAssumption() const { return isWrite() || isAssumption(); }
6018
6019    /// Return true if this is an assumption access.
6020    bool isAssumption() const { return Kind == AK_ASSUMPTION; }
6021
6022    bool isMustAccess() const {
6023      bool MustAccess = Kind & AK_MUST;
6024      assert((!MustAccess || Ranges.size() < 2) &&
6025             "Cannot be a must access if there are multiple ranges.");
6026      return MustAccess;
6027    }
6028
6029    bool isMayAccess() const {
6030      bool MayAccess = Kind & AK_MAY;
6031      assert((MayAccess || Ranges.size() < 2) &&
6032             "Cannot be a must access if there are multiple ranges.");
6033      return MayAccess;
6034    }
6035
6036    /// Return the instruction that causes the access with respect to the local
6037    /// scope of the associated attribute.
6038    Instruction *getLocalInst() const { return LocalI; }
6039
6040    /// Return the actual instruction that causes the access.
6041    Instruction *getRemoteInst() const { return RemoteI; }
6042
6043    /// Return true if the value written is not known yet.
6044    bool isWrittenValueYetUndetermined() const { return !Content; }
6045
6046    /// Return true if the value written cannot be determined at all.
6047    bool isWrittenValueUnknown() const {
6048      return Content.has_value() && !*Content;
6049    }
6050
6051    /// Set the value written to nullptr, i.e., unknown.
6052    void setWrittenValueUnknown() { Content = nullptr; }
6053
6054    /// Return the type associated with the access, if known.
6055    Type *getType() const { return Ty; }
6056
6057    /// Return the value writen, if any.
6058    Value *getWrittenValue() const {
6059      assert(!isWrittenValueYetUndetermined() &&
6060             "Value needs to be determined before accessing it.");
6061      return *Content;
6062    }
6063
6064    /// Return the written value which can be `llvm::null` if it is not yet
6065    /// determined.
6066    std::optional<Value *> getContent() const { return Content; }
6067
6068    bool hasUniqueRange() const { return Ranges.isUnique(); }
6069    const AA::RangeTy &getUniqueRange() const { return Ranges.getUnique(); }
6070
6071    /// Add a range accessed by this Access.
6072    ///
6073    /// If there are multiple ranges, then this is a "may access".
6074    void addRange(int64_t Offset, int64_t Size) {
6075      Ranges.insert({Offset, Size});
6076      if (!hasUniqueRange()) {
6077        Kind = AccessKind(Kind | AK_MAY);
6078        Kind = AccessKind(Kind & ~AK_MUST);
6079      }
6080    }
6081
6082    const RangeList &getRanges() const { return Ranges; }
6083
6084    using const_iterator = RangeList::const_iterator;
6085    const_iterator begin() const { return Ranges.begin(); }
6086    const_iterator end() const { return Ranges.end(); }
6087
6088  private:
6089    /// The instruction responsible for the access with respect to the local
6090    /// scope of the associated attribute.
6091    Instruction *LocalI;
6092
6093    /// The instruction responsible for the access.
6094    Instruction *RemoteI;
6095
6096    /// The value written, if any. `std::nullopt` means "not known yet",
6097    /// `nullptr` cannot be determined.
6098    std::optional<Value *> Content;
6099
6100    /// Set of potential ranges accessed from the base pointer.
6101    RangeList Ranges;
6102
6103    /// The access kind, e.g., READ, as bitset (could be more than one).
6104    AccessKind Kind;
6105
6106    /// The type of the content, thus the type read/written, can be null if not
6107    /// available.
6108    Type *Ty;
6109  };
6110
6111  /// Create an abstract attribute view for the position \p IRP.
6112  static AAPointerInfo &createForPosition(const IRPosition &IRP, Attributor &A);
6113
6114  /// See AbstractAttribute::getName()
6115  const std::string getName() const override { return "AAPointerInfo"; }
6116
6117  /// See AbstractAttribute::getIdAddr()
6118  const char *getIdAddr() const override { return &ID; }
6119
6120  using OffsetBinsTy = DenseMap<AA::RangeTy, SmallSet<unsigned, 4>>;
6121  using const_bin_iterator = OffsetBinsTy::const_iterator;
6122  virtual const_bin_iterator begin() const = 0;
6123  virtual const_bin_iterator end() const = 0;
6124  virtual int64_t numOffsetBins() const = 0;
6125
6126  /// Call \p CB on all accesses that might interfere with \p Range and return
6127  /// true if all such accesses were known and the callback returned true for
6128  /// all of them, false otherwise. An access interferes with an offset-size
6129  /// pair if it might read or write that memory region.
6130  virtual bool forallInterferingAccesses(
6131      AA::RangeTy Range, function_ref<bool(const Access &, bool)> CB) const = 0;
6132
6133  /// Call \p CB on all accesses that might interfere with \p I and
6134  /// return true if all such accesses were known and the callback returned true
6135  /// for all of them, false otherwise. In contrast to forallInterferingAccesses
6136  /// this function will perform reasoning to exclude write accesses that cannot
6137  /// affect the load even if they on the surface look as if they would. The
6138  /// flag \p HasBeenWrittenTo will be set to true if we know that \p I does not
6139  /// read the initial value of the underlying memory. If \p SkipCB is given and
6140  /// returns false for a potentially interfering access, that access is not
6141  /// checked for actual interference.
6142  virtual bool forallInterferingAccesses(
6143      Attributor &A, const AbstractAttribute &QueryingAA, Instruction &I,
6144      bool FindInterferingWrites, bool FindInterferingReads,
6145      function_ref<bool(const Access &, bool)> CB, bool &HasBeenWrittenTo,
6146      AA::RangeTy &Range,
6147      function_ref<bool(const Access &)> SkipCB = nullptr) const = 0;
6148
6149  /// This function should return true if the type of the \p AA is AAPointerInfo
6150  static bool classof(const AbstractAttribute *AA) {
6151    return (AA->getIdAddr() == &ID);
6152  }
6153
6154  /// Unique ID (due to the unique address)
6155  static const char ID;
6156};
6157
6158raw_ostream &operator<<(raw_ostream &, const AAPointerInfo::Access &);
6159
6160/// An abstract attribute for getting assumption information.
6161struct AAAssumptionInfo
6162    : public StateWrapper<SetState<StringRef>, AbstractAttribute,
6163                          DenseSet<StringRef>> {
6164  using Base =
6165      StateWrapper<SetState<StringRef>, AbstractAttribute, DenseSet<StringRef>>;
6166
6167  AAAssumptionInfo(const IRPosition &IRP, Attributor &A,
6168                   const DenseSet<StringRef> &Known)
6169      : Base(IRP, Known) {}
6170
6171  /// Returns true if the assumption set contains the assumption \p Assumption.
6172  virtual bool hasAssumption(const StringRef Assumption) const = 0;
6173
6174  /// Create an abstract attribute view for the position \p IRP.
6175  static AAAssumptionInfo &createForPosition(const IRPosition &IRP,
6176                                             Attributor &A);
6177
6178  /// See AbstractAttribute::getName()
6179  const std::string getName() const override { return "AAAssumptionInfo"; }
6180
6181  /// See AbstractAttribute::getIdAddr()
6182  const char *getIdAddr() const override { return &ID; }
6183
6184  /// This function should return true if the type of the \p AA is
6185  /// AAAssumptionInfo
6186  static bool classof(const AbstractAttribute *AA) {
6187    return (AA->getIdAddr() == &ID);
6188  }
6189
6190  /// Unique ID (due to the unique address)
6191  static const char ID;
6192};
6193
6194/// An abstract attribute for getting all assumption underlying objects.
6195struct AAUnderlyingObjects : AbstractAttribute {
6196  AAUnderlyingObjects(const IRPosition &IRP) : AbstractAttribute(IRP) {}
6197
6198  /// See AbstractAttribute::isValidIRPositionForInit
6199  static bool isValidIRPositionForInit(Attributor &A, const IRPosition &IRP) {
6200    if (!IRP.getAssociatedType()->isPtrOrPtrVectorTy())
6201      return false;
6202    return AbstractAttribute::isValidIRPositionForInit(A, IRP);
6203  }
6204
6205  /// See AbstractAttribute::requiresCallersForArgOrFunction
6206  static bool requiresCallersForArgOrFunction() { return true; }
6207
6208  /// Create an abstract attribute biew for the position \p IRP.
6209  static AAUnderlyingObjects &createForPosition(const IRPosition &IRP,
6210                                                Attributor &A);
6211
6212  /// See AbstractAttribute::getName()
6213  const std::string getName() const override { return "AAUnderlyingObjects"; }
6214
6215  /// See AbstractAttribute::getIdAddr()
6216  const char *getIdAddr() const override { return &ID; }
6217
6218  /// This function should return true if the type of the \p AA is
6219  /// AAUnderlyingObjects.
6220  static bool classof(const AbstractAttribute *AA) {
6221    return (AA->getIdAddr() == &ID);
6222  }
6223
6224  /// Unique ID (due to the unique address)
6225  static const char ID;
6226
6227  /// Check \p Pred on all underlying objects in \p Scope collected so far.
6228  ///
6229  /// This method will evaluate \p Pred on all underlying objects in \p Scope
6230  /// collected so far and return true if \p Pred holds on all of them.
6231  virtual bool
6232  forallUnderlyingObjects(function_ref<bool(Value &)> Pred,
6233                          AA::ValueScope Scope = AA::Interprocedural) const = 0;
6234};
6235
6236/// An abstract interface for address space information.
6237struct AAAddressSpace : public StateWrapper<BooleanState, AbstractAttribute> {
6238  AAAddressSpace(const IRPosition &IRP, Attributor &A)
6239      : StateWrapper<BooleanState, AbstractAttribute>(IRP) {}
6240
6241  /// See AbstractAttribute::isValidIRPositionForInit
6242  static bool isValidIRPositionForInit(Attributor &A, const IRPosition &IRP) {
6243    if (!IRP.getAssociatedType()->isPtrOrPtrVectorTy())
6244      return false;
6245    return AbstractAttribute::isValidIRPositionForInit(A, IRP);
6246  }
6247
6248  /// See AbstractAttribute::requiresCallersForArgOrFunction
6249  static bool requiresCallersForArgOrFunction() { return true; }
6250
6251  /// Return the address space of the associated value. \p NoAddressSpace is
6252  /// returned if the associated value is dead. This functions is not supposed
6253  /// to be called if the AA is invalid.
6254  virtual int32_t getAddressSpace() const = 0;
6255
6256  /// Create an abstract attribute view for the position \p IRP.
6257  static AAAddressSpace &createForPosition(const IRPosition &IRP,
6258                                           Attributor &A);
6259
6260  /// See AbstractAttribute::getName()
6261  const std::string getName() const override { return "AAAddressSpace"; }
6262
6263  /// See AbstractAttribute::getIdAddr()
6264  const char *getIdAddr() const override { return &ID; }
6265
6266  /// This function should return true if the type of the \p AA is
6267  /// AAAssumptionInfo
6268  static bool classof(const AbstractAttribute *AA) {
6269    return (AA->getIdAddr() == &ID);
6270  }
6271
6272  // No address space which indicates the associated value is dead.
6273  static const int32_t NoAddressSpace = -1;
6274
6275  /// Unique ID (due to the unique address)
6276  static const char ID;
6277};
6278
6279struct AAAllocationInfo : public StateWrapper<BooleanState, AbstractAttribute> {
6280  AAAllocationInfo(const IRPosition &IRP, Attributor &A)
6281      : StateWrapper<BooleanState, AbstractAttribute>(IRP) {}
6282
6283  /// See AbstractAttribute::isValidIRPositionForInit
6284  static bool isValidIRPositionForInit(Attributor &A, const IRPosition &IRP) {
6285    if (!IRP.getAssociatedType()->isPtrOrPtrVectorTy())
6286      return false;
6287    return AbstractAttribute::isValidIRPositionForInit(A, IRP);
6288  }
6289
6290  /// Create an abstract attribute view for the position \p IRP.
6291  static AAAllocationInfo &createForPosition(const IRPosition &IRP,
6292                                             Attributor &A);
6293
6294  virtual std::optional<TypeSize> getAllocatedSize() const = 0;
6295
6296  /// See AbstractAttribute::getName()
6297  const std::string getName() const override { return "AAAllocationInfo"; }
6298
6299  /// See AbstractAttribute::getIdAddr()
6300  const char *getIdAddr() const override { return &ID; }
6301
6302  /// This function should return true if the type of the \p AA is
6303  /// AAAllocationInfo
6304  static bool classof(const AbstractAttribute *AA) {
6305    return (AA->getIdAddr() == &ID);
6306  }
6307
6308  constexpr static const std::optional<TypeSize> HasNoAllocationSize =
6309      std::optional<TypeSize>(TypeSize(-1, true));
6310
6311  static const char ID;
6312};
6313
6314/// An abstract interface for llvm::GlobalValue information interference.
6315struct AAGlobalValueInfo
6316    : public StateWrapper<BooleanState, AbstractAttribute> {
6317  AAGlobalValueInfo(const IRPosition &IRP, Attributor &A)
6318      : StateWrapper<BooleanState, AbstractAttribute>(IRP) {}
6319
6320  /// See AbstractAttribute::isValidIRPositionForInit
6321  static bool isValidIRPositionForInit(Attributor &A, const IRPosition &IRP) {
6322    if (IRP.getPositionKind() != IRPosition::IRP_FLOAT)
6323      return false;
6324    auto *GV = dyn_cast<GlobalValue>(&IRP.getAnchorValue());
6325    if (!GV)
6326      return false;
6327    return GV->hasLocalLinkage();
6328  }
6329
6330  /// Create an abstract attribute view for the position \p IRP.
6331  static AAGlobalValueInfo &createForPosition(const IRPosition &IRP,
6332                                              Attributor &A);
6333
6334  /// Return true iff \p U is a potential use of the associated global value.
6335  virtual bool isPotentialUse(const Use &U) const = 0;
6336
6337  /// See AbstractAttribute::getName()
6338  const std::string getName() const override { return "AAGlobalValueInfo"; }
6339
6340  /// See AbstractAttribute::getIdAddr()
6341  const char *getIdAddr() const override { return &ID; }
6342
6343  /// This function should return true if the type of the \p AA is
6344  /// AAGlobalValueInfo
6345  static bool classof(const AbstractAttribute *AA) {
6346    return (AA->getIdAddr() == &ID);
6347  }
6348
6349  /// Unique ID (due to the unique address)
6350  static const char ID;
6351};
6352
6353/// An abstract interface for indirect call information interference.
6354struct AAIndirectCallInfo
6355    : public StateWrapper<BooleanState, AbstractAttribute> {
6356  AAIndirectCallInfo(const IRPosition &IRP, Attributor &A)
6357      : StateWrapper<BooleanState, AbstractAttribute>(IRP) {}
6358
6359  /// See AbstractAttribute::isValidIRPositionForInit
6360  static bool isValidIRPositionForInit(Attributor &A, const IRPosition &IRP) {
6361    if (IRP.getPositionKind() != IRPosition::IRP_CALL_SITE)
6362      return false;
6363    auto *CB = cast<CallBase>(IRP.getCtxI());
6364    return CB->getOpcode() == Instruction::Call && CB->isIndirectCall() &&
6365           !CB->isMustTailCall();
6366  }
6367
6368  /// Create an abstract attribute view for the position \p IRP.
6369  static AAIndirectCallInfo &createForPosition(const IRPosition &IRP,
6370                                               Attributor &A);
6371
6372  /// Call \CB on each potential callee value and return true if all were known
6373  /// and \p CB returned true on all of them. Otherwise, return false.
6374  virtual bool foreachCallee(function_ref<bool(Function *)> CB) const = 0;
6375
6376  /// See AbstractAttribute::getName()
6377  const std::string getName() const override { return "AAIndirectCallInfo"; }
6378
6379  /// See AbstractAttribute::getIdAddr()
6380  const char *getIdAddr() const override { return &ID; }
6381
6382  /// This function should return true if the type of the \p AA is
6383  /// AAIndirectCallInfo
6384  /// This function should return true if the type of the \p AA is
6385  /// AADenormalFPMath.
6386  static bool classof(const AbstractAttribute *AA) {
6387    return (AA->getIdAddr() == &ID);
6388  }
6389
6390  /// Unique ID (due to the unique address)
6391  static const char ID;
6392};
6393
6394/// An abstract Attribute for specializing "dynamic" components of
6395/// "denormal-fp-math" and "denormal-fp-math-f32" to a known denormal mode.
6396struct AADenormalFPMath
6397    : public StateWrapper<DenormalFPMathState, AbstractAttribute> {
6398  using Base = StateWrapper<DenormalFPMathState, AbstractAttribute>;
6399
6400  AADenormalFPMath(const IRPosition &IRP, Attributor &A) : Base(IRP) {}
6401
6402  /// Create an abstract attribute view for the position \p IRP.
6403  static AADenormalFPMath &createForPosition(const IRPosition &IRP,
6404                                             Attributor &A);
6405
6406  /// See AbstractAttribute::getName()
6407  const std::string getName() const override { return "AADenormalFPMath"; }
6408
6409  /// See AbstractAttribute::getIdAddr()
6410  const char *getIdAddr() const override { return &ID; }
6411
6412  /// This function should return true if the type of the \p AA is
6413  /// AADenormalFPMath.
6414  static bool classof(const AbstractAttribute *AA) {
6415    return (AA->getIdAddr() == &ID);
6416  }
6417
6418  /// Unique ID (due to the unique address)
6419  static const char ID;
6420};
6421
6422raw_ostream &operator<<(raw_ostream &, const AAPointerInfo::Access &);
6423
6424/// Run options, used by the pass manager.
6425enum AttributorRunOption {
6426  NONE = 0,
6427  MODULE = 1 << 0,
6428  CGSCC = 1 << 1,
6429  ALL = MODULE | CGSCC
6430};
6431
6432namespace AA {
6433/// Helper to avoid creating an AA for IR Attributes that might already be set.
6434template <Attribute::AttrKind AK, typename AAType = AbstractAttribute>
6435bool hasAssumedIRAttr(Attributor &A, const AbstractAttribute *QueryingAA,
6436                      const IRPosition &IRP, DepClassTy DepClass, bool &IsKnown,
6437                      bool IgnoreSubsumingPositions = false,
6438                      const AAType **AAPtr = nullptr) {
6439  IsKnown = false;
6440  switch (AK) {
6441#define CASE(ATTRNAME, AANAME, ...)                                            \
6442  case Attribute::ATTRNAME: {                                                  \
6443    if (AANAME::isImpliedByIR(A, IRP, AK, IgnoreSubsumingPositions))           \
6444      return IsKnown = true;                                                   \
6445    if (!QueryingAA)                                                           \
6446      return false;                                                            \
6447    const auto *AA = A.getAAFor<AANAME>(*QueryingAA, IRP, DepClass);           \
6448    if (AAPtr)                                                                 \
6449      *AAPtr = reinterpret_cast<const AAType *>(AA);                           \
6450    if (!AA || !AA->isAssumed(__VA_ARGS__))                                    \
6451      return false;                                                            \
6452    IsKnown = AA->isKnown(__VA_ARGS__);                                        \
6453    return true;                                                               \
6454  }
6455    CASE(NoUnwind, AANoUnwind, );
6456    CASE(WillReturn, AAWillReturn, );
6457    CASE(NoFree, AANoFree, );
6458    CASE(NoCapture, AANoCapture, );
6459    CASE(NoRecurse, AANoRecurse, );
6460    CASE(NoReturn, AANoReturn, );
6461    CASE(NoSync, AANoSync, );
6462    CASE(NoAlias, AANoAlias, );
6463    CASE(NonNull, AANonNull, );
6464    CASE(MustProgress, AAMustProgress, );
6465    CASE(NoUndef, AANoUndef, );
6466    CASE(ReadNone, AAMemoryBehavior, AAMemoryBehavior::NO_ACCESSES);
6467    CASE(ReadOnly, AAMemoryBehavior, AAMemoryBehavior::NO_WRITES);
6468    CASE(WriteOnly, AAMemoryBehavior, AAMemoryBehavior::NO_READS);
6469#undef CASE
6470  default:
6471    llvm_unreachable("hasAssumedIRAttr not available for this attribute kind");
6472  };
6473}
6474} // namespace AA
6475
6476} // end namespace llvm
6477
6478#endif // LLVM_TRANSFORMS_IPO_ATTRIBUTOR_H
6479