diff --git a/llvm/lib/Transforms/Coroutines/CoroFrame.cpp b/llvm/lib/Transforms/Coroutines/CoroFrame.cpp index 5b0e0cba13a7..f7146a04e208 100644 --- a/llvm/lib/Transforms/Coroutines/CoroFrame.cpp +++ b/llvm/lib/Transforms/Coroutines/CoroFrame.cpp @@ -2694,7 +2694,7 @@ void coro::buildCoroutineFrame(Function &F, Shape &Shape) { } // Later code makes structural assumptions about single predecessors phis e.g - // that they are not live accross a suspend point. + // that they are not live across a suspend point. cleanupSinglePredPHIs(F); // Transforms multi-edge PHI Nodes, so that any value feeding into a PHI will diff --git a/llvm/lib/Transforms/IPO/Attributor.cpp b/llvm/lib/Transforms/IPO/Attributor.cpp index 37c773bd47d6..c82001c39d96 100644 --- a/llvm/lib/Transforms/IPO/Attributor.cpp +++ b/llvm/lib/Transforms/IPO/Attributor.cpp @@ -3298,7 +3298,7 @@ static bool runAttributorOnFunctions(InformationCache &InfoCache, // Internalize non-exact functions // TODO: for now we eagerly internalize functions without calculating the // cost, we need a cost interface to determine whether internalizing - // a function is "benefitial" + // a function is "beneficial" if (AllowDeepWrapper) { unsigned FunSize = Functions.size(); for (unsigned u = 0; u < FunSize; u++) { diff --git a/llvm/lib/Transforms/IPO/AttributorAttributes.cpp b/llvm/lib/Transforms/IPO/AttributorAttributes.cpp index 6cf3188001d2..276fcc2c61d8 100644 --- a/llvm/lib/Transforms/IPO/AttributorAttributes.cpp +++ b/llvm/lib/Transforms/IPO/AttributorAttributes.cpp @@ -707,7 +707,7 @@ struct State; } // namespace PointerInfo } // namespace AA -/// Helper for AA::PointerInfo::Acccess DenseMap/Set usage. +/// Helper for AA::PointerInfo::Access DenseMap/Set usage. template <> struct DenseMapInfo : DenseMapInfo { using Access = AAPointerInfo::Access; @@ -722,7 +722,7 @@ template <> struct DenseMapInfo : DenseMapInfo> {}; -/// Helper for AA::PointerInfo::Acccess DenseMap/Set usage ignoring everythign +/// Helper for AA::PointerInfo::Access DenseMap/Set usage ignoring everythign /// but the instruction struct AccessAsInstructionInfo : DenseMapInfo { using Base = DenseMapInfo; @@ -7771,7 +7771,7 @@ void AAMemoryLocationImpl::categorizePtrValue( // on the call edge, though, we should. To make that happen we need to // teach various passes, e.g., DSE, about the copy effect of a byval. That // would also allow us to mark functions only accessing byval arguments as - // readnone again, atguably their acceses have no effect outside of the + // readnone again, arguably their accesses have no effect outside of the // function, like accesses to allocas. MLK = NO_ARGUMENT_MEM; } else if (auto *GV = dyn_cast(Obj)) { diff --git a/llvm/lib/Transforms/IPO/IROutliner.cpp b/llvm/lib/Transforms/IPO/IROutliner.cpp index 076076b076f6..909f2fc56802 100644 --- a/llvm/lib/Transforms/IPO/IROutliner.cpp +++ b/llvm/lib/Transforms/IPO/IROutliner.cpp @@ -1210,7 +1210,7 @@ static Optional getGVNForPHINode(OutlinableRegion &Region, // the hash for the PHINode. OGVN = Cand.getGVN(IncomingBlock); - // If there is no number for the incoming block, it is becaause we have + // If there is no number for the incoming block, it is because we have // split the candidate basic blocks. So we use the previous block that it // was split from to find the valid global value numbering for the PHINode. if (!OGVN) { diff --git a/llvm/lib/Transforms/IPO/PartialInlining.cpp b/llvm/lib/Transforms/IPO/PartialInlining.cpp index e42d80a9d48e..ab2be532c056 100644 --- a/llvm/lib/Transforms/IPO/PartialInlining.cpp +++ b/llvm/lib/Transforms/IPO/PartialInlining.cpp @@ -753,7 +753,7 @@ BranchProbability PartialInlinerImpl::getOutliningCallBBRelativeFreq( // is predicted to be less likely, the predicted probablity is usually // higher than the actual. For instance, the actual probability of the // less likely target is only 5%, but the guessed probablity can be - // 40%. In the latter case, there is no need for further adjustement. + // 40%. In the latter case, there is no need for further adjustment. // FIXME: add an option for this. if (OutlineRegionRelFreq < BranchProbability(45, 100)) return OutlineRegionRelFreq; diff --git a/llvm/lib/Transforms/Instrumentation/AddressSanitizer.cpp b/llvm/lib/Transforms/Instrumentation/AddressSanitizer.cpp index 3274e36ab71a..87041b9da7a2 100644 --- a/llvm/lib/Transforms/Instrumentation/AddressSanitizer.cpp +++ b/llvm/lib/Transforms/Instrumentation/AddressSanitizer.cpp @@ -1241,7 +1241,7 @@ bool AddressSanitizer::isInterestingAlloca(const AllocaInst &AI) { } bool AddressSanitizer::ignoreAccess(Instruction *Inst, Value *Ptr) { - // Instrument acesses from different address spaces only for AMDGPU. + // Instrument accesses from different address spaces only for AMDGPU. Type *PtrTy = cast(Ptr->getType()->getScalarType()); if (PtrTy->getPointerAddressSpace() != 0 && !(TargetTriple.isAMDGPU() && !isUnsupportedAMDGPUAddrspace(Ptr))) diff --git a/llvm/lib/Transforms/Instrumentation/DataFlowSanitizer.cpp b/llvm/lib/Transforms/Instrumentation/DataFlowSanitizer.cpp index dfdf9c0337dd..de5ba3101fd7 100644 --- a/llvm/lib/Transforms/Instrumentation/DataFlowSanitizer.cpp +++ b/llvm/lib/Transforms/Instrumentation/DataFlowSanitizer.cpp @@ -1146,7 +1146,7 @@ void DataFlowSanitizer::buildExternWeakCheckIfNeeded(IRBuilder<> &IRB, // but replacing with a known-to-not-be-null wrapper can break this check. // When replacing uses of the extern weak function with the wrapper we try // to avoid replacing uses in conditionals, but this is not perfect. - // In the case where we fail, and accidentially optimize out a null check + // In the case where we fail, and accidentally optimize out a null check // for a extern weak function, add a check here to help identify the issue. if (GlobalValue::isExternalWeakLinkage(F->getLinkage())) { std::vector Args; @@ -1465,7 +1465,7 @@ bool DataFlowSanitizer::runImpl(Module &M) { // label %avoid_my_func // The @"dfsw$my_func" wrapper is never null, so if we replace this use // in the comparison, the icmp will simplify to false and we have - // accidentially optimized away a null check that is necessary. + // accidentally optimized away a null check that is necessary. // This can lead to a crash when the null extern_weak my_func is called. // // To prevent (the most common pattern of) this problem, diff --git a/llvm/lib/Transforms/Instrumentation/HWAddressSanitizer.cpp b/llvm/lib/Transforms/Instrumentation/HWAddressSanitizer.cpp index b01c74320380..bceb3fc2c8d4 100644 --- a/llvm/lib/Transforms/Instrumentation/HWAddressSanitizer.cpp +++ b/llvm/lib/Transforms/Instrumentation/HWAddressSanitizer.cpp @@ -708,7 +708,7 @@ Value *HWAddressSanitizer::getShadowNonTls(IRBuilder<> &IRB) { } bool HWAddressSanitizer::ignoreAccess(Instruction *Inst, Value *Ptr) { - // Do not instrument acesses from different address spaces; we cannot deal + // Do not instrument accesses from different address spaces; we cannot deal // with them. Type *PtrTy = cast(Ptr->getType()->getScalarType()); if (PtrTy->getPointerAddressSpace() != 0) diff --git a/llvm/lib/Transforms/Instrumentation/MemProfiler.cpp b/llvm/lib/Transforms/Instrumentation/MemProfiler.cpp index 01e3b2c20218..47cc2cd3065c 100644 --- a/llvm/lib/Transforms/Instrumentation/MemProfiler.cpp +++ b/llvm/lib/Transforms/Instrumentation/MemProfiler.cpp @@ -385,7 +385,7 @@ MemProfiler::isInterestingMemoryAccess(Instruction *I) const { if (!Access.Addr) return None; - // Do not instrument acesses from different address spaces; we cannot deal + // Do not instrument accesses from different address spaces; we cannot deal // with them. Type *PtrTy = cast(Access.Addr->getType()->getScalarType()); if (PtrTy->getPointerAddressSpace() != 0) diff --git a/llvm/lib/Transforms/Instrumentation/ThreadSanitizer.cpp b/llvm/lib/Transforms/Instrumentation/ThreadSanitizer.cpp index d4aa31db8337..772a1d89e4e9 100644 --- a/llvm/lib/Transforms/Instrumentation/ThreadSanitizer.cpp +++ b/llvm/lib/Transforms/Instrumentation/ThreadSanitizer.cpp @@ -379,7 +379,7 @@ static bool shouldInstrumentReadWriteFromAddress(const Module *M, Value *Addr) { return false; } - // Do not instrument acesses from different address spaces; we cannot deal + // Do not instrument accesses from different address spaces; we cannot deal // with them. if (Addr) { Type *PtrTy = cast(Addr->getType()->getScalarType()); diff --git a/llvm/lib/Transforms/Scalar/CorrelatedValuePropagation.cpp b/llvm/lib/Transforms/Scalar/CorrelatedValuePropagation.cpp index 64bd4241f37c..574577ee598e 100644 --- a/llvm/lib/Transforms/Scalar/CorrelatedValuePropagation.cpp +++ b/llvm/lib/Transforms/Scalar/CorrelatedValuePropagation.cpp @@ -729,7 +729,7 @@ static bool narrowSDivOrSRem(BinaryOperator *Instr, LazyValueInfo *LVI) { // operands. unsigned OrigWidth = Instr->getType()->getIntegerBitWidth(); - // What is the smallest bit width that can accomodate the entire value ranges + // What is the smallest bit width that can accommodate the entire value ranges // of both of the operands? std::array, 2> CRs; unsigned MinSignedBits = 0; @@ -781,7 +781,7 @@ static bool processUDivOrURem(BinaryOperator *Instr, LazyValueInfo *LVI) { // Find the smallest power of two bitwidth that's sufficient to hold Instr's // operands. - // What is the smallest bit width that can accomodate the entire value ranges + // What is the smallest bit width that can accommodate the entire value ranges // of both of the operands? unsigned MaxActiveBits = 0; for (Value *Operand : Instr->operands()) { diff --git a/llvm/lib/Transforms/Scalar/GVN.cpp b/llvm/lib/Transforms/Scalar/GVN.cpp index b460637b7d88..afff8a8cfea1 100644 --- a/llvm/lib/Transforms/Scalar/GVN.cpp +++ b/llvm/lib/Transforms/Scalar/GVN.cpp @@ -471,7 +471,7 @@ uint32_t GVNPass::ValueTable::lookupOrAddCall(CallInst *C) { } if (local_dep.isDef()) { - // For masked load/store intrinsics, the local_dep may actully be + // For masked load/store intrinsics, the local_dep may actually be // a normal load or store instruction. CallInst *local_cdep = dyn_cast(local_dep.getInst()); diff --git a/llvm/lib/Transforms/Scalar/LICM.cpp b/llvm/lib/Transforms/Scalar/LICM.cpp index f54264b1dca6..e7b0eb0838b0 100644 --- a/llvm/lib/Transforms/Scalar/LICM.cpp +++ b/llvm/lib/Transforms/Scalar/LICM.cpp @@ -1988,7 +1988,7 @@ bool llvm::promoteLoopAccessesToScalars( IsKnownThreadLocalObject = !isa(Object); } - // Check that all accesses to pointers in the aliass set use the same type. + // Check that all accesses to pointers in the alias set use the same type. // We cannot (yet) promote a memory location that is loaded and stored in // different sizes. While we are at it, collect alignment and AA info. Type *AccessTy = nullptr; diff --git a/llvm/lib/Transforms/Scalar/LoopFlatten.cpp b/llvm/lib/Transforms/Scalar/LoopFlatten.cpp index e1eeb1d5bd31..aa75af240631 100644 --- a/llvm/lib/Transforms/Scalar/LoopFlatten.cpp +++ b/llvm/lib/Transforms/Scalar/LoopFlatten.cpp @@ -139,7 +139,7 @@ struct FlattenInfo { PHINode *NarrowInnerInductionPHI = nullptr; // Holds the old/narrow induction PHINode *NarrowOuterInductionPHI = nullptr; // phis, i.e. the Phis before IV - // has been apllied. Used to skip + // has been applied. Used to skip // checks on phi nodes. FlattenInfo(Loop *OL, Loop *IL) : OuterLoop(OL), InnerLoop(IL){}; diff --git a/llvm/lib/Transforms/Scalar/LoopFuse.cpp b/llvm/lib/Transforms/Scalar/LoopFuse.cpp index d8736446caf7..19cf0eba0985 100644 --- a/llvm/lib/Transforms/Scalar/LoopFuse.cpp +++ b/llvm/lib/Transforms/Scalar/LoopFuse.cpp @@ -428,7 +428,7 @@ using LoopVector = SmallVector; // order. Thus, if FC0 comes *before* FC1 in a FusionCandidateSet, then FC0 // dominates FC1 and FC1 post-dominates FC0. // std::set was chosen because we want a sorted data structure with stable -// iterators. A subsequent patch to loop fusion will enable fusing non-ajdacent +// iterators. A subsequent patch to loop fusion will enable fusing non-adjacent // loops by moving intervening code around. When this intervening code contains // loops, those loops will be moved also. The corresponding FusionCandidates // will also need to be moved accordingly. As this is done, having stable diff --git a/llvm/lib/Transforms/Scalar/LoopSimplifyCFG.cpp b/llvm/lib/Transforms/Scalar/LoopSimplifyCFG.cpp index 083f87436acd..3ef2980c9065 100644 --- a/llvm/lib/Transforms/Scalar/LoopSimplifyCFG.cpp +++ b/llvm/lib/Transforms/Scalar/LoopSimplifyCFG.cpp @@ -474,7 +474,7 @@ private: NumLoopBlocksDeleted += DeadLoopBlocks.size(); } - /// Constant-fold terminators of blocks acculumated in FoldCandidates into the + /// Constant-fold terminators of blocks accumulated in FoldCandidates into the /// unconditional branches. void foldTerminators() { for (BasicBlock *BB : FoldCandidates) { diff --git a/llvm/lib/Transforms/Scalar/LoopSink.cpp b/llvm/lib/Transforms/Scalar/LoopSink.cpp index dce1af475fb1..250eab8219db 100644 --- a/llvm/lib/Transforms/Scalar/LoopSink.cpp +++ b/llvm/lib/Transforms/Scalar/LoopSink.cpp @@ -300,8 +300,8 @@ static bool sinkLoopInvariantInstructions(Loop &L, AAResults &AA, LoopInfo &LI, return BFI.getBlockFreq(A) < BFI.getBlockFreq(B); }); - // Traverse preheader's instructions in reverse order becaue if A depends - // on B (A appears after B), A needs to be sinked first before B can be + // Traverse preheader's instructions in reverse order because if A depends + // on B (A appears after B), A needs to be sunk first before B can be // sinked. for (Instruction &I : llvm::make_early_inc_range(llvm::reverse(*Preheader))) { if (isa(&I)) diff --git a/llvm/lib/Transforms/Scalar/LoopUnrollPass.cpp b/llvm/lib/Transforms/Scalar/LoopUnrollPass.cpp index de5833f60adc..ca9ab3edcd0d 100644 --- a/llvm/lib/Transforms/Scalar/LoopUnrollPass.cpp +++ b/llvm/lib/Transforms/Scalar/LoopUnrollPass.cpp @@ -1222,7 +1222,7 @@ static LoopUnrollResult tryToUnrollLoop( // Find the smallest exact trip count for any exit. This is an upper bound // on the loop trip count, but an exit at an earlier iteration is still // possible. An unroll by the smallest exact trip count guarantees that all - // brnaches relating to at least one exit can be eliminated. This is unlike + // branches relating to at least one exit can be eliminated. This is unlike // the max trip count, which only guarantees that the backedge can be broken. unsigned TripCount = 0; unsigned TripMultiple = 1; diff --git a/llvm/lib/Transforms/Utils/LoopPeel.cpp b/llvm/lib/Transforms/Utils/LoopPeel.cpp index 9a7f9df7a330..3ccb766b81f6 100644 --- a/llvm/lib/Transforms/Utils/LoopPeel.cpp +++ b/llvm/lib/Transforms/Utils/LoopPeel.cpp @@ -331,7 +331,7 @@ static unsigned countToEliminateCompares(Loop &L, unsigned MaxPeelCount, /// This "heuristic" exactly matches implicit behavior which used to exist /// inside getLoopEstimatedTripCount. It was added here to keep an -/// improvement inside that API from causing peeling to become more agressive. +/// improvement inside that API from causing peeling to become more aggressive. /// This should probably be removed. static bool violatesLegacyMultiExitLoopCheck(Loop *L) { BasicBlock *Latch = L->getLoopLatch(); diff --git a/llvm/lib/Transforms/Utils/LoopUnrollRuntime.cpp b/llvm/lib/Transforms/Utils/LoopUnrollRuntime.cpp index 1c44ccb589bc..0d1ed73d1ed4 100644 --- a/llvm/lib/Transforms/Utils/LoopUnrollRuntime.cpp +++ b/llvm/lib/Transforms/Utils/LoopUnrollRuntime.cpp @@ -218,7 +218,7 @@ static void ConnectEpilog(Loop *L, Value *ModVal, BasicBlock *NewExit, for (PHINode &PN : NewExit->phis()) { // PN should be used in another PHI located in Exit block as // Exit was split by SplitBlockPredecessors into Exit and NewExit - // Basicaly it should look like: + // Basically it should look like: // NewExit: // PN = PHI [I, Latch] // ... diff --git a/llvm/lib/Transforms/Utils/LowerMemIntrinsics.cpp b/llvm/lib/Transforms/Utils/LowerMemIntrinsics.cpp index b4acb1b2ae90..630f586db76c 100644 --- a/llvm/lib/Transforms/Utils/LowerMemIntrinsics.cpp +++ b/llvm/lib/Transforms/Utils/LowerMemIntrinsics.cpp @@ -126,7 +126,7 @@ void llvm::createMemCpyLoopKnownSize(Instruction *InsertBefore, Value *SrcAddr, Align PartSrcAlign(commonAlignment(SrcAlign, BytesCopied)); Align PartDstAlign(commonAlignment(DstAlign, BytesCopied)); - // Calaculate the new index + // Calculate the new index unsigned OperandSize = DL.getTypeStoreSize(OpTy); assert( (!AtomicElementSize || OperandSize % *AtomicElementSize == 0) && diff --git a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp b/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp index 0f993403485f..1d37674aa48f 100644 --- a/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp +++ b/llvm/lib/Transforms/Vectorize/SLPVectorizer.cpp @@ -9394,7 +9394,7 @@ void BoUpSLP::BlockScheduling::calculateDependencies(ScheduleData *SD, WorkList.push_back(DestBundle); }; - // Any instruction which isn't safe to speculate at the begining of the + // Any instruction which isn't safe to speculate at the beginning of the // block is control dependend on any early exit or non-willreturn call // which proceeds it. if (!isGuaranteedToTransferExecutionToSuccessor(BundleMember->Inst)) { diff --git a/llvm/lib/Transforms/Vectorize/VPlanHCFGBuilder.cpp b/llvm/lib/Transforms/Vectorize/VPlanHCFGBuilder.cpp index 84b0dac862b6..af95e7016974 100644 --- a/llvm/lib/Transforms/Vectorize/VPlanHCFGBuilder.cpp +++ b/llvm/lib/Transforms/Vectorize/VPlanHCFGBuilder.cpp @@ -243,7 +243,7 @@ void PlainCFGBuilder::createVPInstructionsForVPBB(VPBasicBlock *VPBB, for (Value *Op : Inst->operands()) VPOperands.push_back(getOrCreateVPOperand(Op)); - // Build VPInstruction for any arbitraty Instruction without specific + // Build VPInstruction for any arbitrary Instruction without specific // representation in VPlan. NewVPV = cast( VPIRBuilder.createNaryOp(Inst->getOpcode(), VPOperands, Inst));