Merge llvm-project 13.0.0 release
This updates llvm, clang, compiler-rt, libc++, libunwind, lld, lldb and openmp to llvmorg-13.0.0-0-gd7b669b3a303, aka 13.0.0 release. PR: 258209 MFC after: 2 weeks
This commit is contained in:
commit
28a41182c0
@ -1804,7 +1804,6 @@ class DeclContext {
|
||||
ObjCContainerDeclBitfields ObjCContainerDeclBits;
|
||||
LinkageSpecDeclBitfields LinkageSpecDeclBits;
|
||||
BlockDeclBitfields BlockDeclBits;
|
||||
};
|
||||
|
||||
static_assert(sizeof(DeclContextBitfields) <= 8,
|
||||
"DeclContextBitfields is larger than 8 bytes!");
|
||||
@ -1828,6 +1827,7 @@ class DeclContext {
|
||||
"LinkageSpecDeclBitfields is larger than 8 bytes!");
|
||||
static_assert(sizeof(BlockDeclBitfields) <= 8,
|
||||
"BlockDeclBitfields is larger than 8 bytes!");
|
||||
};
|
||||
|
||||
/// FirstDecl - The first declaration stored within this declaration
|
||||
/// context.
|
||||
|
@ -1804,40 +1804,6 @@ class alignas(8) Type : public ExtQualsTypeCommonBase {
|
||||
PackExpansionTypeBitfields PackExpansionTypeBits;
|
||||
};
|
||||
|
||||
static_assert(sizeof(TypeBitfields) <= 8,
|
||||
"TypeBitfields is larger than 8 bytes!");
|
||||
static_assert(sizeof(ArrayTypeBitfields) <= 8,
|
||||
"ArrayTypeBitfields is larger than 8 bytes!");
|
||||
static_assert(sizeof(AttributedTypeBitfields) <= 8,
|
||||
"AttributedTypeBitfields is larger than 8 bytes!");
|
||||
static_assert(sizeof(AutoTypeBitfields) <= 8,
|
||||
"AutoTypeBitfields is larger than 8 bytes!");
|
||||
static_assert(sizeof(BuiltinTypeBitfields) <= 8,
|
||||
"BuiltinTypeBitfields is larger than 8 bytes!");
|
||||
static_assert(sizeof(FunctionTypeBitfields) <= 8,
|
||||
"FunctionTypeBitfields is larger than 8 bytes!");
|
||||
static_assert(sizeof(ObjCObjectTypeBitfields) <= 8,
|
||||
"ObjCObjectTypeBitfields is larger than 8 bytes!");
|
||||
static_assert(sizeof(ReferenceTypeBitfields) <= 8,
|
||||
"ReferenceTypeBitfields is larger than 8 bytes!");
|
||||
static_assert(sizeof(TypeWithKeywordBitfields) <= 8,
|
||||
"TypeWithKeywordBitfields is larger than 8 bytes!");
|
||||
static_assert(sizeof(ElaboratedTypeBitfields) <= 8,
|
||||
"ElaboratedTypeBitfields is larger than 8 bytes!");
|
||||
static_assert(sizeof(VectorTypeBitfields) <= 8,
|
||||
"VectorTypeBitfields is larger than 8 bytes!");
|
||||
static_assert(sizeof(SubstTemplateTypeParmPackTypeBitfields) <= 8,
|
||||
"SubstTemplateTypeParmPackTypeBitfields is larger"
|
||||
" than 8 bytes!");
|
||||
static_assert(sizeof(TemplateSpecializationTypeBitfields) <= 8,
|
||||
"TemplateSpecializationTypeBitfields is larger"
|
||||
" than 8 bytes!");
|
||||
static_assert(sizeof(DependentTemplateSpecializationTypeBitfields) <= 8,
|
||||
"DependentTemplateSpecializationTypeBitfields is larger"
|
||||
" than 8 bytes!");
|
||||
static_assert(sizeof(PackExpansionTypeBitfields) <= 8,
|
||||
"PackExpansionTypeBitfields is larger than 8 bytes");
|
||||
|
||||
private:
|
||||
template <class T> friend class TypePropertyCache;
|
||||
|
||||
|
@ -298,8 +298,8 @@ class alignas(8) InitializedEntity {
|
||||
|
||||
/// Create the initialization entity for the result of a function.
|
||||
static InitializedEntity InitializeResult(SourceLocation ReturnLoc,
|
||||
QualType Type, bool NRVO) {
|
||||
return InitializedEntity(EK_Result, ReturnLoc, Type, NRVO);
|
||||
QualType Type) {
|
||||
return InitializedEntity(EK_Result, ReturnLoc, Type);
|
||||
}
|
||||
|
||||
static InitializedEntity InitializeStmtExprResult(SourceLocation ReturnLoc,
|
||||
@ -308,20 +308,20 @@ class alignas(8) InitializedEntity {
|
||||
}
|
||||
|
||||
static InitializedEntity InitializeBlock(SourceLocation BlockVarLoc,
|
||||
QualType Type, bool NRVO) {
|
||||
return InitializedEntity(EK_BlockElement, BlockVarLoc, Type, NRVO);
|
||||
QualType Type) {
|
||||
return InitializedEntity(EK_BlockElement, BlockVarLoc, Type);
|
||||
}
|
||||
|
||||
static InitializedEntity InitializeLambdaToBlock(SourceLocation BlockVarLoc,
|
||||
QualType Type, bool NRVO) {
|
||||
QualType Type) {
|
||||
return InitializedEntity(EK_LambdaToBlockConversionBlockElement,
|
||||
BlockVarLoc, Type, NRVO);
|
||||
BlockVarLoc, Type);
|
||||
}
|
||||
|
||||
/// Create the initialization entity for an exception object.
|
||||
static InitializedEntity InitializeException(SourceLocation ThrowLoc,
|
||||
QualType Type, bool NRVO) {
|
||||
return InitializedEntity(EK_Exception, ThrowLoc, Type, NRVO);
|
||||
QualType Type) {
|
||||
return InitializedEntity(EK_Exception, ThrowLoc, Type);
|
||||
}
|
||||
|
||||
/// Create the initialization entity for an object allocated via new.
|
||||
|
@ -9934,10 +9934,19 @@ bool RecordExprEvaluator::VisitCXXConstructExpr(const CXXConstructExpr *E,
|
||||
return false;
|
||||
|
||||
// Avoid materializing a temporary for an elidable copy/move constructor.
|
||||
if (E->isElidable() && !ZeroInit)
|
||||
if (const MaterializeTemporaryExpr *ME
|
||||
= dyn_cast<MaterializeTemporaryExpr>(E->getArg(0)))
|
||||
if (E->isElidable() && !ZeroInit) {
|
||||
// FIXME: This only handles the simplest case, where the source object
|
||||
// is passed directly as the first argument to the constructor.
|
||||
// This should also handle stepping though implicit casts and
|
||||
// and conversion sequences which involve two steps, with a
|
||||
// conversion operator followed by a converting constructor.
|
||||
const Expr *SrcObj = E->getArg(0);
|
||||
assert(SrcObj->isTemporaryObject(Info.Ctx, FD->getParent()));
|
||||
assert(Info.Ctx.hasSameUnqualifiedType(E->getType(), SrcObj->getType()));
|
||||
if (const MaterializeTemporaryExpr *ME =
|
||||
dyn_cast<MaterializeTemporaryExpr>(SrcObj))
|
||||
return Visit(ME->getSubExpr());
|
||||
}
|
||||
|
||||
if (ZeroInit && !ZeroInitialization(E, T))
|
||||
return false;
|
||||
|
@ -1050,7 +1050,7 @@ class ThreadSafetyAnalyzer {
|
||||
const CFGBlock* PredBlock,
|
||||
const CFGBlock *CurrBlock);
|
||||
|
||||
bool join(const FactEntry &a, const FactEntry &b);
|
||||
bool join(const FactEntry &a, const FactEntry &b, bool CanModify);
|
||||
|
||||
void intersectAndWarn(FactSet &EntrySet, const FactSet &ExitSet,
|
||||
SourceLocation JoinLoc, LockErrorKind EntryLEK,
|
||||
@ -2188,25 +2188,28 @@ void BuildLockset::VisitDeclStmt(const DeclStmt *S) {
|
||||
}
|
||||
}
|
||||
|
||||
/// Given two facts merging on a join point, decide whether to warn and which
|
||||
/// one to keep.
|
||||
/// Given two facts merging on a join point, possibly warn and decide whether to
|
||||
/// keep or replace.
|
||||
///
|
||||
/// \return false if we should keep \p A, true if we should keep \p B.
|
||||
bool ThreadSafetyAnalyzer::join(const FactEntry &A, const FactEntry &B) {
|
||||
/// \param CanModify Whether we can replace \p A by \p B.
|
||||
/// \return false if we should keep \p A, true if we should take \p B.
|
||||
bool ThreadSafetyAnalyzer::join(const FactEntry &A, const FactEntry &B,
|
||||
bool CanModify) {
|
||||
if (A.kind() != B.kind()) {
|
||||
// For managed capabilities, the destructor should unlock in the right mode
|
||||
// anyway. For asserted capabilities no unlocking is needed.
|
||||
if ((A.managed() || A.asserted()) && (B.managed() || B.asserted())) {
|
||||
// The shared capability subsumes the exclusive capability.
|
||||
return B.kind() == LK_Shared;
|
||||
} else {
|
||||
// The shared capability subsumes the exclusive capability, if possible.
|
||||
bool ShouldTakeB = B.kind() == LK_Shared;
|
||||
if (CanModify || !ShouldTakeB)
|
||||
return ShouldTakeB;
|
||||
}
|
||||
Handler.handleExclusiveAndShared("mutex", B.toString(), B.loc(), A.loc());
|
||||
// Take the exclusive capability to reduce further warnings.
|
||||
return B.kind() == LK_Exclusive;
|
||||
}
|
||||
return CanModify && B.kind() == LK_Exclusive;
|
||||
} else {
|
||||
// The non-asserted capability is the one we want to track.
|
||||
return A.asserted() && !B.asserted();
|
||||
return CanModify && A.asserted() && !B.asserted();
|
||||
}
|
||||
}
|
||||
|
||||
@ -2237,8 +2240,8 @@ void ThreadSafetyAnalyzer::intersectAndWarn(FactSet &EntrySet,
|
||||
|
||||
FactSet::iterator EntryIt = EntrySet.findLockIter(FactMan, ExitFact);
|
||||
if (EntryIt != EntrySet.end()) {
|
||||
if (join(FactMan[*EntryIt], ExitFact) &&
|
||||
EntryLEK == LEK_LockedSomePredecessors)
|
||||
if (join(FactMan[*EntryIt], ExitFact,
|
||||
EntryLEK != LEK_LockedSomeLoopIterations))
|
||||
*EntryIt = Fact;
|
||||
} else if (!ExitFact.managed()) {
|
||||
ExitFact.handleRemovalFromIntersection(ExitSet, FactMan, JoinLoc,
|
||||
|
@ -609,16 +609,19 @@ CodeGenFunction::EmitCXXConstructExpr(const CXXConstructExpr *E,
|
||||
return;
|
||||
|
||||
// Elide the constructor if we're constructing from a temporary.
|
||||
// The temporary check is required because Sema sets this on NRVO
|
||||
// returns.
|
||||
if (getLangOpts().ElideConstructors && E->isElidable()) {
|
||||
assert(getContext().hasSameUnqualifiedType(E->getType(),
|
||||
E->getArg(0)->getType()));
|
||||
if (E->getArg(0)->isTemporaryObject(getContext(), CD->getParent())) {
|
||||
EmitAggExpr(E->getArg(0), Dest);
|
||||
// FIXME: This only handles the simplest case, where the source object
|
||||
// is passed directly as the first argument to the constructor.
|
||||
// This should also handle stepping though implicit casts and
|
||||
// conversion sequences which involve two steps, with a
|
||||
// conversion operator followed by a converting constructor.
|
||||
const Expr *SrcObj = E->getArg(0);
|
||||
assert(SrcObj->isTemporaryObject(getContext(), CD->getParent()));
|
||||
assert(
|
||||
getContext().hasSameUnqualifiedType(E->getType(), SrcObj->getType()));
|
||||
EmitAggExpr(SrcObj, Dest);
|
||||
return;
|
||||
}
|
||||
}
|
||||
|
||||
if (const ArrayType *arrayType
|
||||
= getContext().getAsArrayType(E->getType())) {
|
||||
|
@ -649,7 +649,6 @@ void UnwrappedLineParser::parseBlock(bool MustBeDeclaration, unsigned AddLevels,
|
||||
nextToken();
|
||||
|
||||
Line->Level = InitialLevel;
|
||||
FormatTok->setBlockKind(BK_Block);
|
||||
|
||||
if (PPStartHash == PPEndHash) {
|
||||
Line->MatchingOpeningBlockLineIndex = OpeningLineIndex;
|
||||
|
@ -2010,7 +2010,7 @@ static void checkEscapingByref(VarDecl *VD, Sema &S) {
|
||||
Expr *VarRef =
|
||||
new (S.Context) DeclRefExpr(S.Context, VD, false, T, VK_LValue, Loc);
|
||||
ExprResult Result;
|
||||
auto IE = InitializedEntity::InitializeBlock(Loc, T, false);
|
||||
auto IE = InitializedEntity::InitializeBlock(Loc, T);
|
||||
if (S.getLangOpts().CPlusPlus2b) {
|
||||
auto *E = ImplicitCastExpr::Create(S.Context, T, CK_NoOp, VarRef, nullptr,
|
||||
VK_XValue, FPOptionsOverride());
|
||||
|
@ -1533,7 +1533,7 @@ bool CoroutineStmtBuilder::makeGroDeclAndReturnStmt() {
|
||||
if (GroType->isVoidType()) {
|
||||
// Trigger a nice error message.
|
||||
InitializedEntity Entity =
|
||||
InitializedEntity::InitializeResult(Loc, FnRetType, false);
|
||||
InitializedEntity::InitializeResult(Loc, FnRetType);
|
||||
S.PerformCopyInitialization(Entity, SourceLocation(), ReturnValue);
|
||||
noteMemberDeclaredHere(S, ReturnValue, Fn);
|
||||
return false;
|
||||
|
@ -15262,8 +15262,17 @@ Sema::BuildCXXConstructExpr(SourceLocation ConstructLoc, QualType DeclInitType,
|
||||
// can be omitted by constructing the temporary object
|
||||
// directly into the target of the omitted copy/move
|
||||
if (ConstructKind == CXXConstructExpr::CK_Complete && Constructor &&
|
||||
// FIXME: Converting constructors should also be accepted.
|
||||
// But to fix this, the logic that digs down into a CXXConstructExpr
|
||||
// to find the source object needs to handle it.
|
||||
// Right now it assumes the source object is passed directly as the
|
||||
// first argument.
|
||||
Constructor->isCopyOrMoveConstructor() && hasOneRealArgument(ExprArgs)) {
|
||||
Expr *SubExpr = ExprArgs[0];
|
||||
// FIXME: Per above, this is also incorrect if we want to accept
|
||||
// converting constructors, as isTemporaryObject will
|
||||
// reject temporaries with different type from the
|
||||
// CXXRecord itself.
|
||||
Elidable = SubExpr->isTemporaryObject(
|
||||
Context, cast<CXXRecordDecl>(FoundDecl->getDeclContext()));
|
||||
}
|
||||
|
@ -15683,7 +15683,7 @@ ExprResult Sema::ActOnBlockStmtExpr(SourceLocation CaretLoc,
|
||||
if (!Result.isInvalid()) {
|
||||
Result = PerformCopyInitialization(
|
||||
InitializedEntity::InitializeBlock(Var->getLocation(),
|
||||
Cap.getCaptureType(), false),
|
||||
Cap.getCaptureType()),
|
||||
Loc, Result.get());
|
||||
}
|
||||
|
||||
|
@ -893,9 +893,8 @@ ExprResult Sema::BuildCXXThrow(SourceLocation OpLoc, Expr *Ex,
|
||||
if (CheckCXXThrowOperand(OpLoc, ExceptionObjectTy, Ex))
|
||||
return ExprError();
|
||||
|
||||
InitializedEntity Entity = InitializedEntity::InitializeException(
|
||||
OpLoc, ExceptionObjectTy,
|
||||
/*NRVO=*/NRInfo.isCopyElidable());
|
||||
InitializedEntity Entity =
|
||||
InitializedEntity::InitializeException(OpLoc, ExceptionObjectTy);
|
||||
ExprResult Res = PerformMoveOrCopyInitialization(Entity, NRInfo, Ex);
|
||||
if (Res.isInvalid())
|
||||
return ExprError();
|
||||
|
@ -1975,8 +1975,7 @@ ExprResult Sema::BuildBlockForLambdaConversion(SourceLocation CurrentLocation,
|
||||
CallOperator->markUsed(Context);
|
||||
|
||||
ExprResult Init = PerformCopyInitialization(
|
||||
InitializedEntity::InitializeLambdaToBlock(ConvLocation, Src->getType(),
|
||||
/*NRVO=*/false),
|
||||
InitializedEntity::InitializeLambdaToBlock(ConvLocation, Src->getType()),
|
||||
CurrentLocation, Src);
|
||||
if (!Init.isInvalid())
|
||||
Init = ActOnFinishFullExpr(Init.get(), /*DiscardedValue*/ false);
|
||||
|
@ -1467,8 +1467,7 @@ Decl *Sema::ActOnPropertyImplDecl(Scope *S,
|
||||
LoadSelfExpr, true, true);
|
||||
ExprResult Res = PerformCopyInitialization(
|
||||
InitializedEntity::InitializeResult(PropertyDiagLoc,
|
||||
getterMethod->getReturnType(),
|
||||
/*NRVO=*/false),
|
||||
getterMethod->getReturnType()),
|
||||
PropertyDiagLoc, IvarRefExpr);
|
||||
if (!Res.isInvalid()) {
|
||||
Expr *ResExpr = Res.getAs<Expr>();
|
||||
|
@ -3653,8 +3653,8 @@ StmtResult Sema::ActOnCapScopeReturnStmt(SourceLocation ReturnLoc,
|
||||
|
||||
// In C++ the return statement is handled via a copy initialization.
|
||||
// the C version of which boils down to CheckSingleAssignmentConstraints.
|
||||
InitializedEntity Entity = InitializedEntity::InitializeResult(
|
||||
ReturnLoc, FnRetType, NRVOCandidate != nullptr);
|
||||
InitializedEntity Entity =
|
||||
InitializedEntity::InitializeResult(ReturnLoc, FnRetType);
|
||||
ExprResult Res = PerformMoveOrCopyInitialization(
|
||||
Entity, NRInfo, RetValExp, SupressSimplerImplicitMoves);
|
||||
if (Res.isInvalid()) {
|
||||
@ -4085,8 +4085,8 @@ StmtResult Sema::BuildReturnStmt(SourceLocation ReturnLoc, Expr *RetValExp) {
|
||||
// the C version of which boils down to CheckSingleAssignmentConstraints.
|
||||
if (!HasDependentReturnType && !RetValExp->isTypeDependent()) {
|
||||
// we have a non-void function with an expression, continue checking
|
||||
InitializedEntity Entity = InitializedEntity::InitializeResult(
|
||||
ReturnLoc, RetType, NRVOCandidate != nullptr);
|
||||
InitializedEntity Entity =
|
||||
InitializedEntity::InitializeResult(ReturnLoc, RetType);
|
||||
ExprResult Res = PerformMoveOrCopyInitialization(
|
||||
Entity, NRInfo, RetValExp, SupressSimplerImplicitMoves);
|
||||
if (Res.isInvalid()) {
|
||||
|
@ -609,9 +609,6 @@ void Writer::finalizeAddresses() {
|
||||
void Writer::run() {
|
||||
ScopedTimer t1(codeLayoutTimer);
|
||||
|
||||
// First, clear the output sections from previous runs
|
||||
outputSections.clear();
|
||||
|
||||
createImportTables();
|
||||
createSections();
|
||||
appendImportThunks();
|
||||
|
@ -80,7 +80,7 @@ void DemandedBitsWrapperPass::print(raw_ostream &OS, const Module *M) const {
|
||||
|
||||
static bool isAlwaysLive(Instruction *I) {
|
||||
return I->isTerminator() || isa<DbgInfoIntrinsic>(I) || I->isEHPad() ||
|
||||
I->mayHaveSideEffects() || !I->willReturn();
|
||||
I->mayHaveSideEffects();
|
||||
}
|
||||
|
||||
void DemandedBits::determineLiveOperandBits(
|
||||
|
@ -459,7 +459,7 @@ Constant *llvm::ConstantFoldCastInstruction(unsigned opc, Constant *V,
|
||||
V.convertToInteger(IntVal, APFloat::rmTowardZero, &ignored)) {
|
||||
// Undefined behavior invoked - the destination type can't represent
|
||||
// the input constant.
|
||||
return UndefValue::get(DestTy);
|
||||
return PoisonValue::get(DestTy);
|
||||
}
|
||||
return ConstantInt::get(FPC->getContext(), IntVal);
|
||||
}
|
||||
@ -694,7 +694,7 @@ Constant *llvm::ConstantFoldInsertElementInstruction(Constant *Val,
|
||||
|
||||
unsigned NumElts = ValTy->getNumElements();
|
||||
if (CIdx->uge(NumElts))
|
||||
return UndefValue::get(Val->getType());
|
||||
return PoisonValue::get(Val->getType());
|
||||
|
||||
SmallVector<Constant*, 16> Result;
|
||||
Result.reserve(NumElts);
|
||||
@ -922,23 +922,21 @@ Constant *llvm::ConstantFoldBinaryInstruction(unsigned Opcode, Constant *C1,
|
||||
}
|
||||
case Instruction::SDiv:
|
||||
case Instruction::UDiv:
|
||||
// X / undef -> undef
|
||||
if (isa<UndefValue>(C2))
|
||||
return C2;
|
||||
// undef / 0 -> undef
|
||||
// X / undef -> poison
|
||||
// X / 0 -> poison
|
||||
if (match(C2, m_CombineOr(m_Undef(), m_Zero())))
|
||||
return PoisonValue::get(C2->getType());
|
||||
// undef / 1 -> undef
|
||||
if (match(C2, m_Zero()) || match(C2, m_One()))
|
||||
if (match(C2, m_One()))
|
||||
return C1;
|
||||
// undef / X -> 0 otherwise
|
||||
return Constant::getNullValue(C1->getType());
|
||||
case Instruction::URem:
|
||||
case Instruction::SRem:
|
||||
// X % undef -> undef
|
||||
if (match(C2, m_Undef()))
|
||||
return C2;
|
||||
// undef % 0 -> undef
|
||||
if (match(C2, m_Zero()))
|
||||
return C1;
|
||||
// X % undef -> poison
|
||||
// X % 0 -> poison
|
||||
if (match(C2, m_CombineOr(m_Undef(), m_Zero())))
|
||||
return PoisonValue::get(C2->getType());
|
||||
// undef % X -> 0 otherwise
|
||||
return Constant::getNullValue(C1->getType());
|
||||
case Instruction::Or: // X | undef -> -1
|
||||
@ -946,28 +944,28 @@ Constant *llvm::ConstantFoldBinaryInstruction(unsigned Opcode, Constant *C1,
|
||||
return C1;
|
||||
return Constant::getAllOnesValue(C1->getType()); // undef | X -> ~0
|
||||
case Instruction::LShr:
|
||||
// X >>l undef -> undef
|
||||
// X >>l undef -> poison
|
||||
if (isa<UndefValue>(C2))
|
||||
return C2;
|
||||
return PoisonValue::get(C2->getType());
|
||||
// undef >>l 0 -> undef
|
||||
if (match(C2, m_Zero()))
|
||||
return C1;
|
||||
// undef >>l X -> 0
|
||||
return Constant::getNullValue(C1->getType());
|
||||
case Instruction::AShr:
|
||||
// X >>a undef -> undef
|
||||
// X >>a undef -> poison
|
||||
if (isa<UndefValue>(C2))
|
||||
return C2;
|
||||
return PoisonValue::get(C2->getType());
|
||||
// undef >>a 0 -> undef
|
||||
if (match(C2, m_Zero()))
|
||||
return C1;
|
||||
// TODO: undef >>a X -> undef if the shift is exact
|
||||
// TODO: undef >>a X -> poison if the shift is exact
|
||||
// undef >>a X -> 0
|
||||
return Constant::getNullValue(C1->getType());
|
||||
case Instruction::Shl:
|
||||
// X << undef -> undef
|
||||
if (isa<UndefValue>(C2))
|
||||
return C2;
|
||||
return PoisonValue::get(C2->getType());
|
||||
// undef << 0 -> undef
|
||||
if (match(C2, m_Zero()))
|
||||
return C1;
|
||||
@ -1020,14 +1018,14 @@ Constant *llvm::ConstantFoldBinaryInstruction(unsigned Opcode, Constant *C1,
|
||||
if (CI2->isOne())
|
||||
return C1; // X / 1 == X
|
||||
if (CI2->isZero())
|
||||
return UndefValue::get(CI2->getType()); // X / 0 == undef
|
||||
return PoisonValue::get(CI2->getType()); // X / 0 == poison
|
||||
break;
|
||||
case Instruction::URem:
|
||||
case Instruction::SRem:
|
||||
if (CI2->isOne())
|
||||
return Constant::getNullValue(CI2->getType()); // X % 1 == 0
|
||||
if (CI2->isZero())
|
||||
return UndefValue::get(CI2->getType()); // X % 0 == undef
|
||||
return PoisonValue::get(CI2->getType()); // X % 0 == poison
|
||||
break;
|
||||
case Instruction::And:
|
||||
if (CI2->isZero()) return C2; // X & 0 == 0
|
||||
@ -1141,7 +1139,7 @@ Constant *llvm::ConstantFoldBinaryInstruction(unsigned Opcode, Constant *C1,
|
||||
case Instruction::SDiv:
|
||||
assert(!CI2->isZero() && "Div by zero handled above");
|
||||
if (C2V.isAllOnesValue() && C1V.isMinSignedValue())
|
||||
return UndefValue::get(CI1->getType()); // MIN_INT / -1 -> undef
|
||||
return PoisonValue::get(CI1->getType()); // MIN_INT / -1 -> poison
|
||||
return ConstantInt::get(CI1->getContext(), C1V.sdiv(C2V));
|
||||
case Instruction::URem:
|
||||
assert(!CI2->isZero() && "Div by zero handled above");
|
||||
@ -1149,7 +1147,7 @@ Constant *llvm::ConstantFoldBinaryInstruction(unsigned Opcode, Constant *C1,
|
||||
case Instruction::SRem:
|
||||
assert(!CI2->isZero() && "Div by zero handled above");
|
||||
if (C2V.isAllOnesValue() && C1V.isMinSignedValue())
|
||||
return UndefValue::get(CI1->getType()); // MIN_INT % -1 -> undef
|
||||
return PoisonValue::get(CI1->getType()); // MIN_INT % -1 -> poison
|
||||
return ConstantInt::get(CI1->getContext(), C1V.srem(C2V));
|
||||
case Instruction::And:
|
||||
return ConstantInt::get(CI1->getContext(), C1V & C2V);
|
||||
@ -1160,15 +1158,15 @@ Constant *llvm::ConstantFoldBinaryInstruction(unsigned Opcode, Constant *C1,
|
||||
case Instruction::Shl:
|
||||
if (C2V.ult(C1V.getBitWidth()))
|
||||
return ConstantInt::get(CI1->getContext(), C1V.shl(C2V));
|
||||
return UndefValue::get(C1->getType()); // too big shift is undef
|
||||
return PoisonValue::get(C1->getType()); // too big shift is poison
|
||||
case Instruction::LShr:
|
||||
if (C2V.ult(C1V.getBitWidth()))
|
||||
return ConstantInt::get(CI1->getContext(), C1V.lshr(C2V));
|
||||
return UndefValue::get(C1->getType()); // too big shift is undef
|
||||
return PoisonValue::get(C1->getType()); // too big shift is poison
|
||||
case Instruction::AShr:
|
||||
if (C2V.ult(C1V.getBitWidth()))
|
||||
return ConstantInt::get(CI1->getContext(), C1V.ashr(C2V));
|
||||
return UndefValue::get(C1->getType()); // too big shift is undef
|
||||
return PoisonValue::get(C1->getType()); // too big shift is poison
|
||||
}
|
||||
}
|
||||
|
||||
@ -1214,7 +1212,7 @@ Constant *llvm::ConstantFoldBinaryInstruction(unsigned Opcode, Constant *C1,
|
||||
// Fast path for splatted constants.
|
||||
if (Constant *C2Splat = C2->getSplatValue()) {
|
||||
if (Instruction::isIntDivRem(Opcode) && C2Splat->isNullValue())
|
||||
return UndefValue::get(VTy);
|
||||
return PoisonValue::get(VTy);
|
||||
if (Constant *C1Splat = C1->getSplatValue()) {
|
||||
return ConstantVector::getSplat(
|
||||
VTy->getElementCount(),
|
||||
@ -1231,9 +1229,9 @@ Constant *llvm::ConstantFoldBinaryInstruction(unsigned Opcode, Constant *C1,
|
||||
Constant *LHS = ConstantExpr::getExtractElement(C1, ExtractIdx);
|
||||
Constant *RHS = ConstantExpr::getExtractElement(C2, ExtractIdx);
|
||||
|
||||
// If any element of a divisor vector is zero, the whole op is undef.
|
||||
// If any element of a divisor vector is zero, the whole op is poison.
|
||||
if (Instruction::isIntDivRem(Opcode) && RHS->isNullValue())
|
||||
return UndefValue::get(VTy);
|
||||
return PoisonValue::get(VTy);
|
||||
|
||||
Result.push_back(ConstantExpr::get(Opcode, LHS, RHS));
|
||||
}
|
||||
@ -2209,7 +2207,8 @@ Constant *llvm::ConstantFoldGetElementPtr(Type *PointeeTy, Constant *C,
|
||||
return PoisonValue::get(GEPTy);
|
||||
|
||||
if (isa<UndefValue>(C))
|
||||
return UndefValue::get(GEPTy);
|
||||
// If inbounds, we can choose an out-of-bounds pointer as a base pointer.
|
||||
return InBounds ? PoisonValue::get(GEPTy) : UndefValue::get(GEPTy);
|
||||
|
||||
Constant *Idx0 = cast<Constant>(Idxs[0]);
|
||||
if (Idxs.size() == 1 && (Idx0->isNullValue() || isa<UndefValue>(Idx0)))
|
||||
|
@ -1572,9 +1572,6 @@ PassBuilder::buildThinLTOPreLinkDefaultPipeline(OptimizationLevel Level) {
|
||||
for (auto &C : OptimizerLastEPCallbacks)
|
||||
C(MPM, Level);
|
||||
|
||||
if (PGOOpt && PGOOpt->PseudoProbeForProfiling)
|
||||
MPM.addPass(PseudoProbeUpdatePass());
|
||||
|
||||
// Emit annotation remarks.
|
||||
addAnnotationRemarksPass(MPM);
|
||||
|
||||
|
@ -1301,6 +1301,7 @@ static AArch64CC::CondCode changeICMPPredToAArch64CC(CmpInst::Predicate P) {
|
||||
static Register getTestBitReg(Register Reg, uint64_t &Bit, bool &Invert,
|
||||
MachineRegisterInfo &MRI) {
|
||||
assert(Reg.isValid() && "Expected valid register!");
|
||||
bool HasZext = false;
|
||||
while (MachineInstr *MI = getDefIgnoringCopies(Reg, MRI)) {
|
||||
unsigned Opc = MI->getOpcode();
|
||||
|
||||
@ -1314,6 +1315,9 @@ static Register getTestBitReg(Register Reg, uint64_t &Bit, bool &Invert,
|
||||
// on the truncated x is the same as the bit number on x.
|
||||
if (Opc == TargetOpcode::G_ANYEXT || Opc == TargetOpcode::G_ZEXT ||
|
||||
Opc == TargetOpcode::G_TRUNC) {
|
||||
if (Opc == TargetOpcode::G_ZEXT)
|
||||
HasZext = true;
|
||||
|
||||
Register NextReg = MI->getOperand(1).getReg();
|
||||
// Did we find something worth folding?
|
||||
if (!NextReg.isValid() || !MRI.hasOneNonDBGUse(NextReg))
|
||||
@ -1342,8 +1346,12 @@ static Register getTestBitReg(Register Reg, uint64_t &Bit, bool &Invert,
|
||||
std::swap(ConstantReg, TestReg);
|
||||
VRegAndVal = getConstantVRegValWithLookThrough(ConstantReg, MRI);
|
||||
}
|
||||
if (VRegAndVal)
|
||||
if (VRegAndVal) {
|
||||
if (HasZext)
|
||||
C = VRegAndVal->Value.getZExtValue();
|
||||
else
|
||||
C = VRegAndVal->Value.getSExtValue();
|
||||
}
|
||||
break;
|
||||
}
|
||||
case TargetOpcode::G_ASHR:
|
||||
|
@ -35823,7 +35823,7 @@ static SDValue combineX86ShuffleChain(ArrayRef<SDValue> Inputs, SDValue Root,
|
||||
|
||||
// See if the shuffle is a hidden identity shuffle - repeated args in HOPs
|
||||
// etc. can be simplified.
|
||||
if (VT1 == VT2 && VT1.getSizeInBits() == RootSizeInBits) {
|
||||
if (VT1 == VT2 && VT1.getSizeInBits() == RootSizeInBits && VT1.isVector()) {
|
||||
SmallVector<int> ScaledMask, IdentityMask;
|
||||
unsigned NumElts = VT1.getVectorNumElements();
|
||||
if (BaseMask.size() <= NumElts &&
|
||||
|
@ -326,7 +326,7 @@ void AggressiveDeadCodeElimination::initialize() {
|
||||
|
||||
bool AggressiveDeadCodeElimination::isAlwaysLive(Instruction &I) {
|
||||
// TODO -- use llvm::isInstructionTriviallyDead
|
||||
if (I.isEHPad() || I.mayHaveSideEffects() || !I.willReturn()) {
|
||||
if (I.isEHPad() || I.mayHaveSideEffects()) {
|
||||
// Skip any value profile instrumentation calls if they are
|
||||
// instrumenting constants.
|
||||
if (isInstrumentsConstant(I))
|
||||
|
@ -1,14 +1,14 @@
|
||||
// $FreeBSD$
|
||||
|
||||
#define LLVM_REVISION "llvmorg-13.0.0-rc3-8-g08642a395f23"
|
||||
#define LLVM_REVISION "llvmorg-13.0.0-0-gd7b669b3a303"
|
||||
#define LLVM_REPOSITORY "git@github.com:llvm/llvm-project.git"
|
||||
|
||||
#define CLANG_REVISION "llvmorg-13.0.0-rc3-8-g08642a395f23"
|
||||
#define CLANG_REVISION "llvmorg-13.0.0-0-gd7b669b3a303"
|
||||
#define CLANG_REPOSITORY "git@github.com:llvm/llvm-project.git"
|
||||
|
||||
// <Upstream revision at import>-<Local identifier in __FreeBSD_version style>
|
||||
#define LLD_REVISION "llvmorg-13.0.0-rc3-8-g08642a395f23-1400002"
|
||||
#define LLD_REVISION "llvmorg-13.0.0-0-gd7b669b3a303-1400002"
|
||||
#define LLD_REPOSITORY "FreeBSD"
|
||||
|
||||
#define LLDB_REVISION "llvmorg-13.0.0-rc3-8-g08642a395f23"
|
||||
#define LLDB_REVISION "llvmorg-13.0.0-0-gd7b669b3a303"
|
||||
#define LLDB_REPOSITORY "git@github.com:llvm/llvm-project.git"
|
||||
|
@ -1,3 +1,3 @@
|
||||
/* $FreeBSD$ */
|
||||
#define LLVM_REVISION "llvmorg-13.0.0-rc3-8-g08642a395f23"
|
||||
#define LLVM_REVISION "llvmorg-13.0.0-0-gd7b669b3a303"
|
||||
#define LLVM_REPOSITORY "git@github.com:llvm/llvm-project.git"
|
||||
|
@ -18,6 +18,22 @@ SRCS+= Win64EHDumper.cpp
|
||||
SRCS+= WindowsResourceDumper.cpp
|
||||
SRCS+= XCOFFDumper.cpp
|
||||
|
||||
.include "${SRCTOP}/lib/clang/llvm.pre.mk"
|
||||
|
||||
CFLAGS+= -I${.OBJDIR}
|
||||
|
||||
INCFILE= Opts.inc
|
||||
TDFILE= ${LLVM_BASE}/${SRCDIR}/Opts.td
|
||||
GENOPT= -gen-opt-parser-defs
|
||||
${INCFILE}: ${TDFILE}
|
||||
${LLVM_TBLGEN} ${GENOPT} -I ${LLVM_SRCS}/include -d ${.TARGET:C/$/.d/} \
|
||||
-o ${.TARGET} ${TDFILE}
|
||||
TGHDRS+= ${INCFILE}
|
||||
|
||||
DEPENDFILES+= ${TGHDRS:C/$/.d/}
|
||||
DPSRCS+= ${TGHDRS}
|
||||
CLEANFILES+= ${TGHDRS} ${TGHDRS:C/$/.d/}
|
||||
|
||||
LIBADD+= z
|
||||
|
||||
LINKS+= ${BINDIR}/llvm-readobj ${BINDIR}/llvm-readelf
|
||||
|
Loading…
Reference in New Issue
Block a user