Upgrade our copies of clang, llvm and libc++ to r310316 from the

upstream release_50 branch.

MFC after:	2 months
X-MFC-with:	r321369
This commit is contained in:
Dimitry Andric 2017-08-09 17:32:39 +00:00
Notes: svn2git 2020-12-20 02:59:44 +00:00
svn path=/head/; revision=322320
48 changed files with 479 additions and 228 deletions

View File

@ -15,6 +15,10 @@
#ifndef _LIBCPP_BSD_LOCALE_DEFAULTS_H #ifndef _LIBCPP_BSD_LOCALE_DEFAULTS_H
#define _LIBCPP_BSD_LOCALE_DEFAULTS_H #define _LIBCPP_BSD_LOCALE_DEFAULTS_H
#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER)
#pragma GCC system_header
#endif
#define __libcpp_mb_cur_max_l(loc) MB_CUR_MAX_L(loc) #define __libcpp_mb_cur_max_l(loc) MB_CUR_MAX_L(loc)
#define __libcpp_btowc_l(ch, loc) btowc_l(ch, loc) #define __libcpp_btowc_l(ch, loc) btowc_l(ch, loc)
#define __libcpp_wctob_l(wch, loc) wctob_l(wch, loc) #define __libcpp_wctob_l(wch, loc) wctob_l(wch, loc)

View File

@ -18,6 +18,10 @@
#include <stdarg.h> #include <stdarg.h>
#include <memory> #include <memory>
#if !defined(_LIBCPP_HAS_NO_PRAGMA_SYSTEM_HEADER)
#pragma GCC system_header
#endif
_LIBCPP_BEGIN_NAMESPACE_STD _LIBCPP_BEGIN_NAMESPACE_STD
inline _LIBCPP_ALWAYS_INLINE inline _LIBCPP_ALWAYS_INLINE

View File

@ -34,7 +34,7 @@
# include <support/solaris/xlocale.h> # include <support/solaris/xlocale.h>
#elif defined(_NEWLIB_VERSION) #elif defined(_NEWLIB_VERSION)
# include <support/newlib/xlocale.h> # include <support/newlib/xlocale.h>
#elif (defined(__GLIBC__) || defined(__APPLE__) || defined(__FreeBSD__) \ #elif (defined(__APPLE__) || defined(__FreeBSD__) \
|| defined(__EMSCRIPTEN__) || defined(__IBMCPP__)) || defined(__EMSCRIPTEN__) || defined(__IBMCPP__))
# include <xlocale.h> # include <xlocale.h>
#elif defined(__Fuchsia__) #elif defined(__Fuchsia__)

View File

@ -116,7 +116,7 @@ public:
using mutex_type = Mutex; // If MutexTypes... consists of the single type Mutex using mutex_type = Mutex; // If MutexTypes... consists of the single type Mutex
explicit scoped_lock(MutexTypes&... m); explicit scoped_lock(MutexTypes&... m);
scoped_lock(MutexTypes&... m, adopt_lock_t); scoped_lock(adopt_lock_t, MutexTypes&... m);
~scoped_lock(); ~scoped_lock();
scoped_lock(scoped_lock const&) = delete; scoped_lock(scoped_lock const&) = delete;
scoped_lock& operator=(scoped_lock const&) = delete; scoped_lock& operator=(scoped_lock const&) = delete;
@ -500,7 +500,7 @@ public:
~scoped_lock() _LIBCPP_THREAD_SAFETY_ANNOTATION(release_capability()) {__m_.unlock();} ~scoped_lock() _LIBCPP_THREAD_SAFETY_ANNOTATION(release_capability()) {__m_.unlock();}
_LIBCPP_INLINE_VISIBILITY _LIBCPP_INLINE_VISIBILITY
explicit scoped_lock(mutex_type& __m, adopt_lock_t) _LIBCPP_THREAD_SAFETY_ANNOTATION(requires_capability(__m)) explicit scoped_lock(adopt_lock_t, mutex_type& __m) _LIBCPP_THREAD_SAFETY_ANNOTATION(requires_capability(__m))
: __m_(__m) {} : __m_(__m) {}
scoped_lock(scoped_lock const&) = delete; scoped_lock(scoped_lock const&) = delete;
@ -522,7 +522,7 @@ public:
} }
_LIBCPP_INLINE_VISIBILITY _LIBCPP_INLINE_VISIBILITY
scoped_lock(_MArgs&... __margs, adopt_lock_t) scoped_lock(adopt_lock_t, _MArgs&... __margs)
: __t_(__margs...) : __t_(__margs...)
{ {
} }

View File

@ -249,7 +249,8 @@ basic_stringbuf<_CharT, _Traits, _Allocator>::basic_stringbuf(ios_base::openmode
template <class _CharT, class _Traits, class _Allocator> template <class _CharT, class _Traits, class _Allocator>
basic_stringbuf<_CharT, _Traits, _Allocator>::basic_stringbuf(const string_type& __s, basic_stringbuf<_CharT, _Traits, _Allocator>::basic_stringbuf(const string_type& __s,
ios_base::openmode __wch) ios_base::openmode __wch)
: __hm_(0), : __str_(__s.get_allocator()),
__hm_(0),
__mode_(__wch) __mode_(__wch)
{ {
str(__s); str(__s);

View File

@ -312,6 +312,12 @@ template <typename T> class ArrayRef;
const DataLayout &DL, LoopInfo *LI = nullptr, const DataLayout &DL, LoopInfo *LI = nullptr,
unsigned MaxLookup = 6); unsigned MaxLookup = 6);
/// This is a wrapper around GetUnderlyingObjects and adds support for basic
/// ptrtoint+arithmetic+inttoptr sequences.
void getUnderlyingObjectsForCodeGen(const Value *V,
SmallVectorImpl<Value *> &Objects,
const DataLayout &DL);
/// Return true if the only users of this pointer are lifetime markers. /// Return true if the only users of this pointer are lifetime markers.
bool onlyUsedByLifetimeMarkers(const Value *V); bool onlyUsedByLifetimeMarkers(const Value *V);

View File

@ -661,6 +661,12 @@ class MachineFunction {
MachineMemOperand *getMachineMemOperand(const MachineMemOperand *MMO, MachineMemOperand *getMachineMemOperand(const MachineMemOperand *MMO,
int64_t Offset, uint64_t Size); int64_t Offset, uint64_t Size);
/// Allocate a new MachineMemOperand by copying an existing one,
/// replacing only AliasAnalysis information. MachineMemOperands are owned
/// by the MachineFunction and need not be explicitly deallocated.
MachineMemOperand *getMachineMemOperand(const MachineMemOperand *MMO,
const AAMDNodes &AAInfo);
using OperandCapacity = ArrayRecycler<MachineOperand>::Capacity; using OperandCapacity = ArrayRecycler<MachineOperand>::Capacity;
/// Allocate an array of MachineOperands. This is only intended for use by /// Allocate an array of MachineOperands. This is only intended for use by

View File

@ -379,6 +379,9 @@ class MachineInstr
return NumMemRefs == 1; return NumMemRefs == 1;
} }
/// Return the number of memory operands.
unsigned getNumMemOperands() const { return NumMemRefs; }
/// API for querying MachineInstr properties. They are the same as MCInstrDesc /// API for querying MachineInstr properties. They are the same as MCInstrDesc
/// queries but they are bundle aware. /// queries but they are bundle aware.

View File

@ -3277,6 +3277,69 @@ void llvm::GetUnderlyingObjects(Value *V, SmallVectorImpl<Value *> &Objects,
} while (!Worklist.empty()); } while (!Worklist.empty());
} }
/// This is the function that does the work of looking through basic
/// ptrtoint+arithmetic+inttoptr sequences.
static const Value *getUnderlyingObjectFromInt(const Value *V) {
do {
if (const Operator *U = dyn_cast<Operator>(V)) {
// If we find a ptrtoint, we can transfer control back to the
// regular getUnderlyingObjectFromInt.
if (U->getOpcode() == Instruction::PtrToInt)
return U->getOperand(0);
// If we find an add of a constant, a multiplied value, or a phi, it's
// likely that the other operand will lead us to the base
// object. We don't have to worry about the case where the
// object address is somehow being computed by the multiply,
// because our callers only care when the result is an
// identifiable object.
if (U->getOpcode() != Instruction::Add ||
(!isa<ConstantInt>(U->getOperand(1)) &&
Operator::getOpcode(U->getOperand(1)) != Instruction::Mul &&
!isa<PHINode>(U->getOperand(1))))
return V;
V = U->getOperand(0);
} else {
return V;
}
assert(V->getType()->isIntegerTy() && "Unexpected operand type!");
} while (true);
}
/// This is a wrapper around GetUnderlyingObjects and adds support for basic
/// ptrtoint+arithmetic+inttoptr sequences.
void llvm::getUnderlyingObjectsForCodeGen(const Value *V,
SmallVectorImpl<Value *> &Objects,
const DataLayout &DL) {
SmallPtrSet<const Value *, 16> Visited;
SmallVector<const Value *, 4> Working(1, V);
do {
V = Working.pop_back_val();
SmallVector<Value *, 4> Objs;
GetUnderlyingObjects(const_cast<Value *>(V), Objs, DL);
for (Value *V : Objs) {
if (!Visited.insert(V).second)
continue;
if (Operator::getOpcode(V) == Instruction::IntToPtr) {
const Value *O =
getUnderlyingObjectFromInt(cast<User>(V)->getOperand(0));
if (O->getType()->isPointerTy()) {
Working.push_back(O);
continue;
}
}
// If GetUnderlyingObjects fails to find an identifiable object,
// getUnderlyingObjectsForCodeGen also fails for safety.
if (!isIdentifiedObject(V)) {
Objects.clear();
return;
}
Objects.push_back(const_cast<Value *>(V));
}
} while (!Working.empty());
}
/// Return true if the only users of this pointer are lifetime markers. /// Return true if the only users of this pointer are lifetime markers.
bool llvm::onlyUsedByLifetimeMarkers(const Value *V) { bool llvm::onlyUsedByLifetimeMarkers(const Value *V) {
for (const User *U : V->users()) { for (const User *U : V->users()) {

View File

@ -1475,13 +1475,14 @@ bool BranchFolder::OptimizeBlock(MachineBasicBlock *MBB) {
bool PredAnalyzable = bool PredAnalyzable =
!TII->analyzeBranch(*Pred, PredTBB, PredFBB, PredCond, true); !TII->analyzeBranch(*Pred, PredTBB, PredFBB, PredCond, true);
if (PredAnalyzable && !PredCond.empty() && PredTBB == MBB) { if (PredAnalyzable && !PredCond.empty() && PredTBB == MBB &&
PredTBB != PredFBB) {
// The predecessor has a conditional branch to this block which consists // The predecessor has a conditional branch to this block which consists
// of only a tail call. Try to fold the tail call into the conditional // of only a tail call. Try to fold the tail call into the conditional
// branch. // branch.
if (TII->canMakeTailCallConditional(PredCond, TailCall)) { if (TII->canMakeTailCallConditional(PredCond, TailCall)) {
// TODO: It would be nice if analyzeBranch() could provide a pointer // TODO: It would be nice if analyzeBranch() could provide a pointer
// to the branch insturction so replaceBranchWithTailCall() doesn't // to the branch instruction so replaceBranchWithTailCall() doesn't
// have to search for it. // have to search for it.
TII->replaceBranchWithTailCall(*Pred, PredCond, TailCall); TII->replaceBranchWithTailCall(*Pred, PredCond, TailCall);
++NumTailCalls; ++NumTailCalls;

View File

@ -330,6 +330,20 @@ MachineFunction::getMachineMemOperand(const MachineMemOperand *MMO,
MMO->getOrdering(), MMO->getFailureOrdering()); MMO->getOrdering(), MMO->getFailureOrdering());
} }
MachineMemOperand *
MachineFunction::getMachineMemOperand(const MachineMemOperand *MMO,
const AAMDNodes &AAInfo) {
MachinePointerInfo MPI = MMO->getValue() ?
MachinePointerInfo(MMO->getValue(), MMO->getOffset()) :
MachinePointerInfo(MMO->getPseudoValue(), MMO->getOffset());
return new (Allocator)
MachineMemOperand(MPI, MMO->getFlags(), MMO->getSize(),
MMO->getBaseAlignment(), AAInfo,
MMO->getRanges(), MMO->getSyncScopeID(),
MMO->getOrdering(), MMO->getFailureOrdering());
}
MachineInstr::mmo_iterator MachineInstr::mmo_iterator
MachineFunction::allocateMemRefsArray(unsigned long Num) { MachineFunction::allocateMemRefsArray(unsigned long Num) {
return Allocator.Allocate<MachineMemOperand *>(Num); return Allocator.Allocate<MachineMemOperand *>(Num);

View File

@ -578,10 +578,8 @@ bool MachinePointerInfo::isDereferenceable(unsigned Size, LLVMContext &C,
if (BasePtr == nullptr) if (BasePtr == nullptr)
return false; return false;
return isDereferenceableAndAlignedPointer(BasePtr, 1, return isDereferenceableAndAlignedPointer(
APInt(DL.getPointerSize(), BasePtr, 1, APInt(DL.getPointerSizeInBits(), Offset + Size), DL);
Offset + Size),
DL);
} }
/// getConstantPool - Return a MachinePointerInfo record that refers to the /// getConstantPool - Return a MachinePointerInfo record that refers to the

View File

@ -121,63 +121,6 @@ ScheduleDAGInstrs::ScheduleDAGInstrs(MachineFunction &mf,
SchedModel.init(ST.getSchedModel(), &ST, TII); SchedModel.init(ST.getSchedModel(), &ST, TII);
} }
/// This is the function that does the work of looking through basic
/// ptrtoint+arithmetic+inttoptr sequences.
static const Value *getUnderlyingObjectFromInt(const Value *V) {
do {
if (const Operator *U = dyn_cast<Operator>(V)) {
// If we find a ptrtoint, we can transfer control back to the
// regular getUnderlyingObjectFromInt.
if (U->getOpcode() == Instruction::PtrToInt)
return U->getOperand(0);
// If we find an add of a constant, a multiplied value, or a phi, it's
// likely that the other operand will lead us to the base
// object. We don't have to worry about the case where the
// object address is somehow being computed by the multiply,
// because our callers only care when the result is an
// identifiable object.
if (U->getOpcode() != Instruction::Add ||
(!isa<ConstantInt>(U->getOperand(1)) &&
Operator::getOpcode(U->getOperand(1)) != Instruction::Mul &&
!isa<PHINode>(U->getOperand(1))))
return V;
V = U->getOperand(0);
} else {
return V;
}
assert(V->getType()->isIntegerTy() && "Unexpected operand type!");
} while (true);
}
/// This is a wrapper around GetUnderlyingObjects and adds support for basic
/// ptrtoint+arithmetic+inttoptr sequences.
static void getUnderlyingObjects(const Value *V,
SmallVectorImpl<Value *> &Objects,
const DataLayout &DL) {
SmallPtrSet<const Value *, 16> Visited;
SmallVector<const Value *, 4> Working(1, V);
do {
V = Working.pop_back_val();
SmallVector<Value *, 4> Objs;
GetUnderlyingObjects(const_cast<Value *>(V), Objs, DL);
for (Value *V : Objs) {
if (!Visited.insert(V).second)
continue;
if (Operator::getOpcode(V) == Instruction::IntToPtr) {
const Value *O =
getUnderlyingObjectFromInt(cast<User>(V)->getOperand(0));
if (O->getType()->isPointerTy()) {
Working.push_back(O);
continue;
}
}
Objects.push_back(const_cast<Value *>(V));
}
} while (!Working.empty());
}
/// If this machine instr has memory reference information and it can be tracked /// If this machine instr has memory reference information and it can be tracked
/// to a normal reference to a known object, return the Value for that object. /// to a normal reference to a known object, return the Value for that object.
static void getUnderlyingObjectsForInstr(const MachineInstr *MI, static void getUnderlyingObjectsForInstr(const MachineInstr *MI,
@ -208,12 +151,10 @@ static void getUnderlyingObjectsForInstr(const MachineInstr *MI,
Objects.push_back(UnderlyingObjectsVector::value_type(PSV, MayAlias)); Objects.push_back(UnderlyingObjectsVector::value_type(PSV, MayAlias));
} else if (const Value *V = MMO->getValue()) { } else if (const Value *V = MMO->getValue()) {
SmallVector<Value *, 4> Objs; SmallVector<Value *, 4> Objs;
getUnderlyingObjects(V, Objs, DL); getUnderlyingObjectsForCodeGen(V, Objs, DL);
for (Value *V : Objs) { for (Value *V : Objs) {
if (!isIdentifiedObject(V)) assert(isIdentifiedObject(V));
return false;
Objects.push_back(UnderlyingObjectsVector::value_type(V, true)); Objects.push_back(UnderlyingObjectsVector::value_type(V, true));
} }
} else } else

View File

@ -99,6 +99,27 @@ LimitFPPrecision("limit-float-precision",
// store [4096 x i8] %data, [4096 x i8]* %buffer // store [4096 x i8] %data, [4096 x i8]* %buffer
static const unsigned MaxParallelChains = 64; static const unsigned MaxParallelChains = 64;
// True if the Value passed requires ABI mangling as it is a parameter to a
// function or a return value from a function which is not an intrinsic.
static bool isABIRegCopy(const Value * V) {
const bool IsRetInst = V && isa<ReturnInst>(V);
const bool IsCallInst = V && isa<CallInst>(V);
const bool IsInLineAsm =
IsCallInst && static_cast<const CallInst *>(V)->isInlineAsm();
const bool IsIndirectFunctionCall =
IsCallInst && !IsInLineAsm &&
!static_cast<const CallInst *>(V)->getCalledFunction();
// It is possible that the call instruction is an inline asm statement or an
// indirect function call in which case the return value of
// getCalledFunction() would be nullptr.
const bool IsInstrinsicCall =
IsCallInst && !IsInLineAsm && !IsIndirectFunctionCall &&
static_cast<const CallInst *>(V)->getCalledFunction()->getIntrinsicID() !=
Intrinsic::not_intrinsic;
return IsRetInst || (IsCallInst && (!IsInLineAsm && !IsInstrinsicCall));
}
static SDValue getCopyFromPartsVector(SelectionDAG &DAG, const SDLoc &DL, static SDValue getCopyFromPartsVector(SelectionDAG &DAG, const SDLoc &DL,
const SDValue *Parts, unsigned NumParts, const SDValue *Parts, unsigned NumParts,
MVT PartVT, EVT ValueVT, const Value *V, MVT PartVT, EVT ValueVT, const Value *V,
@ -1026,13 +1047,9 @@ SDValue SelectionDAGBuilder::getCopyFromRegs(const Value *V, Type *Ty) {
if (It != FuncInfo.ValueMap.end()) { if (It != FuncInfo.ValueMap.end()) {
unsigned InReg = It->second; unsigned InReg = It->second;
bool IsABIRegCopy =
V && ((isa<CallInst>(V) &&
!(static_cast<const CallInst *>(V))->isInlineAsm()) ||
isa<ReturnInst>(V));
RegsForValue RFV(*DAG.getContext(), DAG.getTargetLoweringInfo(), RegsForValue RFV(*DAG.getContext(), DAG.getTargetLoweringInfo(),
DAG.getDataLayout(), InReg, Ty, IsABIRegCopy); DAG.getDataLayout(), InReg, Ty, isABIRegCopy(V));
SDValue Chain = DAG.getEntryNode(); SDValue Chain = DAG.getEntryNode();
Result = RFV.getCopyFromRegs(DAG, FuncInfo, getCurSDLoc(), Chain, nullptr, Result = RFV.getCopyFromRegs(DAG, FuncInfo, getCurSDLoc(), Chain, nullptr,
V); V);
@ -1221,13 +1238,9 @@ SDValue SelectionDAGBuilder::getValueImpl(const Value *V) {
// If this is an instruction which fast-isel has deferred, select it now. // If this is an instruction which fast-isel has deferred, select it now.
if (const Instruction *Inst = dyn_cast<Instruction>(V)) { if (const Instruction *Inst = dyn_cast<Instruction>(V)) {
unsigned InReg = FuncInfo.InitializeRegForValue(Inst); unsigned InReg = FuncInfo.InitializeRegForValue(Inst);
bool IsABIRegCopy =
V && ((isa<CallInst>(V) &&
!(static_cast<const CallInst *>(V))->isInlineAsm()) ||
isa<ReturnInst>(V));
RegsForValue RFV(*DAG.getContext(), TLI, DAG.getDataLayout(), InReg, RegsForValue RFV(*DAG.getContext(), TLI, DAG.getDataLayout(), InReg,
Inst->getType(), IsABIRegCopy); Inst->getType(), isABIRegCopy(V));
SDValue Chain = DAG.getEntryNode(); SDValue Chain = DAG.getEntryNode();
return RFV.getCopyFromRegs(DAG, FuncInfo, getCurSDLoc(), Chain, nullptr, V); return RFV.getCopyFromRegs(DAG, FuncInfo, getCurSDLoc(), Chain, nullptr, V);
} }
@ -8281,13 +8294,9 @@ SelectionDAGBuilder::CopyValueToVirtualRegister(const Value *V, unsigned Reg) {
const TargetLowering &TLI = DAG.getTargetLoweringInfo(); const TargetLowering &TLI = DAG.getTargetLoweringInfo();
// If this is an InlineAsm we have to match the registers required, not the // If this is an InlineAsm we have to match the registers required, not the
// notional registers required by the type. // notional registers required by the type.
bool IsABIRegCopy =
V && ((isa<CallInst>(V) &&
!(static_cast<const CallInst *>(V))->isInlineAsm()) ||
isa<ReturnInst>(V));
RegsForValue RFV(V->getContext(), TLI, DAG.getDataLayout(), Reg, RegsForValue RFV(V->getContext(), TLI, DAG.getDataLayout(), Reg,
V->getType(), IsABIRegCopy); V->getType(), isABIRegCopy(V));
SDValue Chain = DAG.getEntryNode(); SDValue Chain = DAG.getEntryNode();
ISD::NodeType ExtendType = (FuncInfo.PreferredExtendType.find(V) == ISD::NodeType ExtendType = (FuncInfo.PreferredExtendType.find(V) ==

View File

@ -37,6 +37,7 @@
#include "llvm/CodeGen/MachineRegisterInfo.h" #include "llvm/CodeGen/MachineRegisterInfo.h"
#include "llvm/CodeGen/Passes.h" #include "llvm/CodeGen/Passes.h"
#include "llvm/CodeGen/PseudoSourceValue.h" #include "llvm/CodeGen/PseudoSourceValue.h"
#include "llvm/CodeGen/SelectionDAGNodes.h"
#include "llvm/CodeGen/SlotIndexes.h" #include "llvm/CodeGen/SlotIndexes.h"
#include "llvm/CodeGen/StackProtector.h" #include "llvm/CodeGen/StackProtector.h"
#include "llvm/CodeGen/WinEHFuncInfo.h" #include "llvm/CodeGen/WinEHFuncInfo.h"
@ -889,6 +890,10 @@ void StackColoring::remapInstructions(DenseMap<int, int> &SlotRemap) {
// Keep a list of *allocas* which need to be remapped. // Keep a list of *allocas* which need to be remapped.
DenseMap<const AllocaInst*, const AllocaInst*> Allocas; DenseMap<const AllocaInst*, const AllocaInst*> Allocas;
// Keep a list of allocas which has been affected by the remap.
SmallPtrSet<const AllocaInst*, 32> MergedAllocas;
for (const std::pair<int, int> &SI : SlotRemap) { for (const std::pair<int, int> &SI : SlotRemap) {
const AllocaInst *From = MFI->getObjectAllocation(SI.first); const AllocaInst *From = MFI->getObjectAllocation(SI.first);
const AllocaInst *To = MFI->getObjectAllocation(SI.second); const AllocaInst *To = MFI->getObjectAllocation(SI.second);
@ -908,6 +913,10 @@ void StackColoring::remapInstructions(DenseMap<int, int> &SlotRemap) {
Inst = Cast; Inst = Cast;
} }
// We keep both slots to maintain AliasAnalysis metadata later.
MergedAllocas.insert(From);
MergedAllocas.insert(To);
// Allow the stack protector to adjust its value map to account for the // Allow the stack protector to adjust its value map to account for the
// upcoming replacement. // upcoming replacement.
SP->adjustForColoring(From, To); SP->adjustForColoring(From, To);
@ -939,13 +948,6 @@ void StackColoring::remapInstructions(DenseMap<int, int> &SlotRemap) {
// Update the MachineMemOperand to use the new alloca. // Update the MachineMemOperand to use the new alloca.
for (MachineMemOperand *MMO : I.memoperands()) { for (MachineMemOperand *MMO : I.memoperands()) {
// FIXME: In order to enable the use of TBAA when using AA in CodeGen,
// we'll also need to update the TBAA nodes in MMOs with values
// derived from the merged allocas. When doing this, we'll need to use
// the same variant of GetUnderlyingObjects that is used by the
// instruction scheduler (that can look through ptrtoint/inttoptr
// pairs).
// We've replaced IR-level uses of the remapped allocas, so we only // We've replaced IR-level uses of the remapped allocas, so we only
// need to replace direct uses here. // need to replace direct uses here.
const AllocaInst *AI = dyn_cast_or_null<AllocaInst>(MMO->getValue()); const AllocaInst *AI = dyn_cast_or_null<AllocaInst>(MMO->getValue());
@ -997,6 +999,48 @@ void StackColoring::remapInstructions(DenseMap<int, int> &SlotRemap) {
MO.setIndex(ToSlot); MO.setIndex(ToSlot);
FixedInstr++; FixedInstr++;
} }
// We adjust AliasAnalysis information for merged stack slots.
MachineSDNode::mmo_iterator NewMemOps =
MF->allocateMemRefsArray(I.getNumMemOperands());
unsigned MemOpIdx = 0;
bool ReplaceMemOps = false;
for (MachineMemOperand *MMO : I.memoperands()) {
// If this memory location can be a slot remapped here,
// we remove AA information.
bool MayHaveConflictingAAMD = false;
if (MMO->getAAInfo()) {
if (const Value *MMOV = MMO->getValue()) {
SmallVector<Value *, 4> Objs;
getUnderlyingObjectsForCodeGen(MMOV, Objs, MF->getDataLayout());
if (Objs.empty())
MayHaveConflictingAAMD = true;
else
for (Value *V : Objs) {
// If this memory location comes from a known stack slot
// that is not remapped, we continue checking.
// Otherwise, we need to invalidate AA infomation.
const AllocaInst *AI = dyn_cast_or_null<AllocaInst>(V);
if (AI && MergedAllocas.count(AI)) {
MayHaveConflictingAAMD = true;
break;
}
}
}
}
if (MayHaveConflictingAAMD) {
NewMemOps[MemOpIdx++] = MF->getMachineMemOperand(MMO, AAMDNodes());
ReplaceMemOps = true;
}
else
NewMemOps[MemOpIdx++] = MMO;
}
// If any memory operand is updated, set memory references of
// this instruction.
if (ReplaceMemOps)
I.setMemRefs(std::make_pair(NewMemOps, I.getNumMemOperands()));
} }
// Update the location of C++ catch objects for the MSVC personality routine. // Update the location of C++ catch objects for the MSVC personality routine.

View File

@ -2097,15 +2097,19 @@ Constant *llvm::ConstantFoldGetElementPtr(Type *PointeeTy, Constant *C,
// Subsequent evaluation would get confused and produce erroneous results. // Subsequent evaluation would get confused and produce erroneous results.
// //
// The following prohibits such a GEP from being formed by checking to see // The following prohibits such a GEP from being formed by checking to see
// if the index is in-range with respect to an array or vector. // if the index is in-range with respect to an array.
// TODO: This code may be extended to handle vectors as well.
bool PerformFold = false; bool PerformFold = false;
if (Idx0->isNullValue()) if (Idx0->isNullValue())
PerformFold = true; PerformFold = true;
else if (LastI.isSequential()) else if (LastI.isSequential())
if (ConstantInt *CI = dyn_cast<ConstantInt>(Idx0)) if (ConstantInt *CI = dyn_cast<ConstantInt>(Idx0))
PerformFold = PerformFold = (!LastI.isBoundedSequential() ||
!LastI.isBoundedSequential() || isIndexInRangeOfArrayType(
isIndexInRangeOfArrayType(LastI.getSequentialNumElements(), CI); LastI.getSequentialNumElements(), CI)) &&
!CE->getOperand(CE->getNumOperands() - 1)
->getType()
->isVectorTy();
if (PerformFold) { if (PerformFold) {
SmallVector<Value*, 16> NewIndices; SmallVector<Value*, 16> NewIndices;

View File

@ -542,15 +542,12 @@ NewArchiveMember ObjectFactory::createWeakExternal(StringRef Sym,
SymbolTable[2].Name.Offset.Offset = sizeof(uint32_t); SymbolTable[2].Name.Offset.Offset = sizeof(uint32_t);
//__imp_ String Table //__imp_ String Table
if (Imp) { StringRef Prefix = Imp ? "__imp_" : "";
SymbolTable[3].Name.Offset.Offset = sizeof(uint32_t) + Sym.size() + 7; SymbolTable[3].Name.Offset.Offset =
writeStringTable(Buffer, {std::string("__imp_").append(Sym), sizeof(uint32_t) + Sym.size() + Prefix.size() + 1;
std::string("__imp_").append(Weak)});
} else {
SymbolTable[3].Name.Offset.Offset = sizeof(uint32_t) + Sym.size() + 1;
writeStringTable(Buffer, {Sym, Weak});
}
append(Buffer, SymbolTable); append(Buffer, SymbolTable);
writeStringTable(Buffer, {(Prefix + Sym).str(),
(Prefix + Weak).str()});
// Copied here so we can still use writeStringTable // Copied here so we can still use writeStringTable
char *Buf = Alloc.Allocate<char>(Buffer.size()); char *Buf = Alloc.Allocate<char>(Buffer.size());

View File

@ -946,6 +946,18 @@ bool AArch64ExpandPseudo::expandMI(MachineBasicBlock &MBB,
case AArch64::CMP_SWAP_128: case AArch64::CMP_SWAP_128:
return expandCMP_SWAP_128(MBB, MBBI, NextMBBI); return expandCMP_SWAP_128(MBB, MBBI, NextMBBI);
case AArch64::AESMCrrTied:
case AArch64::AESIMCrrTied: {
MachineInstrBuilder MIB =
BuildMI(MBB, MBBI, MI.getDebugLoc(),
TII->get(Opcode == AArch64::AESMCrrTied ? AArch64::AESMCrr :
AArch64::AESIMCrr))
.add(MI.getOperand(0))
.add(MI.getOperand(1));
transferImpOps(MI, MIB, MIB);
MI.eraseFromParent();
return true;
}
} }
return false; return false;
} }

View File

@ -506,19 +506,23 @@ void AArch64FrameLowering::emitPrologue(MachineFunction &MF,
return; return;
} }
auto CSStackSize = AFI->getCalleeSavedStackSize(); bool IsWin64 =
Subtarget.isCallingConvWin64(MF.getFunction()->getCallingConv());
unsigned FixedObject = IsWin64 ? alignTo(AFI->getVarArgsGPRSize(), 16) : 0;
auto PrologueSaveSize = AFI->getCalleeSavedStackSize() + FixedObject;
// All of the remaining stack allocations are for locals. // All of the remaining stack allocations are for locals.
AFI->setLocalStackSize(NumBytes - CSStackSize); AFI->setLocalStackSize(NumBytes - PrologueSaveSize);
bool CombineSPBump = shouldCombineCSRLocalStackBump(MF, NumBytes); bool CombineSPBump = shouldCombineCSRLocalStackBump(MF, NumBytes);
if (CombineSPBump) { if (CombineSPBump) {
emitFrameOffset(MBB, MBBI, DL, AArch64::SP, AArch64::SP, -NumBytes, TII, emitFrameOffset(MBB, MBBI, DL, AArch64::SP, AArch64::SP, -NumBytes, TII,
MachineInstr::FrameSetup); MachineInstr::FrameSetup);
NumBytes = 0; NumBytes = 0;
} else if (CSStackSize != 0) { } else if (PrologueSaveSize != 0) {
MBBI = convertCalleeSaveRestoreToSPPrePostIncDec(MBB, MBBI, DL, TII, MBBI = convertCalleeSaveRestoreToSPPrePostIncDec(MBB, MBBI, DL, TII,
-CSStackSize); -PrologueSaveSize);
NumBytes -= CSStackSize; NumBytes -= PrologueSaveSize;
} }
assert(NumBytes >= 0 && "Negative stack allocation size!?"); assert(NumBytes >= 0 && "Negative stack allocation size!?");
@ -532,8 +536,9 @@ void AArch64FrameLowering::emitPrologue(MachineFunction &MF,
++MBBI; ++MBBI;
} }
if (HasFP) { if (HasFP) {
// Only set up FP if we actually need to. Frame pointer is fp = sp - 16. // Only set up FP if we actually need to. Frame pointer is fp =
int FPOffset = CSStackSize - 16; // sp - fixedobject - 16.
int FPOffset = AFI->getCalleeSavedStackSize() - 16;
if (CombineSPBump) if (CombineSPBump)
FPOffset += AFI->getLocalStackSize(); FPOffset += AFI->getLocalStackSize();
@ -672,8 +677,8 @@ void AArch64FrameLowering::emitPrologue(MachineFunction &MF,
if (HasFP) { if (HasFP) {
// Define the current CFA rule to use the provided FP. // Define the current CFA rule to use the provided FP.
unsigned Reg = RegInfo->getDwarfRegNum(FramePtr, true); unsigned Reg = RegInfo->getDwarfRegNum(FramePtr, true);
unsigned CFIIndex = MF.addFrameInst( unsigned CFIIndex = MF.addFrameInst(MCCFIInstruction::createDefCfa(
MCCFIInstruction::createDefCfa(nullptr, Reg, 2 * StackGrowth)); nullptr, Reg, 2 * StackGrowth - FixedObject));
BuildMI(MBB, MBBI, DL, TII->get(TargetOpcode::CFI_INSTRUCTION)) BuildMI(MBB, MBBI, DL, TII->get(TargetOpcode::CFI_INSTRUCTION))
.addCFIIndex(CFIIndex) .addCFIIndex(CFIIndex)
.setMIFlags(MachineInstr::FrameSetup); .setMIFlags(MachineInstr::FrameSetup);
@ -759,12 +764,16 @@ void AArch64FrameLowering::emitEpilogue(MachineFunction &MF,
// AArch64TargetLowering::LowerCall figures out ArgumentPopSize and keeps // AArch64TargetLowering::LowerCall figures out ArgumentPopSize and keeps
// it as the 2nd argument of AArch64ISD::TC_RETURN. // it as the 2nd argument of AArch64ISD::TC_RETURN.
auto CSStackSize = AFI->getCalleeSavedStackSize(); bool IsWin64 =
Subtarget.isCallingConvWin64(MF.getFunction()->getCallingConv());
unsigned FixedObject = IsWin64 ? alignTo(AFI->getVarArgsGPRSize(), 16) : 0;
auto PrologueSaveSize = AFI->getCalleeSavedStackSize() + FixedObject;
bool CombineSPBump = shouldCombineCSRLocalStackBump(MF, NumBytes); bool CombineSPBump = shouldCombineCSRLocalStackBump(MF, NumBytes);
if (!CombineSPBump && CSStackSize != 0) if (!CombineSPBump && PrologueSaveSize != 0)
convertCalleeSaveRestoreToSPPrePostIncDec( convertCalleeSaveRestoreToSPPrePostIncDec(
MBB, std::prev(MBB.getFirstTerminator()), DL, TII, CSStackSize); MBB, std::prev(MBB.getFirstTerminator()), DL, TII, PrologueSaveSize);
// Move past the restores of the callee-saved registers. // Move past the restores of the callee-saved registers.
MachineBasicBlock::iterator LastPopI = MBB.getFirstTerminator(); MachineBasicBlock::iterator LastPopI = MBB.getFirstTerminator();
@ -786,7 +795,7 @@ void AArch64FrameLowering::emitEpilogue(MachineFunction &MF,
return; return;
} }
NumBytes -= CSStackSize; NumBytes -= PrologueSaveSize;
assert(NumBytes >= 0 && "Negative stack allocation size!?"); assert(NumBytes >= 0 && "Negative stack allocation size!?");
if (!hasFP(MF)) { if (!hasFP(MF)) {
@ -796,7 +805,7 @@ void AArch64FrameLowering::emitEpilogue(MachineFunction &MF,
if (RedZone && ArgumentPopSize == 0) if (RedZone && ArgumentPopSize == 0)
return; return;
bool NoCalleeSaveRestore = CSStackSize == 0; bool NoCalleeSaveRestore = PrologueSaveSize == 0;
int StackRestoreBytes = RedZone ? 0 : NumBytes; int StackRestoreBytes = RedZone ? 0 : NumBytes;
if (NoCalleeSaveRestore) if (NoCalleeSaveRestore)
StackRestoreBytes += ArgumentPopSize; StackRestoreBytes += ArgumentPopSize;
@ -815,7 +824,8 @@ void AArch64FrameLowering::emitEpilogue(MachineFunction &MF,
// be able to save any instructions. // be able to save any instructions.
if (MFI.hasVarSizedObjects() || AFI->isStackRealigned()) if (MFI.hasVarSizedObjects() || AFI->isStackRealigned())
emitFrameOffset(MBB, LastPopI, DL, AArch64::SP, AArch64::FP, emitFrameOffset(MBB, LastPopI, DL, AArch64::SP, AArch64::FP,
-CSStackSize + 16, TII, MachineInstr::FrameDestroy); -AFI->getCalleeSavedStackSize() + 16, TII,
MachineInstr::FrameDestroy);
else if (NumBytes) else if (NumBytes)
emitFrameOffset(MBB, LastPopI, DL, AArch64::SP, AArch64::SP, NumBytes, TII, emitFrameOffset(MBB, LastPopI, DL, AArch64::SP, AArch64::SP, NumBytes, TII,
MachineInstr::FrameDestroy); MachineInstr::FrameDestroy);
@ -845,7 +855,11 @@ int AArch64FrameLowering::resolveFrameIndexReference(const MachineFunction &MF,
const AArch64RegisterInfo *RegInfo = static_cast<const AArch64RegisterInfo *>( const AArch64RegisterInfo *RegInfo = static_cast<const AArch64RegisterInfo *>(
MF.getSubtarget().getRegisterInfo()); MF.getSubtarget().getRegisterInfo());
const AArch64FunctionInfo *AFI = MF.getInfo<AArch64FunctionInfo>(); const AArch64FunctionInfo *AFI = MF.getInfo<AArch64FunctionInfo>();
int FPOffset = MFI.getObjectOffset(FI) + 16; const AArch64Subtarget &Subtarget = MF.getSubtarget<AArch64Subtarget>();
bool IsWin64 =
Subtarget.isCallingConvWin64(MF.getFunction()->getCallingConv());
unsigned FixedObject = IsWin64 ? alignTo(AFI->getVarArgsGPRSize(), 16) : 0;
int FPOffset = MFI.getObjectOffset(FI) + FixedObject + 16;
int Offset = MFI.getObjectOffset(FI) + MFI.getStackSize(); int Offset = MFI.getObjectOffset(FI) + MFI.getStackSize();
bool isFixed = MFI.isFixedObjectIndex(FI); bool isFixed = MFI.isFixedObjectIndex(FI);
@ -956,12 +970,6 @@ static void computeCalleeSaveRegisterPairs(
"Odd number of callee-saved regs to spill!"); "Odd number of callee-saved regs to spill!");
int Offset = AFI->getCalleeSavedStackSize(); int Offset = AFI->getCalleeSavedStackSize();
unsigned GPRSaveSize = AFI->getVarArgsGPRSize();
const AArch64Subtarget &Subtarget = MF.getSubtarget<AArch64Subtarget>();
bool IsWin64 = Subtarget.isCallingConvWin64(MF.getFunction()->getCallingConv());
if (IsWin64)
Offset -= alignTo(GPRSaveSize, 16);
for (unsigned i = 0; i < Count; ++i) { for (unsigned i = 0; i < Count; ++i) {
RegPairInfo RPI; RegPairInfo RPI;
RPI.Reg1 = CSI[i].getReg(); RPI.Reg1 = CSI[i].getReg();

View File

@ -9586,8 +9586,8 @@ static bool performTBISimplification(SDValue Addr,
SelectionDAG &DAG) { SelectionDAG &DAG) {
APInt DemandedMask = APInt::getLowBitsSet(64, 56); APInt DemandedMask = APInt::getLowBitsSet(64, 56);
KnownBits Known; KnownBits Known;
TargetLowering::TargetLoweringOpt TLO(DAG, DCI.isBeforeLegalize(), TargetLowering::TargetLoweringOpt TLO(DAG, !DCI.isBeforeLegalize(),
DCI.isBeforeLegalizeOps()); !DCI.isBeforeLegalizeOps());
const TargetLowering &TLI = DAG.getTargetLoweringInfo(); const TargetLowering &TLI = DAG.getTargetLoweringInfo();
if (TLI.SimplifyDemandedBits(Addr, DemandedMask, Known, TLO)) { if (TLI.SimplifyDemandedBits(Addr, DemandedMask, Known, TLO)) {
DCI.CommitTargetLoweringOpt(TLO); DCI.CommitTargetLoweringOpt(TLO);

View File

@ -37,6 +37,9 @@ def HasFullFP16 : Predicate<"Subtarget->hasFullFP16()">,
AssemblerPredicate<"FeatureFullFP16", "fullfp16">; AssemblerPredicate<"FeatureFullFP16", "fullfp16">;
def HasSPE : Predicate<"Subtarget->hasSPE()">, def HasSPE : Predicate<"Subtarget->hasSPE()">,
AssemblerPredicate<"FeatureSPE", "spe">; AssemblerPredicate<"FeatureSPE", "spe">;
def HasFuseAES : Predicate<"Subtarget->hasFuseAES()">,
AssemblerPredicate<"FeatureFuseAES",
"fuse-aes">;
def HasSVE : Predicate<"Subtarget->hasSVE()">, def HasSVE : Predicate<"Subtarget->hasSVE()">,
AssemblerPredicate<"FeatureSVE", "sve">; AssemblerPredicate<"FeatureSVE", "sve">;
@ -5304,6 +5307,31 @@ def AESDrr : AESTiedInst<0b0101, "aesd", int_aarch64_crypto_aesd>;
def AESMCrr : AESInst< 0b0110, "aesmc", int_aarch64_crypto_aesmc>; def AESMCrr : AESInst< 0b0110, "aesmc", int_aarch64_crypto_aesmc>;
def AESIMCrr : AESInst< 0b0111, "aesimc", int_aarch64_crypto_aesimc>; def AESIMCrr : AESInst< 0b0111, "aesimc", int_aarch64_crypto_aesimc>;
// Pseudo instructions for AESMCrr/AESIMCrr with a register constraint required
// for AES fusion on some CPUs.
let hasSideEffects = 0, mayStore = 0, mayLoad = 0 in {
def AESMCrrTied: Pseudo<(outs V128:$Rd), (ins V128:$Rn), [], "$Rn = $Rd">,
Sched<[WriteV]>;
def AESIMCrrTied: Pseudo<(outs V128:$Rd), (ins V128:$Rn), [], "$Rn = $Rd">,
Sched<[WriteV]>;
}
// Only use constrained versions of AES(I)MC instructions if they are paired with
// AESE/AESD.
def : Pat<(v16i8 (int_aarch64_crypto_aesmc
(v16i8 (int_aarch64_crypto_aese (v16i8 V128:$src1),
(v16i8 V128:$src2))))),
(v16i8 (AESMCrrTied (v16i8 (AESErr (v16i8 V128:$src1),
(v16i8 V128:$src2)))))>,
Requires<[HasFuseAES]>;
def : Pat<(v16i8 (int_aarch64_crypto_aesimc
(v16i8 (int_aarch64_crypto_aesd (v16i8 V128:$src1),
(v16i8 V128:$src2))))),
(v16i8 (AESIMCrrTied (v16i8 (AESDrr (v16i8 V128:$src1),
(v16i8 V128:$src2)))))>,
Requires<[HasFuseAES]>;
def SHA1Crrr : SHATiedInstQSV<0b000, "sha1c", int_aarch64_crypto_sha1c>; def SHA1Crrr : SHATiedInstQSV<0b000, "sha1c", int_aarch64_crypto_sha1c>;
def SHA1Prrr : SHATiedInstQSV<0b001, "sha1p", int_aarch64_crypto_sha1p>; def SHA1Prrr : SHATiedInstQSV<0b001, "sha1p", int_aarch64_crypto_sha1p>;
def SHA1Mrrr : SHATiedInstQSV<0b010, "sha1m", int_aarch64_crypto_sha1m>; def SHA1Mrrr : SHATiedInstQSV<0b010, "sha1m", int_aarch64_crypto_sha1m>;

View File

@ -118,11 +118,13 @@ static bool shouldScheduleAdjacent(const TargetInstrInfo &TII,
// Fuse AES crypto operations. // Fuse AES crypto operations.
switch(SecondOpcode) { switch(SecondOpcode) {
// AES encode. // AES encode.
case AArch64::AESMCrr : case AArch64::AESMCrr:
case AArch64::AESMCrrTied:
return FirstOpcode == AArch64::AESErr || return FirstOpcode == AArch64::AESErr ||
FirstOpcode == AArch64::INSTRUCTION_LIST_END; FirstOpcode == AArch64::INSTRUCTION_LIST_END;
// AES decode. // AES decode.
case AArch64::AESIMCrr: case AArch64::AESIMCrr:
case AArch64::AESIMCrrTied:
return FirstOpcode == AArch64::AESDrr || return FirstOpcode == AArch64::AESDrr ||
FirstOpcode == AArch64::INSTRUCTION_LIST_END; FirstOpcode == AArch64::INSTRUCTION_LIST_END;
} }

View File

@ -3984,6 +3984,13 @@ bool MatchingStackOffset(SDValue Arg, unsigned Offset, ISD::ArgFlagsTy Flags,
if (Offset != MFI.getObjectOffset(FI)) if (Offset != MFI.getObjectOffset(FI))
return false; return false;
// If this is not byval, check that the argument stack object is immutable.
// inalloca and argument copy elision can create mutable argument stack
// objects. Byval objects can be mutated, but a byval call intends to pass the
// mutated memory.
if (!Flags.isByVal() && !MFI.isImmutableObjectIndex(FI))
return false;
if (VA.getLocVT().getSizeInBits() > Arg.getValueSizeInBits()) { if (VA.getLocVT().getSizeInBits() > Arg.getValueSizeInBits()) {
// If the argument location is wider than the argument type, check that any // If the argument location is wider than the argument type, check that any
// extension flags match. // extension flags match.
@ -30605,8 +30612,8 @@ static SDValue combineSelect(SDNode *N, SelectionDAG &DAG,
assert(BitWidth >= 8 && BitWidth <= 64 && "Invalid mask size"); assert(BitWidth >= 8 && BitWidth <= 64 && "Invalid mask size");
APInt DemandedMask(APInt::getSignMask(BitWidth)); APInt DemandedMask(APInt::getSignMask(BitWidth));
KnownBits Known; KnownBits Known;
TargetLowering::TargetLoweringOpt TLO(DAG, DCI.isBeforeLegalize(), TargetLowering::TargetLoweringOpt TLO(DAG, !DCI.isBeforeLegalize(),
DCI.isBeforeLegalizeOps()); !DCI.isBeforeLegalizeOps());
if (TLI.ShrinkDemandedConstant(Cond, DemandedMask, TLO) || if (TLI.ShrinkDemandedConstant(Cond, DemandedMask, TLO) ||
TLI.SimplifyDemandedBits(Cond, DemandedMask, Known, TLO)) { TLI.SimplifyDemandedBits(Cond, DemandedMask, Known, TLO)) {
// If we changed the computation somewhere in the DAG, this change will // If we changed the computation somewhere in the DAG, this change will

View File

@ -3697,8 +3697,7 @@ let SchedRW = [WriteNop] in {
// Pause. This "instruction" is encoded as "rep; nop", so even though it // Pause. This "instruction" is encoded as "rep; nop", so even though it
// was introduced with SSE2, it's backward compatible. // was introduced with SSE2, it's backward compatible.
def PAUSE : I<0x90, RawFrm, (outs), (ins), def PAUSE : I<0x90, RawFrm, (outs), (ins),
"pause", [(int_x86_sse2_pause)], IIC_SSE_PAUSE>, "pause", [(int_x86_sse2_pause)], IIC_SSE_PAUSE>, OBXS;
OBXS, Requires<[HasSSE2]>;
} }
let SchedRW = [WriteFence] in { let SchedRW = [WriteFence] in {

View File

@ -356,7 +356,7 @@ doPromotion(Function *F, SmallPtrSetImpl<Argument *> &ArgsToPromote,
// Just add all the struct element types. // Just add all the struct element types.
Type *AgTy = cast<PointerType>(I->getType())->getElementType(); Type *AgTy = cast<PointerType>(I->getType())->getElementType();
Value *TheAlloca = new AllocaInst(AgTy, DL.getAllocaAddrSpace(), nullptr, Value *TheAlloca = new AllocaInst(AgTy, DL.getAllocaAddrSpace(), nullptr,
"", InsertPt); I->getParamAlignment(), "", InsertPt);
StructType *STy = cast<StructType>(AgTy); StructType *STy = cast<StructType>(AgTy);
Value *Idxs[2] = {ConstantInt::get(Type::getInt32Ty(F->getContext()), 0), Value *Idxs[2] = {ConstantInt::get(Type::getInt32Ty(F->getContext()), 0),
nullptr}; nullptr};

View File

@ -756,7 +756,8 @@ struct FunctionStackPoisoner : public InstVisitor<FunctionStackPoisoner> {
bool runOnFunction() { bool runOnFunction() {
if (!ClStack) return false; if (!ClStack) return false;
if (ClRedzoneByvalArgs) copyArgsPassedByValToAllocas(); if (ClRedzoneByvalArgs && Mapping.Offset != kDynamicShadowSentinel)
copyArgsPassedByValToAllocas();
// Collect alloca, ret, lifetime instructions etc. // Collect alloca, ret, lifetime instructions etc.
for (BasicBlock *BB : depth_first(&F.getEntryBlock())) visit(*BB); for (BasicBlock *BB : depth_first(&F.getEntryBlock())) visit(*BB);

View File

@ -1790,7 +1790,8 @@ static bool runIPSCCP(Module &M, const DataLayout &DL,
// variables that do not have their 'addresses taken'. If they don't have // variables that do not have their 'addresses taken'. If they don't have
// their addresses taken, we can propagate constants through them. // their addresses taken, we can propagate constants through them.
for (GlobalVariable &G : M.globals()) for (GlobalVariable &G : M.globals())
if (!G.isConstant() && G.hasLocalLinkage() && !AddressIsTaken(&G)) if (!G.isConstant() && G.hasLocalLinkage() &&
G.hasDefinitiveInitializer() && !AddressIsTaken(&G))
Solver.TrackValueOfGlobalVariable(&G); Solver.TrackValueOfGlobalVariable(&G);
// Solve for constants. // Solve for constants.

View File

@ -132,7 +132,8 @@ std::unique_ptr<Module> llvm::CloneModule(
SmallVector<std::pair<unsigned, MDNode *>, 1> MDs; SmallVector<std::pair<unsigned, MDNode *>, 1> MDs;
I->getAllMetadata(MDs); I->getAllMetadata(MDs);
for (auto MD : MDs) for (auto MD : MDs)
GV->addMetadata(MD.first, *MapMetadata(MD.second, VMap)); GV->addMetadata(MD.first,
*MapMetadata(MD.second, VMap, RF_MoveDistinctMDs));
copyComdat(GV, &*I); copyComdat(GV, &*I);
} }

View File

@ -317,6 +317,7 @@ class CoroutineBodyStmt final
unsigned NumParams; unsigned NumParams;
friend class ASTStmtReader; friend class ASTStmtReader;
friend class ASTReader;
friend TrailingObjects; friend TrailingObjects;
Stmt **getStoredStmts() { return getTrailingObjects<Stmt *>(); } Stmt **getStoredStmts() { return getTrailingObjects<Stmt *>(); }
@ -347,6 +348,8 @@ class CoroutineBodyStmt final
public: public:
static CoroutineBodyStmt *Create(const ASTContext &C, CtorArgs const &Args); static CoroutineBodyStmt *Create(const ASTContext &C, CtorArgs const &Args);
static CoroutineBodyStmt *Create(const ASTContext &C, EmptyShell,
unsigned NumParams);
bool hasDependentPromiseType() const { bool hasDependentPromiseType() const {
return getPromiseDecl()->getType()->isDependentType(); return getPromiseDecl()->getType()->isDependentType();
@ -444,6 +447,8 @@ class CoreturnStmt : public Stmt {
SubStmts[SubStmt::PromiseCall] = PromiseCall; SubStmts[SubStmt::PromiseCall] = PromiseCall;
} }
CoreturnStmt(EmptyShell) : CoreturnStmt({}, {}, {}) {}
SourceLocation getKeywordLoc() const { return CoreturnLoc; } SourceLocation getKeywordLoc() const { return CoreturnLoc; }
/// \brief Retrieve the operand of the 'co_return' statement. Will be nullptr /// \brief Retrieve the operand of the 'co_return' statement. Will be nullptr

View File

@ -338,8 +338,8 @@ TARGET_BUILTIN(__builtin_ia32_lfence, "v", "", "sse2")
TARGET_HEADER_BUILTIN(_mm_lfence, "v", "h", "emmintrin.h", ALL_LANGUAGES, "sse2") TARGET_HEADER_BUILTIN(_mm_lfence, "v", "h", "emmintrin.h", ALL_LANGUAGES, "sse2")
TARGET_BUILTIN(__builtin_ia32_mfence, "v", "", "sse2") TARGET_BUILTIN(__builtin_ia32_mfence, "v", "", "sse2")
TARGET_HEADER_BUILTIN(_mm_mfence, "v", "h", "emmintrin.h", ALL_LANGUAGES, "sse2") TARGET_HEADER_BUILTIN(_mm_mfence, "v", "h", "emmintrin.h", ALL_LANGUAGES, "sse2")
TARGET_BUILTIN(__builtin_ia32_pause, "v", "", "sse2") TARGET_BUILTIN(__builtin_ia32_pause, "v", "", "")
TARGET_HEADER_BUILTIN(_mm_pause, "v", "h", "emmintrin.h", ALL_LANGUAGES, "sse2") TARGET_HEADER_BUILTIN(_mm_pause, "v", "h", "emmintrin.h", ALL_LANGUAGES, "")
TARGET_BUILTIN(__builtin_ia32_pmuludq128, "V2LLiV4iV4i", "", "sse2") TARGET_BUILTIN(__builtin_ia32_pmuludq128, "V2LLiV4iV4i", "", "sse2")
TARGET_BUILTIN(__builtin_ia32_psraw128, "V8sV8sV8s", "", "sse2") TARGET_BUILTIN(__builtin_ia32_psraw128, "V8sV8sV8s", "", "sse2")
TARGET_BUILTIN(__builtin_ia32_psrad128, "V4iV4iV4i", "", "sse2") TARGET_BUILTIN(__builtin_ia32_psrad128, "V4iV4iV4i", "", "sse2")

View File

@ -2019,10 +2019,6 @@ def mdspr2 : Flag<["-"], "mdspr2">, Group<m_Group>;
def mno_dspr2 : Flag<["-"], "mno-dspr2">, Group<m_Group>; def mno_dspr2 : Flag<["-"], "mno-dspr2">, Group<m_Group>;
def msingle_float : Flag<["-"], "msingle-float">, Group<m_Group>; def msingle_float : Flag<["-"], "msingle-float">, Group<m_Group>;
def mdouble_float : Flag<["-"], "mdouble-float">, Group<m_Group>; def mdouble_float : Flag<["-"], "mdouble-float">, Group<m_Group>;
def mmadd4 : Flag<["-"], "mmadd4">, Group<m_Group>,
HelpText<"Enable the generation of 4-operand madd.s, madd.d and related instructions.">;
def mno_madd4 : Flag<["-"], "mno-madd4">, Group<m_Group>,
HelpText<"Disable the generation of 4-operand madd.s, madd.d and related instructions.">;
def mmsa : Flag<["-"], "mmsa">, Group<m_Group>, def mmsa : Flag<["-"], "mmsa">, Group<m_Group>,
HelpText<"Enable MSA ASE (MIPS only)">; HelpText<"Enable MSA ASE (MIPS only)">;
def mno_msa : Flag<["-"], "mno-msa">, Group<m_Group>, def mno_msa : Flag<["-"], "mno-msa">, Group<m_Group>,

View File

@ -1545,9 +1545,14 @@ namespace clang {
// ARC // ARC
EXPR_OBJC_BRIDGED_CAST, // ObjCBridgedCastExpr EXPR_OBJC_BRIDGED_CAST, // ObjCBridgedCastExpr
STMT_MS_DEPENDENT_EXISTS, // MSDependentExistsStmt STMT_MS_DEPENDENT_EXISTS, // MSDependentExistsStmt
EXPR_LAMBDA // LambdaExpr EXPR_LAMBDA, // LambdaExpr
STMT_COROUTINE_BODY,
STMT_CORETURN,
EXPR_COAWAIT,
EXPR_COYIELD,
EXPR_DEPENDENT_COAWAIT,
}; };
/// \brief The kinds of designators that can occur in a /// \brief The kinds of designators that can occur in a

View File

@ -378,8 +378,12 @@ void ODRHash::AddCXXRecordDecl(const CXXRecordDecl *Record) {
assert(Record && Record->hasDefinition() && assert(Record && Record->hasDefinition() &&
"Expected non-null record to be a definition."); "Expected non-null record to be a definition.");
if (isa<ClassTemplateSpecializationDecl>(Record)) { const DeclContext *DC = Record;
return; while (DC) {
if (isa<ClassTemplateSpecializationDecl>(DC)) {
return;
}
DC = DC->getParent();
} }
AddDecl(Record); AddDecl(Record);

View File

@ -96,6 +96,20 @@ CoroutineBodyStmt *CoroutineBodyStmt::Create(
return new (Mem) CoroutineBodyStmt(Args); return new (Mem) CoroutineBodyStmt(Args);
} }
CoroutineBodyStmt *CoroutineBodyStmt::Create(const ASTContext &C, EmptyShell,
unsigned NumParams) {
std::size_t Size = totalSizeToAlloc<Stmt *>(
CoroutineBodyStmt::FirstParamMove + NumParams);
void *Mem = C.Allocate(Size, alignof(CoroutineBodyStmt));
auto *Result = new (Mem) CoroutineBodyStmt(CtorArgs());
Result->NumParams = NumParams;
auto *ParamBegin = Result->getStoredStmts() + SubStmt::FirstParamMove;
std::uninitialized_fill(ParamBegin, ParamBegin + NumParams,
static_cast<Stmt *>(nullptr));
return Result;
}
CoroutineBodyStmt::CoroutineBodyStmt(CoroutineBodyStmt::CtorArgs const &Args) CoroutineBodyStmt::CoroutineBodyStmt(CoroutineBodyStmt::CtorArgs const &Args)
: Stmt(CoroutineBodyStmtClass), NumParams(Args.ParamMoves.size()) { : Stmt(CoroutineBodyStmtClass), NumParams(Args.ParamMoves.size()) {
Stmt **SubStmts = getStoredStmts(); Stmt **SubStmts = getStoredStmts();

View File

@ -8050,7 +8050,6 @@ class MipsTargetInfo : public TargetInfo {
NoDSP, DSP1, DSP2 NoDSP, DSP1, DSP2
} DspRev; } DspRev;
bool HasMSA; bool HasMSA;
bool DisableMadd4;
protected: protected:
bool HasFP64; bool HasFP64;
@ -8061,7 +8060,7 @@ class MipsTargetInfo : public TargetInfo {
: TargetInfo(Triple), IsMips16(false), IsMicromips(false), : TargetInfo(Triple), IsMips16(false), IsMicromips(false),
IsNan2008(false), IsSingleFloat(false), IsNoABICalls(false), IsNan2008(false), IsSingleFloat(false), IsNoABICalls(false),
CanUseBSDABICalls(false), FloatABI(HardFloat), DspRev(NoDSP), CanUseBSDABICalls(false), FloatABI(HardFloat), DspRev(NoDSP),
HasMSA(false), DisableMadd4(false), HasFP64(false) { HasMSA(false), HasFP64(false) {
TheCXXABI.set(TargetCXXABI::GenericMIPS); TheCXXABI.set(TargetCXXABI::GenericMIPS);
setABI((getTriple().getArch() == llvm::Triple::mips || setABI((getTriple().getArch() == llvm::Triple::mips ||
@ -8307,9 +8306,6 @@ class MipsTargetInfo : public TargetInfo {
if (HasMSA) if (HasMSA)
Builder.defineMacro("__mips_msa", Twine(1)); Builder.defineMacro("__mips_msa", Twine(1));
if (DisableMadd4)
Builder.defineMacro("__mips_no_madd4", Twine(1));
Builder.defineMacro("_MIPS_SZPTR", Twine(getPointerWidth(0))); Builder.defineMacro("_MIPS_SZPTR", Twine(getPointerWidth(0)));
Builder.defineMacro("_MIPS_SZINT", Twine(getIntWidth())); Builder.defineMacro("_MIPS_SZINT", Twine(getIntWidth()));
Builder.defineMacro("_MIPS_SZLONG", Twine(getLongWidth())); Builder.defineMacro("_MIPS_SZLONG", Twine(getLongWidth()));
@ -8472,8 +8468,6 @@ class MipsTargetInfo : public TargetInfo {
DspRev = std::max(DspRev, DSP2); DspRev = std::max(DspRev, DSP2);
else if (Feature == "+msa") else if (Feature == "+msa")
HasMSA = true; HasMSA = true;
else if (Feature == "+nomadd4")
DisableMadd4 = true;
else if (Feature == "+fp64") else if (Feature == "+fp64")
HasFP64 = true; HasFP64 = true;
else if (Feature == "-fp64") else if (Feature == "-fp64")

View File

@ -297,8 +297,6 @@ void mips::getMIPSTargetFeatures(const Driver &D, const llvm::Triple &Triple,
AddTargetFeature(Args, Features, options::OPT_mno_odd_spreg, AddTargetFeature(Args, Features, options::OPT_mno_odd_spreg,
options::OPT_modd_spreg, "nooddspreg"); options::OPT_modd_spreg, "nooddspreg");
AddTargetFeature(Args, Features, options::OPT_mno_madd4, options::OPT_mmadd4,
"nomadd4");
AddTargetFeature(Args, Features, options::OPT_mlong_calls, AddTargetFeature(Args, Features, options::OPT_mlong_calls,
options::OPT_mno_long_calls, "long-calls"); options::OPT_mno_long_calls, "long-calls");
AddTargetFeature(Args, Features, options::OPT_mmt, options::OPT_mno_mt,"mt"); AddTargetFeature(Args, Features, options::OPT_mmt, options::OPT_mno_mt,"mt");

View File

@ -133,6 +133,8 @@ void openbsd::Linker::ConstructJob(Compilation &C, const JobAction &JA,
} }
} }
if (Args.hasArg(options::OPT_pie))
CmdArgs.push_back("-pie");
if (Args.hasArg(options::OPT_nopie)) if (Args.hasArg(options::OPT_nopie))
CmdArgs.push_back("-nopie"); CmdArgs.push_back("-nopie");

View File

@ -832,7 +832,8 @@ _mm256_xor_si256(__m256i __a, __m256i __b)
static __inline__ __m256i __DEFAULT_FN_ATTRS static __inline__ __m256i __DEFAULT_FN_ATTRS
_mm256_stream_load_si256(__m256i const *__V) _mm256_stream_load_si256(__m256i const *__V)
{ {
return (__m256i)__builtin_nontemporal_load((const __v4di *)__V); typedef __v4di __v4di_aligned __attribute__((aligned(32)));
return (__m256i)__builtin_nontemporal_load((const __v4di_aligned *)__V);
} }
static __inline__ __m128 __DEFAULT_FN_ATTRS static __inline__ __m128 __DEFAULT_FN_ATTRS

View File

@ -4289,7 +4289,7 @@ static __inline__ __m512i __DEFAULT_FN_ATTRS
_mm512_maskz_cvtps_epu32 ( __mmask16 __U, __m512 __A) _mm512_maskz_cvtps_epu32 ( __mmask16 __U, __m512 __A)
{ {
return (__m512i) __builtin_ia32_cvtps2udq512_mask ((__v16sf) __A, return (__m512i) __builtin_ia32_cvtps2udq512_mask ((__v16sf) __A,
(__v16si) (__v16si)
_mm512_setzero_si512 (), _mm512_setzero_si512 (),
(__mmask16) __U , (__mmask16) __U ,
_MM_FROUND_CUR_DIRECTION); _MM_FROUND_CUR_DIRECTION);
@ -9035,25 +9035,29 @@ _mm512_kxor (__mmask16 __A, __mmask16 __B)
static __inline__ void __DEFAULT_FN_ATTRS static __inline__ void __DEFAULT_FN_ATTRS
_mm512_stream_si512 (__m512i * __P, __m512i __A) _mm512_stream_si512 (__m512i * __P, __m512i __A)
{ {
__builtin_nontemporal_store((__v8di)__A, (__v8di*)__P); typedef __v8di __v8di_aligned __attribute__((aligned(64)));
__builtin_nontemporal_store((__v8di_aligned)__A, (__v8di_aligned*)__P);
} }
static __inline__ __m512i __DEFAULT_FN_ATTRS static __inline__ __m512i __DEFAULT_FN_ATTRS
_mm512_stream_load_si512 (void *__P) _mm512_stream_load_si512 (void *__P)
{ {
return (__m512i) __builtin_nontemporal_load((const __v8di *)__P); typedef __v8di __v8di_aligned __attribute__((aligned(64)));
return (__m512i) __builtin_nontemporal_load((const __v8di_aligned *)__P);
} }
static __inline__ void __DEFAULT_FN_ATTRS static __inline__ void __DEFAULT_FN_ATTRS
_mm512_stream_pd (double *__P, __m512d __A) _mm512_stream_pd (double *__P, __m512d __A)
{ {
__builtin_nontemporal_store((__v8df)__A, (__v8df*)__P); typedef __v8df __v8df_aligned __attribute__((aligned(64)));
__builtin_nontemporal_store((__v8df_aligned)__A, (__v8df_aligned*)__P);
} }
static __inline__ void __DEFAULT_FN_ATTRS static __inline__ void __DEFAULT_FN_ATTRS
_mm512_stream_ps (float *__P, __m512 __A) _mm512_stream_ps (float *__P, __m512 __A)
{ {
__builtin_nontemporal_store((__v16sf)__A, (__v16sf*)__P); typedef __v16sf __v16sf_aligned __attribute__((aligned(64)));
__builtin_nontemporal_store((__v16sf_aligned)__A, (__v16sf_aligned*)__P);
} }
static __inline__ __m512d __DEFAULT_FN_ATTRS static __inline__ __m512d __DEFAULT_FN_ATTRS
@ -9217,39 +9221,39 @@ _mm512_maskz_moveldup_ps (__mmask16 __U, __m512 __A)
static __inline__ __m128 __DEFAULT_FN_ATTRS static __inline__ __m128 __DEFAULT_FN_ATTRS
_mm_mask_move_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B) _mm_mask_move_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128 __B)
{ {
__m128 res = __A; __m128 res = __A;
res[0] = (__U & 1) ? __B[0] : __W[0]; res[0] = (__U & 1) ? __B[0] : __W[0];
return res; return res;
} }
static __inline__ __m128 __DEFAULT_FN_ATTRS static __inline__ __m128 __DEFAULT_FN_ATTRS
_mm_maskz_move_ss (__mmask8 __U, __m128 __A, __m128 __B) _mm_maskz_move_ss (__mmask8 __U, __m128 __A, __m128 __B)
{ {
__m128 res = __A; __m128 res = __A;
res[0] = (__U & 1) ? __B[0] : 0; res[0] = (__U & 1) ? __B[0] : 0;
return res; return res;
} }
static __inline__ __m128d __DEFAULT_FN_ATTRS static __inline__ __m128d __DEFAULT_FN_ATTRS
_mm_mask_move_sd (__m128d __W, __mmask8 __U, __m128d __A, __m128d __B) _mm_mask_move_sd (__m128d __W, __mmask8 __U, __m128d __A, __m128d __B)
{ {
__m128d res = __A; __m128d res = __A;
res[0] = (__U & 1) ? __B[0] : __W[0]; res[0] = (__U & 1) ? __B[0] : __W[0];
return res; return res;
} }
static __inline__ __m128d __DEFAULT_FN_ATTRS static __inline__ __m128d __DEFAULT_FN_ATTRS
_mm_maskz_move_sd (__mmask8 __U, __m128d __A, __m128d __B) _mm_maskz_move_sd (__mmask8 __U, __m128d __A, __m128d __B)
{ {
__m128d res = __A; __m128d res = __A;
res[0] = (__U & 1) ? __B[0] : 0; res[0] = (__U & 1) ? __B[0] : 0;
return res; return res;
} }
static __inline__ void __DEFAULT_FN_ATTRS static __inline__ void __DEFAULT_FN_ATTRS
_mm_mask_store_ss (float * __W, __mmask8 __U, __m128 __A) _mm_mask_store_ss (float * __W, __mmask8 __U, __m128 __A)
{ {
__builtin_ia32_storess128_mask ((__v16sf *)__W, __builtin_ia32_storess128_mask ((__v16sf *)__W,
(__v16sf) _mm512_castps128_ps512(__A), (__v16sf) _mm512_castps128_ps512(__A),
(__mmask16) __U & (__mmask16)1); (__mmask16) __U & (__mmask16)1);
} }
@ -9257,7 +9261,7 @@ _mm_mask_store_ss (float * __W, __mmask8 __U, __m128 __A)
static __inline__ void __DEFAULT_FN_ATTRS static __inline__ void __DEFAULT_FN_ATTRS
_mm_mask_store_sd (double * __W, __mmask8 __U, __m128d __A) _mm_mask_store_sd (double * __W, __mmask8 __U, __m128d __A)
{ {
__builtin_ia32_storesd128_mask ((__v8df *)__W, __builtin_ia32_storesd128_mask ((__v8df *)__W,
(__v8df) _mm512_castpd128_pd512(__A), (__v8df) _mm512_castpd128_pd512(__A),
(__mmask8) __U & 1); (__mmask8) __U & 1);
} }
@ -9606,7 +9610,7 @@ _mm_mask_cvtsd_ss (__m128 __W, __mmask8 __U, __m128 __A, __m128d __B)
{ {
return __builtin_ia32_cvtsd2ss_round_mask ((__v4sf)(__A), return __builtin_ia32_cvtsd2ss_round_mask ((__v4sf)(__A),
(__v2df)(__B), (__v2df)(__B),
(__v4sf)(__W), (__v4sf)(__W),
(__mmask8)(__U), _MM_FROUND_CUR_DIRECTION); (__mmask8)(__U), _MM_FROUND_CUR_DIRECTION);
} }
@ -9615,7 +9619,7 @@ _mm_maskz_cvtsd_ss (__mmask8 __U, __m128 __A, __m128d __B)
{ {
return __builtin_ia32_cvtsd2ss_round_mask ((__v4sf)(__A), return __builtin_ia32_cvtsd2ss_round_mask ((__v4sf)(__A),
(__v2df)(__B), (__v2df)(__B),
(__v4sf)_mm_setzero_ps(), (__v4sf)_mm_setzero_ps(),
(__mmask8)(__U), _MM_FROUND_CUR_DIRECTION); (__mmask8)(__U), _MM_FROUND_CUR_DIRECTION);
} }
@ -9680,7 +9684,7 @@ _mm_mask_cvtss_sd (__m128d __W, __mmask8 __U, __m128d __A, __m128 __B)
return __builtin_ia32_cvtss2sd_round_mask((__v2df)(__A), return __builtin_ia32_cvtss2sd_round_mask((__v2df)(__A),
(__v4sf)(__B), (__v4sf)(__B),
(__v2df)(__W), (__v2df)(__W),
(__mmask8)(__U), _MM_FROUND_CUR_DIRECTION); (__mmask8)(__U), _MM_FROUND_CUR_DIRECTION);
} }
static __inline__ __m128d __DEFAULT_FN_ATTRS static __inline__ __m128d __DEFAULT_FN_ATTRS
@ -9688,8 +9692,8 @@ _mm_maskz_cvtss_sd (__mmask8 __U, __m128d __A, __m128 __B)
{ {
return __builtin_ia32_cvtss2sd_round_mask((__v2df)(__A), return __builtin_ia32_cvtss2sd_round_mask((__v2df)(__A),
(__v4sf)(__B), (__v4sf)(__B),
(__v2df)_mm_setzero_pd(), (__v2df)_mm_setzero_pd(),
(__mmask8)(__U), _MM_FROUND_CUR_DIRECTION); (__mmask8)(__U), _MM_FROUND_CUR_DIRECTION);
} }
static __inline__ __m128d __DEFAULT_FN_ATTRS static __inline__ __m128d __DEFAULT_FN_ATTRS
@ -9935,7 +9939,7 @@ static __inline__ double __DEFAULT_FN_ATTRS _mm512_reduce_mul_pd(__m512d __W) {
} }
// Vec512 - Vector with size 512. // Vec512 - Vector with size 512.
// Vec512Neutral - All vector elements set to the identity element. // Vec512Neutral - All vector elements set to the identity element.
// Identity element: {+,0},{*,1},{&,0xFFFFFFFFFFFFFFFF},{|,0} // Identity element: {+,0},{*,1},{&,0xFFFFFFFFFFFFFFFF},{|,0}
// Operator - Can be one of following: +,*,&,| // Operator - Can be one of following: +,*,&,|
// Mask - Intrinsic Mask // Mask - Intrinsic Mask
@ -9965,19 +9969,19 @@ _mm512_mask_reduce_mul_epi64(__mmask8 __M, __m512i __W) {
static __inline__ long long __DEFAULT_FN_ATTRS static __inline__ long long __DEFAULT_FN_ATTRS
_mm512_mask_reduce_and_epi64(__mmask8 __M, __m512i __W) { _mm512_mask_reduce_and_epi64(__mmask8 __M, __m512i __W) {
_mm512_mask_reduce_operator_64bit(__W, _mm512_set1_epi64(0xFFFFFFFFFFFFFFFF), _mm512_mask_reduce_operator_64bit(__W, _mm512_set1_epi64(0xFFFFFFFFFFFFFFFF),
&, __M, i, i, q); &, __M, i, i, q);
} }
static __inline__ long long __DEFAULT_FN_ATTRS static __inline__ long long __DEFAULT_FN_ATTRS
_mm512_mask_reduce_or_epi64(__mmask8 __M, __m512i __W) { _mm512_mask_reduce_or_epi64(__mmask8 __M, __m512i __W) {
_mm512_mask_reduce_operator_64bit(__W, _mm512_set1_epi64(0), |, __M, _mm512_mask_reduce_operator_64bit(__W, _mm512_set1_epi64(0), |, __M,
i, i, q); i, i, q);
} }
static __inline__ double __DEFAULT_FN_ATTRS static __inline__ double __DEFAULT_FN_ATTRS
_mm512_mask_reduce_add_pd(__mmask8 __M, __m512d __W) { _mm512_mask_reduce_add_pd(__mmask8 __M, __m512d __W) {
_mm512_mask_reduce_operator_64bit(__W, _mm512_set1_pd(0), +, __M, _mm512_mask_reduce_operator_64bit(__W, _mm512_set1_pd(0), +, __M,
f, d, pd); f, d, pd);
} }
@ -10039,17 +10043,17 @@ _mm512_reduce_add_epi32(__m512i __W) {
_mm512_reduce_operator_32bit(__W, +, i, i); _mm512_reduce_operator_32bit(__W, +, i, i);
} }
static __inline__ int __DEFAULT_FN_ATTRS static __inline__ int __DEFAULT_FN_ATTRS
_mm512_reduce_mul_epi32(__m512i __W) { _mm512_reduce_mul_epi32(__m512i __W) {
_mm512_reduce_operator_32bit(__W, *, i, i); _mm512_reduce_operator_32bit(__W, *, i, i);
} }
static __inline__ int __DEFAULT_FN_ATTRS static __inline__ int __DEFAULT_FN_ATTRS
_mm512_reduce_and_epi32(__m512i __W) { _mm512_reduce_and_epi32(__m512i __W) {
_mm512_reduce_operator_32bit(__W, &, i, i); _mm512_reduce_operator_32bit(__W, &, i, i);
} }
static __inline__ int __DEFAULT_FN_ATTRS static __inline__ int __DEFAULT_FN_ATTRS
_mm512_reduce_or_epi32(__m512i __W) { _mm512_reduce_or_epi32(__m512i __W) {
_mm512_reduce_operator_32bit(__W, |, i, i); _mm512_reduce_operator_32bit(__W, |, i, i);
} }
@ -10065,7 +10069,7 @@ _mm512_reduce_mul_ps(__m512 __W) {
} }
// Vec512 - Vector with size 512. // Vec512 - Vector with size 512.
// Vec512Neutral - All vector elements set to the identity element. // Vec512Neutral - All vector elements set to the identity element.
// Identity element: {+,0},{*,1},{&,0xFFFFFFFF},{|,0} // Identity element: {+,0},{*,1},{&,0xFFFFFFFF},{|,0}
// Operator - Can be one of following: +,*,&,| // Operator - Can be one of following: +,*,&,|
// Mask - Intrinsic Mask // Mask - Intrinsic Mask
@ -10095,7 +10099,7 @@ _mm512_mask_reduce_mul_epi32( __mmask16 __M, __m512i __W) {
static __inline__ int __DEFAULT_FN_ATTRS static __inline__ int __DEFAULT_FN_ATTRS
_mm512_mask_reduce_and_epi32( __mmask16 __M, __m512i __W) { _mm512_mask_reduce_and_epi32( __mmask16 __M, __m512i __W) {
_mm512_mask_reduce_operator_32bit(__W, _mm512_set1_epi32(0xFFFFFFFF), &, __M, _mm512_mask_reduce_operator_32bit(__W, _mm512_set1_epi32(0xFFFFFFFF), &, __M,
i, i, d); i, i, d);
} }
@ -10158,7 +10162,7 @@ _mm512_mask_reduce_mul_ps(__mmask16 __M, __m512 __W) {
return Vec512[0]; \ return Vec512[0]; \
}) })
static __inline__ long long __DEFAULT_FN_ATTRS static __inline__ long long __DEFAULT_FN_ATTRS
_mm512_reduce_max_epi64(__m512i __V) { _mm512_reduce_max_epi64(__m512i __V) {
_mm512_reduce_maxMin_64bit(__V, max_epi64, i, i); _mm512_reduce_maxMin_64bit(__V, max_epi64, i, i);
} }
@ -10168,7 +10172,7 @@ _mm512_reduce_max_epu64(__m512i __V) {
_mm512_reduce_maxMin_64bit(__V, max_epu64, i, i); _mm512_reduce_maxMin_64bit(__V, max_epu64, i, i);
} }
static __inline__ double __DEFAULT_FN_ATTRS static __inline__ double __DEFAULT_FN_ATTRS
_mm512_reduce_max_pd(__m512d __V) { _mm512_reduce_max_pd(__m512d __V) {
_mm512_reduce_maxMin_64bit(__V, max_pd, d, f); _mm512_reduce_maxMin_64bit(__V, max_pd, d, f);
} }
@ -10183,7 +10187,7 @@ _mm512_reduce_min_epu64(__m512i __V) {
_mm512_reduce_maxMin_64bit(__V, min_epu64, i, i); _mm512_reduce_maxMin_64bit(__V, min_epu64, i, i);
} }
static __inline__ double __DEFAULT_FN_ATTRS static __inline__ double __DEFAULT_FN_ATTRS
_mm512_reduce_min_pd(__m512d __V) { _mm512_reduce_min_pd(__m512d __V) {
_mm512_reduce_maxMin_64bit(__V, min_pd, d, f); _mm512_reduce_maxMin_64bit(__V, min_pd, d, f);
} }

View File

@ -3590,7 +3590,8 @@ _mm_maskstore_ps(float *__p, __m128i __m, __m128 __a)
static __inline void __DEFAULT_FN_ATTRS static __inline void __DEFAULT_FN_ATTRS
_mm256_stream_si256(__m256i *__a, __m256i __b) _mm256_stream_si256(__m256i *__a, __m256i __b)
{ {
__builtin_nontemporal_store((__v4di)__b, (__v4di*)__a); typedef __v4di __v4di_aligned __attribute__((aligned(32)));
__builtin_nontemporal_store((__v4di_aligned)__b, (__v4di_aligned*)__a);
} }
/// \brief Moves double-precision values from a 256-bit vector of [4 x double] /// \brief Moves double-precision values from a 256-bit vector of [4 x double]
@ -3609,7 +3610,8 @@ _mm256_stream_si256(__m256i *__a, __m256i __b)
static __inline void __DEFAULT_FN_ATTRS static __inline void __DEFAULT_FN_ATTRS
_mm256_stream_pd(double *__a, __m256d __b) _mm256_stream_pd(double *__a, __m256d __b)
{ {
__builtin_nontemporal_store((__v4df)__b, (__v4df*)__a); typedef __v4df __v4df_aligned __attribute__((aligned(32)));
__builtin_nontemporal_store((__v4df_aligned)__b, (__v4df_aligned*)__a);
} }
/// \brief Moves single-precision floating point values from a 256-bit vector /// \brief Moves single-precision floating point values from a 256-bit vector
@ -3629,7 +3631,8 @@ _mm256_stream_pd(double *__a, __m256d __b)
static __inline void __DEFAULT_FN_ATTRS static __inline void __DEFAULT_FN_ATTRS
_mm256_stream_ps(float *__p, __m256 __a) _mm256_stream_ps(float *__p, __m256 __a)
{ {
__builtin_nontemporal_store((__v8sf)__a, (__v8sf*)__p); typedef __v8sf __v8sf_aligned __attribute__((aligned(32)));
__builtin_nontemporal_store((__v8sf_aligned)__a, (__v8sf_aligned*)__p);
} }
/* Create vectors */ /* Create vectors */

View File

@ -33,6 +33,15 @@
*/ */
#if (defined(__APPLE__) || (defined(__MINGW32__) || defined(_MSC_VER))) && \ #if (defined(__APPLE__) || (defined(__MINGW32__) || defined(_MSC_VER))) && \
__STDC_HOSTED__ && __has_include_next(<float.h>) __STDC_HOSTED__ && __has_include_next(<float.h>)
/* Prior to Apple's 10.7 SDK, float.h SDK header used to apply an extra level
* of #include_next<float.h> to keep Metrowerks compilers happy. Avoid this
* extra indirection.
*/
#ifdef __APPLE__
#define _FLOAT_H_
#endif
# include_next <float.h> # include_next <float.h>
/* Undefine anything that we'll be redefining below. */ /* Undefine anything that we'll be redefining below. */

View File

@ -542,6 +542,9 @@ void Sema::getUndefinedButUsed(
// __attribute__((weakref)) is basically a definition. // __attribute__((weakref)) is basically a definition.
if (ND->hasAttr<WeakRefAttr>()) continue; if (ND->hasAttr<WeakRefAttr>()) continue;
if (isa<CXXDeductionGuideDecl>(ND))
continue;
if (FunctionDecl *FD = dyn_cast<FunctionDecl>(ND)) { if (FunctionDecl *FD = dyn_cast<FunctionDecl>(ND)) {
if (FD->isDefined()) if (FD->isDefined())
continue; continue;

View File

@ -8288,7 +8288,7 @@ QualType Sema::CheckVectorOperands(ExprResult &LHS, ExprResult &RHS,
// type. Note that this is already done by non-compound assignments in // type. Note that this is already done by non-compound assignments in
// CheckAssignmentConstraints. If it's a scalar type, only bitcast for // CheckAssignmentConstraints. If it's a scalar type, only bitcast for
// <1 x T> -> T. The result is also a vector type. // <1 x T> -> T. The result is also a vector type.
} else if (OtherType->isExtVectorType() || } else if (OtherType->isExtVectorType() || OtherType->isVectorType() ||
(OtherType->isScalarType() && VT->getNumElements() == 1)) { (OtherType->isScalarType() && VT->getNumElements() == 1)) {
ExprResult *RHSExpr = &RHS; ExprResult *RHSExpr = &RHS;
*RHSExpr = ImpCastExprToType(RHSExpr->get(), LHSType, CK_BitCast); *RHSExpr = ImpCastExprToType(RHSExpr->get(), LHSType, CK_BitCast);

View File

@ -367,28 +367,45 @@ void ASTStmtReader::VisitMSAsmStmt(MSAsmStmt *S) {
} }
void ASTStmtReader::VisitCoroutineBodyStmt(CoroutineBodyStmt *S) { void ASTStmtReader::VisitCoroutineBodyStmt(CoroutineBodyStmt *S) {
// FIXME: Implement coroutine serialization. VisitStmt(S);
llvm_unreachable("unimplemented"); assert(Record.peekInt() == S->NumParams);
Record.skipInts(1);
auto *StoredStmts = S->getStoredStmts();
for (unsigned i = 0;
i < CoroutineBodyStmt::SubStmt::FirstParamMove + S->NumParams; ++i)
StoredStmts[i] = Record.readSubStmt();
} }
void ASTStmtReader::VisitCoreturnStmt(CoreturnStmt *S) { void ASTStmtReader::VisitCoreturnStmt(CoreturnStmt *S) {
// FIXME: Implement coroutine serialization. VisitStmt(S);
llvm_unreachable("unimplemented"); S->CoreturnLoc = Record.readSourceLocation();
for (auto &SubStmt: S->SubStmts)
SubStmt = Record.readSubStmt();
S->IsImplicit = Record.readInt() != 0;
} }
void ASTStmtReader::VisitCoawaitExpr(CoawaitExpr *S) { void ASTStmtReader::VisitCoawaitExpr(CoawaitExpr *E) {
// FIXME: Implement coroutine serialization. VisitExpr(E);
llvm_unreachable("unimplemented"); E->KeywordLoc = ReadSourceLocation();
for (auto &SubExpr: E->SubExprs)
SubExpr = Record.readSubStmt();
E->OpaqueValue = cast_or_null<OpaqueValueExpr>(Record.readSubStmt());
E->setIsImplicit(Record.readInt() != 0);
} }
void ASTStmtReader::VisitDependentCoawaitExpr(DependentCoawaitExpr *S) { void ASTStmtReader::VisitCoyieldExpr(CoyieldExpr *E) {
// FIXME: Implement coroutine serialization. VisitExpr(E);
llvm_unreachable("unimplemented"); E->KeywordLoc = ReadSourceLocation();
for (auto &SubExpr: E->SubExprs)
SubExpr = Record.readSubStmt();
E->OpaqueValue = cast_or_null<OpaqueValueExpr>(Record.readSubStmt());
} }
void ASTStmtReader::VisitCoyieldExpr(CoyieldExpr *S) { void ASTStmtReader::VisitDependentCoawaitExpr(DependentCoawaitExpr *E) {
// FIXME: Implement coroutine serialization. VisitExpr(E);
llvm_unreachable("unimplemented"); E->KeywordLoc = ReadSourceLocation();
for (auto &SubExpr: E->SubExprs)
SubExpr = Record.readSubStmt();
} }
void ASTStmtReader::VisitCapturedStmt(CapturedStmt *S) { void ASTStmtReader::VisitCapturedStmt(CapturedStmt *S) {
@ -3947,6 +3964,29 @@ Stmt *ASTReader::ReadStmtFromStream(ModuleFile &F) {
S = LambdaExpr::CreateDeserialized(Context, NumCaptures); S = LambdaExpr::CreateDeserialized(Context, NumCaptures);
break; break;
} }
case STMT_COROUTINE_BODY: {
unsigned NumParams = Record[ASTStmtReader::NumStmtFields];
S = CoroutineBodyStmt::Create(Context, Empty, NumParams);
break;
}
case STMT_CORETURN:
S = new (Context) CoreturnStmt(Empty);
break;
case EXPR_COAWAIT:
S = new (Context) CoawaitExpr(Empty);
break;
case EXPR_COYIELD:
S = new (Context) CoyieldExpr(Empty);
break;
case EXPR_DEPENDENT_COAWAIT:
S = new (Context) DependentCoawaitExpr(Empty);
break;
} }
// We hit a STMT_STOP, so we're done with this expression. // We hit a STMT_STOP, so we're done with this expression.

View File

@ -286,7 +286,7 @@ void ASTStmtWriter::VisitMSAsmStmt(MSAsmStmt *S) {
} }
// Outputs // Outputs
for (unsigned I = 0, N = S->getNumOutputs(); I != N; ++I) { for (unsigned I = 0, N = S->getNumOutputs(); I != N; ++I) {
Record.AddStmt(S->getOutputExpr(I)); Record.AddStmt(S->getOutputExpr(I));
Record.AddString(S->getOutputConstraint(I)); Record.AddString(S->getOutputConstraint(I));
} }
@ -300,29 +300,48 @@ void ASTStmtWriter::VisitMSAsmStmt(MSAsmStmt *S) {
Code = serialization::STMT_MSASM; Code = serialization::STMT_MSASM;
} }
void ASTStmtWriter::VisitCoroutineBodyStmt(CoroutineBodyStmt *S) { void ASTStmtWriter::VisitCoroutineBodyStmt(CoroutineBodyStmt *CoroStmt) {
// FIXME: Implement coroutine serialization. VisitStmt(CoroStmt);
llvm_unreachable("unimplemented"); Record.push_back(CoroStmt->getParamMoves().size());
for (Stmt *S : CoroStmt->children())
Record.AddStmt(S);
Code = serialization::STMT_COROUTINE_BODY;
} }
void ASTStmtWriter::VisitCoreturnStmt(CoreturnStmt *S) { void ASTStmtWriter::VisitCoreturnStmt(CoreturnStmt *S) {
// FIXME: Implement coroutine serialization. VisitStmt(S);
llvm_unreachable("unimplemented"); Record.AddSourceLocation(S->getKeywordLoc());
Record.AddStmt(S->getOperand());
Record.AddStmt(S->getPromiseCall());
Record.push_back(S->isImplicit());
Code = serialization::STMT_CORETURN;
} }
void ASTStmtWriter::VisitCoawaitExpr(CoawaitExpr *S) { void ASTStmtWriter::VisitCoroutineSuspendExpr(CoroutineSuspendExpr *E) {
// FIXME: Implement coroutine serialization. VisitExpr(E);
llvm_unreachable("unimplemented"); Record.AddSourceLocation(E->getKeywordLoc());
for (Stmt *S : E->children())
Record.AddStmt(S);
Record.AddStmt(E->getOpaqueValue());
} }
void ASTStmtWriter::VisitDependentCoawaitExpr(DependentCoawaitExpr *S) { void ASTStmtWriter::VisitCoawaitExpr(CoawaitExpr *E) {
// FIXME: Implement coroutine serialization. VisitCoroutineSuspendExpr(E);
llvm_unreachable("unimplemented"); Record.push_back(E->isImplicit());
Code = serialization::EXPR_COAWAIT;
} }
void ASTStmtWriter::VisitCoyieldExpr(CoyieldExpr *S) { void ASTStmtWriter::VisitCoyieldExpr(CoyieldExpr *E) {
// FIXME: Implement coroutine serialization. VisitCoroutineSuspendExpr(E);
llvm_unreachable("unimplemented"); Code = serialization::EXPR_COYIELD;
}
void ASTStmtWriter::VisitDependentCoawaitExpr(DependentCoawaitExpr *E) {
VisitExpr(E);
Record.AddSourceLocation(E->getKeywordLoc());
for (Stmt *S : E->children())
Record.AddStmt(S);
Code = serialization::EXPR_DEPENDENT_COAWAIT;
} }
void ASTStmtWriter::VisitCapturedStmt(CapturedStmt *S) { void ASTStmtWriter::VisitCapturedStmt(CapturedStmt *S) {

View File

@ -8,4 +8,4 @@
#define CLANG_VENDOR "FreeBSD " #define CLANG_VENDOR "FreeBSD "
#define SVN_REVISION "309439" #define SVN_REVISION "310316"

View File

@ -4,5 +4,5 @@
#define LLD_VERSION_STRING "5.0.0" #define LLD_VERSION_STRING "5.0.0"
#define LLD_VERSION_MAJOR 5 #define LLD_VERSION_MAJOR 5
#define LLD_VERSION_MINOR 0 #define LLD_VERSION_MINOR 0
#define LLD_REVISION_STRING "309439" #define LLD_REVISION_STRING "310316"
#define LLD_REPOSITORY_STRING "FreeBSD" #define LLD_REPOSITORY_STRING "FreeBSD"

View File

@ -1,2 +1,2 @@
/* $FreeBSD$ */ /* $FreeBSD$ */
#define LLVM_REVISION "svn-r309439" #define LLVM_REVISION "svn-r310316"