//=- AArch64MachineFunctionInfo.cpp - AArch64 Machine Function Info ---------=// // // Part of the LLVM Project, under the Apache License v2.0 with LLVM Exceptions. // See https://llvm.org/LICENSE.txt for license information. // SPDX-License-Identifier: Apache-2.0 WITH LLVM-exception // //===----------------------------------------------------------------------===// /// /// \file /// This file implements AArch64-specific per-machine-function /// information. /// //===----------------------------------------------------------------------===// #include "AArch64MachineFunctionInfo.h" #include "AArch64InstrInfo.h" #include "AArch64Subtarget.h" #include "llvm/IR/Constants.h" #include "llvm/IR/Metadata.h" #include "llvm/IR/Module.h" #include "llvm/MC/MCAsmInfo.h" using namespace llvm; yaml::AArch64FunctionInfo::AArch64FunctionInfo( const llvm::AArch64FunctionInfo &MFI) : HasRedZone(MFI.hasRedZone()) {} void yaml::AArch64FunctionInfo::mappingImpl(yaml::IO &YamlIO) { MappingTraits::mapping(YamlIO, *this); } void AArch64FunctionInfo::initializeBaseYamlFields( const yaml::AArch64FunctionInfo &YamlMFI) { if (YamlMFI.HasRedZone) HasRedZone = YamlMFI.HasRedZone; } static std::pair GetSignReturnAddress(const Function &F) { if (F.hasFnAttribute("ptrauth-returns")) return {true, false}; // non-leaf // The function should be signed in the following situations: // - sign-return-address=all // - sign-return-address=non-leaf and the functions spills the LR if (!F.hasFnAttribute("sign-return-address")) return {false, false}; StringRef Scope = F.getFnAttribute("sign-return-address").getValueAsString(); if (Scope == "none") return {false, false}; if (Scope == "all") return {true, true}; assert(Scope == "non-leaf"); return {true, false}; } static bool ShouldSignWithBKey(const Function &F, const AArch64Subtarget &STI) { if (F.hasFnAttribute("ptrauth-returns")) return true; if (!F.hasFnAttribute("sign-return-address-key")) { if (STI.getTargetTriple().isOSWindows()) return true; return false; } const StringRef Key = F.getFnAttribute("sign-return-address-key").getValueAsString(); assert(Key == "a_key" || Key == "b_key"); return Key == "b_key"; } AArch64FunctionInfo::AArch64FunctionInfo(const Function &F, const AArch64Subtarget *STI) { // If we already know that the function doesn't have a redzone, set // HasRedZone here. if (F.hasFnAttribute(Attribute::NoRedZone)) HasRedZone = false; std::tie(SignReturnAddress, SignReturnAddressAll) = GetSignReturnAddress(F); SignWithBKey = ShouldSignWithBKey(F, *STI); // TODO: skip functions that have no instrumented allocas for optimization IsMTETagged = F.hasFnAttribute(Attribute::SanitizeMemTag); // BTI/PAuthLR are set on the function attribute. BranchTargetEnforcement = F.hasFnAttribute("branch-target-enforcement"); BranchProtectionPAuthLR = F.hasFnAttribute("branch-protection-pauth-lr"); // The default stack probe size is 4096 if the function has no // stack-probe-size attribute. This is a safe default because it is the // smallest possible guard page size. uint64_t ProbeSize = 4096; if (F.hasFnAttribute("stack-probe-size")) ProbeSize = F.getFnAttributeAsParsedInteger("stack-probe-size"); else if (const auto *PS = mdconst::extract_or_null( F.getParent()->getModuleFlag("stack-probe-size"))) ProbeSize = PS->getZExtValue(); assert(int64_t(ProbeSize) > 0 && "Invalid stack probe size"); if (STI->isTargetWindows()) { if (!F.hasFnAttribute("no-stack-arg-probe")) StackProbeSize = ProbeSize; } else { // Round down to the stack alignment. uint64_t StackAlign = STI->getFrameLowering()->getTransientStackAlign().value(); ProbeSize = std::max(StackAlign, ProbeSize & ~(StackAlign - 1U)); StringRef ProbeKind; if (F.hasFnAttribute("probe-stack")) ProbeKind = F.getFnAttribute("probe-stack").getValueAsString(); else if (const auto *PS = dyn_cast_or_null( F.getParent()->getModuleFlag("probe-stack"))) ProbeKind = PS->getString(); if (ProbeKind.size()) { if (ProbeKind != "inline-asm") report_fatal_error("Unsupported stack probing method"); StackProbeSize = ProbeSize; } } } MachineFunctionInfo *AArch64FunctionInfo::clone( BumpPtrAllocator &Allocator, MachineFunction &DestMF, const DenseMap &Src2DstMBB) const { return DestMF.cloneInfo(*this); } bool AArch64FunctionInfo::shouldSignReturnAddress(bool SpillsLR) const { if (!SignReturnAddress) return false; if (SignReturnAddressAll) return true; return SpillsLR; } static bool isLRSpilled(const MachineFunction &MF) { return llvm::any_of( MF.getFrameInfo().getCalleeSavedInfo(), [](const auto &Info) { return Info.getReg() == AArch64::LR; }); } bool AArch64FunctionInfo::shouldSignReturnAddress( const MachineFunction &MF) const { return shouldSignReturnAddress(isLRSpilled(MF)); } bool AArch64FunctionInfo::needsShadowCallStackPrologueEpilogue( MachineFunction &MF) const { if (!(isLRSpilled(MF) && MF.getFunction().hasFnAttribute(Attribute::ShadowCallStack))) return false; if (!MF.getSubtarget().isXRegisterReserved(18)) report_fatal_error("Must reserve x18 to use shadow call stack"); return true; } bool AArch64FunctionInfo::needsDwarfUnwindInfo( const MachineFunction &MF) const { if (!NeedsDwarfUnwindInfo) NeedsDwarfUnwindInfo = MF.needsFrameMoves() && !MF.getTarget().getMCAsmInfo()->usesWindowsCFI(); return *NeedsDwarfUnwindInfo; } bool AArch64FunctionInfo::needsAsyncDwarfUnwindInfo( const MachineFunction &MF) const { if (!NeedsAsyncDwarfUnwindInfo) { const Function &F = MF.getFunction(); const AArch64FunctionInfo *AFI = MF.getInfo(); // The check got "minsize" is because epilogue unwind info is not emitted // (yet) for homogeneous epilogues, outlined functions, and functions // outlined from. NeedsAsyncDwarfUnwindInfo = needsDwarfUnwindInfo(MF) && ((F.getUWTableKind() == UWTableKind::Async && !F.hasMinSize()) || AFI->hasStreamingModeChanges()); } return *NeedsAsyncDwarfUnwindInfo; }