void setArchInfo() {
StringRef ArchName = getTriple().getArchName();
- ArchISA = llvm::ARMTargetParser::parseArchISA(ArchName);
+ ArchISA = llvm::ARM::parseArchISA(ArchName);
DefaultCPU = getDefaultCPU(ArchName);
- unsigned ArchKind = llvm::ARMTargetParser::parseArch(ArchName);
+ unsigned ArchKind = llvm::ARM::parseArch(ArchName);
if (ArchKind == llvm::ARM::AK_INVALID)
// set arch of the CPU, either provided explicitly or hardcoded default
- ArchKind = llvm::ARMTargetParser::parseCPUArch(CPU);
+ ArchKind = llvm::ARM::parseCPUArch(CPU);
setArchInfo(ArchKind);
}
// cache TargetParser info
ArchKind = Kind;
- SubArch = llvm::ARMTargetParser::getSubArch(ArchKind);
- ArchProfile = llvm::ARMTargetParser::parseArchProfile(SubArch);
- ArchVersion = llvm::ARMTargetParser::parseArchVersion(SubArch);
+ SubArch = llvm::ARM::getSubArch(ArchKind);
+ ArchProfile = llvm::ARM::parseArchProfile(SubArch);
+ ArchVersion = llvm::ARM::parseArchVersion(SubArch);
// cache CPU related strings
CPUAttr = getCPUAttr();
}
StringRef getDefaultCPU(StringRef ArchName) const {
- const char *DefaultCPU = llvm::ARMTargetParser::getDefaultCPU(ArchName);
+ const char *DefaultCPU = llvm::ARM::getDefaultCPU(ArchName);
return DefaultCPU ? DefaultCPU : "";
}
// For Cortex variants, it's slightly different.
switch(ArchKind) {
default:
- CPUAttr = llvm::ARMTargetParser::getCPUAttr(ArchKind);
+ CPUAttr = llvm::ARM::getCPUAttr(ArchKind);
return CPUAttr ? CPUAttr : "";
case llvm::ARM::AK_ARMV6M:
case llvm::ARM::AK_ARMV6SM:
bool setCPU(const std::string &Name) override {
if (Name != "generic")
- setArchInfo(llvm::ARMTargetParser::parseCPUArch(Name));
+ setArchInfo(llvm::ARM::parseCPUArch(Name));
if (ArchKind == llvm::ARM::AK_INVALID)
return false;
// Get SubArch (vN).
static int getARMSubArchVersionNumber(const llvm::Triple &Triple) {
llvm::StringRef Arch = Triple.getArchName();
- return llvm::ARMTargetParser::parseArchVersion(Arch);
+ return llvm::ARM::parseArchVersion(Arch);
}
// True if M-profile.
static bool isARMMProfile(const llvm::Triple &Triple) {
llvm::StringRef Arch = Triple.getArchName();
- unsigned Profile = llvm::ARMTargetParser::parseArchProfile(Arch);
+ unsigned Profile = llvm::ARM::parseArchProfile(Arch);
return Profile == llvm::ARM::PK_M;
}
static void getARMHWDivFeatures(const Driver &D, const Arg *A,
const ArgList &Args, StringRef HWDiv,
std::vector<const char *> &Features) {
- unsigned HWDivID = llvm::ARMTargetParser::parseHWDiv(HWDiv);
- if (!llvm::ARMTargetParser::getHWDivFeatures(HWDivID, Features))
+ unsigned HWDivID = llvm::ARM::parseHWDiv(HWDiv);
+ if (!llvm::ARM::getHWDivFeatures(HWDivID, Features))
D.Diag(diag::err_drv_clang_unsupported) << A->getAsString(Args);
}
static void getARMFPUFeatures(const Driver &D, const Arg *A,
const ArgList &Args, StringRef FPU,
std::vector<const char *> &Features) {
- unsigned FPUID = llvm::ARMTargetParser::parseFPU(FPU);
- if (!llvm::ARMTargetParser::getFPUFeatures(FPUID, Features))
+ unsigned FPUID = llvm::ARM::parseFPU(FPU);
+ if (!llvm::ARM::getFPUFeatures(FPUID, Features))
D.Diag(diag::err_drv_clang_unsupported) << A->getAsString(Args);
}
llvm::StringRef ArchName,
const llvm::Triple &Triple) {
std::string MArch = arm::getARMArch(ArchName, Triple);
- if (llvm::ARMTargetParser::parseArch(MArch) == llvm::ARM::AK_INVALID)
+ if (llvm::ARM::parseArch(MArch) == llvm::ARM::AK_INVALID)
D.Diag(diag::err_drv_clang_unsupported) << A->getAsString(Args);
}
// FIXME: This is redundant with -mcpu, why does LLVM use this.
const char *arm::getLLVMArchSuffixForARM(StringRef CPU, StringRef Arch) {
if (CPU == "generic")
- return llvm::ARMTargetParser::getSubArch(
- llvm::ARMTargetParser::parseArch(Arch));
+ return llvm::ARM::getSubArch(
+ llvm::ARM::parseArch(Arch));
- unsigned ArchKind = llvm::ARMTargetParser::parseCPUArch(CPU);
+ unsigned ArchKind = llvm::ARM::parseCPUArch(CPU);
if (ArchKind == llvm::ARM::AK_INVALID)
return "";
- return llvm::ARMTargetParser::getSubArch(ArchKind);
+ return llvm::ARM::getSubArch(ArchKind);
}
void arm::appendEBLinkFlags(const ArgList &Args, ArgStringList &CmdArgs,