diff --git a/llvm/lib/Target/AArch64/GISel/AArch64InstructionSelector.cpp b/llvm/lib/Target/AArch64/GISel/AArch64InstructionSelector.cpp index decf71fa8af0d..bf44fa73e53db 100644 --- a/llvm/lib/Target/AArch64/GISel/AArch64InstructionSelector.cpp +++ b/llvm/lib/Target/AArch64/GISel/AArch64InstructionSelector.cpp @@ -5219,7 +5219,7 @@ bool AArch64InstructionSelector::isWorthFoldingIntoExtendedReg( // Always fold if there is one use, or if we're optimizing for size. Register DefReg = MI.getOperand(0).getReg(); if (MRI.hasOneNonDBGUse(DefReg) || - MI.getParent()->getParent()->getFunction().hasMinSize()) + MI.getParent()->getParent()->getFunction().hasOptSize()) return true; // It's better to avoid folding and recomputing shifts when we don't have a diff --git a/llvm/test/CodeGen/AArch64/GlobalISel/load-addressing-modes.mir b/llvm/test/CodeGen/AArch64/GlobalISel/load-addressing-modes.mir index 8b8d665414693..91b681b9b9bb7 100644 --- a/llvm/test/CodeGen/AArch64/GlobalISel/load-addressing-modes.mir +++ b/llvm/test/CodeGen/AArch64/GlobalISel/load-addressing-modes.mir @@ -23,7 +23,7 @@ define void @ldrhrox(i64* %addr) { ret void } define void @ldbbrox(i64* %addr) { ret void } define void @ldrqrox(i64* %addr) { ret void } - attributes #0 = { optsize minsize } + attributes #0 = { optsize } attributes #1 = { "target-features"="+lsl-fast" } ...