//! ARM 64-bit Instruction Set Architecture. use crate::ir::condcodes::IntCC; use crate::ir::Function; use crate::isa::aarch64::settings as aarch64_settings; use crate::isa::Builder as IsaBuilder; use crate::machinst::{compile, MachBackend, MachCompileResult, TargetIsaAdapter, VCode}; use crate::result::CodegenResult; use crate::settings as shared_settings; use alloc::{boxed::Box, vec::Vec}; use core::hash::{Hash, Hasher}; use regalloc::{PrettyPrint, RealRegUniverse}; use target_lexicon::{Aarch64Architecture, Architecture, Triple}; // New backend: mod abi; pub(crate) mod inst; mod lower; mod lower_inst; mod settings; use inst::create_reg_universe; use self::inst::EmitInfo; /// An AArch64 backend. pub struct AArch64Backend { triple: Triple, flags: shared_settings::Flags, isa_flags: aarch64_settings::Flags, reg_universe: RealRegUniverse, } impl AArch64Backend { /// Create a new AArch64 backend with the given (shared) flags. pub fn new_with_flags( triple: Triple, flags: shared_settings::Flags, isa_flags: aarch64_settings::Flags, ) -> AArch64Backend { let reg_universe = create_reg_universe(&flags); AArch64Backend { triple, flags, isa_flags, reg_universe, } } /// This performs lowering to VCode, register-allocates the code, computes block layout and /// finalizes branches. The result is ready for binary emission. fn compile_vcode( &self, func: &Function, flags: shared_settings::Flags, ) -> CodegenResult> { let emit_info = EmitInfo::new(flags.clone()); let abi = Box::new(abi::AArch64ABICallee::new(func, flags)?); compile::compile::(func, self, abi, emit_info) } } impl MachBackend for AArch64Backend { fn compile_function( &self, func: &Function, want_disasm: bool, ) -> CodegenResult { let flags = self.flags(); let vcode = self.compile_vcode(func, flags.clone())?; let buffer = vcode.emit(); let frame_size = vcode.frame_size(); let stackslot_offsets = vcode.stackslot_offsets().clone(); let disasm = if want_disasm { Some(vcode.show_rru(Some(&create_reg_universe(flags)))) } else { None }; let buffer = buffer.finish(); Ok(MachCompileResult { buffer, frame_size, disasm, value_labels_ranges: Default::default(), stackslot_offsets, }) } fn name(&self) -> &'static str { "aarch64" } fn triple(&self) -> Triple { self.triple.clone() } fn flags(&self) -> &shared_settings::Flags { &self.flags } fn isa_flags(&self) -> Vec { self.isa_flags.iter().collect() } fn hash_all_flags(&self, mut hasher: &mut dyn Hasher) { self.flags.hash(&mut hasher); self.isa_flags.hash(&mut hasher); } fn reg_universe(&self) -> &RealRegUniverse { &self.reg_universe } fn unsigned_add_overflow_condition(&self) -> IntCC { // Unsigned `>=`; this corresponds to the carry flag set on aarch64, which happens on // overflow of an add. IntCC::UnsignedGreaterThanOrEqual } fn unsigned_sub_overflow_condition(&self) -> IntCC { // unsigned `<`; this corresponds to the carry flag cleared on aarch64, which happens on // underflow of a subtract (aarch64 follows a carry-cleared-on-borrow convention, the // opposite of x86). IntCC::UnsignedLessThan } #[cfg(feature = "unwind")] fn emit_unwind_info( &self, result: &MachCompileResult, kind: crate::machinst::UnwindInfoKind, ) -> CodegenResult> { use crate::isa::unwind::UnwindInfo; use crate::machinst::UnwindInfoKind; Ok(match kind { UnwindInfoKind::SystemV => { let mapper = self::inst::unwind::systemv::RegisterMapper; Some(UnwindInfo::SystemV( crate::isa::unwind::systemv::create_unwind_info_from_insts( &result.buffer.unwind_info[..], result.buffer.data.len(), &mapper, )?, )) } UnwindInfoKind::Windows => { // TODO: support Windows unwind info on AArch64 None } _ => None, }) } #[cfg(feature = "unwind")] fn create_systemv_cie(&self) -> Option { Some(inst::unwind::systemv::create_cie()) } } /// Create a new `isa::Builder`. pub fn isa_builder(triple: Triple) -> IsaBuilder { assert!(triple.architecture == Architecture::Aarch64(Aarch64Architecture::Aarch64)); IsaBuilder { triple, setup: aarch64_settings::builder(), constructor: |triple, shared_flags, builder| { let isa_flags = aarch64_settings::Flags::new(&shared_flags, builder); let backend = AArch64Backend::new_with_flags(triple, shared_flags, isa_flags); Box::new(TargetIsaAdapter::new(backend)) }, } } #[cfg(test)] mod test { use super::*; use crate::cursor::{Cursor, FuncCursor}; use crate::ir::types::*; use crate::ir::{AbiParam, ExternalName, Function, InstBuilder, Signature}; use crate::isa::CallConv; use crate::settings; use crate::settings::Configurable; use core::str::FromStr; use target_lexicon::Triple; #[test] fn test_compile_function() { let name = ExternalName::testcase("test0"); let mut sig = Signature::new(CallConv::SystemV); sig.params.push(AbiParam::new(I32)); sig.returns.push(AbiParam::new(I32)); let mut func = Function::with_name_signature(name, sig); let bb0 = func.dfg.make_block(); let arg0 = func.dfg.append_block_param(bb0, I32); let mut pos = FuncCursor::new(&mut func); pos.insert_block(bb0); let v0 = pos.ins().iconst(I32, 0x1234); let v1 = pos.ins().iadd(arg0, v0); pos.ins().return_(&[v1]); let mut shared_flags_builder = settings::builder(); shared_flags_builder.set("opt_level", "none").unwrap(); let shared_flags = settings::Flags::new(shared_flags_builder); let isa_flags = aarch64_settings::Flags::new(&shared_flags, aarch64_settings::builder()); let backend = AArch64Backend::new_with_flags( Triple::from_str("aarch64").unwrap(), shared_flags, isa_flags, ); let buffer = backend.compile_function(&mut func, false).unwrap().buffer; let code = &buffer.data[..]; // stp x29, x30, [sp, #-16]! // mov x29, sp // mov x1, #0x1234 // add w0, w0, w1 // ldp x29, x30, [sp], #16 // ret let golden = vec![ 0xfd, 0x7b, 0xbf, 0xa9, 0xfd, 0x03, 0x00, 0x91, 0x81, 0x46, 0x82, 0xd2, 0x00, 0x00, 0x01, 0x0b, 0xfd, 0x7b, 0xc1, 0xa8, 0xc0, 0x03, 0x5f, 0xd6, ]; assert_eq!(code, &golden[..]); } #[test] fn test_branch_lowering() { let name = ExternalName::testcase("test0"); let mut sig = Signature::new(CallConv::SystemV); sig.params.push(AbiParam::new(I32)); sig.returns.push(AbiParam::new(I32)); let mut func = Function::with_name_signature(name, sig); let bb0 = func.dfg.make_block(); let arg0 = func.dfg.append_block_param(bb0, I32); let bb1 = func.dfg.make_block(); let bb2 = func.dfg.make_block(); let bb3 = func.dfg.make_block(); let mut pos = FuncCursor::new(&mut func); pos.insert_block(bb0); let v0 = pos.ins().iconst(I32, 0x1234); let v1 = pos.ins().iadd(arg0, v0); pos.ins().brnz(v1, bb1, &[]); pos.ins().jump(bb2, &[]); pos.insert_block(bb1); pos.ins().brnz(v1, bb2, &[]); pos.ins().jump(bb3, &[]); pos.insert_block(bb2); let v2 = pos.ins().iadd(v1, v0); pos.ins().brnz(v2, bb2, &[]); pos.ins().jump(bb1, &[]); pos.insert_block(bb3); let v3 = pos.ins().isub(v1, v0); pos.ins().return_(&[v3]); let mut shared_flags_builder = settings::builder(); shared_flags_builder.set("opt_level", "none").unwrap(); let shared_flags = settings::Flags::new(shared_flags_builder); let isa_flags = aarch64_settings::Flags::new(&shared_flags, aarch64_settings::builder()); let backend = AArch64Backend::new_with_flags( Triple::from_str("aarch64").unwrap(), shared_flags, isa_flags, ); let result = backend .compile_function(&mut func, /* want_disasm = */ false) .unwrap(); let code = &result.buffer.data[..]; // stp x29, x30, [sp, #-16]! // mov x29, sp // mov x1, #0x1234 // #4660 // add w0, w0, w1 // mov w1, w0 // cbnz x1, 0x28 // mov x1, #0x1234 // #4660 // add w1, w0, w1 // mov w1, w1 // cbnz x1, 0x18 // mov w1, w0 // cbnz x1, 0x18 // mov x1, #0x1234 // #4660 // sub w0, w0, w1 // ldp x29, x30, [sp], #16 // ret let golden = vec![ 253, 123, 191, 169, 253, 3, 0, 145, 129, 70, 130, 210, 0, 0, 1, 11, 225, 3, 0, 42, 161, 0, 0, 181, 129, 70, 130, 210, 1, 0, 1, 11, 225, 3, 1, 42, 161, 255, 255, 181, 225, 3, 0, 42, 97, 255, 255, 181, 129, 70, 130, 210, 0, 0, 1, 75, 253, 123, 193, 168, 192, 3, 95, 214, ]; assert_eq!(code, &golden[..]); } }