| // Copyright 2014 the V8 project authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| #ifndef V8_COMPILER_BACKEND_ARM64_INSTRUCTION_CODES_ARM64_H_ |
| #define V8_COMPILER_BACKEND_ARM64_INSTRUCTION_CODES_ARM64_H_ |
| |
| namespace v8 { |
| namespace internal { |
| namespace compiler { |
| |
| // ARM64-specific opcodes that specify which assembly sequence to emit. |
| // Most opcodes specify a single instruction. |
| #define TARGET_ARCH_OPCODE_LIST(V) \ |
| V(Arm64Add) \ |
| V(Arm64Add32) \ |
| V(Arm64And) \ |
| V(Arm64And32) \ |
| V(Arm64Bic) \ |
| V(Arm64Bic32) \ |
| V(Arm64Clz) \ |
| V(Arm64Clz32) \ |
| V(Arm64Cmp) \ |
| V(Arm64Cmp32) \ |
| V(Arm64Cmn) \ |
| V(Arm64Cmn32) \ |
| V(Arm64Cnt) \ |
| V(Arm64Tst) \ |
| V(Arm64Tst32) \ |
| V(Arm64Or) \ |
| V(Arm64Or32) \ |
| V(Arm64Orn) \ |
| V(Arm64Orn32) \ |
| V(Arm64Eor) \ |
| V(Arm64Eor32) \ |
| V(Arm64Eon) \ |
| V(Arm64Eon32) \ |
| V(Arm64Saddlp) \ |
| V(Arm64Sub) \ |
| V(Arm64Sub32) \ |
| V(Arm64Mul) \ |
| V(Arm64Mul32) \ |
| V(Arm64Smull) \ |
| V(Arm64Smull2) \ |
| V(Arm64Uaddlp) \ |
| V(Arm64Umull) \ |
| V(Arm64Umull2) \ |
| V(Arm64Madd) \ |
| V(Arm64Madd32) \ |
| V(Arm64Msub) \ |
| V(Arm64Msub32) \ |
| V(Arm64Mneg) \ |
| V(Arm64Mneg32) \ |
| V(Arm64Idiv) \ |
| V(Arm64Idiv32) \ |
| V(Arm64Udiv) \ |
| V(Arm64Udiv32) \ |
| V(Arm64Imod) \ |
| V(Arm64Imod32) \ |
| V(Arm64Umod) \ |
| V(Arm64Umod32) \ |
| V(Arm64Not) \ |
| V(Arm64Not32) \ |
| V(Arm64Lsl) \ |
| V(Arm64Lsl32) \ |
| V(Arm64Lsr) \ |
| V(Arm64Lsr32) \ |
| V(Arm64Asr) \ |
| V(Arm64Asr32) \ |
| V(Arm64Ror) \ |
| V(Arm64Ror32) \ |
| V(Arm64Mov32) \ |
| V(Arm64Sxtb32) \ |
| V(Arm64Sxth32) \ |
| V(Arm64Sxtb) \ |
| V(Arm64Sxth) \ |
| V(Arm64Sxtw) \ |
| V(Arm64Sbfx) \ |
| V(Arm64Sbfx32) \ |
| V(Arm64Ubfx) \ |
| V(Arm64Ubfx32) \ |
| V(Arm64Ubfiz32) \ |
| V(Arm64Bfi) \ |
| V(Arm64Rbit) \ |
| V(Arm64Rbit32) \ |
| V(Arm64Rev) \ |
| V(Arm64Rev32) \ |
| V(Arm64TestAndBranch32) \ |
| V(Arm64TestAndBranch) \ |
| V(Arm64CompareAndBranch32) \ |
| V(Arm64CompareAndBranch) \ |
| V(Arm64Claim) \ |
| V(Arm64Poke) \ |
| V(Arm64PokePair) \ |
| V(Arm64Peek) \ |
| V(Arm64Float32Cmp) \ |
| V(Arm64Float32Add) \ |
| V(Arm64Float32Sub) \ |
| V(Arm64Float32Mul) \ |
| V(Arm64Float32Div) \ |
| V(Arm64Float32Abs) \ |
| V(Arm64Float32Neg) \ |
| V(Arm64Float32Sqrt) \ |
| V(Arm64Float32Fnmul) \ |
| V(Arm64Float32RoundDown) \ |
| V(Arm64Float32Max) \ |
| V(Arm64Float32Min) \ |
| V(Arm64Float64Cmp) \ |
| V(Arm64Float64Add) \ |
| V(Arm64Float64Sub) \ |
| V(Arm64Float64Mul) \ |
| V(Arm64Float64Div) \ |
| V(Arm64Float64Mod) \ |
| V(Arm64Float64Max) \ |
| V(Arm64Float64Min) \ |
| V(Arm64Float64Abs) \ |
| V(Arm64Float64Neg) \ |
| V(Arm64Float64Sqrt) \ |
| V(Arm64Float64Fnmul) \ |
| V(Arm64Float64RoundDown) \ |
| V(Arm64Float32RoundUp) \ |
| V(Arm64Float64RoundUp) \ |
| V(Arm64Float64RoundTiesAway) \ |
| V(Arm64Float32RoundTruncate) \ |
| V(Arm64Float64RoundTruncate) \ |
| V(Arm64Float32RoundTiesEven) \ |
| V(Arm64Float64RoundTiesEven) \ |
| V(Arm64Float64SilenceNaN) \ |
| V(Arm64Float32ToFloat64) \ |
| V(Arm64Float64ToFloat32) \ |
| V(Arm64Float32ToInt32) \ |
| V(Arm64Float64ToInt32) \ |
| V(Arm64Float32ToUint32) \ |
| V(Arm64Float64ToUint32) \ |
| V(Arm64Float32ToInt64) \ |
| V(Arm64Float64ToInt64) \ |
| V(Arm64Float32ToUint64) \ |
| V(Arm64Float64ToUint64) \ |
| V(Arm64Int32ToFloat32) \ |
| V(Arm64Int32ToFloat64) \ |
| V(Arm64Int64ToFloat32) \ |
| V(Arm64Int64ToFloat64) \ |
| V(Arm64Uint32ToFloat32) \ |
| V(Arm64Uint32ToFloat64) \ |
| V(Arm64Uint64ToFloat32) \ |
| V(Arm64Uint64ToFloat64) \ |
| V(Arm64Float64ExtractLowWord32) \ |
| V(Arm64Float64ExtractHighWord32) \ |
| V(Arm64Float64InsertLowWord32) \ |
| V(Arm64Float64InsertHighWord32) \ |
| V(Arm64Float64MoveU64) \ |
| V(Arm64U64MoveFloat64) \ |
| V(Arm64LdrS) \ |
| V(Arm64StrS) \ |
| V(Arm64LdrD) \ |
| V(Arm64StrD) \ |
| V(Arm64LdrQ) \ |
| V(Arm64StrQ) \ |
| V(Arm64Ldrb) \ |
| V(Arm64Ldrsb) \ |
| V(Arm64Strb) \ |
| V(Arm64Ldrh) \ |
| V(Arm64Ldrsh) \ |
| V(Arm64Strh) \ |
| V(Arm64Ldrsw) \ |
| V(Arm64LdrW) \ |
| V(Arm64StrW) \ |
| V(Arm64Ldr) \ |
| V(Arm64LdrDecompressTaggedSigned) \ |
| V(Arm64LdrDecompressTaggedPointer) \ |
| V(Arm64LdrDecompressAnyTagged) \ |
| V(Arm64Str) \ |
| V(Arm64StrCompressTagged) \ |
| V(Arm64DmbIsh) \ |
| V(Arm64DsbIsb) \ |
| V(Arm64Sxtl) \ |
| V(Arm64Sxtl2) \ |
| V(Arm64Uxtl) \ |
| V(Arm64Uxtl2) \ |
| V(Arm64F64x2Splat) \ |
| V(Arm64F64x2ExtractLane) \ |
| V(Arm64F64x2ReplaceLane) \ |
| V(Arm64F64x2Abs) \ |
| V(Arm64F64x2Neg) \ |
| V(Arm64F64x2Sqrt) \ |
| V(Arm64F64x2Add) \ |
| V(Arm64F64x2Sub) \ |
| V(Arm64F64x2Mul) \ |
| V(Arm64F64x2Div) \ |
| V(Arm64F64x2Min) \ |
| V(Arm64F64x2Max) \ |
| V(Arm64F64x2Eq) \ |
| V(Arm64F64x2Ne) \ |
| V(Arm64F64x2Lt) \ |
| V(Arm64F64x2Le) \ |
| V(Arm64F64x2Qfma) \ |
| V(Arm64F64x2Qfms) \ |
| V(Arm64F64x2Pmin) \ |
| V(Arm64F64x2Pmax) \ |
| V(Arm64F32x4Splat) \ |
| V(Arm64F32x4ExtractLane) \ |
| V(Arm64F32x4ReplaceLane) \ |
| V(Arm64F32x4SConvertI32x4) \ |
| V(Arm64F32x4UConvertI32x4) \ |
| V(Arm64F32x4Abs) \ |
| V(Arm64F32x4Neg) \ |
| V(Arm64F32x4Sqrt) \ |
| V(Arm64F32x4RecipApprox) \ |
| V(Arm64F32x4RecipSqrtApprox) \ |
| V(Arm64F32x4Add) \ |
| V(Arm64F32x4AddHoriz) \ |
| V(Arm64F32x4Sub) \ |
| V(Arm64F32x4Mul) \ |
| V(Arm64F32x4Div) \ |
| V(Arm64F32x4Min) \ |
| V(Arm64F32x4Max) \ |
| V(Arm64F32x4Eq) \ |
| V(Arm64F32x4Ne) \ |
| V(Arm64F32x4Lt) \ |
| V(Arm64F32x4Le) \ |
| V(Arm64F32x4Qfma) \ |
| V(Arm64F32x4Qfms) \ |
| V(Arm64F32x4Pmin) \ |
| V(Arm64F32x4Pmax) \ |
| V(Arm64I64x2Splat) \ |
| V(Arm64I64x2ExtractLane) \ |
| V(Arm64I64x2ReplaceLane) \ |
| V(Arm64I64x2Neg) \ |
| V(Arm64I64x2Shl) \ |
| V(Arm64I64x2ShrS) \ |
| V(Arm64I64x2Add) \ |
| V(Arm64I64x2Sub) \ |
| V(Arm64I64x2Mul) \ |
| V(Arm64I64x2Eq) \ |
| V(Arm64I64x2ShrU) \ |
| V(Arm64I32x4Splat) \ |
| V(Arm64I32x4ExtractLane) \ |
| V(Arm64I32x4ReplaceLane) \ |
| V(Arm64I32x4SConvertF32x4) \ |
| V(Arm64I32x4Neg) \ |
| V(Arm64I32x4Shl) \ |
| V(Arm64I32x4ShrS) \ |
| V(Arm64I32x4Add) \ |
| V(Arm64I32x4AddHoriz) \ |
| V(Arm64I32x4Sub) \ |
| V(Arm64I32x4Mul) \ |
| V(Arm64I32x4Mla) \ |
| V(Arm64I32x4Mls) \ |
| V(Arm64I32x4MinS) \ |
| V(Arm64I32x4MaxS) \ |
| V(Arm64I32x4Eq) \ |
| V(Arm64I32x4Ne) \ |
| V(Arm64I32x4GtS) \ |
| V(Arm64I32x4GeS) \ |
| V(Arm64I32x4UConvertF32x4) \ |
| V(Arm64I32x4ShrU) \ |
| V(Arm64I32x4MinU) \ |
| V(Arm64I32x4MaxU) \ |
| V(Arm64I32x4GtU) \ |
| V(Arm64I32x4GeU) \ |
| V(Arm64I32x4Abs) \ |
| V(Arm64I32x4BitMask) \ |
| V(Arm64I32x4DotI16x8S) \ |
| V(Arm64I16x8Splat) \ |
| V(Arm64I16x8ExtractLaneU) \ |
| V(Arm64I16x8ExtractLaneS) \ |
| V(Arm64I16x8ReplaceLane) \ |
| V(Arm64I16x8Neg) \ |
| V(Arm64I16x8Shl) \ |
| V(Arm64I16x8ShrS) \ |
| V(Arm64I16x8SConvertI32x4) \ |
| V(Arm64I16x8Add) \ |
| V(Arm64I16x8AddSatS) \ |
| V(Arm64I16x8AddHoriz) \ |
| V(Arm64I16x8Sub) \ |
| V(Arm64I16x8SubSatS) \ |
| V(Arm64I16x8Mul) \ |
| V(Arm64I16x8Mla) \ |
| V(Arm64I16x8Mls) \ |
| V(Arm64I16x8MinS) \ |
| V(Arm64I16x8MaxS) \ |
| V(Arm64I16x8Eq) \ |
| V(Arm64I16x8Ne) \ |
| V(Arm64I16x8GtS) \ |
| V(Arm64I16x8GeS) \ |
| V(Arm64I16x8ShrU) \ |
| V(Arm64I16x8UConvertI32x4) \ |
| V(Arm64I16x8AddSatU) \ |
| V(Arm64I16x8SubSatU) \ |
| V(Arm64I16x8MinU) \ |
| V(Arm64I16x8MaxU) \ |
| V(Arm64I16x8GtU) \ |
| V(Arm64I16x8GeU) \ |
| V(Arm64I16x8RoundingAverageU) \ |
| V(Arm64I16x8Q15MulRSatS) \ |
| V(Arm64I16x8Abs) \ |
| V(Arm64I16x8BitMask) \ |
| V(Arm64I8x16Splat) \ |
| V(Arm64I8x16ExtractLaneU) \ |
| V(Arm64I8x16ExtractLaneS) \ |
| V(Arm64I8x16ReplaceLane) \ |
| V(Arm64I8x16Neg) \ |
| V(Arm64I8x16Shl) \ |
| V(Arm64I8x16ShrS) \ |
| V(Arm64I8x16SConvertI16x8) \ |
| V(Arm64I8x16Add) \ |
| V(Arm64I8x16AddSatS) \ |
| V(Arm64I8x16Sub) \ |
| V(Arm64I8x16SubSatS) \ |
| V(Arm64I8x16Mul) \ |
| V(Arm64I8x16Mla) \ |
| V(Arm64I8x16Mls) \ |
| V(Arm64I8x16MinS) \ |
| V(Arm64I8x16MaxS) \ |
| V(Arm64I8x16Eq) \ |
| V(Arm64I8x16Ne) \ |
| V(Arm64I8x16GtS) \ |
| V(Arm64I8x16GeS) \ |
| V(Arm64I8x16ShrU) \ |
| V(Arm64I8x16UConvertI16x8) \ |
| V(Arm64I8x16AddSatU) \ |
| V(Arm64I8x16SubSatU) \ |
| V(Arm64I8x16MinU) \ |
| V(Arm64I8x16MaxU) \ |
| V(Arm64I8x16GtU) \ |
| V(Arm64I8x16GeU) \ |
| V(Arm64I8x16RoundingAverageU) \ |
| V(Arm64I8x16Abs) \ |
| V(Arm64I8x16BitMask) \ |
| V(Arm64S128Const) \ |
| V(Arm64S128Zero) \ |
| V(Arm64S128Dup) \ |
| V(Arm64S128And) \ |
| V(Arm64S128Or) \ |
| V(Arm64S128Xor) \ |
| V(Arm64S128Not) \ |
| V(Arm64S128Select) \ |
| V(Arm64S128AndNot) \ |
| V(Arm64S32x4ZipLeft) \ |
| V(Arm64S32x4ZipRight) \ |
| V(Arm64S32x4UnzipLeft) \ |
| V(Arm64S32x4UnzipRight) \ |
| V(Arm64S32x4TransposeLeft) \ |
| V(Arm64S32x4TransposeRight) \ |
| V(Arm64S32x4Shuffle) \ |
| V(Arm64S16x8ZipLeft) \ |
| V(Arm64S16x8ZipRight) \ |
| V(Arm64S16x8UnzipLeft) \ |
| V(Arm64S16x8UnzipRight) \ |
| V(Arm64S16x8TransposeLeft) \ |
| V(Arm64S16x8TransposeRight) \ |
| V(Arm64S8x16ZipLeft) \ |
| V(Arm64S8x16ZipRight) \ |
| V(Arm64S8x16UnzipLeft) \ |
| V(Arm64S8x16UnzipRight) \ |
| V(Arm64S8x16TransposeLeft) \ |
| V(Arm64S8x16TransposeRight) \ |
| V(Arm64S8x16Concat) \ |
| V(Arm64I8x16Swizzle) \ |
| V(Arm64I8x16Shuffle) \ |
| V(Arm64S32x2Reverse) \ |
| V(Arm64S16x4Reverse) \ |
| V(Arm64S16x2Reverse) \ |
| V(Arm64S8x8Reverse) \ |
| V(Arm64S8x4Reverse) \ |
| V(Arm64S8x2Reverse) \ |
| V(Arm64V128AnyTrue) \ |
| V(Arm64V32x4AllTrue) \ |
| V(Arm64V16x8AllTrue) \ |
| V(Arm64V8x16AllTrue) \ |
| V(Arm64LoadSplat) \ |
| V(Arm64S128Load8x8S) \ |
| V(Arm64S128Load8x8U) \ |
| V(Arm64S128Load16x4S) \ |
| V(Arm64S128Load16x4U) \ |
| V(Arm64S128Load32x2S) \ |
| V(Arm64S128Load32x2U) \ |
| V(Arm64S128Load32Zero) \ |
| V(Arm64S128Load64Zero) \ |
| V(Arm64Word64AtomicLoadUint8) \ |
| V(Arm64Word64AtomicLoadUint16) \ |
| V(Arm64Word64AtomicLoadUint32) \ |
| V(Arm64Word64AtomicLoadUint64) \ |
| V(Arm64Word64AtomicStoreWord8) \ |
| V(Arm64Word64AtomicStoreWord16) \ |
| V(Arm64Word64AtomicStoreWord32) \ |
| V(Arm64Word64AtomicStoreWord64) \ |
| V(Arm64Word64AtomicAddUint8) \ |
| V(Arm64Word64AtomicAddUint16) \ |
| V(Arm64Word64AtomicAddUint32) \ |
| V(Arm64Word64AtomicAddUint64) \ |
| V(Arm64Word64AtomicSubUint8) \ |
| V(Arm64Word64AtomicSubUint16) \ |
| V(Arm64Word64AtomicSubUint32) \ |
| V(Arm64Word64AtomicSubUint64) \ |
| V(Arm64Word64AtomicAndUint8) \ |
| V(Arm64Word64AtomicAndUint16) \ |
| V(Arm64Word64AtomicAndUint32) \ |
| V(Arm64Word64AtomicAndUint64) \ |
| V(Arm64Word64AtomicOrUint8) \ |
| V(Arm64Word64AtomicOrUint16) \ |
| V(Arm64Word64AtomicOrUint32) \ |
| V(Arm64Word64AtomicOrUint64) \ |
| V(Arm64Word64AtomicXorUint8) \ |
| V(Arm64Word64AtomicXorUint16) \ |
| V(Arm64Word64AtomicXorUint32) \ |
| V(Arm64Word64AtomicXorUint64) \ |
| V(Arm64Word64AtomicExchangeUint8) \ |
| V(Arm64Word64AtomicExchangeUint16) \ |
| V(Arm64Word64AtomicExchangeUint32) \ |
| V(Arm64Word64AtomicExchangeUint64) \ |
| V(Arm64Word64AtomicCompareExchangeUint8) \ |
| V(Arm64Word64AtomicCompareExchangeUint16) \ |
| V(Arm64Word64AtomicCompareExchangeUint32) \ |
| V(Arm64Word64AtomicCompareExchangeUint64) |
| |
| // Addressing modes represent the "shape" of inputs to an instruction. |
| // Many instructions support multiple addressing modes. Addressing modes |
| // are encoded into the InstructionCode of the instruction and tell the |
| // code generator after register allocation which assembler method to call. |
| // |
| // We use the following local notation for addressing modes: |
| // |
| // R = register |
| // O = register or stack slot |
| // D = double register |
| // I = immediate (handle, external, int32) |
| // MRI = [register + immediate] |
| // MRR = [register + register] |
| #define TARGET_ADDRESSING_MODE_LIST(V) \ |
| V(MRI) /* [%r0 + K] */ \ |
| V(MRR) /* [%r0 + %r1] */ \ |
| V(Operand2_R_LSL_I) /* %r0 LSL K */ \ |
| V(Operand2_R_LSR_I) /* %r0 LSR K */ \ |
| V(Operand2_R_ASR_I) /* %r0 ASR K */ \ |
| V(Operand2_R_ROR_I) /* %r0 ROR K */ \ |
| V(Operand2_R_UXTB) /* %r0 UXTB (unsigned extend byte) */ \ |
| V(Operand2_R_UXTH) /* %r0 UXTH (unsigned extend halfword) */ \ |
| V(Operand2_R_SXTB) /* %r0 SXTB (signed extend byte) */ \ |
| V(Operand2_R_SXTH) /* %r0 SXTH (signed extend halfword) */ \ |
| V(Operand2_R_SXTW) /* %r0 SXTW (signed extend word) */ \ |
| V(Root) /* [%rr + K] */ |
| |
| } // namespace compiler |
| } // namespace internal |
| } // namespace v8 |
| |
| #endif // V8_COMPILER_BACKEND_ARM64_INSTRUCTION_CODES_ARM64_H_ |