| // Copyright 2014 the V8 project authors. All rights reserved. |
| // Use of this source code is governed by a BSD-style license that can be |
| // found in the LICENSE file. |
| |
| #ifndef V8_COMPILER_BACKEND_INSTRUCTION_H_ |
| #define V8_COMPILER_BACKEND_INSTRUCTION_H_ |
| |
| #include <deque> |
| #include <iosfwd> |
| #include <map> |
| #include <set> |
| |
| #include "src/base/compiler-specific.h" |
| #include "src/codegen/external-reference.h" |
| #include "src/codegen/register-arch.h" |
| #include "src/codegen/source-position.h" |
| #include "src/common/globals.h" |
| #include "src/compiler/backend/instruction-codes.h" |
| #include "src/compiler/common-operator.h" |
| #include "src/compiler/frame.h" |
| #include "src/compiler/opcodes.h" |
| #include "src/numbers/double.h" |
| #include "src/zone/zone-allocator.h" |
| |
| namespace v8 { |
| namespace internal { |
| |
| class RegisterConfiguration; |
| |
| namespace compiler { |
| |
| class Schedule; |
| class SourcePositionTable; |
| |
| class V8_EXPORT_PRIVATE InstructionOperand { |
| public: |
| static const int kInvalidVirtualRegister = -1; |
| |
| enum Kind { |
| INVALID, |
| UNALLOCATED, |
| CONSTANT, |
| IMMEDIATE, |
| // Location operand kinds. |
| EXPLICIT, |
| ALLOCATED, |
| FIRST_LOCATION_OPERAND_KIND = EXPLICIT |
| // Location operand kinds must be last. |
| }; |
| |
| InstructionOperand() : InstructionOperand(INVALID) {} |
| |
| Kind kind() const { return KindField::decode(value_); } |
| |
| #define INSTRUCTION_OPERAND_PREDICATE(name, type) \ |
| bool Is##name() const { return kind() == type; } |
| INSTRUCTION_OPERAND_PREDICATE(Invalid, INVALID) |
| // UnallocatedOperands are place-holder operands created before register |
| // allocation. They later are assigned registers and become AllocatedOperands. |
| INSTRUCTION_OPERAND_PREDICATE(Unallocated, UNALLOCATED) |
| // Constant operands participate in register allocation. They are allocated to |
| // registers but have a special "spilling" behavior. When a ConstantOperand |
| // value must be rematerialized, it is loaded from an immediate constant |
| // rather from an unspilled slot. |
| INSTRUCTION_OPERAND_PREDICATE(Constant, CONSTANT) |
| // ImmediateOperands do not participate in register allocation and are only |
| // embedded directly in instructions, e.g. small integers and on some |
| // platforms Objects. |
| INSTRUCTION_OPERAND_PREDICATE(Immediate, IMMEDIATE) |
| // ExplicitOperands do not participate in register allocation. They are |
| // created by the instruction selector for direct access to registers and |
| // stack slots, completely bypassing the register allocator. They are never |
| // associated with a virtual register |
| INSTRUCTION_OPERAND_PREDICATE(Explicit, EXPLICIT) |
| // AllocatedOperands are registers or stack slots that are assigned by the |
| // register allocator and are always associated with a virtual register. |
| INSTRUCTION_OPERAND_PREDICATE(Allocated, ALLOCATED) |
| #undef INSTRUCTION_OPERAND_PREDICATE |
| |
| inline bool IsAnyLocationOperand() const; |
| inline bool IsLocationOperand() const; |
| inline bool IsFPLocationOperand() const; |
| inline bool IsAnyRegister() const; |
| inline bool IsRegister() const; |
| inline bool IsFPRegister() const; |
| inline bool IsFloatRegister() const; |
| inline bool IsDoubleRegister() const; |
| inline bool IsSimd128Register() const; |
| inline bool IsAnyStackSlot() const; |
| inline bool IsStackSlot() const; |
| inline bool IsFPStackSlot() const; |
| inline bool IsFloatStackSlot() const; |
| inline bool IsDoubleStackSlot() const; |
| inline bool IsSimd128StackSlot() const; |
| |
| template <typename SubKindOperand> |
| static SubKindOperand* New(Zone* zone, const SubKindOperand& op) { |
| void* buffer = zone->New(sizeof(op)); |
| return new (buffer) SubKindOperand(op); |
| } |
| |
| static void ReplaceWith(InstructionOperand* dest, |
| const InstructionOperand* src) { |
| *dest = *src; |
| } |
| |
| bool Equals(const InstructionOperand& that) const { |
| return this->value_ == that.value_; |
| } |
| |
| bool Compare(const InstructionOperand& that) const { |
| return this->value_ < that.value_; |
| } |
| |
| bool EqualsCanonicalized(const InstructionOperand& that) const { |
| return this->GetCanonicalizedValue() == that.GetCanonicalizedValue(); |
| } |
| |
| bool CompareCanonicalized(const InstructionOperand& that) const { |
| return this->GetCanonicalizedValue() < that.GetCanonicalizedValue(); |
| } |
| |
| bool InterferesWith(const InstructionOperand& other) const; |
| |
| // APIs to aid debugging. For general-stream APIs, use operator<<. |
| void Print() const; |
| |
| protected: |
| explicit InstructionOperand(Kind kind) : value_(KindField::encode(kind)) {} |
| |
| inline uint64_t GetCanonicalizedValue() const; |
| |
| class KindField : public BitField64<Kind, 0, 3> {}; |
| |
| uint64_t value_; |
| }; |
| |
| using InstructionOperandVector = ZoneVector<InstructionOperand>; |
| |
| std::ostream& operator<<(std::ostream&, const InstructionOperand&); |
| |
| #define INSTRUCTION_OPERAND_CASTS(OperandType, OperandKind) \ |
| \ |
| static OperandType* cast(InstructionOperand* op) { \ |
| DCHECK_EQ(OperandKind, op->kind()); \ |
| return static_cast<OperandType*>(op); \ |
| } \ |
| \ |
| static const OperandType* cast(const InstructionOperand* op) { \ |
| DCHECK_EQ(OperandKind, op->kind()); \ |
| return static_cast<const OperandType*>(op); \ |
| } \ |
| \ |
| static OperandType cast(const InstructionOperand& op) { \ |
| DCHECK_EQ(OperandKind, op.kind()); \ |
| return *static_cast<const OperandType*>(&op); \ |
| } |
| |
| class UnallocatedOperand final : public InstructionOperand { |
| public: |
| enum BasicPolicy { FIXED_SLOT, EXTENDED_POLICY }; |
| |
| enum ExtendedPolicy { |
| NONE, |
| REGISTER_OR_SLOT, |
| REGISTER_OR_SLOT_OR_CONSTANT, |
| FIXED_REGISTER, |
| FIXED_FP_REGISTER, |
| MUST_HAVE_REGISTER, |
| MUST_HAVE_SLOT, |
| SAME_AS_FIRST_INPUT |
| }; |
| |
| // Lifetime of operand inside the instruction. |
| enum Lifetime { |
| // USED_AT_START operand is guaranteed to be live only at instruction start. |
| // The register allocator is free to assign the same register to some other |
| // operand used inside instruction (i.e. temporary or output). |
| USED_AT_START, |
| |
| // USED_AT_END operand is treated as live until the end of instruction. |
| // This means that register allocator will not reuse its register for any |
| // other operand inside instruction. |
| USED_AT_END |
| }; |
| |
| UnallocatedOperand(ExtendedPolicy policy, int virtual_register) |
| : UnallocatedOperand(virtual_register) { |
| value_ |= BasicPolicyField::encode(EXTENDED_POLICY); |
| value_ |= ExtendedPolicyField::encode(policy); |
| value_ |= LifetimeField::encode(USED_AT_END); |
| } |
| |
| UnallocatedOperand(BasicPolicy policy, int index, int virtual_register) |
| : UnallocatedOperand(virtual_register) { |
| DCHECK(policy == FIXED_SLOT); |
| value_ |= BasicPolicyField::encode(policy); |
| value_ |= static_cast<uint64_t>(static_cast<int64_t>(index)) |
| << FixedSlotIndexField::kShift; |
| DCHECK(this->fixed_slot_index() == index); |
| } |
| |
| UnallocatedOperand(ExtendedPolicy policy, int index, int virtual_register) |
| : UnallocatedOperand(virtual_register) { |
| DCHECK(policy == FIXED_REGISTER || policy == FIXED_FP_REGISTER); |
| value_ |= BasicPolicyField::encode(EXTENDED_POLICY); |
| value_ |= ExtendedPolicyField::encode(policy); |
| value_ |= LifetimeField::encode(USED_AT_END); |
| value_ |= FixedRegisterField::encode(index); |
| } |
| |
| UnallocatedOperand(ExtendedPolicy policy, Lifetime lifetime, |
| int virtual_register) |
| : UnallocatedOperand(virtual_register) { |
| value_ |= BasicPolicyField::encode(EXTENDED_POLICY); |
| value_ |= ExtendedPolicyField::encode(policy); |
| value_ |= LifetimeField::encode(lifetime); |
| } |
| |
| UnallocatedOperand(int reg_id, int slot_id, int virtual_register) |
| : UnallocatedOperand(FIXED_REGISTER, reg_id, virtual_register) { |
| value_ |= HasSecondaryStorageField::encode(true); |
| value_ |= SecondaryStorageField::encode(slot_id); |
| } |
| |
| UnallocatedOperand(const UnallocatedOperand& other, int virtual_register) { |
| DCHECK_NE(kInvalidVirtualRegister, virtual_register); |
| value_ = VirtualRegisterField::update( |
| other.value_, static_cast<uint32_t>(virtual_register)); |
| } |
| |
| // Predicates for the operand policy. |
| bool HasRegisterOrSlotPolicy() const { |
| return basic_policy() == EXTENDED_POLICY && |
| extended_policy() == REGISTER_OR_SLOT; |
| } |
| bool HasRegisterOrSlotOrConstantPolicy() const { |
| return basic_policy() == EXTENDED_POLICY && |
| extended_policy() == REGISTER_OR_SLOT_OR_CONSTANT; |
| } |
| bool HasFixedPolicy() const { |
| return basic_policy() == FIXED_SLOT || |
| extended_policy() == FIXED_REGISTER || |
| extended_policy() == FIXED_FP_REGISTER; |
| } |
| bool HasRegisterPolicy() const { |
| return basic_policy() == EXTENDED_POLICY && |
| extended_policy() == MUST_HAVE_REGISTER; |
| } |
| bool HasSlotPolicy() const { |
| return basic_policy() == EXTENDED_POLICY && |
| extended_policy() == MUST_HAVE_SLOT; |
| } |
| bool HasSameAsInputPolicy() const { |
| return basic_policy() == EXTENDED_POLICY && |
| extended_policy() == SAME_AS_FIRST_INPUT; |
| } |
| bool HasFixedSlotPolicy() const { return basic_policy() == FIXED_SLOT; } |
| bool HasFixedRegisterPolicy() const { |
| return basic_policy() == EXTENDED_POLICY && |
| extended_policy() == FIXED_REGISTER; |
| } |
| bool HasFixedFPRegisterPolicy() const { |
| return basic_policy() == EXTENDED_POLICY && |
| extended_policy() == FIXED_FP_REGISTER; |
| } |
| bool HasSecondaryStorage() const { |
| return basic_policy() == EXTENDED_POLICY && |
| extended_policy() == FIXED_REGISTER && |
| HasSecondaryStorageField::decode(value_); |
| } |
| int GetSecondaryStorage() const { |
| DCHECK(HasSecondaryStorage()); |
| return SecondaryStorageField::decode(value_); |
| } |
| |
| // [basic_policy]: Distinguish between FIXED_SLOT and all other policies. |
| BasicPolicy basic_policy() const { return BasicPolicyField::decode(value_); } |
| |
| // [extended_policy]: Only for non-FIXED_SLOT. The finer-grained policy. |
| ExtendedPolicy extended_policy() const { |
| DCHECK(basic_policy() == EXTENDED_POLICY); |
| return ExtendedPolicyField::decode(value_); |
| } |
| |
| // [fixed_slot_index]: Only for FIXED_SLOT. |
| int fixed_slot_index() const { |
| DCHECK(HasFixedSlotPolicy()); |
| return static_cast<int>(static_cast<int64_t>(value_) >> |
| FixedSlotIndexField::kShift); |
| } |
| |
| // [fixed_register_index]: Only for FIXED_REGISTER or FIXED_FP_REGISTER. |
| int fixed_register_index() const { |
| DCHECK(HasFixedRegisterPolicy() || HasFixedFPRegisterPolicy()); |
| return FixedRegisterField::decode(value_); |
| } |
| |
| // [virtual_register]: The virtual register ID for this operand. |
| int32_t virtual_register() const { |
| return static_cast<int32_t>(VirtualRegisterField::decode(value_)); |
| } |
| |
| // [lifetime]: Only for non-FIXED_SLOT. |
| bool IsUsedAtStart() const { |
| DCHECK(basic_policy() == EXTENDED_POLICY); |
| return LifetimeField::decode(value_) == USED_AT_START; |
| } |
| |
| INSTRUCTION_OPERAND_CASTS(UnallocatedOperand, UNALLOCATED) |
| |
| // The encoding used for UnallocatedOperand operands depends on the policy |
| // that is |
| // stored within the operand. The FIXED_SLOT policy uses a compact encoding |
| // because it accommodates a larger pay-load. |
| // |
| // For FIXED_SLOT policy: |
| // +------------------------------------------------+ |
| // | slot_index | 0 | virtual_register | 001 | |
| // +------------------------------------------------+ |
| // |
| // For all other (extended) policies: |
| // +-----------------------------------------------------+ |
| // | reg_index | L | PPP | 1 | virtual_register | 001 | |
| // +-----------------------------------------------------+ |
| // L ... Lifetime |
| // P ... Policy |
| // |
| // The slot index is a signed value which requires us to decode it manually |
| // instead of using the BitField utility class. |
| |
| STATIC_ASSERT(KindField::kSize == 3); |
| |
| class VirtualRegisterField : public BitField64<uint32_t, 3, 32> {}; |
| |
| // BitFields for all unallocated operands. |
| class BasicPolicyField : public BitField64<BasicPolicy, 35, 1> {}; |
| |
| // BitFields specific to BasicPolicy::FIXED_SLOT. |
| class FixedSlotIndexField : public BitField64<int, 36, 28> {}; |
| |
| // BitFields specific to BasicPolicy::EXTENDED_POLICY. |
| class ExtendedPolicyField : public BitField64<ExtendedPolicy, 36, 3> {}; |
| class LifetimeField : public BitField64<Lifetime, 39, 1> {}; |
| class HasSecondaryStorageField : public BitField64<bool, 40, 1> {}; |
| class FixedRegisterField : public BitField64<int, 41, 6> {}; |
| class SecondaryStorageField : public BitField64<int, 47, 3> {}; |
| |
| private: |
| explicit UnallocatedOperand(int virtual_register) |
| : InstructionOperand(UNALLOCATED) { |
| value_ |= |
| VirtualRegisterField::encode(static_cast<uint32_t>(virtual_register)); |
| } |
| }; |
| |
| class ConstantOperand : public InstructionOperand { |
| public: |
| explicit ConstantOperand(int virtual_register) |
| : InstructionOperand(CONSTANT) { |
| value_ |= |
| VirtualRegisterField::encode(static_cast<uint32_t>(virtual_register)); |
| } |
| |
| int32_t virtual_register() const { |
| return static_cast<int32_t>(VirtualRegisterField::decode(value_)); |
| } |
| |
| static ConstantOperand* New(Zone* zone, int virtual_register) { |
| return InstructionOperand::New(zone, ConstantOperand(virtual_register)); |
| } |
| |
| INSTRUCTION_OPERAND_CASTS(ConstantOperand, CONSTANT) |
| |
| STATIC_ASSERT(KindField::kSize == 3); |
| class VirtualRegisterField : public BitField64<uint32_t, 3, 32> {}; |
| }; |
| |
| class ImmediateOperand : public InstructionOperand { |
| public: |
| enum ImmediateType { INLINE, INDEXED }; |
| |
| explicit ImmediateOperand(ImmediateType type, int32_t value) |
| : InstructionOperand(IMMEDIATE) { |
| value_ |= TypeField::encode(type); |
| value_ |= static_cast<uint64_t>(static_cast<int64_t>(value)) |
| << ValueField::kShift; |
| } |
| |
| ImmediateType type() const { return TypeField::decode(value_); } |
| |
| int32_t inline_value() const { |
| DCHECK_EQ(INLINE, type()); |
| return static_cast<int64_t>(value_) >> ValueField::kShift; |
| } |
| |
| int32_t indexed_value() const { |
| DCHECK_EQ(INDEXED, type()); |
| return static_cast<int64_t>(value_) >> ValueField::kShift; |
| } |
| |
| static ImmediateOperand* New(Zone* zone, ImmediateType type, int32_t value) { |
| return InstructionOperand::New(zone, ImmediateOperand(type, value)); |
| } |
| |
| INSTRUCTION_OPERAND_CASTS(ImmediateOperand, IMMEDIATE) |
| |
| STATIC_ASSERT(KindField::kSize == 3); |
| class TypeField : public BitField64<ImmediateType, 3, 1> {}; |
| class ValueField : public BitField64<int32_t, 32, 32> {}; |
| }; |
| |
| class LocationOperand : public InstructionOperand { |
| public: |
| enum LocationKind { REGISTER, STACK_SLOT }; |
| |
| LocationOperand(InstructionOperand::Kind operand_kind, |
| LocationOperand::LocationKind location_kind, |
| MachineRepresentation rep, int index) |
| : InstructionOperand(operand_kind) { |
| DCHECK_IMPLIES(location_kind == REGISTER, index >= 0); |
| DCHECK(IsSupportedRepresentation(rep)); |
| value_ |= LocationKindField::encode(location_kind); |
| value_ |= RepresentationField::encode(rep); |
| value_ |= static_cast<uint64_t>(static_cast<int64_t>(index)) |
| << IndexField::kShift; |
| } |
| |
| int index() const { |
| DCHECK(IsStackSlot() || IsFPStackSlot()); |
| return static_cast<int64_t>(value_) >> IndexField::kShift; |
| } |
| |
| int register_code() const { |
| DCHECK(IsRegister() || IsFPRegister()); |
| return static_cast<int64_t>(value_) >> IndexField::kShift; |
| } |
| |
| Register GetRegister() const { |
| DCHECK(IsRegister()); |
| return Register::from_code(register_code()); |
| } |
| |
| FloatRegister GetFloatRegister() const { |
| DCHECK(IsFloatRegister()); |
| return FloatRegister::from_code(register_code()); |
| } |
| |
| DoubleRegister GetDoubleRegister() const { |
| // On platforms where FloatRegister, DoubleRegister, and Simd128Register |
| // are all the same type, it's convenient to treat everything as a |
| // DoubleRegister, so be lax about type checking here. |
| DCHECK(IsFPRegister()); |
| return DoubleRegister::from_code(register_code()); |
| } |
| |
| Simd128Register GetSimd128Register() const { |
| DCHECK(IsSimd128Register()); |
| return Simd128Register::from_code(register_code()); |
| } |
| |
| LocationKind location_kind() const { |
| return LocationKindField::decode(value_); |
| } |
| |
| MachineRepresentation representation() const { |
| return RepresentationField::decode(value_); |
| } |
| |
| static bool IsSupportedRepresentation(MachineRepresentation rep) { |
| switch (rep) { |
| case MachineRepresentation::kWord32: |
| case MachineRepresentation::kWord64: |
| case MachineRepresentation::kFloat32: |
| case MachineRepresentation::kFloat64: |
| case MachineRepresentation::kSimd128: |
| case MachineRepresentation::kTaggedSigned: |
| case MachineRepresentation::kTaggedPointer: |
| case MachineRepresentation::kTagged: |
| case MachineRepresentation::kCompressedSigned: |
| case MachineRepresentation::kCompressedPointer: |
| case MachineRepresentation::kCompressed: |
| return true; |
| case MachineRepresentation::kBit: |
| case MachineRepresentation::kWord8: |
| case MachineRepresentation::kWord16: |
| case MachineRepresentation::kNone: |
| return false; |
| } |
| UNREACHABLE(); |
| } |
| |
| // Return true if the locations can be moved to one another. |
| bool IsCompatible(LocationOperand* op); |
| |
| static LocationOperand* cast(InstructionOperand* op) { |
| DCHECK(op->IsAnyLocationOperand()); |
| return static_cast<LocationOperand*>(op); |
| } |
| |
| static const LocationOperand* cast(const InstructionOperand* op) { |
| DCHECK(op->IsAnyLocationOperand()); |
| return static_cast<const LocationOperand*>(op); |
| } |
| |
| static LocationOperand cast(const InstructionOperand& op) { |
| DCHECK(op.IsAnyLocationOperand()); |
| return *static_cast<const LocationOperand*>(&op); |
| } |
| |
| STATIC_ASSERT(KindField::kSize == 3); |
| class LocationKindField : public BitField64<LocationKind, 3, 2> {}; |
| class RepresentationField : public BitField64<MachineRepresentation, 5, 8> {}; |
| class IndexField : public BitField64<int32_t, 35, 29> {}; |
| }; |
| |
| class V8_EXPORT_PRIVATE ExplicitOperand |
| : public NON_EXPORTED_BASE(LocationOperand) { |
| public: |
| ExplicitOperand(LocationKind kind, MachineRepresentation rep, int index); |
| |
| static ExplicitOperand* New(Zone* zone, LocationKind kind, |
| MachineRepresentation rep, int index) { |
| return InstructionOperand::New(zone, ExplicitOperand(kind, rep, index)); |
| } |
| |
| INSTRUCTION_OPERAND_CASTS(ExplicitOperand, EXPLICIT) |
| }; |
| |
| class AllocatedOperand : public LocationOperand { |
| public: |
| AllocatedOperand(LocationKind kind, MachineRepresentation rep, int index) |
| : LocationOperand(ALLOCATED, kind, rep, index) {} |
| |
| static AllocatedOperand* New(Zone* zone, LocationKind kind, |
| MachineRepresentation rep, int index) { |
| return InstructionOperand::New(zone, AllocatedOperand(kind, rep, index)); |
| } |
| |
| INSTRUCTION_OPERAND_CASTS(AllocatedOperand, ALLOCATED) |
| }; |
| |
| #undef INSTRUCTION_OPERAND_CASTS |
| |
| bool InstructionOperand::IsAnyLocationOperand() const { |
| return this->kind() >= FIRST_LOCATION_OPERAND_KIND; |
| } |
| |
| bool InstructionOperand::IsLocationOperand() const { |
| return IsAnyLocationOperand() && |
| !IsFloatingPoint(LocationOperand::cast(this)->representation()); |
| } |
| |
| bool InstructionOperand::IsFPLocationOperand() const { |
| return IsAnyLocationOperand() && |
| IsFloatingPoint(LocationOperand::cast(this)->representation()); |
| } |
| |
| bool InstructionOperand::IsAnyRegister() const { |
| return IsAnyLocationOperand() && |
| LocationOperand::cast(this)->location_kind() == |
| LocationOperand::REGISTER; |
| } |
| |
| bool InstructionOperand::IsRegister() const { |
| return IsAnyRegister() && |
| !IsFloatingPoint(LocationOperand::cast(this)->representation()); |
| } |
| |
| bool InstructionOperand::IsFPRegister() const { |
| return IsAnyRegister() && |
| IsFloatingPoint(LocationOperand::cast(this)->representation()); |
| } |
| |
| bool InstructionOperand::IsFloatRegister() const { |
| return IsAnyRegister() && LocationOperand::cast(this)->representation() == |
| MachineRepresentation::kFloat32; |
| } |
| |
| bool InstructionOperand::IsDoubleRegister() const { |
| return IsAnyRegister() && LocationOperand::cast(this)->representation() == |
| MachineRepresentation::kFloat64; |
| } |
| |
| bool InstructionOperand::IsSimd128Register() const { |
| return IsAnyRegister() && LocationOperand::cast(this)->representation() == |
| MachineRepresentation::kSimd128; |
| } |
| |
| bool InstructionOperand::IsAnyStackSlot() const { |
| return IsAnyLocationOperand() && |
| LocationOperand::cast(this)->location_kind() == |
| LocationOperand::STACK_SLOT; |
| } |
| |
| bool InstructionOperand::IsStackSlot() const { |
| return IsAnyStackSlot() && |
| !IsFloatingPoint(LocationOperand::cast(this)->representation()); |
| } |
| |
| bool InstructionOperand::IsFPStackSlot() const { |
| return IsAnyStackSlot() && |
| IsFloatingPoint(LocationOperand::cast(this)->representation()); |
| } |
| |
| bool InstructionOperand::IsFloatStackSlot() const { |
| return IsAnyLocationOperand() && |
| LocationOperand::cast(this)->location_kind() == |
| LocationOperand::STACK_SLOT && |
| LocationOperand::cast(this)->representation() == |
| MachineRepresentation::kFloat32; |
| } |
| |
| bool InstructionOperand::IsDoubleStackSlot() const { |
| return IsAnyLocationOperand() && |
| LocationOperand::cast(this)->location_kind() == |
| LocationOperand::STACK_SLOT && |
| LocationOperand::cast(this)->representation() == |
| MachineRepresentation::kFloat64; |
| } |
| |
| bool InstructionOperand::IsSimd128StackSlot() const { |
| return IsAnyLocationOperand() && |
| LocationOperand::cast(this)->location_kind() == |
| LocationOperand::STACK_SLOT && |
| LocationOperand::cast(this)->representation() == |
| MachineRepresentation::kSimd128; |
| } |
| |
| uint64_t InstructionOperand::GetCanonicalizedValue() const { |
| if (IsAnyLocationOperand()) { |
| MachineRepresentation canonical = MachineRepresentation::kNone; |
| if (IsFPRegister()) { |
| if (kSimpleFPAliasing) { |
| // We treat all FP register operands the same for simple aliasing. |
| canonical = MachineRepresentation::kFloat64; |
| } else { |
| // We need to distinguish FP register operands of different reps when |
| // aliasing is not simple (e.g. ARM). |
| canonical = LocationOperand::cast(this)->representation(); |
| } |
| } |
| return InstructionOperand::KindField::update( |
| LocationOperand::RepresentationField::update(this->value_, canonical), |
| LocationOperand::EXPLICIT); |
| } |
| return this->value_; |
| } |
| |
| // Required for maps that don't care about machine type. |
| struct CompareOperandModuloType { |
| bool operator()(const InstructionOperand& a, |
| const InstructionOperand& b) const { |
| return a.CompareCanonicalized(b); |
| } |
| }; |
| |
| class V8_EXPORT_PRIVATE MoveOperands final |
| : public NON_EXPORTED_BASE(ZoneObject) { |
| public: |
| MoveOperands(const InstructionOperand& source, |
| const InstructionOperand& destination) |
| : source_(source), destination_(destination) { |
| DCHECK(!source.IsInvalid() && !destination.IsInvalid()); |
| } |
| |
| const InstructionOperand& source() const { return source_; } |
| InstructionOperand& source() { return source_; } |
| void set_source(const InstructionOperand& operand) { source_ = operand; } |
| |
| const InstructionOperand& destination() const { return destination_; } |
| InstructionOperand& destination() { return destination_; } |
| void set_destination(const InstructionOperand& operand) { |
| destination_ = operand; |
| } |
| |
| // The gap resolver marks moves as "in-progress" by clearing the |
| // destination (but not the source). |
| bool IsPending() const { |
| return destination_.IsInvalid() && !source_.IsInvalid(); |
| } |
| void SetPending() { destination_ = InstructionOperand(); } |
| |
| // A move is redundant if it's been eliminated or if its source and |
| // destination are the same. |
| bool IsRedundant() const { |
| DCHECK_IMPLIES(!destination_.IsInvalid(), !destination_.IsConstant()); |
| return IsEliminated() || source_.EqualsCanonicalized(destination_); |
| } |
| |
| // We clear both operands to indicate move that's been eliminated. |
| void Eliminate() { source_ = destination_ = InstructionOperand(); } |
| bool IsEliminated() const { |
| DCHECK_IMPLIES(source_.IsInvalid(), destination_.IsInvalid()); |
| return source_.IsInvalid(); |
| } |
| |
| // APIs to aid debugging. For general-stream APIs, use operator<<. |
| void Print() const; |
| |
| private: |
| InstructionOperand source_; |
| InstructionOperand destination_; |
| |
| DISALLOW_COPY_AND_ASSIGN(MoveOperands); |
| }; |
| |
| V8_EXPORT_PRIVATE std::ostream& operator<<(std::ostream&, const MoveOperands&); |
| |
| class V8_EXPORT_PRIVATE ParallelMove final |
| : public NON_EXPORTED_BASE(ZoneVector<MoveOperands*>), |
| public NON_EXPORTED_BASE(ZoneObject) { |
| public: |
| explicit ParallelMove(Zone* zone) : ZoneVector<MoveOperands*>(zone) {} |
| |
| MoveOperands* AddMove(const InstructionOperand& from, |
| const InstructionOperand& to) { |
| Zone* zone = get_allocator().zone(); |
| return AddMove(from, to, zone); |
| } |
| |
| MoveOperands* AddMove(const InstructionOperand& from, |
| const InstructionOperand& to, |
| Zone* operand_allocation_zone) { |
| if (from.EqualsCanonicalized(to)) return nullptr; |
| MoveOperands* move = new (operand_allocation_zone) MoveOperands(from, to); |
| if (empty()) reserve(4); |
| push_back(move); |
| return move; |
| } |
| |
| bool IsRedundant() const; |
| |
| // Prepare this ParallelMove to insert move as if it happened in a subsequent |
| // ParallelMove. move->source() may be changed. Any MoveOperands added to |
| // to_eliminate must be Eliminated. |
| void PrepareInsertAfter(MoveOperands* move, |
| ZoneVector<MoveOperands*>* to_eliminate) const; |
| |
| private: |
| DISALLOW_COPY_AND_ASSIGN(ParallelMove); |
| }; |
| |
| std::ostream& operator<<(std::ostream&, const ParallelMove&); |
| |
| class ReferenceMap final : public ZoneObject { |
| public: |
| explicit ReferenceMap(Zone* zone) |
| : reference_operands_(8, zone), instruction_position_(-1) {} |
| |
| const ZoneVector<InstructionOperand>& reference_operands() const { |
| return reference_operands_; |
| } |
| int instruction_position() const { return instruction_position_; } |
| |
| void set_instruction_position(int pos) { |
| DCHECK_EQ(-1, instruction_position_); |
| instruction_position_ = pos; |
| } |
| |
| void RecordReference(const AllocatedOperand& op); |
| |
| private: |
| friend std::ostream& operator<<(std::ostream&, const ReferenceMap&); |
| |
| ZoneVector<InstructionOperand> reference_operands_; |
| int instruction_position_; |
| }; |
| |
| std::ostream& operator<<(std::ostream&, const ReferenceMap&); |
| |
| class InstructionBlock; |
| |
| class V8_EXPORT_PRIVATE Instruction final { |
| public: |
| size_t OutputCount() const { return OutputCountField::decode(bit_field_); } |
| const InstructionOperand* OutputAt(size_t i) const { |
| DCHECK(i < OutputCount()); |
| return &operands_[i]; |
| } |
| InstructionOperand* OutputAt(size_t i) { |
| DCHECK(i < OutputCount()); |
| return &operands_[i]; |
| } |
| |
| bool HasOutput() const { return OutputCount() > 0; } |
| const InstructionOperand* Output() const { return OutputAt(0); } |
| InstructionOperand* Output() { return OutputAt(0); } |
| |
| size_t InputCount() const { return InputCountField::decode(bit_field_); } |
| const InstructionOperand* InputAt(size_t i) const { |
| DCHECK(i < InputCount()); |
| return &operands_[OutputCount() + i]; |
| } |
| InstructionOperand* InputAt(size_t i) { |
| DCHECK(i < InputCount()); |
| return &operands_[OutputCount() + i]; |
| } |
| |
| size_t TempCount() const { return TempCountField::decode(bit_field_); } |
| const InstructionOperand* TempAt(size_t i) const { |
| DCHECK(i < TempCount()); |
| return &operands_[OutputCount() + InputCount() + i]; |
| } |
| InstructionOperand* TempAt(size_t i) { |
| DCHECK(i < TempCount()); |
| return &operands_[OutputCount() + InputCount() + i]; |
| } |
| |
| InstructionCode opcode() const { return opcode_; } |
| ArchOpcode arch_opcode() const { return ArchOpcodeField::decode(opcode()); } |
| AddressingMode addressing_mode() const { |
| return AddressingModeField::decode(opcode()); |
| } |
| FlagsMode flags_mode() const { return FlagsModeField::decode(opcode()); } |
| FlagsCondition flags_condition() const { |
| return FlagsConditionField::decode(opcode()); |
| } |
| |
| static Instruction* New(Zone* zone, InstructionCode opcode) { |
| return New(zone, opcode, 0, nullptr, 0, nullptr, 0, nullptr); |
| } |
| |
| static Instruction* New(Zone* zone, InstructionCode opcode, |
| size_t output_count, InstructionOperand* outputs, |
| size_t input_count, InstructionOperand* inputs, |
| size_t temp_count, InstructionOperand* temps) { |
| DCHECK_LE(0, opcode); |
| DCHECK(output_count == 0 || outputs != nullptr); |
| DCHECK(input_count == 0 || inputs != nullptr); |
| DCHECK(temp_count == 0 || temps != nullptr); |
| // TODO(jarin/mstarzinger): Handle this gracefully. See crbug.com/582702. |
| CHECK(InputCountField::is_valid(input_count)); |
| |
| size_t total_extra_ops = output_count + input_count + temp_count; |
| if (total_extra_ops != 0) total_extra_ops--; |
| int size = static_cast<int>( |
| RoundUp(sizeof(Instruction), sizeof(InstructionOperand)) + |
| total_extra_ops * sizeof(InstructionOperand)); |
| return new (zone->New(size)) Instruction( |
| opcode, output_count, outputs, input_count, inputs, temp_count, temps); |
| } |
| |
| Instruction* MarkAsCall() { |
| bit_field_ = IsCallField::update(bit_field_, true); |
| return this; |
| } |
| bool IsCall() const { return IsCallField::decode(bit_field_); } |
| bool NeedsReferenceMap() const { return IsCall(); } |
| bool HasReferenceMap() const { return reference_map_ != nullptr; } |
| |
| bool ClobbersRegisters() const { return IsCall(); } |
| bool ClobbersTemps() const { return IsCall(); } |
| bool ClobbersDoubleRegisters() const { return IsCall(); } |
| ReferenceMap* reference_map() const { return reference_map_; } |
| |
| void set_reference_map(ReferenceMap* map) { |
| DCHECK(NeedsReferenceMap()); |
| DCHECK(!reference_map_); |
| reference_map_ = map; |
| } |
| |
| void OverwriteWithNop() { |
| opcode_ = ArchOpcodeField::encode(kArchNop); |
| bit_field_ = 0; |
| reference_map_ = nullptr; |
| } |
| |
| bool IsNop() const { return arch_opcode() == kArchNop; } |
| |
| bool IsDeoptimizeCall() const { |
| return arch_opcode() == ArchOpcode::kArchDeoptimize || |
| FlagsModeField::decode(opcode()) == kFlags_deoptimize || |
| FlagsModeField::decode(opcode()) == kFlags_deoptimize_and_poison; |
| } |
| |
| bool IsTrap() const { |
| return FlagsModeField::decode(opcode()) == kFlags_trap; |
| } |
| |
| bool IsJump() const { return arch_opcode() == ArchOpcode::kArchJmp; } |
| bool IsRet() const { return arch_opcode() == ArchOpcode::kArchRet; } |
| bool IsTailCall() const { |
| return arch_opcode() == ArchOpcode::kArchTailCallCodeObject || |
| arch_opcode() == ArchOpcode::kArchTailCallCodeObjectFromJSFunction || |
| arch_opcode() == ArchOpcode::kArchTailCallAddress || |
| arch_opcode() == ArchOpcode::kArchTailCallWasm; |
| } |
| bool IsThrow() const { |
| return arch_opcode() == ArchOpcode::kArchThrowTerminator; |
| } |
| |
| enum GapPosition { |
| START, |
| END, |
| FIRST_GAP_POSITION = START, |
| LAST_GAP_POSITION = END |
| }; |
| |
| ParallelMove* GetOrCreateParallelMove(GapPosition pos, Zone* zone) { |
| if (parallel_moves_[pos] == nullptr) { |
| parallel_moves_[pos] = new (zone) ParallelMove(zone); |
| } |
| return parallel_moves_[pos]; |
| } |
| |
| ParallelMove* GetParallelMove(GapPosition pos) { |
| return parallel_moves_[pos]; |
| } |
| |
| const ParallelMove* GetParallelMove(GapPosition pos) const { |
| return parallel_moves_[pos]; |
| } |
| |
| bool AreMovesRedundant() const; |
| |
| ParallelMove* const* parallel_moves() const { return ¶llel_moves_[0]; } |
| ParallelMove** parallel_moves() { return ¶llel_moves_[0]; } |
| |
| // The block_id may be invalidated in JumpThreading. It is only important for |
| // register allocation, to avoid searching for blocks from instruction |
| // indexes. |
| InstructionBlock* block() const { return block_; } |
| void set_block(InstructionBlock* block) { |
| DCHECK_NOT_NULL(block); |
| block_ = block; |
| } |
| |
| // APIs to aid debugging. For general-stream APIs, use operator<<. |
| void Print() const; |
| |
| using OutputCountField = BitField<size_t, 0, 8>; |
| using InputCountField = BitField<size_t, 8, 16>; |
| using TempCountField = BitField<size_t, 24, 6>; |
| |
| static const size_t kMaxOutputCount = OutputCountField::kMax; |
| static const size_t kMaxInputCount = InputCountField::kMax; |
| static const size_t kMaxTempCount = TempCountField::kMax; |
| |
| private: |
| explicit Instruction(InstructionCode opcode); |
| |
| Instruction(InstructionCode opcode, size_t output_count, |
| InstructionOperand* outputs, size_t input_count, |
| InstructionOperand* inputs, size_t temp_count, |
| InstructionOperand* temps); |
| |
| using IsCallField = BitField<bool, 30, 1>; |
| |
| InstructionCode opcode_; |
| uint32_t bit_field_; |
| ParallelMove* parallel_moves_[2]; |
| ReferenceMap* reference_map_; |
| InstructionBlock* block_; |
| InstructionOperand operands_[1]; |
| |
| DISALLOW_COPY_AND_ASSIGN(Instruction); |
| }; |
| |
| std::ostream& operator<<(std::ostream&, const Instruction&); |
| |
| class RpoNumber final { |
| public: |
| static const int kInvalidRpoNumber = -1; |
| int ToInt() const { |
| DCHECK(IsValid()); |
| return index_; |
| } |
| size_t ToSize() const { |
| DCHECK(IsValid()); |
| return static_cast<size_t>(index_); |
| } |
| bool IsValid() const { return index_ >= 0; } |
| static RpoNumber FromInt(int index) { return RpoNumber(index); } |
| static RpoNumber Invalid() { return RpoNumber(kInvalidRpoNumber); } |
| |
| bool IsNext(const RpoNumber other) const { |
| DCHECK(IsValid()); |
| return other.index_ == this->index_ + 1; |
| } |
| |
| RpoNumber Next() const { |
| DCHECK(IsValid()); |
| return RpoNumber(index_ + 1); |
| } |
| |
| // Comparison operators. |
| bool operator==(RpoNumber other) const { return index_ == other.index_; } |
| bool operator!=(RpoNumber other) const { return index_ != other.index_; } |
| bool operator>(RpoNumber other) const { return index_ > other.index_; } |
| bool operator<(RpoNumber other) const { return index_ < other.index_; } |
| bool operator<=(RpoNumber other) const { return index_ <= other.index_; } |
| bool operator>=(RpoNumber other) const { return index_ >= other.index_; } |
| |
| private: |
| explicit RpoNumber(int32_t index) : index_(index) {} |
| int32_t index_; |
| }; |
| |
| std::ostream& operator<<(std::ostream&, const RpoNumber&); |
| |
| class V8_EXPORT_PRIVATE Constant final { |
| public: |
| enum Type { |
| kInt32, |
| kInt64, |
| kFloat32, |
| kFloat64, |
| kExternalReference, |
| kCompressedHeapObject, |
| kHeapObject, |
| kRpoNumber, |
| kDelayedStringConstant |
| }; |
| |
| explicit Constant(int32_t v); |
| explicit Constant(int64_t v) : type_(kInt64), value_(v) {} |
| explicit Constant(float v) : type_(kFloat32), value_(bit_cast<int32_t>(v)) {} |
| explicit Constant(double v) : type_(kFloat64), value_(bit_cast<int64_t>(v)) {} |
| explicit Constant(ExternalReference ref) |
| : type_(kExternalReference), value_(bit_cast<intptr_t>(ref.address())) {} |
| explicit Constant(Handle<HeapObject> obj, bool is_compressed = false) |
| : type_(is_compressed ? kCompressedHeapObject : kHeapObject), |
| value_(bit_cast<intptr_t>(obj)) {} |
| explicit Constant(RpoNumber rpo) : type_(kRpoNumber), value_(rpo.ToInt()) {} |
| explicit Constant(const StringConstantBase* str) |
| : type_(kDelayedStringConstant), value_(bit_cast<intptr_t>(str)) {} |
| explicit Constant(RelocatablePtrConstantInfo info); |
| |
| Type type() const { return type_; } |
| |
| RelocInfo::Mode rmode() const { return rmode_; } |
| |
| int32_t ToInt32() const { |
| DCHECK(type() == kInt32 || type() == kInt64); |
| const int32_t value = static_cast<int32_t>(value_); |
| DCHECK_EQ(value_, static_cast<int64_t>(value)); |
| return value; |
| } |
| |
| int64_t ToInt64() const { |
| if (type() == kInt32) return ToInt32(); |
| DCHECK_EQ(kInt64, type()); |
| return value_; |
| } |
| |
| float ToFloat32() const { |
| // TODO(ahaas): We should remove this function. If value_ has the bit |
| // representation of a signalling NaN, then returning it as float can cause |
| // the signalling bit to flip, and value_ is returned as a quiet NaN. |
| DCHECK_EQ(kFloat32, type()); |
| return bit_cast<float>(static_cast<int32_t>(value_)); |
| } |
| |
| uint32_t ToFloat32AsInt() const { |
| DCHECK_EQ(kFloat32, type()); |
| return bit_cast<uint32_t>(static_cast<int32_t>(value_)); |
| } |
| |
| Double ToFloat64() const { |
| DCHECK_EQ(kFloat64, type()); |
| return Double(bit_cast<uint64_t>(value_)); |
| } |
| |
| ExternalReference ToExternalReference() const { |
| DCHECK_EQ(kExternalReference, type()); |
| return ExternalReference::FromRawAddress(static_cast<Address>(value_)); |
| } |
| |
| RpoNumber ToRpoNumber() const { |
| DCHECK_EQ(kRpoNumber, type()); |
| return RpoNumber::FromInt(static_cast<int>(value_)); |
| } |
| |
| Handle<HeapObject> ToHeapObject() const; |
| Handle<Code> ToCode() const; |
| const StringConstantBase* ToDelayedStringConstant() const; |
| |
| private: |
| Type type_; |
| RelocInfo::Mode rmode_ = RelocInfo::NONE; |
| int64_t value_; |
| }; |
| |
| std::ostream& operator<<(std::ostream&, const Constant&); |
| |
| // Forward declarations. |
| class FrameStateDescriptor; |
| |
| enum class StateValueKind : uint8_t { |
| kArgumentsElements, |
| kArgumentsLength, |
| kPlain, |
| kOptimizedOut, |
| kNested, |
| kDuplicate |
| }; |
| |
| class StateValueDescriptor { |
| public: |
| StateValueDescriptor() |
| : kind_(StateValueKind::kPlain), type_(MachineType::AnyTagged()) {} |
| |
| static StateValueDescriptor ArgumentsElements(ArgumentsStateType type) { |
| StateValueDescriptor descr(StateValueKind::kArgumentsElements, |
| MachineType::AnyTagged()); |
| descr.args_type_ = type; |
| return descr; |
| } |
| static StateValueDescriptor ArgumentsLength(ArgumentsStateType type) { |
| StateValueDescriptor descr(StateValueKind::kArgumentsLength, |
| MachineType::AnyTagged()); |
| descr.args_type_ = type; |
| return descr; |
| } |
| static StateValueDescriptor Plain(MachineType type) { |
| return StateValueDescriptor(StateValueKind::kPlain, type); |
| } |
| static StateValueDescriptor OptimizedOut() { |
| return StateValueDescriptor(StateValueKind::kOptimizedOut, |
| MachineType::AnyTagged()); |
| } |
| static StateValueDescriptor Recursive(size_t id) { |
| StateValueDescriptor descr(StateValueKind::kNested, |
| MachineType::AnyTagged()); |
| descr.id_ = id; |
| return descr; |
| } |
| static StateValueDescriptor Duplicate(size_t id) { |
| StateValueDescriptor descr(StateValueKind::kDuplicate, |
| MachineType::AnyTagged()); |
| descr.id_ = id; |
| return descr; |
| } |
| |
| bool IsArgumentsElements() const { |
| return kind_ == StateValueKind::kArgumentsElements; |
| } |
| bool IsArgumentsLength() const { |
| return kind_ == StateValueKind::kArgumentsLength; |
| } |
| bool IsPlain() const { return kind_ == StateValueKind::kPlain; } |
| bool IsOptimizedOut() const { return kind_ == StateValueKind::kOptimizedOut; } |
| bool IsNested() const { return kind_ == StateValueKind::kNested; } |
| bool IsDuplicate() const { return kind_ == StateValueKind::kDuplicate; } |
| MachineType type() const { return type_; } |
| size_t id() const { |
| DCHECK(kind_ == StateValueKind::kDuplicate || |
| kind_ == StateValueKind::kNested); |
| return id_; |
| } |
| ArgumentsStateType arguments_type() const { |
| DCHECK(kind_ == StateValueKind::kArgumentsElements || |
| kind_ == StateValueKind::kArgumentsLength); |
| return args_type_; |
| } |
| |
| private: |
| StateValueDescriptor(StateValueKind kind, MachineType type) |
| : kind_(kind), type_(type) {} |
| |
| StateValueKind kind_; |
| MachineType type_; |
| union { |
| size_t id_; |
| ArgumentsStateType args_type_; |
| }; |
| }; |
| |
| class StateValueList { |
| public: |
| explicit StateValueList(Zone* zone) : fields_(zone), nested_(zone) {} |
| |
| size_t size() { return fields_.size(); } |
| |
| struct Value { |
| StateValueDescriptor* desc; |
| StateValueList* nested; |
| |
| Value(StateValueDescriptor* desc, StateValueList* nested) |
| : desc(desc), nested(nested) {} |
| }; |
| |
| class iterator { |
| public: |
| // Bare minimum of operators needed for range iteration. |
| bool operator!=(const iterator& other) const { |
| return field_iterator != other.field_iterator; |
| } |
| bool operator==(const iterator& other) const { |
| return field_iterator == other.field_iterator; |
| } |
| iterator& operator++() { |
| if (field_iterator->IsNested()) { |
| nested_iterator++; |
| } |
| ++field_iterator; |
| return *this; |
| } |
| Value operator*() { |
| StateValueDescriptor* desc = &(*field_iterator); |
| StateValueList* nested = desc->IsNested() ? *nested_iterator : nullptr; |
| return Value(desc, nested); |
| } |
| |
| private: |
| friend class StateValueList; |
| |
| iterator(ZoneVector<StateValueDescriptor>::iterator it, |
| ZoneVector<StateValueList*>::iterator nested) |
| : field_iterator(it), nested_iterator(nested) {} |
| |
| ZoneVector<StateValueDescriptor>::iterator field_iterator; |
| ZoneVector<StateValueList*>::iterator nested_iterator; |
| }; |
| |
| void ReserveSize(size_t size) { fields_.reserve(size); } |
| |
| StateValueList* PushRecursiveField(Zone* zone, size_t id) { |
| fields_.push_back(StateValueDescriptor::Recursive(id)); |
| StateValueList* nested = |
| new (zone->New(sizeof(StateValueList))) StateValueList(zone); |
| nested_.push_back(nested); |
| return nested; |
| } |
| void PushArgumentsElements(ArgumentsStateType type) { |
| fields_.push_back(StateValueDescriptor::ArgumentsElements(type)); |
| } |
| void PushArgumentsLength(ArgumentsStateType type) { |
| fields_.push_back(StateValueDescriptor::ArgumentsLength(type)); |
| } |
| void PushDuplicate(size_t id) { |
| fields_.push_back(StateValueDescriptor::Duplicate(id)); |
| } |
| void PushPlain(MachineType type) { |
| fields_.push_back(StateValueDescriptor::Plain(type)); |
| } |
| void PushOptimizedOut() { |
| fields_.push_back(StateValueDescriptor::OptimizedOut()); |
| } |
| |
| iterator begin() { return iterator(fields_.begin(), nested_.begin()); } |
| iterator end() { return iterator(fields_.end(), nested_.end()); } |
| |
| private: |
| ZoneVector<StateValueDescriptor> fields_; |
| ZoneVector<StateValueList*> nested_; |
| }; |
| |
| class FrameStateDescriptor : public ZoneObject { |
| public: |
| FrameStateDescriptor(Zone* zone, FrameStateType type, BailoutId bailout_id, |
| OutputFrameStateCombine state_combine, |
| size_t parameters_count, size_t locals_count, |
| size_t stack_count, |
| MaybeHandle<SharedFunctionInfo> shared_info, |
| FrameStateDescriptor* outer_state = nullptr); |
| |
| FrameStateType type() const { return type_; } |
| BailoutId bailout_id() const { return bailout_id_; } |
| OutputFrameStateCombine state_combine() const { return frame_state_combine_; } |
| size_t parameters_count() const { return parameters_count_; } |
| size_t locals_count() const { return locals_count_; } |
| size_t stack_count() const { return stack_count_; } |
| MaybeHandle<SharedFunctionInfo> shared_info() const { return shared_info_; } |
| FrameStateDescriptor* outer_state() const { return outer_state_; } |
| bool HasContext() const { |
| return FrameStateFunctionInfo::IsJSFunctionType(type_) || |
| type_ == FrameStateType::kBuiltinContinuation || |
| type_ == FrameStateType::kConstructStub; |
| } |
| |
| size_t GetSize() const; |
| size_t GetTotalSize() const; |
| size_t GetFrameCount() const; |
| size_t GetJSFrameCount() const; |
| |
| StateValueList* GetStateValueDescriptors() { return &values_; } |
| |
| static const int kImpossibleValue = 0xdead; |
| |
| private: |
| FrameStateType type_; |
| BailoutId bailout_id_; |
| OutputFrameStateCombine frame_state_combine_; |
| size_t parameters_count_; |
| size_t locals_count_; |
| size_t stack_count_; |
| StateValueList values_; |
| MaybeHandle<SharedFunctionInfo> const shared_info_; |
| FrameStateDescriptor* outer_state_; |
| }; |
| |
| // A deoptimization entry is a pair of the reason why we deoptimize and the |
| // frame state descriptor that we have to go back to. |
| class DeoptimizationEntry final { |
| public: |
| DeoptimizationEntry() = default; |
| DeoptimizationEntry(FrameStateDescriptor* descriptor, DeoptimizeKind kind, |
| DeoptimizeReason reason, VectorSlotPair const& feedback) |
| : descriptor_(descriptor), |
| kind_(kind), |
| reason_(reason), |
| feedback_(feedback) {} |
| |
| FrameStateDescriptor* descriptor() const { return descriptor_; } |
| DeoptimizeKind kind() const { return kind_; } |
| DeoptimizeReason reason() const { return reason_; } |
| VectorSlotPair const& feedback() const { return feedback_; } |
| |
| private: |
| FrameStateDescriptor* descriptor_ = nullptr; |
| DeoptimizeKind kind_ = DeoptimizeKind::kEager; |
| DeoptimizeReason reason_ = DeoptimizeReason::kUnknown; |
| VectorSlotPair feedback_ = VectorSlotPair(); |
| }; |
| |
| using DeoptimizationVector = ZoneVector<DeoptimizationEntry>; |
| |
| class V8_EXPORT_PRIVATE PhiInstruction final |
| : public NON_EXPORTED_BASE(ZoneObject) { |
| public: |
| using Inputs = ZoneVector<InstructionOperand>; |
| |
| PhiInstruction(Zone* zone, int virtual_register, size_t input_count); |
| |
| void SetInput(size_t offset, int virtual_register); |
| void RenameInput(size_t offset, int virtual_register); |
| |
| int virtual_register() const { return virtual_register_; } |
| const IntVector& operands() const { return operands_; } |
| |
| // TODO(dcarney): this has no real business being here, since it's internal to |
| // the register allocator, but putting it here was convenient. |
| const InstructionOperand& output() const { return output_; } |
| InstructionOperand& output() { return output_; } |
| |
| private: |
| const int virtual_register_; |
| InstructionOperand output_; |
| IntVector operands_; |
| }; |
| |
| // Analogue of BasicBlock for Instructions instead of Nodes. |
| class V8_EXPORT_PRIVATE InstructionBlock final |
| : public NON_EXPORTED_BASE(ZoneObject) { |
| public: |
| InstructionBlock(Zone* zone, RpoNumber rpo_number, RpoNumber loop_header, |
| RpoNumber loop_end, bool deferred, bool handler); |
| |
| // Instruction indexes (used by the register allocator). |
| int first_instruction_index() const { |
| DCHECK_LE(0, code_start_); |
| DCHECK_LT(0, code_end_); |
| DCHECK_GE(code_end_, code_start_); |
| return code_start_; |
| } |
| int last_instruction_index() const { |
| DCHECK_LE(0, code_start_); |
| DCHECK_LT(0, code_end_); |
| DCHECK_GE(code_end_, code_start_); |
| return code_end_ - 1; |
| } |
| |
| int32_t code_start() const { return code_start_; } |
| void set_code_start(int32_t start) { code_start_ = start; } |
| |
| int32_t code_end() const { return code_end_; } |
| void set_code_end(int32_t end) { code_end_ = end; } |
| |
| bool IsDeferred() const { return deferred_; } |
| bool IsHandler() const { return handler_; } |
| |
| RpoNumber ao_number() const { return ao_number_; } |
| RpoNumber rpo_number() const { return rpo_number_; } |
| RpoNumber loop_header() const { return loop_header_; } |
| RpoNumber loop_end() const { |
| DCHECK(IsLoopHeader()); |
| return loop_end_; |
| } |
| inline bool IsLoopHeader() const { return loop_end_.IsValid(); } |
| inline bool IsSwitchTarget() const { return switch_target_; } |
| inline bool ShouldAlign() const { return alignment_; } |
| |
| using Predecessors = ZoneVector<RpoNumber>; |
| Predecessors& predecessors() { return predecessors_; } |
| const Predecessors& predecessors() const { return predecessors_; } |
| size_t PredecessorCount() const { return predecessors_.size(); } |
| size_t PredecessorIndexOf(RpoNumber rpo_number) const; |
| |
| using Successors = ZoneVector<RpoNumber>; |
| Successors& successors() { return successors_; } |
| const Successors& successors() const { return successors_; } |
| size_t SuccessorCount() const { return successors_.size(); } |
| |
| using PhiInstructions = ZoneVector<PhiInstruction*>; |
| const PhiInstructions& phis() const { return phis_; } |
| PhiInstruction* PhiAt(size_t i) const { return phis_[i]; } |
| void AddPhi(PhiInstruction* phi) { phis_.push_back(phi); } |
| |
| void set_ao_number(RpoNumber ao_number) { ao_number_ = ao_number; } |
| |
| void set_alignment(bool val) { alignment_ = val; } |
| |
| void set_switch_target(bool val) { switch_target_ = val; } |
| |
| bool needs_frame() const { return needs_frame_; } |
| void mark_needs_frame() { needs_frame_ = true; } |
| |
| bool must_construct_frame() const { return must_construct_frame_; } |
| void mark_must_construct_frame() { must_construct_frame_ = true; } |
| |
| bool must_deconstruct_frame() const { return must_deconstruct_frame_; } |
| void mark_must_deconstruct_frame() { must_deconstruct_frame_ = true; } |
| |
| private: |
| Successors successors_; |
| Predecessors predecessors_; |
| PhiInstructions phis_; |
| RpoNumber ao_number_; // Assembly order number. |
| const RpoNumber rpo_number_; |
| const RpoNumber loop_header_; |
| const RpoNumber loop_end_; |
| int32_t code_start_; // start index of arch-specific code. |
| int32_t code_end_ = -1; // end index of arch-specific code. |
| const bool deferred_ = 0; // Block contains deferred code. |
| const bool handler_; // Block is a handler entry point. |
| bool switch_target_ = false; |
| bool alignment_ = false; // insert alignment before this block |
| bool needs_frame_ = false; |
| bool must_construct_frame_ = false; |
| bool must_deconstruct_frame_ = false; |
| }; |
| |
| class InstructionSequence; |
| |
| struct PrintableInstructionBlock { |
| const InstructionBlock* block_; |
| const InstructionSequence* code_; |
| }; |
| |
| std::ostream& operator<<(std::ostream&, const PrintableInstructionBlock&); |
| |
| using ConstantDeque = ZoneDeque<Constant>; |
| using ConstantMap = std::map<int, Constant, std::less<int>, |
| ZoneAllocator<std::pair<const int, Constant> > >; |
| |
| using InstructionDeque = ZoneDeque<Instruction*>; |
| using ReferenceMapDeque = ZoneDeque<ReferenceMap*>; |
| using InstructionBlocks = ZoneVector<InstructionBlock*>; |
| |
| // Represents architecture-specific generated code before, during, and after |
| // register allocation. |
| class V8_EXPORT_PRIVATE InstructionSequence final |
| : public NON_EXPORTED_BASE(ZoneObject) { |
| public: |
| static InstructionBlocks* InstructionBlocksFor(Zone* zone, |
| const Schedule* schedule); |
| InstructionSequence(Isolate* isolate, Zone* zone, |
| InstructionBlocks* instruction_blocks); |
| |
| int NextVirtualRegister(); |
| int VirtualRegisterCount() const { return next_virtual_register_; } |
| |
| const InstructionBlocks& instruction_blocks() const { |
| return *instruction_blocks_; |
| } |
| |
| const InstructionBlocks& ao_blocks() const { return *ao_blocks_; } |
| |
| int InstructionBlockCount() const { |
| return static_cast<int>(instruction_blocks_->size()); |
| } |
| |
| InstructionBlock* InstructionBlockAt(RpoNumber rpo_number) { |
| return instruction_blocks_->at(rpo_number.ToSize()); |
| } |
| |
| int LastLoopInstructionIndex(const InstructionBlock* block) { |
| return instruction_blocks_->at(block->loop_end().ToSize() - 1) |
| ->last_instruction_index(); |
| } |
| |
| const InstructionBlock* InstructionBlockAt(RpoNumber rpo_number) const { |
| return instruction_blocks_->at(rpo_number.ToSize()); |
| } |
| |
| InstructionBlock* GetInstructionBlock(int instruction_index) const; |
| |
| static MachineRepresentation DefaultRepresentation() { |
| return MachineType::PointerRepresentation(); |
| } |
| MachineRepresentation GetRepresentation(int virtual_register) const; |
| void MarkAsRepresentation(MachineRepresentation rep, int virtual_register); |
| |
| bool IsReference(int virtual_register) const { |
| return CanBeTaggedOrCompressedPointer(GetRepresentation(virtual_register)); |
| } |
| bool IsFP(int virtual_register) const { |
| return IsFloatingPoint(GetRepresentation(virtual_register)); |
| } |
| int representation_mask() const { return representation_mask_; } |
| bool HasFPVirtualRegisters() const { |
| constexpr int kFPRepMask = |
| RepresentationBit(MachineRepresentation::kFloat32) | |
| RepresentationBit(MachineRepresentation::kFloat64) | |
| RepresentationBit(MachineRepresentation::kSimd128); |
| return (representation_mask() & kFPRepMask) != 0; |
| } |
| |
| Instruction* GetBlockStart(RpoNumber rpo) const; |
| |
| using const_iterator = InstructionDeque::const_iterator; |
| const_iterator begin() const { return instructions_.begin(); } |
| const_iterator end() const { return instructions_.end(); } |
| const InstructionDeque& instructions() const { return instructions_; } |
| int LastInstructionIndex() const { |
| return static_cast<int>(instructions().size()) - 1; |
| } |
| |
| Instruction* InstructionAt(int index) const { |
| DCHECK_LE(0, index); |
| DCHECK_GT(instructions_.size(), index); |
| return instructions_[index]; |
| } |
| |
| Isolate* isolate() const { return isolate_; } |
| const ReferenceMapDeque* reference_maps() const { return &reference_maps_; } |
| Zone* zone() const { return zone_; } |
| |
| // Used by the instruction selector while adding instructions. |
| int AddInstruction(Instruction* instr); |
| void StartBlock(RpoNumber rpo); |
| void EndBlock(RpoNumber rpo); |
| |
| int AddConstant(int virtual_register, Constant constant) { |
| // TODO(titzer): allow RPO numbers as constants? |
| DCHECK_NE(Constant::kRpoNumber, constant.type()); |
| DCHECK(virtual_register >= 0 && virtual_register < next_virtual_register_); |
| DCHECK(constants_.find(virtual_register) == constants_.end()); |
| constants_.insert(std::make_pair(virtual_register, constant)); |
| return virtual_register; |
| } |
| Constant GetConstant(int virtual_register) const { |
| ConstantMap::const_iterator it = constants_.find(virtual_register); |
| DCHECK(it != constants_.end()); |
| DCHECK_EQ(virtual_register, it->first); |
| return it->second; |
| } |
| |
| using Immediates = ZoneVector<Constant>; |
| Immediates& immediates() { return immediates_; } |
| |
| ImmediateOperand AddImmediate(const Constant& constant) { |
| if (constant.type() == Constant::kInt32 && |
| RelocInfo::IsNone(constant.rmode())) { |
| return ImmediateOperand(ImmediateOperand::INLINE, constant.ToInt32()); |
| } |
| int index = static_cast<int>(immediates_.size()); |
| immediates_.push_back(constant); |
| return ImmediateOperand(ImmediateOperand::INDEXED, index); |
| } |
| |
| Constant GetImmediate(const ImmediateOperand* op) const { |
| switch (op->type()) { |
| case ImmediateOperand::INLINE: |
| return Constant(op->inline_value()); |
| case ImmediateOperand::INDEXED: { |
| int index = op->indexed_value(); |
| DCHECK_LE(0, index); |
| DCHECK_GT(immediates_.size(), index); |
| return immediates_[index]; |
| } |
| } |
| UNREACHABLE(); |
| } |
| |
| int AddDeoptimizationEntry(FrameStateDescriptor* descriptor, |
| DeoptimizeKind kind, DeoptimizeReason reason, |
| VectorSlotPair const& feedback); |
| DeoptimizationEntry const& GetDeoptimizationEntry(int deoptimization_id); |
| int GetDeoptimizationEntryCount() const { |
| return static_cast<int>(deoptimization_entries_.size()); |
| } |
| |
| RpoNumber InputRpo(Instruction* instr, size_t index); |
| |
| bool GetSourcePosition(const Instruction* instr, |
| SourcePosition* result) const; |
| void SetSourcePosition(const Instruction* instr, SourcePosition value); |
| |
| bool ContainsCall() const { |
| for (Instruction* instr : instructions_) { |
| if (instr->IsCall()) return true; |
| } |
| return false; |
| } |
| |
| // APIs to aid debugging. For general-stream APIs, use operator<<. |
| void Print() const; |
| |
| void PrintBlock(int block_id) const; |
| |
| void ValidateEdgeSplitForm() const; |
| void ValidateDeferredBlockExitPaths() const; |
| void ValidateDeferredBlockEntryPaths() const; |
| void ValidateSSA() const; |
| |
| static void SetRegisterConfigurationForTesting( |
| const RegisterConfiguration* regConfig); |
| static void ClearRegisterConfigurationForTesting(); |
| |
| void RecomputeAssemblyOrderForTesting(); |
| |
| private: |
| friend V8_EXPORT_PRIVATE std::ostream& operator<<(std::ostream&, |
| const InstructionSequence&); |
| |
| using SourcePositionMap = ZoneMap<const Instruction*, SourcePosition>; |
| |
| static const RegisterConfiguration* RegisterConfigurationForTesting(); |
| static const RegisterConfiguration* registerConfigurationForTesting_; |
| |
| // Puts the deferred blocks last and may rotate loops. |
| void ComputeAssemblyOrder(); |
| |
| Isolate* isolate_; |
| Zone* const zone_; |
| InstructionBlocks* const instruction_blocks_; |
| InstructionBlocks* ao_blocks_; |
| SourcePositionMap source_positions_; |
| ConstantMap constants_; |
| Immediates immediates_; |
| InstructionDeque instructions_; |
| int next_virtual_register_; |
| ReferenceMapDeque reference_maps_; |
| ZoneVector<MachineRepresentation> representations_; |
| int representation_mask_; |
| DeoptimizationVector deoptimization_entries_; |
| |
| // Used at construction time |
| InstructionBlock* current_block_; |
| |
| DISALLOW_COPY_AND_ASSIGN(InstructionSequence); |
| }; |
| |
| V8_EXPORT_PRIVATE std::ostream& operator<<(std::ostream&, |
| const InstructionSequence&); |
| |
| } // namespace compiler |
| } // namespace internal |
| } // namespace v8 |
| |
| #endif // V8_COMPILER_BACKEND_INSTRUCTION_H_ |