blob: e1c7601174aa2f34226937b4d241c0e491f3a515 [file] [log] [blame]
/* -*- Mode: C++; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 4 -*-
* vim: set ts=8 sts=4 et sw=4 tw=99:
* This Source Code Form is subject to the terms of the Mozilla Public
* License, v. 2.0. If a copy of the MPL was not distributed with this
* file, You can obtain one at http://mozilla.org/MPL/2.0/. */
#include "jit/mips/MacroAssembler-mips.h"
#include "mozilla/DebugOnly.h"
#include "mozilla/MathAlgorithms.h"
#include "jit/Bailouts.h"
#include "jit/BaselineFrame.h"
#include "jit/BaselineRegisters.h"
#include "jit/IonFrames.h"
#include "jit/MoveEmitter.h"
using namespace js;
using namespace jit;
using mozilla::Abs;
static const int32_t PAYLOAD_OFFSET = NUNBOX32_PAYLOAD_OFFSET;
static const int32_t TAG_OFFSET = NUNBOX32_TYPE_OFFSET;
static_assert(sizeof(intptr_t) == 4, "Not 64-bit clean.");
void
MacroAssemblerMIPS::convertBoolToInt32(Register src, Register dest)
{
// Note that C++ bool is only 1 byte, so zero extend it to clear the
// higher-order bits.
ma_and(dest, src, Imm32(0xff));
}
void
MacroAssemblerMIPS::convertInt32ToDouble(const Register &src, const FloatRegister &dest)
{
as_mtc1(src, dest);
as_cvtdw(dest, dest);
}
void
MacroAssemblerMIPS::convertInt32ToDouble(const Address &src, FloatRegister dest)
{
ma_lw(ScratchRegister, src);
as_mtc1(ScratchRegister, dest);
as_cvtdw(dest, dest);
}
void
MacroAssemblerMIPS::convertUInt32ToDouble(const Register &src, const FloatRegister &dest)
{
// We use SecondScratchFloatReg because MacroAssembler::loadFromTypedArray
// calls with ScratchFloatReg as dest.
MOZ_ASSERT(dest != SecondScratchFloatReg);
// Subtract INT32_MIN to get a positive number
ma_subu(ScratchRegister, src, Imm32(INT32_MIN));
// Convert value
as_mtc1(ScratchRegister, dest);
as_cvtdw(dest, dest);
// Add unsigned value of INT32_MIN
ma_lid(SecondScratchFloatReg, 2147483648.0);
as_addd(dest, dest, SecondScratchFloatReg);
}
// Convert the floating point value to an integer, if it did not fit, then it
// was clamped to INT32_MIN/INT32_MAX, and we can test it.
// NOTE: if the value really was supposed to be INT32_MAX / INT32_MIN then it
// will be wrong.
void
MacroAssemblerMIPS::branchTruncateDouble(const FloatRegister &src, const Register &dest,
Label *fail)
{
Label test, success;
as_truncwd(ScratchFloatReg, src);
as_mfc1(dest, ScratchFloatReg);
ma_b(dest, Imm32(INT32_MAX), fail, Assembler::Equal);
}
// Checks whether a double is representable as a 32-bit integer. If so, the
// integer is written to the output register. Otherwise, a bailout is taken to
// the given snapshot. This function overwrites the scratch float register.
void
MacroAssemblerMIPS::convertDoubleToInt32(const FloatRegister &src, const Register &dest,
Label *fail, bool negativeZeroCheck)
{
// Convert double to int, then convert back and check if we have the
// same number.
as_cvtwd(ScratchFloatReg, src);
as_mfc1(dest, ScratchFloatReg);
as_cvtdw(ScratchFloatReg, ScratchFloatReg);
ma_bc1d(src, ScratchFloatReg, fail, Assembler::DoubleNotEqualOrUnordered);
if (negativeZeroCheck) {
Label notZero;
ma_b(dest, Imm32(0), &notZero, Assembler::NotEqual, ShortJump);
// Test and bail for -0.0, when integer result is 0
// Move the top word of the double into the output reg, if it is
// non-zero, then the original value was -0.0
moveFromDoubleHi(src, dest);
ma_b(dest, Imm32(INT32_MIN), fail, Assembler::Equal);
bind(&notZero);
}
}
void
MacroAssemblerMIPS::addDouble(FloatRegister src, FloatRegister dest)
{
as_addd(dest, dest, src);
}
void
MacroAssemblerMIPS::subDouble(FloatRegister src, FloatRegister dest)
{
as_subd(dest, dest, src);
}
void
MacroAssemblerMIPS::mulDouble(FloatRegister src, FloatRegister dest)
{
as_muld(dest, dest, src);
}
void
MacroAssemblerMIPS::divDouble(FloatRegister src, FloatRegister dest)
{
as_divd(dest, dest, src);
}
void
MacroAssemblerMIPS::negateDouble(FloatRegister reg)
{
as_negd(reg, reg);
}
void
MacroAssemblerMIPS::inc64(AbsoluteAddress dest)
{
ma_li(ScratchRegister, Imm32((int32_t)dest.addr));
as_lw(SecondScratchReg, ScratchRegister, 0);
as_addiu(SecondScratchReg, SecondScratchReg, 1);
as_sw(SecondScratchReg, ScratchRegister, 0);
as_sltiu(SecondScratchReg, SecondScratchReg, 1);
as_lw(ScratchRegister, ScratchRegister, 4);
as_addu(SecondScratchReg, ScratchRegister, SecondScratchReg);
ma_li(ScratchRegister, Imm32((int32_t)dest.addr));
as_sw(SecondScratchReg, ScratchRegister, 4);
}
void
MacroAssemblerMIPS::ma_move(Register rd, Register rs)
{
as_or(rd, rs, zero);
}
void
MacroAssemblerMIPS::ma_li(Register dest, const ImmGCPtr &ptr)
{
writeDataRelocation(ptr);
ma_liPatchable(dest, Imm32(ptr.value));
}
void
MacroAssemblerMIPS::ma_li(const Register &dest, AbsoluteLabel *label)
{
MOZ_ASSERT(!label->bound());
// Thread the patch list through the unpatched address word in the
// instruction stream.
BufferOffset bo = m_buffer.nextOffset();
ma_liPatchable(dest, Imm32(label->prev()));
label->setPrev(bo.getOffset());
}
void
MacroAssemblerMIPS::ma_li(Register dest, Imm32 imm)
{
if (Imm16::isInSignedRange(imm.value)) {
as_addiu(dest, zero, imm.value);
} else if (Imm16::isInUnsignedRange(imm.value)) {
as_ori(dest, zero, Imm16::lower(imm).encode());
} else if (Imm16::lower(imm).encode() == 0) {
as_lui(dest, Imm16::upper(imm).encode());
} else {
as_lui(dest, Imm16::upper(imm).encode());
as_ori(dest, dest, Imm16::lower(imm).encode());
}
}
// This method generates lui and ori instruction pair that can be modified by
// updateLuiOriValue, either during compilation (eg. Assembler::bind), or
// during execution (eg. jit::PatchJump).
void
MacroAssemblerMIPS::ma_liPatchable(Register dest, Imm32 imm)
{
m_buffer.ensureSpace(2 * sizeof(uint32_t));
as_lui(dest, Imm16::upper(imm).encode());
as_ori(dest, dest, Imm16::lower(imm).encode());
}
void
MacroAssemblerMIPS::ma_liPatchable(Register dest, ImmWord imm)
{
return ma_liPatchable(dest, Imm32(int32_t(imm.value)));
}
// Shifts
void
MacroAssemblerMIPS::ma_sll(Register rd, Register rt, Imm32 shift)
{
as_sll(rd, rt, shift.value % 32);
}
void
MacroAssemblerMIPS::ma_srl(Register rd, Register rt, Imm32 shift)
{
as_srl(rd, rt, shift.value % 32);
}
void
MacroAssemblerMIPS::ma_sra(Register rd, Register rt, Imm32 shift)
{
as_sra(rd, rt, shift.value % 32);
}
void
MacroAssemblerMIPS::ma_ror(Register rd, Register rt, Imm32 shift)
{
as_rotr(rd, rt, shift.value % 32);
}
void
MacroAssemblerMIPS::ma_rol(Register rd, Register rt, Imm32 shift)
{
as_rotr(rd, rt, 32 - (shift.value % 32));
}
void
MacroAssemblerMIPS::ma_sll(Register rd, Register rt, Register shift)
{
as_sllv(rd, rt, shift);
}
void
MacroAssemblerMIPS::ma_srl(Register rd, Register rt, Register shift)
{
as_srlv(rd, rt, shift);
}
void
MacroAssemblerMIPS::ma_sra(Register rd, Register rt, Register shift)
{
as_srav(rd, rt, shift);
}
void
MacroAssemblerMIPS::ma_ror(Register rd, Register rt, Register shift)
{
as_rotrv(rd, rt, shift);
}
void
MacroAssemblerMIPS::ma_rol(Register rd, Register rt, Register shift)
{
ma_negu(ScratchRegister, shift);
as_rotrv(rd, rt, ScratchRegister);
}
void
MacroAssemblerMIPS::ma_negu(Register rd, Register rs)
{
as_subu(rd, zero, rs);
}
void
MacroAssemblerMIPS::ma_not(Register rd, Register rs)
{
as_nor(rd, rs, zero);
}
// And.
void
MacroAssemblerMIPS::ma_and(Register rd, Register rs)
{
as_and(rd, rd, rs);
}
void
MacroAssemblerMIPS::ma_and(Register rd, Register rs, Register rt)
{
as_and(rd, rs, rt);
}
void
MacroAssemblerMIPS::ma_and(Register rd, Imm32 imm)
{
ma_and(rd, rd, imm);
}
void
MacroAssemblerMIPS::ma_and(Register rd, Register rs, Imm32 imm)
{
if (Imm16::isInUnsignedRange(imm.value)) {
as_andi(rd, rs, imm.value);
} else {
ma_li(ScratchRegister, imm);
as_and(rd, rs, ScratchRegister);
}
}
// Or.
void
MacroAssemblerMIPS::ma_or(Register rd, Register rs)
{
as_or(rd, rd, rs);
}
void
MacroAssemblerMIPS::ma_or(Register rd, Register rs, Register rt)
{
as_or(rd, rs, rt);
}
void
MacroAssemblerMIPS::ma_or(Register rd, Imm32 imm)
{
ma_or(rd, rd, imm);
}
void
MacroAssemblerMIPS::ma_or(Register rd, Register rs, Imm32 imm)
{
if (Imm16::isInUnsignedRange(imm.value)) {
as_ori(rd, rs, imm.value);
} else {
ma_li(ScratchRegister, imm);
as_or(rd, rs, ScratchRegister);
}
}
// xor
void
MacroAssemblerMIPS::ma_xor(Register rd, Register rs)
{
as_xor(rd, rd, rs);
}
void
MacroAssemblerMIPS::ma_xor(Register rd, Register rs, Register rt)
{
as_xor(rd, rs, rt);
}
void
MacroAssemblerMIPS::ma_xor(Register rd, Imm32 imm)
{
ma_xor(rd, rd, imm);
}
void
MacroAssemblerMIPS::ma_xor(Register rd, Register rs, Imm32 imm)
{
if (Imm16::isInUnsignedRange(imm.value)) {
as_xori(rd, rs, imm.value);
} else {
ma_li(ScratchRegister, imm);
as_xor(rd, rs, ScratchRegister);
}
}
// Arithmetic-based ops.
// Add.
void
MacroAssemblerMIPS::ma_addu(Register rd, Register rs, Imm32 imm)
{
if (Imm16::isInSignedRange(imm.value)) {
as_addiu(rd, rs, imm.value);
} else {
ma_li(ScratchRegister, imm);
as_addu(rd, rs, ScratchRegister);
}
}
void
MacroAssemblerMIPS::ma_addu(Register rd, Register rs)
{
as_addu(rd, rd, rs);
}
void
MacroAssemblerMIPS::ma_addu(Register rd, Imm32 imm)
{
ma_addu(rd, rd, imm);
}
void
MacroAssemblerMIPS::ma_addTestOverflow(Register rd, Register rs, Register rt, Label *overflow)
{
Label goodAddition;
as_addu(SecondScratchReg, rs, rt);
as_xor(ScratchRegister, rs, rt); // If different sign, no overflow
ma_b(ScratchRegister, Imm32(0), &goodAddition, Assembler::LessThan, ShortJump);
// If different sign, then overflow
as_xor(ScratchRegister, rs, SecondScratchReg);
ma_b(ScratchRegister, Imm32(0), overflow, Assembler::LessThan);
bind(&goodAddition);
ma_move(rd, SecondScratchReg);
}
void
MacroAssemblerMIPS::ma_addTestOverflow(Register rd, Register rs, Imm32 imm, Label *overflow)
{
// Check for signed range because of as_addiu
// Check for unsigned range because of as_xori
if (Imm16::isInSignedRange(imm.value) && Imm16::isInUnsignedRange(imm.value)) {
Label goodAddition;
as_addiu(SecondScratchReg, rs, imm.value);
// If different sign, no overflow
as_xori(ScratchRegister, rs, imm.value);
ma_b(ScratchRegister, Imm32(0), &goodAddition, Assembler::LessThan, ShortJump);
// If different sign, then overflow
as_xor(ScratchRegister, rs, SecondScratchReg);
ma_b(ScratchRegister, Imm32(0), overflow, Assembler::LessThan);
bind(&goodAddition);
ma_move(rd, SecondScratchReg);
} else {
ma_li(ScratchRegister, imm);
ma_addTestOverflow(rd, rs, ScratchRegister, overflow);
}
}
// Subtract.
void
MacroAssemblerMIPS::ma_subu(Register rd, Register rs, Register rt)
{
as_subu(rd, rs, rt);
}
void
MacroAssemblerMIPS::ma_subu(Register rd, Register rs, Imm32 imm)
{
if (Imm16::isInSignedRange(-imm.value)) {
as_addiu(rd, rs, -imm.value);
} else {
ma_li(ScratchRegister, imm);
as_subu(rd, rs, ScratchRegister);
}
}
void
MacroAssemblerMIPS::ma_subu(Register rd, Imm32 imm)
{
ma_subu(rd, rd, imm);
}
void
MacroAssemblerMIPS::ma_subTestOverflow(Register rd, Register rs, Register rt, Label *overflow)
{
Label goodSubtraction;
// Use second scratch. The instructions generated by ma_b don't use the
// second scratch register.
ma_subu(SecondScratchReg, rs, rt);
as_xor(ScratchRegister, rs, rt); // If same sign, no overflow
ma_b(ScratchRegister, Imm32(0), &goodSubtraction, Assembler::GreaterThanOrEqual, ShortJump);
// If different sign, then overflow
as_xor(ScratchRegister, rs, SecondScratchReg);
ma_b(ScratchRegister, Imm32(0), overflow, Assembler::LessThan);
bind(&goodSubtraction);
ma_move(rd, SecondScratchReg);
}
void
MacroAssemblerMIPS::ma_subTestOverflow(Register rd, Register rs, Imm32 imm, Label *overflow)
{
if (imm.value != INT32_MIN) {
ma_addTestOverflow(rd, rs, Imm32(-imm.value), overflow);
} else {
ma_li(ScratchRegister, Imm32(imm.value));
ma_subTestOverflow(rd, rs, ScratchRegister, overflow);
}
}
void
MacroAssemblerMIPS::ma_mult(Register rs, Imm32 imm)
{
ma_li(ScratchRegister, imm);
as_mult(rs, ScratchRegister);
}
void
MacroAssemblerMIPS::ma_mul_branch_overflow(Register rd, Register rs, Register rt, Label *overflow)
{
as_mult(rs, rt);
as_mflo(rd);
as_sra(ScratchRegister, rd, 31);
as_mfhi(SecondScratchReg);
ma_b(ScratchRegister, SecondScratchReg, overflow, Assembler::NotEqual);
}
void
MacroAssemblerMIPS::ma_mul_branch_overflow(Register rd, Register rs, Imm32 imm, Label *overflow)
{
ma_li(ScratchRegister, imm);
ma_mul_branch_overflow(rd, rs, ScratchRegister, overflow);
}
void
MacroAssemblerMIPS::ma_div_branch_overflow(Register rd, Register rs, Register rt, Label *overflow)
{
as_div(rs, rt);
as_mflo(rd);
as_mfhi(ScratchRegister);
ma_b(ScratchRegister, ScratchRegister, overflow, Assembler::NonZero);
}
void
MacroAssemblerMIPS::ma_div_branch_overflow(Register rd, Register rs, Imm32 imm, Label *overflow)
{
ma_li(ScratchRegister, imm);
ma_div_branch_overflow(rd, rs, ScratchRegister, overflow);
}
// Memory.
void
MacroAssemblerMIPS::ma_load(const Register &dest, Address address,
LoadStoreSize size, LoadStoreExtension extension)
{
int16_t encodedOffset;
Register base;
if (!Imm16::isInSignedRange(address.offset)) {
ma_li(ScratchRegister, Imm32(address.offset));
as_addu(ScratchRegister, address.base, ScratchRegister);
base = ScratchRegister;
encodedOffset = Imm16(0).encode();
} else {
encodedOffset = Imm16(address.offset).encode();
base = address.base;
}
switch (size) {
case SizeByte:
if (ZeroExtend == extension)
as_lbu(dest, base, encodedOffset);
else
as_lb(dest, base, encodedOffset);
break;
case SizeHalfWord:
if (ZeroExtend == extension)
as_lhu(dest, base, encodedOffset);
else
as_lh(dest, base, encodedOffset);
break;
case SizeWord:
as_lw(dest, base, encodedOffset);
break;
default:
JS_NOT_REACHED("Invalid argument for ma_load");
break;
}
}
void
MacroAssemblerMIPS::ma_load(const Register &dest, const BaseIndex &src,
LoadStoreSize size, LoadStoreExtension extension)
{
computeScaledAddress(src, SecondScratchReg);
ma_load(dest, Address(SecondScratchReg, src.offset), size, extension);
}
void
MacroAssemblerMIPS::ma_store(const Register &data, Address address, LoadStoreSize size,
LoadStoreExtension extension)
{
int16_t encodedOffset;
Register base;
if (!Imm16::isInSignedRange(address.offset)) {
ma_li(ScratchRegister, Imm32(address.offset));
as_addu(ScratchRegister, address.base, ScratchRegister);
base = ScratchRegister;
encodedOffset = Imm16(0).encode();
} else {
encodedOffset = Imm16(address.offset).encode();
base = address.base;
}
switch (size) {
case SizeByte:
as_sb(data, base, encodedOffset);
break;
case SizeHalfWord:
as_sh(data, base, encodedOffset);
break;
case SizeWord:
as_sw(data, base, encodedOffset);
break;
default:
JS_NOT_REACHED("Invalid argument for ma_store");
break;
}
}
void
MacroAssemblerMIPS::ma_store(const Register &data, const BaseIndex &dest,
LoadStoreSize size, LoadStoreExtension extension)
{
computeScaledAddress(dest, SecondScratchReg);
ma_store(data, Address(SecondScratchReg, dest.offset), size, extension);
}
void
MacroAssemblerMIPS::ma_store(const Imm32 &imm, const BaseIndex &dest,
LoadStoreSize size, LoadStoreExtension extension)
{
// Make sure that SecondScratchReg contains absolute address so that
// offset is 0.
computeEffectiveAddress(dest, SecondScratchReg);
// Scrach register is free now, use it for loading imm value
ma_li(ScratchRegister, imm);
// with offset=0 ScratchRegister will not be used in ma_store()
// so we can use it as a parameter here
ma_store(ScratchRegister, Address(SecondScratchReg, 0), size, extension);
}
void
MacroAssemblerMIPS::computeScaledAddress(const BaseIndex &address, Register dest)
{
int32_t shift = Imm32::ShiftOf(address.scale).value;
if (shift) {
ma_sll(dest, address.index, Imm32(shift));
as_addu(dest, address.base, dest);
} else {
as_addu(dest, address.base, address.index);
}
}
// Shortcut for when we know we're transferring 32 bits of data.
void
MacroAssemblerMIPS::ma_lw(Register data, Address address)
{
ma_load(data, address, SizeWord);
}
void
MacroAssemblerMIPS::ma_sw(Register data, Address address)
{
ma_store(data, address, SizeWord);
}
void
MacroAssemblerMIPS::ma_sw(Imm32 imm, Address address)
{
MOZ_ASSERT(address.base != ScratchRegister);
ma_li(ScratchRegister, imm);
if (Imm16::isInSignedRange(address.offset)) {
as_sw(ScratchRegister, address.base, Imm16(address.offset).encode());
} else {
MOZ_ASSERT(address.base != SecondScratchReg);
ma_li(SecondScratchReg, Imm32(address.offset));
as_addu(SecondScratchReg, address.base, SecondScratchReg);
as_sw(ScratchRegister, SecondScratchReg, 0);
}
}
void
MacroAssemblerMIPS::ma_pop(Register r)
{
as_lw(r, StackPointer, 0);
as_addiu(StackPointer, StackPointer, sizeof(intptr_t));
}
void
MacroAssemblerMIPS::ma_push(Register r)
{
if (r == sp) {
// Pushing sp requires one more instruction.
ma_move(ScratchRegister, sp);
r = ScratchRegister;
}
as_addiu(StackPointer, StackPointer, -sizeof(intptr_t));
as_sw(r, StackPointer, 0);
}
// Branches when done from within mips-specific code.
void
MacroAssemblerMIPS::ma_b(Register lhs, Register rhs, Label *label, Condition c, JumpKind jumpKind)
{
switch (c) {
case Equal :
case NotEqual:
branchWithCode(getBranchCode(lhs, rhs, c), label, jumpKind);
break;
case Always:
ma_b(label, jumpKind);
break;
case Zero:
case NonZero:
case Signed:
case NotSigned:
MOZ_ASSERT(lhs == rhs);
branchWithCode(getBranchCode(lhs, c), label, jumpKind);
break;
default:
Condition cond = ma_cmp(ScratchRegister, lhs, rhs, c);
branchWithCode(getBranchCode(ScratchRegister, cond), label, jumpKind);
break;
}
}
void
MacroAssemblerMIPS::ma_b(Register lhs, Imm32 imm, Label *label, Condition c, JumpKind jumpKind)
{
MOZ_ASSERT(c != Overflow);
if (imm.value == 0) {
if (c == Always || c == AboveOrEqual)
ma_b(label, jumpKind);
else if (c == Below)
; // This condition is always false. No branch required.
else
branchWithCode(getBranchCode(lhs, c), label, jumpKind);
} else {
MOZ_ASSERT(lhs != ScratchRegister);
ma_li(ScratchRegister, imm);
ma_b(lhs, ScratchRegister, label, c, jumpKind);
}
}
void
MacroAssemblerMIPS::ma_b(Register lhs, Address addr, Label *label, Condition c, JumpKind jumpKind)
{
MOZ_ASSERT(lhs != ScratchRegister);
ma_lw(ScratchRegister, addr);
ma_b(lhs, ScratchRegister, label, c, jumpKind);
}
void
MacroAssemblerMIPS::ma_b(Address addr, Imm32 imm, Label *label, Condition c, JumpKind jumpKind)
{
ma_lw(SecondScratchReg, addr);
ma_b(SecondScratchReg, imm, label, c, jumpKind);
}
void
MacroAssemblerMIPS::ma_b(Label *label, JumpKind jumpKind)
{
branchWithCode(getBranchCode(BranchIsJump), label, jumpKind);
}
void
MacroAssemblerMIPS::ma_bal(Label *label, JumpKind jumpKind)
{
branchWithCode(getBranchCode(BranchIsCall), label, jumpKind);
}
void
MacroAssemblerMIPS::branchWithCode(InstImm code, Label *label, JumpKind jumpKind)
{
InstImm inst_bgezal = InstImm(op_regimm, zero, rt_bgezal, BOffImm16(0));
InstImm inst_beq = InstImm(op_beq, zero, zero, BOffImm16(0));
if (label->bound()) {
int32_t offset = label->offset() - m_buffer.nextOffset().getOffset();
if (BOffImm16::isInRange(offset))
jumpKind = ShortJump;
if (jumpKind == ShortJump) {
MOZ_ASSERT(BOffImm16::isInRange(offset));
code.setBOffImm16(BOffImm16(offset));
writeInst(code.encode());
as_nop();
return;
}
// Generate long jump because target is out of range of short jump.
if (code.encode() == inst_bgezal.encode()) {
// Handle long call
addLongJump(nextOffset());
ma_liPatchable(ScratchRegister, Imm32(label->offset()));
as_jalr(ScratchRegister);
as_nop();
return;
}
if (code.encode() == inst_beq.encode()) {
// Handle long jump
addLongJump(nextOffset());
ma_liPatchable(ScratchRegister, Imm32(label->offset()));
as_jr(ScratchRegister);
as_nop();
return;
}
// Handle long conditional branch
writeInst(invertBranch(code, BOffImm16(5 * sizeof(uint32_t))).encode());
// No need for a "nop" here because we can clobber scratch.
addLongJump(nextOffset());
ma_liPatchable(ScratchRegister, Imm32(label->offset()));
as_jr(ScratchRegister);
as_nop();
return;
}
// Generate open jump and link it to a label.
// Second word holds a pointer to the next branch in label's chain.
uint32_t nextInChain = label->used() ? label->offset() : LabelBase::INVALID_OFFSET;
if (jumpKind == ShortJump) {
// Make the whole branch continous in the buffer.
m_buffer.ensureSpace(2 * sizeof(uint32_t));
// Indicate that this is short jump with offset 4.
code.setBOffImm16(BOffImm16(4));
BufferOffset bo = writeInst(code.encode());
writeInst(nextInChain);
label->use(bo.getOffset());
return;
}
bool conditional = (code.encode() != inst_bgezal.encode() &&
code.encode() != inst_beq.encode());
// Make the whole branch continous in the buffer.
m_buffer.ensureSpace((conditional ? 5 : 4) * sizeof(uint32_t));
BufferOffset bo = writeInst(code.encode());
writeInst(nextInChain);
label->use(bo.getOffset());
// Leave space for potential long jump.
as_nop();
as_nop();
if (conditional)
as_nop();
}
Assembler::Condition
MacroAssemblerMIPS::ma_cmp(Register scratch, Register lhs, Register rhs, Condition c)
{
switch (c) {
case Above:
// bgtu s,t,label =>
// sltu at,t,s
// bne at,$zero,offs
as_sltu(scratch, rhs, lhs);
return NotEqual;
case AboveOrEqual:
// bgeu s,t,label =>
// sltu at,s,t
// beq at,$zero,offs
as_sltu(scratch, lhs, rhs);
return Equal;
case Below:
// bltu s,t,label =>
// sltu at,s,t
// bne at,$zero,offs
as_sltu(scratch, lhs, rhs);
return NotEqual;
case BelowOrEqual:
// bleu s,t,label =>
// sltu at,t,s
// beq at,$zero,offs
as_sltu(scratch, rhs, lhs);
return Equal;
case GreaterThan:
// bgt s,t,label =>
// slt at,t,s
// bne at,$zero,offs
as_slt(scratch, rhs, lhs);
return NotEqual;
case GreaterThanOrEqual:
// bge s,t,label =>
// slt at,s,t
// beq at,$zero,offs
as_slt(scratch, lhs, rhs);
return Equal;
case LessThan:
// blt s,t,label =>
// slt at,s,t
// bne at,$zero,offs
as_slt(scratch, lhs, rhs);
return NotEqual;
case LessThanOrEqual:
// ble s,t,label =>
// slt at,t,s
// beq at,$zero,offs
as_slt(scratch, rhs, lhs);
return Equal;
case Equal :
case NotEqual:
case Zero:
case NonZero:
case Always:
case Signed:
case NotSigned:
JS_NOT_REACHED("There is a better way to compare for equality.");
break;
case Overflow:
JS_NOT_REACHED("Overflow condition not supported for MIPS.");
break;
default:
JS_NOT_REACHED("Invalid condition for branch.");
}
return Always;
}
void
MacroAssemblerMIPS::ma_cmp_set(Register rd, Register rs, Register rt, Condition c)
{
switch (c) {
case Equal :
// seq d,s,t =>
// xor d,s,t
// sltiu d,d,1
as_xor(rd, rs, rt);
as_sltiu(rd, rd, 1);
break;
case NotEqual:
// sne d,s,t =>
// xor d,s,t
// sltu d,$zero,d
as_xor(rd, rs, rt);
as_sltu(rd, zero, rd);
break;
case Above:
// sgtu d,s,t =>
// sltu d,t,s
as_sltu(rd, rt, rs);
break;
case AboveOrEqual:
// sgeu d,s,t =>
// sltu d,s,t
// xori d,d,1
as_sltu(rd, rs, rt);
as_xori(rd, rd, 1);
break;
case Below:
// sltu d,s,t
as_sltu(rd, rs, rt);
break;
case BelowOrEqual:
// sleu d,s,t =>
// sltu d,t,s
// xori d,d,1
as_sltu(rd, rt, rs);
as_xori(rd, rd, 1);
break;
case GreaterThan:
// sgt d,s,t =>
// slt d,t,s
as_slt(rd, rt, rs);
break;
case GreaterThanOrEqual:
// sge d,s,t =>
// slt d,s,t
// xori d,d,1
as_slt(rd, rs, rt);
as_xori(rd, rd, 1);
break;
case LessThan:
// slt d,s,t
as_slt(rd, rs, rt);
break;
case LessThanOrEqual:
// sle d,s,t =>
// slt d,t,s
// xori d,d,1
as_slt(rd, rt, rs);
as_xori(rd, rd, 1);
break;
case Zero:
MOZ_ASSERT(rs == rt);
// seq d,s,$zero =>
// xor d,s,$zero
// sltiu d,d,1
as_xor(rd, rs, zero);
as_sltiu(rd, rd, 1);
break;
case NonZero:
// sne d,s,$zero =>
// xor d,s,$zero
// sltu d,$zero,d
as_xor(rd, rs, zero);
as_sltu(rd, zero, rd);
break;
case Signed:
as_slt(rd, rs, zero);
break;
case NotSigned:
// sge d,s,$zero =>
// slt d,s,$zero
// xori d,d,1
as_slt(rd, rs, zero);
as_xori(rd, rd, 1);
break;
default:
JS_NOT_REACHED("Invalid condition for ma_cmp_set.");
break;
}
}
void
MacroAssemblerMIPS::compareFloatingPoint(FloatFormat fmt, FloatRegister lhs, FloatRegister rhs,
DoubleCondition c, FloatTestKind *testKind,
FPConditionBit fcc)
{
switch (c) {
case DoubleOrdered:
as_cun(fmt, lhs, rhs, fcc);
*testKind = TestForFalse;
break;
case DoubleEqual:
as_ceq(fmt, lhs, rhs, fcc);
*testKind = TestForTrue;
break;
case DoubleNotEqual:
as_cueq(fmt, lhs, rhs, fcc);
*testKind = TestForFalse;
break;
case DoubleGreaterThan:
as_colt(fmt, rhs, lhs, fcc);
*testKind = TestForTrue;
break;
case DoubleGreaterThanOrEqual:
as_cole(fmt, rhs, lhs, fcc);
*testKind = TestForTrue;
break;
case DoubleLessThan:
as_colt(fmt, lhs, rhs, fcc);
*testKind = TestForTrue;
break;
case DoubleLessThanOrEqual:
as_cole(fmt, lhs, rhs, fcc);
*testKind = TestForTrue;
break;
case DoubleUnordered:
as_cun(fmt, lhs, rhs, fcc);
*testKind = TestForTrue;
break;
case DoubleEqualOrUnordered:
as_cueq(fmt, lhs, rhs, fcc);
*testKind = TestForTrue;
break;
case DoubleNotEqualOrUnordered:
as_ceq(fmt, lhs, rhs, fcc);
*testKind = TestForFalse;
break;
case DoubleGreaterThanOrUnordered:
as_cult(fmt, rhs, lhs, fcc);
*testKind = TestForTrue;
break;
case DoubleGreaterThanOrEqualOrUnordered:
as_cule(fmt, rhs, lhs, fcc);
*testKind = TestForTrue;
break;
case DoubleLessThanOrUnordered:
as_cult(fmt, lhs, rhs, fcc);
*testKind = TestForTrue;
break;
case DoubleLessThanOrEqualOrUnordered:
as_cule(fmt, lhs, rhs, fcc);
*testKind = TestForTrue;
break;
default:
JS_NOT_REACHED("Invalid DoubleCondition.");
break;
}
}
void
MacroAssemblerMIPS::ma_cmp_set_double(Register dest, FloatRegister lhs, FloatRegister rhs,
DoubleCondition c)
{
ma_li(dest, Imm32(0));
ma_li(ScratchRegister, Imm32(1));
FloatTestKind moveCondition;
compareFloatingPoint(DoubleFloat, lhs, rhs, c, &moveCondition);
if (moveCondition == TestForTrue)
as_movt(dest, ScratchRegister);
else
as_movf(dest, ScratchRegister);
}
void
MacroAssemblerMIPS::ma_cmp_set_float32(Register dest, FloatRegister lhs, FloatRegister rhs,
DoubleCondition c)
{
ma_li(dest, Imm32(0));
ma_li(ScratchRegister, Imm32(1));
FloatTestKind moveCondition;
compareFloatingPoint(SingleFloat, lhs, rhs, c, &moveCondition);
if (moveCondition == TestForTrue)
as_movt(dest, ScratchRegister);
else
as_movf(dest, ScratchRegister);
}
void
MacroAssemblerMIPS::ma_cmp_set(Register rd, Register rs, Imm32 imm, Condition c)
{
ma_li(ScratchRegister, imm);
ma_cmp_set(rd, rs, ScratchRegister, c);
}
void
MacroAssemblerMIPS::ma_cmp_set(Register rd, Register rs, Address addr, Condition c)
{
ma_lw(ScratchRegister, addr);
ma_cmp_set(rd, rs, ScratchRegister, c);
}
void
MacroAssemblerMIPS::ma_cmp_set(Register dst, Address lhs, Register rhs, Condition c)
{
ma_lw(ScratchRegister, lhs);
ma_cmp_set(dst, ScratchRegister, rhs, c);
}
// fp instructions
void
MacroAssemblerMIPS::ma_lis(FloatRegister dest, float value)
{
Imm32 imm(mozilla::BitwiseCast<uint32_t>(value));
ma_li(ScratchRegister, imm);
moveToFloat32(ScratchRegister, dest);
}
void
MacroAssemblerMIPS::ma_lid(FloatRegister dest, double value)
{
struct DoubleStruct {
uint32_t lo;
uint32_t hi;
} ;
DoubleStruct intStruct = mozilla::BitwiseCast<DoubleStruct>(value);
// put hi part of 64 bit value into the odd register
if (intStruct.hi == 0) {
moveToDoubleHi(zero, dest);
} else {
ma_li(ScratchRegister, Imm32(intStruct.hi));
moveToDoubleHi(ScratchRegister, dest);
}
// put low part of 64 bit value into the even register
if (intStruct.lo == 0) {
moveToDoubleLo(zero, dest);
} else {
ma_li(ScratchRegister, Imm32(intStruct.lo));
moveToDoubleLo(ScratchRegister, dest);
}
}
void
MacroAssemblerMIPS::ma_liNegZero(FloatRegister dest)
{
moveToDoubleLo(zero, dest);
ma_li(ScratchRegister, Imm32(INT_MIN));
moveToDoubleHi(ScratchRegister, dest);
}
void
MacroAssemblerMIPS::ma_mv(FloatRegister src, ValueOperand dest)
{
moveFromDoubleLo(src, dest.payloadReg());
moveFromDoubleHi(src, dest.typeReg());
}
void
MacroAssemblerMIPS::ma_mv(ValueOperand src, FloatRegister dest)
{
moveToDoubleLo(src.payloadReg(), dest);
moveToDoubleHi(src.typeReg(), dest);
}
void
MacroAssemblerMIPS::ma_ls(FloatRegister ft, Address address)
{
if (Imm16::isInSignedRange(address.offset)) {
as_ls(ft, address.base, Imm16(address.offset).encode());
} else {
MOZ_ASSERT(address.base != ScratchRegister);
ma_li(ScratchRegister, Imm32(address.offset));
as_addu(ScratchRegister, address.base, ScratchRegister);
as_ls(ft, ScratchRegister, 0);
}
}
void
MacroAssemblerMIPS::ma_ld(FloatRegister ft, Address address)
{
// Use single precision load instructions so we don't have to worry about
// alignment.
int32_t off2 = address.offset + TAG_OFFSET;
if (Imm16::isInSignedRange(address.offset) && Imm16::isInSignedRange(off2)) {
as_ls(ft, address.base, Imm16(address.offset).encode());
as_ls(getOddPair(ft), address.base, Imm16(off2).encode());
} else {
ma_li(ScratchRegister, Imm32(address.offset));
as_addu(ScratchRegister, address.base, ScratchRegister);
as_ls(ft, ScratchRegister, PAYLOAD_OFFSET);
as_ls(getOddPair(ft), ScratchRegister, TAG_OFFSET);
}
}
void
MacroAssemblerMIPS::ma_sd(FloatRegister ft, Address address)
{
int32_t off2 = address.offset + TAG_OFFSET;
if (Imm16::isInSignedRange(address.offset) && Imm16::isInSignedRange(off2)) {
as_ss(ft, address.base, Imm16(address.offset).encode());
as_ss(getOddPair(ft), address.base, Imm16(off2).encode());
} else {
ma_li(ScratchRegister, Imm32(address.offset));
as_addu(ScratchRegister, address.base, ScratchRegister);
as_ss(ft, ScratchRegister, PAYLOAD_OFFSET);
as_ss(getOddPair(ft), ScratchRegister, TAG_OFFSET);
}
}
void
MacroAssemblerMIPS::ma_sd(FloatRegister ft, BaseIndex address)
{
computeScaledAddress(address, SecondScratchReg);
ma_sd(ft, Address(SecondScratchReg, address.offset));
}
void
MacroAssemblerMIPS::ma_ss(FloatRegister ft, Address address)
{
if (Imm16::isInSignedRange(address.offset)) {
as_ss(ft, address.base, Imm16(address.offset).encode());
} else {
ma_li(ScratchRegister, Imm32(address.offset));
as_addu(ScratchRegister, address.base, ScratchRegister);
as_ss(ft, ScratchRegister, 0);
}
}
void
MacroAssemblerMIPS::ma_ss(FloatRegister ft, BaseIndex address)
{
computeScaledAddress(address, SecondScratchReg);
ma_ss(ft, Address(SecondScratchReg, address.offset));
}
void
MacroAssemblerMIPS::ma_pop(FloatRegister fs)
{
ma_ld(fs, Address(StackPointer, 0));
as_addiu(StackPointer, StackPointer, sizeof(double));
}
void
MacroAssemblerMIPS::ma_push(FloatRegister fs)
{
as_addiu(StackPointer, StackPointer, -sizeof(double));
ma_sd(fs, Address(StackPointer, 0));
}
void
MacroAssemblerMIPS::ma_bc1s(FloatRegister lhs, FloatRegister rhs, Label *label,
DoubleCondition c, JumpKind jumpKind, FPConditionBit fcc)
{
FloatTestKind testKind;
compareFloatingPoint(SingleFloat, lhs, rhs, c, &testKind, fcc);
branchWithCode(getBranchCode(testKind, fcc), label, jumpKind);
}
void
MacroAssemblerMIPS::ma_bc1d(FloatRegister lhs, FloatRegister rhs, Label *label,
DoubleCondition c, JumpKind jumpKind, FPConditionBit fcc)
{
FloatTestKind testKind;
compareFloatingPoint(DoubleFloat, lhs, rhs, c, &testKind, fcc);
branchWithCode(getBranchCode(testKind, fcc), label, jumpKind);
}
bool
MacroAssemblerMIPSCompat::buildFakeExitFrame(const Register &scratch, uint32_t *offset)
{
mozilla::DebugOnly<uint32_t> initialDepth = framePushed();
CodeLabel cl;
ma_li(scratch, cl.dest());
uint32_t descriptor = MakeFrameDescriptor(framePushed(), IonFrame_OptimizedJS);
Push(Imm32(descriptor));
Push(scratch);
bind(cl.src());
*offset = currentOffset();
MOZ_ASSERT(framePushed() == initialDepth + IonExitFrameLayout::Size());
return addCodeLabel(cl);
}
bool
MacroAssemblerMIPSCompat::buildOOLFakeExitFrame(void *fakeReturnAddr)
{
// DebugOnly<uint32_t> initialDepth = framePushed();
uint32_t descriptor = MakeFrameDescriptor(framePushed(), IonFrame_OptimizedJS);
Push(Imm32(descriptor)); // descriptor_
Push(ImmWord(fakeReturnAddr));
return true;
}
void
MacroAssemblerMIPSCompat::callWithExitFrame(IonCode *target)
{
uint32_t descriptor = MakeFrameDescriptor(framePushed(), IonFrame_OptimizedJS);
Push(Imm32(descriptor)); // descriptor
addPendingJump(m_buffer.nextOffset(), target->raw(), Relocation::IONCODE);
ma_liPatchable(ScratchRegister, ImmWord(target->raw()));
ma_callIonHalfPush(ScratchRegister);
}
void
MacroAssemblerMIPSCompat::callWithExitFrame(IonCode *target, Register dynStack)
{
ma_addu(dynStack, dynStack, Imm32(framePushed()));
makeFrameDescriptor(dynStack, IonFrame_OptimizedJS);
Push(dynStack); // descriptor
addPendingJump(m_buffer.nextOffset(), target->raw(), Relocation::IONCODE);
ma_liPatchable(ScratchRegister, ImmWord(target->raw()));
ma_callIonHalfPush(ScratchRegister);
}
void
MacroAssemblerMIPSCompat::callIon(const Register &callee)
{
MOZ_ASSERT((framePushed() & 3) == 0);
if ((framePushed() & 7) == 4) {
ma_callIonHalfPush(callee);
} else {
adjustFrame(sizeof(uint32_t));
ma_callIon(callee);
}
}
void
MacroAssemblerMIPSCompat::reserveStack(uint32_t amount)
{
if (amount)
ma_subu(StackPointer, StackPointer, Imm32(amount));
adjustFrame(amount);
}
void
MacroAssemblerMIPSCompat::freeStack(uint32_t amount)
{
MOZ_ASSERT(amount <= framePushed_);
if (amount)
ma_addu(StackPointer, StackPointer, Imm32(amount));
adjustFrame(-amount);
}
void
MacroAssemblerMIPSCompat::freeStack(Register amount)
{
as_addu(StackPointer, StackPointer, amount);
}
void
MacroAssemblerMIPSCompat::add32(Register src, Register dest)
{
as_addu(dest, dest, src);
}
void
MacroAssemblerMIPSCompat::add32(Imm32 imm, Register dest)
{
ma_addu(dest, dest, imm);
}
void
MacroAssemblerMIPSCompat::add32(Imm32 imm, const Address &dest)
{
load32(dest, SecondScratchReg);
ma_addu(SecondScratchReg, imm);
store32(SecondScratchReg, dest);
}
void
MacroAssemblerMIPSCompat::sub32(Imm32 imm, Register dest)
{
ma_subu(dest, dest, imm);
}
void
MacroAssemblerMIPSCompat::sub32(Register src, Register dest)
{
ma_subu(dest, dest, src);
}
void
MacroAssemblerMIPSCompat::addPtr(Register src, Register dest)
{
ma_addu(dest, src);
}
void
MacroAssemblerMIPSCompat::addPtr(const Address &src, Register dest)
{
loadPtr(src, ScratchRegister);
ma_addu(dest, ScratchRegister);
}
void
MacroAssemblerMIPSCompat::subPtr(Register src, Register dest)
{
ma_subu(dest, dest, src);
}
void
MacroAssemblerMIPSCompat::not32(Register reg)
{
ma_not(reg, reg);
}
// Logical operations
void
MacroAssemblerMIPSCompat::and32(Imm32 imm, Register dest)
{
ma_and(dest, imm);
}
void
MacroAssemblerMIPSCompat::and32(Imm32 imm, const Address &dest)
{
load32(dest, SecondScratchReg);
ma_and(SecondScratchReg, imm);
store32(SecondScratchReg, dest);
}
void
MacroAssemblerMIPSCompat::or32(Imm32 imm, const Address &dest)
{
load32(dest, SecondScratchReg);
ma_or(SecondScratchReg, imm);
store32(SecondScratchReg, dest);
}
void
MacroAssemblerMIPSCompat::xor32(Imm32 imm, Register dest)
{
ma_xor(dest, imm);
}
void
MacroAssemblerMIPSCompat::xorPtr(Imm32 imm, Register dest)
{
ma_xor(dest, imm);
}
void
MacroAssemblerMIPSCompat::xorPtr(Register src, Register dest)
{
ma_xor(dest, src);
}
void
MacroAssemblerMIPSCompat::orPtr(Imm32 imm, Register dest)
{
ma_or(dest, imm);
}
void
MacroAssemblerMIPSCompat::orPtr(Register src, Register dest)
{
ma_or(dest, src);
}
void
MacroAssemblerMIPSCompat::andPtr(Imm32 imm, Register dest)
{
ma_and(dest, imm);
}
void
MacroAssemblerMIPSCompat::andPtr(Register src, Register dest)
{
ma_and(dest, src);
}
void
MacroAssemblerMIPSCompat::move32(const Imm32 &imm, const Register &dest)
{
ma_li(dest, imm);
}
void
MacroAssemblerMIPSCompat::move32(const Register &src, const Register &dest)
{
ma_move(dest, src);
}
void
MacroAssemblerMIPSCompat::movePtr(const Register &src, const Register &dest)
{
ma_move(dest, src);
}
void
MacroAssemblerMIPSCompat::movePtr(const ImmWord &imm, const Register &dest)
{
ma_li(dest, Imm32(imm.value));
}
void
MacroAssemblerMIPSCompat::movePtr(const ImmGCPtr &imm, const Register &dest)
{
ma_li(dest, imm);
}
void
MacroAssemblerMIPSCompat::load8ZeroExtend(const Address &address, const Register &dest)
{
ma_load(dest, address, SizeByte, ZeroExtend);
}
void
MacroAssemblerMIPSCompat::load8ZeroExtend(const BaseIndex &src, const Register &dest)
{
ma_load(dest, src, SizeByte, ZeroExtend);
}
void
MacroAssemblerMIPSCompat::load8SignExtend(const Address &address, const Register &dest)
{
ma_load(dest, address, SizeByte, SignExtend);
}
void
MacroAssemblerMIPSCompat::load8SignExtend(const BaseIndex &src, const Register &dest)
{
ma_load(dest, src, SizeByte, SignExtend);
}
void
MacroAssemblerMIPSCompat::load16ZeroExtend(const Address &address, const Register &dest)
{
ma_load(dest, address, SizeHalfWord, ZeroExtend);
}
void
MacroAssemblerMIPSCompat::load16ZeroExtend(const BaseIndex &src, const Register &dest)
{
ma_load(dest, src, SizeHalfWord, ZeroExtend);
}
void
MacroAssemblerMIPSCompat::load16SignExtend(const Address &address, const Register &dest)
{
ma_load(dest, address, SizeHalfWord, SignExtend);
}
void
MacroAssemblerMIPSCompat::load16SignExtend(const BaseIndex &src, const Register &dest)
{
ma_load(dest, src, SizeHalfWord, SignExtend);
}
void
MacroAssemblerMIPSCompat::load32(const Address &address, const Register &dest)
{
ma_lw(dest, address);
}
void
MacroAssemblerMIPSCompat::load32(const BaseIndex &address, const Register &dest)
{
ma_load(dest, address, SizeWord);
}
void
MacroAssemblerMIPSCompat::load32(const AbsoluteAddress &address, const Register &dest)
{
ma_li(ScratchRegister, Imm32((uint32_t)address.addr));
as_lw(dest, ScratchRegister, 0);
}
void
MacroAssemblerMIPSCompat::loadPtr(const Address &address, const Register &dest)
{
ma_lw(dest, address);
}
void
MacroAssemblerMIPSCompat::loadPtr(const BaseIndex &src, const Register &dest)
{
load32(src, dest);
}
void
MacroAssemblerMIPSCompat::loadPtr(const AbsoluteAddress &address, const Register &dest)
{
ma_li(ScratchRegister, Imm32((uint32_t)address.addr));
as_lw(dest, ScratchRegister, 0);
}
void
MacroAssemblerMIPSCompat::loadPrivate(const Address &address, const Register &dest)
{
ma_lw(dest, Address(address.base, address.offset + PAYLOAD_OFFSET));
}
void
MacroAssemblerMIPSCompat::loadDouble(const Address &address, const FloatRegister &dest)
{
ma_ld(dest, address);
}
void
MacroAssemblerMIPSCompat::loadDouble(const BaseIndex &src, const FloatRegister &dest)
{
computeScaledAddress(src, SecondScratchReg);
ma_ld(dest, Address(SecondScratchReg, src.offset));
}
void
MacroAssemblerMIPSCompat::loadFloatAsDouble(const Address &address, const FloatRegister &dest)
{
ma_ls(dest, address);
as_cvtds(dest, dest);
}
void
MacroAssemblerMIPSCompat::loadFloatAsDouble(const BaseIndex &src, const FloatRegister &dest)
{
computeScaledAddress(src, SecondScratchReg);
ma_ls(dest, Address(SecondScratchReg, src.offset));
as_cvtds(dest, dest);
}
void
MacroAssemblerMIPSCompat::store8(const Imm32 &imm, const Address &address)
{
ma_li(SecondScratchReg, imm);
ma_store(SecondScratchReg, address, SizeByte);
}
void
MacroAssemblerMIPSCompat::store8(const Register &src, const Address &address)
{
ma_store(src, address, SizeByte);
}
void
MacroAssemblerMIPSCompat::store8(const Imm32 &imm, const BaseIndex &dest)
{
ma_store(imm, dest, SizeByte);
}
void
MacroAssemblerMIPSCompat::store8(const Register &src, const BaseIndex &dest)
{
ma_store(src, dest, SizeByte);
}
void
MacroAssemblerMIPSCompat::store16(const Imm32 &imm, const Address &address)
{
ma_li(SecondScratchReg, imm);
ma_store(SecondScratchReg, address, SizeHalfWord);
}
void
MacroAssemblerMIPSCompat::store16(const Register &src, const Address &address)
{
ma_store(src, address, SizeHalfWord);
}
void
MacroAssemblerMIPSCompat::store16(const Imm32 &imm, const BaseIndex &dest)
{
ma_store(imm, dest, SizeHalfWord);
}
void
MacroAssemblerMIPSCompat::store16(const Register &src, const BaseIndex &address)
{
ma_store(src, address, SizeHalfWord);
}
void
MacroAssemblerMIPSCompat::store32(const Register &src, const AbsoluteAddress &address)
{
storePtr(src, address);
}
void
MacroAssemblerMIPSCompat::store32(const Register &src, const Address &address)
{
storePtr(src, address);
}
void
MacroAssemblerMIPSCompat::store32(const Imm32 &src, const Address &address)
{
move32(src, ScratchRegister);
storePtr(ScratchRegister, address);
}
void
MacroAssemblerMIPSCompat::store32(const Register &src, const BaseIndex &dest)
{
ma_store(src, dest, SizeWord);
}
void
MacroAssemblerMIPSCompat::storePtr(ImmWord imm, const Address &address)
{
ma_li(ScratchRegister, Imm32(imm.value));
ma_sw(ScratchRegister, address);
}
void
MacroAssemblerMIPSCompat::storePtr(ImmGCPtr imm, const Address &address)
{
ma_li(ScratchRegister, imm);
ma_sw(ScratchRegister, address);
}
void
MacroAssemblerMIPSCompat::storePtr(Register src, const Address &address)
{
ma_sw(src, address);
}
void
MacroAssemblerMIPSCompat::storePtr(const Register &src, const AbsoluteAddress &dest)
{
ma_li(ScratchRegister, Imm32((uint32_t)dest.addr));
as_sw(src, ScratchRegister, 0);
}
void
MacroAssemblerMIPSCompat::subPtr(Imm32 imm, const Register dest)
{
ma_subu(dest, dest, imm);
}
void
MacroAssemblerMIPSCompat::addPtr(Imm32 imm, const Register dest)
{
ma_addu(dest, imm);
}
void
MacroAssemblerMIPSCompat::addPtr(Imm32 imm, const Address &dest)
{
loadPtr(dest, ScratchRegister);
addPtr(imm, ScratchRegister);
storePtr(ScratchRegister, dest);
}
void
MacroAssemblerMIPSCompat::branchDouble(DoubleCondition cond, const FloatRegister &lhs,
const FloatRegister &rhs, Label *label)
{
ma_bc1d(lhs, rhs, label, cond);
}
// higher level tag testing code
Operand
ToPayload(Operand base)
{
return Operand(Register::FromCode(base.base()), base.disp() + PAYLOAD_OFFSET);
}
Operand
ToType(Operand base)
{
return Operand(Register::FromCode(base.base()), base.disp() + TAG_OFFSET);
}
void
MacroAssemblerMIPSCompat::branchTestGCThing(Condition cond, const Address &address, Label *label)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
extractTag(address, SecondScratchReg);
ma_b(SecondScratchReg, ImmTag(JSVAL_LOWER_INCL_TAG_OF_GCTHING_SET), label,
(cond == Equal) ? AboveOrEqual : Below);
}
void
MacroAssemblerMIPSCompat::branchTestGCThing(Condition cond, const BaseIndex &src, Label *label)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
extractTag(src, SecondScratchReg);
ma_b(SecondScratchReg, ImmTag(JSVAL_LOWER_INCL_TAG_OF_GCTHING_SET), label,
(cond == Equal) ? AboveOrEqual : Below);
}
void
MacroAssemblerMIPSCompat::branchTestPrimitive(Condition cond, const ValueOperand &value,
Label *label)
{
branchTestPrimitive(cond, value.typeReg(), label);
}
void
MacroAssemblerMIPSCompat::branchTestPrimitive(Condition cond, const Register &tag, Label *label)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
ma_b(tag, ImmTag(JSVAL_UPPER_EXCL_TAG_OF_PRIMITIVE_SET), label,
(cond == Equal) ? Below : AboveOrEqual);
}
void
MacroAssemblerMIPSCompat::branchTestInt32(Condition cond, const ValueOperand &value, Label *label)
{
MOZ_ASSERT(cond == Assembler::Equal || cond == Assembler::NotEqual);
ma_b(value.typeReg(), ImmType(JSVAL_TYPE_INT32), label, cond);
}
void
MacroAssemblerMIPSCompat::branchTestInt32(Condition cond, const Register &tag, Label *label)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
ma_b(tag, ImmTag(JSVAL_TAG_INT32), label, cond);
}
void
MacroAssemblerMIPSCompat::branchTestInt32(Condition cond, const Address &address, Label *label)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
extractTag(address, SecondScratchReg);
ma_b(SecondScratchReg, ImmTag(JSVAL_TAG_INT32), label, cond);
}
void
MacroAssemblerMIPSCompat::branchTestInt32(Condition cond, const BaseIndex &src, Label *label)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
extractTag(src, SecondScratchReg);
ma_b(SecondScratchReg, ImmTag(JSVAL_TAG_INT32), label, cond);
}
void
MacroAssemblerMIPSCompat:: branchTestBoolean(Condition cond, const ValueOperand &value,
Label *label)
{
MOZ_ASSERT(cond == Assembler::Equal || cond == Assembler::NotEqual);
ma_b(value.typeReg(), ImmType(JSVAL_TYPE_BOOLEAN), label, cond);
}
void
MacroAssemblerMIPSCompat:: branchTestBoolean(Condition cond, const Register &tag, Label *label)
{
MOZ_ASSERT(cond == Assembler::Equal || cond == Assembler::NotEqual);
ma_b(tag, ImmType(JSVAL_TYPE_BOOLEAN), label, cond);
}
void
MacroAssemblerMIPSCompat::branchTestBoolean(Condition cond, const BaseIndex &src, Label *label)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
extractTag(src, SecondScratchReg);
ma_b(SecondScratchReg, ImmType(JSVAL_TYPE_BOOLEAN), label, cond);
}
void
MacroAssemblerMIPSCompat::branchTestDouble(Condition cond, const ValueOperand &value, Label *label)
{
MOZ_ASSERT(cond == Assembler::Equal || cond == Assembler::NotEqual);
Assembler::Condition actual = (cond == Equal) ? Below : AboveOrEqual;
ma_b(value.typeReg(), ImmTag(JSVAL_TAG_CLEAR), label, actual);
}
void
MacroAssemblerMIPSCompat::branchTestDouble(Condition cond, const Register &tag, Label *label)
{
MOZ_ASSERT(cond == Assembler::Equal || cond == NotEqual);
Condition actual = (cond == Equal) ? Below : AboveOrEqual;
ma_b(tag, ImmTag(JSVAL_TAG_CLEAR), label, actual);
}
void
MacroAssemblerMIPSCompat::branchTestDouble(Condition cond, const Address &address, Label *label)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
extractTag(address, SecondScratchReg);
ma_b(SecondScratchReg, ImmTag(JSVAL_TAG_CLEAR), label, cond);
}
void
MacroAssemblerMIPSCompat::branchTestDouble(Condition cond, const BaseIndex &src, Label *label)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
Condition actual = (cond == Equal) ? Below : AboveOrEqual;
extractTag(src, SecondScratchReg);
ma_b(SecondScratchReg, ImmTag(JSVAL_TAG_CLEAR), label, actual);
}
void
MacroAssemblerMIPSCompat::branchTestNull(Condition cond, const ValueOperand &value, Label *label)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
ma_b(value.typeReg(), ImmType(JSVAL_TYPE_NULL), label, cond);
}
void
MacroAssemblerMIPSCompat::branchTestNull(Condition cond, const Register &tag, Label *label)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
ma_b(tag, ImmTag(JSVAL_TAG_NULL), label, cond);
}
void
MacroAssemblerMIPSCompat::branchTestNull(Condition cond, const BaseIndex &src, Label *label)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
extractTag(src, SecondScratchReg);
ma_b(SecondScratchReg, ImmTag(JSVAL_TAG_NULL), label, cond);
}
void
MacroAssemblerMIPSCompat::branchTestObject(Condition cond, const ValueOperand &value, Label *label)
{
branchTestObject(cond, value.typeReg(), label);
}
void
MacroAssemblerMIPSCompat::branchTestObject(Condition cond, const Register &tag, Label *label)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
ma_b(tag, ImmTag(JSVAL_TAG_OBJECT), label, cond);
}
void
MacroAssemblerMIPSCompat::branchTestObject(Condition cond, const BaseIndex &src, Label *label)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
extractTag(src, SecondScratchReg);
ma_b(SecondScratchReg, ImmTag(JSVAL_TAG_OBJECT), label, cond);
}
void
MacroAssemblerMIPSCompat::branchTestString(Condition cond, const ValueOperand &value, Label *label)
{
branchTestString(cond, value.typeReg(), label);
}
void
MacroAssemblerMIPSCompat::branchTestString(Condition cond, const Register &tag, Label *label)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
ma_b(tag, ImmTag(JSVAL_TAG_STRING), label, cond);
}
void
MacroAssemblerMIPSCompat::branchTestString(Condition cond, const BaseIndex &src, Label *label)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
extractTag(src, SecondScratchReg);
ma_b(SecondScratchReg, ImmTag(JSVAL_TAG_STRING), label, cond);
}
void
MacroAssemblerMIPSCompat::branchTestUndefined(Condition cond, const ValueOperand &value,
Label *label)
{
MOZ_ASSERT(cond == Assembler::Equal || cond == Assembler::NotEqual);
ma_b(value.typeReg(), ImmType(JSVAL_TYPE_UNDEFINED), label, cond);
}
void
MacroAssemblerMIPSCompat::branchTestUndefined(Condition cond, const Register &tag, Label *label)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
ma_b(tag, ImmTag(JSVAL_TAG_UNDEFINED), label, cond);
}
void
MacroAssemblerMIPSCompat::branchTestUndefined(Condition cond, const BaseIndex &src, Label *label)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
extractTag(src, SecondScratchReg);
ma_b(SecondScratchReg, ImmTag(JSVAL_TAG_UNDEFINED), label, cond);
}
void
MacroAssemblerMIPSCompat::branchTestUndefined(Condition cond, const Address &address, Label *label)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
extractTag(address, SecondScratchReg);
ma_b(SecondScratchReg, ImmTag(JSVAL_TAG_UNDEFINED), label, cond);
}
void
MacroAssemblerMIPSCompat::branchTestNumber(Condition cond, const ValueOperand &value, Label *label)
{
branchTestNumber(cond, value.typeReg(), label);
}
void
MacroAssemblerMIPSCompat::branchTestNumber(Condition cond, const Register &tag, Label *label)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
ma_b(tag, ImmTag(JSVAL_UPPER_INCL_TAG_OF_NUMBER_SET), label,
cond == Equal ? BelowOrEqual : Above);
}
void
MacroAssemblerMIPSCompat::branchTestMagic(Condition cond, const ValueOperand &value, Label *label)
{
branchTestMagic(cond, value.typeReg(), label);
}
void
MacroAssemblerMIPSCompat::branchTestMagic(Condition cond, const Register &tag, Label *label)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
ma_b(tag, ImmTag(JSVAL_TAG_MAGIC), label, cond);
}
void
MacroAssemblerMIPSCompat::branchTestMagic(Condition cond, const Address &address, Label *label)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
extractTag(address, SecondScratchReg);
ma_b(SecondScratchReg, ImmTag(JSVAL_TAG_MAGIC), label, cond);
}
void
MacroAssemblerMIPSCompat::branchTestMagic(Condition cond, const BaseIndex &src, Label *label)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
extractTag(src, SecondScratchReg);
ma_b(SecondScratchReg, ImmTag(JSVAL_TAG_MAGIC), label, cond);
}
void
MacroAssemblerMIPSCompat::branchTestValue(Condition cond, const ValueOperand &value,
const Value &v, Label *label)
{
moveData(v, ScratchRegister);
if (cond == Equal) {
Label done;
ma_b(value.payloadReg(), ScratchRegister, &done, NotEqual, ShortJump);
{
ma_b(value.typeReg(), Imm32(getType(v)), label, Equal);
}
bind(&done);
} else {
MOZ_ASSERT(cond == NotEqual);
ma_b(value.payloadReg(), ScratchRegister, label, NotEqual);
ma_b(value.typeReg(), Imm32(getType(v)), label, NotEqual);
}
}
void
MacroAssemblerMIPSCompat::branchTestValue(Condition cond, const Address &valaddr,
const ValueOperand &value, Label *label)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
// Load tag.
ma_lw(ScratchRegister, Address(valaddr.base, valaddr.offset + TAG_OFFSET));
branchPtr(cond, ScratchRegister, value.typeReg(), label);
// Load payload
ma_lw(ScratchRegister, Address(valaddr.base, valaddr.offset + PAYLOAD_OFFSET));
branchPtr(cond, ScratchRegister, value.payloadReg(), label);
}
// unboxing code
void
MacroAssemblerMIPSCompat::unboxInt32(const ValueOperand &operand, const Register &dest)
{
ma_move(dest, operand.payloadReg());
}
void
MacroAssemblerMIPSCompat::unboxInt32(const Address &src, const Register &dest)
{
ma_lw(dest, Address(src.base, src.offset + PAYLOAD_OFFSET));
}
void
MacroAssemblerMIPSCompat::unboxBoolean(const ValueOperand &operand, const Register &dest)
{
ma_move(dest, operand.payloadReg());
}
void
MacroAssemblerMIPSCompat::unboxBoolean(const Address &src, const Register &dest)
{
ma_lw(dest, Address(src.base, src.offset + PAYLOAD_OFFSET));
}
void
MacroAssemblerMIPSCompat::unboxDouble(const ValueOperand &operand, const FloatRegister &dest)
{
MOZ_ASSERT(dest != ScratchFloatReg);
moveToDoubleLo(operand.payloadReg(), dest);
moveToDoubleHi(operand.typeReg(), dest);
}
void
MacroAssemblerMIPSCompat::unboxDouble(const Address &src, const FloatRegister &dest)
{
ma_lw(ScratchRegister, Address(src.base, src.offset + PAYLOAD_OFFSET));
moveToDoubleLo(ScratchRegister, dest);
ma_lw(ScratchRegister, Address(src.base, src.offset + TAG_OFFSET));
moveToDoubleHi(ScratchRegister, dest);
}
void
MacroAssemblerMIPSCompat::unboxValue(const ValueOperand &src, AnyRegister dest)
{
if (dest.isFloat()) {
Label notInt32, end;
branchTestInt32(Assembler::NotEqual, src, &notInt32);
convertInt32ToDouble(src.payloadReg(), dest.fpu());
ma_b(&end, ShortJump);
bind(&notInt32);
unboxDouble(src, dest.fpu());
bind(&end);
} else if (src.payloadReg() != dest.gpr()) {
ma_move(dest.gpr(), src.payloadReg());
}
}
void
MacroAssemblerMIPSCompat::unboxPrivate(const ValueOperand &src, Register dest)
{
ma_move(dest, src.payloadReg());
}
void
MacroAssemblerMIPSCompat::boxDouble(const FloatRegister &src, const ValueOperand &dest)
{
moveFromDoubleLo(src, dest.payloadReg());
moveFromDoubleHi(src, dest.typeReg());
}
void
MacroAssemblerMIPSCompat::boxNonDouble(JSValueType type, const Register &src,
const ValueOperand &dest)
{
if (src != dest.payloadReg())
ma_move(dest.payloadReg(), src);
ma_li(dest.typeReg(), ImmType(type));
}
void
MacroAssemblerMIPSCompat::boolValueToDouble(const ValueOperand &operand, const FloatRegister &dest)
{
convertBoolToInt32(ScratchRegister, operand.payloadReg());
convertInt32ToDouble(ScratchRegister, dest);
}
void
MacroAssemblerMIPSCompat::int32ValueToDouble(const ValueOperand &operand,
const FloatRegister &dest)
{
convertInt32ToDouble(operand.payloadReg(), dest);
}
void
MacroAssemblerMIPSCompat::loadInt32OrDouble(const Address &src, const FloatRegister &dest)
{
Label notInt32, end;
// If it's an int, convert it to double.
ma_lw(SecondScratchReg, Address(src.base, src.offset + TAG_OFFSET));
branchTestInt32(Assembler::NotEqual, SecondScratchReg, &notInt32);
ma_lw(SecondScratchReg, Address(src.base, src.offset + PAYLOAD_OFFSET));
convertInt32ToDouble(SecondScratchReg, dest);
ma_b(&end, ShortJump);
// Not an int, just load as double.
bind(&notInt32);
ma_ld(dest, src);
bind(&end);
}
void
MacroAssemblerMIPSCompat::loadInt32OrDouble(Register base, Register index,
const FloatRegister &dest, int32_t shift)
{
Label notInt32, end;
// If it's an int, convert it to double.
computeScaledAddress(BaseIndex(base, index, ShiftToScale(shift)), SecondScratchReg);
// Since we only have one scratch, we need to stomp over it with the tag.
load32(Address(SecondScratchReg, TAG_OFFSET), SecondScratchReg);
branchTestInt32(Assembler::NotEqual, SecondScratchReg, &notInt32);
computeScaledAddress(BaseIndex(base, index, ShiftToScale(shift)), SecondScratchReg);
load32(Address(SecondScratchReg, PAYLOAD_OFFSET), SecondScratchReg);
convertInt32ToDouble(SecondScratchReg, dest);
ma_b(&end, ShortJump);
// Not an int, just load as double.
bind(&notInt32);
// First, recompute the offset that had been stored in the scratch register
// since the scratch register was overwritten loading in the type.
computeScaledAddress(BaseIndex(base, index, ShiftToScale(shift)), SecondScratchReg);
loadDouble(Address(SecondScratchReg, 0), dest);
bind(&end);
}
void
MacroAssemblerMIPSCompat::loadConstantDouble(double dp, const FloatRegister &dest)
{
ma_lid(dest, dp);
}
Register
MacroAssemblerMIPSCompat::extractObject(const Address &address, Register scratch)
{
ma_lw(scratch, Address(address.base, address.offset + PAYLOAD_OFFSET));
return scratch;
}
Register
MacroAssemblerMIPSCompat::extractTag(const Address &address, Register scratch)
{
ma_lw(scratch, Address(address.base, address.offset + TAG_OFFSET));
return scratch;
}
Register
MacroAssemblerMIPSCompat::extractTag(const BaseIndex &address, Register scratch)
{
computeScaledAddress(address, scratch);
return extractTag(Address(scratch, address.offset), scratch);
}
uint32_t
MacroAssemblerMIPSCompat::getType(const Value &val)
{
jsval_layout jv = JSVAL_TO_IMPL(val);
return jv.s.tag;
}
void
MacroAssemblerMIPSCompat::moveData(const Value &val, Register data)
{
jsval_layout jv = JSVAL_TO_IMPL(val);
if (val.isMarkable())
ma_li(data, ImmGCPtr(reinterpret_cast<gc::Cell *>(val.toGCThing())));
else
ma_li(data, Imm32(jv.s.payload.i32));
}
void
MacroAssemblerMIPSCompat::moveValue(const Value &val, Register type, Register data)
{
MOZ_ASSERT(type != data);
ma_li(type, Imm32(getType(val)));
moveData(val, data);
}
void
MacroAssemblerMIPSCompat::moveValue(const Value &val, const ValueOperand &dest)
{
moveValue(val, dest.typeReg(), dest.payloadReg());
}
CodeOffsetJump
MacroAssemblerMIPSCompat::jumpWithPatch(RepatchLabel *label)
{
// Only one branch per label.
MOZ_ASSERT(!label->used());
uint32_t dest = label->bound() ? label->offset() : LabelBase::INVALID_OFFSET;
BufferOffset bo = nextOffset();
label->use(bo.getOffset());
addLongJump(bo);
ma_liPatchable(ScratchRegister, Imm32(dest));
as_jr(ScratchRegister);
as_nop();
return CodeOffsetJump(bo.getOffset());
}
/////////////////////////////////////////////////////////////////
// X86/X64-common/ARM/MIPS interface.
/////////////////////////////////////////////////////////////////
void
MacroAssemblerMIPSCompat::storeValue(ValueOperand val, Operand dst)
{
storeValue(val, Address(Register::FromCode(dst.base()), dst.disp()));
}
void
MacroAssemblerMIPSCompat::storeValue(ValueOperand val, const BaseIndex &dest)
{
computeScaledAddress(dest, SecondScratchReg);
storeValue(val, Address(SecondScratchReg, dest.offset));
}
void
MacroAssemblerMIPSCompat::storeValue(JSValueType type, Register reg, BaseIndex dest)
{
computeScaledAddress(dest, ScratchRegister);
// Make sure that ma_sw doesn't clobber ScratchRegister
int32_t offset = dest.offset;
if (!Imm16::isInSignedRange(offset)) {
ma_li(SecondScratchReg, Imm32(offset));
as_addu(ScratchRegister, ScratchRegister, SecondScratchReg);
offset = 0;
}
storeValue(type, reg, Address(ScratchRegister, offset));
}
void
MacroAssemblerMIPSCompat::storeValue(ValueOperand val, const Address &dest)
{
ma_sw(val.payloadReg(), Address(dest.base, dest.offset + PAYLOAD_OFFSET));
ma_sw(val.typeReg(), Address(dest.base, dest.offset + TAG_OFFSET));
}
void
MacroAssemblerMIPSCompat::storeValue(JSValueType type, Register reg, Address dest)
{
MOZ_ASSERT(dest.base != SecondScratchReg);
ma_sw(reg, Address(dest.base, dest.offset + PAYLOAD_OFFSET));
ma_li(SecondScratchReg, ImmTag(JSVAL_TYPE_TO_TAG(type)));
ma_sw(SecondScratchReg, Address(dest.base, dest.offset + TAG_OFFSET));
}
void
MacroAssemblerMIPSCompat::storeValue(const Value &val, Address dest)
{
MOZ_ASSERT(dest.base != SecondScratchReg);
ma_li(SecondScratchReg, Imm32(getType(val)));
ma_sw(SecondScratchReg, Address(dest.base, dest.offset + TAG_OFFSET));
moveData(val, SecondScratchReg);
ma_sw(SecondScratchReg, Address(dest.base, dest.offset + PAYLOAD_OFFSET));
}
void
MacroAssemblerMIPSCompat::storeValue(const Value &val, BaseIndex dest)
{
computeScaledAddress(dest, ScratchRegister);
// Make sure that ma_sw doesn't clobber ScratchRegister
int32_t offset = dest.offset;
if (!Imm16::isInSignedRange(offset)) {
ma_li(SecondScratchReg, Imm32(offset));
as_addu(ScratchRegister, ScratchRegister, SecondScratchReg);
offset = 0;
}
storeValue(val, Address(ScratchRegister, offset));
}
void
MacroAssemblerMIPSCompat::loadValue(const BaseIndex &addr, ValueOperand val)
{
computeScaledAddress(addr, SecondScratchReg);
loadValue(Address(SecondScratchReg, addr.offset), val);
}
void
MacroAssemblerMIPSCompat::loadValue(Address src, ValueOperand val)
{
// Ensure that loading the payload does not erase the pointer to the
// Value in memory.
if (src.base != val.payloadReg()) {
ma_lw(val.payloadReg(), Address(src.base, src.offset + PAYLOAD_OFFSET));
ma_lw(val.typeReg(), Address(src.base, src.offset + TAG_OFFSET));
} else {
ma_lw(val.typeReg(), Address(src.base, src.offset + TAG_OFFSET));
ma_lw(val.payloadReg(), Address(src.base, src.offset + PAYLOAD_OFFSET));
}
}
void
MacroAssemblerMIPSCompat::tagValue(JSValueType type, Register payload, ValueOperand dest)
{
MOZ_ASSERT(payload != dest.typeReg());
ma_li(dest.typeReg(), ImmType(type));
if (payload != dest.payloadReg())
ma_move(dest.payloadReg(), payload);
}
void
MacroAssemblerMIPSCompat::pushValue(ValueOperand val)
{
// Allocate stack slots for type and payload. One for each.
ma_subu(StackPointer, StackPointer, Imm32(sizeof(Value)));
// Store type and payload.
storeValue(val, Address(StackPointer, 0));
}
void
MacroAssemblerMIPSCompat::pushValue(const Address &addr)
{
// Allocate stack slots for type and payload. One for each.
ma_subu(StackPointer, StackPointer, Imm32(sizeof(Value)));
// Store type and payload.
ma_lw(ScratchRegister, Address(addr.base, addr.offset + TAG_OFFSET));
ma_sw(ScratchRegister, Address(StackPointer, TAG_OFFSET));
ma_lw(ScratchRegister, Address(addr.base, addr.offset + PAYLOAD_OFFSET));
ma_sw(ScratchRegister, Address(StackPointer, PAYLOAD_OFFSET));
}
void
MacroAssemblerMIPSCompat::popValue(ValueOperand val)
{
// Load payload and type.
as_lw(val.payloadReg(), StackPointer, PAYLOAD_OFFSET);
as_lw(val.typeReg(), StackPointer, TAG_OFFSET);
// Free stack.
as_addiu(StackPointer, StackPointer, sizeof(Value));
}
void
MacroAssemblerMIPSCompat::storePayload(const Value &val, Address dest)
{
moveData(val, SecondScratchReg);
ma_sw(SecondScratchReg, Address(dest.base, dest.offset + PAYLOAD_OFFSET));
}
void
MacroAssemblerMIPSCompat::storePayload(Register src, Address dest)
{
ma_sw(src, Address(dest.base, dest.offset + PAYLOAD_OFFSET));
return;
}
void
MacroAssemblerMIPSCompat::storePayload(const Value &val, Register base, Register index,
int32_t shift)
{
computeScaledAddress(BaseIndex(base, index, ShiftToScale(shift)), SecondScratchReg);
moveData(val, ScratchRegister);
as_sw(ScratchRegister, SecondScratchReg, NUNBOX32_PAYLOAD_OFFSET);
}
void
MacroAssemblerMIPSCompat::storePayload(Register src, Register base, Register index, int32_t shift)
{
computeScaledAddress(BaseIndex(base, index, ShiftToScale(shift)), SecondScratchReg);
as_sw(src, SecondScratchReg, NUNBOX32_PAYLOAD_OFFSET);
}
void
MacroAssemblerMIPSCompat::storeTypeTag(ImmTag tag, Address dest)
{
ma_li(SecondScratchReg, tag);
ma_sw(SecondScratchReg, Address(dest.base, dest.offset + TAG_OFFSET));
}
void
MacroAssemblerMIPSCompat::storeTypeTag(ImmTag tag, Register base, Register index, int32_t shift)
{
computeScaledAddress(BaseIndex(base, index, ShiftToScale(shift)), SecondScratchReg);
ma_li(ScratchRegister, tag);
as_sw(ScratchRegister, SecondScratchReg, TAG_OFFSET);
}
void
MacroAssemblerMIPSCompat::linkExitFrame()
{
uint8_t *dest = ((uint8_t*)GetIonContext()->compartment->rt) + offsetof(JSRuntime, mainThread.ionTop);
movePtr(ImmWord(dest), ScratchRegister);
ma_sw(StackPointer, Address(ScratchRegister, 0));
}
void
MacroAssemblerMIPSCompat::linkParallelExitFrame(const Register &pt)
{
ma_sw(StackPointer, Address(pt, offsetof(PerThreadData, ionTop)));
}
// This macrosintruction calls the ion code and pushes the return address to
// the stack in the case when stack is alligned.
void
MacroAssemblerMIPS::ma_callIon(const Register r)
{
// This is a MIPS hack to push return address during jalr delay slot.
as_addiu(StackPointer, StackPointer, -2 * sizeof(intptr_t));
as_jalr(r);
as_sw(ra, StackPointer, 0);
}
// This macrosintruction calls the ion code and pushes the return address to
// the stack in the case when stack is not alligned.
void
MacroAssemblerMIPS::ma_callIonHalfPush(const Register r)
{
// This is a MIPS hack to push return address during jalr delay slot.
as_addiu(StackPointer, StackPointer, -sizeof(intptr_t));
as_jalr(r);
as_sw(ra, StackPointer, 0);
}
void
MacroAssemblerMIPS::ma_call(void* dest)
{
MOZ_STATIC_ASSERT(sizeof(void*) == 4, "Must be 32 bit arch.");
ma_liPatchable(CallReg, Imm32(reinterpret_cast<uint32_t>(dest)));
as_jalr(CallReg);
as_nop();
}
void
MacroAssemblerMIPSCompat::breakpoint()
{
as_break(0);
}
void
MacroAssemblerMIPSCompat::ensureDouble(const ValueOperand &source, FloatRegister dest,
Label *failure)
{
Label isDouble, done;
branchTestDouble(Assembler::Equal, source.typeReg(), &isDouble);
branchTestInt32(Assembler::NotEqual, source.typeReg(), failure);
convertInt32ToDouble(source.payloadReg(), dest);
jump(&done);
bind(&isDouble);
unboxDouble(source, dest);
bind(&done);
}
void
MacroAssemblerMIPSCompat::setupABICall(uint32_t args)
{
MOZ_ASSERT(!inCall_);
inCall_ = true;
args_ = args;
passedArgs_ = 0;
usedArgSlots_ = 0;
firstArgType = GENERAL;
}
void
MacroAssemblerMIPSCompat::setupAlignedABICall(uint32_t args)
{
setupABICall(args);
dynamicAlignment_ = false;
}
void
MacroAssemblerMIPSCompat::setupUnalignedABICall(uint32_t args, const Register &scratch)
{
setupABICall(args);
dynamicAlignment_ = true;
ma_move(scratch, StackPointer);
// Force sp to be aligned
ma_subu(StackPointer, StackPointer, Imm32(sizeof(uint32_t)));
ma_and(StackPointer, StackPointer, Imm32(~(StackAlignment - 1)));
as_sw(scratch, StackPointer, 0);
}
void MacroAssemblerMIPSCompat::checkStackAlignment()
{
#ifdef DEBUG
Label aligned;
as_andi(ScratchRegister, sp, StackAlignment - 1);
ma_b(ScratchRegister, zero, &aligned, Equal, ShortJump);
as_break(MAX_BREAK_CODE);
bind(&aligned);
#endif
}
void
MacroAssemblerMIPSCompat::alignPointerUp(Register src, Register dest, uint32_t alignment)
{
MOZ_ASSERT(alignment > 1);
ma_addu(dest, src, Imm32(alignment - 1));
ma_and(dest, dest, Imm32(~(alignment - 1)));
}
void
MacroAssemblerMIPSCompat::callWithABIPre(uint32_t *stackAdjust)
{
MOZ_ASSERT(inCall_);
// Reserve place for $ra.
*stackAdjust = sizeof(intptr_t);
*stackAdjust += usedArgSlots_ > NumIntArgRegs ?
usedArgSlots_ * sizeof(intptr_t) :
NumIntArgRegs * sizeof(intptr_t);
if (dynamicAlignment_) {
*stackAdjust += ComputeByteAlignment(*stackAdjust, StackAlignment);
} else {
*stackAdjust += ComputeByteAlignment(framePushed_ + *stackAdjust, StackAlignment);
}
reserveStack(*stackAdjust);
// Save $ra because call is going to clobber it. Restore it in
// callWithABIPost. NOTE: This is needed for calls from BaselineIC.
// Maybe we can do this differently.
ma_sw(ra, Address(StackPointer, *stackAdjust - sizeof(intptr_t)));
// Position all arguments.
{
enoughMemory_ = enoughMemory_ && moveResolver_.resolve();
if (!enoughMemory_)
return;
MoveEmitter emitter(*this);
emitter.emit(moveResolver_);
emitter.finish();
}
checkStackAlignment();
}
void
MacroAssemblerMIPSCompat::callWithABIPost(uint32_t stackAdjust, Result result)
{
// Restore ra value (as stored in callWithABIPre()).
ma_lw(ra, Address(StackPointer, stackAdjust - sizeof(intptr_t)));
if (dynamicAlignment_) {
// Restore sp value from stack (as stored in setupUnalignedABICall()).
ma_lw(StackPointer, Address(StackPointer, stackAdjust));
// Use adjustFrame instead of freeStack because we already restored sp.
adjustFrame(-stackAdjust);
} else {
freeStack(stackAdjust);
}
MOZ_ASSERT(inCall_);
inCall_ = false;
}
void
MacroAssemblerMIPSCompat::callWithABI(void *fun, Result result)
{
uint32_t stackAdjust;
callWithABIPre(&stackAdjust);
ma_call(fun);
callWithABIPost(stackAdjust, result);
}
void
MacroAssemblerMIPSCompat::callWithABI(const Address &fun, Result result)
{
// Load the callee in t9, no instruction between the lw and call
// should clobber it. Note that we can't use fun.base because it may
// be one of the IntArg registers clobbered before the call.
ma_lw(t9, Address(fun.base, fun.offset));
uint32_t stackAdjust;
callWithABIPre(&stackAdjust);
call(t9);
callWithABIPost(stackAdjust, result);
}
CodeOffsetLabel
MacroAssemblerMIPSCompat::toggledJump(Label *label)
{
CodeOffsetLabel ret(nextOffset().getOffset());
ma_b(label);
return ret;
}
CodeOffsetLabel
MacroAssemblerMIPSCompat::toggledCall(IonCode *target, bool enabled)
{
BufferOffset bo = nextOffset();
CodeOffsetLabel offset(bo.getOffset());
addPendingJump(bo, target->raw(), Relocation::IONCODE);
ma_liPatchable(ScratchRegister, ImmWord(target->raw()));
if (enabled) {
as_jalr(ScratchRegister);
as_nop();
} else {
as_nop();
as_nop();
}
MOZ_ASSERT(nextOffset().getOffset() - offset.offset() == ToggledCallSize());
return offset;
}
void MacroAssemblerMIPSCompat::call(const Register reg) {
as_jalr(reg);
as_nop();
}
void MacroAssemblerMIPSCompat::call(Label *label) {
ma_bal(label);
}
void MacroAssemblerMIPSCompat::call(ImmWord imm) {
// call(imm.asPointer());
// 31 version converts this to an ImmPtr and then calls that.
BufferOffset bo = m_buffer.nextOffset();
// addPendingJump(bo, imm, Relocation::HARDCODED);
addPendingJump(bo, imm.asPointer(), Relocation::HARDCODED);
ma_call(imm.asPointer());
}
void
MacroAssemblerMIPSCompat::call(IonCode *c) {
BufferOffset bo = m_buffer.nextOffset();
addPendingJump(bo, c->raw(), Relocation::IONCODE);
ma_liPatchable(ScratchRegister, Imm32((uint32_t)c->raw()));
ma_callIonHalfPush(ScratchRegister);
}
void
MacroAssemblerMIPSCompat::storePtr(Register src, const BaseIndex& address)
{
ma_store(src, address, SizeWord);
}
void
MacroAssemblerMIPSCompat::testUndefinedSet(Condition cond, const ValueOperand& value, Register dest)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
ma_cmp_set(dest, value.typeReg(), ImmType(JSVAL_TYPE_UNDEFINED), cond);
}
void
MacroAssemblerMIPSCompat::testNullSet(Condition cond, const ValueOperand& value, Register dest)
{
MOZ_ASSERT(cond == Equal || cond == NotEqual);
ma_cmp_set(dest, value.typeReg(), ImmType(JSVAL_TYPE_NULL), cond);
}
void
MacroAssemblerMIPSCompat::PopRegsInMaskIgnore(RegisterSet set, RegisterSet ignore)
{
// int32_t diffG = set.gprs().size() * sizeof(intptr_t);
int32_t diffG = set.gprs().size() * STACK_SLOT_SIZE;
int32_t diffF = set.fpus().size() * sizeof(double);
const int32_t reservedG = diffG;
const int32_t reservedF = diffF;
// Read the buffer form the first aligned location.
ma_addu(SecondScratchReg, sp, Imm32(reservedF + sizeof(double)));
ma_and(SecondScratchReg, SecondScratchReg, Imm32(~(StackAlignment - 1)));
for (FloatRegisterIterator iter(set.fpus()); iter.more(); iter++) {
// :TODO: (Bug 972836) Fix this once odd regs can be used as
// float32 only. For now we skip loading odd regs for O32 ABI.
// :TODO: (Bug 985881) Make a switch for N32 ABI.
if (!ignore.has(*iter) && ((*iter).code() % 2 == 0)) {
// Use assembly l.d because we have alligned the stack.
as_ld(*iter, SecondScratchReg, -diffF);
}
diffF -= sizeof(double);
}
freeStack(reservedF + sizeof(double));
MOZ_ASSERT(diffF == 0);
for (GeneralRegisterIterator iter(set.gprs()); iter.more(); iter++) {
diffG -= STACK_SLOT_SIZE;
if (!ignore.has(*iter)) {
loadPtr(Address(StackPointer, diffG), *iter);
}
}
freeStack(reservedG);
MOZ_ASSERT(diffG == 0);
}
void
MacroAssemblerMIPSCompat::passABIArg(const MoveOperand &from)
{
++passedArgs_;
if (!enoughMemory_)
{
return;
}
if (from.isDouble())
{
Move::Kind type = Move::Kind::DOUBLE;
if (!usedArgSlots_)
{
if (from.floatReg() != f12)
{
enoughMemory_ = moveResolver_.addMove(from, MoveOperand(f12), type);
}
usedArgSlots_ = 2;
firstArgType = DOUBLE;
}
else if (usedArgSlots_ <= 2)
{
if ((usedArgSlots_ == 1 && false) ||
(usedArgSlots_ == 2 && firstArgType == DOUBLE))
{
if (from.floatReg() != f14)
{
enoughMemory_ = moveResolver_.addMove(from, MoveOperand(f14), type);
}
}
else
{
// Create two moves so that cycles are found. Move emitter
// will have special case to handle this.
enoughMemory_ = moveResolver_.addMove(from, MoveOperand(a2), type);
enoughMemory_ = moveResolver_.addMove(from, MoveOperand(a3), type);
}
usedArgSlots_ = 4;
}
else
{
// Align if necessary
usedArgSlots_ += usedArgSlots_ % 2;
uint32_t disp = GetArgStackDisp(usedArgSlots_);
enoughMemory_ = moveResolver_.addMove(from, MoveOperand(sp, disp), type);
usedArgSlots_ += 2;
}
}
else
{
Move::Kind type = Move::Kind::GENERAL;
Register destReg;
if (GetIntArgReg(usedArgSlots_, &destReg))
{
if (from.isGeneralReg() && from.reg() == destReg)
{
// Nothing to do. Value is in the right register already
}
else
{
enoughMemory_ = moveResolver_.addMove(from, MoveOperand(destReg), type);
}
}
else
{
uint32_t disp = GetArgStackDisp(usedArgSlots_);
enoughMemory_ = moveResolver_.addMove(from, MoveOperand(sp, disp), type);
}
usedArgSlots_++;
}
}
void
MacroAssemblerMIPSCompat::passABIArg(const Register &reg)
{
passABIArg(MoveOperand(reg));
}
void
MacroAssemblerMIPSCompat::passABIArg(const FloatRegister &freg)
{
passABIArg(MoveOperand(freg));
}
void
MacroAssemblerMIPSCompat::handleFailureWithHandler(void *handler)
{
// Reserve space for exception information.
int size = (sizeof(ResumeFromException) + StackAlignment) & ~(StackAlignment - 1);
ma_subu(StackPointer, StackPointer, Imm32(size));
ma_move(a0, StackPointer); // Use a0 since it is a first function argument
// Ask for an exception handler.
setupUnalignedABICall(1, a1);
passABIArg(a0);
callWithABI(handler);
IonCode* excTail = GetIonContext()->runtime->ionRuntime()->getExceptionTail();
branch(excTail);
}
void
MacroAssemblerMIPSCompat::handleFailureWithHandlerTail()
{
Label entryFrame;
Label catch_;
Label finally;
Label return_;
Label bailout;
// Already clobbered a0, so use it...
ma_lw(a0, Address(StackPointer, offsetof(ResumeFromException, kind)));
branch32(Assembler::Equal, a0, Imm32(ResumeFromException::RESUME_ENTRY_FRAME), &entryFrame);
branch32(Assembler::Equal, a0, Imm32(ResumeFromException::RESUME_CATCH), &catch_);
branch32(Assembler::Equal, a0, Imm32(ResumeFromException::RESUME_FINALLY), &finally);
branch32(Assembler::Equal, a0, Imm32(ResumeFromException::RESUME_FORCED_RETURN), &return_);
// branch32(Assembler::Equal, a0, Imm32(ResumeFromException::RESUME_BAILOUT), &bailout);
breakpoint(); // Invalid kind.
// No exception handler. Load the error value, load the new stack pointer
// and return from the entry frame.
bind(&entryFrame);
moveValue(MagicValue(JS_ION_ERROR), JSReturnOperand);
ma_lw(StackPointer, Address(StackPointer, offsetof(ResumeFromException, stackPointer)));
// We're going to be returning by the ion calling convention
ma_pop(ra);
as_jr(ra);
as_nop();
// If we found a catch handler, this must be a baseline frame. Restore
// state and jump to the catch block.
bind(&catch_);
ma_lw(a0, Address(StackPointer, offsetof(ResumeFromException, target)));
ma_lw(BaselineFrameReg, Address(StackPointer, offsetof(ResumeFromException, framePointer)));
ma_lw(StackPointer, Address(StackPointer, offsetof(ResumeFromException, stackPointer)));
jump(a0);
// If we found a finally block, this must be a baseline frame. Push
// two values expected by JSOP_RETSUB: BooleanValue(true) and the
// exception.
bind(&finally);
ValueOperand exception = ValueOperand(a1, a2);
loadValue(Address(sp, offsetof(ResumeFromException, exception)), exception);
ma_lw(a0, Address(sp, offsetof(ResumeFromException, target)));
ma_lw(BaselineFrameReg, Address(sp, offsetof(ResumeFromException, framePointer)));
ma_lw(sp, Address(sp, offsetof(ResumeFromException, stackPointer)));
pushValue(BooleanValue(true));
pushValue(exception);
jump(a0);
// Only used in debug mode. Return BaselineFrame->returnValue() to the
// caller.
bind(&return_);
ma_lw(BaselineFrameReg, Address(StackPointer, offsetof(ResumeFromException, framePointer)));
ma_lw(StackPointer, Address(StackPointer, offsetof(ResumeFromException, stackPointer)));
loadValue(Address(BaselineFrameReg, BaselineFrame::reverseOffsetOfReturnValue()),
JSReturnOperand);
ma_move(StackPointer, BaselineFrameReg);
pop(BaselineFrameReg);
ret();
// If we are bailing out to baseline to handle an excep