blob: 2da80a1fbb53e397893f3363f63f6a4dac29ded3 [file] [log] [blame]
/* -*- Mode: C++; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 4 -*-
* vim: set ts=8 sts=4 et sw=4 tw=99:
* This Source Code Form is subject to the terms of the Mozilla Public
* License, v. 2.0. If a copy of the MPL was not distributed with this
* file, You can obtain one at http://mozilla.org/MPL/2.0/. */
#include "jit/mips-shared/MacroAssembler-mips-shared.h"
// Unified leak fix:
#include "jit/MacroAssembler.h"
#include "jit/MacroAssembler-inl.h"
#include "jit/mips-shared/MacroAssembler-mips-shared-inl.h"
using namespace js;
using namespace jit;
void
MacroAssemblerMIPSShared::ma_move(Register rd, Register rs)
{
as_or(rd, rs, zero);
}
void
MacroAssemblerMIPSShared::ma_li(Register dest, ImmGCPtr ptr)
{
writeDataRelocation(ptr);
asMasm().ma_liPatchable(dest, ImmPtr(ptr.value));
}
void
MacroAssemblerMIPSShared::ma_li(Register dest, Imm32 imm)
{
if (Imm16::IsInSignedRange(imm.value)) {
as_addiu(dest, zero, imm.value);
} else if (Imm16::IsInUnsignedRange(imm.value)) {
as_ori(dest, zero, Imm16::Lower(imm).encode());
} else if (Imm16::Lower(imm).encode() == 0) {
as_lui(dest, Imm16::Upper(imm).encode());
} else {
as_lui(dest, Imm16::Upper(imm).encode());
as_ori(dest, dest, Imm16::Lower(imm).encode());
}
}
// Shifts
void
MacroAssemblerMIPSShared::ma_sll(Register rd, Register rt, Imm32 shift)
{
as_sll(rd, rt, shift.value % 32);
}
void
MacroAssemblerMIPSShared::ma_srl(Register rd, Register rt, Imm32 shift)
{
as_srl(rd, rt, shift.value % 32);
}
void
MacroAssemblerMIPSShared::ma_sra(Register rd, Register rt, Imm32 shift)
{
as_sra(rd, rt, shift.value % 32);
}
void
MacroAssemblerMIPSShared::ma_ror(Register rd, Register rt, Imm32 shift)
{
as_rotr(rd, rt, shift.value % 32);
}
void
MacroAssemblerMIPSShared::ma_rol(Register rd, Register rt, Imm32 shift)
{
as_rotr(rd, rt, 32 - (shift.value % 32));
}
void
MacroAssemblerMIPSShared::ma_sll(Register rd, Register rt, Register shift)
{
as_sllv(rd, rt, shift);
}
void
MacroAssemblerMIPSShared::ma_srl(Register rd, Register rt, Register shift)
{
as_srlv(rd, rt, shift);
}
void
MacroAssemblerMIPSShared::ma_sra(Register rd, Register rt, Register shift)
{
as_srav(rd, rt, shift);
}
void
MacroAssemblerMIPSShared::ma_ror(Register rd, Register rt, Register shift)
{
as_rotrv(rd, rt, shift);
}
void
MacroAssemblerMIPSShared::ma_rol(Register rd, Register rt, Register shift)
{
ma_negu(ScratchRegister, shift);
as_rotrv(rd, rt, ScratchRegister);
}
void
MacroAssemblerMIPSShared::ma_negu(Register rd, Register rs)
{
as_subu(rd, zero, rs);
}
void
MacroAssemblerMIPSShared::ma_not(Register rd, Register rs)
{
as_nor(rd, rs, zero);
}
// And.
void
MacroAssemblerMIPSShared::ma_and(Register rd, Register rs)
{
as_and(rd, rd, rs);
}
void
MacroAssemblerMIPSShared::ma_and(Register rd, Imm32 imm)
{
ma_and(rd, rd, imm);
}
void
MacroAssemblerMIPSShared::ma_and(Register rd, Register rs, Imm32 imm)
{
if (Imm16::IsInUnsignedRange(imm.value)) {
as_andi(rd, rs, imm.value);
} else {
ma_li(ScratchRegister, imm);
as_and(rd, rs, ScratchRegister);
}
}
// Or.
void
MacroAssemblerMIPSShared::ma_or(Register rd, Register rs)
{
as_or(rd, rd, rs);
}
void
MacroAssemblerMIPSShared::ma_or(Register rd, Imm32 imm)
{
ma_or(rd, rd, imm);
}
void
MacroAssemblerMIPSShared::ma_or(Register rd, Register rs, Imm32 imm)
{
if (Imm16::IsInUnsignedRange(imm.value)) {
as_ori(rd, rs, imm.value);
} else {
ma_li(ScratchRegister, imm);
as_or(rd, rs, ScratchRegister);
}
}
// xor
void
MacroAssemblerMIPSShared::ma_xor(Register rd, Register rs)
{
as_xor(rd, rd, rs);
}
void
MacroAssemblerMIPSShared::ma_xor(Register rd, Imm32 imm)
{
ma_xor(rd, rd, imm);
}
void
MacroAssemblerMIPSShared::ma_xor(Register rd, Register rs, Imm32 imm)
{
if (Imm16::IsInUnsignedRange(imm.value)) {
as_xori(rd, rs, imm.value);
} else {
ma_li(ScratchRegister, imm);
as_xor(rd, rs, ScratchRegister);
}
}
// Arithmetic-based ops.
// Add.
void
MacroAssemblerMIPSShared::ma_addu(Register rd, Register rs, Imm32 imm)
{
if (Imm16::IsInSignedRange(imm.value)) {
as_addiu(rd, rs, imm.value);
} else {
ma_li(ScratchRegister, imm);
as_addu(rd, rs, ScratchRegister);
}
}
void
MacroAssemblerMIPSShared::ma_addu(Register rd, Register rs)
{
as_addu(rd, rd, rs);
}
void
MacroAssemblerMIPSShared::ma_addu(Register rd, Imm32 imm)
{
ma_addu(rd, rd, imm);
}
// Subtract.
void
MacroAssemblerMIPSShared::ma_subu(Register rd, Register rs, Imm32 imm)
{
if (Imm16::IsInSignedRange(-imm.value)) {
as_addiu(rd, rs, -imm.value);
} else {
ma_li(ScratchRegister, imm);
as_subu(rd, rs, ScratchRegister);
}
}
void
MacroAssemblerMIPSShared::ma_subu(Register rd, Imm32 imm)
{
ma_subu(rd, rd, imm);
}
void
MacroAssemblerMIPSShared::ma_subu(Register rd, Register rs)
{
as_subu(rd, rd, rs);
}
void
MacroAssemblerMIPSShared::ma_subTestOverflow(Register rd, Register rs, Imm32 imm, Label* overflow)
{
if (imm.value != INT32_MIN) {
asMasm().ma_addTestOverflow(rd, rs, Imm32(-imm.value), overflow);
} else {
ma_li(ScratchRegister, Imm32(imm.value));
asMasm().ma_subTestOverflow(rd, rs, ScratchRegister, overflow);
}
}
void
MacroAssemblerMIPSShared::ma_mult(Register rs, Imm32 imm)
{
ma_li(ScratchRegister, imm);
as_mult(rs, ScratchRegister);
}
void
MacroAssemblerMIPSShared::ma_mul_branch_overflow(Register rd, Register rs, Register rt, Label* overflow)
{
as_mult(rs, rt);
as_mflo(rd);
as_sra(ScratchRegister, rd, 31);
as_mfhi(SecondScratchReg);
ma_b(ScratchRegister, SecondScratchReg, overflow, Assembler::NotEqual);
}
void
MacroAssemblerMIPSShared::ma_mul_branch_overflow(Register rd, Register rs, Imm32 imm, Label* overflow)
{
ma_li(ScratchRegister, imm);
ma_mul_branch_overflow(rd, rs, ScratchRegister, overflow);
}
void
MacroAssemblerMIPSShared::ma_div_branch_overflow(Register rd, Register rs, Register rt, Label* overflow)
{
as_div(rs, rt);
as_mflo(rd);
as_mfhi(ScratchRegister);
ma_b(ScratchRegister, ScratchRegister, overflow, Assembler::NonZero);
}
void
MacroAssemblerMIPSShared::ma_div_branch_overflow(Register rd, Register rs, Imm32 imm, Label* overflow)
{
ma_li(ScratchRegister, imm);
ma_div_branch_overflow(rd, rs, ScratchRegister, overflow);
}
void
MacroAssemblerMIPSShared::ma_mod_mask(Register src, Register dest, Register hold, Register remain,
int32_t shift, Label* negZero)
{
// MATH:
// We wish to compute x % (1<<y) - 1 for a known constant, y.
// First, let b = (1<<y) and C = (1<<y)-1, then think of the 32 bit
// dividend as a number in base b, namely
// c_0*1 + c_1*b + c_2*b^2 ... c_n*b^n
// now, since both addition and multiplication commute with modulus,
// x % C == (c_0 + c_1*b + ... + c_n*b^n) % C ==
// (c_0 % C) + (c_1%C) * (b % C) + (c_2 % C) * (b^2 % C)...
// now, since b == C + 1, b % C == 1, and b^n % C == 1
// this means that the whole thing simplifies to:
// c_0 + c_1 + c_2 ... c_n % C
// each c_n can easily be computed by a shift/bitextract, and the modulus
// can be maintained by simply subtracting by C whenever the number gets
// over C.
int32_t mask = (1 << shift) - 1;
Label head, negative, sumSigned, done;
// hold holds -1 if the value was negative, 1 otherwise.
// remain holds the remaining bits that have not been processed
// SecondScratchReg serves as a temporary location to store extracted bits
// into as well as holding the trial subtraction as a temp value dest is
// the accumulator (and holds the final result)
// move the whole value into the remain.
ma_move(remain, src);
// Zero out the dest.
ma_li(dest, Imm32(0));
// Set the hold appropriately.
ma_b(remain, remain, &negative, Signed, ShortJump);
ma_li(hold, Imm32(1));
ma_b(&head, ShortJump);
bind(&negative);
ma_li(hold, Imm32(-1));
ma_negu(remain, remain);
// Begin the main loop.
bind(&head);
// Extract the bottom bits into SecondScratchReg.
ma_and(SecondScratchReg, remain, Imm32(mask));
// Add those bits to the accumulator.
as_addu(dest, dest, SecondScratchReg);
// Do a trial subtraction
ma_subu(SecondScratchReg, dest, Imm32(mask));
// If (sum - C) > 0, store sum - C back into sum, thus performing a
// modulus.
ma_b(SecondScratchReg, SecondScratchReg, &sumSigned, Signed, ShortJump);
ma_move(dest, SecondScratchReg);
bind(&sumSigned);
// Get rid of the bits that we extracted before.
as_srl(remain, remain, shift);
// If the shift produced zero, finish, otherwise, continue in the loop.
ma_b(remain, remain, &head, NonZero, ShortJump);
// Check the hold to see if we need to negate the result.
ma_b(hold, hold, &done, NotSigned, ShortJump);
// If the hold was non-zero, negate the result to be in line with
// what JS wants
if (negZero != nullptr) {
// Jump out in case of negative zero.
ma_b(hold, hold, negZero, Zero);
ma_negu(dest, dest);
} else {
ma_negu(dest, dest);
}
bind(&done);
}
// Memory.
void
MacroAssemblerMIPSShared::ma_load(Register dest, const BaseIndex& src,
LoadStoreSize size, LoadStoreExtension extension)
{
asMasm().computeScaledAddress(src, SecondScratchReg);
asMasm().ma_load(dest, Address(SecondScratchReg, src.offset), size, extension);
}
void
MacroAssemblerMIPSShared::ma_store(Register data, const BaseIndex& dest,
LoadStoreSize size, LoadStoreExtension extension)
{
asMasm().computeScaledAddress(dest, SecondScratchReg);
asMasm().ma_store(data, Address(SecondScratchReg, dest.offset), size, extension);
}
void
MacroAssemblerMIPSShared::ma_store(Imm32 imm, const BaseIndex& dest,
LoadStoreSize size, LoadStoreExtension extension)
{
// Make sure that SecondScratchReg contains absolute address so that
// offset is 0.
asMasm().computeEffectiveAddress(dest, SecondScratchReg);
// Scrach register is free now, use it for loading imm value
ma_li(ScratchRegister, imm);
// with offset=0 ScratchRegister will not be used in ma_store()
// so we can use it as a parameter here
asMasm().ma_store(ScratchRegister, Address(SecondScratchReg, 0), size, extension);
}
// Branches when done from within mips-specific code.
void
MacroAssemblerMIPSShared::ma_b(Register lhs, Register rhs, Label* label, Condition c, JumpKind jumpKind)
{
switch (c) {
case Equal :
case NotEqual:
asMasm().branchWithCode(getBranchCode(lhs, rhs, c), label, jumpKind);
break;
case Always:
ma_b(label, jumpKind);
break;
case Zero:
case NonZero:
case Signed:
case NotSigned:
MOZ_ASSERT(lhs == rhs);
asMasm().branchWithCode(getBranchCode(lhs, c), label, jumpKind);
break;
default:
Condition cond = ma_cmp(ScratchRegister, lhs, rhs, c);
asMasm().branchWithCode(getBranchCode(ScratchRegister, cond), label, jumpKind);
break;
}
}
void
MacroAssemblerMIPSShared::ma_b(Register lhs, Imm32 imm, Label* label, Condition c, JumpKind jumpKind)
{
MOZ_ASSERT(c != Overflow);
if (imm.value == 0) {
if (c == Always || c == AboveOrEqual)
ma_b(label, jumpKind);
else if (c == Below)
; // This condition is always false. No branch required.
else
asMasm().branchWithCode(getBranchCode(lhs, c), label, jumpKind);
} else {
MOZ_ASSERT(lhs != ScratchRegister);
ma_li(ScratchRegister, imm);
ma_b(lhs, ScratchRegister, label, c, jumpKind);
}
}
void
MacroAssemblerMIPSShared::ma_b(Register lhs, ImmPtr imm, Label* l, Condition c, JumpKind jumpKind)
{
asMasm().ma_b(lhs, ImmWord(uintptr_t(imm.value)), l, c, jumpKind);
}
void
MacroAssemblerMIPSShared::ma_b(Label* label, JumpKind jumpKind)
{
asMasm().branchWithCode(getBranchCode(BranchIsJump), label, jumpKind);
}
Assembler::Condition
MacroAssemblerMIPSShared::ma_cmp(Register scratch, Register lhs, Register rhs, Condition c)
{
switch (c) {
case Above:
// bgtu s,t,label =>
// sltu at,t,s
// bne at,$zero,offs
as_sltu(scratch, rhs, lhs);
return NotEqual;
case AboveOrEqual:
// bgeu s,t,label =>
// sltu at,s,t
// beq at,$zero,offs
as_sltu(scratch, lhs, rhs);
return Equal;
case Below:
// bltu s,t,label =>
// sltu at,s,t
// bne at,$zero,offs
as_sltu(scratch, lhs, rhs);
return NotEqual;
case BelowOrEqual:
// bleu s,t,label =>
// sltu at,t,s
// beq at,$zero,offs
as_sltu(scratch, rhs, lhs);
return Equal;
case GreaterThan:
// bgt s,t,label =>
// slt at,t,s
// bne at,$zero,offs
as_slt(scratch, rhs, lhs);
return NotEqual;
case GreaterThanOrEqual:
// bge s,t,label =>
// slt at,s,t
// beq at,$zero,offs
as_slt(scratch, lhs, rhs);
return Equal;
case LessThan:
// blt s,t,label =>
// slt at,s,t
// bne at,$zero,offs
as_slt(scratch, lhs, rhs);
return NotEqual;
case LessThanOrEqual:
// ble s,t,label =>
// slt at,t,s
// beq at,$zero,offs
as_slt(scratch, rhs, lhs);
return Equal;
case Equal :
case NotEqual:
case Zero:
case NonZero:
case Always:
case Signed:
case NotSigned:
MOZ_CRASH("There is a better way to compare for equality.");
break;
case Overflow:
MOZ_CRASH("Overflow condition not supported for MIPS.");
break;
default:
MOZ_CRASH("Invalid condition for branch.");
}
return Always;
}
void
MacroAssemblerMIPSShared::ma_cmp_set(Register rd, Register rs, Register rt, Condition c)
{
switch (c) {
case Equal :
// seq d,s,t =>
// xor d,s,t
// sltiu d,d,1
as_xor(rd, rs, rt);
as_sltiu(rd, rd, 1);
break;
case NotEqual:
// sne d,s,t =>
// xor d,s,t
// sltu d,$zero,d
as_xor(rd, rs, rt);
as_sltu(rd, zero, rd);
break;
case Above:
// sgtu d,s,t =>
// sltu d,t,s
as_sltu(rd, rt, rs);
break;
case AboveOrEqual:
// sgeu d,s,t =>
// sltu d,s,t
// xori d,d,1
as_sltu(rd, rs, rt);
as_xori(rd, rd, 1);
break;
case Below:
// sltu d,s,t
as_sltu(rd, rs, rt);
break;
case BelowOrEqual:
// sleu d,s,t =>
// sltu d,t,s
// xori d,d,1
as_sltu(rd, rt, rs);
as_xori(rd, rd, 1);
break;
case GreaterThan:
// sgt d,s,t =>
// slt d,t,s
as_slt(rd, rt, rs);
break;
case GreaterThanOrEqual:
// sge d,s,t =>
// slt d,s,t
// xori d,d,1
as_slt(rd, rs, rt);
as_xori(rd, rd, 1);
break;
case LessThan:
// slt d,s,t
as_slt(rd, rs, rt);
break;
case LessThanOrEqual:
// sle d,s,t =>
// slt d,t,s
// xori d,d,1
as_slt(rd, rt, rs);
as_xori(rd, rd, 1);
break;
case Zero:
MOZ_ASSERT(rs == rt);
// seq d,s,$zero =>
// xor d,s,$zero
// sltiu d,d,1
as_xor(rd, rs, zero);
as_sltiu(rd, rd, 1);
break;
case NonZero:
// sne d,s,$zero =>
// xor d,s,$zero
// sltu d,$zero,d
as_xor(rd, rs, zero);
as_sltu(rd, zero, rd);
break;
case Signed:
as_slt(rd, rs, zero);
break;
case NotSigned:
// sge d,s,$zero =>
// slt d,s,$zero
// xori d,d,1
as_slt(rd, rs, zero);
as_xori(rd, rd, 1);
break;
default:
MOZ_CRASH("Invalid condition for ma_cmp_set.");
}
}
void
MacroAssemblerMIPSShared::compareFloatingPoint(FloatFormat fmt, FloatRegister lhs, FloatRegister rhs,
DoubleCondition c, FloatTestKind* testKind,
FPConditionBit fcc)
{
switch (c) {
case DoubleOrdered:
as_cun(fmt, lhs, rhs, fcc);
*testKind = TestForFalse;
break;
case DoubleEqual:
as_ceq(fmt, lhs, rhs, fcc);
*testKind = TestForTrue;
break;
case DoubleNotEqual:
as_cueq(fmt, lhs, rhs, fcc);
*testKind = TestForFalse;
break;
case DoubleGreaterThan:
as_colt(fmt, rhs, lhs, fcc);
*testKind = TestForTrue;
break;
case DoubleGreaterThanOrEqual:
as_cole(fmt, rhs, lhs, fcc);
*testKind = TestForTrue;
break;
case DoubleLessThan:
as_colt(fmt, lhs, rhs, fcc);
*testKind = TestForTrue;
break;
case DoubleLessThanOrEqual:
as_cole(fmt, lhs, rhs, fcc);
*testKind = TestForTrue;
break;
case DoubleUnordered:
as_cun(fmt, lhs, rhs, fcc);
*testKind = TestForTrue;
break;
case DoubleEqualOrUnordered:
as_cueq(fmt, lhs, rhs, fcc);
*testKind = TestForTrue;
break;
case DoubleNotEqualOrUnordered:
as_ceq(fmt, lhs, rhs, fcc);
*testKind = TestForFalse;
break;
case DoubleGreaterThanOrUnordered:
as_cult(fmt, rhs, lhs, fcc);
*testKind = TestForTrue;
break;
case DoubleGreaterThanOrEqualOrUnordered:
as_cule(fmt, rhs, lhs, fcc);
*testKind = TestForTrue;
break;
case DoubleLessThanOrUnordered:
as_cult(fmt, lhs, rhs, fcc);
*testKind = TestForTrue;
break;
case DoubleLessThanOrEqualOrUnordered:
as_cule(fmt, lhs, rhs, fcc);
*testKind = TestForTrue;
break;
default:
MOZ_CRASH("Invalid DoubleCondition.");
}
}
void
MacroAssemblerMIPSShared::ma_cmp_set_double(Register dest, FloatRegister lhs, FloatRegister rhs,
DoubleCondition c)
{
ma_li(dest, Imm32(0));
ma_li(ScratchRegister, Imm32(1));
FloatTestKind moveCondition;
compareFloatingPoint(DoubleFloat, lhs, rhs, c, &moveCondition);
if (moveCondition == TestForTrue)
as_movt(dest, ScratchRegister);
else
as_movf(dest, ScratchRegister);
}
void
MacroAssemblerMIPSShared::ma_cmp_set_float32(Register dest, FloatRegister lhs, FloatRegister rhs,
DoubleCondition c)
{
ma_li(dest, Imm32(0));
ma_li(ScratchRegister, Imm32(1));
FloatTestKind moveCondition;
compareFloatingPoint(SingleFloat, lhs, rhs, c, &moveCondition);
if (moveCondition == TestForTrue)
as_movt(dest, ScratchRegister);
else
as_movf(dest, ScratchRegister);
}
void
MacroAssemblerMIPSShared::ma_cmp_set(Register rd, Register rs, Imm32 imm, Condition c)
{
ma_li(ScratchRegister, imm);
ma_cmp_set(rd, rs, ScratchRegister, c);
}
// fp instructions
void
MacroAssemblerMIPSShared::ma_lis(FloatRegister dest, float value)
{
Imm32 imm(mozilla::BitwiseCast<uint32_t>(value));
ma_li(ScratchRegister, imm);
moveToFloat32(ScratchRegister, dest);
}
void
MacroAssemblerMIPSShared::ma_liNegZero(FloatRegister dest)
{
moveToDoubleLo(zero, dest);
ma_li(ScratchRegister, Imm32(INT_MIN));
asMasm().moveToDoubleHi(ScratchRegister, dest);
}
void
MacroAssemblerMIPSShared::ma_sd(FloatRegister ft, BaseIndex address)
{
asMasm().computeScaledAddress(address, SecondScratchReg);
asMasm().ma_sd(ft, Address(SecondScratchReg, address.offset));
}
void
MacroAssemblerMIPSShared::ma_ss(FloatRegister ft, BaseIndex address)
{
asMasm().computeScaledAddress(address, SecondScratchReg);
asMasm().ma_ss(ft, Address(SecondScratchReg, address.offset));
}
void
MacroAssemblerMIPSShared::ma_bc1s(FloatRegister lhs, FloatRegister rhs, Label* label,
DoubleCondition c, JumpKind jumpKind, FPConditionBit fcc)
{
FloatTestKind testKind;
compareFloatingPoint(SingleFloat, lhs, rhs, c, &testKind, fcc);
asMasm().branchWithCode(getBranchCode(testKind, fcc), label, jumpKind);
}
void
MacroAssemblerMIPSShared::ma_bc1d(FloatRegister lhs, FloatRegister rhs, Label* label,
DoubleCondition c, JumpKind jumpKind, FPConditionBit fcc)
{
FloatTestKind testKind;
compareFloatingPoint(DoubleFloat, lhs, rhs, c, &testKind, fcc);
asMasm().branchWithCode(getBranchCode(testKind, fcc), label, jumpKind);
}
void
MacroAssemblerMIPSShared::ma_call(ImmPtr dest)
{
asMasm().ma_liPatchable(CallReg, dest);
as_jalr(CallReg);
as_nop();
}
void
MacroAssemblerMIPSShared::ma_jump(ImmPtr dest)
{
asMasm().ma_liPatchable(ScratchRegister, dest);
as_jr(ScratchRegister);
as_nop();
}
MacroAssembler&
MacroAssemblerMIPSShared::asMasm()
{
return *static_cast<MacroAssembler*>(this);
}
const MacroAssembler&
MacroAssemblerMIPSShared::asMasm() const
{
return *static_cast<const MacroAssembler*>(this);
}
void
MacroAssemblerMIPSShared::atomicEffectOpMIPSr2(int nbytes, AtomicOp op,
const Register& value, const Register& addr,
Register flagTemp, Register valueTemp,
Register offsetTemp, Register maskTemp)
{
atomicFetchOpMIPSr2(nbytes, false, op, value, addr, flagTemp,
valueTemp, offsetTemp, maskTemp, InvalidReg);
}
void
MacroAssemblerMIPSShared::atomicFetchOpMIPSr2(int nbytes, bool signExtend, AtomicOp op, const Register& value,
const Register& addr, Register flagTemp, Register valueTemp,
Register offsetTemp, Register maskTemp, Register output)
{
Label again;
as_andi(offsetTemp, addr, 3);
asMasm().subPtr(offsetTemp, addr);
as_sll(offsetTemp, offsetTemp, 3);
ma_li(maskTemp, Imm32(UINT32_MAX >> ((4 - nbytes) * 8)));
as_sllv(maskTemp, maskTemp, offsetTemp);
bind(&again);
as_sync(16);
as_ll(flagTemp, addr, 0);
as_sllv(valueTemp, value, offsetTemp);
if (output != InvalidReg) {
as_and(output, flagTemp, maskTemp);
as_srlv(output, output, offsetTemp);
if (signExtend) {
switch (nbytes) {
case 1:
as_seb(output, output);
break;
case 2:
as_seh(output, output);
break;
case 4:
break;
default:
MOZ_CRASH("NYI");
}
}
}
switch (op) {
case AtomicFetchAddOp:
as_addu(valueTemp, flagTemp, valueTemp);
break;
case AtomicFetchSubOp:
as_subu(valueTemp, flagTemp, valueTemp);
break;
case AtomicFetchAndOp:
as_and(valueTemp, flagTemp, valueTemp);
break;
case AtomicFetchOrOp:
as_or(valueTemp, flagTemp, valueTemp);
break;
case AtomicFetchXorOp:
as_xor(valueTemp, flagTemp, valueTemp);
break;
default:
MOZ_CRASH("NYI");
}
as_and(valueTemp, valueTemp, maskTemp);
as_or(flagTemp, flagTemp, maskTemp);
as_xor(flagTemp, flagTemp, maskTemp);
as_or(flagTemp, flagTemp, valueTemp);
as_sc(flagTemp, addr, 0);
ma_b(flagTemp, flagTemp, &again, Zero, ShortJump);
as_sync(0);
}
void
MacroAssemblerMIPSShared::atomicEffectOp(int nbytes, AtomicOp op, const Imm32& value,
const Address& address, Register flagTemp,
Register valueTemp, Register offsetTemp, Register maskTemp)
{
ma_li(SecondScratchReg, value);
asMasm().computeEffectiveAddress(address, ScratchRegister);
atomicEffectOpMIPSr2(nbytes, op, SecondScratchReg, ScratchRegister,
flagTemp, valueTemp, offsetTemp, maskTemp);
}
void
MacroAssemblerMIPSShared::atomicEffectOp(int nbytes, AtomicOp op, const Imm32& value,
const BaseIndex& address, Register flagTemp,
Register valueTemp, Register offsetTemp, Register maskTemp)
{
ma_li(SecondScratchReg, value);
asMasm().computeEffectiveAddress(address, ScratchRegister);
atomicEffectOpMIPSr2(nbytes, op, SecondScratchReg, ScratchRegister,
flagTemp, valueTemp, offsetTemp, maskTemp);
}
void
MacroAssemblerMIPSShared::atomicEffectOp(int nbytes, AtomicOp op, const Register& value,
const Address& address, Register flagTemp,
Register valueTemp, Register offsetTemp, Register maskTemp)
{
asMasm().computeEffectiveAddress(address, ScratchRegister);
atomicEffectOpMIPSr2(nbytes, op, value, ScratchRegister,
flagTemp, valueTemp, offsetTemp, maskTemp);
}
void
MacroAssemblerMIPSShared::atomicEffectOp(int nbytes, AtomicOp op, const Register& value,
const BaseIndex& address, Register flagTemp,
Register valueTemp, Register offsetTemp, Register maskTemp)
{
asMasm().computeEffectiveAddress(address, ScratchRegister);
atomicEffectOpMIPSr2(nbytes, op, value, ScratchRegister,
flagTemp, valueTemp, offsetTemp, maskTemp);
}
void
MacroAssemblerMIPSShared::atomicFetchOp(int nbytes, bool signExtend, AtomicOp op, const Imm32& value,
const Address& address, Register flagTemp, Register valueTemp,
Register offsetTemp, Register maskTemp, Register output)
{
ma_li(SecondScratchReg, value);
asMasm().computeEffectiveAddress(address, ScratchRegister);
atomicFetchOpMIPSr2(nbytes, signExtend, op, SecondScratchReg, ScratchRegister,
flagTemp, valueTemp, offsetTemp, maskTemp, output);
}
void
MacroAssemblerMIPSShared::atomicFetchOp(int nbytes, bool signExtend, AtomicOp op, const Imm32& value,
const BaseIndex& address, Register flagTemp, Register valueTemp,
Register offsetTemp, Register maskTemp, Register output)
{
ma_li(SecondScratchReg, value);
asMasm().computeEffectiveAddress(address, ScratchRegister);
atomicFetchOpMIPSr2(nbytes, signExtend, op, SecondScratchReg, ScratchRegister,
flagTemp, valueTemp, offsetTemp, maskTemp, output);
}
void
MacroAssemblerMIPSShared::atomicFetchOp(int nbytes, bool signExtend, AtomicOp op, const Register& value,
const Address& address, Register flagTemp, Register valueTemp,
Register offsetTemp, Register maskTemp, Register output)
{
asMasm().computeEffectiveAddress(address, ScratchRegister);
atomicFetchOpMIPSr2(nbytes, signExtend, op, value, ScratchRegister,
flagTemp, valueTemp, offsetTemp, maskTemp, output);
}
void
MacroAssemblerMIPSShared::atomicFetchOp(int nbytes, bool signExtend, AtomicOp op, const Register& value,
const BaseIndex& address, Register flagTemp, Register valueTemp,
Register offsetTemp, Register maskTemp, Register output)
{
asMasm().computeEffectiveAddress(address, ScratchRegister);
atomicFetchOpMIPSr2(nbytes, signExtend, op, value, ScratchRegister,
flagTemp, valueTemp, offsetTemp, maskTemp, output);
}
void
MacroAssemblerMIPSShared::compareExchangeMIPSr2(int nbytes, bool signExtend, const Register& addr,
Register oldval, Register newval, Register flagTemp,
Register valueTemp, Register offsetTemp, Register maskTemp,
Register output)
{
Label again, end;
as_andi(offsetTemp, addr, 3);
asMasm().subPtr(offsetTemp, addr);
as_sll(offsetTemp, offsetTemp, 3);
ma_li(maskTemp, Imm32(UINT32_MAX >> ((4 - nbytes) * 8)));
as_sllv(maskTemp, maskTemp, offsetTemp);
bind(&again);
as_sync(16);
as_ll(flagTemp, addr, 0);
as_and(output, flagTemp, maskTemp);
// If oldval is valid register, do compareExchange
if (InvalidReg != oldval) {
as_sllv(valueTemp, oldval, offsetTemp);
as_and(valueTemp, valueTemp, maskTemp);
ma_b(output, valueTemp, &end, NotEqual, ShortJump);
}
as_sllv(valueTemp, newval, offsetTemp);
as_and(valueTemp, valueTemp, maskTemp);
as_or(flagTemp, flagTemp, maskTemp);
as_xor(flagTemp, flagTemp, maskTemp);
as_or(flagTemp, flagTemp, valueTemp);
as_sc(flagTemp, addr, 0);
ma_b(flagTemp, flagTemp, &again, Zero, ShortJump);
as_sync(0);
bind(&end);
as_srlv(output, output, offsetTemp);
if (signExtend) {
switch (nbytes) {
case 1:
as_seb(output, output);
break;
case 2:
as_seh(output, output);
break;
case 4:
break;
default:
MOZ_CRASH("NYI");
}
}
}
void
MacroAssemblerMIPSShared::compareExchange(int nbytes, bool signExtend, const Address& address,
Register oldval, Register newval, Register valueTemp,
Register offsetTemp, Register maskTemp, Register output)
{
asMasm().computeEffectiveAddress(address, ScratchRegister);
compareExchangeMIPSr2(nbytes, signExtend, ScratchRegister, oldval, newval, SecondScratchReg,
valueTemp, offsetTemp, maskTemp, output);
}
void
MacroAssemblerMIPSShared::compareExchange(int nbytes, bool signExtend, const BaseIndex& address,
Register oldval, Register newval, Register valueTemp,
Register offsetTemp, Register maskTemp, Register output)
{
asMasm().computeEffectiveAddress(address, ScratchRegister);
compareExchangeMIPSr2(nbytes, signExtend, ScratchRegister, oldval, newval, SecondScratchReg,
valueTemp, offsetTemp, maskTemp, output);
}
void
MacroAssemblerMIPSShared::atomicExchange(int nbytes, bool signExtend, const Address& address,
Register value, Register valueTemp, Register offsetTemp,
Register maskTemp, Register output)
{
asMasm().computeEffectiveAddress(address, ScratchRegister);
compareExchangeMIPSr2(nbytes, signExtend, ScratchRegister, InvalidReg, value, SecondScratchReg,
valueTemp, offsetTemp, maskTemp, output);
}
void
MacroAssemblerMIPSShared::atomicExchange(int nbytes, bool signExtend, const BaseIndex& address,
Register value, Register valueTemp, Register offsetTemp,
Register maskTemp, Register output)
{
asMasm().computeEffectiveAddress(address, ScratchRegister);
compareExchangeMIPSr2(nbytes, signExtend, ScratchRegister, InvalidReg, value, SecondScratchReg,
valueTemp, offsetTemp, maskTemp, output);
}
//{{{ check_macroassembler_style
// ===============================================================
// Stack manipulation functions.
void
MacroAssembler::Push(Register reg)
{
ma_push(reg);
adjustFrame(sizeof(intptr_t));
}
void
MacroAssembler::Push(const Imm32 imm)
{
ma_li(ScratchRegister, imm);
ma_push(ScratchRegister);
adjustFrame(sizeof(intptr_t));
}
void
MacroAssembler::Push(const ImmWord imm)
{
ma_li(ScratchRegister, imm);
ma_push(ScratchRegister);
adjustFrame(sizeof(intptr_t));
}
void
MacroAssembler::Push(const ImmPtr imm)
{
Push(ImmWord(uintptr_t(imm.value)));
}
void
MacroAssembler::Push(const ImmGCPtr ptr)
{
ma_li(ScratchRegister, ptr);
ma_push(ScratchRegister);
adjustFrame(sizeof(intptr_t));
}
void
MacroAssembler::Push(FloatRegister f)
{
ma_push(f);
adjustFrame(sizeof(double));
}
void
MacroAssembler::Pop(Register reg)
{
ma_pop(reg);
adjustFrame(-sizeof(intptr_t));
}
void
MacroAssembler::Pop(const ValueOperand& val)
{
popValue(val);
framePushed_ -= sizeof(Value);
}
// ===============================================================
// Simple call functions.
CodeOffset
MacroAssembler::call(Register reg)
{
as_jalr(reg);
as_nop();
return CodeOffset(currentOffset());
}
CodeOffset
MacroAssembler::call(Label* label)
{
ma_bal(label);
return CodeOffset(currentOffset());
}
CodeOffset
MacroAssembler::callWithPatch()
{
addLongJump(nextOffset());
ma_liPatchable(ScratchRegister, ImmWord(0));
return call(ScratchRegister);
}
void
MacroAssembler::patchCall(uint32_t callerOffset, uint32_t calleeOffset)
{
BufferOffset li(callerOffset - Assembler::PatchWrite_NearCallSize());
Assembler::PatchInstructionImmediate((uint8_t*)editSrc(li),
PatchedImmPtr((const void*)calleeOffset));
}
void
MacroAssembler::call(wasm::SymbolicAddress target)
{
movePtr(target, CallReg);
call(CallReg);
}
void
MacroAssembler::call(ImmWord target)
{
call(ImmPtr((void*)target.value));
}
void
MacroAssembler::call(ImmPtr target)
{
BufferOffset bo = m_buffer.nextOffset();
addPendingJump(bo, target, Relocation::HARDCODED);
ma_call(target);
}
void
MacroAssembler::call(JitCode* c)
{
BufferOffset bo = m_buffer.nextOffset();
addPendingJump(bo, ImmPtr(c->raw()), Relocation::JITCODE);
ma_liPatchable(ScratchRegister, ImmPtr(c->raw()));
callJitNoProfiler(ScratchRegister);
}
void
MacroAssembler::callAndPushReturnAddress(Register callee)
{
// Push return address during jalr delay slot.
subPtr(Imm32(sizeof(intptr_t)), StackPointer);
as_jalr(callee);
storePtr(ra, Address(StackPointer, 0));
}
void
MacroAssembler::callAndPushReturnAddress(Label* label)
{
// Push return address during bal delay slot.
subPtr(Imm32(sizeof(intptr_t)), StackPointer);
ma_bal(label, DontFillDelaySlot);
storePtr(ra, Address(StackPointer, 0));
}
// ===============================================================
// Jit Frames.
uint32_t
MacroAssembler::pushFakeReturnAddress(Register scratch)
{
CodeLabel cl;
ma_li(scratch, cl.patchAt());
Push(scratch);
bind(cl.target());
uint32_t retAddr = currentOffset();
addCodeLabel(cl);
return retAddr;
}
//}}} check_macroassembler_style