| // Copyright 2012 the V8 project authors. All rights reserved. |
| // Redistribution and use in source and binary forms, with or without |
| // modification, are permitted provided that the following conditions are |
| // met: |
| // |
| // * Redistributions of source code must retain the above copyright |
| // notice, this list of conditions and the following disclaimer. |
| // * Redistributions in binary form must reproduce the above |
| // copyright notice, this list of conditions and the following |
| // disclaimer in the documentation and/or other materials provided |
| // with the distribution. |
| // * Neither the name of Google Inc. nor the names of its |
| // contributors may be used to endorse or promote products derived |
| // from this software without specific prior written permission. |
| // |
| // THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS |
| // "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT |
| // LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR |
| // A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT |
| // OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, |
| // SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT |
| // LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, |
| // DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY |
| // THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
| // (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
| // OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| |
| #include <iostream> // NOLINT(readability/streams) |
| |
| #include "src/init/v8.h" |
| |
| #include "src/base/utils/random-number-generator.h" |
| #include "src/codegen/assembler-inl.h" |
| #include "src/codegen/macro-assembler.h" |
| #include "src/diagnostics/disassembler.h" |
| #include "src/execution/simulator.h" |
| #include "src/heap/factory.h" |
| |
| #include "test/cctest/cctest.h" |
| |
| namespace v8 { |
| namespace internal { |
| |
| // Define these function prototypes to match JSEntryFunction in execution.cc. |
| // TODO(mips): Refine these signatures per test case. |
| using F1 = void*(int x, int p1, int p2, int p3, int p4); |
| using F2 = void*(int x, int y, int p2, int p3, int p4); |
| using F3 = void*(void* p, int p1, int p2, int p3, int p4); |
| using F4 = void*(void* p0, void* p1, int p2, int p3, int p4); |
| |
| #define __ assm. |
| |
| TEST(MIPS0) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| // Addition. |
| __ addu(v0, a0, a1); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F2>::FromCode(*code); |
| int res = reinterpret_cast<int>(f.Call(0xAB0, 0xC, 0, 0, 0)); |
| CHECK_EQ(static_cast<int32_t>(0xABC), res); |
| } |
| |
| |
| TEST(MIPS1) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| Label L, C; |
| |
| __ mov(a1, a0); |
| __ li(v0, 0); |
| __ b(&C); |
| __ nop(); |
| |
| __ bind(&L); |
| __ addu(v0, v0, a1); |
| __ addiu(a1, a1, -1); |
| |
| __ bind(&C); |
| __ xori(v1, a1, 0); |
| __ Branch(&L, ne, v1, Operand(0)); |
| __ nop(); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F1>::FromCode(*code); |
| int res = reinterpret_cast<int>(f.Call(50, 0, 0, 0, 0)); |
| CHECK_EQ(1275, res); |
| } |
| |
| |
| TEST(MIPS2) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| Label exit, error; |
| |
| // ----- Test all instructions. |
| |
| // Test lui, ori, and addiu, used in the li pseudo-instruction. |
| // This way we can then safely load registers with chosen values. |
| |
| __ ori(t0, zero_reg, 0); |
| __ lui(t0, 0x1234); |
| __ ori(t0, t0, 0); |
| __ ori(t0, t0, 0x0F0F); |
| __ ori(t0, t0, 0xF0F0); |
| __ addiu(t1, t0, 1); |
| __ addiu(t2, t1, -0x10); |
| |
| // Load values in temporary registers. |
| __ li(t0, 0x00000004); |
| __ li(t1, 0x00001234); |
| __ li(t2, 0x12345678); |
| __ li(t3, 0x7FFFFFFF); |
| __ li(t4, 0xFFFFFFFC); |
| __ li(t5, 0xFFFFEDCC); |
| __ li(t6, 0xEDCBA988); |
| __ li(t7, 0x80000000); |
| |
| // SPECIAL class. |
| __ srl(v0, t2, 8); // 0x00123456 |
| __ sll(v0, v0, 11); // 0x91A2B000 |
| __ sra(v0, v0, 3); // 0xF2345600 |
| __ srav(v0, v0, t0); // 0xFF234560 |
| __ sllv(v0, v0, t0); // 0xF2345600 |
| __ srlv(v0, v0, t0); // 0x0F234560 |
| __ Branch(&error, ne, v0, Operand(0x0F234560)); |
| __ nop(); |
| |
| __ addu(v0, t0, t1); // 0x00001238 |
| __ subu(v0, v0, t0); // 0x00001234 |
| __ Branch(&error, ne, v0, Operand(0x00001234)); |
| __ nop(); |
| __ addu(v1, t3, t0); |
| __ Branch(&error, ne, v1, Operand(0x80000003)); |
| __ nop(); |
| __ subu(v1, t7, t0); // 0x7FFFFFFC |
| __ Branch(&error, ne, v1, Operand(0x7FFFFFFC)); |
| __ nop(); |
| |
| __ and_(v0, t1, t2); // 0x00001230 |
| __ or_(v0, v0, t1); // 0x00001234 |
| __ xor_(v0, v0, t2); // 0x1234444C |
| __ nor(v0, v0, t2); // 0xEDCBA987 |
| __ Branch(&error, ne, v0, Operand(0xEDCBA983)); |
| __ nop(); |
| |
| __ slt(v0, t7, t3); |
| __ Branch(&error, ne, v0, Operand(0x1)); |
| __ nop(); |
| __ sltu(v0, t7, t3); |
| __ Branch(&error, ne, v0, Operand(zero_reg)); |
| __ nop(); |
| // End of SPECIAL class. |
| |
| __ addiu(v0, zero_reg, 0x7421); // 0x00007421 |
| __ addiu(v0, v0, -0x1); // 0x00007420 |
| __ addiu(v0, v0, -0x20); // 0x00007400 |
| __ Branch(&error, ne, v0, Operand(0x00007400)); |
| __ nop(); |
| __ addiu(v1, t3, 0x1); // 0x80000000 |
| __ Branch(&error, ne, v1, Operand(0x80000000)); |
| __ nop(); |
| |
| __ slti(v0, t1, 0x00002000); // 0x1 |
| __ slti(v0, v0, 0xFFFF8000); // 0x0 |
| __ Branch(&error, ne, v0, Operand(zero_reg)); |
| __ nop(); |
| __ sltiu(v0, t1, 0x00002000); // 0x1 |
| __ sltiu(v0, v0, 0x00008000); // 0x1 |
| __ Branch(&error, ne, v0, Operand(0x1)); |
| __ nop(); |
| |
| __ andi(v0, t1, 0xF0F0); // 0x00001030 |
| __ ori(v0, v0, 0x8A00); // 0x00009A30 |
| __ xori(v0, v0, 0x83CC); // 0x000019FC |
| __ Branch(&error, ne, v0, Operand(0x000019FC)); |
| __ nop(); |
| __ lui(v1, 0x8123); // 0x81230000 |
| __ Branch(&error, ne, v1, Operand(0x81230000)); |
| __ nop(); |
| |
| // Bit twiddling instructions & conditional moves. |
| // Uses t0-t7 as set above. |
| __ Clz(v0, t0); // 29 |
| __ Clz(v1, t1); // 19 |
| __ addu(v0, v0, v1); // 48 |
| __ Clz(v1, t2); // 3 |
| __ addu(v0, v0, v1); // 51 |
| __ Clz(v1, t7); // 0 |
| __ addu(v0, v0, v1); // 51 |
| __ Branch(&error, ne, v0, Operand(51)); |
| __ Movn(a0, t3, t0); // Move a0<-t3 (t0 is NOT 0). |
| __ Ins(a0, t1, 12, 8); // 0x7FF34FFF |
| __ Branch(&error, ne, a0, Operand(0x7FF34FFF)); |
| __ Movz(a0, t6, t7); // a0 not updated (t7 is NOT 0). |
| __ Ext(a1, a0, 8, 12); // 0x34F |
| __ Branch(&error, ne, a1, Operand(0x34F)); |
| __ Movz(a0, t6, v1); // a0<-t6, v0 is 0, from 8 instr back. |
| __ Branch(&error, ne, a0, Operand(t6)); |
| |
| // Everything was correctly executed. Load the expected result. |
| __ li(v0, 0x31415926); |
| __ b(&exit); |
| __ nop(); |
| |
| __ bind(&error); |
| // Got an error. Return a wrong result. |
| __ li(v0, 666); |
| |
| __ bind(&exit); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F2>::FromCode(*code); |
| int res = reinterpret_cast<int>(f.Call(0xAB0, 0xC, 0, 0, 0)); |
| CHECK_EQ(static_cast<int32_t>(0x31415926), res); |
| } |
| |
| |
| TEST(MIPS3) { |
| // Test floating point instructions. |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| struct T { |
| double a; |
| double b; |
| double c; |
| double d; |
| double e; |
| double f; |
| double g; |
| double h; |
| double i; |
| float fa; |
| float fb; |
| float fc; |
| float fd; |
| float fe; |
| float ff; |
| float fg; |
| }; |
| T t; |
| |
| // Create a function that accepts &t, and loads, manipulates, and stores |
| // the doubles t.a ... t.f. |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| // Double precision floating point instructions. |
| __ Ldc1(f4, MemOperand(a0, offsetof(T, a))); |
| __ Ldc1(f6, MemOperand(a0, offsetof(T, b))); |
| __ add_d(f8, f4, f6); |
| __ Sdc1(f8, MemOperand(a0, offsetof(T, c))); // c = a + b. |
| |
| __ mov_d(f10, f8); // c |
| __ neg_d(f12, f6); // -b |
| __ sub_d(f10, f10, f12); |
| __ Sdc1(f10, MemOperand(a0, offsetof(T, d))); // d = c - (-b). |
| |
| __ Sdc1(f4, MemOperand(a0, offsetof(T, b))); // b = a. |
| |
| __ li(t0, 120); |
| __ mtc1(t0, f14); |
| __ cvt_d_w(f14, f14); // f14 = 120.0. |
| __ mul_d(f10, f10, f14); |
| __ Sdc1(f10, MemOperand(a0, offsetof(T, e))); // e = d * 120 = 1.8066e16. |
| |
| __ div_d(f12, f10, f4); |
| __ Sdc1(f12, MemOperand(a0, offsetof(T, f))); // f = e / a = 120.44. |
| |
| __ sqrt_d(f14, f12); |
| __ Sdc1(f14, MemOperand(a0, offsetof(T, g))); |
| // g = sqrt(f) = 10.97451593465515908537 |
| |
| if (IsMipsArchVariant(kMips32r2)) { |
| __ Ldc1(f4, MemOperand(a0, offsetof(T, h))); |
| __ Ldc1(f6, MemOperand(a0, offsetof(T, i))); |
| __ madd_d(f14, f6, f4, f6); |
| __ Sdc1(f14, MemOperand(a0, offsetof(T, h))); |
| } |
| |
| // Single precision floating point instructions. |
| __ lwc1(f4, MemOperand(a0, offsetof(T, fa)) ); |
| __ lwc1(f6, MemOperand(a0, offsetof(T, fb)) ); |
| __ add_s(f8, f4, f6); |
| __ swc1(f8, MemOperand(a0, offsetof(T, fc)) ); // fc = fa + fb. |
| |
| __ neg_s(f10, f6); // -fb |
| __ sub_s(f10, f8, f10); |
| __ swc1(f10, MemOperand(a0, offsetof(T, fd)) ); // fd = fc - (-fb). |
| |
| __ swc1(f4, MemOperand(a0, offsetof(T, fb)) ); // fb = fa. |
| |
| __ li(t0, 120); |
| __ mtc1(t0, f14); |
| __ cvt_s_w(f14, f14); // f14 = 120.0. |
| __ mul_s(f10, f10, f14); |
| __ swc1(f10, MemOperand(a0, offsetof(T, fe)) ); // fe = fd * 120 |
| |
| __ div_s(f12, f10, f4); |
| __ swc1(f12, MemOperand(a0, offsetof(T, ff)) ); // ff = fe / fa |
| |
| __ sqrt_s(f14, f12); |
| __ swc1(f14, MemOperand(a0, offsetof(T, fg)) ); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| // Double test values. |
| t.a = 1.5e14; |
| t.b = 2.75e11; |
| t.c = 0.0; |
| t.d = 0.0; |
| t.e = 0.0; |
| t.f = 0.0; |
| t.h = 1.5; |
| t.i = 2.75; |
| // Single test values. |
| t.fa = 1.5e6; |
| t.fb = 2.75e4; |
| t.fc = 0.0; |
| t.fd = 0.0; |
| t.fe = 0.0; |
| t.ff = 0.0; |
| f.Call(&t, 0, 0, 0, 0); |
| // Expected double results. |
| CHECK_EQ(1.5e14, t.a); |
| CHECK_EQ(1.5e14, t.b); |
| CHECK_EQ(1.50275e14, t.c); |
| CHECK_EQ(1.50550e14, t.d); |
| CHECK_EQ(1.8066e16, t.e); |
| CHECK_EQ(120.44, t.f); |
| CHECK_EQ(10.97451593465515908537, t.g); |
| if (IsMipsArchVariant(kMips32r2)) { |
| CHECK_EQ(6.875, t.h); |
| } |
| // Expected single results. |
| CHECK_EQ(1.5e6, t.fa); |
| CHECK_EQ(1.5e6, t.fb); |
| CHECK_EQ(1.5275e06, t.fc); |
| CHECK_EQ(1.5550e06, t.fd); |
| CHECK_EQ(1.866e08, t.fe); |
| CHECK_EQ(124.40000152587890625, t.ff); |
| CHECK_EQ(11.1534748077392578125, t.fg); |
| } |
| |
| |
| TEST(MIPS4) { |
| // Exchange between GP anf FP registers is done through memory |
| // on FPXX compiled binaries and architectures that do not support |
| // MTHC1 and MTFC1. If this is the case, skipping this test. |
| if (IsFpxxMode() && |
| (IsMipsArchVariant(kMips32r1) || IsMipsArchVariant(kLoongson))) { |
| return; |
| } |
| |
| // Test moves between floating point and integer registers. |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| struct T { |
| double a; |
| double b; |
| double c; |
| }; |
| T t; |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| __ Ldc1(f4, MemOperand(a0, offsetof(T, a))); |
| __ Ldc1(f6, MemOperand(a0, offsetof(T, b))); |
| |
| // Swap f4 and f6, by using four integer registers, t0-t3. |
| if (IsFp32Mode()) { |
| __ mfc1(t0, f4); |
| __ mfc1(t1, f5); |
| __ mfc1(t2, f6); |
| __ mfc1(t3, f7); |
| |
| __ mtc1(t0, f6); |
| __ mtc1(t1, f7); |
| __ mtc1(t2, f4); |
| __ mtc1(t3, f5); |
| } else { |
| CHECK(!IsMipsArchVariant(kMips32r1) && !IsMipsArchVariant(kLoongson)); |
| DCHECK(IsFp64Mode() || IsFpxxMode()); |
| __ mfc1(t0, f4); |
| __ mfhc1(t1, f4); |
| __ mfc1(t2, f6); |
| __ mfhc1(t3, f6); |
| |
| __ mtc1(t0, f6); |
| __ mthc1(t1, f6); |
| __ mtc1(t2, f4); |
| __ mthc1(t3, f4); |
| } |
| |
| // Store the swapped f4 and f5 back to memory. |
| __ Sdc1(f4, MemOperand(a0, offsetof(T, a))); |
| __ Sdc1(f6, MemOperand(a0, offsetof(T, c))); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| t.a = 1.5e22; |
| t.b = 2.75e11; |
| t.c = 17.17; |
| f.Call(&t, 0, 0, 0, 0); |
| |
| CHECK_EQ(2.75e11, t.a); |
| CHECK_EQ(2.75e11, t.b); |
| CHECK_EQ(1.5e22, t.c); |
| } |
| |
| |
| TEST(MIPS5) { |
| // Test conversions between doubles and integers. |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| struct T { |
| double a; |
| double b; |
| int i; |
| int j; |
| }; |
| T t; |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| // Load all structure elements to registers. |
| __ Ldc1(f4, MemOperand(a0, offsetof(T, a))); |
| __ Ldc1(f6, MemOperand(a0, offsetof(T, b))); |
| __ lw(t0, MemOperand(a0, offsetof(T, i)) ); |
| __ lw(t1, MemOperand(a0, offsetof(T, j)) ); |
| |
| // Convert double in f4 to int in element i. |
| __ cvt_w_d(f8, f4); |
| __ mfc1(t2, f8); |
| __ sw(t2, MemOperand(a0, offsetof(T, i)) ); |
| |
| // Convert double in f6 to int in element j. |
| __ cvt_w_d(f10, f6); |
| __ mfc1(t3, f10); |
| __ sw(t3, MemOperand(a0, offsetof(T, j)) ); |
| |
| // Convert int in original i (t0) to double in a. |
| __ mtc1(t0, f12); |
| __ cvt_d_w(f0, f12); |
| __ Sdc1(f0, MemOperand(a0, offsetof(T, a))); |
| |
| // Convert int in original j (t1) to double in b. |
| __ mtc1(t1, f14); |
| __ cvt_d_w(f2, f14); |
| __ Sdc1(f2, MemOperand(a0, offsetof(T, b))); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| t.a = 1.5e4; |
| t.b = 2.75e8; |
| t.i = 12345678; |
| t.j = -100000; |
| f.Call(&t, 0, 0, 0, 0); |
| |
| CHECK_EQ(12345678.0, t.a); |
| CHECK_EQ(-100000.0, t.b); |
| CHECK_EQ(15000, t.i); |
| CHECK_EQ(275000000, t.j); |
| } |
| |
| |
| TEST(MIPS6) { |
| // Test simple memory loads and stores. |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| struct T { |
| uint32_t ui; |
| int32_t si; |
| int32_t r1; |
| int32_t r2; |
| int32_t r3; |
| int32_t r4; |
| int32_t r5; |
| int32_t r6; |
| }; |
| T t; |
| |
| Assembler assm(AssemblerOptions{}); |
| |
| // Basic word load/store. |
| __ lw(t0, MemOperand(a0, offsetof(T, ui)) ); |
| __ sw(t0, MemOperand(a0, offsetof(T, r1)) ); |
| |
| // lh with positive data. |
| __ lh(t1, MemOperand(a0, offsetof(T, ui)) ); |
| __ sw(t1, MemOperand(a0, offsetof(T, r2)) ); |
| |
| // lh with negative data. |
| __ lh(t2, MemOperand(a0, offsetof(T, si)) ); |
| __ sw(t2, MemOperand(a0, offsetof(T, r3)) ); |
| |
| // lhu with negative data. |
| __ lhu(t3, MemOperand(a0, offsetof(T, si)) ); |
| __ sw(t3, MemOperand(a0, offsetof(T, r4)) ); |
| |
| // lb with negative data. |
| __ lb(t4, MemOperand(a0, offsetof(T, si)) ); |
| __ sw(t4, MemOperand(a0, offsetof(T, r5)) ); |
| |
| // sh writes only 1/2 of word. |
| __ lui(t5, 0x3333); |
| __ ori(t5, t5, 0x3333); |
| __ sw(t5, MemOperand(a0, offsetof(T, r6)) ); |
| __ lhu(t5, MemOperand(a0, offsetof(T, si)) ); |
| __ sh(t5, MemOperand(a0, offsetof(T, r6)) ); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| t.ui = 0x11223344; |
| t.si = 0x99AABBCC; |
| f.Call(&t, 0, 0, 0, 0); |
| |
| CHECK_EQ(static_cast<int32_t>(0x11223344), t.r1); |
| #if __BYTE_ORDER == __LITTLE_ENDIAN |
| CHECK_EQ(static_cast<int32_t>(0x3344), t.r2); |
| CHECK_EQ(static_cast<int32_t>(0xFFFFBBCC), t.r3); |
| CHECK_EQ(static_cast<int32_t>(0x0000BBCC), t.r4); |
| CHECK_EQ(static_cast<int32_t>(0xFFFFFFCC), t.r5); |
| CHECK_EQ(static_cast<int32_t>(0x3333BBCC), t.r6); |
| #elif __BYTE_ORDER == __BIG_ENDIAN |
| CHECK_EQ(static_cast<int32_t>(0x1122), t.r2); |
| CHECK_EQ(static_cast<int32_t>(0xFFFF99AA), t.r3); |
| CHECK_EQ(static_cast<int32_t>(0x000099AA), t.r4); |
| CHECK_EQ(static_cast<int32_t>(0xFFFFFF99), t.r5); |
| CHECK_EQ(static_cast<int32_t>(0x99AA3333), t.r6); |
| #else |
| #error Unknown endianness |
| #endif |
| } |
| |
| |
| TEST(MIPS7) { |
| // Test floating point compare and branch instructions. |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| struct T { |
| double a; |
| double b; |
| double c; |
| double d; |
| double e; |
| double f; |
| int32_t result; |
| }; |
| T t; |
| |
| // Create a function that accepts &t, and loads, manipulates, and stores |
| // the doubles t.a ... t.f. |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| Label neither_is_nan, less_than, outa_here; |
| |
| __ Ldc1(f4, MemOperand(a0, offsetof(T, a))); |
| __ Ldc1(f6, MemOperand(a0, offsetof(T, b))); |
| if (!IsMipsArchVariant(kMips32r6)) { |
| __ c(UN, D, f4, f6); |
| __ bc1f(&neither_is_nan); |
| } else { |
| __ cmp(UN, L, f2, f4, f6); |
| __ bc1eqz(&neither_is_nan, f2); |
| } |
| __ nop(); |
| __ sw(zero_reg, MemOperand(a0, offsetof(T, result)) ); |
| __ Branch(&outa_here); |
| |
| __ bind(&neither_is_nan); |
| |
| if (IsMipsArchVariant(kLoongson)) { |
| __ c(OLT, D, f6, f4); |
| __ bc1t(&less_than); |
| } else if (IsMipsArchVariant(kMips32r6)) { |
| __ cmp(OLT, L, f2, f6, f4); |
| __ bc1nez(&less_than, f2); |
| } else { |
| __ c(OLT, D, f6, f4, 2); |
| __ bc1t(&less_than, 2); |
| } |
| |
| __ nop(); |
| __ sw(zero_reg, MemOperand(a0, offsetof(T, result)) ); |
| __ Branch(&outa_here); |
| |
| __ bind(&less_than); |
| __ Addu(t0, zero_reg, Operand(1)); |
| __ sw(t0, MemOperand(a0, offsetof(T, result)) ); // Set true. |
| |
| |
| // This test-case should have additional tests. |
| |
| __ bind(&outa_here); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| t.a = 1.5e14; |
| t.b = 2.75e11; |
| t.c = 2.0; |
| t.d = -4.0; |
| t.e = 0.0; |
| t.f = 0.0; |
| t.result = 0; |
| f.Call(&t, 0, 0, 0, 0); |
| CHECK_EQ(1.5e14, t.a); |
| CHECK_EQ(2.75e11, t.b); |
| CHECK_EQ(1, t.result); |
| } |
| |
| |
| TEST(MIPS8) { |
| // Test ROTR and ROTRV instructions. |
| if (IsMipsArchVariant(kMips32r2)) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| struct T { |
| int32_t input; |
| int32_t result_rotr_4; |
| int32_t result_rotr_8; |
| int32_t result_rotr_12; |
| int32_t result_rotr_16; |
| int32_t result_rotr_20; |
| int32_t result_rotr_24; |
| int32_t result_rotr_28; |
| int32_t result_rotrv_4; |
| int32_t result_rotrv_8; |
| int32_t result_rotrv_12; |
| int32_t result_rotrv_16; |
| int32_t result_rotrv_20; |
| int32_t result_rotrv_24; |
| int32_t result_rotrv_28; |
| }; |
| T t; |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| // Basic word load. |
| __ lw(t0, MemOperand(a0, offsetof(T, input)) ); |
| |
| // ROTR instruction (called through the Ror macro). |
| __ Ror(t1, t0, 0x0004); |
| __ Ror(t2, t0, 0x0008); |
| __ Ror(t3, t0, 0x000C); |
| __ Ror(t4, t0, 0x0010); |
| __ Ror(t5, t0, 0x0014); |
| __ Ror(t6, t0, 0x0018); |
| __ Ror(t7, t0, 0x001C); |
| |
| // Basic word store. |
| __ sw(t1, MemOperand(a0, offsetof(T, result_rotr_4)) ); |
| __ sw(t2, MemOperand(a0, offsetof(T, result_rotr_8)) ); |
| __ sw(t3, MemOperand(a0, offsetof(T, result_rotr_12)) ); |
| __ sw(t4, MemOperand(a0, offsetof(T, result_rotr_16)) ); |
| __ sw(t5, MemOperand(a0, offsetof(T, result_rotr_20)) ); |
| __ sw(t6, MemOperand(a0, offsetof(T, result_rotr_24)) ); |
| __ sw(t7, MemOperand(a0, offsetof(T, result_rotr_28)) ); |
| |
| // ROTRV instruction (called through the Ror macro). |
| __ li(t7, 0x0004); |
| __ Ror(t1, t0, t7); |
| __ li(t7, 0x0008); |
| __ Ror(t2, t0, t7); |
| __ li(t7, 0x000C); |
| __ Ror(t3, t0, t7); |
| __ li(t7, 0x0010); |
| __ Ror(t4, t0, t7); |
| __ li(t7, 0x0014); |
| __ Ror(t5, t0, t7); |
| __ li(t7, 0x0018); |
| __ Ror(t6, t0, t7); |
| __ li(t7, 0x001C); |
| __ Ror(t7, t0, t7); |
| |
| // Basic word store. |
| __ sw(t1, MemOperand(a0, offsetof(T, result_rotrv_4)) ); |
| __ sw(t2, MemOperand(a0, offsetof(T, result_rotrv_8)) ); |
| __ sw(t3, MemOperand(a0, offsetof(T, result_rotrv_12)) ); |
| __ sw(t4, MemOperand(a0, offsetof(T, result_rotrv_16)) ); |
| __ sw(t5, MemOperand(a0, offsetof(T, result_rotrv_20)) ); |
| __ sw(t6, MemOperand(a0, offsetof(T, result_rotrv_24)) ); |
| __ sw(t7, MemOperand(a0, offsetof(T, result_rotrv_28)) ); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| t.input = 0x12345678; |
| f.Call(&t, 0x0, 0, 0, 0); |
| CHECK_EQ(static_cast<int32_t>(0x81234567), t.result_rotr_4); |
| CHECK_EQ(static_cast<int32_t>(0x78123456), t.result_rotr_8); |
| CHECK_EQ(static_cast<int32_t>(0x67812345), t.result_rotr_12); |
| CHECK_EQ(static_cast<int32_t>(0x56781234), t.result_rotr_16); |
| CHECK_EQ(static_cast<int32_t>(0x45678123), t.result_rotr_20); |
| CHECK_EQ(static_cast<int32_t>(0x34567812), t.result_rotr_24); |
| CHECK_EQ(static_cast<int32_t>(0x23456781), t.result_rotr_28); |
| |
| CHECK_EQ(static_cast<int32_t>(0x81234567), t.result_rotrv_4); |
| CHECK_EQ(static_cast<int32_t>(0x78123456), t.result_rotrv_8); |
| CHECK_EQ(static_cast<int32_t>(0x67812345), t.result_rotrv_12); |
| CHECK_EQ(static_cast<int32_t>(0x56781234), t.result_rotrv_16); |
| CHECK_EQ(static_cast<int32_t>(0x45678123), t.result_rotrv_20); |
| CHECK_EQ(static_cast<int32_t>(0x34567812), t.result_rotrv_24); |
| CHECK_EQ(static_cast<int32_t>(0x23456781), t.result_rotrv_28); |
| } |
| } |
| |
| |
| TEST(MIPS9) { |
| // Test BRANCH improvements. |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| Label exit, exit2, exit3; |
| |
| __ Branch(&exit, ge, a0, Operand(zero_reg)); |
| __ Branch(&exit2, ge, a0, Operand(0x00001FFF)); |
| __ Branch(&exit3, ge, a0, Operand(0x0001FFFF)); |
| |
| __ bind(&exit); |
| __ bind(&exit2); |
| __ bind(&exit3); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| USE(code); |
| } |
| |
| |
| TEST(MIPS10) { |
| // Test conversions between doubles and words. |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| struct T { |
| double a; |
| double b; |
| int32_t dbl_mant; |
| int32_t dbl_exp; |
| int32_t word; |
| int32_t b_word; |
| }; |
| T t; |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| if (IsMipsArchVariant(kMips32r1) || IsMipsArchVariant(kLoongson)) return; |
| |
| // Load all structure elements to registers. |
| // (f0, f1) = a (fp32), f0 = a (fp64) |
| __ Ldc1(f0, MemOperand(a0, offsetof(T, a))); |
| |
| __ mfc1(t0, f0); // t0 = f0(31..0) |
| __ mfhc1(t1, f0); // t1 = sign_extend(f0(63..32)) |
| __ sw(t0, MemOperand(a0, offsetof(T, dbl_mant))); // dbl_mant = t0 |
| __ sw(t1, MemOperand(a0, offsetof(T, dbl_exp))); // dbl_exp = t1 |
| |
| // Convert double in f0 to word, save hi/lo parts. |
| __ cvt_w_d(f0, f0); // a_word = (word)a |
| __ mfc1(t0, f0); // f0 has a 32-bits word. t0 = a_word |
| __ sw(t0, MemOperand(a0, offsetof(T, word))); // word = a_word |
| |
| // Convert the b word to double b. |
| __ lw(t0, MemOperand(a0, offsetof(T, b_word))); |
| __ mtc1(t0, f8); // f8 has a 32-bits word. |
| __ cvt_d_w(f10, f8); |
| __ Sdc1(f10, MemOperand(a0, offsetof(T, b))); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| t.a = 2.147483646e+09; // 0x7FFFFFFE -> 0xFF80000041DFFFFF as double. |
| t.b_word = 0x0FF00FF0; // 0x0FF00FF0 -> 0x as double. |
| f.Call(&t, 0, 0, 0, 0); |
| CHECK_EQ(static_cast<int32_t>(0x41DFFFFF), t.dbl_exp); |
| CHECK_EQ(static_cast<int32_t>(0xFF800000), t.dbl_mant); |
| CHECK_EQ(static_cast<int32_t>(0x7FFFFFFE), t.word); |
| // 0x0FF00FF0 -> 2.6739096+e08 |
| CHECK_EQ(2.6739096e08, t.b); |
| } |
| |
| |
| TEST(MIPS11) { |
| // Do not run test on MIPS32r6, as these instructions are removed. |
| if (IsMipsArchVariant(kMips32r6)) return; |
| // Test LWL, LWR, SWL and SWR instructions. |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| struct T { |
| int32_t reg_init; |
| int32_t mem_init; |
| int32_t lwl_0; |
| int32_t lwl_1; |
| int32_t lwl_2; |
| int32_t lwl_3; |
| int32_t lwr_0; |
| int32_t lwr_1; |
| int32_t lwr_2; |
| int32_t lwr_3; |
| int32_t swl_0; |
| int32_t swl_1; |
| int32_t swl_2; |
| int32_t swl_3; |
| int32_t swr_0; |
| int32_t swr_1; |
| int32_t swr_2; |
| int32_t swr_3; |
| }; |
| T t; |
| |
| Assembler assm(AssemblerOptions{}); |
| |
| // Test all combinations of LWL and vAddr. |
| __ lw(t0, MemOperand(a0, offsetof(T, reg_init)) ); |
| __ lwl(t0, MemOperand(a0, offsetof(T, mem_init)) ); |
| __ sw(t0, MemOperand(a0, offsetof(T, lwl_0)) ); |
| |
| __ lw(t1, MemOperand(a0, offsetof(T, reg_init)) ); |
| __ lwl(t1, MemOperand(a0, offsetof(T, mem_init) + 1) ); |
| __ sw(t1, MemOperand(a0, offsetof(T, lwl_1)) ); |
| |
| __ lw(t2, MemOperand(a0, offsetof(T, reg_init)) ); |
| __ lwl(t2, MemOperand(a0, offsetof(T, mem_init) + 2) ); |
| __ sw(t2, MemOperand(a0, offsetof(T, lwl_2)) ); |
| |
| __ lw(t3, MemOperand(a0, offsetof(T, reg_init)) ); |
| __ lwl(t3, MemOperand(a0, offsetof(T, mem_init) + 3) ); |
| __ sw(t3, MemOperand(a0, offsetof(T, lwl_3)) ); |
| |
| // Test all combinations of LWR and vAddr. |
| __ lw(t0, MemOperand(a0, offsetof(T, reg_init)) ); |
| __ lwr(t0, MemOperand(a0, offsetof(T, mem_init)) ); |
| __ sw(t0, MemOperand(a0, offsetof(T, lwr_0)) ); |
| |
| __ lw(t1, MemOperand(a0, offsetof(T, reg_init)) ); |
| __ lwr(t1, MemOperand(a0, offsetof(T, mem_init) + 1) ); |
| __ sw(t1, MemOperand(a0, offsetof(T, lwr_1)) ); |
| |
| __ lw(t2, MemOperand(a0, offsetof(T, reg_init)) ); |
| __ lwr(t2, MemOperand(a0, offsetof(T, mem_init) + 2) ); |
| __ sw(t2, MemOperand(a0, offsetof(T, lwr_2)) ); |
| |
| __ lw(t3, MemOperand(a0, offsetof(T, reg_init)) ); |
| __ lwr(t3, MemOperand(a0, offsetof(T, mem_init) + 3) ); |
| __ sw(t3, MemOperand(a0, offsetof(T, lwr_3)) ); |
| |
| // Test all combinations of SWL and vAddr. |
| __ lw(t0, MemOperand(a0, offsetof(T, mem_init)) ); |
| __ sw(t0, MemOperand(a0, offsetof(T, swl_0)) ); |
| __ lw(t0, MemOperand(a0, offsetof(T, reg_init)) ); |
| __ swl(t0, MemOperand(a0, offsetof(T, swl_0)) ); |
| |
| __ lw(t1, MemOperand(a0, offsetof(T, mem_init)) ); |
| __ sw(t1, MemOperand(a0, offsetof(T, swl_1)) ); |
| __ lw(t1, MemOperand(a0, offsetof(T, reg_init)) ); |
| __ swl(t1, MemOperand(a0, offsetof(T, swl_1) + 1) ); |
| |
| __ lw(t2, MemOperand(a0, offsetof(T, mem_init)) ); |
| __ sw(t2, MemOperand(a0, offsetof(T, swl_2)) ); |
| __ lw(t2, MemOperand(a0, offsetof(T, reg_init)) ); |
| __ swl(t2, MemOperand(a0, offsetof(T, swl_2) + 2) ); |
| |
| __ lw(t3, MemOperand(a0, offsetof(T, mem_init)) ); |
| __ sw(t3, MemOperand(a0, offsetof(T, swl_3)) ); |
| __ lw(t3, MemOperand(a0, offsetof(T, reg_init)) ); |
| __ swl(t3, MemOperand(a0, offsetof(T, swl_3) + 3) ); |
| |
| // Test all combinations of SWR and vAddr. |
| __ lw(t0, MemOperand(a0, offsetof(T, mem_init)) ); |
| __ sw(t0, MemOperand(a0, offsetof(T, swr_0)) ); |
| __ lw(t0, MemOperand(a0, offsetof(T, reg_init)) ); |
| __ swr(t0, MemOperand(a0, offsetof(T, swr_0)) ); |
| |
| __ lw(t1, MemOperand(a0, offsetof(T, mem_init)) ); |
| __ sw(t1, MemOperand(a0, offsetof(T, swr_1)) ); |
| __ lw(t1, MemOperand(a0, offsetof(T, reg_init)) ); |
| __ swr(t1, MemOperand(a0, offsetof(T, swr_1) + 1) ); |
| |
| __ lw(t2, MemOperand(a0, offsetof(T, mem_init)) ); |
| __ sw(t2, MemOperand(a0, offsetof(T, swr_2)) ); |
| __ lw(t2, MemOperand(a0, offsetof(T, reg_init)) ); |
| __ swr(t2, MemOperand(a0, offsetof(T, swr_2) + 2) ); |
| |
| __ lw(t3, MemOperand(a0, offsetof(T, mem_init)) ); |
| __ sw(t3, MemOperand(a0, offsetof(T, swr_3)) ); |
| __ lw(t3, MemOperand(a0, offsetof(T, reg_init)) ); |
| __ swr(t3, MemOperand(a0, offsetof(T, swr_3) + 3) ); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| t.reg_init = 0xAABBCCDD; |
| t.mem_init = 0x11223344; |
| |
| f.Call(&t, 0, 0, 0, 0); |
| |
| #if __BYTE_ORDER == __LITTLE_ENDIAN |
| CHECK_EQ(static_cast<int32_t>(0x44BBCCDD), t.lwl_0); |
| CHECK_EQ(static_cast<int32_t>(0x3344CCDD), t.lwl_1); |
| CHECK_EQ(static_cast<int32_t>(0x223344DD), t.lwl_2); |
| CHECK_EQ(static_cast<int32_t>(0x11223344), t.lwl_3); |
| |
| CHECK_EQ(static_cast<int32_t>(0x11223344), t.lwr_0); |
| CHECK_EQ(static_cast<int32_t>(0xAA112233), t.lwr_1); |
| CHECK_EQ(static_cast<int32_t>(0xAABB1122), t.lwr_2); |
| CHECK_EQ(static_cast<int32_t>(0xAABBCC11), t.lwr_3); |
| |
| CHECK_EQ(static_cast<int32_t>(0x112233AA), t.swl_0); |
| CHECK_EQ(static_cast<int32_t>(0x1122AABB), t.swl_1); |
| CHECK_EQ(static_cast<int32_t>(0x11AABBCC), t.swl_2); |
| CHECK_EQ(static_cast<int32_t>(0xAABBCCDD), t.swl_3); |
| |
| CHECK_EQ(static_cast<int32_t>(0xAABBCCDD), t.swr_0); |
| CHECK_EQ(static_cast<int32_t>(0xBBCCDD44), t.swr_1); |
| CHECK_EQ(static_cast<int32_t>(0xCCDD3344), t.swr_2); |
| CHECK_EQ(static_cast<int32_t>(0xDD223344), t.swr_3); |
| #elif __BYTE_ORDER == __BIG_ENDIAN |
| CHECK_EQ(static_cast<int32_t>(0x11223344), t.lwl_0); |
| CHECK_EQ(static_cast<int32_t>(0x223344DD), t.lwl_1); |
| CHECK_EQ(static_cast<int32_t>(0x3344CCDD), t.lwl_2); |
| CHECK_EQ(static_cast<int32_t>(0x44BBCCDD), t.lwl_3); |
| |
| CHECK_EQ(static_cast<int32_t>(0xAABBCC11), t.lwr_0); |
| CHECK_EQ(static_cast<int32_t>(0xAABB1122), t.lwr_1); |
| CHECK_EQ(static_cast<int32_t>(0xAA112233), t.lwr_2); |
| CHECK_EQ(static_cast<int32_t>(0x11223344), t.lwr_3); |
| |
| CHECK_EQ(static_cast<int32_t>(0xAABBCCDD), t.swl_0); |
| CHECK_EQ(static_cast<int32_t>(0x11AABBCC), t.swl_1); |
| CHECK_EQ(static_cast<int32_t>(0x1122AABB), t.swl_2); |
| CHECK_EQ(static_cast<int32_t>(0x112233AA), t.swl_3); |
| |
| CHECK_EQ(static_cast<int32_t>(0xDD223344), t.swr_0); |
| CHECK_EQ(static_cast<int32_t>(0xCCDD3344), t.swr_1); |
| CHECK_EQ(static_cast<int32_t>(0xBBCCDD44), t.swr_2); |
| CHECK_EQ(static_cast<int32_t>(0xAABBCCDD), t.swr_3); |
| #else |
| #error Unknown endianness |
| #endif |
| } |
| |
| |
| TEST(MIPS12) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| struct T { |
| int32_t x; |
| int32_t y; |
| int32_t y1; |
| int32_t y2; |
| int32_t y3; |
| int32_t y4; |
| }; |
| T t; |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| __ mov(t6, fp); // Save frame pointer. |
| __ mov(fp, a0); // Access struct T by fp. |
| __ lw(t0, MemOperand(a0, offsetof(T, y)) ); |
| __ lw(t3, MemOperand(a0, offsetof(T, y4)) ); |
| |
| __ addu(t1, t0, t3); |
| __ subu(t4, t0, t3); |
| __ nop(); |
| __ push(t0); // These instructions disappear after opt. |
| __ Pop(); |
| __ addu(t0, t0, t0); |
| __ nop(); |
| __ Pop(); // These instructions disappear after opt. |
| __ push(t3); |
| __ nop(); |
| __ push(t3); // These instructions disappear after opt. |
| __ pop(t3); |
| __ nop(); |
| __ push(t3); |
| __ pop(t4); |
| __ nop(); |
| __ sw(t0, MemOperand(fp, offsetof(T, y)) ); |
| __ lw(t0, MemOperand(fp, offsetof(T, y)) ); |
| __ nop(); |
| __ sw(t0, MemOperand(fp, offsetof(T, y)) ); |
| __ lw(t1, MemOperand(fp, offsetof(T, y)) ); |
| __ nop(); |
| __ push(t1); |
| __ lw(t1, MemOperand(fp, offsetof(T, y)) ); |
| __ pop(t1); |
| __ nop(); |
| __ push(t1); |
| __ lw(t2, MemOperand(fp, offsetof(T, y)) ); |
| __ pop(t1); |
| __ nop(); |
| __ push(t1); |
| __ lw(t2, MemOperand(fp, offsetof(T, y)) ); |
| __ pop(t2); |
| __ nop(); |
| __ push(t2); |
| __ lw(t2, MemOperand(fp, offsetof(T, y)) ); |
| __ pop(t1); |
| __ nop(); |
| __ push(t1); |
| __ lw(t2, MemOperand(fp, offsetof(T, y)) ); |
| __ pop(t3); |
| __ nop(); |
| |
| __ mov(fp, t6); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| t.x = 1; |
| t.y = 2; |
| t.y1 = 3; |
| t.y2 = 4; |
| t.y3 = 0XBABA; |
| t.y4 = 0xDEDA; |
| |
| f.Call(&t, 0, 0, 0, 0); |
| |
| CHECK_EQ(3, t.y1); |
| } |
| |
| |
| TEST(MIPS13) { |
| // Test Cvt_d_uw and Trunc_uw_d macros. |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| struct T { |
| double cvt_big_out; |
| double cvt_small_out; |
| uint32_t trunc_big_out; |
| uint32_t trunc_small_out; |
| uint32_t cvt_big_in; |
| uint32_t cvt_small_in; |
| }; |
| T t; |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| __ sw(t0, MemOperand(a0, offsetof(T, cvt_small_in))); |
| __ Cvt_d_uw(f10, t0, f4); |
| __ Sdc1(f10, MemOperand(a0, offsetof(T, cvt_small_out))); |
| |
| __ Trunc_uw_d(f10, f10, f4); |
| __ swc1(f10, MemOperand(a0, offsetof(T, trunc_small_out))); |
| |
| __ sw(t0, MemOperand(a0, offsetof(T, cvt_big_in))); |
| __ Cvt_d_uw(f8, t0, f4); |
| __ Sdc1(f8, MemOperand(a0, offsetof(T, cvt_big_out))); |
| |
| __ Trunc_uw_d(f8, f8, f4); |
| __ swc1(f8, MemOperand(a0, offsetof(T, trunc_big_out))); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| |
| t.cvt_big_in = 0xFFFFFFFF; |
| t.cvt_small_in = 333; |
| |
| f.Call(&t, 0, 0, 0, 0); |
| |
| CHECK_EQ(t.cvt_big_out, static_cast<double>(t.cvt_big_in)); |
| CHECK_EQ(t.cvt_small_out, static_cast<double>(t.cvt_small_in)); |
| |
| CHECK_EQ(static_cast<int>(t.trunc_big_out), static_cast<int>(t.cvt_big_in)); |
| CHECK_EQ(static_cast<int>(t.trunc_small_out), |
| static_cast<int>(t.cvt_small_in)); |
| } |
| |
| |
| TEST(MIPS14) { |
| // Test round, floor, ceil, trunc, cvt. |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| #define ROUND_STRUCT_ELEMENT(x) \ |
| uint32_t x##_isNaN2008; \ |
| int32_t x##_up_out; \ |
| int32_t x##_down_out; \ |
| int32_t neg_##x##_up_out; \ |
| int32_t neg_##x##_down_out; \ |
| uint32_t x##_err1_out; \ |
| uint32_t x##_err2_out; \ |
| uint32_t x##_err3_out; \ |
| uint32_t x##_err4_out; \ |
| int32_t x##_invalid_result; |
| |
| struct T { |
| double round_up_in; |
| double round_down_in; |
| double neg_round_up_in; |
| double neg_round_down_in; |
| double err1_in; |
| double err2_in; |
| double err3_in; |
| double err4_in; |
| |
| ROUND_STRUCT_ELEMENT(round) |
| ROUND_STRUCT_ELEMENT(floor) |
| ROUND_STRUCT_ELEMENT(ceil) |
| ROUND_STRUCT_ELEMENT(trunc) |
| ROUND_STRUCT_ELEMENT(cvt) |
| }; |
| T t; |
| |
| #undef ROUND_STRUCT_ELEMENT |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| // Save FCSR. |
| __ cfc1(a1, FCSR); |
| // Disable FPU exceptions. |
| __ ctc1(zero_reg, FCSR); |
| #define RUN_ROUND_TEST(x) \ |
| __ cfc1(t0, FCSR); \ |
| __ sw(t0, MemOperand(a0, offsetof(T, x##_isNaN2008))); \ |
| __ Ldc1(f0, MemOperand(a0, offsetof(T, round_up_in))); \ |
| __ x##_w_d(f0, f0); \ |
| __ swc1(f0, MemOperand(a0, offsetof(T, x##_up_out))); \ |
| \ |
| __ Ldc1(f0, MemOperand(a0, offsetof(T, round_down_in))); \ |
| __ x##_w_d(f0, f0); \ |
| __ swc1(f0, MemOperand(a0, offsetof(T, x##_down_out))); \ |
| \ |
| __ Ldc1(f0, MemOperand(a0, offsetof(T, neg_round_up_in))); \ |
| __ x##_w_d(f0, f0); \ |
| __ swc1(f0, MemOperand(a0, offsetof(T, neg_##x##_up_out))); \ |
| \ |
| __ Ldc1(f0, MemOperand(a0, offsetof(T, neg_round_down_in))); \ |
| __ x##_w_d(f0, f0); \ |
| __ swc1(f0, MemOperand(a0, offsetof(T, neg_##x##_down_out))); \ |
| \ |
| __ Ldc1(f0, MemOperand(a0, offsetof(T, err1_in))); \ |
| __ ctc1(zero_reg, FCSR); \ |
| __ x##_w_d(f0, f0); \ |
| __ cfc1(a2, FCSR); \ |
| __ sw(a2, MemOperand(a0, offsetof(T, x##_err1_out))); \ |
| \ |
| __ Ldc1(f0, MemOperand(a0, offsetof(T, err2_in))); \ |
| __ ctc1(zero_reg, FCSR); \ |
| __ x##_w_d(f0, f0); \ |
| __ cfc1(a2, FCSR); \ |
| __ sw(a2, MemOperand(a0, offsetof(T, x##_err2_out))); \ |
| \ |
| __ Ldc1(f0, MemOperand(a0, offsetof(T, err3_in))); \ |
| __ ctc1(zero_reg, FCSR); \ |
| __ x##_w_d(f0, f0); \ |
| __ cfc1(a2, FCSR); \ |
| __ sw(a2, MemOperand(a0, offsetof(T, x##_err3_out))); \ |
| \ |
| __ Ldc1(f0, MemOperand(a0, offsetof(T, err4_in))); \ |
| __ ctc1(zero_reg, FCSR); \ |
| __ x##_w_d(f0, f0); \ |
| __ cfc1(a2, FCSR); \ |
| __ sw(a2, MemOperand(a0, offsetof(T, x##_err4_out))); \ |
| __ swc1(f0, MemOperand(a0, offsetof(T, x##_invalid_result))); |
| |
| RUN_ROUND_TEST(round) |
| RUN_ROUND_TEST(floor) |
| RUN_ROUND_TEST(ceil) |
| RUN_ROUND_TEST(trunc) |
| RUN_ROUND_TEST(cvt) |
| |
| // Restore FCSR. |
| __ ctc1(a1, FCSR); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| |
| t.round_up_in = 123.51; |
| t.round_down_in = 123.49; |
| t.neg_round_up_in = -123.5; |
| t.neg_round_down_in = -123.49; |
| t.err1_in = 123.51; |
| t.err2_in = 1; |
| t.err3_in = static_cast<double>(1) + 0xFFFFFFFF; |
| t.err4_in = NAN; |
| |
| f.Call(&t, 0, 0, 0, 0); |
| |
| #define GET_FPU_ERR(x) (static_cast<int>(x & kFCSRFlagMask)) |
| #define CHECK_NAN2008(x) (x & kFCSRNaN2008FlagMask) |
| #define CHECK_ROUND_RESULT(type) \ |
| CHECK(GET_FPU_ERR(t.type##_err1_out) & kFCSRInexactFlagMask); \ |
| CHECK_EQ(0, GET_FPU_ERR(t.type##_err2_out)); \ |
| CHECK(GET_FPU_ERR(t.type##_err3_out) & kFCSRInvalidOpFlagMask); \ |
| CHECK(GET_FPU_ERR(t.type##_err4_out) & kFCSRInvalidOpFlagMask); \ |
| if (CHECK_NAN2008(t.type##_isNaN2008) && kArchVariant == kMips32r6) {\ |
| CHECK_EQ(static_cast<int32_t>(0), t.type##_invalid_result);\ |
| } else {\ |
| CHECK_EQ(static_cast<int32_t>(kFPUInvalidResult), t.type##_invalid_result);\ |
| } |
| |
| |
| CHECK_ROUND_RESULT(round); |
| CHECK_ROUND_RESULT(floor); |
| CHECK_ROUND_RESULT(ceil); |
| CHECK_ROUND_RESULT(cvt); |
| } |
| |
| |
| TEST(MIPS15) { |
| // Test chaining of label usages within instructions (issue 1644). |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| Assembler assm(AssemblerOptions{}); |
| |
| Label target; |
| __ beq(v0, v1, &target); |
| __ nop(); |
| __ bne(v0, v1, &target); |
| __ nop(); |
| __ bind(&target); |
| __ nop(); |
| } |
| |
| |
| // ----------------------mips32r6 specific tests---------------------- |
| TEST(seleqz_selnez) { |
| if (IsMipsArchVariant(kMips32r6)) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| struct Test { |
| int a; |
| int b; |
| int c; |
| int d; |
| double e; |
| double f; |
| double g; |
| double h; |
| float i; |
| float j; |
| float k; |
| float l; |
| }; |
| |
| Test test; |
| // Integer part of test. |
| __ addiu(t1, zero_reg, 1); // t1 = 1 |
| __ seleqz(t3, t1, zero_reg); // t3 = 1 |
| __ sw(t3, MemOperand(a0, offsetof(Test, a))); // a = 1 |
| __ seleqz(t2, t1, t1); // t2 = 0 |
| __ sw(t2, MemOperand(a0, offsetof(Test, b))); // b = 0 |
| __ selnez(t3, t1, zero_reg); // t3 = 1; |
| __ sw(t3, MemOperand(a0, offsetof(Test, c))); // c = 0 |
| __ selnez(t3, t1, t1); // t3 = 1 |
| __ sw(t3, MemOperand(a0, offsetof(Test, d))); // d = 1 |
| // Floating point part of test. |
| __ Ldc1(f0, MemOperand(a0, offsetof(Test, e))); // src |
| __ Ldc1(f2, MemOperand(a0, offsetof(Test, f))); // test |
| __ lwc1(f8, MemOperand(a0, offsetof(Test, i)) ); // src |
| __ lwc1(f10, MemOperand(a0, offsetof(Test, j)) ); // test |
| __ seleqz_d(f4, f0, f2); |
| __ selnez_d(f6, f0, f2); |
| __ seleqz_s(f12, f8, f10); |
| __ selnez_s(f14, f8, f10); |
| __ Sdc1(f4, MemOperand(a0, offsetof(Test, g))); // src |
| __ Sdc1(f6, MemOperand(a0, offsetof(Test, h))); // src |
| __ swc1(f12, MemOperand(a0, offsetof(Test, k)) ); // src |
| __ swc1(f14, MemOperand(a0, offsetof(Test, l)) ); // src |
| __ jr(ra); |
| __ nop(); |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| |
| (f.Call(&test, 0, 0, 0, 0)); |
| |
| CHECK_EQ(1, test.a); |
| CHECK_EQ(0, test.b); |
| CHECK_EQ(0, test.c); |
| CHECK_EQ(1, test.d); |
| |
| const int test_size = 3; |
| const int input_size = 5; |
| |
| double inputs_D[input_size] = {0.0, 65.2, -70.32, |
| 18446744073709551621.0, -18446744073709551621.0}; |
| double outputs_D[input_size] = {0.0, 65.2, -70.32, |
| 18446744073709551621.0, -18446744073709551621.0}; |
| double tests_D[test_size*2] = {2.8, 2.9, -2.8, -2.9, |
| 18446744073709551616.0, 18446744073709555712.0}; |
| float inputs_S[input_size] = {0.0, 65.2, -70.32, |
| 18446744073709551621.0, -18446744073709551621.0}; |
| float outputs_S[input_size] = {0.0, 65.2, -70.32, |
| 18446744073709551621.0, -18446744073709551621.0}; |
| float tests_S[test_size*2] = {2.9, 2.8, -2.9, -2.8, |
| 18446744073709551616.0, 18446746272732807168.0}; |
| for (int j=0; j < test_size; j+=2) { |
| for (int i=0; i < input_size; i++) { |
| test.e = inputs_D[i]; |
| test.f = tests_D[j]; |
| test.i = inputs_S[i]; |
| test.j = tests_S[j]; |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(outputs_D[i], test.g); |
| CHECK_EQ(0, test.h); |
| CHECK_EQ(outputs_S[i], test.k); |
| CHECK_EQ(0, test.l); |
| |
| test.f = tests_D[j+1]; |
| test.j = tests_S[j+1]; |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(0, test.g); |
| CHECK_EQ(outputs_D[i], test.h); |
| CHECK_EQ(0, test.k); |
| CHECK_EQ(outputs_S[i], test.l); |
| } |
| } |
| } |
| } |
| |
| |
| TEST(min_max) { |
| if (IsMipsArchVariant(kMips32r6)) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| struct TestFloat { |
| double a; |
| double b; |
| double c; |
| double d; |
| float e; |
| float f; |
| float g; |
| float h; |
| }; |
| |
| TestFloat test; |
| const double dnan = std::numeric_limits<double>::quiet_NaN(); |
| const double dinf = std::numeric_limits<double>::infinity(); |
| const double dminf = -std::numeric_limits<double>::infinity(); |
| const float fnan = std::numeric_limits<float>::quiet_NaN(); |
| const float finf = std::numeric_limits<float>::infinity(); |
| const float fminf = std::numeric_limits<float>::infinity(); |
| const int kTableLength = 13; |
| double inputsa[kTableLength] = {2.0, 3.0, dnan, 3.0, -0.0, 0.0, dinf, |
| dnan, 42.0, dinf, dminf, dinf, dnan}; |
| double inputsb[kTableLength] = {3.0, 2.0, 3.0, dnan, 0.0, -0.0, dnan, |
| dinf, dinf, 42.0, dinf, dminf, dnan}; |
| double outputsdmin[kTableLength] = {2.0, 2.0, 3.0, 3.0, -0.0, |
| -0.0, dinf, dinf, 42.0, 42.0, |
| dminf, dminf, dnan}; |
| double outputsdmax[kTableLength] = {3.0, 3.0, 3.0, 3.0, 0.0, 0.0, dinf, |
| dinf, dinf, dinf, dinf, dinf, dnan}; |
| |
| float inputse[kTableLength] = {2.0, 3.0, fnan, 3.0, -0.0, 0.0, finf, |
| fnan, 42.0, finf, fminf, finf, fnan}; |
| float inputsf[kTableLength] = {3.0, 2.0, 3.0, fnan, 0.0, -0.0, fnan, |
| finf, finf, 42.0, finf, fminf, fnan}; |
| float outputsfmin[kTableLength] = {2.0, 2.0, 3.0, 3.0, -0.0, |
| -0.0, finf, finf, 42.0, 42.0, |
| fminf, fminf, fnan}; |
| float outputsfmax[kTableLength] = {3.0, 3.0, 3.0, 3.0, 0.0, 0.0, finf, |
| finf, finf, finf, finf, finf, fnan}; |
| |
| __ Ldc1(f4, MemOperand(a0, offsetof(TestFloat, a))); |
| __ Ldc1(f8, MemOperand(a0, offsetof(TestFloat, b))); |
| __ lwc1(f2, MemOperand(a0, offsetof(TestFloat, e))); |
| __ lwc1(f6, MemOperand(a0, offsetof(TestFloat, f))); |
| __ min_d(f10, f4, f8); |
| __ max_d(f12, f4, f8); |
| __ min_s(f14, f2, f6); |
| __ max_s(f16, f2, f6); |
| __ Sdc1(f10, MemOperand(a0, offsetof(TestFloat, c))); |
| __ Sdc1(f12, MemOperand(a0, offsetof(TestFloat, d))); |
| __ swc1(f14, MemOperand(a0, offsetof(TestFloat, g))); |
| __ swc1(f16, MemOperand(a0, offsetof(TestFloat, h))); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| for (int i = 0; i < kTableLength; i++) { |
| test.a = inputsa[i]; |
| test.b = inputsb[i]; |
| test.e = inputse[i]; |
| test.f = inputsf[i]; |
| |
| f.Call(&test, 0, 0, 0, 0); |
| |
| CHECK_EQ(0, memcmp(&test.c, &outputsdmin[i], sizeof(test.c))); |
| CHECK_EQ(0, memcmp(&test.d, &outputsdmax[i], sizeof(test.d))); |
| CHECK_EQ(0, memcmp(&test.g, &outputsfmin[i], sizeof(test.g))); |
| CHECK_EQ(0, memcmp(&test.h, &outputsfmax[i], sizeof(test.h))); |
| } |
| } |
| } |
| |
| |
| TEST(rint_d) { |
| if (IsMipsArchVariant(kMips32r6)) { |
| const int kTableLength = 30; |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| struct TestFloat { |
| double a; |
| double b; |
| int fcsr; |
| }; |
| |
| TestFloat test; |
| double inputs[kTableLength] = {18446744073709551617.0, |
| 4503599627370496.0, -4503599627370496.0, |
| 1.26782468584154733584017312973E30, 1.44860108245951772690707170478E147, |
| 1.7976931348623157E+308, 6.27463370218383111104242366943E-307, |
| 309485009821345068724781056.89, |
| 2.1, 2.6, 2.5, 3.1, 3.6, 3.5, |
| -2.1, -2.6, -2.5, -3.1, -3.6, -3.5, |
| 37778931862957161709568.0, 37778931862957161709569.0, |
| 37778931862957161709580.0, 37778931862957161709581.0, |
| 37778931862957161709582.0, 37778931862957161709583.0, |
| 37778931862957161709584.0, 37778931862957161709585.0, |
| 37778931862957161709586.0, 37778931862957161709587.0}; |
| double outputs_RN[kTableLength] = {18446744073709551617.0, |
| 4503599627370496.0, -4503599627370496.0, |
| 1.26782468584154733584017312973E30, 1.44860108245951772690707170478E147, |
| 1.7976931348623157E308, 0, |
| 309485009821345068724781057.0, |
| 2.0, 3.0, 2.0, 3.0, 4.0, 4.0, |
| -2.0, -3.0, -2.0, -3.0, -4.0, -4.0, |
| 37778931862957161709568.0, 37778931862957161709569.0, |
| 37778931862957161709580.0, 37778931862957161709581.0, |
| 37778931862957161709582.0, 37778931862957161709583.0, |
| 37778931862957161709584.0, 37778931862957161709585.0, |
| 37778931862957161709586.0, 37778931862957161709587.0}; |
| double outputs_RZ[kTableLength] = {18446744073709551617.0, |
| 4503599627370496.0, -4503599627370496.0, |
| 1.26782468584154733584017312973E30, 1.44860108245951772690707170478E147, |
| 1.7976931348623157E308, 0, |
| 309485009821345068724781057.0, |
| 2.0, 2.0, 2.0, 3.0, 3.0, 3.0, |
| -2.0, -2.0, -2.0, -3.0, -3.0, -3.0, |
| 37778931862957161709568.0, 37778931862957161709569.0, |
| 37778931862957161709580.0, 37778931862957161709581.0, |
| 37778931862957161709582.0, 37778931862957161709583.0, |
| 37778931862957161709584.0, 37778931862957161709585.0, |
| 37778931862957161709586.0, 37778931862957161709587.0}; |
| double outputs_RP[kTableLength] = {18446744073709551617.0, |
| 4503599627370496.0, -4503599627370496.0, |
| 1.26782468584154733584017312973E30, 1.44860108245951772690707170478E147, |
| 1.7976931348623157E308, 1, |
| 309485009821345068724781057.0, |
| 3.0, 3.0, 3.0, 4.0, 4.0, 4.0, |
| -2.0, -2.0, -2.0, -3.0, -3.0, -3.0, |
| 37778931862957161709568.0, 37778931862957161709569.0, |
| 37778931862957161709580.0, 37778931862957161709581.0, |
| 37778931862957161709582.0, 37778931862957161709583.0, |
| 37778931862957161709584.0, 37778931862957161709585.0, |
| 37778931862957161709586.0, 37778931862957161709587.0}; |
| double outputs_RM[kTableLength] = {18446744073709551617.0, |
| 4503599627370496.0, -4503599627370496.0, |
| 1.26782468584154733584017312973E30, 1.44860108245951772690707170478E147, |
| 1.7976931348623157E308, 0, |
| 309485009821345068724781057.0, |
| 2.0, 2.0, 2.0, 3.0, 3.0, 3.0, |
| -3.0, -3.0, -3.0, -4.0, -4.0, -4.0, |
| 37778931862957161709568.0, 37778931862957161709569.0, |
| 37778931862957161709580.0, 37778931862957161709581.0, |
| 37778931862957161709582.0, 37778931862957161709583.0, |
| 37778931862957161709584.0, 37778931862957161709585.0, |
| 37778931862957161709586.0, 37778931862957161709587.0}; |
| int fcsr_inputs[4] = |
| {kRoundToNearest, kRoundToZero, kRoundToPlusInf, kRoundToMinusInf}; |
| double* outputs[4] = {outputs_RN, outputs_RZ, outputs_RP, outputs_RM}; |
| __ Ldc1(f4, MemOperand(a0, offsetof(TestFloat, a))); |
| __ lw(t0, MemOperand(a0, offsetof(TestFloat, fcsr)) ); |
| __ cfc1(t1, FCSR); |
| __ ctc1(t0, FCSR); |
| __ rint_d(f8, f4); |
| __ Sdc1(f8, MemOperand(a0, offsetof(TestFloat, b))); |
| __ ctc1(t1, FCSR); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| |
| for (int j = 0; j < 4; j++) { |
| test.fcsr = fcsr_inputs[j]; |
| for (int i = 0; i < kTableLength; i++) { |
| test.a = inputs[i]; |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.b, outputs[j][i]); |
| } |
| } |
| } |
| } |
| |
| |
| TEST(sel) { |
| if (IsMipsArchVariant(kMips32r6)) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| struct Test { |
| double dd; |
| double ds; |
| double dt; |
| float fd; |
| float fs; |
| float ft; |
| }; |
| |
| Test test; |
| __ Ldc1(f0, MemOperand(a0, offsetof(Test, dd))); // test |
| __ Ldc1(f2, MemOperand(a0, offsetof(Test, ds))); // src1 |
| __ Ldc1(f4, MemOperand(a0, offsetof(Test, dt))); // src2 |
| __ lwc1(f6, MemOperand(a0, offsetof(Test, fd)) ); // test |
| __ lwc1(f8, MemOperand(a0, offsetof(Test, fs)) ); // src1 |
| __ lwc1(f10, MemOperand(a0, offsetof(Test, ft)) ); // src2 |
| __ sel_d(f0, f2, f4); |
| __ sel_s(f6, f8, f10); |
| __ Sdc1(f0, MemOperand(a0, offsetof(Test, dd))); |
| __ swc1(f6, MemOperand(a0, offsetof(Test, fd)) ); |
| __ jr(ra); |
| __ nop(); |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| |
| const int test_size = 3; |
| const int input_size = 5; |
| |
| double inputs_dt[input_size] = {0.0, 65.2, -70.32, |
| 18446744073709551621.0, -18446744073709551621.0}; |
| double inputs_ds[input_size] = {0.1, 69.88, -91.325, |
| 18446744073709551625.0, -18446744073709551625.0}; |
| float inputs_ft[input_size] = {0.0, 65.2, -70.32, |
| 18446744073709551621.0, -18446744073709551621.0}; |
| float inputs_fs[input_size] = {0.1, 69.88, -91.325, |
| 18446744073709551625.0, -18446744073709551625.0}; |
| double tests_D[test_size*2] = {2.8, 2.9, -2.8, -2.9, |
| 18446744073709551616.0, 18446744073709555712.0}; |
| float tests_S[test_size*2] = {2.9, 2.8, -2.9, -2.8, |
| 18446744073709551616.0, 18446746272732807168.0}; |
| for (int j=0; j < test_size; j+=2) { |
| for (int i=0; i < input_size; i++) { |
| test.dt = inputs_dt[i]; |
| test.dd = tests_D[j]; |
| test.ds = inputs_ds[i]; |
| test.ft = inputs_ft[i]; |
| test.fd = tests_S[j]; |
| test.fs = inputs_fs[i]; |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.dd, inputs_ds[i]); |
| CHECK_EQ(test.fd, inputs_fs[i]); |
| |
| test.dd = tests_D[j+1]; |
| test.fd = tests_S[j+1]; |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.dd, inputs_dt[i]); |
| CHECK_EQ(test.fd, inputs_ft[i]); |
| } |
| } |
| } |
| } |
| |
| |
| TEST(rint_s) { |
| if (IsMipsArchVariant(kMips32r6)) { |
| const int kTableLength = 30; |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| struct TestFloat { |
| float a; |
| float b; |
| int fcsr; |
| }; |
| |
| TestFloat test; |
| float inputs[kTableLength] = {18446744073709551617.0, |
| 4503599627370496.0, -4503599627370496.0, |
| 1.26782468584154733584017312973E30, 1.44860108245951772690707170478E37, |
| 1.7976931348623157E+38, 6.27463370218383111104242366943E-37, |
| 309485009821345068724781056.89, |
| 2.1, 2.6, 2.5, 3.1, 3.6, 3.5, |
| -2.1, -2.6, -2.5, -3.1, -3.6, -3.5, |
| 37778931862957161709568.0, 37778931862957161709569.0, |
| 37778931862957161709580.0, 37778931862957161709581.0, |
| 37778931862957161709582.0, 37778931862957161709583.0, |
| 37778931862957161709584.0, 37778931862957161709585.0, |
| 37778931862957161709586.0, 37778931862957161709587.0}; |
| float outputs_RN[kTableLength] = {18446744073709551617.0, |
| 4503599627370496.0, -4503599627370496.0, |
| 1.26782468584154733584017312973E30, 1.44860108245951772690707170478E37, |
| 1.7976931348623157E38, 0, |
| 309485009821345068724781057.0, |
| 2.0, 3.0, 2.0, 3.0, 4.0, 4.0, |
| -2.0, -3.0, -2.0, -3.0, -4.0, -4.0, |
| 37778931862957161709568.0, 37778931862957161709569.0, |
| 37778931862957161709580.0, 37778931862957161709581.0, |
| 37778931862957161709582.0, 37778931862957161709583.0, |
| 37778931862957161709584.0, 37778931862957161709585.0, |
| 37778931862957161709586.0, 37778931862957161709587.0}; |
| float outputs_RZ[kTableLength] = {18446744073709551617.0, |
| 4503599627370496.0, -4503599627370496.0, |
| 1.26782468584154733584017312973E30, 1.44860108245951772690707170478E37, |
| 1.7976931348623157E38, 0, |
| 309485009821345068724781057.0, |
| 2.0, 2.0, 2.0, 3.0, 3.0, 3.0, |
| -2.0, -2.0, -2.0, -3.0, -3.0, -3.0, |
| 37778931862957161709568.0, 37778931862957161709569.0, |
| 37778931862957161709580.0, 37778931862957161709581.0, |
| 37778931862957161709582.0, 37778931862957161709583.0, |
| 37778931862957161709584.0, 37778931862957161709585.0, |
| 37778931862957161709586.0, 37778931862957161709587.0}; |
| float outputs_RP[kTableLength] = {18446744073709551617.0, |
| 4503599627370496.0, -4503599627370496.0, |
| 1.26782468584154733584017312973E30, 1.44860108245951772690707170478E37, |
| 1.7976931348623157E38, 1, |
| 309485009821345068724781057.0, |
| 3.0, 3.0, 3.0, 4.0, 4.0, 4.0, |
| -2.0, -2.0, -2.0, -3.0, -3.0, -3.0, |
| 37778931862957161709568.0, 37778931862957161709569.0, |
| 37778931862957161709580.0, 37778931862957161709581.0, |
| 37778931862957161709582.0, 37778931862957161709583.0, |
| 37778931862957161709584.0, 37778931862957161709585.0, |
| 37778931862957161709586.0, 37778931862957161709587.0}; |
| float outputs_RM[kTableLength] = {18446744073709551617.0, |
| 4503599627370496.0, -4503599627370496.0, |
| 1.26782468584154733584017312973E30, 1.44860108245951772690707170478E37, |
| 1.7976931348623157E38, 0, |
| 309485009821345068724781057.0, |
| 2.0, 2.0, 2.0, 3.0, 3.0, 3.0, |
| -3.0, -3.0, -3.0, -4.0, -4.0, -4.0, |
| 37778931862957161709568.0, 37778931862957161709569.0, |
| 37778931862957161709580.0, 37778931862957161709581.0, |
| 37778931862957161709582.0, 37778931862957161709583.0, |
| 37778931862957161709584.0, 37778931862957161709585.0, |
| 37778931862957161709586.0, 37778931862957161709587.0}; |
| int fcsr_inputs[4] = |
| {kRoundToNearest, kRoundToZero, kRoundToPlusInf, kRoundToMinusInf}; |
| float* outputs[4] = {outputs_RN, outputs_RZ, outputs_RP, outputs_RM}; |
| __ lwc1(f4, MemOperand(a0, offsetof(TestFloat, a)) ); |
| __ lw(t0, MemOperand(a0, offsetof(TestFloat, fcsr)) ); |
| __ cfc1(t1, FCSR); |
| __ ctc1(t0, FCSR); |
| __ rint_s(f8, f4); |
| __ swc1(f8, MemOperand(a0, offsetof(TestFloat, b)) ); |
| __ ctc1(t1, FCSR); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| |
| for (int j = 0; j < 4; j++) { |
| test.fcsr = fcsr_inputs[j]; |
| for (int i = 0; i < kTableLength; i++) { |
| test.a = inputs[i]; |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.b, outputs[j][i]); |
| } |
| } |
| } |
| } |
| |
| |
| TEST(Cvt_d_uw) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| struct TestStruct { |
| unsigned input; |
| uint64_t output; |
| }; |
| |
| unsigned inputs[] = {0x0, 0xFFFFFFFF, 0x80000000, 0x7FFFFFFF}; |
| |
| uint64_t outputs[] = {0x0, 0x41EFFFFFFFE00000, 0x41E0000000000000, |
| 0x41DFFFFFFFC00000}; |
| |
| int kTableLength = sizeof(inputs)/sizeof(inputs[0]); |
| |
| TestStruct test; |
| |
| __ lw(t1, MemOperand(a0, offsetof(TestStruct, input))); |
| __ Cvt_d_uw(f4, t1, f6); |
| __ Sdc1(f4, MemOperand(a0, offsetof(TestStruct, output))); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| for (int i = 0; i < kTableLength; i++) { |
| test.input = inputs[i]; |
| (f.Call(&test, 0, 0, 0, 0)); |
| // Check outputs |
| CHECK_EQ(test.output, outputs[i]); |
| } |
| } |
| |
| |
| TEST(mina_maxa) { |
| if (IsMipsArchVariant(kMips32r6)) { |
| const int kTableLength = 23; |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| const double dnan = std::numeric_limits<double>::quiet_NaN(); |
| const double dinf = std::numeric_limits<double>::infinity(); |
| const double dminf = -std::numeric_limits<double>::infinity(); |
| const float fnan = std::numeric_limits<float>::quiet_NaN(); |
| const float finf = std::numeric_limits<float>::infinity(); |
| const float fminf = std::numeric_limits<float>::infinity(); |
| |
| struct TestFloat { |
| double a; |
| double b; |
| double resd; |
| double resd1; |
| float c; |
| float d; |
| float resf; |
| float resf1; |
| }; |
| |
| TestFloat test; |
| double inputsa[kTableLength] = { |
| 5.3, 4.8, 6.1, 9.8, 9.8, 9.8, -10.0, -8.9, -9.8, -10.0, -8.9, -9.8, |
| dnan, 3.0, -0.0, 0.0, dinf, dnan, 42.0, dinf, dminf, dinf, dnan}; |
| double inputsb[kTableLength] = { |
| 4.8, 5.3, 6.1, -10.0, -8.9, -9.8, 9.8, 9.8, 9.8, -9.8, -11.2, -9.8, |
| 3.0, dnan, 0.0, -0.0, dnan, dinf, dinf, 42.0, dinf, dminf, dnan}; |
| double resd[kTableLength] = { |
| 4.8, 4.8, 6.1, 9.8, -8.9, -9.8, 9.8, -8.9, -9.8, -9.8, -8.9, -9.8, |
| 3.0, 3.0, -0.0, -0.0, dinf, dinf, 42.0, 42.0, dminf, dminf, dnan}; |
| double resd1[kTableLength] = { |
| 5.3, 5.3, 6.1, -10.0, 9.8, 9.8, -10.0, 9.8, 9.8, -10.0, -11.2, -9.8, |
| 3.0, 3.0, 0.0, 0.0, dinf, dinf, dinf, dinf, dinf, dinf, dnan}; |
| float inputsc[kTableLength] = { |
| 5.3, 4.8, 6.1, 9.8, 9.8, 9.8, -10.0, -8.9, -9.8, -10.0, -8.9, -9.8, |
| fnan, 3.0, -0.0, 0.0, finf, fnan, 42.0, finf, fminf, finf, fnan}; |
| float inputsd[kTableLength] = {4.8, 5.3, 6.1, -10.0, -8.9, -9.8, |
| 9.8, 9.8, 9.8, -9.8, -11.2, -9.8, |
| 3.0, fnan, -0.0, 0.0, fnan, finf, |
| finf, 42.0, finf, fminf, fnan}; |
| float resf[kTableLength] = { |
| 4.8, 4.8, 6.1, 9.8, -8.9, -9.8, 9.8, -8.9, -9.8, -9.8, -8.9, -9.8, |
| 3.0, 3.0, -0.0, -0.0, finf, finf, 42.0, 42.0, fminf, fminf, fnan}; |
| float resf1[kTableLength] = { |
| 5.3, 5.3, 6.1, -10.0, 9.8, 9.8, -10.0, 9.8, 9.8, -10.0, -11.2, -9.8, |
| 3.0, 3.0, 0.0, 0.0, finf, finf, finf, finf, finf, finf, fnan}; |
| |
| __ Ldc1(f2, MemOperand(a0, offsetof(TestFloat, a))); |
| __ Ldc1(f4, MemOperand(a0, offsetof(TestFloat, b))); |
| __ lwc1(f8, MemOperand(a0, offsetof(TestFloat, c)) ); |
| __ lwc1(f10, MemOperand(a0, offsetof(TestFloat, d)) ); |
| __ mina_d(f6, f2, f4); |
| __ mina_s(f12, f8, f10); |
| __ maxa_d(f14, f2, f4); |
| __ maxa_s(f16, f8, f10); |
| __ swc1(f12, MemOperand(a0, offsetof(TestFloat, resf)) ); |
| __ Sdc1(f6, MemOperand(a0, offsetof(TestFloat, resd))); |
| __ swc1(f16, MemOperand(a0, offsetof(TestFloat, resf1)) ); |
| __ Sdc1(f14, MemOperand(a0, offsetof(TestFloat, resd1))); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| for (int i = 0; i < kTableLength; i++) { |
| test.a = inputsa[i]; |
| test.b = inputsb[i]; |
| test.c = inputsc[i]; |
| test.d = inputsd[i]; |
| (f.Call(&test, 0, 0, 0, 0)); |
| if (i < kTableLength - 1) { |
| CHECK_EQ(test.resd, resd[i]); |
| CHECK_EQ(test.resf, resf[i]); |
| CHECK_EQ(test.resd1, resd1[i]); |
| CHECK_EQ(test.resf1, resf1[i]); |
| } else { |
| CHECK(std::isnan(test.resd)); |
| CHECK(std::isnan(test.resf)); |
| CHECK(std::isnan(test.resd1)); |
| CHECK(std::isnan(test.resf1)); |
| } |
| } |
| } |
| } |
| |
| |
| // ----------------------mips32r2 specific tests---------------------- |
| TEST(trunc_l) { |
| if (IsMipsArchVariant(kMips32r2) && IsFp64Mode()) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| const double dFPU64InvalidResult = static_cast<double>(kFPU64InvalidResult); |
| struct Test { |
| uint32_t isNaN2008; |
| double a; |
| float b; |
| int64_t c; // a trunc result |
| int64_t d; // b trunc result |
| }; |
| const int kTableLength = 15; |
| double inputs_D[kTableLength] = { |
| 2.1, 2.6, 2.5, 3.1, 3.6, 3.5, |
| -2.1, -2.6, -2.5, -3.1, -3.6, -3.5, |
| 2147483648.0, |
| std::numeric_limits<double>::quiet_NaN(), |
| std::numeric_limits<double>::infinity() |
| }; |
| float inputs_S[kTableLength] = { |
| 2.1, 2.6, 2.5, 3.1, 3.6, 3.5, |
| -2.1, -2.6, -2.5, -3.1, -3.6, -3.5, |
| 2147483648.0, |
| std::numeric_limits<float>::quiet_NaN(), |
| std::numeric_limits<float>::infinity() |
| }; |
| double outputs[kTableLength] = { |
| 2.0, 2.0, 2.0, 3.0, 3.0, 3.0, |
| -2.0, -2.0, -2.0, -3.0, -3.0, -3.0, |
| 2147483648.0, dFPU64InvalidResult, |
| dFPU64InvalidResult}; |
| double outputsNaN2008[kTableLength] = { |
| 2.0, 2.0, 2.0, 3.0, 3.0, 3.0, |
| -2.0, -2.0, -2.0, -3.0, -3.0, -3.0, |
| 2147483648.0, |
| 0, |
| dFPU64InvalidResult}; |
| |
| __ cfc1(t1, FCSR); |
| __ sw(t1, MemOperand(a0, offsetof(Test, isNaN2008))); |
| __ Ldc1(f4, MemOperand(a0, offsetof(Test, a))); |
| __ lwc1(f6, MemOperand(a0, offsetof(Test, b)) ); |
| __ trunc_l_d(f8, f4); |
| __ trunc_l_s(f10, f6); |
| __ Sdc1(f8, MemOperand(a0, offsetof(Test, c))); |
| __ Sdc1(f10, MemOperand(a0, offsetof(Test, d))); |
| __ jr(ra); |
| __ nop(); |
| Test test; |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| for (int i = 0; i < kTableLength; i++) { |
| test.a = inputs_D[i]; |
| test.b = inputs_S[i]; |
| (f.Call(&test, 0, 0, 0, 0)); |
| if ((test.isNaN2008 & kFCSRNaN2008FlagMask) && |
| kArchVariant == kMips32r6) { |
| CHECK_EQ(test.c, outputsNaN2008[i]); |
| } else { |
| CHECK_EQ(test.c, outputs[i]); |
| } |
| CHECK_EQ(test.d, test.c); |
| } |
| } |
| } |
| |
| |
| TEST(movz_movn) { |
| if (IsMipsArchVariant(kMips32r2)) { |
| const int kTableLength = 4; |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| struct TestFloat { |
| int32_t rt; |
| double a; |
| double b; |
| double bold; |
| double b1; |
| double bold1; |
| float c; |
| float d; |
| float dold; |
| float d1; |
| float dold1; |
| }; |
| |
| TestFloat test; |
| double inputs_D[kTableLength] = { |
| 5.3, -5.3, 5.3, -2.9 |
| }; |
| double inputs_S[kTableLength] = { |
| 4.8, 4.8, -4.8, -0.29 |
| }; |
| |
| float outputs_S[kTableLength] = { |
| 4.8, 4.8, -4.8, -0.29 |
| }; |
| double outputs_D[kTableLength] = { |
| 5.3, -5.3, 5.3, -2.9 |
| }; |
| |
| __ Ldc1(f2, MemOperand(a0, offsetof(TestFloat, a))); |
| __ lwc1(f6, MemOperand(a0, offsetof(TestFloat, c)) ); |
| __ lw(t0, MemOperand(a0, offsetof(TestFloat, rt)) ); |
| __ Move(f12, 0.0); |
| __ Move(f10, 0.0); |
| __ Move(f16, 0.0); |
| __ Move(f14, 0.0); |
| __ Sdc1(f12, MemOperand(a0, offsetof(TestFloat, bold))); |
| __ swc1(f10, MemOperand(a0, offsetof(TestFloat, dold)) ); |
| __ Sdc1(f16, MemOperand(a0, offsetof(TestFloat, bold1))); |
| __ swc1(f14, MemOperand(a0, offsetof(TestFloat, dold1)) ); |
| __ movz_s(f10, f6, t0); |
| __ movz_d(f12, f2, t0); |
| __ movn_s(f14, f6, t0); |
| __ movn_d(f16, f2, t0); |
| __ swc1(f10, MemOperand(a0, offsetof(TestFloat, d)) ); |
| __ Sdc1(f12, MemOperand(a0, offsetof(TestFloat, b))); |
| __ swc1(f14, MemOperand(a0, offsetof(TestFloat, d1)) ); |
| __ Sdc1(f16, MemOperand(a0, offsetof(TestFloat, b1))); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| for (int i = 0; i < kTableLength; i++) { |
| test.a = inputs_D[i]; |
| test.c = inputs_S[i]; |
| |
| test.rt = 1; |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.b, test.bold); |
| CHECK_EQ(test.d, test.dold); |
| CHECK_EQ(test.b1, outputs_D[i]); |
| CHECK_EQ(test.d1, outputs_S[i]); |
| |
| test.rt = 0; |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.b, outputs_D[i]); |
| CHECK_EQ(test.d, outputs_S[i]); |
| CHECK_EQ(test.b1, test.bold1); |
| CHECK_EQ(test.d1, test.dold1); |
| } |
| } |
| } |
| |
| |
| TEST(movt_movd) { |
| if (IsMipsArchVariant(kMips32r2)) { |
| const int kTableLength = 4; |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| |
| struct TestFloat { |
| double srcd; |
| double dstd; |
| double dstdold; |
| double dstd1; |
| double dstdold1; |
| float srcf; |
| float dstf; |
| float dstfold; |
| float dstf1; |
| float dstfold1; |
| int32_t cc; |
| int32_t fcsr; |
| }; |
| |
| TestFloat test; |
| double inputs_D[kTableLength] = { |
| 5.3, -5.3, 20.8, -2.9 |
| }; |
| double inputs_S[kTableLength] = { |
| 4.88, 4.8, -4.8, -0.29 |
| }; |
| |
| float outputs_S[kTableLength] = { |
| 4.88, 4.8, -4.8, -0.29 |
| }; |
| double outputs_D[kTableLength] = { |
| 5.3, -5.3, 20.8, -2.9 |
| }; |
| int condition_flags[8] = {0, 1, 2, 3, 4, 5, 6, 7}; |
| |
| for (int i = 0; i < kTableLength; i++) { |
| test.srcd = inputs_D[i]; |
| test.srcf = inputs_S[i]; |
| |
| for (int j = 0; j< 8; j++) { |
| test.cc = condition_flags[j]; |
| if (test.cc == 0) { |
| test.fcsr = 1 << 23; |
| } else { |
| test.fcsr = 1 << (24+condition_flags[j]); |
| } |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| __ Ldc1(f2, MemOperand(a0, offsetof(TestFloat, srcd))); |
| __ lwc1(f4, MemOperand(a0, offsetof(TestFloat, srcf)) ); |
| __ lw(t1, MemOperand(a0, offsetof(TestFloat, fcsr)) ); |
| __ cfc1(t0, FCSR); |
| __ ctc1(t1, FCSR); |
| __ li(t2, 0x0); |
| __ mtc1(t2, f12); |
| __ mtc1(t2, f10); |
| __ Sdc1(f10, MemOperand(a0, offsetof(TestFloat, dstdold))); |
| __ swc1(f12, MemOperand(a0, offsetof(TestFloat, dstfold)) ); |
| __ movt_s(f12, f4, test.cc); |
| __ movt_d(f10, f2, test.cc); |
| __ swc1(f12, MemOperand(a0, offsetof(TestFloat, dstf)) ); |
| __ Sdc1(f10, MemOperand(a0, offsetof(TestFloat, dstd))); |
| __ Sdc1(f10, MemOperand(a0, offsetof(TestFloat, dstdold1))); |
| __ swc1(f12, MemOperand(a0, offsetof(TestFloat, dstfold1)) ); |
| __ movf_s(f12, f4, test.cc); |
| __ movf_d(f10, f2, test.cc); |
| __ swc1(f12, MemOperand(a0, offsetof(TestFloat, dstf1)) ); |
| __ Sdc1(f10, MemOperand(a0, offsetof(TestFloat, dstd1))); |
| __ ctc1(t0, FCSR); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = |
| Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.dstf, outputs_S[i]); |
| CHECK_EQ(test.dstd, outputs_D[i]); |
| CHECK_EQ(test.dstf1, test.dstfold1); |
| CHECK_EQ(test.dstd1, test.dstdold1); |
| test.fcsr = 0; |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.dstf, test.dstfold); |
| CHECK_EQ(test.dstd, test.dstdold); |
| CHECK_EQ(test.dstf1, outputs_S[i]); |
| CHECK_EQ(test.dstd1, outputs_D[i]); |
| } |
| } |
| } |
| } |
| |
| |
| // ----------------------tests for all archs-------------------------- |
| TEST(cvt_w_d) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| struct Test { |
| double a; |
| int32_t b; |
| int32_t fcsr; |
| }; |
| const int kTableLength = 24; |
| double inputs[kTableLength] = { |
| 2.1, 2.6, 2.5, 3.1, 3.6, 3.5, |
| -2.1, -2.6, -2.5, -3.1, -3.6, -3.5, |
| 2147483637.0, 2147483638.0, 2147483639.0, |
| 2147483640.0, 2147483641.0, 2147483642.0, |
| 2147483643.0, 2147483644.0, 2147483645.0, |
| 2147483646.0, 2147483647.0, 2147483653.0 |
| }; |
| double outputs_RN[kTableLength] = { |
| 2.0, 3.0, 2.0, 3.0, 4.0, 4.0, |
| -2.0, -3.0, -2.0, -3.0, -4.0, -4.0, |
| 2147483637.0, 2147483638.0, 2147483639.0, |
| 2147483640.0, 2147483641.0, 2147483642.0, |
| 2147483643.0, 2147483644.0, 2147483645.0, |
| 2147483646.0, 2147483647.0, kFPUInvalidResult}; |
| double outputs_RZ[kTableLength] = { |
| 2.0, 2.0, 2.0, 3.0, 3.0, 3.0, |
| -2.0, -2.0, -2.0, -3.0, -3.0, -3.0, |
| 2147483637.0, 2147483638.0, 2147483639.0, |
| 2147483640.0, 2147483641.0, 2147483642.0, |
| 2147483643.0, 2147483644.0, 2147483645.0, |
| 2147483646.0, 2147483647.0, kFPUInvalidResult}; |
| double outputs_RP[kTableLength] = { |
| 3.0, 3.0, 3.0, 4.0, 4.0, 4.0, |
| -2.0, -2.0, -2.0, -3.0, -3.0, -3.0, |
| 2147483637.0, 2147483638.0, 2147483639.0, |
| 2147483640.0, 2147483641.0, 2147483642.0, |
| 2147483643.0, 2147483644.0, 2147483645.0, |
| 2147483646.0, 2147483647.0, kFPUInvalidResult}; |
| double outputs_RM[kTableLength] = { |
| 2.0, 2.0, 2.0, 3.0, 3.0, 3.0, |
| -3.0, -3.0, -3.0, -4.0, -4.0, -4.0, |
| 2147483637.0, 2147483638.0, 2147483639.0, |
| 2147483640.0, 2147483641.0, 2147483642.0, |
| 2147483643.0, 2147483644.0, 2147483645.0, |
| 2147483646.0, 2147483647.0, kFPUInvalidResult}; |
| int fcsr_inputs[4] = |
| {kRoundToNearest, kRoundToZero, kRoundToPlusInf, kRoundToMinusInf}; |
| double* outputs[4] = {outputs_RN, outputs_RZ, outputs_RP, outputs_RM}; |
| __ Ldc1(f4, MemOperand(a0, offsetof(Test, a))); |
| __ lw(t0, MemOperand(a0, offsetof(Test, fcsr)) ); |
| __ cfc1(t1, FCSR); |
| __ ctc1(t0, FCSR); |
| __ cvt_w_d(f8, f4); |
| __ swc1(f8, MemOperand(a0, offsetof(Test, b)) ); |
| __ ctc1(t1, FCSR); |
| __ jr(ra); |
| __ nop(); |
| Test test; |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| for (int j = 0; j < 4; j++) { |
| test.fcsr = fcsr_inputs[j]; |
| for (int i = 0; i < kTableLength; i++) { |
| test.a = inputs[i]; |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.b, outputs[j][i]); |
| } |
| } |
| } |
| |
| |
| TEST(trunc_w) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| struct Test { |
| uint32_t isNaN2008; |
| double a; |
| float b; |
| int32_t c; // a trunc result |
| int32_t d; // b trunc result |
| }; |
| const int kTableLength = 15; |
| double inputs_D[kTableLength] = { |
| 2.1, 2.6, 2.5, 3.1, 3.6, 3.5, |
| -2.1, -2.6, -2.5, -3.1, -3.6, -3.5, |
| 2147483648.0, |
| std::numeric_limits<double>::quiet_NaN(), |
| std::numeric_limits<double>::infinity() |
| }; |
| float inputs_S[kTableLength] = { |
| 2.1, 2.6, 2.5, 3.1, 3.6, 3.5, |
| -2.1, -2.6, -2.5, -3.1, -3.6, -3.5, |
| 2147483648.0, |
| std::numeric_limits<float>::quiet_NaN(), |
| std::numeric_limits<float>::infinity() |
| }; |
| double outputs[kTableLength] = { |
| 2.0, 2.0, 2.0, 3.0, 3.0, 3.0, |
| -2.0, -2.0, -2.0, -3.0, -3.0, -3.0, |
| kFPUInvalidResult, kFPUInvalidResult, |
| kFPUInvalidResult}; |
| double outputsNaN2008[kTableLength] = { |
| 2.0, 2.0, 2.0, 3.0, 3.0, 3.0, |
| -2.0, -2.0, -2.0, -3.0, -3.0, -3.0, |
| kFPUInvalidResult, |
| 0, |
| kFPUInvalidResult}; |
| |
| __ cfc1(t1, FCSR); |
| __ sw(t1, MemOperand(a0, offsetof(Test, isNaN2008))); |
| __ Ldc1(f4, MemOperand(a0, offsetof(Test, a))); |
| __ lwc1(f6, MemOperand(a0, offsetof(Test, b)) ); |
| __ trunc_w_d(f8, f4); |
| __ trunc_w_s(f10, f6); |
| __ swc1(f8, MemOperand(a0, offsetof(Test, c)) ); |
| __ swc1(f10, MemOperand(a0, offsetof(Test, d)) ); |
| __ jr(ra); |
| __ nop(); |
| Test test; |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| for (int i = 0; i < kTableLength; i++) { |
| test.a = inputs_D[i]; |
| test.b = inputs_S[i]; |
| (f.Call(&test, 0, 0, 0, 0)); |
| if ((test.isNaN2008 & kFCSRNaN2008FlagMask) && kArchVariant == kMips32r6) { |
| CHECK_EQ(test.c, outputsNaN2008[i]); |
| } else { |
| CHECK_EQ(test.c, outputs[i]); |
| } |
| CHECK_EQ(test.d, test.c); |
| } |
| } |
| |
| |
| TEST(round_w) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| struct Test { |
| uint32_t isNaN2008; |
| double a; |
| float b; |
| int32_t c; // a trunc result |
| int32_t d; // b trunc result |
| }; |
| const int kTableLength = 15; |
| double inputs_D[kTableLength] = { |
| 2.1, 2.6, 2.5, 3.1, 3.6, 3.5, |
| -2.1, -2.6, -2.5, -3.1, -3.6, -3.5, |
| 2147483648.0, |
| std::numeric_limits<double>::quiet_NaN(), |
| std::numeric_limits<double>::infinity() |
| }; |
| float inputs_S[kTableLength] = { |
| 2.1, 2.6, 2.5, 3.1, 3.6, 3.5, |
| -2.1, -2.6, -2.5, -3.1, -3.6, -3.5, |
| 2147483648.0, |
| std::numeric_limits<float>::quiet_NaN(), |
| std::numeric_limits<float>::infinity() |
| }; |
| double outputs[kTableLength] = { |
| 2.0, 3.0, 2.0, 3.0, 4.0, 4.0, |
| -2.0, -3.0, -2.0, -3.0, -4.0, -4.0, |
| kFPUInvalidResult, kFPUInvalidResult, |
| kFPUInvalidResult}; |
| double outputsNaN2008[kTableLength] = { |
| 2.0, 3.0, 2.0, 3.0, 4.0, 4.0, |
| -2.0, -3.0, -2.0, -3.0, -4.0, -4.0, |
| kFPUInvalidResult, 0, |
| kFPUInvalidResult}; |
| |
| __ cfc1(t1, FCSR); |
| __ sw(t1, MemOperand(a0, offsetof(Test, isNaN2008))); |
| __ Ldc1(f4, MemOperand(a0, offsetof(Test, a))); |
| __ lwc1(f6, MemOperand(a0, offsetof(Test, b)) ); |
| __ round_w_d(f8, f4); |
| __ round_w_s(f10, f6); |
| __ swc1(f8, MemOperand(a0, offsetof(Test, c)) ); |
| __ swc1(f10, MemOperand(a0, offsetof(Test, d)) ); |
| __ jr(ra); |
| __ nop(); |
| Test test; |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| for (int i = 0; i < kTableLength; i++) { |
| test.a = inputs_D[i]; |
| test.b = inputs_S[i]; |
| (f.Call(&test, 0, 0, 0, 0)); |
| if ((test.isNaN2008 & kFCSRNaN2008FlagMask) && kArchVariant == kMips32r6) { |
| CHECK_EQ(test.c, outputsNaN2008[i]); |
| } else { |
| CHECK_EQ(test.c, outputs[i]); |
| } |
| CHECK_EQ(test.d, test.c); |
| } |
| } |
| |
| |
| TEST(round_l) { |
| if (IsFp64Mode()) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| const double dFPU64InvalidResult = static_cast<double>(kFPU64InvalidResult); |
| struct Test { |
| uint32_t isNaN2008; |
| double a; |
| float b; |
| int64_t c; |
| int64_t d; |
| }; |
| const int kTableLength = 15; |
| double inputs_D[kTableLength] = { |
| 2.1, 2.6, 2.5, 3.1, 3.6, 3.5, |
| -2.1, -2.6, -2.5, -3.1, -3.6, -3.5, |
| 2147483648.0, |
| std::numeric_limits<double>::quiet_NaN(), |
| std::numeric_limits<double>::infinity() |
| }; |
| float inputs_S[kTableLength] = { |
| 2.1, 2.6, 2.5, 3.1, 3.6, 3.5, |
| -2.1, -2.6, -2.5, -3.1, -3.6, -3.5, |
| 2147483648.0, |
| std::numeric_limits<float>::quiet_NaN(), |
| std::numeric_limits<float>::infinity() |
| }; |
| double outputs[kTableLength] = { |
| 2.0, 3.0, 2.0, 3.0, 4.0, 4.0, |
| -2.0, -3.0, -2.0, -3.0, -4.0, -4.0, |
| 2147483648.0, dFPU64InvalidResult, |
| dFPU64InvalidResult}; |
| double outputsNaN2008[kTableLength] = { |
| 2.0, 3.0, 2.0, 3.0, 4.0, 4.0, |
| -2.0, -3.0, -2.0, -3.0, -4.0, -4.0, |
| 2147483648.0, |
| 0, |
| dFPU64InvalidResult}; |
| |
| __ cfc1(t1, FCSR); |
| __ sw(t1, MemOperand(a0, offsetof(Test, isNaN2008))); |
| __ Ldc1(f4, MemOperand(a0, offsetof(Test, a))); |
| __ lwc1(f6, MemOperand(a0, offsetof(Test, b)) ); |
| __ round_l_d(f8, f4); |
| __ round_l_s(f10, f6); |
| __ Sdc1(f8, MemOperand(a0, offsetof(Test, c))); |
| __ Sdc1(f10, MemOperand(a0, offsetof(Test, d))); |
| __ jr(ra); |
| __ nop(); |
| Test test; |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| for (int i = 0; i < kTableLength; i++) { |
| test.a = inputs_D[i]; |
| test.b = inputs_S[i]; |
| (f.Call(&test, 0, 0, 0, 0)); |
| if ((test.isNaN2008 & kFCSRNaN2008FlagMask) && |
| kArchVariant == kMips32r6) { |
| CHECK_EQ(test.c, outputsNaN2008[i]); |
| } else { |
| CHECK_EQ(test.c, outputs[i]); |
| } |
| CHECK_EQ(test.d, test.c); |
| } |
| } |
| } |
| |
| |
| TEST(sub) { |
| const int kTableLength = 12; |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| struct TestFloat { |
| float a; |
| float b; |
| float resultS; |
| double c; |
| double d; |
| double resultD; |
| }; |
| |
| TestFloat test; |
| double inputfs_D[kTableLength] = { |
| 5.3, 4.8, 2.9, -5.3, -4.8, -2.9, |
| 5.3, 4.8, 2.9, -5.3, -4.8, -2.9 |
| }; |
| double inputft_D[kTableLength] = { |
| 4.8, 5.3, 2.9, 4.8, 5.3, 2.9, |
| -4.8, -5.3, -2.9, -4.8, -5.3, -2.9 |
| }; |
| double outputs_D[kTableLength] = { |
| 0.5, -0.5, 0.0, -10.1, -10.1, -5.8, |
| 10.1, 10.1, 5.8, -0.5, 0.5, 0.0 |
| }; |
| float inputfs_S[kTableLength] = { |
| 5.3, 4.8, 2.9, -5.3, -4.8, -2.9, |
| 5.3, 4.8, 2.9, -5.3, -4.8, -2.9 |
| }; |
| float inputft_S[kTableLength] = { |
| 4.8, 5.3, 2.9, 4.8, 5.3, 2.9, |
| -4.8, -5.3, -2.9, -4.8, -5.3, -2.9 |
| }; |
| float outputs_S[kTableLength] = { |
| 0.5, -0.5, 0.0, -10.1, -10.1, -5.8, |
| 10.1, 10.1, 5.8, -0.5, 0.5, 0.0 |
| }; |
| __ lwc1(f2, MemOperand(a0, offsetof(TestFloat, a)) ); |
| __ lwc1(f4, MemOperand(a0, offsetof(TestFloat, b)) ); |
| __ Ldc1(f8, MemOperand(a0, offsetof(TestFloat, c))); |
| __ Ldc1(f10, MemOperand(a0, offsetof(TestFloat, d))); |
| __ sub_s(f6, f2, f4); |
| __ sub_d(f12, f8, f10); |
| __ swc1(f6, MemOperand(a0, offsetof(TestFloat, resultS)) ); |
| __ Sdc1(f12, MemOperand(a0, offsetof(TestFloat, resultD))); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| for (int i = 0; i < kTableLength; i++) { |
| test.a = inputfs_S[i]; |
| test.b = inputft_S[i]; |
| test.c = inputfs_D[i]; |
| test.d = inputft_D[i]; |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.resultS, outputs_S[i]); |
| CHECK_EQ(test.resultD, outputs_D[i]); |
| } |
| } |
| |
| |
| TEST(sqrt_rsqrt_recip) { |
| const int kTableLength = 4; |
| const double deltaDouble = 2E-15; |
| const float deltaFloat = 2E-7; |
| const float sqrt2_s = sqrt(2); |
| const double sqrt2_d = sqrt(2); |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| struct TestFloat { |
| float a; |
| float resultS; |
| float resultS1; |
| float resultS2; |
| double c; |
| double resultD; |
| double resultD1; |
| double resultD2; |
| }; |
| TestFloat test; |
| |
| double inputs_D[kTableLength] = { |
| 0.0L, 4.0L, 2.0L, 4e-28L |
| }; |
| |
| double outputs_D[kTableLength] = { |
| 0.0L, 2.0L, sqrt2_d, 2e-14L |
| }; |
| float inputs_S[kTableLength] = { |
| 0.0, 4.0, 2.0, 4e-28 |
| }; |
| |
| float outputs_S[kTableLength] = { |
| 0.0, 2.0, sqrt2_s, 2e-14 |
| }; |
| |
| |
| __ lwc1(f2, MemOperand(a0, offsetof(TestFloat, a)) ); |
| __ Ldc1(f8, MemOperand(a0, offsetof(TestFloat, c))); |
| __ sqrt_s(f6, f2); |
| __ sqrt_d(f12, f8); |
| |
| if (IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) { |
| __ rsqrt_d(f14, f8); |
| __ rsqrt_s(f16, f2); |
| __ recip_d(f18, f8); |
| __ recip_s(f4, f2); |
| } |
| __ swc1(f6, MemOperand(a0, offsetof(TestFloat, resultS)) ); |
| __ Sdc1(f12, MemOperand(a0, offsetof(TestFloat, resultD))); |
| |
| if (IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) { |
| __ swc1(f16, MemOperand(a0, offsetof(TestFloat, resultS1)) ); |
| __ Sdc1(f14, MemOperand(a0, offsetof(TestFloat, resultD1))); |
| __ swc1(f4, MemOperand(a0, offsetof(TestFloat, resultS2)) ); |
| __ Sdc1(f18, MemOperand(a0, offsetof(TestFloat, resultD2))); |
| } |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| |
| for (int i = 0; i < kTableLength; i++) { |
| float f1; |
| double d1; |
| test.a = inputs_S[i]; |
| test.c = inputs_D[i]; |
| |
| (f.Call(&test, 0, 0, 0, 0)); |
| |
| CHECK_EQ(test.resultS, outputs_S[i]); |
| CHECK_EQ(test.resultD, outputs_D[i]); |
| |
| if (IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) { |
| if (i != 0) { |
| f1 = test.resultS1 - 1.0F/outputs_S[i]; |
| f1 = (f1 < 0) ? f1 : -f1; |
| CHECK(f1 <= deltaFloat); |
| d1 = test.resultD1 - 1.0L/outputs_D[i]; |
| d1 = (d1 < 0) ? d1 : -d1; |
| CHECK(d1 <= deltaDouble); |
| f1 = test.resultS2 - 1.0F/inputs_S[i]; |
| f1 = (f1 < 0) ? f1 : -f1; |
| CHECK(f1 <= deltaFloat); |
| d1 = test.resultD2 - 1.0L/inputs_D[i]; |
| d1 = (d1 < 0) ? d1 : -d1; |
| CHECK(d1 <= deltaDouble); |
| } else { |
| CHECK_EQ(test.resultS1, 1.0F/outputs_S[i]); |
| CHECK_EQ(test.resultD1, 1.0L/outputs_D[i]); |
| CHECK_EQ(test.resultS2, 1.0F/inputs_S[i]); |
| CHECK_EQ(test.resultD2, 1.0L/inputs_D[i]); |
| } |
| } |
| } |
| } |
| |
| |
| TEST(neg) { |
| const int kTableLength = 3; |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| struct TestFloat { |
| float a; |
| float resultS; |
| double c; |
| double resultD; |
| }; |
| |
| TestFloat test; |
| double inputs_D[kTableLength] = { |
| 0.0, 4.0, -2.0 |
| }; |
| |
| double outputs_D[kTableLength] = { |
| 0.0, -4.0, 2.0 |
| }; |
| float inputs_S[kTableLength] = { |
| 0.0, 4.0, -2.0 |
| }; |
| |
| float outputs_S[kTableLength] = { |
| 0.0, -4.0, 2.0 |
| }; |
| __ lwc1(f2, MemOperand(a0, offsetof(TestFloat, a)) ); |
| __ Ldc1(f8, MemOperand(a0, offsetof(TestFloat, c))); |
| __ neg_s(f6, f2); |
| __ neg_d(f12, f8); |
| __ swc1(f6, MemOperand(a0, offsetof(TestFloat, resultS)) ); |
| __ Sdc1(f12, MemOperand(a0, offsetof(TestFloat, resultD))); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| for (int i = 0; i < kTableLength; i++) { |
| test.a = inputs_S[i]; |
| test.c = inputs_D[i]; |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.resultS, outputs_S[i]); |
| CHECK_EQ(test.resultD, outputs_D[i]); |
| } |
| } |
| |
| |
| TEST(mul) { |
| const int kTableLength = 4; |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| struct TestFloat { |
| float a; |
| float b; |
| float resultS; |
| double c; |
| double d; |
| double resultD; |
| }; |
| |
| TestFloat test; |
| double inputfs_D[kTableLength] = { |
| 5.3, -5.3, 5.3, -2.9 |
| }; |
| double inputft_D[kTableLength] = { |
| 4.8, 4.8, -4.8, -0.29 |
| }; |
| |
| float inputfs_S[kTableLength] = { |
| 5.3, -5.3, 5.3, -2.9 |
| }; |
| float inputft_S[kTableLength] = { |
| 4.8, 4.8, -4.8, -0.29 |
| }; |
| |
| __ lwc1(f2, MemOperand(a0, offsetof(TestFloat, a)) ); |
| __ lwc1(f4, MemOperand(a0, offsetof(TestFloat, b)) ); |
| __ Ldc1(f6, MemOperand(a0, offsetof(TestFloat, c))); |
| __ Ldc1(f8, MemOperand(a0, offsetof(TestFloat, d))); |
| __ mul_s(f10, f2, f4); |
| __ mul_d(f12, f6, f8); |
| __ swc1(f10, MemOperand(a0, offsetof(TestFloat, resultS)) ); |
| __ Sdc1(f12, MemOperand(a0, offsetof(TestFloat, resultD))); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| for (int i = 0; i < kTableLength; i++) { |
| test.a = inputfs_S[i]; |
| test.b = inputft_S[i]; |
| test.c = inputfs_D[i]; |
| test.d = inputft_D[i]; |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.resultS, inputfs_S[i]*inputft_S[i]); |
| CHECK_EQ(test.resultD, inputfs_D[i]*inputft_D[i]); |
| } |
| } |
| |
| |
| TEST(mov) { |
| const int kTableLength = 4; |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| struct TestFloat { |
| double a; |
| double b; |
| float c; |
| float d; |
| }; |
| |
| TestFloat test; |
| double inputs_D[kTableLength] = { |
| 5.3, -5.3, 5.3, -2.9 |
| }; |
| double inputs_S[kTableLength] = { |
| 4.8, 4.8, -4.8, -0.29 |
| }; |
| |
| float outputs_S[kTableLength] = { |
| 4.8, 4.8, -4.8, -0.29 |
| }; |
| double outputs_D[kTableLength] = { |
| 5.3, -5.3, 5.3, -2.9 |
| }; |
| |
| __ Ldc1(f4, MemOperand(a0, offsetof(TestFloat, a))); |
| __ lwc1(f6, MemOperand(a0, offsetof(TestFloat, c)) ); |
| __ mov_s(f8, f6); |
| __ mov_d(f10, f4); |
| __ swc1(f8, MemOperand(a0, offsetof(TestFloat, d)) ); |
| __ Sdc1(f10, MemOperand(a0, offsetof(TestFloat, b))); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| for (int i = 0; i < kTableLength; i++) { |
| test.a = inputs_D[i]; |
| test.c = inputs_S[i]; |
| |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.b, outputs_D[i]); |
| CHECK_EQ(test.d, outputs_S[i]); |
| } |
| } |
| |
| |
| TEST(floor_w) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| struct Test { |
| uint32_t isNaN2008; |
| double a; |
| float b; |
| int32_t c; // a floor result |
| int32_t d; // b floor result |
| }; |
| const int kTableLength = 15; |
| double inputs_D[kTableLength] = { |
| 2.1, 2.6, 2.5, 3.1, 3.6, 3.5, |
| -2.1, -2.6, -2.5, -3.1, -3.6, -3.5, |
| 2147483648.0, |
| std::numeric_limits<double>::quiet_NaN(), |
| std::numeric_limits<double>::infinity() |
| }; |
| float inputs_S[kTableLength] = { |
| 2.1, 2.6, 2.5, 3.1, 3.6, 3.5, |
| -2.1, -2.6, -2.5, -3.1, -3.6, -3.5, |
| 2147483648.0, |
| std::numeric_limits<float>::quiet_NaN(), |
| std::numeric_limits<float>::infinity() |
| }; |
| double outputs[kTableLength] = { |
| 2.0, 2.0, 2.0, 3.0, 3.0, 3.0, |
| -3.0, -3.0, -3.0, -4.0, -4.0, -4.0, |
| kFPUInvalidResult, kFPUInvalidResult, |
| kFPUInvalidResult}; |
| double outputsNaN2008[kTableLength] = { |
| 2.0, 2.0, 2.0, 3.0, 3.0, 3.0, |
| -3.0, -3.0, -3.0, -4.0, -4.0, -4.0, |
| kFPUInvalidResult, |
| 0, |
| kFPUInvalidResult}; |
| |
| __ cfc1(t1, FCSR); |
| __ sw(t1, MemOperand(a0, offsetof(Test, isNaN2008))); |
| __ Ldc1(f4, MemOperand(a0, offsetof(Test, a))); |
| __ lwc1(f6, MemOperand(a0, offsetof(Test, b)) ); |
| __ floor_w_d(f8, f4); |
| __ floor_w_s(f10, f6); |
| __ swc1(f8, MemOperand(a0, offsetof(Test, c)) ); |
| __ swc1(f10, MemOperand(a0, offsetof(Test, d)) ); |
| __ jr(ra); |
| __ nop(); |
| Test test; |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| for (int i = 0; i < kTableLength; i++) { |
| test.a = inputs_D[i]; |
| test.b = inputs_S[i]; |
| (f.Call(&test, 0, 0, 0, 0)); |
| if ((test.isNaN2008 & kFCSRNaN2008FlagMask) && kArchVariant == kMips32r6) { |
| CHECK_EQ(test.c, outputsNaN2008[i]); |
| } else { |
| CHECK_EQ(test.c, outputs[i]); |
| } |
| CHECK_EQ(test.d, test.c); |
| } |
| } |
| |
| |
| TEST(floor_l) { |
| if (IsFp64Mode()) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| const double dFPU64InvalidResult = static_cast<double>(kFPU64InvalidResult); |
| struct Test { |
| uint32_t isNaN2008; |
| double a; |
| float b; |
| int64_t c; |
| int64_t d; |
| }; |
| const int kTableLength = 15; |
| double inputs_D[kTableLength] = { |
| 2.1, 2.6, 2.5, 3.1, 3.6, 3.5, |
| -2.1, -2.6, -2.5, -3.1, -3.6, -3.5, |
| 2147483648.0, |
| std::numeric_limits<double>::quiet_NaN(), |
| std::numeric_limits<double>::infinity() |
| }; |
| float inputs_S[kTableLength] = { |
| 2.1, 2.6, 2.5, 3.1, 3.6, 3.5, |
| -2.1, -2.6, -2.5, -3.1, -3.6, -3.5, |
| 2147483648.0, |
| std::numeric_limits<float>::quiet_NaN(), |
| std::numeric_limits<float>::infinity() |
| }; |
| double outputs[kTableLength] = { |
| 2.0, 2.0, 2.0, 3.0, 3.0, 3.0, |
| -3.0, -3.0, -3.0, -4.0, -4.0, -4.0, |
| 2147483648.0, dFPU64InvalidResult, |
| dFPU64InvalidResult}; |
| double outputsNaN2008[kTableLength] = { |
| 2.0, 2.0, 2.0, 3.0, 3.0, 3.0, |
| -3.0, -3.0, -3.0, -4.0, -4.0, -4.0, |
| 2147483648.0, |
| 0, |
| dFPU64InvalidResult}; |
| |
| __ cfc1(t1, FCSR); |
| __ sw(t1, MemOperand(a0, offsetof(Test, isNaN2008))); |
| __ Ldc1(f4, MemOperand(a0, offsetof(Test, a))); |
| __ lwc1(f6, MemOperand(a0, offsetof(Test, b)) ); |
| __ floor_l_d(f8, f4); |
| __ floor_l_s(f10, f6); |
| __ Sdc1(f8, MemOperand(a0, offsetof(Test, c))); |
| __ Sdc1(f10, MemOperand(a0, offsetof(Test, d))); |
| __ jr(ra); |
| __ nop(); |
| Test test; |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| for (int i = 0; i < kTableLength; i++) { |
| test.a = inputs_D[i]; |
| test.b = inputs_S[i]; |
| (f.Call(&test, 0, 0, 0, 0)); |
| if ((test.isNaN2008 & kFCSRNaN2008FlagMask) && |
| kArchVariant == kMips32r6) { |
| CHECK_EQ(test.c, outputsNaN2008[i]); |
| } else { |
| CHECK_EQ(test.c, outputs[i]); |
| } |
| CHECK_EQ(test.d, test.c); |
| } |
| } |
| } |
| |
| |
| TEST(ceil_w) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| struct Test { |
| uint32_t isNaN2008; |
| double a; |
| float b; |
| int32_t c; // a floor result |
| int32_t d; // b floor result |
| }; |
| const int kTableLength = 15; |
| double inputs_D[kTableLength] = { |
| 2.1, 2.6, 2.5, 3.1, 3.6, 3.5, |
| -2.1, -2.6, -2.5, -3.1, -3.6, -3.5, |
| 2147483648.0, |
| std::numeric_limits<double>::quiet_NaN(), |
| std::numeric_limits<double>::infinity() |
| }; |
| float inputs_S[kTableLength] = { |
| 2.1, 2.6, 2.5, 3.1, 3.6, 3.5, |
| -2.1, -2.6, -2.5, -3.1, -3.6, -3.5, |
| 2147483648.0, |
| std::numeric_limits<float>::quiet_NaN(), |
| std::numeric_limits<float>::infinity() |
| }; |
| double outputs[kTableLength] = { |
| 3.0, 3.0, 3.0, 4.0, 4.0, 4.0, |
| -2.0, -2.0, -2.0, -3.0, -3.0, -3.0, |
| kFPUInvalidResult, kFPUInvalidResult, |
| kFPUInvalidResult}; |
| double outputsNaN2008[kTableLength] = { |
| 3.0, 3.0, 3.0, 4.0, 4.0, 4.0, |
| -2.0, -2.0, -2.0, -3.0, -3.0, -3.0, |
| kFPUInvalidResult, |
| 0, |
| kFPUInvalidResult}; |
| |
| __ cfc1(t1, FCSR); |
| __ sw(t1, MemOperand(a0, offsetof(Test, isNaN2008))); |
| __ Ldc1(f4, MemOperand(a0, offsetof(Test, a))); |
| __ lwc1(f6, MemOperand(a0, offsetof(Test, b)) ); |
| __ ceil_w_d(f8, f4); |
| __ ceil_w_s(f10, f6); |
| __ swc1(f8, MemOperand(a0, offsetof(Test, c)) ); |
| __ swc1(f10, MemOperand(a0, offsetof(Test, d)) ); |
| __ jr(ra); |
| __ nop(); |
| Test test; |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| for (int i = 0; i < kTableLength; i++) { |
| test.a = inputs_D[i]; |
| test.b = inputs_S[i]; |
| (f.Call(&test, 0, 0, 0, 0)); |
| if ((test.isNaN2008 & kFCSRNaN2008FlagMask) && kArchVariant == kMips32r6) { |
| CHECK_EQ(test.c, outputsNaN2008[i]); |
| } else { |
| CHECK_EQ(test.c, outputs[i]); |
| } |
| CHECK_EQ(test.d, test.c); |
| } |
| } |
| |
| |
| TEST(ceil_l) { |
| if (IsFp64Mode()) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| const double dFPU64InvalidResult = static_cast<double>(kFPU64InvalidResult); |
| struct Test { |
| uint32_t isNaN2008; |
| double a; |
| float b; |
| int64_t c; |
| int64_t d; |
| }; |
| const int kTableLength = 15; |
| double inputs_D[kTableLength] = { |
| 2.1, 2.6, 2.5, 3.1, 3.6, 3.5, |
| -2.1, -2.6, -2.5, -3.1, -3.6, -3.5, |
| 2147483648.0, |
| std::numeric_limits<double>::quiet_NaN(), |
| std::numeric_limits<double>::infinity() |
| }; |
| float inputs_S[kTableLength] = { |
| 2.1, 2.6, 2.5, 3.1, 3.6, 3.5, |
| -2.1, -2.6, -2.5, -3.1, -3.6, -3.5, |
| 2147483648.0, |
| std::numeric_limits<float>::quiet_NaN(), |
| std::numeric_limits<float>::infinity() |
| }; |
| double outputs[kTableLength] = { |
| 3.0, 3.0, 3.0, 4.0, 4.0, 4.0, |
| -2.0, -2.0, -2.0, -3.0, -3.0, -3.0, |
| 2147483648.0, dFPU64InvalidResult, |
| dFPU64InvalidResult}; |
| double outputsNaN2008[kTableLength] = { |
| 3.0, 3.0, 3.0, 4.0, 4.0, 4.0, |
| -2.0, -2.0, -2.0, -3.0, -3.0, -3.0, |
| 2147483648.0, |
| 0, |
| dFPU64InvalidResult}; |
| |
| __ cfc1(t1, FCSR); |
| __ sw(t1, MemOperand(a0, offsetof(Test, isNaN2008))); |
| __ Ldc1(f4, MemOperand(a0, offsetof(Test, a))); |
| __ lwc1(f6, MemOperand(a0, offsetof(Test, b)) ); |
| __ ceil_l_d(f8, f4); |
| __ ceil_l_s(f10, f6); |
| __ Sdc1(f8, MemOperand(a0, offsetof(Test, c))); |
| __ Sdc1(f10, MemOperand(a0, offsetof(Test, d))); |
| __ jr(ra); |
| __ nop(); |
| Test test; |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| for (int i = 0; i < kTableLength; i++) { |
| test.a = inputs_D[i]; |
| test.b = inputs_S[i]; |
| (f.Call(&test, 0, 0, 0, 0)); |
| if ((test.isNaN2008 & kFCSRNaN2008FlagMask) && |
| kArchVariant == kMips32r6) { |
| CHECK_EQ(test.c, outputsNaN2008[i]); |
| } else { |
| CHECK_EQ(test.c, outputs[i]); |
| } |
| CHECK_EQ(test.d, test.c); |
| } |
| } |
| } |
| |
| |
| TEST(jump_tables1) { |
| // Test jump tables with forward jumps. |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| Assembler assm(AssemblerOptions{}); |
| |
| const int kNumCases = 512; |
| int values[kNumCases]; |
| isolate->random_number_generator()->NextBytes(values, sizeof(values)); |
| Label labels[kNumCases]; |
| |
| __ addiu(sp, sp, -4); |
| __ sw(ra, MemOperand(sp)); |
| |
| Label done; |
| { |
| __ BlockTrampolinePoolFor(kNumCases + 7); |
| |
| __ nal(); |
| __ nop(); |
| __ sll(at, a0, 2); |
| __ addu(at, at, ra); |
| __ lw(at, MemOperand(at, 5 * kInstrSize)); |
| __ jr(at); |
| __ nop(); |
| for (int i = 0; i < kNumCases; ++i) { |
| __ dd(&labels[i]); |
| } |
| } |
| |
| for (int i = 0; i < kNumCases; ++i) { |
| __ bind(&labels[i]); |
| __ lui(v0, (values[i] >> 16) & 0xFFFF); |
| __ ori(v0, v0, values[i] & 0xFFFF); |
| __ b(&done); |
| __ nop(); |
| } |
| |
| __ bind(&done); |
| __ lw(ra, MemOperand(sp)); |
| __ addiu(sp, sp, 4); |
| __ jr(ra); |
| __ nop(); |
| |
| CHECK_EQ(0, assm.UnboundLabelsCount()); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| #ifdef OBJECT_PRINT |
| code->Print(std::cout); |
| #endif |
| auto f = GeneratedCode<F1>::FromCode(*code); |
| for (int i = 0; i < kNumCases; ++i) { |
| int res = reinterpret_cast<int>(f.Call(i, 0, 0, 0, 0)); |
| ::printf("f(%d) = %d\n", i, res); |
| CHECK_EQ(values[i], res); |
| } |
| } |
| |
| |
| TEST(jump_tables2) { |
| // Test jump tables with backward jumps. |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| Assembler assm(AssemblerOptions{}); |
| |
| const int kNumCases = 512; |
| int values[kNumCases]; |
| isolate->random_number_generator()->NextBytes(values, sizeof(values)); |
| Label labels[kNumCases]; |
| |
| __ addiu(sp, sp, -4); |
| __ sw(ra, MemOperand(sp)); |
| |
| Label done, dispatch; |
| __ b(&dispatch); |
| __ nop(); |
| |
| for (int i = 0; i < kNumCases; ++i) { |
| __ bind(&labels[i]); |
| __ lui(v0, (values[i] >> 16) & 0xFFFF); |
| __ ori(v0, v0, values[i] & 0xFFFF); |
| __ b(&done); |
| __ nop(); |
| } |
| |
| __ bind(&dispatch); |
| { |
| __ BlockTrampolinePoolFor(kNumCases + 7); |
| |
| __ nal(); |
| __ nop(); |
| __ sll(at, a0, 2); |
| __ addu(at, at, ra); |
| __ lw(at, MemOperand(at, 5 * kInstrSize)); |
| __ jr(at); |
| __ nop(); |
| for (int i = 0; i < kNumCases; ++i) { |
| __ dd(&labels[i]); |
| } |
| } |
| |
| __ bind(&done); |
| __ lw(ra, MemOperand(sp)); |
| __ addiu(sp, sp, 4); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| #ifdef OBJECT_PRINT |
| code->Print(std::cout); |
| #endif |
| auto f = GeneratedCode<F1>::FromCode(*code); |
| for (int i = 0; i < kNumCases; ++i) { |
| int res = reinterpret_cast<int>(f.Call(i, 0, 0, 0, 0)); |
| ::printf("f(%d) = %d\n", i, res); |
| CHECK_EQ(values[i], res); |
| } |
| } |
| |
| |
| TEST(jump_tables3) { |
| // Test jump tables with backward jumps and embedded heap objects. |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| Assembler assm(AssemblerOptions{}); |
| |
| const int kNumCases = 256; |
| Handle<Object> values[kNumCases]; |
| for (int i = 0; i < kNumCases; ++i) { |
| double value = isolate->random_number_generator()->NextDouble(); |
| values[i] = isolate->factory()->NewHeapNumber(value, AllocationType::kOld); |
| } |
| Label labels[kNumCases]; |
| Object obj; |
| int32_t imm32; |
| |
| __ addiu(sp, sp, -4); |
| __ sw(ra, MemOperand(sp)); |
| |
| Label done, dispatch; |
| __ b(&dispatch); |
| |
| |
| for (int i = 0; i < kNumCases; ++i) { |
| __ bind(&labels[i]); |
| obj = *values[i]; |
| imm32 = obj.ptr(); |
| __ lui(v0, (imm32 >> 16) & 0xFFFF); |
| __ ori(v0, v0, imm32 & 0xFFFF); |
| __ b(&done); |
| __ nop(); |
| } |
| |
| __ bind(&dispatch); |
| { |
| __ BlockTrampolinePoolFor(kNumCases + 7); |
| |
| __ nal(); |
| __ nop(); |
| __ sll(at, a0, 2); |
| __ addu(at, at, ra); |
| __ lw(at, MemOperand(at, 5 * kInstrSize)); |
| __ jr(at); |
| __ nop(); |
| for (int i = 0; i < kNumCases; ++i) { |
| __ dd(&labels[i]); |
| } |
| } |
| |
| __ bind(&done); |
| __ lw(ra, MemOperand(sp)); |
| __ addiu(sp, sp, 4); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| #ifdef OBJECT_PRINT |
| code->Print(std::cout); |
| #endif |
| auto f = GeneratedCode<F1>::FromCode(*code); |
| for (int i = 0; i < kNumCases; ++i) { |
| Handle<Object> result( |
| Object(reinterpret_cast<Address>(f.Call(i, 0, 0, 0, 0))), isolate); |
| #ifdef OBJECT_PRINT |
| ::printf("f(%d) = ", i); |
| result->Print(std::cout); |
| ::printf("\n"); |
| #endif |
| CHECK(values[i].is_identical_to(result)); |
| } |
| } |
| |
| |
| TEST(BITSWAP) { |
| // Test BITSWAP |
| if (IsMipsArchVariant(kMips32r6)) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| struct T { |
| int32_t r1; |
| int32_t r2; |
| int32_t r3; |
| int32_t r4; |
| }; |
| T t; |
| |
| Assembler assm(AssemblerOptions{}); |
| |
| __ lw(a2, MemOperand(a0, offsetof(T, r1))); |
| __ nop(); |
| __ bitswap(a1, a2); |
| __ sw(a1, MemOperand(a0, offsetof(T, r1))); |
| |
| __ lw(a2, MemOperand(a0, offsetof(T, r2))); |
| __ nop(); |
| __ bitswap(a1, a2); |
| __ sw(a1, MemOperand(a0, offsetof(T, r2))); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| t.r1 = 0x781A15C3; |
| t.r2 = 0x8B71FCDE; |
| f.Call(&t, 0, 0, 0, 0); |
| |
| CHECK_EQ(static_cast<int32_t>(0x1E58A8C3), t.r1); |
| CHECK_EQ(static_cast<int32_t>(0xD18E3F7B), t.r2); |
| } |
| } |
| |
| |
| TEST(class_fmt) { |
| if (IsMipsArchVariant(kMips32r6)) { |
| // Test CLASS.fmt instruction. |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| struct T { |
| double dSignalingNan; |
| double dQuietNan; |
| double dNegInf; |
| double dNegNorm; |
| double dNegSubnorm; |
| double dNegZero; |
| double dPosInf; |
| double dPosNorm; |
| double dPosSubnorm; |
| double dPosZero; |
| float fSignalingNan; |
| float fQuietNan; |
| float fNegInf; |
| float fNegNorm; |
| float fNegSubnorm; |
| float fNegZero; |
| float fPosInf; |
| float fPosNorm; |
| float fPosSubnorm; |
| float fPosZero; |
| }; |
| T t; |
| |
| // Create a function that accepts &t, and loads, manipulates, and stores |
| // the doubles t.a ... t.f. |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| __ Ldc1(f4, MemOperand(a0, offsetof(T, dSignalingNan))); |
| __ class_d(f6, f4); |
| __ Sdc1(f6, MemOperand(a0, offsetof(T, dSignalingNan))); |
| |
| __ Ldc1(f4, MemOperand(a0, offsetof(T, dQuietNan))); |
| __ class_d(f6, f4); |
| __ Sdc1(f6, MemOperand(a0, offsetof(T, dQuietNan))); |
| |
| __ Ldc1(f4, MemOperand(a0, offsetof(T, dNegInf))); |
| __ class_d(f6, f4); |
| __ Sdc1(f6, MemOperand(a0, offsetof(T, dNegInf))); |
| |
| __ Ldc1(f4, MemOperand(a0, offsetof(T, dNegNorm))); |
| __ class_d(f6, f4); |
| __ Sdc1(f6, MemOperand(a0, offsetof(T, dNegNorm))); |
| |
| __ Ldc1(f4, MemOperand(a0, offsetof(T, dNegSubnorm))); |
| __ class_d(f6, f4); |
| __ Sdc1(f6, MemOperand(a0, offsetof(T, dNegSubnorm))); |
| |
| __ Ldc1(f4, MemOperand(a0, offsetof(T, dNegZero))); |
| __ class_d(f6, f4); |
| __ Sdc1(f6, MemOperand(a0, offsetof(T, dNegZero))); |
| |
| __ Ldc1(f4, MemOperand(a0, offsetof(T, dPosInf))); |
| __ class_d(f6, f4); |
| __ Sdc1(f6, MemOperand(a0, offsetof(T, dPosInf))); |
| |
| __ Ldc1(f4, MemOperand(a0, offsetof(T, dPosNorm))); |
| __ class_d(f6, f4); |
| __ Sdc1(f6, MemOperand(a0, offsetof(T, dPosNorm))); |
| |
| __ Ldc1(f4, MemOperand(a0, offsetof(T, dPosSubnorm))); |
| __ class_d(f6, f4); |
| __ Sdc1(f6, MemOperand(a0, offsetof(T, dPosSubnorm))); |
| |
| __ Ldc1(f4, MemOperand(a0, offsetof(T, dPosZero))); |
| __ class_d(f6, f4); |
| __ Sdc1(f6, MemOperand(a0, offsetof(T, dPosZero))); |
| |
| // Testing instruction CLASS.S |
| __ lwc1(f4, MemOperand(a0, offsetof(T, fSignalingNan))); |
| __ class_s(f6, f4); |
| __ swc1(f6, MemOperand(a0, offsetof(T, fSignalingNan))); |
| |
| __ lwc1(f4, MemOperand(a0, offsetof(T, fQuietNan))); |
| __ class_s(f6, f4); |
| __ swc1(f6, MemOperand(a0, offsetof(T, fQuietNan))); |
| |
| __ lwc1(f4, MemOperand(a0, offsetof(T, fNegInf))); |
| __ class_s(f6, f4); |
| __ swc1(f6, MemOperand(a0, offsetof(T, fNegInf))); |
| |
| __ lwc1(f4, MemOperand(a0, offsetof(T, fNegNorm))); |
| __ class_s(f6, f4); |
| __ swc1(f6, MemOperand(a0, offsetof(T, fNegNorm))); |
| |
| __ lwc1(f4, MemOperand(a0, offsetof(T, fNegSubnorm))); |
| __ class_s(f6, f4); |
| __ swc1(f6, MemOperand(a0, offsetof(T, fNegSubnorm))); |
| |
| __ lwc1(f4, MemOperand(a0, offsetof(T, fNegZero))); |
| __ class_s(f6, f4); |
| __ swc1(f6, MemOperand(a0, offsetof(T, fNegZero))); |
| |
| __ lwc1(f4, MemOperand(a0, offsetof(T, fPosInf))); |
| __ class_s(f6, f4); |
| __ swc1(f6, MemOperand(a0, offsetof(T, fPosInf))); |
| |
| __ lwc1(f4, MemOperand(a0, offsetof(T, fPosNorm))); |
| __ class_s(f6, f4); |
| __ swc1(f6, MemOperand(a0, offsetof(T, fPosNorm))); |
| |
| __ lwc1(f4, MemOperand(a0, offsetof(T, fPosSubnorm))); |
| __ class_s(f6, f4); |
| __ swc1(f6, MemOperand(a0, offsetof(T, fPosSubnorm))); |
| |
| __ lwc1(f4, MemOperand(a0, offsetof(T, fPosZero))); |
| __ class_s(f6, f4); |
| __ swc1(f6, MemOperand(a0, offsetof(T, fPosZero))); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| |
| t.dSignalingNan = std::numeric_limits<double>::signaling_NaN(); |
| t.dQuietNan = std::numeric_limits<double>::quiet_NaN(); |
| t.dNegInf = -1.0 / 0.0; |
| t.dNegNorm = -5.0; |
| t.dNegSubnorm = -DBL_MIN / 2.0; |
| t.dNegZero = -0.0; |
| t.dPosInf = 2.0 / 0.0; |
| t.dPosNorm = 275.35; |
| t.dPosSubnorm = DBL_MIN / 2.0; |
| t.dPosZero = +0.0; |
| // Float test values |
| |
| t.fSignalingNan = std::numeric_limits<float>::signaling_NaN(); |
| t.fQuietNan = std::numeric_limits<float>::quiet_NaN(); |
| t.fNegInf = -0.5/0.0; |
| t.fNegNorm = -FLT_MIN; |
| t.fNegSubnorm = -FLT_MIN / 1.5; |
| t.fNegZero = -0.0; |
| t.fPosInf = 100000.0 / 0.0; |
| t.fPosNorm = FLT_MAX; |
| t.fPosSubnorm = FLT_MIN / 20.0; |
| t.fPosZero = +0.0; |
| |
| f.Call(&t, 0, 0, 0, 0); |
| // Expected double results. |
| CHECK_EQ(bit_cast<int64_t>(t.dSignalingNan), 0x001); |
| CHECK_EQ(bit_cast<int64_t>(t.dQuietNan), 0x002); |
| CHECK_EQ(bit_cast<int64_t>(t.dNegInf), 0x004); |
| CHECK_EQ(bit_cast<int64_t>(t.dNegNorm), 0x008); |
| CHECK_EQ(bit_cast<int64_t>(t.dNegSubnorm), 0x010); |
| CHECK_EQ(bit_cast<int64_t>(t.dNegZero), 0x020); |
| CHECK_EQ(bit_cast<int64_t>(t.dPosInf), 0x040); |
| CHECK_EQ(bit_cast<int64_t>(t.dPosNorm), 0x080); |
| CHECK_EQ(bit_cast<int64_t>(t.dPosSubnorm), 0x100); |
| CHECK_EQ(bit_cast<int64_t>(t.dPosZero), 0x200); |
| |
| // Expected float results. |
| CHECK_EQ(bit_cast<int32_t>(t.fSignalingNan), 0x001); |
| CHECK_EQ(bit_cast<int32_t>(t.fQuietNan), 0x002); |
| CHECK_EQ(bit_cast<int32_t>(t.fNegInf), 0x004); |
| CHECK_EQ(bit_cast<int32_t>(t.fNegNorm), 0x008); |
| CHECK_EQ(bit_cast<int32_t>(t.fNegSubnorm), 0x010); |
| CHECK_EQ(bit_cast<int32_t>(t.fNegZero), 0x020); |
| CHECK_EQ(bit_cast<int32_t>(t.fPosInf), 0x040); |
| CHECK_EQ(bit_cast<int32_t>(t.fPosNorm), 0x080); |
| CHECK_EQ(bit_cast<int32_t>(t.fPosSubnorm), 0x100); |
| CHECK_EQ(bit_cast<int32_t>(t.fPosZero), 0x200); |
| } |
| } |
| |
| |
| TEST(ABS) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| struct TestFloat { |
| int64_t fir; |
| double a; |
| float b; |
| double fcsr; |
| }; |
| |
| TestFloat test; |
| |
| // Save FIR. |
| __ cfc1(a1, FCSR); |
| // Disable FPU exceptions. |
| __ ctc1(zero_reg, FCSR); |
| |
| __ Ldc1(f4, MemOperand(a0, offsetof(TestFloat, a))); |
| __ abs_d(f10, f4); |
| __ Sdc1(f10, MemOperand(a0, offsetof(TestFloat, a))); |
| |
| __ lwc1(f4, MemOperand(a0, offsetof(TestFloat, b))); |
| __ abs_s(f10, f4); |
| __ swc1(f10, MemOperand(a0, offsetof(TestFloat, b))); |
| |
| // Restore FCSR. |
| __ ctc1(a1, FCSR); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| test.a = -2.0; |
| test.b = -2.0; |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.a, 2.0); |
| CHECK_EQ(test.b, 2.0); |
| |
| test.a = 2.0; |
| test.b = 2.0; |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.a, 2.0); |
| CHECK_EQ(test.b, 2.0); |
| |
| // Testing biggest positive number |
| test.a = std::numeric_limits<double>::max(); |
| test.b = std::numeric_limits<float>::max(); |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.a, std::numeric_limits<double>::max()); |
| CHECK_EQ(test.b, std::numeric_limits<float>::max()); |
| |
| // Testing smallest negative number |
| test.a = -std::numeric_limits<double>::max(); // lowest() |
| test.b = -std::numeric_limits<float>::max(); // lowest() |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.a, std::numeric_limits<double>::max()); |
| CHECK_EQ(test.b, std::numeric_limits<float>::max()); |
| |
| // Testing smallest positive number |
| test.a = -std::numeric_limits<double>::min(); |
| test.b = -std::numeric_limits<float>::min(); |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.a, std::numeric_limits<double>::min()); |
| CHECK_EQ(test.b, std::numeric_limits<float>::min()); |
| |
| // Testing infinity |
| test.a = -std::numeric_limits<double>::max() |
| / std::numeric_limits<double>::min(); |
| test.b = -std::numeric_limits<float>::max() |
| / std::numeric_limits<float>::min(); |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.a, std::numeric_limits<double>::max() |
| / std::numeric_limits<double>::min()); |
| CHECK_EQ(test.b, std::numeric_limits<float>::max() |
| / std::numeric_limits<float>::min()); |
| |
| test.a = std::numeric_limits<double>::quiet_NaN(); |
| test.b = std::numeric_limits<float>::quiet_NaN(); |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK(std::isnan(test.a)); |
| CHECK(std::isnan(test.b)); |
| |
| test.a = std::numeric_limits<double>::signaling_NaN(); |
| test.b = std::numeric_limits<float>::signaling_NaN(); |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK(std::isnan(test.a)); |
| CHECK(std::isnan(test.b)); |
| } |
| |
| |
| TEST(ADD_FMT) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| struct TestFloat { |
| double a; |
| double b; |
| double c; |
| float fa; |
| float fb; |
| float fc; |
| }; |
| |
| TestFloat test; |
| |
| __ Ldc1(f4, MemOperand(a0, offsetof(TestFloat, a))); |
| __ Ldc1(f8, MemOperand(a0, offsetof(TestFloat, b))); |
| __ add_d(f10, f8, f4); |
| __ Sdc1(f10, MemOperand(a0, offsetof(TestFloat, c))); |
| |
| __ lwc1(f4, MemOperand(a0, offsetof(TestFloat, fa))); |
| __ lwc1(f8, MemOperand(a0, offsetof(TestFloat, fb))); |
| __ add_s(f10, f8, f4); |
| __ swc1(f10, MemOperand(a0, offsetof(TestFloat, fc))); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| test.a = 2.0; |
| test.b = 3.0; |
| test.fa = 2.0; |
| test.fb = 3.0; |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.c, 5.0); |
| CHECK_EQ(test.fc, 5.0); |
| |
| test.a = std::numeric_limits<double>::max(); |
| test.b = -std::numeric_limits<double>::max(); // lowest() |
| test.fa = std::numeric_limits<float>::max(); |
| test.fb = -std::numeric_limits<float>::max(); // lowest() |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.c, 0.0); |
| CHECK_EQ(test.fc, 0.0); |
| |
| test.a = std::numeric_limits<double>::max(); |
| test.b = std::numeric_limits<double>::max(); |
| test.fa = std::numeric_limits<float>::max(); |
| test.fb = std::numeric_limits<float>::max(); |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK(!std::isfinite(test.c)); |
| CHECK(!std::isfinite(test.fc)); |
| |
| test.a = 5.0; |
| test.b = std::numeric_limits<double>::signaling_NaN(); |
| test.fa = 5.0; |
| test.fb = std::numeric_limits<float>::signaling_NaN(); |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK(std::isnan(test.c)); |
| CHECK(std::isnan(test.fc)); |
| } |
| |
| |
| TEST(C_COND_FMT) { |
| if ((IsMipsArchVariant(kMips32r1)) || (IsMipsArchVariant(kMips32r2))) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| struct TestFloat { |
| double dOp1; |
| double dOp2; |
| uint32_t dF; |
| uint32_t dUn; |
| uint32_t dEq; |
| uint32_t dUeq; |
| uint32_t dOlt; |
| uint32_t dUlt; |
| uint32_t dOle; |
| uint32_t dUle; |
| float fOp1; |
| float fOp2; |
| uint32_t fF; |
| uint32_t fUn; |
| uint32_t fEq; |
| uint32_t fUeq; |
| uint32_t fOlt; |
| uint32_t fUlt; |
| uint32_t fOle; |
| uint32_t fUle; |
| }; |
| |
| TestFloat test; |
| |
| __ li(t1, 1); |
| |
| __ Ldc1(f4, MemOperand(a0, offsetof(TestFloat, dOp1))); |
| __ Ldc1(f6, MemOperand(a0, offsetof(TestFloat, dOp2))); |
| |
| __ lwc1(f14, MemOperand(a0, offsetof(TestFloat, fOp1))); |
| __ lwc1(f16, MemOperand(a0, offsetof(TestFloat, fOp2))); |
| |
| __ mov(t2, zero_reg); |
| __ mov(t3, zero_reg); |
| __ c_d(F, f4, f6, 0); |
| __ c_s(F, f14, f16, 2); |
| __ movt(t2, t1, 0); |
| __ movt(t3, t1, 2); |
| __ sw(t2, MemOperand(a0, offsetof(TestFloat, dF)) ); |
| __ sw(t3, MemOperand(a0, offsetof(TestFloat, fF)) ); |
| |
| __ mov(t2, zero_reg); |
| __ mov(t3, zero_reg); |
| __ c_d(UN, f4, f6, 2); |
| __ c_s(UN, f14, f16, 4); |
| __ movt(t2, t1, 2); |
| __ movt(t3, t1, 4); |
| __ sw(t2, MemOperand(a0, offsetof(TestFloat, dUn)) ); |
| __ sw(t3, MemOperand(a0, offsetof(TestFloat, fUn)) ); |
| |
| __ mov(t2, zero_reg); |
| __ mov(t3, zero_reg); |
| __ c_d(EQ, f4, f6, 4); |
| __ c_s(EQ, f14, f16, 6); |
| __ movt(t2, t1, 4); |
| __ movt(t3, t1, 6); |
| __ sw(t2, MemOperand(a0, offsetof(TestFloat, dEq)) ); |
| __ sw(t3, MemOperand(a0, offsetof(TestFloat, fEq)) ); |
| |
| __ mov(t2, zero_reg); |
| __ mov(t3, zero_reg); |
| __ c_d(UEQ, f4, f6, 6); |
| __ c_s(UEQ, f14, f16, 0); |
| __ movt(t2, t1, 6); |
| __ movt(t3, t1, 0); |
| __ sw(t2, MemOperand(a0, offsetof(TestFloat, dUeq)) ); |
| __ sw(t3, MemOperand(a0, offsetof(TestFloat, fUeq)) ); |
| |
| __ mov(t2, zero_reg); |
| __ mov(t3, zero_reg); |
| __ c_d(OLT, f4, f6, 0); |
| __ c_s(OLT, f14, f16, 2); |
| __ movt(t2, t1, 0); |
| __ movt(t3, t1, 2); |
| __ sw(t2, MemOperand(a0, offsetof(TestFloat, dOlt)) ); |
| __ sw(t3, MemOperand(a0, offsetof(TestFloat, fOlt)) ); |
| |
| __ mov(t2, zero_reg); |
| __ mov(t3, zero_reg); |
| __ c_d(ULT, f4, f6, 2); |
| __ c_s(ULT, f14, f16, 4); |
| __ movt(t2, t1, 2); |
| __ movt(t3, t1, 4); |
| __ sw(t2, MemOperand(a0, offsetof(TestFloat, dUlt)) ); |
| __ sw(t3, MemOperand(a0, offsetof(TestFloat, fUlt)) ); |
| |
| __ mov(t2, zero_reg); |
| __ mov(t3, zero_reg); |
| __ c_d(OLE, f4, f6, 4); |
| __ c_s(OLE, f14, f16, 6); |
| __ movt(t2, t1, 4); |
| __ movt(t3, t1, 6); |
| __ sw(t2, MemOperand(a0, offsetof(TestFloat, dOle)) ); |
| __ sw(t3, MemOperand(a0, offsetof(TestFloat, fOle)) ); |
| |
| __ mov(t2, zero_reg); |
| __ mov(t3, zero_reg); |
| __ c_d(ULE, f4, f6, 6); |
| __ c_s(ULE, f14, f16, 0); |
| __ movt(t2, t1, 6); |
| __ movt(t3, t1, 0); |
| __ sw(t2, MemOperand(a0, offsetof(TestFloat, dUle)) ); |
| __ sw(t3, MemOperand(a0, offsetof(TestFloat, fUle)) ); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| test.dOp1 = 2.0; |
| test.dOp2 = 3.0; |
| test.fOp1 = 2.0; |
| test.fOp2 = 3.0; |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.dF, 0U); |
| CHECK_EQ(test.dUn, 0U); |
| CHECK_EQ(test.dEq, 0U); |
| CHECK_EQ(test.dUeq, 0U); |
| CHECK_EQ(test.dOlt, 1U); |
| CHECK_EQ(test.dUlt, 1U); |
| CHECK_EQ(test.dOle, 1U); |
| CHECK_EQ(test.dUle, 1U); |
| CHECK_EQ(test.fF, 0U); |
| CHECK_EQ(test.fUn, 0U); |
| CHECK_EQ(test.fEq, 0U); |
| CHECK_EQ(test.fUeq, 0U); |
| CHECK_EQ(test.fOlt, 1U); |
| CHECK_EQ(test.fUlt, 1U); |
| CHECK_EQ(test.fOle, 1U); |
| CHECK_EQ(test.fUle, 1U); |
| |
| test.dOp1 = std::numeric_limits<double>::max(); |
| test.dOp2 = std::numeric_limits<double>::min(); |
| test.fOp1 = std::numeric_limits<float>::min(); |
| test.fOp2 = -std::numeric_limits<float>::max(); // lowest() |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.dF, 0U); |
| CHECK_EQ(test.dUn, 0U); |
| CHECK_EQ(test.dEq, 0U); |
| CHECK_EQ(test.dUeq, 0U); |
| CHECK_EQ(test.dOlt, 0U); |
| CHECK_EQ(test.dUlt, 0U); |
| CHECK_EQ(test.dOle, 0U); |
| CHECK_EQ(test.dUle, 0U); |
| CHECK_EQ(test.fF, 0U); |
| CHECK_EQ(test.fUn, 0U); |
| CHECK_EQ(test.fEq, 0U); |
| CHECK_EQ(test.fUeq, 0U); |
| CHECK_EQ(test.fOlt, 0U); |
| CHECK_EQ(test.fUlt, 0U); |
| CHECK_EQ(test.fOle, 0U); |
| CHECK_EQ(test.fUle, 0U); |
| |
| test.dOp1 = -std::numeric_limits<double>::max(); // lowest() |
| test.dOp2 = -std::numeric_limits<double>::max(); // lowest() |
| test.fOp1 = std::numeric_limits<float>::max(); |
| test.fOp2 = std::numeric_limits<float>::max(); |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.dF, 0U); |
| CHECK_EQ(test.dUn, 0U); |
| CHECK_EQ(test.dEq, 1U); |
| CHECK_EQ(test.dUeq, 1U); |
| CHECK_EQ(test.dOlt, 0U); |
| CHECK_EQ(test.dUlt, 0U); |
| CHECK_EQ(test.dOle, 1U); |
| CHECK_EQ(test.dUle, 1U); |
| CHECK_EQ(test.fF, 0U); |
| CHECK_EQ(test.fUn, 0U); |
| CHECK_EQ(test.fEq, 1U); |
| CHECK_EQ(test.fUeq, 1U); |
| CHECK_EQ(test.fOlt, 0U); |
| CHECK_EQ(test.fUlt, 0U); |
| CHECK_EQ(test.fOle, 1U); |
| CHECK_EQ(test.fUle, 1U); |
| |
| test.dOp1 = std::numeric_limits<double>::quiet_NaN(); |
| test.dOp2 = 0.0; |
| test.fOp1 = std::numeric_limits<float>::quiet_NaN(); |
| test.fOp2 = 0.0; |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.dF, 0U); |
| CHECK_EQ(test.dUn, 1U); |
| CHECK_EQ(test.dEq, 0U); |
| CHECK_EQ(test.dUeq, 1U); |
| CHECK_EQ(test.dOlt, 0U); |
| CHECK_EQ(test.dUlt, 1U); |
| CHECK_EQ(test.dOle, 0U); |
| CHECK_EQ(test.dUle, 1U); |
| CHECK_EQ(test.fF, 0U); |
| CHECK_EQ(test.fUn, 1U); |
| CHECK_EQ(test.fEq, 0U); |
| CHECK_EQ(test.fUeq, 1U); |
| CHECK_EQ(test.fOlt, 0U); |
| CHECK_EQ(test.fUlt, 1U); |
| CHECK_EQ(test.fOle, 0U); |
| CHECK_EQ(test.fUle, 1U); |
| } |
| } |
| |
| |
| TEST(CMP_COND_FMT) { |
| if (IsMipsArchVariant(kMips32r6)) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| struct TestFloat { |
| double dOp1; |
| double dOp2; |
| double dF; |
| double dUn; |
| double dEq; |
| double dUeq; |
| double dOlt; |
| double dUlt; |
| double dOle; |
| double dUle; |
| double dOr; |
| double dUne; |
| double dNe; |
| float fOp1; |
| float fOp2; |
| float fF; |
| float fUn; |
| float fEq; |
| float fUeq; |
| float fOlt; |
| float fUlt; |
| float fOle; |
| float fUle; |
| float fOr; |
| float fUne; |
| float fNe; |
| }; |
| |
| TestFloat test; |
| |
| __ li(t1, 1); |
| |
| __ Ldc1(f4, MemOperand(a0, offsetof(TestFloat, dOp1))); |
| __ Ldc1(f6, MemOperand(a0, offsetof(TestFloat, dOp2))); |
| |
| __ lwc1(f14, MemOperand(a0, offsetof(TestFloat, fOp1))); |
| __ lwc1(f16, MemOperand(a0, offsetof(TestFloat, fOp2))); |
| |
| __ cmp_d(F, f2, f4, f6); |
| __ cmp_s(F, f12, f14, f16); |
| __ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dF))); |
| __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fF)) ); |
| |
| __ cmp_d(UN, f2, f4, f6); |
| __ cmp_s(UN, f12, f14, f16); |
| __ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dUn))); |
| __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fUn)) ); |
| |
| __ cmp_d(EQ, f2, f4, f6); |
| __ cmp_s(EQ, f12, f14, f16); |
| __ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dEq))); |
| __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fEq)) ); |
| |
| __ cmp_d(UEQ, f2, f4, f6); |
| __ cmp_s(UEQ, f12, f14, f16); |
| __ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dUeq))); |
| __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fUeq)) ); |
| |
| __ cmp_d(LT, f2, f4, f6); |
| __ cmp_s(LT, f12, f14, f16); |
| __ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dOlt))); |
| __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fOlt)) ); |
| |
| __ cmp_d(ULT, f2, f4, f6); |
| __ cmp_s(ULT, f12, f14, f16); |
| __ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dUlt))); |
| __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fUlt)) ); |
| |
| __ cmp_d(LE, f2, f4, f6); |
| __ cmp_s(LE, f12, f14, f16); |
| __ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dOle))); |
| __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fOle)) ); |
| |
| __ cmp_d(ULE, f2, f4, f6); |
| __ cmp_s(ULE, f12, f14, f16); |
| __ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dUle))); |
| __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fUle)) ); |
| |
| __ cmp_d(ORD, f2, f4, f6); |
| __ cmp_s(ORD, f12, f14, f16); |
| __ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dOr))); |
| __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fOr)) ); |
| |
| __ cmp_d(UNE, f2, f4, f6); |
| __ cmp_s(UNE, f12, f14, f16); |
| __ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dUne))); |
| __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fUne)) ); |
| |
| __ cmp_d(NE, f2, f4, f6); |
| __ cmp_s(NE, f12, f14, f16); |
| __ Sdc1(f2, MemOperand(a0, offsetof(TestFloat, dNe))); |
| __ swc1(f12, MemOperand(a0, offsetof(TestFloat, fNe)) ); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| uint64_t dTrue = 0xFFFFFFFFFFFFFFFF; |
| uint64_t dFalse = 0x0000000000000000; |
| uint32_t fTrue = 0xFFFFFFFF; |
| uint32_t fFalse = 0x00000000; |
| |
| test.dOp1 = 2.0; |
| test.dOp2 = 3.0; |
| test.fOp1 = 2.0; |
| test.fOp2 = 3.0; |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(bit_cast<uint64_t>(test.dF), dFalse); |
| CHECK_EQ(bit_cast<uint64_t>(test.dUn), dFalse); |
| CHECK_EQ(bit_cast<uint64_t>(test.dEq), dFalse); |
| CHECK_EQ(bit_cast<uint64_t>(test.dUeq), dFalse); |
| CHECK_EQ(bit_cast<uint64_t>(test.dOlt), dTrue); |
| CHECK_EQ(bit_cast<uint64_t>(test.dUlt), dTrue); |
| CHECK_EQ(bit_cast<uint64_t>(test.dOle), dTrue); |
| CHECK_EQ(bit_cast<uint64_t>(test.dUle), dTrue); |
| CHECK_EQ(bit_cast<uint64_t>(test.dOr), dTrue); |
| CHECK_EQ(bit_cast<uint64_t>(test.dUne), dTrue); |
| CHECK_EQ(bit_cast<uint64_t>(test.dNe), dTrue); |
| CHECK_EQ(bit_cast<uint32_t>(test.fF), fFalse); |
| CHECK_EQ(bit_cast<uint32_t>(test.fUn), fFalse); |
| CHECK_EQ(bit_cast<uint32_t>(test.fEq), fFalse); |
| CHECK_EQ(bit_cast<uint32_t>(test.fUeq), fFalse); |
| CHECK_EQ(bit_cast<uint32_t>(test.fOlt), fTrue); |
| CHECK_EQ(bit_cast<uint32_t>(test.fUlt), fTrue); |
| CHECK_EQ(bit_cast<uint32_t>(test.fOle), fTrue); |
| CHECK_EQ(bit_cast<uint32_t>(test.fUle), fTrue); |
| |
| test.dOp1 = std::numeric_limits<double>::max(); |
| test.dOp2 = std::numeric_limits<double>::min(); |
| test.fOp1 = std::numeric_limits<float>::min(); |
| test.fOp2 = -std::numeric_limits<float>::max(); // lowest() |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(bit_cast<uint64_t>(test.dF), dFalse); |
| CHECK_EQ(bit_cast<uint64_t>(test.dUn), dFalse); |
| CHECK_EQ(bit_cast<uint64_t>(test.dEq), dFalse); |
| CHECK_EQ(bit_cast<uint64_t>(test.dUeq), dFalse); |
| CHECK_EQ(bit_cast<uint64_t>(test.dOlt), dFalse); |
| CHECK_EQ(bit_cast<uint64_t>(test.dUlt), dFalse); |
| CHECK_EQ(bit_cast<uint64_t>(test.dOle), dFalse); |
| CHECK_EQ(bit_cast<uint64_t>(test.dUle), dFalse); |
| CHECK_EQ(bit_cast<uint64_t>(test.dOr), dTrue); |
| CHECK_EQ(bit_cast<uint64_t>(test.dUne), dTrue); |
| CHECK_EQ(bit_cast<uint64_t>(test.dNe), dTrue); |
| CHECK_EQ(bit_cast<uint32_t>(test.fF), fFalse); |
| CHECK_EQ(bit_cast<uint32_t>(test.fUn), fFalse); |
| CHECK_EQ(bit_cast<uint32_t>(test.fEq), fFalse); |
| CHECK_EQ(bit_cast<uint32_t>(test.fUeq), fFalse); |
| CHECK_EQ(bit_cast<uint32_t>(test.fOlt), fFalse); |
| CHECK_EQ(bit_cast<uint32_t>(test.fUlt), fFalse); |
| CHECK_EQ(bit_cast<uint32_t>(test.fOle), fFalse); |
| CHECK_EQ(bit_cast<uint32_t>(test.fUle), fFalse); |
| |
| test.dOp1 = -std::numeric_limits<double>::max(); // lowest() |
| test.dOp2 = -std::numeric_limits<double>::max(); // lowest() |
| test.fOp1 = std::numeric_limits<float>::max(); |
| test.fOp2 = std::numeric_limits<float>::max(); |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(bit_cast<uint64_t>(test.dF), dFalse); |
| CHECK_EQ(bit_cast<uint64_t>(test.dUn), dFalse); |
| CHECK_EQ(bit_cast<uint64_t>(test.dEq), dTrue); |
| CHECK_EQ(bit_cast<uint64_t>(test.dUeq), dTrue); |
| CHECK_EQ(bit_cast<uint64_t>(test.dOlt), dFalse); |
| CHECK_EQ(bit_cast<uint64_t>(test.dUlt), dFalse); |
| CHECK_EQ(bit_cast<uint64_t>(test.dOle), dTrue); |
| CHECK_EQ(bit_cast<uint64_t>(test.dUle), dTrue); |
| CHECK_EQ(bit_cast<uint64_t>(test.dOr), dTrue); |
| CHECK_EQ(bit_cast<uint64_t>(test.dUne), dFalse); |
| CHECK_EQ(bit_cast<uint64_t>(test.dNe), dFalse); |
| CHECK_EQ(bit_cast<uint32_t>(test.fF), fFalse); |
| CHECK_EQ(bit_cast<uint32_t>(test.fUn), fFalse); |
| CHECK_EQ(bit_cast<uint32_t>(test.fEq), fTrue); |
| CHECK_EQ(bit_cast<uint32_t>(test.fUeq), fTrue); |
| CHECK_EQ(bit_cast<uint32_t>(test.fOlt), fFalse); |
| CHECK_EQ(bit_cast<uint32_t>(test.fUlt), fFalse); |
| CHECK_EQ(bit_cast<uint32_t>(test.fOle), fTrue); |
| CHECK_EQ(bit_cast<uint32_t>(test.fUle), fTrue); |
| |
| test.dOp1 = std::numeric_limits<double>::quiet_NaN(); |
| test.dOp2 = 0.0; |
| test.fOp1 = std::numeric_limits<float>::quiet_NaN(); |
| test.fOp2 = 0.0; |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(bit_cast<uint64_t>(test.dF), dFalse); |
| CHECK_EQ(bit_cast<uint64_t>(test.dUn), dTrue); |
| CHECK_EQ(bit_cast<uint64_t>(test.dEq), dFalse); |
| CHECK_EQ(bit_cast<uint64_t>(test.dUeq), dTrue); |
| CHECK_EQ(bit_cast<uint64_t>(test.dOlt), dFalse); |
| CHECK_EQ(bit_cast<uint64_t>(test.dUlt), dTrue); |
| CHECK_EQ(bit_cast<uint64_t>(test.dOle), dFalse); |
| CHECK_EQ(bit_cast<uint64_t>(test.dUle), dTrue); |
| CHECK_EQ(bit_cast<uint64_t>(test.dOr), dFalse); |
| CHECK_EQ(bit_cast<uint64_t>(test.dUne), dTrue); |
| CHECK_EQ(bit_cast<uint64_t>(test.dNe), dFalse); |
| CHECK_EQ(bit_cast<uint32_t>(test.fF), fFalse); |
| CHECK_EQ(bit_cast<uint32_t>(test.fUn), fTrue); |
| CHECK_EQ(bit_cast<uint32_t>(test.fEq), fFalse); |
| CHECK_EQ(bit_cast<uint32_t>(test.fUeq), fTrue); |
| CHECK_EQ(bit_cast<uint32_t>(test.fOlt), fFalse); |
| CHECK_EQ(bit_cast<uint32_t>(test.fUlt), fTrue); |
| CHECK_EQ(bit_cast<uint32_t>(test.fOle), fFalse); |
| CHECK_EQ(bit_cast<uint32_t>(test.fUle), fTrue); |
| } |
| } |
| |
| |
| TEST(CVT) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| struct TestFloat { |
| float cvt_d_s_in; |
| double cvt_d_s_out; |
| int32_t cvt_d_w_in; |
| double cvt_d_w_out; |
| int64_t cvt_d_l_in; |
| double cvt_d_l_out; |
| |
| float cvt_l_s_in; |
| int64_t cvt_l_s_out; |
| double cvt_l_d_in; |
| int64_t cvt_l_d_out; |
| |
| double cvt_s_d_in; |
| float cvt_s_d_out; |
| int32_t cvt_s_w_in; |
| float cvt_s_w_out; |
| int64_t cvt_s_l_in; |
| float cvt_s_l_out; |
| |
| float cvt_w_s_in; |
| int32_t cvt_w_s_out; |
| double cvt_w_d_in; |
| int32_t cvt_w_d_out; |
| }; |
| |
| TestFloat test; |
| |
| // Save FCSR. |
| __ cfc1(a1, FCSR); |
| // Disable FPU exceptions. |
| __ ctc1(zero_reg, FCSR); |
| |
| #define GENERATE_CVT_TEST(x, y, z) \ |
| __ y##c1(f0, MemOperand(a0, offsetof(TestFloat, x##_in))); \ |
| __ x(f0, f0); \ |
| __ nop(); \ |
| __ z##c1(f0, MemOperand(a0, offsetof(TestFloat, x##_out))); |
| |
| GENERATE_CVT_TEST(cvt_d_s, lw, Sd) |
| GENERATE_CVT_TEST(cvt_d_w, lw, Sd) |
| if ((IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) && |
| IsFp64Mode()) { |
| GENERATE_CVT_TEST(cvt_d_l, Ld, Sd) |
| } |
| |
| if (IsFp64Mode()) { |
| GENERATE_CVT_TEST(cvt_l_s, lw, Sd) |
| GENERATE_CVT_TEST(cvt_l_d, Ld, Sd) |
| } |
| |
| GENERATE_CVT_TEST(cvt_s_d, Ld, sw) |
| GENERATE_CVT_TEST(cvt_s_w, lw, sw) |
| if ((IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) && |
| IsFp64Mode()) { |
| GENERATE_CVT_TEST(cvt_s_l, Ld, sw) |
| } |
| |
| GENERATE_CVT_TEST(cvt_w_s, lw, sw) |
| GENERATE_CVT_TEST(cvt_w_d, Ld, sw) |
| |
| // Restore FCSR. |
| __ ctc1(a1, FCSR); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| |
| test.cvt_d_s_in = -0.51; |
| test.cvt_d_w_in = -1; |
| test.cvt_d_l_in = -1; |
| test.cvt_l_s_in = -0.51; |
| test.cvt_l_d_in = -0.51; |
| test.cvt_s_d_in = -0.51; |
| test.cvt_s_w_in = -1; |
| test.cvt_s_l_in = -1; |
| test.cvt_w_s_in = -0.51; |
| test.cvt_w_d_in = -0.51; |
| |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.cvt_d_s_out, static_cast<double>(test.cvt_d_s_in)); |
| CHECK_EQ(test.cvt_d_w_out, static_cast<double>(test.cvt_d_w_in)); |
| if ((IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) && |
| IsFp64Mode()) { |
| CHECK_EQ(test.cvt_d_l_out, static_cast<double>(test.cvt_d_l_in)); |
| } |
| if (IsFp64Mode()) { |
| CHECK_EQ(-1, test.cvt_l_s_out); |
| CHECK_EQ(-1, test.cvt_l_d_out); |
| } |
| CHECK_EQ(test.cvt_s_d_out, static_cast<float>(test.cvt_s_d_in)); |
| CHECK_EQ(test.cvt_s_w_out, static_cast<float>(test.cvt_s_w_in)); |
| if ((IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) && |
| IsFp64Mode()) { |
| CHECK_EQ(test.cvt_s_l_out, static_cast<float>(test.cvt_s_l_in)); |
| } |
| CHECK_EQ(-1, test.cvt_w_s_out); |
| CHECK_EQ(-1, test.cvt_w_d_out); |
| |
| test.cvt_d_s_in = 0.49; |
| test.cvt_d_w_in = 1; |
| test.cvt_d_l_in = 1; |
| test.cvt_l_s_in = 0.49; |
| test.cvt_l_d_in = 0.49; |
| test.cvt_s_d_in = 0.49; |
| test.cvt_s_w_in = 1; |
| test.cvt_s_l_in = 1; |
| test.cvt_w_s_in = 0.49; |
| test.cvt_w_d_in = 0.49; |
| |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.cvt_d_s_out, static_cast<double>(test.cvt_d_s_in)); |
| CHECK_EQ(test.cvt_d_w_out, static_cast<double>(test.cvt_d_w_in)); |
| if ((IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) && |
| IsFp64Mode()) { |
| CHECK_EQ(test.cvt_d_l_out, static_cast<double>(test.cvt_d_l_in)); |
| } |
| if (IsFp64Mode()) { |
| CHECK_EQ(0, test.cvt_l_s_out); |
| CHECK_EQ(0, test.cvt_l_d_out); |
| } |
| CHECK_EQ(test.cvt_s_d_out, static_cast<float>(test.cvt_s_d_in)); |
| CHECK_EQ(test.cvt_s_w_out, static_cast<float>(test.cvt_s_w_in)); |
| if ((IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) && |
| IsFp64Mode()) { |
| CHECK_EQ(test.cvt_s_l_out, static_cast<float>(test.cvt_s_l_in)); |
| } |
| CHECK_EQ(0, test.cvt_w_s_out); |
| CHECK_EQ(0, test.cvt_w_d_out); |
| |
| test.cvt_d_s_in = std::numeric_limits<float>::max(); |
| test.cvt_d_w_in = std::numeric_limits<int32_t>::max(); |
| test.cvt_d_l_in = std::numeric_limits<int64_t>::max(); |
| test.cvt_l_s_in = std::numeric_limits<float>::max(); |
| test.cvt_l_d_in = std::numeric_limits<double>::max(); |
| test.cvt_s_d_in = std::numeric_limits<double>::max(); |
| test.cvt_s_w_in = std::numeric_limits<int32_t>::max(); |
| test.cvt_s_l_in = std::numeric_limits<int64_t>::max(); |
| test.cvt_w_s_in = std::numeric_limits<float>::max(); |
| test.cvt_w_d_in = std::numeric_limits<double>::max(); |
| |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.cvt_d_s_out, static_cast<double>(test.cvt_d_s_in)); |
| CHECK_EQ(test.cvt_d_w_out, static_cast<double>(test.cvt_d_w_in)); |
| if ((IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) && |
| IsFp64Mode()) { |
| CHECK_EQ(test.cvt_d_l_out, static_cast<double>(test.cvt_d_l_in)); |
| } |
| if (IsFp64Mode()) { |
| CHECK_EQ(test.cvt_l_s_out, std::numeric_limits<int64_t>::max()); |
| CHECK_EQ(test.cvt_l_d_out, std::numeric_limits<int64_t>::max()); |
| } |
| CHECK_EQ(test.cvt_s_d_out, static_cast<float>(test.cvt_s_d_in)); |
| CHECK_EQ(test.cvt_s_w_out, static_cast<float>(test.cvt_s_w_in)); |
| if ((IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) && |
| IsFp64Mode()) { |
| CHECK_EQ(test.cvt_s_l_out, static_cast<float>(test.cvt_s_l_in)); |
| } |
| CHECK_EQ(test.cvt_w_s_out, std::numeric_limits<int32_t>::max()); |
| CHECK_EQ(test.cvt_w_d_out, std::numeric_limits<int32_t>::max()); |
| |
| |
| test.cvt_d_s_in = -std::numeric_limits<float>::max(); // lowest() |
| test.cvt_d_w_in = std::numeric_limits<int32_t>::min(); // lowest() |
| test.cvt_d_l_in = std::numeric_limits<int64_t>::min(); // lowest() |
| test.cvt_l_s_in = -std::numeric_limits<float>::max(); // lowest() |
| test.cvt_l_d_in = -std::numeric_limits<double>::max(); // lowest() |
| test.cvt_s_d_in = -std::numeric_limits<double>::max(); // lowest() |
| test.cvt_s_w_in = std::numeric_limits<int32_t>::min(); // lowest() |
| test.cvt_s_l_in = std::numeric_limits<int64_t>::min(); // lowest() |
| test.cvt_w_s_in = -std::numeric_limits<float>::max(); // lowest() |
| test.cvt_w_d_in = -std::numeric_limits<double>::max(); // lowest() |
| |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.cvt_d_s_out, static_cast<double>(test.cvt_d_s_in)); |
| CHECK_EQ(test.cvt_d_w_out, static_cast<double>(test.cvt_d_w_in)); |
| if ((IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) && |
| IsFp64Mode()) { |
| CHECK_EQ(test.cvt_d_l_out, static_cast<double>(test.cvt_d_l_in)); |
| } |
| // The returned value when converting from fixed-point to float-point |
| // is not consistent between board, simulator and specification |
| // in this test case, therefore modifying the test |
| if (IsFp64Mode()) { |
| CHECK(test.cvt_l_s_out == std::numeric_limits<int64_t>::min() || |
| test.cvt_l_s_out == std::numeric_limits<int64_t>::max()); |
| CHECK(test.cvt_l_d_out == std::numeric_limits<int64_t>::min() || |
| test.cvt_l_d_out == std::numeric_limits<int64_t>::max()); |
| } |
| CHECK_EQ(test.cvt_s_d_out, static_cast<float>(test.cvt_s_d_in)); |
| CHECK_EQ(test.cvt_s_w_out, static_cast<float>(test.cvt_s_w_in)); |
| if ((IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) && |
| IsFp64Mode()) { |
| CHECK_EQ(test.cvt_s_l_out, static_cast<float>(test.cvt_s_l_in)); |
| } |
| CHECK(test.cvt_w_s_out == std::numeric_limits<int32_t>::min() || |
| test.cvt_w_s_out == std::numeric_limits<int32_t>::max()); |
| CHECK(test.cvt_w_d_out == std::numeric_limits<int32_t>::min() || |
| test.cvt_w_d_out == std::numeric_limits<int32_t>::max()); |
| |
| |
| test.cvt_d_s_in = std::numeric_limits<float>::min(); |
| test.cvt_d_w_in = std::numeric_limits<int32_t>::min(); |
| test.cvt_d_l_in = std::numeric_limits<int64_t>::min(); |
| test.cvt_l_s_in = std::numeric_limits<float>::min(); |
| test.cvt_l_d_in = std::numeric_limits<double>::min(); |
| test.cvt_s_d_in = std::numeric_limits<double>::min(); |
| test.cvt_s_w_in = std::numeric_limits<int32_t>::min(); |
| test.cvt_s_l_in = std::numeric_limits<int64_t>::min(); |
| test.cvt_w_s_in = std::numeric_limits<float>::min(); |
| test.cvt_w_d_in = std::numeric_limits<double>::min(); |
| |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.cvt_d_s_out, static_cast<double>(test.cvt_d_s_in)); |
| CHECK_EQ(test.cvt_d_w_out, static_cast<double>(test.cvt_d_w_in)); |
| if ((IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) && |
| IsFp64Mode()) { |
| CHECK_EQ(test.cvt_d_l_out, static_cast<double>(test.cvt_d_l_in)); |
| } |
| if (IsFp64Mode()) { |
| CHECK_EQ(0, test.cvt_l_s_out); |
| CHECK_EQ(0, test.cvt_l_d_out); |
| } |
| CHECK_EQ(test.cvt_s_d_out, static_cast<float>(test.cvt_s_d_in)); |
| CHECK_EQ(test.cvt_s_w_out, static_cast<float>(test.cvt_s_w_in)); |
| if ((IsMipsArchVariant(kMips32r2) || IsMipsArchVariant(kMips32r6)) && |
| IsFp64Mode()) { |
| CHECK_EQ(test.cvt_s_l_out, static_cast<float>(test.cvt_s_l_in)); |
| } |
| CHECK_EQ(0, test.cvt_w_s_out); |
| CHECK_EQ(0, test.cvt_w_d_out); |
| } |
| |
| |
| TEST(DIV_FMT) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| struct Test { |
| double dOp1; |
| double dOp2; |
| double dRes; |
| float fOp1; |
| float fOp2; |
| float fRes; |
| }; |
| |
| Test test; |
| |
| // Save FCSR. |
| __ cfc1(a1, FCSR); |
| // Disable FPU exceptions. |
| __ ctc1(zero_reg, FCSR); |
| |
| __ Ldc1(f4, MemOperand(a0, offsetof(Test, dOp1))); |
| __ Ldc1(f2, MemOperand(a0, offsetof(Test, dOp2))); |
| __ nop(); |
| __ div_d(f6, f4, f2); |
| __ Sdc1(f6, MemOperand(a0, offsetof(Test, dRes))); |
| |
| __ lwc1(f4, MemOperand(a0, offsetof(Test, fOp1)) ); |
| __ lwc1(f2, MemOperand(a0, offsetof(Test, fOp2)) ); |
| __ nop(); |
| __ div_s(f6, f4, f2); |
| __ swc1(f6, MemOperand(a0, offsetof(Test, fRes)) ); |
| |
| // Restore FCSR. |
| __ ctc1(a1, FCSR); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| |
| (f.Call(&test, 0, 0, 0, 0)); |
| |
| const int test_size = 3; |
| |
| double dOp1[test_size] = { |
| 5.0, |
| DBL_MAX, |
| DBL_MAX, |
| }; |
| double dOp2[test_size] = { |
| 2.0, |
| 2.0, |
| -DBL_MAX, |
| }; |
| double dRes[test_size] = { |
| 2.5, |
| DBL_MAX / 2.0, |
| -1.0, |
| }; |
| float fOp1[test_size] = { |
| 5.0, |
| FLT_MAX, |
| FLT_MAX, |
| }; |
| float fOp2[test_size] = { |
| 2.0, |
| 2.0, |
| -FLT_MAX, |
| }; |
| float fRes[test_size] = { |
| 2.5, |
| FLT_MAX / 2.0, |
| -1.0, |
| }; |
| |
| for (int i = 0; i < test_size; i++) { |
| test.dOp1 = dOp1[i]; |
| test.dOp2 = dOp2[i]; |
| test.fOp1 = fOp1[i]; |
| test.fOp2 = fOp2[i]; |
| |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK_EQ(test.dRes, dRes[i]); |
| CHECK_EQ(test.fRes, fRes[i]); |
| } |
| |
| test.dOp1 = DBL_MAX; |
| test.dOp2 = -0.0; |
| test.fOp1 = FLT_MAX; |
| test.fOp2 = -0.0; |
| |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK(!std::isfinite(test.dRes)); |
| CHECK(!std::isfinite(test.fRes)); |
| |
| test.dOp1 = 0.0; |
| test.dOp2 = -0.0; |
| test.fOp1 = 0.0; |
| test.fOp2 = -0.0; |
| |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK(std::isnan(test.dRes)); |
| CHECK(std::isnan(test.fRes)); |
| |
| test.dOp1 = std::numeric_limits<double>::quiet_NaN(); |
| test.dOp2 = -5.0; |
| test.fOp1 = std::numeric_limits<float>::quiet_NaN(); |
| test.fOp2 = -5.0; |
| |
| (f.Call(&test, 0, 0, 0, 0)); |
| CHECK(std::isnan(test.dRes)); |
| CHECK(std::isnan(test.fRes)); |
| } |
| |
| |
| uint32_t run_align(uint32_t rs_value, uint32_t rt_value, uint8_t bp) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| __ align(v0, a0, a1, bp); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| |
| auto f = GeneratedCode<F2>::FromCode(*code); |
| |
| uint32_t res = |
| reinterpret_cast<uint32_t>(f.Call(rs_value, rt_value, 0, 0, 0)); |
| |
| return res; |
| } |
| |
| |
| TEST(r6_align) { |
| if (IsMipsArchVariant(kMips32r6)) { |
| CcTest::InitializeVM(); |
| |
| struct TestCaseAlign { |
| uint32_t rs_value; |
| uint32_t rt_value; |
| uint8_t bp; |
| uint32_t expected_res; |
| }; |
| |
| // clang-format off |
| struct TestCaseAlign tc[] = { |
| // rs_value, rt_value, bp, expected_res |
| {0x11223344, 0xAABBCCDD, 0, 0xAABBCCDD}, |
| {0x11223344, 0xAABBCCDD, 1, 0xBBCCDD11}, |
| {0x11223344, 0xAABBCCDD, 2, 0xCCDD1122}, |
| {0x11223344, 0xAABBCCDD, 3, 0xDD112233}, |
| }; |
| // clang-format on |
| |
| size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseAlign); |
| for (size_t i = 0; i < nr_test_cases; ++i) { |
| CHECK_EQ(tc[i].expected_res, run_align(tc[i].rs_value, |
| tc[i].rt_value, tc[i].bp)); |
| } |
| } |
| } |
| |
| uint32_t PC; // The program counter. |
| |
| uint32_t run_aluipc(int16_t offset) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| __ aluipc(v0, offset); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| |
| auto f = GeneratedCode<F2>::FromCode(*code); |
| PC = (uint32_t)code->entry(); // Set the program counter. |
| |
| uint32_t res = reinterpret_cast<uint32_t>(f.Call(0, 0, 0, 0, 0)); |
| |
| return res; |
| } |
| |
| |
| TEST(r6_aluipc) { |
| if (IsMipsArchVariant(kMips32r6)) { |
| CcTest::InitializeVM(); |
| |
| struct TestCaseAluipc { |
| int16_t offset; |
| }; |
| |
| struct TestCaseAluipc tc[] = { |
| // offset |
| { -32768 }, // 0x8000 |
| { -1 }, // 0xFFFF |
| { 0 }, |
| { 1 }, |
| { 32767 }, // 0x7FFF |
| }; |
| |
| size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseAluipc); |
| for (size_t i = 0; i < nr_test_cases; ++i) { |
| PC = 0; |
| uint32_t res = run_aluipc(tc[i].offset); |
| // Now, the program_counter (PC) is set. |
| uint32_t expected_res = ~0x0FFFF & (PC + (tc[i].offset << 16)); |
| CHECK_EQ(expected_res, res); |
| } |
| } |
| } |
| |
| |
| uint32_t run_auipc(int16_t offset) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| __ auipc(v0, offset); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| |
| auto f = GeneratedCode<F2>::FromCode(*code); |
| PC = (uint32_t)code->entry(); // Set the program counter. |
| |
| uint32_t res = reinterpret_cast<uint32_t>(f.Call(0, 0, 0, 0, 0)); |
| |
| return res; |
| } |
| |
| |
| TEST(r6_auipc) { |
| if (IsMipsArchVariant(kMips32r6)) { |
| CcTest::InitializeVM(); |
| |
| struct TestCaseAuipc { |
| int16_t offset; |
| }; |
| |
| struct TestCaseAuipc tc[] = { |
| // offset |
| { -32768 }, // 0x8000 |
| { -1 }, // 0xFFFF |
| { 0 }, |
| { 1 }, |
| { 32767 }, // 0x7FFF |
| }; |
| |
| size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseAuipc); |
| for (size_t i = 0; i < nr_test_cases; ++i) { |
| PC = 0; |
| uint32_t res = run_auipc(tc[i].offset); |
| // Now, the program_counter (PC) is set. |
| uint32_t expected_res = PC + (tc[i].offset << 16); |
| CHECK_EQ(expected_res, res); |
| } |
| } |
| } |
| |
| |
| uint32_t run_lwpc(int offset) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| // 256k instructions; 2^8k |
| // addiu t7, t0, 0xFFFF; (0x250FFFFF) |
| // ... |
| // addiu t4, t0, 0x0000; (0x250C0000) |
| uint32_t addiu_start_1 = 0x25000000; |
| for (int32_t i = 0xFFFFF; i >= 0xC0000; --i) { |
| uint32_t addiu_new = addiu_start_1 + i; |
| __ dd(addiu_new); |
| } |
| |
| __ lwpc(t8, offset); // offset 0; 0xEF080000 (t8 register) |
| __ mov(v0, t8); |
| |
| // 256k instructions; 2^8k |
| // addiu t0, t0, 0x0000; (0x25080000) |
| // ... |
| // addiu t3, t0, 0xFFFF; (0x250BFFFF) |
| uint32_t addiu_start_2 = 0x25000000; |
| for (int32_t i = 0x80000; i <= 0xBFFFF; ++i) { |
| uint32_t addiu_new = addiu_start_2 + i; |
| __ dd(addiu_new); |
| } |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| |
| auto f = GeneratedCode<F2>::FromCode(*code); |
| |
| uint32_t res = reinterpret_cast<uint32_t>(f.Call(0, 0, 0, 0, 0)); |
| |
| return res; |
| } |
| |
| |
| TEST(r6_lwpc) { |
| if (IsMipsArchVariant(kMips32r6)) { |
| CcTest::InitializeVM(); |
| |
| struct TestCaseLwpc { |
| int offset; |
| uint32_t expected_res; |
| }; |
| |
| // clang-format off |
| struct TestCaseLwpc tc[] = { |
| // offset, expected_res |
| { -262144, 0x250FFFFF }, // offset 0x40000 |
| { -4, 0x250C0003 }, |
| { -1, 0x250C0000 }, |
| { 0, 0xEF080000 }, |
| { 1, 0x03001025 }, // mov(v0, t8) |
| { 2, 0x25080000 }, |
| { 4, 0x25080002 }, |
| { 262143, 0x250BFFFD }, // offset 0x3FFFF |
| }; |
| // clang-format on |
| |
| size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseLwpc); |
| for (size_t i = 0; i < nr_test_cases; ++i) { |
| uint32_t res = run_lwpc(tc[i].offset); |
| CHECK_EQ(tc[i].expected_res, res); |
| } |
| } |
| } |
| |
| |
| uint32_t run_jic(int16_t offset) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| Label stop_execution; |
| __ push(ra); |
| __ li(v0, 0); |
| __ li(t1, 0x66); |
| |
| __ addiu(v0, v0, 0x1); // <-- offset = -32 |
| __ addiu(v0, v0, 0x2); |
| __ addiu(v0, v0, 0x10); |
| __ addiu(v0, v0, 0x20); |
| __ beq(v0, t1, &stop_execution); |
| __ nop(); |
| |
| __ nal(); // t0 <- program counter |
| __ mov(t0, ra); |
| __ jic(t0, offset); |
| |
| __ addiu(v0, v0, 0x100); |
| __ addiu(v0, v0, 0x200); |
| __ addiu(v0, v0, 0x1000); |
| __ addiu(v0, v0, 0x2000); // <--- offset = 16 |
| __ pop(ra); |
| __ jr(ra); |
| __ nop(); |
| |
| __ bind(&stop_execution); |
| __ pop(ra); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| |
| auto f = GeneratedCode<F2>::FromCode(*code); |
| |
| uint32_t res = reinterpret_cast<uint32_t>(f.Call(0, 0, 0, 0, 0)); |
| |
| return res; |
| } |
| |
| |
| TEST(r6_jic) { |
| if (IsMipsArchVariant(kMips32r6)) { |
| CcTest::InitializeVM(); |
| |
| struct TestCaseJic { |
| // As rt will be used t0 register which will have value of |
| // the program counter for the jic instruction. |
| int16_t offset; |
| uint32_t expected_res; |
| }; |
| |
| struct TestCaseJic tc[] = { |
| // offset, expected_result |
| { 16, 0x2033 }, |
| { 4, 0x3333 }, |
| { -32, 0x66 }, |
| }; |
| |
| size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseJic); |
| for (size_t i = 0; i < nr_test_cases; ++i) { |
| uint32_t res = run_jic(tc[i].offset); |
| CHECK_EQ(tc[i].expected_res, res); |
| } |
| } |
| } |
| |
| |
| uint64_t run_beqzc(int32_t value, int32_t offset) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| Label stop_execution; |
| __ li(v0, 0); |
| __ li(t1, 0x66); |
| |
| __ addiu(v0, v0, 0x1); // <-- offset = -32 |
| __ addiu(v0, v0, 0x2); |
| __ addiu(v0, v0, 0x10); |
| __ addiu(v0, v0, 0x20); |
| __ beq(v0, t1, &stop_execution); |
| __ nop(); |
| |
| __ beqzc(a0, offset); // BEQZC rs, offset |
| |
| __ addiu(v0, v0, 0x1); |
| __ addiu(v0, v0, 0x100); |
| __ addiu(v0, v0, 0x200); |
| __ addiu(v0, v0, 0x1000); |
| __ addiu(v0, v0, 0x2000); // <--- offset = 16 |
| __ jr(ra); |
| __ nop(); |
| |
| __ bind(&stop_execution); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| |
| auto f = GeneratedCode<F2>::FromCode(*code); |
| |
| uint32_t res = reinterpret_cast<uint32_t>(f.Call(value, 0, 0, 0, 0)); |
| |
| return res; |
| } |
| |
| |
| TEST(r6_beqzc) { |
| if (IsMipsArchVariant(kMips32r6)) { |
| CcTest::InitializeVM(); |
| |
| struct TestCaseBeqzc { |
| uint32_t value; |
| int32_t offset; |
| uint32_t expected_res; |
| }; |
| |
| // clang-format off |
| struct TestCaseBeqzc tc[] = { |
| // value, offset, expected_res |
| { 0x0, -8, 0x66 }, |
| { 0x0, 0, 0x3334 }, |
| { 0x0, 1, 0x3333 }, |
| { 0xABC, 1, 0x3334 }, |
| { 0x0, 4, 0x2033 }, |
| }; |
| // clang-format on |
| |
| size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseBeqzc); |
| for (size_t i = 0; i < nr_test_cases; ++i) { |
| uint32_t res = run_beqzc(tc[i].value, tc[i].offset); |
| CHECK_EQ(tc[i].expected_res, res); |
| } |
| } |
| } |
| |
| void load_elements_of_vector( |
| MacroAssembler& assm, // NOLINT(runtime/references) |
| const uint64_t elements[], MSARegister w, Register t0, Register t1) { |
| __ li(t0, static_cast<uint32_t>(elements[0] & 0xFFFFFFFF)); |
| __ li(t1, static_cast<uint32_t>((elements[0] >> 32) & 0xFFFFFFFF)); |
| __ insert_w(w, 0, t0); |
| __ insert_w(w, 1, t1); |
| __ li(t0, static_cast<uint32_t>(elements[1] & 0xFFFFFFFF)); |
| __ li(t1, static_cast<uint32_t>((elements[1] >> 32) & 0xFFFFFFFF)); |
| __ insert_w(w, 2, t0); |
| __ insert_w(w, 3, t1); |
| } |
| |
| inline void store_elements_of_vector( |
| MacroAssembler& assm, // NOLINT(runtime/references) |
| MSARegister w, Register a) { |
| __ st_d(w, MemOperand(a, 0)); |
| } |
| |
| union msa_reg_t { |
| uint8_t b[16]; |
| uint16_t h[8]; |
| uint32_t w[4]; |
| uint64_t d[2]; |
| }; |
| |
| struct TestCaseMsaBranch { |
| uint64_t wt_lo; |
| uint64_t wt_hi; |
| }; |
| |
| template <typename Branch> |
| void run_bz_bnz(TestCaseMsaBranch* input, Branch GenerateBranch, |
| bool branched) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| CpuFeatureScope fscope(&assm, MIPS_SIMD); |
| |
| struct T { |
| uint64_t ws_lo; |
| uint64_t ws_hi; |
| uint64_t wd_lo; |
| uint64_t wd_hi; |
| }; |
| T t = {0x20B9CC4F1A83E0C5, 0xA27E1B5F2F5BB18A, 0x0000000000000000, |
| 0x0000000000000000}; |
| msa_reg_t res; |
| Label do_not_move_w0_to_w2; |
| |
| load_elements_of_vector(assm, &t.ws_lo, w0, t0, t1); |
| load_elements_of_vector(assm, &t.wd_lo, w2, t0, t1); |
| load_elements_of_vector(assm, &input->wt_lo, w1, t0, t1); |
| GenerateBranch(assm, do_not_move_w0_to_w2); |
| __ nop(); |
| __ move_v(w2, w0); |
| |
| __ bind(&do_not_move_w0_to_w2); |
| store_elements_of_vector(assm, w2, a0); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| #ifdef OBJECT_PRINT |
| code->Print(std::cout); |
| #endif |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| |
| (f.Call(&res, 0, 0, 0, 0)); |
| if (branched) { |
| CHECK_EQ(t.wd_lo, res.d[0]); |
| CHECK_EQ(t.wd_hi, res.d[1]); |
| } else { |
| CHECK_EQ(t.ws_lo, res.d[0]); |
| CHECK_EQ(t.ws_hi, res.d[1]); |
| } |
| } |
| |
| TEST(MSA_bz_bnz) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| TestCaseMsaBranch tz_v[] = { |
| {0x0, 0x0}, {0xABC, 0x0}, {0x0, 0xABC}, {0xABC, 0xABC}}; |
| for (unsigned i = 0; i < arraysize(tz_v); ++i) { |
| run_bz_bnz( |
| &tz_v[i], |
| [](MacroAssembler& assm, Label& br_target) { __ bz_v(w1, &br_target); }, |
| tz_v[i].wt_lo == 0 && tz_v[i].wt_hi == 0); |
| } |
| |
| #define TEST_BZ_DF(input_array, lanes, instruction, int_type) \ |
| for (unsigned i = 0; i < arraysize(input_array); ++i) { \ |
| int j; \ |
| int_type* element = reinterpret_cast<int_type*>(&input_array[i]); \ |
| for (j = 0; j < lanes; ++j) { \ |
| if (element[j] == 0) { \ |
| break; \ |
| } \ |
| } \ |
| run_bz_bnz(&input_array[i], \ |
| [](MacroAssembler& assm, Label& br_target) { \ |
| __ instruction(w1, &br_target); \ |
| }, \ |
| j != lanes); \ |
| } |
| TestCaseMsaBranch tz_b[] = {{0x0, 0x0}, |
| {0xBC0000, 0x0}, |
| {0x0, 0xAB000000000000CD}, |
| {0x123456789ABCDEF0, 0xAAAAAAAAAAAAAAAA}}; |
| TEST_BZ_DF(tz_b, kMSALanesByte, bz_b, int8_t) |
| |
| TestCaseMsaBranch tz_h[] = {{0x0, 0x0}, |
| {0xBCDE0000, 0x0}, |
| {0x0, 0xABCD00000000ABCD}, |
| {0x123456789ABCDEF0, 0xAAAAAAAAAAAAAAAA}}; |
| TEST_BZ_DF(tz_h, kMSALanesHalf, bz_h, int16_t) |
| |
| TestCaseMsaBranch tz_w[] = {{0x0, 0x0}, |
| {0xBCDE123400000000, 0x0}, |
| {0x0, 0x000000001234ABCD}, |
| {0x123456789ABCDEF0, 0xAAAAAAAAAAAAAAAA}}; |
| TEST_BZ_DF(tz_w, kMSALanesWord, bz_w, int32_t) |
| |
| TestCaseMsaBranch tz_d[] = {{0x0, 0x0}, |
| {0xBCDE0000, 0x0}, |
| {0x0, 0xABCD00000000ABCD}, |
| {0x123456789ABCDEF0, 0xAAAAAAAAAAAAAAAA}}; |
| TEST_BZ_DF(tz_d, kMSALanesDword, bz_d, int64_t) |
| #undef TEST_BZ_DF |
| |
| TestCaseMsaBranch tnz_v[] = { |
| {0x0, 0x0}, {0xABC, 0x0}, {0x0, 0xABC}, {0xABC, 0xABC}}; |
| for (unsigned i = 0; i < arraysize(tnz_v); ++i) { |
| run_bz_bnz(&tnz_v[i], |
| [](MacroAssembler& assm, Label& br_target) { |
| __ bnz_v(w1, &br_target); |
| }, |
| tnz_v[i].wt_lo != 0 || tnz_v[i].wt_hi != 0); |
| } |
| |
| #define TEST_BNZ_DF(input_array, lanes, instruction, int_type) \ |
| for (unsigned i = 0; i < arraysize(input_array); ++i) { \ |
| int j; \ |
| int_type* element = reinterpret_cast<int_type*>(&input_array[i]); \ |
| for (j = 0; j < lanes; ++j) { \ |
| if (element[j] == 0) { \ |
| break; \ |
| } \ |
| } \ |
| run_bz_bnz(&input_array[i], \ |
| [](MacroAssembler& assm, Label& br_target) { \ |
| __ instruction(w1, &br_target); \ |
| }, \ |
| j == lanes); \ |
| } |
| TestCaseMsaBranch tnz_b[] = {{0x0, 0x0}, |
| {0xBC0000, 0x0}, |
| {0x0, 0xAB000000000000CD}, |
| {0x123456789ABCDEF0, 0xAAAAAAAAAAAAAAAA}}; |
| TEST_BNZ_DF(tnz_b, 16, bnz_b, int8_t) |
| |
| TestCaseMsaBranch tnz_h[] = {{0x0, 0x0}, |
| {0xBCDE0000, 0x0}, |
| {0x0, 0xABCD00000000ABCD}, |
| {0x123456789ABCDEF0, 0xAAAAAAAAAAAAAAAA}}; |
| TEST_BNZ_DF(tnz_h, 8, bnz_h, int16_t) |
| |
| TestCaseMsaBranch tnz_w[] = {{0x0, 0x0}, |
| {0xBCDE123400000000, 0x0}, |
| {0x0, 0x000000001234ABCD}, |
| {0x123456789ABCDEF0, 0xAAAAAAAAAAAAAAAA}}; |
| TEST_BNZ_DF(tnz_w, 4, bnz_w, int32_t) |
| |
| TestCaseMsaBranch tnz_d[] = {{0x0, 0x0}, |
| {0xBCDE0000, 0x0}, |
| {0x0, 0xABCD00000000ABCD}, |
| {0x123456789ABCDEF0, 0xAAAAAAAAAAAAAAAA}}; |
| TEST_BNZ_DF(tnz_d, 2, bnz_d, int64_t) |
| #undef TEST_BNZ_DF |
| } |
| |
| uint32_t run_jialc(int16_t offset) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| Label main_block; |
| __ push(ra); |
| __ li(v0, 0); |
| __ beq(v0, v0, &main_block); |
| __ nop(); |
| |
| // Block 1 |
| __ addiu(v0, v0, 0x1); // <-- offset = -40 |
| __ addiu(v0, v0, 0x2); |
| __ jr(ra); |
| __ nop(); |
| |
| // Block 2 |
| __ addiu(v0, v0, 0x10); // <-- offset = -24 |
| __ addiu(v0, v0, 0x20); |
| __ jr(ra); |
| __ nop(); |
| |
| // Block 3 (Main) |
| __ bind(&main_block); |
| __ nal(); // t0 <- program counter |
| __ mov(t0, ra); |
| __ jialc(t0, offset); |
| __ addiu(v0, v0, 0x4); |
| __ pop(ra); |
| __ jr(ra); |
| __ nop(); |
| |
| // Block 4 |
| __ addiu(v0, v0, 0x100); // <-- offset = 20 |
| __ addiu(v0, v0, 0x200); |
| __ jr(ra); |
| __ nop(); |
| |
| // Block 5 |
| __ addiu(v0, v0, 0x1000); // <--- offset = 36 |
| __ addiu(v0, v0, 0x2000); |
| __ jr(ra); |
| __ nop(); |
| |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| |
| auto f = GeneratedCode<F2>::FromCode(*code); |
| |
| uint32_t res = reinterpret_cast<uint32_t>(f.Call(0, 0, 0, 0, 0)); |
| |
| return res; |
| } |
| |
| |
| TEST(r6_jialc) { |
| if (IsMipsArchVariant(kMips32r6)) { |
| CcTest::InitializeVM(); |
| |
| struct TestCaseJialc { |
| int16_t offset; |
| uint32_t expected_res; |
| }; |
| |
| struct TestCaseJialc tc[] = { |
| // offset, expected_res |
| { -40, 0x7 }, |
| { -24, 0x34 }, |
| { 20, 0x304 }, |
| { 36, 0x3004 } |
| }; |
| |
| size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseJialc); |
| for (size_t i = 0; i < nr_test_cases; ++i) { |
| uint32_t res = run_jialc(tc[i].offset); |
| CHECK_EQ(tc[i].expected_res, res); |
| } |
| } |
| } |
| |
| static uint32_t run_addiupc(int32_t imm19) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| __ addiupc(v0, imm19); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| |
| auto f = GeneratedCode<F2>::FromCode(*code); |
| PC = (uint32_t)code->entry(); // Set the program counter. |
| |
| uint32_t rs = reinterpret_cast<uint32_t>(f.Call(imm19, 0, 0, 0, 0)); |
| |
| return rs; |
| } |
| |
| |
| TEST(r6_addiupc) { |
| if (IsMipsArchVariant(kMips32r6)) { |
| CcTest::InitializeVM(); |
| |
| struct TestCaseAddiupc { |
| int32_t imm19; |
| }; |
| |
| TestCaseAddiupc tc[] = { |
| // imm19 |
| {-262144}, // 0x40000 |
| {-1}, // 0x7FFFF |
| {0}, |
| {1}, // 0x00001 |
| {262143} // 0x3FFFF |
| }; |
| |
| size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseAddiupc); |
| for (size_t i = 0; i < nr_test_cases; ++i) { |
| PC = 0; |
| uint32_t res = run_addiupc(tc[i].imm19); |
| // Now, the program_counter (PC) is set. |
| uint32_t expected_res = PC + (tc[i].imm19 << 2); |
| CHECK_EQ(expected_res, res); |
| } |
| } |
| } |
| |
| |
| int32_t run_bc(int32_t offset) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| Label continue_1, stop_execution; |
| __ push(ra); |
| __ li(v0, 0); |
| __ li(t8, 0); |
| __ li(t9, 2); // A condition for stopping execution. |
| |
| for (int32_t i = -100; i <= -11; ++i) { |
| __ addiu(v0, v0, 1); |
| } |
| |
| __ addiu(t8, t8, 1); // -10 |
| |
| __ beq(t8, t9, &stop_execution); // -9 |
| __ nop(); // -8 |
| __ beq(t8, t8, &continue_1); // -7 |
| __ nop(); // -6 |
| |
| __ bind(&stop_execution); |
| __ pop(ra); // -5, -4 |
| __ jr(ra); // -3 |
| __ nop(); // -2 |
| |
| __ bind(&continue_1); |
| __ bc(offset); // -1 |
| |
| for (int32_t i = 0; i <= 99; ++i) { |
| __ addiu(v0, v0, 1); |
| } |
| |
| __ pop(ra); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| |
| auto f = GeneratedCode<F2>::FromCode(*code); |
| |
| int32_t res = reinterpret_cast<int32_t>(f.Call(0, 0, 0, 0, 0)); |
| |
| return res; |
| } |
| |
| |
| TEST(r6_bc) { |
| if (IsMipsArchVariant(kMips32r6)) { |
| CcTest::InitializeVM(); |
| |
| struct TestCaseBc { |
| int32_t offset; |
| int32_t expected_res; |
| }; |
| |
| struct TestCaseBc tc[] = { |
| // offset, expected_result |
| { -100, (abs(-100) - 10) * 2 }, |
| { -11, (abs(-100) - 10 + 1) }, |
| { 0, (abs(-100) - 10 + 1 + 99) }, |
| { 1, (abs(-100) - 10 + 99) }, |
| { 99, (abs(-100) - 10 + 1) }, |
| }; |
| |
| size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseBc); |
| for (size_t i = 0; i < nr_test_cases; ++i) { |
| int32_t res = run_bc(tc[i].offset); |
| CHECK_EQ(tc[i].expected_res, res); |
| } |
| } |
| } |
| |
| |
| int32_t run_balc(int32_t offset) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| Label continue_1; |
| __ push(ra); |
| __ li(v0, 0); |
| __ li(t8, 0); |
| __ li(t9, 2); // A condition for stopping execution. |
| |
| __ beq(t8, t8, &continue_1); |
| __ nop(); |
| |
| uint32_t instruction_addiu = 0x24420001; // addiu v0, v0, 1 |
| for (int32_t i = -117; i <= -57; ++i) { |
| __ dd(instruction_addiu); |
| } |
| __ jr(ra); // -56 |
| __ nop(); // -55 |
| |
| for (int32_t i = -54; i <= -4; ++i) { |
| __ dd(instruction_addiu); |
| } |
| __ jr(ra); // -3 |
| __ nop(); // -2 |
| |
| __ bind(&continue_1); |
| __ balc(offset); // -1 |
| |
| __ pop(ra); // 0, 1 |
| __ jr(ra); // 2 |
| __ nop(); // 3 |
| |
| for (int32_t i = 4; i <= 44; ++i) { |
| __ dd(instruction_addiu); |
| } |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| |
| auto f = GeneratedCode<F2>::FromCode(*code); |
| |
| int32_t res = reinterpret_cast<int32_t>(f.Call(0, 0, 0, 0, 0)); |
| |
| return res; |
| } |
| |
| |
| uint32_t run_aui(uint32_t rs, uint16_t offset) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| __ li(t0, rs); |
| __ aui(v0, t0, offset); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| |
| auto f = GeneratedCode<F2>::FromCode(*code); |
| |
| uint32_t res = reinterpret_cast<uint32_t>(f.Call(0, 0, 0, 0, 0)); |
| |
| return res; |
| } |
| |
| |
| TEST(r6_aui) { |
| if (IsMipsArchVariant(kMips32r6)) { |
| CcTest::InitializeVM(); |
| |
| struct TestCaseAui { |
| uint32_t rs; |
| uint16_t offset; |
| uint32_t ref_res; |
| }; |
| |
| struct TestCaseAui tc[] = { |
| // input, offset, result |
| {0xFFFEFFFF, 1, 0xFFFFFFFF}, |
| {0xFFFFFFFF, 0, 0xFFFFFFFF}, |
| {0, 0xFFFF, 0xFFFF0000}, |
| {0x0008FFFF, 0xFFF7, 0xFFFFFFFF}, |
| {32767, 32767, 0x7FFF7FFF}, |
| // overflow cases |
| {0xFFFFFFFF, 0x1, 0x0000FFFF}, |
| {0xFFFFFFFF, 0xFFFF, 0xFFFEFFFF}, |
| }; |
| |
| size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseAui); |
| for (size_t i = 0; i < nr_test_cases; ++i) { |
| PC = 0; |
| uint32_t res = run_aui(tc[i].rs, tc[i].offset); |
| CHECK_EQ(tc[i].ref_res, res); |
| } |
| } |
| } |
| |
| |
| TEST(r6_balc) { |
| if (IsMipsArchVariant(kMips32r6)) { |
| CcTest::InitializeVM(); |
| |
| struct TestCaseBalc { |
| int32_t offset; |
| int32_t expected_res; |
| }; |
| |
| struct TestCaseBalc tc[] = { |
| // offset, expected_result |
| { -117, 61 }, |
| { -54, 51 }, |
| { 0, 0 }, |
| { 4, 41 }, |
| }; |
| |
| size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseBalc); |
| for (size_t i = 0; i < nr_test_cases; ++i) { |
| int32_t res = run_balc(tc[i].offset); |
| CHECK_EQ(tc[i].expected_res, res); |
| } |
| } |
| } |
| |
| |
| uint32_t run_bal(int16_t offset) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| __ mov(t0, ra); |
| __ bal(offset); // Equivalent for "BGEZAL zero_reg, offset". |
| __ nop(); |
| |
| __ mov(ra, t0); |
| __ jr(ra); |
| __ nop(); |
| |
| __ li(v0, 1); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| |
| auto f = GeneratedCode<F2>::FromCode(*code); |
| |
| uint32_t res = reinterpret_cast<uint32_t>(f.Call(0, 0, 0, 0, 0)); |
| |
| return res; |
| } |
| |
| |
| TEST(bal) { |
| CcTest::InitializeVM(); |
| |
| struct TestCaseBal { |
| int16_t offset; |
| uint32_t expected_res; |
| }; |
| |
| struct TestCaseBal tc[] = { |
| // offset, expected_res |
| { 4, 1 }, |
| }; |
| |
| size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseBal); |
| for (size_t i = 0; i < nr_test_cases; ++i) { |
| CHECK_EQ(tc[i].expected_res, run_bal(tc[i].offset)); |
| } |
| } |
| |
| |
| TEST(Trampoline) { |
| // Private member of Assembler class. |
| static const int kMaxBranchOffset = (1 << (18 - 1)) - 1; |
| |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| Label done; |
| size_t nr_calls = kMaxBranchOffset / (2 * kInstrSize) + 2; |
| |
| for (size_t i = 0; i < nr_calls; ++i) { |
| __ BranchShort(&done, eq, a0, Operand(a1)); |
| } |
| __ bind(&done); |
| __ Ret(USE_DELAY_SLOT); |
| __ mov(v0, zero_reg); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F2>::FromCode(*code); |
| |
| int32_t res = reinterpret_cast<int32_t>(f.Call(42, 42, 0, 0, 0)); |
| CHECK_EQ(0, res); |
| } |
| |
| template <class T> |
| struct TestCaseMaddMsub { |
| T fr, fs, ft, fd_add, fd_sub; |
| }; |
| |
| template <typename T, typename F> |
| void helper_madd_msub_maddf_msubf(F func) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| T x = std::sqrt(static_cast<T>(2.0)); |
| T y = std::sqrt(static_cast<T>(3.0)); |
| T z = std::sqrt(static_cast<T>(5.0)); |
| T x2 = 11.11, y2 = 22.22, z2 = 33.33; |
| TestCaseMaddMsub<T> test_cases[] = { |
| {x, y, z, 0.0, 0.0}, |
| {x, y, -z, 0.0, 0.0}, |
| {x, -y, z, 0.0, 0.0}, |
| {x, -y, -z, 0.0, 0.0}, |
| {-x, y, z, 0.0, 0.0}, |
| {-x, y, -z, 0.0, 0.0}, |
| {-x, -y, z, 0.0, 0.0}, |
| {-x, -y, -z, 0.0, 0.0}, |
| {-3.14, 0.2345, -123.000056, 0.0, 0.0}, |
| {7.3, -23.257, -357.1357, 0.0, 0.0}, |
| {x2, y2, z2, 0.0, 0.0}, |
| {x2, y2, -z2, 0.0, 0.0}, |
| {x2, -y2, z2, 0.0, 0.0}, |
| {x2, -y2, -z2, 0.0, 0.0}, |
| {-x2, y2, z2, 0.0, 0.0}, |
| {-x2, y2, -z2, 0.0, 0.0}, |
| {-x2, -y2, z2, 0.0, 0.0}, |
| {-x2, -y2, -z2, 0.0, 0.0}, |
| }; |
| |
| if (std::is_same<T, float>::value) { |
| __ lwc1(f4, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, fr))); |
| __ lwc1(f6, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, fs))); |
| __ lwc1(f8, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, ft))); |
| __ lwc1(f16, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, fr))); |
| } else if (std::is_same<T, double>::value) { |
| __ Ldc1(f4, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, fr))); |
| __ Ldc1(f6, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, fs))); |
| __ Ldc1(f8, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, ft))); |
| __ Ldc1(f16, MemOperand(a0, offsetof(TestCaseMaddMsub<T>, fr))); |
| } else { |
| UNREACHABLE(); |
| } |
| |
| func(assm); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| |
| const size_t kTableLength = sizeof(test_cases) / sizeof(TestCaseMaddMsub<T>); |
| TestCaseMaddMsub<T> tc; |
| for (size_t i = 0; i < kTableLength; i++) { |
| tc.fr = test_cases[i].fr; |
| tc.fs = test_cases[i].fs; |
| tc.ft = test_cases[i].ft; |
| |
| (f.Call(&tc, 0, 0, 0, 0)); |
| |
| T res_add = 0; |
| T res_sub = 0; |
| if (IsMipsArchVariant(kMips32r2)) { |
| res_add = (tc.fs * tc.ft) + tc.fr; |
| res_sub = (tc.fs * tc.ft) - tc.fr; |
| } else if (IsMipsArchVariant(kMips32r6)) { |
| res_add = std::fma(tc.fs, tc.ft, tc.fr); |
| res_sub = std::fma(-tc.fs, tc.ft, tc.fr); |
| } else { |
| UNREACHABLE(); |
| } |
| |
| CHECK_EQ(tc.fd_add, res_add); |
| CHECK_EQ(tc.fd_sub, res_sub); |
| } |
| } |
| |
| TEST(madd_msub_s) { |
| if (!IsMipsArchVariant(kMips32r2)) return; |
| helper_madd_msub_maddf_msubf<float>([](MacroAssembler& assm) { |
| __ madd_s(f10, f4, f6, f8); |
| __ swc1(f10, MemOperand(a0, offsetof(TestCaseMaddMsub<float>, fd_add))); |
| __ msub_s(f16, f4, f6, f8); |
| __ swc1(f16, MemOperand(a0, offsetof(TestCaseMaddMsub<float>, fd_sub))); |
| }); |
| } |
| |
| TEST(madd_msub_d) { |
| if (!IsMipsArchVariant(kMips32r2)) return; |
| helper_madd_msub_maddf_msubf<double>([](MacroAssembler& assm) { |
| __ madd_d(f10, f4, f6, f8); |
| __ Sdc1(f10, MemOperand(a0, offsetof(TestCaseMaddMsub<double>, fd_add))); |
| __ msub_d(f16, f4, f6, f8); |
| __ Sdc1(f16, MemOperand(a0, offsetof(TestCaseMaddMsub<double>, fd_sub))); |
| }); |
| } |
| |
| TEST(maddf_msubf_s) { |
| if (!IsMipsArchVariant(kMips32r6)) return; |
| helper_madd_msub_maddf_msubf<float>([](MacroAssembler& assm) { |
| __ maddf_s(f4, f6, f8); |
| __ swc1(f4, MemOperand(a0, offsetof(TestCaseMaddMsub<float>, fd_add))); |
| __ msubf_s(f16, f6, f8); |
| __ swc1(f16, MemOperand(a0, offsetof(TestCaseMaddMsub<float>, fd_sub))); |
| }); |
| } |
| |
| TEST(maddf_msubf_d) { |
| if (!IsMipsArchVariant(kMips32r6)) return; |
| helper_madd_msub_maddf_msubf<double>([](MacroAssembler& assm) { |
| __ maddf_d(f4, f6, f8); |
| __ Sdc1(f4, MemOperand(a0, offsetof(TestCaseMaddMsub<double>, fd_add))); |
| __ msubf_d(f16, f6, f8); |
| __ Sdc1(f16, MemOperand(a0, offsetof(TestCaseMaddMsub<double>, fd_sub))); |
| }); |
| } |
| |
| uint32_t run_Subu(uint32_t imm, int32_t num_instr) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| Label code_start; |
| __ bind(&code_start); |
| __ Subu(v0, zero_reg, imm); |
| CHECK_EQ(assm.SizeOfCodeGeneratedSince(&code_start), num_instr * kInstrSize); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F2>::FromCode(*code); |
| |
| uint32_t res = reinterpret_cast<uint32_t>(f.Call(0, 0, 0, 0, 0)); |
| |
| return res; |
| } |
| |
| TEST(Subu) { |
| CcTest::InitializeVM(); |
| |
| // Test Subu macro-instruction for min_int16 and max_int16 border cases. |
| // For subtracting int16 immediate values we use addiu. |
| |
| struct TestCaseSubu { |
| uint32_t imm; |
| uint32_t expected_res; |
| int32_t num_instr; |
| }; |
| |
| // We call Subu(v0, zero_reg, imm) to test cases listed below. |
| // 0 - imm = expected_res |
| struct TestCaseSubu tc[] = { |
| // imm, expected_res, num_instr |
| {0xFFFF8000, 0x8000, 2}, // min_int16 |
| // Generates ori + addu |
| // We can't have just addiu because -min_int16 > max_int16 so use |
| // register. We can load min_int16 to at register with addiu and then |
| // subtract at with subu, but now we use ori + addu because -min_int16 can |
| // be loaded using ori. |
| {0x8000, 0xFFFF8000, 1}, // max_int16 + 1 |
| // Generates addiu |
| // max_int16 + 1 is not int16 but -(max_int16 + 1) is, just use addiu. |
| {0xFFFF7FFF, 0x8001, 2}, // min_int16 - 1 |
| // Generates ori + addu |
| // To load this value to at we need two instructions and another one to |
| // subtract, lui + ori + subu. But we can load -value to at using just |
| // ori and then add at register with addu. |
| {0x8001, 0xFFFF7FFF, 2}, // max_int16 + 2 |
| // Generates ori + subu |
| // Not int16 but is uint16, load value to at with ori and subtract with |
| // subu. |
| {0x00010000, 0xFFFF0000, 2}, |
| // Generates lui + subu |
| // Load value using lui to at and subtract with subu. |
| {0x00010001, 0xFFFEFFFF, 3}, |
| // Generates lui + ori + subu |
| // We have to generate three instructions in this case. |
| }; |
| |
| size_t nr_test_cases = sizeof(tc) / sizeof(TestCaseSubu); |
| for (size_t i = 0; i < nr_test_cases; ++i) { |
| CHECK_EQ(tc[i].expected_res, run_Subu(tc[i].imm, tc[i].num_instr)); |
| } |
| } |
| |
| TEST(MSA_fill_copy) { |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| struct T { |
| uint32_t u8; |
| uint32_t u16; |
| uint32_t u32; |
| uint32_t s8; |
| uint32_t s16; |
| uint32_t s32; |
| }; |
| T t; |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| { |
| CpuFeatureScope fscope(&assm, MIPS_SIMD); |
| |
| __ li(t0, 0xA512B683); |
| |
| __ fill_b(w0, t0); |
| __ fill_h(w2, t0); |
| __ fill_w(w4, t0); |
| __ copy_u_b(t1, w0, 11); |
| __ sw(t1, MemOperand(a0, offsetof(T, u8))); |
| __ copy_u_h(t1, w2, 6); |
| __ sw(t1, MemOperand(a0, offsetof(T, u16))); |
| __ copy_u_w(t1, w4, 3); |
| __ sw(t1, MemOperand(a0, offsetof(T, u32))); |
| |
| __ copy_s_b(t1, w0, 8); |
| __ sw(t1, MemOperand(a0, offsetof(T, s8))); |
| __ copy_s_h(t1, w2, 5); |
| __ sw(t1, MemOperand(a0, offsetof(T, s16))); |
| __ copy_s_w(t1, w4, 1); |
| __ sw(t1, MemOperand(a0, offsetof(T, s32))); |
| |
| __ jr(ra); |
| __ nop(); |
| } |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| #ifdef OBJECT_PRINT |
| code->Print(std::cout); |
| #endif |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| |
| f.Call(&t, 0, 0, 0, 0); |
| |
| CHECK_EQ(0x83u, t.u8); |
| CHECK_EQ(0xB683u, t.u16); |
| CHECK_EQ(0xA512B683u, t.u32); |
| CHECK_EQ(0xFFFFFF83u, t.s8); |
| CHECK_EQ(0xFFFFB683u, t.s16); |
| CHECK_EQ(0xA512B683u, t.s32); |
| } |
| |
| TEST(MSA_fill_copy_2) { |
| // Similar to MSA_fill_copy test, but also check overlaping between MSA and |
| // FPU registers with same numbers |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| struct T { |
| uint32_t w0; |
| uint32_t w1; |
| uint32_t w2; |
| uint32_t w3; |
| }; |
| T t[2]; |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| { |
| CpuFeatureScope fscope(&assm, MIPS_SIMD); |
| |
| __ li(t0, 0xAAAAAAAA); |
| __ li(t1, 0x55555555); |
| |
| __ fill_w(w0, t0); |
| __ fill_w(w2, t0); |
| |
| __ FmoveLow(f0, t1); |
| __ FmoveHigh(f2, t1); |
| |
| #define STORE_MSA_REG(w_reg, base, scratch) \ |
| __ copy_u_w(scratch, w_reg, 0); \ |
| __ sw(scratch, MemOperand(base, offsetof(T, w0))); \ |
| __ copy_u_w(scratch, w_reg, 1); \ |
| __ sw(scratch, MemOperand(base, offsetof(T, w1))); \ |
| __ copy_u_w(scratch, w_reg, 2); \ |
| __ sw(scratch, MemOperand(base, offsetof(T, w2))); \ |
| __ copy_u_w(scratch, w_reg, 3); \ |
| __ sw(scratch, MemOperand(base, offsetof(T, w3))); |
| |
| STORE_MSA_REG(w0, a0, t2) |
| STORE_MSA_REG(w2, a1, t2) |
| #undef STORE_MSA_REG |
| |
| __ jr(ra); |
| __ nop(); |
| } |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| #ifdef OBJECT_PRINT |
| code->Print(std::cout); |
| #endif |
| auto f = GeneratedCode<F4>::FromCode(*code); |
| |
| f.Call(&t[0], &t[1], 0, 0, 0); |
| |
| CHECK_EQ(0x55555555, t[0].w0); |
| CHECK_EQ(0xAAAAAAAA, t[0].w1); |
| CHECK_EQ(0xAAAAAAAA, t[0].w2); |
| CHECK_EQ(0xAAAAAAAA, t[0].w3); |
| CHECK_EQ(0xAAAAAAAA, t[1].w0); |
| CHECK_EQ(0x55555555, t[1].w1); |
| CHECK_EQ(0xAAAAAAAA, t[1].w2); |
| CHECK_EQ(0xAAAAAAAA, t[1].w3); |
| } |
| |
| TEST(MSA_fill_copy_3) { |
| // Similar to MSA_fill_copy test, but also check overlaping between MSA and |
| // FPU registers with same numbers |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| struct T { |
| uint64_t d0; |
| uint64_t d1; |
| }; |
| T t[2]; |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| { |
| CpuFeatureScope fscope(&assm, MIPS_SIMD); |
| |
| __ li(t0, 0xAAAAAAAA); |
| __ li(t1, 0x55555555); |
| |
| __ Move(f0, t0, t0); |
| __ Move(f2, t0, t0); |
| |
| __ fill_w(w0, t1); |
| __ fill_w(w2, t1); |
| |
| __ Sdc1(f0, MemOperand(a0, offsetof(T, d0))); |
| __ Sdc1(f2, MemOperand(a1, offsetof(T, d0))); |
| |
| __ jr(ra); |
| __ nop(); |
| } |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| #ifdef OBJECT_PRINT |
| code->Print(std::cout); |
| #endif |
| auto f = GeneratedCode<F4>::FromCode(*code); |
| |
| f.Call(&t[0], &t[1], 0, 0, 0); |
| |
| CHECK_EQ(0x5555555555555555, t[0].d0); |
| CHECK_EQ(0x5555555555555555, t[1].d0); |
| } |
| |
| template <typename T> |
| void run_msa_insert(int32_t rs_value, int n, msa_reg_t* w) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| CpuFeatureScope fscope(&assm, MIPS_SIMD); |
| |
| __ li(t0, -1); |
| __ li(t1, rs_value); |
| __ fill_w(w0, t0); |
| |
| if (std::is_same<T, int8_t>::value) { |
| DCHECK_LT(n, 16); |
| __ insert_b(w0, n, t1); |
| } else if (std::is_same<T, int16_t>::value) { |
| DCHECK_LT(n, 8); |
| __ insert_h(w0, n, t1); |
| } else if (std::is_same<T, int32_t>::value) { |
| DCHECK_LT(n, 4); |
| __ insert_w(w0, n, t1); |
| } else { |
| UNREACHABLE(); |
| } |
| |
| store_elements_of_vector(assm, w0, a0); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| #ifdef OBJECT_PRINT |
| code->Print(std::cout); |
| #endif |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| |
| (f.Call(w, 0, 0, 0, 0)); |
| } |
| |
| TEST(MSA_insert) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| struct TestCaseInsert { |
| uint32_t input; |
| int n; |
| uint64_t exp_res_lo; |
| uint64_t exp_res_hi; |
| }; |
| |
| struct TestCaseInsert tc_b[] = { |
| // input, n, exp_res_lo, exp_res_hi |
| {0xA2, 13, 0xFFFFFFFFFFFFFFFFu, 0xFFFFA2FFFFFFFFFFu}, |
| {0x73, 10, 0xFFFFFFFFFFFFFFFFu, 0xFFFFFFFFFF73FFFFu}, |
| {0x3494, 5, 0xFFFF94FFFFFFFFFFu, 0xFFFFFFFFFFFFFFFFu}, |
| {0xA6B8, 1, 0xFFFFFFFFFFFFB8FFu, 0xFFFFFFFFFFFFFFFFu}}; |
| |
| for (size_t i = 0; i < sizeof(tc_b) / sizeof(TestCaseInsert); ++i) { |
| msa_reg_t res; |
| run_msa_insert<int8_t>(tc_b[i].input, tc_b[i].n, &res); |
| CHECK_EQ(tc_b[i].exp_res_lo, res.d[0]); |
| CHECK_EQ(tc_b[i].exp_res_hi, res.d[1]); |
| } |
| |
| struct TestCaseInsert tc_h[] = { |
| // input, n, exp_res_lo, exp_res_hi |
| {0x85A2, 7, 0xFFFFFFFFFFFFFFFFu, 0x85A2FFFFFFFFFFFFu}, |
| {0xE873, 5, 0xFFFFFFFFFFFFFFFFu, 0xFFFFFFFFE873FFFFu}, |
| {0x3494, 3, 0x3494FFFFFFFFFFFFu, 0xFFFFFFFFFFFFFFFFu}, |
| {0xA6B8, 1, 0xFFFFFFFFA6B8FFFFu, 0xFFFFFFFFFFFFFFFFu}}; |
| |
| for (size_t i = 0; i < sizeof(tc_h) / sizeof(TestCaseInsert); ++i) { |
| msa_reg_t res; |
| run_msa_insert<int16_t>(tc_h[i].input, tc_h[i].n, &res); |
| CHECK_EQ(tc_h[i].exp_res_lo, res.d[0]); |
| CHECK_EQ(tc_h[i].exp_res_hi, res.d[1]); |
| } |
| |
| struct TestCaseInsert tc_w[] = { |
| // input, n, exp_res_lo, exp_res_hi |
| {0xD2F085A2u, 3, 0xFFFFFFFFFFFFFFFFu, 0xD2F085A2FFFFFFFFu}, |
| {0x4567E873u, 2, 0xFFFFFFFFFFFFFFFFu, 0xFFFFFFFF4567E873u}, |
| {0xACDB3494u, 1, 0xACDB3494FFFFFFFFu, 0xFFFFFFFFFFFFFFFFu}, |
| {0x89ABA6B8u, 0, 0xFFFFFFFF89ABA6B8u, 0xFFFFFFFFFFFFFFFFu}}; |
| |
| for (size_t i = 0; i < sizeof(tc_w) / sizeof(TestCaseInsert); ++i) { |
| msa_reg_t res; |
| run_msa_insert<int32_t>(tc_w[i].input, tc_w[i].n, &res); |
| CHECK_EQ(tc_w[i].exp_res_lo, res.d[0]); |
| CHECK_EQ(tc_w[i].exp_res_hi, res.d[1]); |
| } |
| } |
| |
| TEST(MSA_move_v) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| CcTest::InitializeVM(); |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| struct T { |
| uint64_t ws_lo; |
| uint64_t ws_hi; |
| uint64_t wd_lo; |
| uint64_t wd_hi; |
| }; |
| T t[] = {{0x20B9CC4F1A83E0C5, 0xA27E1B5F2F5BB18A, 0x1E86678B52F8E1FF, |
| 0x706E51290AC76FB9}, |
| {0x4414AED7883FFD18, 0x047D183A06B67016, 0x4EF258CF8D822870, |
| 0x2686B73484C2E843}, |
| {0xD38FF9D048884FFC, 0x6DC63A57C0943CA7, 0x8520CA2F3E97C426, |
| 0xA9913868FB819C59}}; |
| |
| for (unsigned i = 0; i < arraysize(t); ++i) { |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| CpuFeatureScope fscope(&assm, MIPS_SIMD); |
| |
| load_elements_of_vector(assm, &t[i].ws_lo, w0, t0, t1); |
| load_elements_of_vector(assm, &t[i].wd_lo, w2, t0, t1); |
| __ move_v(w2, w0); |
| store_elements_of_vector(assm, w2, a0); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| #ifdef OBJECT_PRINT |
| code->Print(std::cout); |
| #endif |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| (f.Call(&t[i].wd_lo, 0, 0, 0, 0)); |
| CHECK_EQ(t[i].ws_lo, t[i].wd_lo); |
| CHECK_EQ(t[i].ws_hi, t[i].wd_hi); |
| } |
| } |
| |
| template <typename ExpectFunc, typename OperFunc> |
| void run_msa_sldi(OperFunc GenerateOperation, |
| ExpectFunc GenerateExpectedResult) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| struct T { |
| uint64_t ws_lo; |
| uint64_t ws_hi; |
| uint64_t wd_lo; |
| uint64_t wd_hi; |
| }; |
| T t[] = {{0x20B9CC4F1A83E0C5, 0xA27E1B5F2F5BB18A, 0x1E86678B52F8E1FF, |
| 0x706E51290AC76FB9}, |
| {0x4414AED7883FFD18, 0x047D183A06B67016, 0x4EF258CF8D822870, |
| 0x2686B73484C2E843}, |
| {0xD38FF9D048884FFC, 0x6DC63A57C0943CA7, 0x8520CA2F3E97C426, |
| 0xA9913868FB819C59}}; |
| uint64_t res[2]; |
| |
| for (unsigned i = 0; i < arraysize(t); ++i) { |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| CpuFeatureScope fscope(&assm, MIPS_SIMD); |
| load_elements_of_vector(assm, &t[i].ws_lo, w0, t0, t1); |
| load_elements_of_vector(assm, &t[i].wd_lo, w2, t0, t1); |
| GenerateOperation(assm); |
| store_elements_of_vector(assm, w2, a0); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| #ifdef OBJECT_PRINT |
| code->Print(std::cout); |
| #endif |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| (f.Call(&res[0], 0, 0, 0, 0)); |
| GenerateExpectedResult(reinterpret_cast<uint8_t*>(&t[i].ws_lo), |
| reinterpret_cast<uint8_t*>(&t[i].wd_lo)); |
| CHECK_EQ(res[0], t[i].wd_lo); |
| CHECK_EQ(res[1], t[i].wd_hi); |
| } |
| } |
| |
| TEST(MSA_sldi) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| CcTest::InitializeVM(); |
| |
| #define SLDI_DF(s, k) \ |
| uint8_t v[32]; \ |
| for (unsigned i = 0; i < s; i++) { \ |
| v[i] = ws[s * k + i]; \ |
| v[i + s] = wd[s * k + i]; \ |
| } \ |
| for (unsigned i = 0; i < s; i++) { \ |
| wd[s * k + i] = v[i + n]; \ |
| } |
| |
| for (int n = 0; n < 16; ++n) { |
| run_msa_sldi([n](MacroAssembler& assm) { __ sldi_b(w2, w0, n); }, |
| [n](uint8_t* ws, uint8_t* wd) { |
| SLDI_DF(kMSARegSize / sizeof(int8_t) / kBitsPerByte, 0) |
| }); |
| } |
| |
| for (int n = 0; n < 8; ++n) { |
| run_msa_sldi([n](MacroAssembler& assm) { __ sldi_h(w2, w0, n); }, |
| [n](uint8_t* ws, uint8_t* wd) { |
| for (int k = 0; k < 2; ++k) { |
| SLDI_DF(kMSARegSize / sizeof(int16_t) / kBitsPerByte, k) |
| } |
| }); |
| } |
| |
| for (int n = 0; n < 4; ++n) { |
| run_msa_sldi([n](MacroAssembler& assm) { __ sldi_w(w2, w0, n); }, |
| [n](uint8_t* ws, uint8_t* wd) { |
| for (int k = 0; k < 4; ++k) { |
| SLDI_DF(kMSARegSize / sizeof(int32_t) / kBitsPerByte, k) |
| } |
| }); |
| } |
| |
| for (int n = 0; n < 2; ++n) { |
| run_msa_sldi([n](MacroAssembler& assm) { __ sldi_d(w2, w0, n); }, |
| [n](uint8_t* ws, uint8_t* wd) { |
| for (int k = 0; k < 8; ++k) { |
| SLDI_DF(kMSARegSize / sizeof(int64_t) / kBitsPerByte, k) |
| } |
| }); |
| } |
| #undef SLDI_DF |
| } |
| |
| void run_msa_ctc_cfc(uint32_t value) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| CpuFeatureScope fscope(&assm, MIPS_SIMD); |
| |
| MSAControlRegister msareg = {kMSACSRRegister}; |
| __ li(t0, value); |
| __ li(t2, 0); |
| __ cfcmsa(t1, msareg); |
| __ ctcmsa(msareg, t0); |
| __ cfcmsa(t2, msareg); |
| __ ctcmsa(msareg, t1); |
| __ sw(t2, MemOperand(a0, 0)); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| #ifdef OBJECT_PRINT |
| code->Print(std::cout); |
| #endif |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| |
| uint32_t res; |
| (f.Call(&res, 0, 0, 0, 0)); |
| |
| CHECK_EQ(value & 0x0167FFFF, res); |
| } |
| |
| TEST(MSA_cfc_ctc) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| const uint32_t mask_without_cause = 0xFF9C0FFF; |
| const uint32_t mask_always_zero = 0x0167FFFF; |
| const uint32_t mask_enables = 0x00000F80; |
| uint32_t test_case[] = {0x2D5EDE31, 0x07955425, 0x15B7DBE3, 0x2BF8BC37, |
| 0xE6AAE923, 0x24D0F68D, 0x41AFA84C, 0x2D6BF64F, |
| 0x925014BD, 0x4DBA7E61}; |
| for (unsigned i = 0; i < arraysize(test_case); i++) { |
| // Setting enable bits and corresponding cause bits could result in |
| // exception raised and this prevents that from happening |
| test_case[i] = (~test_case[i] & mask_enables) << 5 | |
| (test_case[i] & mask_without_cause); |
| run_msa_ctc_cfc(test_case[i] & mask_always_zero); |
| } |
| } |
| |
| struct ExpResShf { |
| uint8_t i8; |
| uint64_t lo; |
| uint64_t hi; |
| }; |
| |
| void run_msa_i8(SecondaryField opcode, uint64_t ws_lo, uint64_t ws_hi, |
| uint8_t i8) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| CpuFeatureScope fscope(&assm, MIPS_SIMD); |
| msa_reg_t res; |
| uint64_t wd_lo = 0xF35862E13E38F8B0; |
| uint64_t wd_hi = 0x4F41FFDEF2BFE636; |
| |
| #define LOAD_W_REG(lo, hi, w_reg) \ |
| __ li(t0, static_cast<uint32_t>(lo & 0xFFFFFFFF)); \ |
| __ li(t1, static_cast<uint32_t>((lo >> 32) & 0xFFFFFFFF)); \ |
| __ insert_w(w_reg, 0, t0); \ |
| __ insert_w(w_reg, 1, t1); \ |
| __ li(t0, static_cast<uint32_t>(hi & 0xFFFFFFFF)); \ |
| __ li(t1, static_cast<uint32_t>((hi >> 32) & 0xFFFFFFFF)); \ |
| __ insert_w(w_reg, 2, t0); \ |
| __ insert_w(w_reg, 3, t1); |
| |
| LOAD_W_REG(ws_lo, ws_hi, w0) |
| |
| switch (opcode) { |
| case ANDI_B: |
| __ andi_b(w2, w0, i8); |
| break; |
| case ORI_B: |
| __ ori_b(w2, w0, i8); |
| break; |
| case NORI_B: |
| __ nori_b(w2, w0, i8); |
| break; |
| case XORI_B: |
| __ xori_b(w2, w0, i8); |
| break; |
| case BMNZI_B: |
| LOAD_W_REG(wd_lo, wd_hi, w2); |
| __ bmnzi_b(w2, w0, i8); |
| break; |
| case BMZI_B: |
| LOAD_W_REG(wd_lo, wd_hi, w2); |
| __ bmzi_b(w2, w0, i8); |
| break; |
| case BSELI_B: |
| LOAD_W_REG(wd_lo, wd_hi, w2); |
| __ bseli_b(w2, w0, i8); |
| break; |
| case SHF_B: |
| __ shf_b(w2, w0, i8); |
| break; |
| case SHF_H: |
| __ shf_h(w2, w0, i8); |
| break; |
| case SHF_W: |
| __ shf_w(w2, w0, i8); |
| break; |
| default: |
| UNREACHABLE(); |
| } |
| |
| store_elements_of_vector(assm, w2, a0); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| #undef LOAD_W_REG |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| #ifdef OBJECT_PRINT |
| code->Print(std::cout); |
| #endif |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| |
| (f.Call(&res, 0, 0, 0, 0)); |
| |
| uint64_t mask = i8 * 0x0101010101010101ull; |
| switch (opcode) { |
| case ANDI_B: |
| CHECK_EQ(ws_lo & mask, res.d[0]); |
| CHECK_EQ(ws_hi & mask, res.d[1]); |
| break; |
| case ORI_B: |
| CHECK_EQ(ws_lo | mask, res.d[0]); |
| CHECK_EQ(ws_hi | mask, res.d[1]); |
| break; |
| case NORI_B: |
| CHECK_EQ(~(ws_lo | mask), res.d[0]); |
| CHECK_EQ(~(ws_hi | mask), res.d[1]); |
| break; |
| case XORI_B: |
| CHECK_EQ(ws_lo ^ mask, res.d[0]); |
| CHECK_EQ(ws_hi ^ mask, res.d[1]); |
| break; |
| case BMNZI_B: |
| CHECK_EQ((ws_lo & mask) | (wd_lo & ~mask), res.d[0]); |
| CHECK_EQ((ws_hi & mask) | (wd_hi & ~mask), res.d[1]); |
| break; |
| case BMZI_B: |
| CHECK_EQ((ws_lo & ~mask) | (wd_lo & mask), res.d[0]); |
| CHECK_EQ((ws_hi & ~mask) | (wd_hi & mask), res.d[1]); |
| break; |
| case BSELI_B: |
| CHECK_EQ((ws_lo & ~wd_lo) | (mask & wd_lo), res.d[0]); |
| CHECK_EQ((ws_hi & ~wd_hi) | (mask & wd_hi), res.d[1]); |
| break; |
| case SHF_B: { |
| struct ExpResShf exp_b[] = { |
| // i8, exp_lo, exp_hi |
| {0xFFu, 0x11111111B9B9B9B9, 0xF7F7F7F7C8C8C8C8}, |
| {0x0u, 0x62626262DFDFDFDF, 0xD6D6D6D6C8C8C8C8}, |
| {0xE4u, 0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636}, |
| {0x1Bu, 0x1B756911C3D9A7B9, 0xAE94A5F79C8AEFC8}, |
| {0xB1u, 0x662B6253E8C4DF12, 0x0D3AD6803F8BC88B}, |
| {0x4Eu, 0x62E1F358F8B03E38, 0xFFDE4F41E636F2BF}, |
| {0x27u, 0x1B697511C3A7D9B9, 0xAEA594F79CEF8AC8}}; |
| for (size_t i = 0; i < sizeof(exp_b) / sizeof(ExpResShf); ++i) { |
| if (exp_b[i].i8 == i8) { |
| CHECK_EQ(exp_b[i].lo, res.d[0]); |
| CHECK_EQ(exp_b[i].hi, res.d[1]); |
| } |
| } |
| } break; |
| case SHF_H: { |
| struct ExpResShf exp_h[] = { |
| // i8, exp_lo, exp_hi |
| {0xFFu, 0x1169116911691169, 0xF7A5F7A5F7A5F7A5}, |
| {0x0u, 0x12DF12DF12DF12DF, 0x8BC88BC88BC88BC8}, |
| {0xE4u, 0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636}, |
| {0x1Bu, 0xD9C3B9A7751B1169, 0x8A9CC8EF94AEF7A5}, |
| {0xB1u, 0x53622B6612DFC4E8, 0x80D63A0D8BC88B3F}, |
| {0x4Eu, 0x3E38F8B0F35862E1, 0xF2BFE6364F41FFDE}, |
| {0x27u, 0xD9C3751BB9A71169, 0x8A9C94AEC8EFF7A5}}; |
| for (size_t i = 0; i < sizeof(exp_h) / sizeof(ExpResShf); ++i) { |
| if (exp_h[i].i8 == i8) { |
| CHECK_EQ(exp_h[i].lo, res.d[0]); |
| CHECK_EQ(exp_h[i].hi, res.d[1]); |
| } |
| } |
| } break; |
| case SHF_W: { |
| struct ExpResShf exp_w[] = { |
| // i8, exp_lo, exp_hi |
| {0xFFu, 0xF7A594AEF7A594AE, 0xF7A594AEF7A594AE}, |
| {0x0u, 0xC4E812DFC4E812DF, 0xC4E812DFC4E812DF}, |
| {0xE4u, 0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636}, |
| {0x1Bu, 0xC8EF8A9CF7A594AE, 0xB9A7D9C31169751B}, |
| {0xB1u, 0xC4E812DF2B665362, 0x8B3F8BC83A0D80D6}, |
| {0x4Eu, 0x4F41FFDEF2BFE636, 0xF35862E13E38F8B0}, |
| {0x27u, 0x1169751BF7A594AE, 0xB9A7D9C3C8EF8A9C}}; |
| for (size_t i = 0; i < sizeof(exp_w) / sizeof(ExpResShf); ++i) { |
| if (exp_w[i].i8 == i8) { |
| CHECK_EQ(exp_w[i].lo, res.d[0]); |
| CHECK_EQ(exp_w[i].hi, res.d[1]); |
| } |
| } |
| } break; |
| default: |
| UNREACHABLE(); |
| } |
| } |
| |
| struct TestCaseMsaI8 { |
| uint64_t input_lo; |
| uint64_t input_hi; |
| uint8_t i8; |
| }; |
| |
| TEST(MSA_andi_ori_nori_xori) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| struct TestCaseMsaI8 tc[] = {// input_lo, input_hi, i8 |
| {0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0xFFu}, |
| {0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0x0u}, |
| {0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0x3Bu}, |
| {0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0xD9u}}; |
| |
| for (size_t i = 0; i < sizeof(tc) / sizeof(TestCaseMsaI8); ++i) { |
| run_msa_i8(ANDI_B, tc[i].input_lo, tc[i].input_hi, tc[i].i8); |
| run_msa_i8(ORI_B, tc[i].input_lo, tc[i].input_hi, tc[i].i8); |
| run_msa_i8(NORI_B, tc[i].input_lo, tc[i].input_hi, tc[i].i8); |
| run_msa_i8(XORI_B, tc[i].input_lo, tc[i].input_hi, tc[i].i8); |
| } |
| } |
| |
| TEST(MSA_bmnzi_bmzi_bseli) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| struct TestCaseMsaI8 tc[] = {// input_lo, input_hi, i8 |
| {0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0xFFu}, |
| {0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0x0u}, |
| {0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0x3Bu}, |
| {0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0xD9u}}; |
| |
| for (size_t i = 0; i < sizeof(tc) / sizeof(TestCaseMsaI8); ++i) { |
| run_msa_i8(BMNZI_B, tc[i].input_lo, tc[i].input_hi, tc[i].i8); |
| run_msa_i8(BMZI_B, tc[i].input_lo, tc[i].input_hi, tc[i].i8); |
| run_msa_i8(BSELI_B, tc[i].input_lo, tc[i].input_hi, tc[i].i8); |
| } |
| } |
| |
| TEST(MSA_shf) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| struct TestCaseMsaI8 tc[] = { |
| // input_lo, input_hi, i8 |
| {0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0xFFu}, // 3333 |
| {0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0x0u}, // 0000 |
| {0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636, 0xE4u}, // 3210 |
| {0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0x1Bu}, // 0123 |
| {0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0xB1u}, // 2301 |
| {0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636, 0x4Eu}, // 1032 |
| {0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0x27u} // 0213 |
| }; |
| |
| for (size_t i = 0; i < sizeof(tc) / sizeof(TestCaseMsaI8); ++i) { |
| run_msa_i8(SHF_B, tc[i].input_lo, tc[i].input_hi, tc[i].i8); |
| run_msa_i8(SHF_H, tc[i].input_lo, tc[i].input_hi, tc[i].i8); |
| run_msa_i8(SHF_W, tc[i].input_lo, tc[i].input_hi, tc[i].i8); |
| } |
| } |
| |
| uint32_t run_Ins(uint32_t imm, uint32_t source, uint16_t pos, uint16_t size) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| __ li(v0, imm); |
| __ li(t0, source); |
| __ Ins(v0, t0, pos, size); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F2>::FromCode(*code); |
| |
| uint32_t res = reinterpret_cast<uint32_t>(f.Call(0, 0, 0, 0, 0)); |
| |
| return res; |
| } |
| |
| TEST(Ins) { |
| CcTest::InitializeVM(); |
| |
| // run_Ins(rt_value, rs_value, pos, size), expected_result |
| CHECK_EQ(run_Ins(0x55555555, 0xABCDEF01, 31, 1), 0xD5555555); |
| CHECK_EQ(run_Ins(0x55555555, 0xABCDEF02, 30, 2), 0x95555555); |
| CHECK_EQ(run_Ins(0x01234567, 0xFABCDEFF, 0, 32), 0xFABCDEFF); |
| |
| // Results with positive sign. |
| CHECK_EQ(run_Ins(0x55555550, 0x80000001, 0, 1), 0x55555551); |
| CHECK_EQ(run_Ins(0x55555555, 0x40000001, 0, 32), 0x40000001); |
| CHECK_EQ(run_Ins(0x55555555, 0x20000001, 1, 31), 0x40000003); |
| CHECK_EQ(run_Ins(0x55555555, 0x80700001, 8, 24), 0x70000155); |
| CHECK_EQ(run_Ins(0x55555555, 0x80007001, 16, 16), 0x70015555); |
| CHECK_EQ(run_Ins(0x55555555, 0x80000071, 24, 8), 0x71555555); |
| CHECK_EQ(run_Ins(0x75555555, 0x40000000, 31, 1), 0x75555555); |
| |
| // Results with negative sign. |
| CHECK_EQ(run_Ins(0x85555550, 0x80000001, 0, 1), 0x85555551); |
| CHECK_EQ(run_Ins(0x55555555, 0x80000001, 0, 32), 0x80000001); |
| CHECK_EQ(run_Ins(0x55555555, 0x40000001, 1, 31), 0x80000003); |
| CHECK_EQ(run_Ins(0x55555555, 0x80800001, 8, 24), 0x80000155); |
| CHECK_EQ(run_Ins(0x55555555, 0x80008001, 16, 16), 0x80015555); |
| CHECK_EQ(run_Ins(0x55555555, 0x80000081, 24, 8), 0x81555555); |
| CHECK_EQ(run_Ins(0x75555555, 0x00000001, 31, 1), 0xF5555555); |
| } |
| |
| uint32_t run_Ext(uint32_t source, uint16_t pos, uint16_t size) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| |
| __ li(v0, 0xFFFFFFFF); |
| __ li(t0, source); |
| __ Ext(v0, t0, pos, size); |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| auto f = GeneratedCode<F2>::FromCode(*code); |
| |
| uint32_t res = reinterpret_cast<uint32_t>(f.Call(0, 0, 0, 0, 0)); |
| |
| return res; |
| } |
| |
| TEST(Ext) { |
| CcTest::InitializeVM(); |
| |
| // Source values with negative sign. |
| // run_Ext(rs_value, pos, size), expected_result |
| CHECK_EQ(run_Ext(0x80000001, 0, 1), 0x00000001); |
| CHECK_EQ(run_Ext(0x80000001, 0, 32), 0x80000001); |
| CHECK_EQ(run_Ext(0x80000002, 1, 31), 0x40000001); |
| CHECK_EQ(run_Ext(0x80000100, 8, 24), 0x00800001); |
| CHECK_EQ(run_Ext(0x80010000, 16, 16), 0x00008001); |
| CHECK_EQ(run_Ext(0x81000000, 24, 8), 0x00000081); |
| CHECK_EQ(run_Ext(0x80000000, 31, 1), 0x00000001); |
| |
| // Source values with positive sign. |
| CHECK_EQ(run_Ext(0x00000001, 0, 1), 0x00000001); |
| CHECK_EQ(run_Ext(0x40000001, 0, 32), 0x40000001); |
| CHECK_EQ(run_Ext(0x40000002, 1, 31), 0x20000001); |
| CHECK_EQ(run_Ext(0x40000100, 8, 24), 0x00400001); |
| CHECK_EQ(run_Ext(0x40010000, 16, 16), 0x00004001); |
| CHECK_EQ(run_Ext(0x41000000, 24, 8), 0x00000041); |
| CHECK_EQ(run_Ext(0x40000000, 31, 1), 0x00000000); |
| } |
| |
| struct TestCaseMsaI5 { |
| uint64_t ws_lo; |
| uint64_t ws_hi; |
| uint32_t i5; |
| }; |
| |
| template <typename InstFunc, typename OperFunc> |
| void run_msa_i5(struct TestCaseMsaI5* input, bool i5_sign_ext, |
| InstFunc GenerateI5InstructionFunc, |
| OperFunc GenerateOperationFunc) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| CpuFeatureScope fscope(&assm, MIPS_SIMD); |
| msa_reg_t res; |
| int32_t i5 = |
| i5_sign_ext ? static_cast<int32_t>(input->i5 << 27) >> 27 : input->i5; |
| |
| load_elements_of_vector(assm, &(input->ws_lo), w0, t0, t1); |
| |
| GenerateI5InstructionFunc(assm, i5); |
| |
| store_elements_of_vector(assm, w2, a0); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| #ifdef OBJECT_PRINT |
| code->Print(std::cout); |
| #endif |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| |
| (f.Call(&res, 0, 0, 0, 0)); |
| |
| CHECK_EQ(GenerateOperationFunc(input->ws_lo, input->i5), res.d[0]); |
| CHECK_EQ(GenerateOperationFunc(input->ws_hi, input->i5), res.d[1]); |
| } |
| |
| TEST(MSA_addvi_subvi) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| struct TestCaseMsaI5 tc[] = { |
| // ws_lo, ws_hi, i5 |
| {0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0x0000001F}, |
| {0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0x0000000F}, |
| {0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0x00000005}, |
| {0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0x00000010}, |
| {0xFFAB807F807FFFCD, 0x7F23FF80FF567F80, 0x0000000F}, |
| {0x80FFEFFF7F12807F, 0x807F80FF7FDEFF78, 0x00000010}}; |
| |
| #define ADDVI_DF(lanes, mask) \ |
| uint64_t res = 0; \ |
| for (int i = 0; i < lanes / 2; ++i) { \ |
| int shift = (kMSARegSize / lanes) * i; \ |
| res |= ((((ws >> shift) & mask) + i5) & mask) << shift; \ |
| } \ |
| return res |
| |
| #define SUBVI_DF(lanes, mask) \ |
| uint64_t res = 0; \ |
| for (int i = 0; i < lanes / 2; ++i) { \ |
| int shift = (kMSARegSize / lanes) * i; \ |
| res |= ((((ws >> shift) & mask) - i5) & mask) << shift; \ |
| } \ |
| return res |
| |
| for (size_t i = 0; i < sizeof(tc) / sizeof(TestCaseMsaI5); ++i) { |
| run_msa_i5( |
| &tc[i], false, |
| [](MacroAssembler& assm, int32_t i5) { __ addvi_b(w2, w0, i5); }, |
| [](uint64_t ws, uint32_t i5) { ADDVI_DF(kMSALanesByte, UINT8_MAX); }); |
| |
| run_msa_i5( |
| &tc[i], false, |
| [](MacroAssembler& assm, int32_t i5) { __ addvi_h(w2, w0, i5); }, |
| [](uint64_t ws, uint32_t i5) { ADDVI_DF(kMSALanesHalf, UINT16_MAX); }); |
| |
| run_msa_i5( |
| &tc[i], false, |
| [](MacroAssembler& assm, int32_t i5) { __ addvi_w(w2, w0, i5); }, |
| [](uint64_t ws, uint32_t i5) { ADDVI_DF(kMSALanesWord, UINT32_MAX); }); |
| |
| run_msa_i5( |
| &tc[i], false, |
| [](MacroAssembler& assm, int32_t i5) { __ addvi_d(w2, w0, i5); }, |
| [](uint64_t ws, uint32_t i5) { ADDVI_DF(kMSALanesDword, UINT64_MAX); }); |
| |
| run_msa_i5( |
| &tc[i], false, |
| [](MacroAssembler& assm, int32_t i5) { __ subvi_b(w2, w0, i5); }, |
| [](uint64_t ws, uint32_t i5) { SUBVI_DF(kMSALanesByte, UINT8_MAX); }); |
| |
| run_msa_i5( |
| &tc[i], false, |
| [](MacroAssembler& assm, int32_t i5) { __ subvi_h(w2, w0, i5); }, |
| [](uint64_t ws, uint32_t i5) { SUBVI_DF(kMSALanesHalf, UINT16_MAX); }); |
| |
| run_msa_i5( |
| &tc[i], false, |
| [](MacroAssembler& assm, int32_t i5) { __ subvi_w(w2, w0, i5); }, |
| [](uint64_t ws, uint32_t i5) { SUBVI_DF(kMSALanesWord, UINT32_MAX); }); |
| |
| run_msa_i5( |
| &tc[i], false, |
| [](MacroAssembler& assm, int32_t i5) { __ subvi_d(w2, w0, i5); }, |
| [](uint64_t ws, uint32_t i5) { SUBVI_DF(kMSALanesDword, UINT64_MAX); }); |
| } |
| #undef ADDVI_DF |
| #undef SUBVI_DF |
| } |
| |
| TEST(MSA_maxi_mini) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| struct TestCaseMsaI5 tc[] = { |
| // ws_lo, ws_hi, i5 |
| {0x7F80FF3480FF7F00, 0x8D7FFF80FF7F6780, 0x0000001F}, |
| {0x7F80FF3480FF7F00, 0x8D7FFF80FF7F6780, 0x0000000F}, |
| {0x7F80FF3480FF7F00, 0x8D7FFF80FF7F6780, 0x00000010}, |
| {0x80007FFF91DAFFFF, 0x7FFF8000FFFF5678, 0x0000001F}, |
| {0x80007FFF91DAFFFF, 0x7FFF8000FFFF5678, 0x0000000F}, |
| {0x80007FFF91DAFFFF, 0x7FFF8000FFFF5678, 0x00000010}, |
| {0x7FFFFFFF80000000, 0x12345678FFFFFFFF, 0x0000001F}, |
| {0x7FFFFFFF80000000, 0x12345678FFFFFFFF, 0x0000000F}, |
| {0x7FFFFFFF80000000, 0x12345678FFFFFFFF, 0x00000010}, |
| {0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0x0000001F}, |
| {0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0x0000000F}, |
| {0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636, 0x00000010}, |
| {0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0x00000015}, |
| {0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0x00000009}, |
| {0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636, 0x00000003}}; |
| |
| #define MAXI_MINI_S_DF(lanes, mask, func) \ |
| [](uint64_t ws, uint32_t ui5) { \ |
| uint64_t res = 0; \ |
| int64_t i5 = ArithmeticShiftRight(static_cast<int64_t>(ui5) << 59, 59); \ |
| int elem_size = kMSARegSize / lanes; \ |
| for (int i = 0; i < lanes / 2; ++i) { \ |
| int shift = elem_size * i; \ |
| int64_t elem = \ |
| static_cast<int64_t>(((ws >> shift) & mask) << (64 - elem_size)) >> \ |
| (64 - elem_size); \ |
| res |= static_cast<uint64_t>(func(elem, i5) & mask) << shift; \ |
| } \ |
| return res; \ |
| } |
| |
| #define MAXI_MINI_U_DF(lanes, mask, func) \ |
| [](uint64_t ws, uint32_t ui5) { \ |
| uint64_t res = 0; \ |
| int elem_size = kMSARegSize / lanes; \ |
| for (int i = 0; i < lanes / 2; ++i) { \ |
| int shift = elem_size * i; \ |
| uint64_t elem = (ws >> shift) & mask; \ |
| res |= (func(elem, static_cast<uint64_t>(ui5)) & mask) << shift; \ |
| } \ |
| return res; \ |
| } |
| |
| for (size_t i = 0; i < sizeof(tc) / sizeof(TestCaseMsaI5); ++i) { |
| run_msa_i5( |
| &tc[i], true, |
| [](MacroAssembler& assm, int32_t i5) { __ maxi_s_b(w2, w0, i5); }, |
| MAXI_MINI_S_DF(kMSALanesByte, UINT8_MAX, Max)); |
| |
| run_msa_i5( |
| &tc[i], true, |
| [](MacroAssembler& assm, int32_t i5) { __ maxi_s_h(w2, w0, i5); }, |
| MAXI_MINI_S_DF(kMSALanesHalf, UINT16_MAX, Max)); |
| |
| run_msa_i5( |
| &tc[i], true, |
| [](MacroAssembler& assm, int32_t i5) { __ maxi_s_w(w2, w0, i5); }, |
| MAXI_MINI_S_DF(kMSALanesWord, UINT32_MAX, Max)); |
| |
| run_msa_i5( |
| &tc[i], true, |
| [](MacroAssembler& assm, int32_t i5) { __ maxi_s_d(w2, w0, i5); }, |
| MAXI_MINI_S_DF(kMSALanesDword, UINT64_MAX, Max)); |
| |
| run_msa_i5( |
| &tc[i], true, |
| [](MacroAssembler& assm, int32_t i5) { __ mini_s_b(w2, w0, i5); }, |
| MAXI_MINI_S_DF(kMSALanesByte, UINT8_MAX, Min)); |
| |
| run_msa_i5( |
| &tc[i], true, |
| [](MacroAssembler& assm, int32_t i5) { __ mini_s_h(w2, w0, i5); }, |
| MAXI_MINI_S_DF(kMSALanesHalf, UINT16_MAX, Min)); |
| |
| run_msa_i5( |
| &tc[i], true, |
| [](MacroAssembler& assm, int32_t i5) { __ mini_s_w(w2, w0, i5); }, |
| MAXI_MINI_S_DF(kMSALanesWord, UINT32_MAX, Min)); |
| |
| run_msa_i5( |
| &tc[i], true, |
| [](MacroAssembler& assm, int32_t i5) { __ mini_s_d(w2, w0, i5); }, |
| MAXI_MINI_S_DF(kMSALanesDword, UINT64_MAX, Min)); |
| |
| run_msa_i5( |
| &tc[i], false, |
| [](MacroAssembler& assm, int32_t i5) { __ maxi_u_b(w2, w0, i5); }, |
| MAXI_MINI_U_DF(kMSALanesByte, UINT8_MAX, Max)); |
| |
| run_msa_i5( |
| &tc[i], false, |
| [](MacroAssembler& assm, int32_t i5) { __ maxi_u_h(w2, w0, i5); }, |
| MAXI_MINI_U_DF(kMSALanesHalf, UINT16_MAX, Max)); |
| |
| run_msa_i5( |
| &tc[i], false, |
| [](MacroAssembler& assm, int32_t i5) { __ maxi_u_w(w2, w0, i5); }, |
| MAXI_MINI_U_DF(kMSALanesWord, UINT32_MAX, Max)); |
| |
| run_msa_i5( |
| &tc[i], false, |
| [](MacroAssembler& assm, int32_t i5) { __ maxi_u_d(w2, w0, i5); }, |
| MAXI_MINI_U_DF(kMSALanesDword, UINT64_MAX, Max)); |
| |
| run_msa_i5( |
| &tc[i], false, |
| [](MacroAssembler& assm, int32_t i5) { __ mini_u_b(w2, w0, i5); }, |
| MAXI_MINI_U_DF(kMSALanesByte, UINT8_MAX, Min)); |
| |
| run_msa_i5( |
| &tc[i], false, |
| [](MacroAssembler& assm, int32_t i5) { __ mini_u_h(w2, w0, i5); }, |
| MAXI_MINI_U_DF(kMSALanesHalf, UINT16_MAX, Min)); |
| |
| run_msa_i5( |
| &tc[i], false, |
| [](MacroAssembler& assm, int32_t i5) { __ mini_u_w(w2, w0, i5); }, |
| MAXI_MINI_U_DF(kMSALanesWord, UINT32_MAX, Min)); |
| |
| run_msa_i5( |
| &tc[i], false, |
| [](MacroAssembler& assm, int32_t i5) { __ mini_u_d(w2, w0, i5); }, |
| MAXI_MINI_U_DF(kMSALanesDword, UINT64_MAX, Min)); |
| } |
| #undef MAXI_MINI_S_DF |
| #undef MAXI_MINI_U_DF |
| } |
| |
| TEST(MSA_ceqi_clti_clei) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| struct TestCaseMsaI5 tc[] = { |
| {0xFF69751BB9A7D9C3, 0xF7A594AEC8FF8A9C, 0x0000001F}, |
| {0xE669FFFFB9A7D9C3, 0xF7A594AEFFFF8A9C, 0x0000001F}, |
| {0xFFFFFFFFB9A7D9C3, 0xF7A594AEFFFFFFFF, 0x0000001F}, |
| {0x2B0B5362C4E812DF, 0x3A0D80D68B3F0BC8, 0x0000000B}, |
| {0x2B66000BC4E812DF, 0x3A0D000B8B3F8BC8, 0x0000000B}, |
| {0x0000000BC4E812DF, 0x3A0D80D60000000B, 0x0000000B}, |
| {0xF38062E13E38F8B0, 0x8041FFDEF2BFE636, 0x00000010}, |
| {0xF35880003E38F8B0, 0x4F41FFDEF2BF8000, 0x00000010}, |
| {0xF35862E180000000, 0x80000000F2BFE636, 0x00000010}, |
| {0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0x00000015}, |
| {0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0x00000009}, |
| {0xF30062E13E38F800, 0x4F00FFDEF2BF0036, 0x00000000}}; |
| |
| #define CEQI_CLTI_CLEI_S_DF(lanes, mask, func) \ |
| [](uint64_t ws, uint32_t ui5) { \ |
| uint64_t res = 0; \ |
| int elem_size = kMSARegSize / lanes; \ |
| int64_t i5 = ArithmeticShiftRight(static_cast<int64_t>(ui5) << 59, 59); \ |
| for (int i = 0; i < lanes / 2; ++i) { \ |
| int shift = elem_size * i; \ |
| int64_t elem = \ |
| static_cast<int64_t>(((ws >> shift) & mask) << (64 - elem_size)) >> \ |
| (64 - elem_size); \ |
| res |= static_cast<uint64_t>((func)&mask) << shift; \ |
| } \ |
| return res; \ |
| } |
| |
| #define CEQI_CLTI_CLEI_U_DF(lanes, mask, func) \ |
| [](uint64_t ws, uint64_t ui5) { \ |
| uint64_t res = 0; \ |
| int elem_size = kMSARegSize / lanes; \ |
| for (int i = 0; i < lanes / 2; ++i) { \ |
| int shift = elem_size * i; \ |
| uint64_t elem = (ws >> shift) & mask; \ |
| res |= ((func)&mask) << shift; \ |
| } \ |
| return res; \ |
| } |
| |
| for (size_t i = 0; i < sizeof(tc) / sizeof(TestCaseMsaI5); ++i) { |
| run_msa_i5(&tc[i], true, |
| [](MacroAssembler& assm, int32_t i5) { __ ceqi_b(w2, w0, i5); }, |
| CEQI_CLTI_CLEI_S_DF(kMSALanesByte, UINT8_MAX, |
| !Compare(elem, i5) ? -1u : 0u)); |
| |
| run_msa_i5(&tc[i], true, |
| [](MacroAssembler& assm, int32_t i5) { __ ceqi_h(w2, w0, i5); }, |
| CEQI_CLTI_CLEI_S_DF(kMSALanesHalf, UINT16_MAX, |
| !Compare(elem, i5) ? -1u : 0u)); |
| |
| run_msa_i5(&tc[i], true, |
| [](MacroAssembler& assm, int32_t i5) { __ ceqi_w(w2, w0, i5); }, |
| CEQI_CLTI_CLEI_S_DF(kMSALanesWord, UINT32_MAX, |
| !Compare(elem, i5) ? -1u : 0u)); |
| |
| run_msa_i5(&tc[i], true, |
| [](MacroAssembler& assm, int32_t i5) { __ ceqi_d(w2, w0, i5); }, |
| CEQI_CLTI_CLEI_S_DF(kMSALanesDword, UINT64_MAX, |
| !Compare(elem, i5) ? -1u : 0u)); |
| |
| run_msa_i5( |
| &tc[i], true, |
| [](MacroAssembler& assm, int32_t i5) { __ clti_s_b(w2, w0, i5); }, |
| CEQI_CLTI_CLEI_S_DF(kMSALanesByte, UINT8_MAX, |
| (Compare(elem, i5) == -1) ? -1u : 0u)); |
| |
| run_msa_i5( |
| &tc[i], true, |
| [](MacroAssembler& assm, int32_t i5) { __ clti_s_h(w2, w0, i5); }, |
| CEQI_CLTI_CLEI_S_DF(kMSALanesHalf, UINT16_MAX, |
| (Compare(elem, i5) == -1) ? -1u : 0u)); |
| |
| run_msa_i5( |
| &tc[i], true, |
| [](MacroAssembler& assm, int32_t i5) { __ clti_s_w(w2, w0, i5); }, |
| CEQI_CLTI_CLEI_S_DF(kMSALanesWord, UINT32_MAX, |
| (Compare(elem, i5) == -1) ? -1u : 0u)); |
| |
| run_msa_i5( |
| &tc[i], true, |
| [](MacroAssembler& assm, int32_t i5) { __ clti_s_d(w2, w0, i5); }, |
| CEQI_CLTI_CLEI_S_DF(kMSALanesDword, UINT64_MAX, |
| (Compare(elem, i5) == -1) ? -1ull : 0ull)); |
| |
| run_msa_i5( |
| &tc[i], true, |
| [](MacroAssembler& assm, int32_t i5) { __ clei_s_b(w2, w0, i5); }, |
| CEQI_CLTI_CLEI_S_DF(kMSALanesByte, UINT8_MAX, |
| (Compare(elem, i5) != 1) ? -1u : 0u)); |
| |
| run_msa_i5( |
| &tc[i], true, |
| [](MacroAssembler& assm, int32_t i5) { __ clei_s_h(w2, w0, i5); }, |
| CEQI_CLTI_CLEI_S_DF(kMSALanesHalf, UINT16_MAX, |
| (Compare(elem, i5) != 1) ? -1u : 0u)); |
| |
| run_msa_i5( |
| &tc[i], true, |
| [](MacroAssembler& assm, int32_t i5) { __ clei_s_w(w2, w0, i5); }, |
| CEQI_CLTI_CLEI_S_DF(kMSALanesWord, UINT32_MAX, |
| (Compare(elem, i5) != 1) ? -1u : 0u)); |
| |
| run_msa_i5( |
| &tc[i], true, |
| [](MacroAssembler& assm, int32_t i5) { __ clei_s_d(w2, w0, i5); }, |
| CEQI_CLTI_CLEI_S_DF(kMSALanesDword, UINT64_MAX, |
| (Compare(elem, i5) != 1) ? -1ull : 0ull)); |
| |
| run_msa_i5( |
| &tc[i], false, |
| [](MacroAssembler& assm, int32_t i5) { __ clti_u_b(w2, w0, i5); }, |
| CEQI_CLTI_CLEI_U_DF(kMSALanesByte, UINT8_MAX, |
| (Compare(elem, ui5) == -1) ? -1ull : 0ull)); |
| |
| run_msa_i5( |
| &tc[i], false, |
| [](MacroAssembler& assm, int32_t i5) { __ clti_u_h(w2, w0, i5); }, |
| CEQI_CLTI_CLEI_U_DF(kMSALanesHalf, UINT16_MAX, |
| (Compare(elem, ui5) == -1) ? -1ull : 0ull)); |
| |
| run_msa_i5( |
| &tc[i], false, |
| [](MacroAssembler& assm, int32_t i5) { __ clti_u_w(w2, w0, i5); }, |
| CEQI_CLTI_CLEI_U_DF(kMSALanesWord, UINT32_MAX, |
| (Compare(elem, ui5) == -1) ? -1ull : 0ull)); |
| |
| run_msa_i5( |
| &tc[i], false, |
| [](MacroAssembler& assm, int32_t i5) { __ clti_u_d(w2, w0, i5); }, |
| CEQI_CLTI_CLEI_U_DF(kMSALanesDword, UINT64_MAX, |
| (Compare(elem, ui5) == -1) ? -1ull : 0ull)); |
| |
| run_msa_i5( |
| &tc[i], false, |
| [](MacroAssembler& assm, int32_t i5) { __ clei_u_b(w2, w0, i5); }, |
| CEQI_CLTI_CLEI_U_DF(kMSALanesByte, UINT8_MAX, |
| (Compare(elem, ui5) != 1) ? -1ull : 0ull)); |
| |
| run_msa_i5( |
| &tc[i], false, |
| [](MacroAssembler& assm, int32_t i5) { __ clei_u_h(w2, w0, i5); }, |
| CEQI_CLTI_CLEI_U_DF(kMSALanesHalf, UINT16_MAX, |
| (Compare(elem, ui5) != 1) ? -1ull : 0ull)); |
| |
| run_msa_i5( |
| &tc[i], false, |
| [](MacroAssembler& assm, int32_t i5) { __ clei_u_w(w2, w0, i5); }, |
| CEQI_CLTI_CLEI_U_DF(kMSALanesWord, UINT32_MAX, |
| (Compare(elem, ui5) != 1) ? -1ull : 0ull)); |
| |
| run_msa_i5( |
| &tc[i], false, |
| [](MacroAssembler& assm, int32_t i5) { __ clei_u_d(w2, w0, i5); }, |
| CEQI_CLTI_CLEI_U_DF(kMSALanesDword, UINT64_MAX, |
| (Compare(elem, ui5) != 1) ? -1ull : 0ull)); |
| } |
| #undef CEQI_CLTI_CLEI_S_DF |
| #undef CEQI_CLTI_CLEI_U_DF |
| } |
| |
| struct TestCaseMsa2R { |
| uint64_t ws_lo; |
| uint64_t ws_hi; |
| uint64_t exp_res_lo; |
| uint64_t exp_res_hi; |
| }; |
| |
| template <typename Func> |
| void run_msa_2r(const struct TestCaseMsa2R* input, |
| Func Generate2RInstructionFunc) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| CpuFeatureScope fscope(&assm, MIPS_SIMD); |
| msa_reg_t res; |
| |
| load_elements_of_vector(assm, reinterpret_cast<const uint64_t*>(input), w0, |
| t0, t1); |
| Generate2RInstructionFunc(assm); |
| store_elements_of_vector(assm, w2, a0); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| #ifdef OBJECT_PRINT |
| code->Print(std::cout); |
| #endif |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| |
| (f.Call(&res, 0, 0, 0, 0)); |
| |
| CHECK_EQ(input->exp_res_lo, res.d[0]); |
| CHECK_EQ(input->exp_res_hi, res.d[1]); |
| } |
| |
| TEST(MSA_pcnt) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| struct TestCaseMsa2R tc_b[] = {// ws_lo, ws_hi, exp_res_lo, exp_res_hi |
| {0x0000000000000000, 0x0000000000000000, 0, 0}, |
| {0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF, |
| 0x0808080808080808, 0x0808080808080808}, |
| {0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, |
| 0x0204050405050504, 0x0704030503070304}, |
| {0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, |
| 0x0404040303040207, 0x0403010504060403}, |
| {0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636, |
| 0x0603030405030503, 0x0502080605070504}}; |
| |
| struct TestCaseMsa2R tc_h[] = {// ws_lo, ws_hi, exp_res_lo, exp_res_hi |
| {0x0000000000000000, 0x0000000000000000, 0, 0}, |
| {0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF, |
| 0x0010001000100010, 0x0010001000100010}, |
| {0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, |
| 0x00060009000A0009, 0x000B0008000A0007}, |
| {0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, |
| 0x0008000700070009, 0x00070006000A0007}, |
| {0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636, |
| 0x0009000700080008, 0x0007000E000C0009}}; |
| |
| struct TestCaseMsa2R tc_w[] = {// ws_lo, ws_hi, exp_res_lo, exp_res_hi |
| {0x0000000000000000, 0x0000000000000000, 0, 0}, |
| {0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF, |
| 0x0000002000000020, 0x0000002000000020}, |
| {0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, |
| 0x0000000F00000013, 0x0000001300000011}, |
| {0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, |
| 0x0000000F00000010, 0x0000000D00000011}, |
| {0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636, |
| 0x0000001000000010, 0x0000001500000015}}; |
| |
| struct TestCaseMsa2R tc_d[] = { |
| // ws_lo, ws_hi, exp_res_lo, exp_res_hi |
| {0x0000000000000000, 0x0000000000000000, 0, 0}, |
| {0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF, 0x40, 0x40}, |
| {0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0x22, 0x24}, |
| {0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0x1F, 0x1E}, |
| {0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636, 0x20, 0x2A}}; |
| |
| for (size_t i = 0; i < sizeof(tc_b) / sizeof(TestCaseMsa2R); ++i) { |
| run_msa_2r(&tc_b[i], [](MacroAssembler& assm) { __ pcnt_b(w2, w0); }); |
| run_msa_2r(&tc_h[i], [](MacroAssembler& assm) { __ pcnt_h(w2, w0); }); |
| run_msa_2r(&tc_w[i], [](MacroAssembler& assm) { __ pcnt_w(w2, w0); }); |
| run_msa_2r(&tc_d[i], [](MacroAssembler& assm) { __ pcnt_d(w2, w0); }); |
| } |
| } |
| |
| TEST(MSA_nlzc) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| struct TestCaseMsa2R tc_b[] = {// ws_lo, ws_hi, exp_res_lo, exp_res_hi |
| {0x0000000000000000, 0x0000000000000000, |
| 0x0808080808080808, 0x0808080808080808}, |
| {0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF, 0, 0}, |
| {0x1169350B07030100, 0x7F011402381F0A6C, |
| 0x0301020405060708, 0x0107030602030401}, |
| {0x010806003478121F, 0x03013016073F7B08, |
| 0x0704050802010303, 0x0607020305020104}, |
| {0x0168321100083803, 0x07113F03013F1676, |
| 0x0701020308040206, 0x0503020607020301}}; |
| |
| struct TestCaseMsa2R tc_h[] = {// ws_lo, ws_hi, exp_res_lo, exp_res_hi |
| {0x0000000000000000, 0x0000000000000000, |
| 0x0010001000100010, 0x0010001000100010}, |
| {0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF, 0, 0}, |
| {0x00010007000A003C, 0x37A5001E00010002, |
| 0x000F000D000C000A, 0x0002000B000F000E}, |
| {0x0026066200780EDF, 0x003D0003000F00C8, |
| 0x000A000500090004, 0x000A000E000C0008}, |
| {0x335807E100480030, 0x01410FDE12BF5636, |
| 0x000200050009000A, 0x0007000400030001}}; |
| |
| struct TestCaseMsa2R tc_w[] = {// ws_lo, ws_hi, exp_res_lo, exp_res_hi |
| {0x0000000000000000, 0x0000000000000000, |
| 0x0000002000000020, 0x0000002000000020}, |
| {0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF, 0, 0}, |
| {0x00000005000007C3, 0x000014AE00006A9C, |
| 0x0000001D00000015, 0x0000001300000011}, |
| {0x00009362000112DF, 0x000380D6003F8BC8, |
| 0x000000100000000F, 0x0000000E0000000A}, |
| {0x135862E17E38F8B0, 0x0061FFDE03BFE636, |
| 0x0000000300000001, 0x0000000900000006}}; |
| |
| struct TestCaseMsa2R tc_d[] = { |
| // ws_lo, ws_hi, exp_res_lo, exp_res_hi |
| {0x0000000000000000, 0x0000000000000000, 0x40, 0x40}, |
| {0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF, 0, 0}, |
| {0x000000000000014E, 0x00000000000176DA, 0x37, 0x2F}, |
| {0x00000062C4E812DF, 0x000065D68B3F8BC8, 0x19, 0x11}, |
| {0x00000000E338F8B0, 0x0754534ACAB32654, 0x20, 0x5}}; |
| |
| for (size_t i = 0; i < sizeof(tc_b) / sizeof(TestCaseMsa2R); ++i) { |
| run_msa_2r(&tc_b[i], [](MacroAssembler& assm) { __ nlzc_b(w2, w0); }); |
| run_msa_2r(&tc_h[i], [](MacroAssembler& assm) { __ nlzc_h(w2, w0); }); |
| run_msa_2r(&tc_w[i], [](MacroAssembler& assm) { __ nlzc_w(w2, w0); }); |
| run_msa_2r(&tc_d[i], [](MacroAssembler& assm) { __ nlzc_d(w2, w0); }); |
| } |
| } |
| |
| TEST(MSA_nloc) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| struct TestCaseMsa2R tc_b[] = {// ws_lo, ws_hi, exp_res_lo, exp_res_hi |
| {0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF, |
| 0x0808080808080808, 0x0808080808080808}, |
| {0x0000000000000000, 0x0000000000000000, 0, 0}, |
| {0xEE96CAF4F8FCFEFF, 0x80FEEBFDC7E0F593, |
| 0x0301020405060708, 0x0107030602030401}, |
| {0xFEF7F9FFCB87EDE0, 0xFCFECFE9F8C084F7, |
| 0x0704050802010303, 0x0607020305020104}, |
| {0xFE97CDEEFFF7C7FC, 0xF8EEC0FCFEC0E989, |
| 0x0701020308040206, 0x0503020607020301}}; |
| |
| struct TestCaseMsa2R tc_h[] = {// ws_lo, ws_hi, exp_res_lo, exp_res_hi |
| {0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF, |
| 0x0010001000100010, 0x0010001000100010}, |
| {0x0000000000000000, 0x0000000000000000, 0, 0}, |
| {0xFFFEFFF8FFF5FFC3, 0xC85AFFE1FFFEFFFD, |
| 0x000F000D000C000A, 0x0002000B000F000E}, |
| {0xFFD9F99DFF87F120, 0xFFC2FFFCFFF0FF37, |
| 0x000A000500090004, 0x000A000E000C0008}, |
| {0xCCA7F81EFFB7FFCF, 0xFEBEF021ED40A9C9, |
| 0x000200050009000A, 0x0007000400030001}}; |
| |
| struct TestCaseMsa2R tc_w[] = {// ws_lo, ws_hi, exp_res_lo, exp_res_hi |
| {0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF, |
| 0x0000002000000020, 0x0000002000000020}, |
| {0x0000000000000000, 0x0000000000000000, 0, 0}, |
| {0xFFFFFFFAFFFFF83C, 0xFFFFEB51FFFF9563, |
| 0x0000001D00000015, 0x0000001300000011}, |
| {0xFFFF6C9DFFFEED20, 0xFFFC7F29FFC07437, |
| 0x000000100000000F, 0x0000000E0000000A}, |
| {0xECA79D1E81C7074F, 0xFF9E0021FC4019C9, |
| 0x0000000300000001, 0x0000000900000006}}; |
| |
| struct TestCaseMsa2R tc_d[] = { |
| // ws_lo, ws_hi, exp_res_lo, exp_res_hi |
| {0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF, 0x40, 0x40}, |
| {0x0000000000000000, 0x0000000000000000, 0, 0}, |
| {0xFFFFFFFFFFFFFEB1, 0xFFFFFFFFFFFE8925, 0x37, 0x2F}, |
| {0xFFFFFF9D3B17ED20, 0xFFFF9A2974C07437, 0x19, 0x11}, |
| {0xFFFFFFFF1CC7074F, 0xF8ABACB5354CD9AB, 0x20, 0x5}}; |
| |
| for (size_t i = 0; i < sizeof(tc_b) / sizeof(TestCaseMsa2R); ++i) { |
| run_msa_2r(&tc_b[i], [](MacroAssembler& assm) { __ nloc_b(w2, w0); }); |
| run_msa_2r(&tc_h[i], [](MacroAssembler& assm) { __ nloc_h(w2, w0); }); |
| run_msa_2r(&tc_w[i], [](MacroAssembler& assm) { __ nloc_w(w2, w0); }); |
| run_msa_2r(&tc_d[i], [](MacroAssembler& assm) { __ nloc_d(w2, w0); }); |
| } |
| } |
| |
| struct TestCaseMsa2RF_F_U { |
| float ws1; |
| float ws2; |
| float ws3; |
| float ws4; |
| uint32_t exp_res_1; |
| uint32_t exp_res_2; |
| uint32_t exp_res_3; |
| uint32_t exp_res_4; |
| }; |
| |
| struct TestCaseMsa2RF_D_U { |
| double ws1; |
| double ws2; |
| uint64_t exp_res_1; |
| uint64_t exp_res_2; |
| }; |
| |
| TEST(MSA_fclass) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| #define BIT(n) (0x1 << n) |
| #define SNAN_BIT BIT(0) |
| #define QNAN_BIT BIT(1) |
| #define NEG_INFINITY_BIT BIT((2)) |
| #define NEG_NORMAL_BIT BIT(3) |
| #define NEG_SUBNORMAL_BIT BIT(4) |
| #define NEG_ZERO_BIT BIT(5) |
| #define POS_INFINITY_BIT BIT(6) |
| #define POS_NORMAL_BIT BIT(7) |
| #define POS_SUBNORMAL_BIT BIT(8) |
| #define POS_ZERO_BIT BIT(9) |
| |
| const float inf_float = std::numeric_limits<float>::infinity(); |
| const double inf_double = std::numeric_limits<double>::infinity(); |
| |
| const struct TestCaseMsa2RF_F_U tc_s[] = { |
| {1.f, -0.00001, 208e10f, -34.8e-30f, POS_NORMAL_BIT, NEG_NORMAL_BIT, |
| POS_NORMAL_BIT, NEG_NORMAL_BIT}, |
| {inf_float, -inf_float, 0, -0.f, POS_INFINITY_BIT, NEG_INFINITY_BIT, |
| POS_ZERO_BIT, NEG_ZERO_BIT}, |
| {3.036e-40f, -6.392e-43f, 1.41e-45f, -1.17e-38f, POS_SUBNORMAL_BIT, |
| NEG_SUBNORMAL_BIT, POS_SUBNORMAL_BIT, NEG_SUBNORMAL_BIT}}; |
| |
| const struct TestCaseMsa2RF_D_U tc_d[] = { |
| {1., -0.00000001, POS_NORMAL_BIT, NEG_NORMAL_BIT}, |
| {208e10, -34.8e-300, POS_NORMAL_BIT, NEG_NORMAL_BIT}, |
| {inf_double, -inf_double, POS_INFINITY_BIT, NEG_INFINITY_BIT}, |
| {0, -0., POS_ZERO_BIT, NEG_ZERO_BIT}, |
| {1.036e-308, -6.392e-309, POS_SUBNORMAL_BIT, NEG_SUBNORMAL_BIT}, |
| {1.41e-323, -3.17e208, POS_SUBNORMAL_BIT, NEG_NORMAL_BIT}}; |
| |
| for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_F_U); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]), |
| [](MacroAssembler& assm) { __ fclass_w(w2, w0); }); |
| } |
| for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_D_U); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]), |
| [](MacroAssembler& assm) { __ fclass_d(w2, w0); }); |
| } |
| |
| #undef BIT |
| #undef SNAN_BIT |
| #undef QNAN_BIT |
| #undef NEG_INFINITY_BIT |
| #undef NEG_NORMAL_BIT |
| #undef NEG_SUBNORMAL_BIT |
| #undef NEG_ZERO_BIT |
| #undef POS_INFINITY_BIT |
| #undef POS_NORMAL_BIT |
| #undef POS_SUBNORMAL_BIT |
| #undef POS_ZERO_BIT |
| } |
| |
| struct TestCaseMsa2RF_F_I { |
| float ws1; |
| float ws2; |
| float ws3; |
| float ws4; |
| int32_t exp_res_1; |
| int32_t exp_res_2; |
| int32_t exp_res_3; |
| int32_t exp_res_4; |
| }; |
| |
| struct TestCaseMsa2RF_D_I { |
| double ws1; |
| double ws2; |
| int64_t exp_res_1; |
| int64_t exp_res_2; |
| }; |
| |
| TEST(MSA_ftrunc_s) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| const float inf_float = std::numeric_limits<float>::infinity(); |
| const float qNaN_float = std::numeric_limits<float>::quiet_NaN(); |
| const double inf_double = std::numeric_limits<double>::infinity(); |
| const double qNaN_double = std::numeric_limits<double>::quiet_NaN(); |
| const int32_t max_int32 = std::numeric_limits<int32_t>::max(); |
| const int32_t min_int32 = std::numeric_limits<int32_t>::min(); |
| const int64_t max_int64 = std::numeric_limits<int64_t>::max(); |
| const int64_t min_int64 = std::numeric_limits<int64_t>::min(); |
| |
| const struct TestCaseMsa2RF_F_I tc_s[] = { |
| {inf_float, 2.345f, -324.9235f, 30004.51f, max_int32, 2, -324, 30004}, |
| {-inf_float, -0.983f, 0.0832f, static_cast<float>(max_int32) * 3.f, |
| min_int32, 0, 0, max_int32}, |
| {-23.125f, qNaN_float, 2 * static_cast<float>(min_int32), -0.f, -23, 0, |
| min_int32, 0}}; |
| |
| const struct TestCaseMsa2RF_D_I tc_d[] = { |
| {inf_double, 2.345, max_int64, 2}, |
| {-324.9235, 246569139.51, -324, 246569139}, |
| {-inf_double, -0.983, min_int64, 0}, |
| {0.0832, 6 * static_cast<double>(max_int64), 0, max_int64}, |
| {-21453889872.94, qNaN_double, -21453889872, 0}, |
| {2 * static_cast<double>(min_int64), -0., min_int64, 0}}; |
| |
| for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_F_I); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]), |
| [](MacroAssembler& assm) { __ ftrunc_s_w(w2, w0); }); |
| } |
| for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_D_I); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]), |
| [](MacroAssembler& assm) { __ ftrunc_s_d(w2, w0); }); |
| } |
| } |
| |
| TEST(MSA_ftrunc_u) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| const float inf_float = std::numeric_limits<float>::infinity(); |
| const float qNaN_float = std::numeric_limits<float>::quiet_NaN(); |
| const double inf_double = std::numeric_limits<double>::infinity(); |
| const double qNaN_double = std::numeric_limits<double>::quiet_NaN(); |
| const uint32_t max_uint32 = std::numeric_limits<uint32_t>::max(); |
| const uint64_t max_uint64 = std::numeric_limits<uint64_t>::max(); |
| |
| const struct TestCaseMsa2RF_F_U tc_s[] = { |
| {inf_float, 2.345f, -324.9235f, 30004.51f, max_uint32, 2, 0, 30004}, |
| {-inf_float, 0.983f, 0.0832f, static_cast<float>(max_uint32) * 3., 0, 0, |
| 0, max_uint32}, |
| {23.125f, qNaN_float, -0.982, -0.f, 23, 0, 0, 0}}; |
| |
| const struct TestCaseMsa2RF_D_U tc_d[] = { |
| {inf_double, 2.345, max_uint64, 2}, |
| {-324.9235, 246569139.51, 0, 246569139}, |
| {-inf_double, -0.983, 0, 0}, |
| {0.0832, 6 * static_cast<double>(max_uint64), 0, max_uint64}, |
| {21453889872.94, qNaN_double, 21453889872, 0}, |
| {0.9889, -0., 0, 0}}; |
| |
| for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_F_U); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]), |
| [](MacroAssembler& assm) { __ ftrunc_u_w(w2, w0); }); |
| } |
| for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_D_U); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]), |
| [](MacroAssembler& assm) { __ ftrunc_u_d(w2, w0); }); |
| } |
| } |
| |
| struct TestCaseMsa2RF_F_F { |
| float ws1; |
| float ws2; |
| float ws3; |
| float ws4; |
| float exp_res_1; |
| float exp_res_2; |
| float exp_res_3; |
| float exp_res_4; |
| }; |
| |
| struct TestCaseMsa2RF_D_D { |
| double ws1; |
| double ws2; |
| double exp_res_1; |
| double exp_res_2; |
| }; |
| |
| TEST(MSA_fsqrt) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| const float inf_float = std::numeric_limits<float>::infinity(); |
| const double inf_double = std::numeric_limits<double>::infinity(); |
| |
| const struct TestCaseMsa2RF_F_F tc_s[] = { |
| {81.f, 576.f, inf_float, -0.f, 9.f, 24.f, inf_float, -0.f}}; |
| |
| const struct TestCaseMsa2RF_D_D tc_d[] = {{81., inf_double, 9., inf_double}, |
| {331776., -0., 576, -0.}}; |
| |
| for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_F_F); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]), |
| [](MacroAssembler& assm) { __ fsqrt_w(w2, w0); }); |
| } |
| for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_D_D); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]), |
| [](MacroAssembler& assm) { __ fsqrt_d(w2, w0); }); |
| } |
| } |
| |
| TEST(MSA_frsqrt) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| const float inf_float = std::numeric_limits<float>::infinity(); |
| const double inf_double = std::numeric_limits<double>::infinity(); |
| |
| const struct TestCaseMsa2RF_F_F tc_s[] = { |
| {81.f, 576.f, inf_float, -0.f, 1.f / 9.f, 1.f / 24.f, 0.f, -inf_float}, |
| {0.f, 1.f / 576.f, 1.f / 81.f, 1.f / 4.f, inf_float, 24.f, 9.f, 2.f}}; |
| |
| const struct TestCaseMsa2RF_D_D tc_d[] = { |
| {81., inf_double, 1. / 9., 0.}, |
| {331776., -0., 1. / 576., -inf_double}, |
| {0., 1. / 81, inf_double, 9.}}; |
| |
| for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_F_F); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]), |
| [](MacroAssembler& assm) { __ frsqrt_w(w2, w0); }); |
| } |
| for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_D_D); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]), |
| [](MacroAssembler& assm) { __ frsqrt_d(w2, w0); }); |
| } |
| } |
| |
| TEST(MSA_frcp) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| const float inf_float = std::numeric_limits<float>::infinity(); |
| const double inf_double = std::numeric_limits<double>::infinity(); |
| |
| const struct TestCaseMsa2RF_F_F tc_s[] = { |
| {12.f, 576.f, inf_float, -0.f, 1.f / 12.f, 1.f / 576.f, 0.f, -inf_float}, |
| {0.f, 1.f / 576.f, -inf_float, 1.f / 400.f, inf_float, 576.f, -0.f, |
| 400.f}}; |
| |
| const struct TestCaseMsa2RF_D_D tc_d[] = { |
| {81., inf_double, 1. / 81., 0.}, |
| {331777., -0., 1. / 331777., -inf_double}, |
| {0., 1. / 80, inf_double, 80.}, |
| {1. / 40000., -inf_double, 40000., -0.}}; |
| |
| for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_F_F); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]), |
| [](MacroAssembler& assm) { __ frcp_w(w2, w0); }); |
| } |
| for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_D_D); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]), |
| [](MacroAssembler& assm) { __ frcp_d(w2, w0); }); |
| } |
| } |
| |
| void test_frint_s(size_t data_size, TestCaseMsa2RF_F_F tc_d[], |
| int rounding_mode) { |
| for (size_t i = 0; i < data_size / sizeof(TestCaseMsa2RF_F_F); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]), |
| [&rounding_mode](MacroAssembler& assm) { |
| MSAControlRegister msareg = {kMSACSRRegister}; |
| __ li(t0, static_cast<uint32_t>(rounding_mode)); |
| __ cfcmsa(t1, msareg); |
| __ ctcmsa(msareg, t0); |
| __ frint_w(w2, w0); |
| __ ctcmsa(msareg, t1); |
| }); |
| } |
| } |
| |
| void test_frint_d(size_t data_size, TestCaseMsa2RF_D_D tc_d[], |
| int rounding_mode) { |
| for (size_t i = 0; i < data_size / sizeof(TestCaseMsa2RF_D_D); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]), |
| [&rounding_mode](MacroAssembler& assm) { |
| MSAControlRegister msareg = {kMSACSRRegister}; |
| __ li(t0, static_cast<uint32_t>(rounding_mode)); |
| __ cfcmsa(t1, msareg); |
| __ ctcmsa(msareg, t0); |
| __ frint_d(w2, w0); |
| __ ctcmsa(msareg, t1); |
| }); |
| } |
| } |
| |
| TEST(MSA_frint) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| struct TestCaseMsa2RF_F_F tc_s1[] = { |
| {0.f, 4.51f, 1.49f, -12.51f, 0.f, 5.f, 1.f, -13.f}, |
| {-1.32f, -23.38f, 2.8f, -32.5f, -1.f, -23.f, 3.f, -32.f}}; |
| |
| struct TestCaseMsa2RF_D_D tc_d1[] = {{0., 4.51, 0., 5.}, |
| {1.49, -12.51, 1., -13.}, |
| {-1.32, -23.38, -1., -23.}, |
| {2.8, -32.6, 3., -33.}}; |
| |
| test_frint_s(sizeof(tc_s1), tc_s1, kRoundToNearest); |
| test_frint_d(sizeof(tc_d1), tc_d1, kRoundToNearest); |
| |
| struct TestCaseMsa2RF_F_F tc_s2[] = { |
| {0.f, 4.5f, 1.49f, -12.51f, 0.f, 4.f, 1.f, -12.f}, |
| {-1.f, -23.38f, 2.8f, -32.6f, -1.f, -23.f, 2.f, -32.f}}; |
| |
| struct TestCaseMsa2RF_D_D tc_d2[] = {{0., 4.5, 0., 4.}, |
| {1.49, -12.51, 1., -12.}, |
| {-1., -23.38, -1., -23.}, |
| {2.8, -32.6, 2., -32.}}; |
| |
| test_frint_s(sizeof(tc_s2), tc_s2, kRoundToZero); |
| test_frint_d(sizeof(tc_d2), tc_d2, kRoundToZero); |
| |
| struct TestCaseMsa2RF_F_F tc_s3[] = { |
| {0.f, 4.5f, 1.49f, -12.51f, 0.f, 5.f, 2.f, -12.f}, |
| {-1.f, -23.38f, 2.8f, -32.6f, -1.f, -23.f, 3.f, -32.f}}; |
| |
| struct TestCaseMsa2RF_D_D tc_d3[] = {{0., 4.5, 0., 5.}, |
| {1.49, -12.51, 2., -12.}, |
| {-1., -23.38, -1., -23.}, |
| {2.8, -32.6, 3., -32.}}; |
| |
| test_frint_s(sizeof(tc_s3), tc_s3, kRoundToPlusInf); |
| test_frint_d(sizeof(tc_d3), tc_d3, kRoundToPlusInf); |
| |
| struct TestCaseMsa2RF_F_F tc_s4[] = { |
| {0.f, 4.5f, 1.49f, -12.51f, 0.f, 4.f, 1.f, -13.f}, |
| {-1.f, -23.38f, 2.8f, -32.6f, -1.f, -24.f, 2.f, -33.f}}; |
| |
| struct TestCaseMsa2RF_D_D tc_d4[] = {{0., 4.5, 0., 4.}, |
| {1.49, -12.51, 1., -13.}, |
| {-1., -23.38, -1., -24.}, |
| {2.8, -32.6, 2., -33.}}; |
| |
| test_frint_s(sizeof(tc_s4), tc_s4, kRoundToMinusInf); |
| test_frint_d(sizeof(tc_d4), tc_d4, kRoundToMinusInf); |
| } |
| |
| TEST(MSA_flog2) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| const float inf_float = std::numeric_limits<float>::infinity(); |
| const double inf_double = std::numeric_limits<double>::infinity(); |
| |
| struct TestCaseMsa2RF_F_F tc_s[] = { |
| {std::ldexp(0.58f, -48), std::ldexp(0.5f, 110), std::ldexp(1.11f, -130), |
| inf_float, -49.f, 109.f, -130.f, inf_float}, |
| {0.f, -0.f, std::ldexp(0.89f, -12), std::ldexp(0.32f, 126), -inf_float, |
| -inf_float, -13.f, 124.f}}; |
| |
| struct TestCaseMsa2RF_D_D tc_d[] = { |
| {std::ldexp(0.58, -48), std::ldexp(0.5, 110), -49., 109.}, |
| {std::ldexp(1.11, -1050), inf_double, -1050., inf_double}, |
| {0., -0., -inf_double, -inf_double}, |
| {std::ldexp(0.32, 1021), std::ldexp(1.23, -123), 1019., -123.}}; |
| |
| for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_F_F); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]), |
| [](MacroAssembler& assm) { __ flog2_w(w2, w0); }); |
| } |
| |
| for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_D_D); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]), |
| [](MacroAssembler& assm) { __ flog2_d(w2, w0); }); |
| } |
| } |
| |
| void test_ftint_s_s(size_t data_size, TestCaseMsa2RF_F_I tc_d[], |
| int rounding_mode) { |
| for (size_t i = 0; i < data_size / sizeof(TestCaseMsa2RF_F_I); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]), |
| [&rounding_mode](MacroAssembler& assm) { |
| MSAControlRegister msareg = {kMSACSRRegister}; |
| __ li(t0, static_cast<uint32_t>(rounding_mode)); |
| __ cfcmsa(t1, msareg); |
| __ ctcmsa(msareg, t0); |
| __ ftint_s_w(w2, w0); |
| __ ctcmsa(msareg, t1); |
| }); |
| } |
| } |
| |
| void test_ftint_s_d(size_t data_size, TestCaseMsa2RF_D_I tc_d[], |
| int rounding_mode) { |
| for (size_t i = 0; i < data_size / sizeof(TestCaseMsa2RF_D_I); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]), |
| [&rounding_mode](MacroAssembler& assm) { |
| MSAControlRegister msareg = {kMSACSRRegister}; |
| __ li(t0, static_cast<uint32_t>(rounding_mode)); |
| __ cfcmsa(t1, msareg); |
| __ ctcmsa(msareg, t0); |
| __ ftint_s_d(w2, w0); |
| __ ctcmsa(msareg, t1); |
| }); |
| } |
| } |
| |
| TEST(MSA_ftint_s) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| const float inf_float = std::numeric_limits<float>::infinity(); |
| const double inf_double = std::numeric_limits<double>::infinity(); |
| const int32_t int32_max = std::numeric_limits<int32_t>::max(); |
| const int32_t int32_min = std::numeric_limits<int32_t>::min(); |
| const int64_t int64_max = std::numeric_limits<int64_t>::max(); |
| const int64_t int64_min = std::numeric_limits<int64_t>::min(); |
| |
| struct TestCaseMsa2RF_F_I tc_s1[] = { |
| {0.f, 4.51f, 1.49f, -12.51f, 0, 5, 1, -13}, |
| {-0.32f, -23.38f, 2.8f, -32.6f, 0, -23, 3, -33}, |
| {inf_float, -inf_float, 3.f * int32_min, 4.f * int32_max, int32_max, |
| int32_min, int32_min, int32_max}}; |
| |
| struct TestCaseMsa2RF_D_I tc_d1[] = { |
| {0., 4.51, 0, 5}, |
| {1.49, -12.51, 1, -13}, |
| {-0.32, -23.38, 0, -23}, |
| {2.8, -32.6, 3, -33}, |
| {inf_double, -inf_double, int64_max, int64_min}, |
| {33.23 * int64_min, 4000. * int64_max, int64_min, int64_max}}; |
| |
| test_ftint_s_s(sizeof(tc_s1), tc_s1, kRoundToNearest); |
| test_ftint_s_d(sizeof(tc_d1), tc_d1, kRoundToNearest); |
| |
| struct TestCaseMsa2RF_F_I tc_s2[] = { |
| {0.f, 4.5f, 1.49f, -12.51f, 0, 4, 1, -12}, |
| {-0.f, -23.38f, 2.8f, -32.6f, -0, -23, 2, -32}, |
| {inf_float, -inf_float, 3.f * int32_min, 4.f * int32_max, int32_max, |
| int32_min, int32_min, int32_max}}; |
| |
| struct TestCaseMsa2RF_D_I tc_d2[] = { |
| {0., 4.5, 0, 4}, |
| {1.49, -12.51, 1, -12}, |
| {-0., -23.38, -0, -23}, |
| {2.8, -32.6, 2, -32}, |
| {inf_double, -inf_double, int64_max, int64_min}, |
| {33.23 * int64_min, 4000. * int64_max, int64_min, int64_max}}; |
| |
| test_ftint_s_s(sizeof(tc_s2), tc_s2, kRoundToZero); |
| test_ftint_s_d(sizeof(tc_d2), tc_d2, kRoundToZero); |
| |
| struct TestCaseMsa2RF_F_I tc_s3[] = { |
| {0.f, 4.5f, 1.49f, -12.51f, 0, 5, 2, -12}, |
| {-0.f, -23.38f, 2.8f, -32.6f, -0, -23, 3, -32}, |
| {inf_float, -inf_float, 3.f * int32_min, 4.f * int32_max, int32_max, |
| int32_min, int32_min, int32_max}}; |
| |
| struct TestCaseMsa2RF_D_I tc_d3[] = { |
| {0., 4.5, 0, 5}, |
| {1.49, -12.51, 2, -12}, |
| {-0., -23.38, -0, -23}, |
| {2.8, -32.6, 3, -32}, |
| {inf_double, -inf_double, int64_max, int64_min}, |
| {33.23 * int64_min, 4000. * int64_max, int64_min, int64_max}}; |
| |
| test_ftint_s_s(sizeof(tc_s3), tc_s3, kRoundToPlusInf); |
| test_ftint_s_d(sizeof(tc_d3), tc_d3, kRoundToPlusInf); |
| |
| struct TestCaseMsa2RF_F_I tc_s4[] = { |
| {0.f, 4.5f, 1.49f, -12.51f, 0, 4, 1, -13}, |
| {-0.f, -23.38f, 2.8f, -32.6f, -0, -24, 2, -33}, |
| {inf_float, -inf_float, 3.f * int32_min, 4.f * int32_max, int32_max, |
| int32_min, int32_min, int32_max}}; |
| |
| struct TestCaseMsa2RF_D_I tc_d4[] = { |
| {0., 4.5, 0, 4}, |
| {1.49, -12.51, 1, -13}, |
| {-0., -23.38, -0, -24}, |
| {2.8, -32.6, 2, -33}, |
| {inf_double, -inf_double, int64_max, int64_min}, |
| {33.23 * int64_min, 4000. * int64_max, int64_min, int64_max}}; |
| |
| test_ftint_s_s(sizeof(tc_s4), tc_s4, kRoundToMinusInf); |
| test_ftint_s_d(sizeof(tc_d4), tc_d4, kRoundToMinusInf); |
| } |
| |
| void test_ftint_u_s(size_t data_size, TestCaseMsa2RF_F_U tc_d[], |
| int rounding_mode) { |
| for (size_t i = 0; i < data_size / sizeof(TestCaseMsa2RF_F_U); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]), |
| [&rounding_mode](MacroAssembler& assm) { |
| MSAControlRegister msareg = {kMSACSRRegister}; |
| __ li(t0, static_cast<uint32_t>(rounding_mode)); |
| __ cfcmsa(t1, msareg); |
| __ ctcmsa(msareg, t0); |
| __ ftint_u_w(w2, w0); |
| __ ctcmsa(msareg, t1); |
| }); |
| } |
| } |
| |
| void test_ftint_u_d(size_t data_size, TestCaseMsa2RF_D_U tc_d[], |
| int rounding_mode) { |
| for (size_t i = 0; i < data_size / sizeof(TestCaseMsa2RF_D_U); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]), |
| [&rounding_mode](MacroAssembler& assm) { |
| MSAControlRegister msareg = {kMSACSRRegister}; |
| __ li(t0, static_cast<uint32_t>(rounding_mode)); |
| __ cfcmsa(t1, msareg); |
| __ ctcmsa(msareg, t0); |
| __ ftint_u_d(w2, w0); |
| __ ctcmsa(msareg, t1); |
| }); |
| } |
| } |
| |
| TEST(MSA_ftint_u) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| const float inf_float = std::numeric_limits<float>::infinity(); |
| const double inf_double = std::numeric_limits<double>::infinity(); |
| const uint32_t uint32_max = std::numeric_limits<uint32_t>::max(); |
| const uint64_t uint64_max = std::numeric_limits<uint64_t>::max(); |
| |
| struct TestCaseMsa2RF_F_U tc_s1[] = { |
| {0.f, 4.51f, 1.49f, -12.51f, 0, 5, 1, 0}, |
| {-0.32f, 23.38f, 2.8f, 32.6f, 0, 23, 3, 33}, |
| {inf_float, -inf_float, 0, 4.f * uint32_max, uint32_max, 0, 0, |
| uint32_max}}; |
| |
| struct TestCaseMsa2RF_D_U tc_d1[] = { |
| {0., 4.51, 0, 5}, |
| {1.49, -12.51, 1, 0}, |
| {-0.32, 23.38, 0, 23}, |
| {2.8, 32.6, 3, 33}, |
| {inf_double, -inf_double, uint64_max, 0}, |
| {-0., 4000. * uint64_max, 0, uint64_max}}; |
| |
| test_ftint_u_s(sizeof(tc_s1), tc_s1, kRoundToNearest); |
| test_ftint_u_d(sizeof(tc_d1), tc_d1, kRoundToNearest); |
| |
| struct TestCaseMsa2RF_F_U tc_s2[] = { |
| {0.f, 4.5f, 1.49f, -12.51f, 0, 4, 1, 0}, |
| {-0.f, 23.38f, 2.8f, 32.6f, 0, 23, 2, 32}, |
| {inf_float, -inf_float, 0., 4.f * uint32_max, uint32_max, 0, 0, |
| uint32_max}}; |
| |
| struct TestCaseMsa2RF_D_U tc_d2[] = { |
| {0., 4.5, 0, 4}, |
| {1.49, -12.51, 1, 0}, |
| {-0., 23.38, 0, 23}, |
| {2.8, 32.6, 2, 32}, |
| {inf_double, -inf_double, uint64_max, 0}, |
| {-0.2345, 4000. * uint64_max, 0, uint64_max}}; |
| |
| test_ftint_u_s(sizeof(tc_s2), tc_s2, kRoundToZero); |
| test_ftint_u_d(sizeof(tc_d2), tc_d2, kRoundToZero); |
| |
| struct TestCaseMsa2RF_F_U tc_s3[] = { |
| {0.f, 4.5f, 1.49f, -12.51f, 0, 5, 2, 0}, |
| {-0.f, 23.38f, 2.8f, 32.6f, 0, 24, 3, 33}, |
| {inf_float, -inf_float, 0, 4.f * uint32_max, uint32_max, 0, 0, |
| uint32_max}}; |
| |
| struct TestCaseMsa2RF_D_U tc_d3[] = { |
| {0., 4.5, 0, 5}, |
| {1.49, -12.51, 2, 0}, |
| {-0., 23.38, -0, 24}, |
| {2.8, 32.6, 3, 33}, |
| {inf_double, -inf_double, uint64_max, 0}, |
| {-0.5252, 4000. * uint64_max, 0, uint64_max}}; |
| |
| test_ftint_u_s(sizeof(tc_s3), tc_s3, kRoundToPlusInf); |
| test_ftint_u_d(sizeof(tc_d3), tc_d3, kRoundToPlusInf); |
| |
| struct TestCaseMsa2RF_F_U tc_s4[] = { |
| {0.f, 4.5f, 1.49f, -12.51f, 0, 4, 1, 0}, |
| {-0.f, 23.38f, 2.8f, 32.6f, 0, 23, 2, 32}, |
| {inf_float, -inf_float, 0, 4.f * uint32_max, uint32_max, 0, 0, |
| uint32_max}}; |
| |
| struct TestCaseMsa2RF_D_U tc_d4[] = { |
| {0., 4.5, 0, 4}, |
| {1.49, -12.51, 1, 0}, |
| {-0., 23.38, -0, 23}, |
| {2.8, 32.6, 2, 32}, |
| {inf_double, -inf_double, uint64_max, 0}, |
| {-0.098797, 4000. * uint64_max, 0, uint64_max}}; |
| |
| test_ftint_u_s(sizeof(tc_s4), tc_s4, kRoundToMinusInf); |
| test_ftint_u_d(sizeof(tc_d4), tc_d4, kRoundToMinusInf); |
| } |
| |
| struct TestCaseMsa2RF_U_F { |
| uint32_t ws1; |
| uint32_t ws2; |
| uint32_t ws3; |
| uint32_t ws4; |
| float exp_res_1; |
| float exp_res_2; |
| float exp_res_3; |
| float exp_res_4; |
| }; |
| |
| struct TestCaseMsa2RF_U_D { |
| uint64_t ws1; |
| uint64_t ws2; |
| double exp_res_1; |
| double exp_res_2; |
| }; |
| |
| TEST(MSA_ffint_u) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| struct TestCaseMsa2RF_U_F tc_s[] = { |
| {0, 345, 234, 1000, 0.f, 345.f, 234.f, 1000.f}}; |
| |
| struct TestCaseMsa2RF_U_D tc_d[] = {{0, 345, 0., 345.}, |
| {234, 1000, 234., 1000.}}; |
| |
| for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_U_F); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]), |
| [](MacroAssembler& assm) { __ ffint_u_w(w2, w0); }); |
| } |
| for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_U_D); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]), |
| [](MacroAssembler& assm) { __ ffint_u_d(w2, w0); }); |
| } |
| } |
| |
| struct TestCaseMsa2RF_I_F { |
| int32_t ws1; |
| int32_t ws2; |
| int32_t ws3; |
| int32_t ws4; |
| float exp_res_1; |
| float exp_res_2; |
| float exp_res_3; |
| float exp_res_4; |
| }; |
| |
| struct TestCaseMsa2RF_I_D { |
| int64_t ws1; |
| int64_t ws2; |
| double exp_res_1; |
| double exp_res_2; |
| }; |
| |
| TEST(MSA_ffint_s) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| struct TestCaseMsa2RF_I_F tc_s[] = { |
| {0, 345, -234, 1000, 0.f, 345.f, -234.f, 1000.f}}; |
| |
| struct TestCaseMsa2RF_I_D tc_d[] = {{0, 345, 0., 345.}, |
| {-234, 1000, -234., 1000.}}; |
| |
| for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_I_F); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]), |
| [](MacroAssembler& assm) { __ ffint_s_w(w2, w0); }); |
| } |
| for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_I_D); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]), |
| [](MacroAssembler& assm) { __ ffint_s_d(w2, w0); }); |
| } |
| } |
| |
| struct TestCaseMsa2RF_U16_F { |
| uint16_t ws1; |
| uint16_t ws2; |
| uint16_t ws3; |
| uint16_t ws4; |
| uint16_t ws5; |
| uint16_t ws6; |
| uint16_t ws7; |
| uint16_t ws8; |
| float exp_res_1; |
| float exp_res_2; |
| float exp_res_3; |
| float exp_res_4; |
| }; |
| |
| struct TestCaseMsa2RF_F_D { |
| float ws1; |
| float ws2; |
| float ws3; |
| float ws4; |
| double exp_res_1; |
| double exp_res_2; |
| }; |
| |
| TEST(MSA_fexupl) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| const float inf_float = std::numeric_limits<float>::infinity(); |
| const double inf_double = std::numeric_limits<double>::infinity(); |
| |
| struct TestCaseMsa2RF_U16_F tc_s[] = { |
| {1, 2, 0x7C00, 0x0C00, 0, 0x7C00, 0xFC00, 0x8000, 0.f, inf_float, |
| -inf_float, -0.f}, |
| {0xFC00, 0xFFFF, 0x00FF, 0x8000, 0x81FE, 0x8000, 0x0345, 0xAAAA, |
| -3.0398368835e-5f, -0.f, 4.9889088e-5f, -5.2062988281e-2f}, |
| {3, 4, 0x5555, 6, 0x2AAA, 0x8700, 0x7777, 0x6A8B, 5.2062988281e-2f, |
| -1.06811523458e-4f, 3.0576e4f, 3.35e3f}}; |
| |
| struct TestCaseMsa2RF_F_D tc_d[] = { |
| {0.f, 123.456f, inf_float, -0.f, inf_double, -0.}, |
| {-inf_float, -3.f, 0.f, -inf_float, 0., -inf_double}, |
| {2.3f, 3., 1.37747639043129518071e-41f, -3.22084585277826e35f, |
| 1.37747639043129518071e-41, -3.22084585277826e35}}; |
| |
| for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_U16_F); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]), |
| [](MacroAssembler& assm) { __ fexupl_w(w2, w0); }); |
| } |
| for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_F_D); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]), |
| [](MacroAssembler& assm) { __ fexupl_d(w2, w0); }); |
| } |
| } |
| |
| TEST(MSA_fexupr) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| const float inf_float = std::numeric_limits<float>::infinity(); |
| const double inf_double = std::numeric_limits<double>::infinity(); |
| |
| struct TestCaseMsa2RF_U16_F tc_s[] = { |
| {0, 0x7C00, 0xFC00, 0x8000, 1, 2, 0x7C00, 0x0C00, 0.f, inf_float, |
| -inf_float, -0.f}, |
| {0x81FE, 0x8000, 0x0345, 0xAAAA, 0xFC00, 0xFFFF, 0x00FF, 0x8000, |
| -3.0398368835e-5f, -0.f, 4.9889088e-5f, -5.2062988281e-2f}, |
| {0x2AAA, 0x8700, 0x7777, 0x6A8B, 3, 4, 0x5555, 6, 5.2062988281e-2f, |
| -1.06811523458e-4f, 3.0576e4f, 3.35e3f}}; |
| |
| struct TestCaseMsa2RF_F_D tc_d[] = { |
| {inf_float, -0.f, 0.f, 123.456f, inf_double, -0.}, |
| {0.f, -inf_float, -inf_float, -3.f, 0., -inf_double}, |
| {1.37747639043129518071e-41f, -3.22084585277826e35f, 2.3f, 3., |
| 1.37747639043129518071e-41, -3.22084585277826e35}}; |
| |
| for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_U16_F); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]), |
| [](MacroAssembler& assm) { __ fexupr_w(w2, w0); }); |
| } |
| for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_F_D); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]), |
| [](MacroAssembler& assm) { __ fexupr_d(w2, w0); }); |
| } |
| } |
| |
| struct TestCaseMsa2RF_U32_D { |
| uint32_t ws1; |
| uint32_t ws2; |
| uint32_t ws3; |
| uint32_t ws4; |
| double exp_res_1; |
| double exp_res_2; |
| }; |
| |
| TEST(MSA_ffql) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| struct TestCaseMsa2RF_U16_F tc_s[] = {{0, 3, 0xFFFF, 0x8000, 0x8000, 0xE000, |
| 0x0FF0, 0, -1.f, -0.25f, |
| 0.12451171875f, 0.f}}; |
| |
| struct TestCaseMsa2RF_U32_D tc_d[] = { |
| {0, 45, 0x80000000, 0xE0000000, -1., -0.25}, |
| {0x28379, 0xAAAA5555, 0x024903D3, 0, 17.853239085525274277e-3, 0.}}; |
| |
| for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_U16_F); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]), |
| [](MacroAssembler& assm) { __ ffql_w(w2, w0); }); |
| } |
| for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_U32_D); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]), |
| [](MacroAssembler& assm) { __ ffql_d(w2, w0); }); |
| } |
| } |
| |
| TEST(MSA_ffqr) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| struct TestCaseMsa2RF_U16_F tc_s[] = {{0x8000, 0xE000, 0x0FF0, 0, 0, 3, |
| 0xFFFF, 0x8000, -1.f, -0.25f, |
| 0.12451171875f, 0.f}}; |
| |
| struct TestCaseMsa2RF_U32_D tc_d[] = { |
| {0x80000000, 0xE0000000, 0, 45, -1., -0.25}, |
| {0x024903D3, 0, 0x28379, 0xAAAA5555, 17.853239085525274277e-3, 0.}}; |
| |
| for (size_t i = 0; i < sizeof(tc_s) / sizeof(TestCaseMsa2RF_U16_F); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_s[i]), |
| [](MacroAssembler& assm) { __ ffqr_w(w2, w0); }); |
| } |
| for (size_t i = 0; i < sizeof(tc_d) / sizeof(TestCaseMsa2RF_U32_D); ++i) { |
| run_msa_2r(reinterpret_cast<const TestCaseMsa2R*>(&tc_d[i]), |
| [](MacroAssembler& assm) { __ ffqr_d(w2, w0); }); |
| } |
| } |
| |
| struct TestCaseMsaVector { |
| uint64_t wd_lo; |
| uint64_t wd_hi; |
| uint64_t ws_lo; |
| uint64_t ws_hi; |
| uint64_t wt_lo; |
| uint64_t wt_hi; |
| }; |
| |
| template <typename InstFunc, typename OperFunc> |
| void run_msa_vector(struct TestCaseMsaVector* input, |
| InstFunc GenerateVectorInstructionFunc, |
| OperFunc GenerateOperationFunc) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| CpuFeatureScope fscope(&assm, MIPS_SIMD); |
| msa_reg_t res; |
| |
| load_elements_of_vector(assm, &(input->ws_lo), w0, t0, t1); |
| load_elements_of_vector(assm, &(input->wt_lo), w2, t0, t1); |
| load_elements_of_vector(assm, &(input->wd_lo), w4, t0, t1); |
| |
| GenerateVectorInstructionFunc(assm); |
| |
| store_elements_of_vector(assm, w4, a0); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| #ifdef OBJECT_PRINT |
| code->Print(std::cout); |
| #endif |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| |
| (f.Call(&res, 0, 0, 0, 0)); |
| |
| CHECK_EQ(GenerateOperationFunc(input->wd_lo, input->ws_lo, input->wt_lo), |
| res.d[0]); |
| CHECK_EQ(GenerateOperationFunc(input->wd_hi, input->ws_hi, input->wt_hi), |
| res.d[1]); |
| } |
| |
| TEST(MSA_vector) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| struct TestCaseMsaVector tc[] = { |
| // wd_lo, wd_hi, ws_lo, ws_hi, wt_lo, wt_hi |
| {0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636, 0xDCD39D91F9057627, |
| 0x64BE4F6DBE9CAA51, 0x6B23DE1A687D9CB9, 0x49547AAD691DA4CA}, |
| {0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636, 0x401614523D830549, |
| 0xD7C46D613F50EDDD, 0x52284CBC60A1562B, 0x1756ED510D8849CD}, |
| {0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636, 0xD6E2D2EBCB40D72F, |
| 0x13A619AFCE67B079, 0x36CCE284343E40F9, 0xB4E8F44FD148BF7F}}; |
| |
| for (size_t i = 0; i < sizeof(tc) / sizeof(TestCaseMsaVector); ++i) { |
| run_msa_vector( |
| &tc[i], [](MacroAssembler& assm) { __ and_v(w4, w0, w2); }, |
| [](uint64_t wd, uint64_t ws, uint64_t wt) { return ws & wt; }); |
| run_msa_vector( |
| &tc[i], [](MacroAssembler& assm) { __ or_v(w4, w0, w2); }, |
| [](uint64_t wd, uint64_t ws, uint64_t wt) { return ws | wt; }); |
| run_msa_vector( |
| &tc[i], [](MacroAssembler& assm) { __ nor_v(w4, w0, w2); }, |
| [](uint64_t wd, uint64_t ws, uint64_t wt) { return ~(ws | wt); }); |
| run_msa_vector( |
| &tc[i], [](MacroAssembler& assm) { __ xor_v(w4, w0, w2); }, |
| [](uint64_t wd, uint64_t ws, uint64_t wt) { return ws ^ wt; }); |
| run_msa_vector(&tc[i], [](MacroAssembler& assm) { __ bmnz_v(w4, w0, w2); }, |
| [](uint64_t wd, uint64_t ws, uint64_t wt) { |
| return (ws & wt) | (wd & ~wt); |
| }); |
| run_msa_vector(&tc[i], [](MacroAssembler& assm) { __ bmz_v(w4, w0, w2); }, |
| [](uint64_t wd, uint64_t ws, uint64_t wt) { |
| return (ws & ~wt) | (wd & wt); |
| }); |
| run_msa_vector(&tc[i], [](MacroAssembler& assm) { __ bsel_v(w4, w0, w2); }, |
| [](uint64_t wd, uint64_t ws, uint64_t wt) { |
| return (ws & ~wd) | (wt & wd); |
| }); |
| } |
| } |
| |
| struct TestCaseMsaBit { |
| uint64_t wd_lo; |
| uint64_t wd_hi; |
| uint64_t ws_lo; |
| uint64_t ws_hi; |
| uint32_t m; |
| }; |
| |
| template <typename InstFunc, typename OperFunc> |
| void run_msa_bit(struct TestCaseMsaBit* input, InstFunc GenerateInstructionFunc, |
| OperFunc GenerateOperationFunc) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| CpuFeatureScope fscope(&assm, MIPS_SIMD); |
| msa_reg_t res; |
| |
| load_elements_of_vector(assm, &(input->ws_lo), w0, t0, t1); |
| load_elements_of_vector(assm, &(input->wd_lo), w2, t0, t1); |
| |
| GenerateInstructionFunc(assm, input->m); |
| |
| store_elements_of_vector(assm, w2, a0); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| #ifdef OBJECT_PRINT |
| code->Print(std::cout); |
| #endif |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| |
| (f.Call(&res, 0, 0, 0, 0)); |
| |
| CHECK_EQ(GenerateOperationFunc(input->wd_lo, input->ws_lo, input->m), |
| res.d[0]); |
| CHECK_EQ(GenerateOperationFunc(input->wd_hi, input->ws_hi, input->m), |
| res.d[1]); |
| } |
| |
| TEST(MSA_slli_srai_srli) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| struct TestCaseMsaBit tc[] = { |
| // wd_lo, wd_hi ws_lo, ws_hi, m |
| {0, 0, 0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636, 3}, |
| {0, 0, 0x64BE4F6DBE9CAA51, 0x6B23DE1A687D9CB9, 5}, |
| {0, 0, 0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 9}, |
| {0, 0, 0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 13}, |
| {0, 0, 0x566BE7BA4365B70A, 0x01EBBC1937D76CB4, 21}, |
| {0, 0, 0x380E2DEB9D3F8AAE, 0x017E0DE0BCC6CA42, 30}, |
| {0, 0, 0xA46A3A9BCB43F4E5, 0x1C62C8473BDFCFFB, 45}, |
| {0, 0, 0xF6759D85F23B5A2B, 0x5C042AE42C6D12C1, 61}}; |
| |
| #define SLLI_SRLI_DF(lanes, mask, func) \ |
| [](uint64_t wd, uint64_t ws, uint32_t m) { \ |
| uint64_t res = 0; \ |
| int elem_size = kMSARegSize / lanes; \ |
| for (int i = 0; i < lanes / 2; ++i) { \ |
| int shift = elem_size * i; \ |
| uint64_t elem = (ws >> shift) & mask; \ |
| res |= ((func)&mask) << shift; \ |
| } \ |
| return res; \ |
| } |
| |
| #define SRAI_DF(lanes, mask, func) \ |
| [](uint64_t wd, uint64_t ws, uint32_t m) { \ |
| uint64_t res = 0; \ |
| int elem_size = kMSARegSize / lanes; \ |
| for (int i = 0; i < lanes / 2; ++i) { \ |
| int shift = elem_size * i; \ |
| int64_t elem = \ |
| static_cast<int64_t>(((ws >> shift) & mask) << (64 - elem_size)) >> \ |
| (64 - elem_size); \ |
| res |= static_cast<uint64_t>((func)&mask) << shift; \ |
| } \ |
| return res; \ |
| } |
| |
| for (size_t i = 0; i < sizeof(tc) / sizeof(TestCaseMsaBit); ++i) { |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ slli_b(w2, w0, m % 8); }, |
| SLLI_SRLI_DF(kMSALanesByte, UINT8_MAX, (elem << (m % elem_size)))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ slli_h(w2, w0, m % 16); }, |
| SLLI_SRLI_DF(kMSALanesHalf, UINT16_MAX, (elem << (m % elem_size)))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ slli_w(w2, w0, m % 32); }, |
| SLLI_SRLI_DF(kMSALanesWord, UINT32_MAX, (elem << (m % elem_size)))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ slli_d(w2, w0, m % 64); }, |
| SLLI_SRLI_DF(kMSALanesDword, UINT64_MAX, (elem << (m % elem_size)))); |
| |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ srli_b(w2, w0, m % 8); }, |
| SLLI_SRLI_DF(kMSALanesByte, UINT8_MAX, (elem >> (m % elem_size)))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ srli_h(w2, w0, m % 16); }, |
| SLLI_SRLI_DF(kMSALanesHalf, UINT16_MAX, (elem >> (m % elem_size)))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ srli_w(w2, w0, m % 32); }, |
| SLLI_SRLI_DF(kMSALanesWord, UINT32_MAX, (elem >> (m % elem_size)))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ srli_d(w2, w0, m % 64); }, |
| SLLI_SRLI_DF(kMSALanesDword, UINT64_MAX, (elem >> (m % elem_size)))); |
| |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ srlri_b(w2, w0, m % 8); }, |
| SLLI_SRLI_DF( |
| kMSALanesByte, UINT8_MAX, |
| (elem >> (m % elem_size)) + ((elem >> (m % elem_size - 1)) & 0x1))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ srlri_h(w2, w0, m % 16); }, |
| SLLI_SRLI_DF( |
| kMSALanesHalf, UINT16_MAX, |
| (elem >> (m % elem_size)) + ((elem >> (m % elem_size - 1)) & 0x1))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ srlri_w(w2, w0, m % 32); }, |
| SLLI_SRLI_DF( |
| kMSALanesWord, UINT32_MAX, |
| (elem >> (m % elem_size)) + ((elem >> (m % elem_size - 1)) & 0x1))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ srlri_d(w2, w0, m % 64); }, |
| SLLI_SRLI_DF( |
| kMSALanesDword, UINT64_MAX, |
| (elem >> (m % elem_size)) + ((elem >> (m % elem_size - 1)) & 0x1))); |
| |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ srai_b(w2, w0, m % 8); }, |
| SRAI_DF(kMSALanesByte, UINT8_MAX, |
| ArithmeticShiftRight(elem, m % elem_size))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ srai_h(w2, w0, m % 16); }, |
| SRAI_DF(kMSALanesHalf, UINT16_MAX, |
| ArithmeticShiftRight(elem, m % elem_size))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ srai_w(w2, w0, m % 32); }, |
| SRAI_DF(kMSALanesWord, UINT32_MAX, |
| ArithmeticShiftRight(elem, m % elem_size))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ srai_d(w2, w0, m % 64); }, |
| SRAI_DF(kMSALanesDword, UINT64_MAX, |
| ArithmeticShiftRight(elem, m % elem_size))); |
| |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ srari_b(w2, w0, m % 8); }, |
| SRAI_DF(kMSALanesByte, UINT8_MAX, |
| ArithmeticShiftRight(elem, m % elem_size) + |
| ((elem >> (m % elem_size - 1)) & 0x1))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ srari_h(w2, w0, m % 16); }, |
| SRAI_DF(kMSALanesHalf, UINT16_MAX, |
| ArithmeticShiftRight(elem, m % elem_size) + |
| ((elem >> (m % elem_size - 1)) & 0x1))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ srari_w(w2, w0, m % 32); }, |
| SRAI_DF(kMSALanesWord, UINT32_MAX, |
| ArithmeticShiftRight(elem, m % elem_size) + |
| ((elem >> (m % elem_size - 1)) & 0x1))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ srari_d(w2, w0, m % 64); }, |
| SRAI_DF(kMSALanesDword, UINT64_MAX, |
| ArithmeticShiftRight(elem, m % elem_size) + |
| ((elem >> (m % elem_size - 1)) & 0x1))); |
| } |
| #undef SLLI_SRLI_DF |
| #undef SRAI_DF |
| } |
| |
| TEST(MSA_bclri_bseti_bnegi) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| struct TestCaseMsaBit tc[] = { |
| // wd_lo, wd_hi, ws_lo, ws_hi, m |
| {0, 0, 0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636, 3}, |
| {0, 0, 0x64BE4F6DBE9CAA51, 0x6B23DE1A687D9CB9, 5}, |
| {0, 0, 0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 9}, |
| {0, 0, 0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 13}, |
| {0, 0, 0x566BE7BA4365B70A, 0x01EBBC1937D76CB4, 21}, |
| {0, 0, 0x380E2DEB9D3F8AAE, 0x017E0DE0BCC6CA42, 30}, |
| {0, 0, 0xA46A3A9BCB43F4E5, 0x1C62C8473BDFCFFB, 45}, |
| {0, 0, 0xF6759D85F23B5A2B, 0x5C042AE42C6D12C1, 61}}; |
| |
| #define BCLRI_BSETI_BNEGI_DF(lanes, mask, func) \ |
| [](uint64_t wd, uint64_t ws, uint32_t m) { \ |
| uint64_t res = 0; \ |
| int elem_size = kMSARegSize / lanes; \ |
| for (int i = 0; i < lanes / 2; ++i) { \ |
| int shift = elem_size * i; \ |
| uint64_t elem = (ws >> shift) & mask; \ |
| res |= ((func)&mask) << shift; \ |
| } \ |
| return res; \ |
| } |
| |
| for (size_t i = 0; i < sizeof(tc) / sizeof(TestCaseMsaBit); ++i) { |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ bclri_b(w2, w0, m % 8); }, |
| BCLRI_BSETI_BNEGI_DF(kMSALanesByte, UINT8_MAX, |
| (~(1ull << (m % elem_size)) & elem))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ bclri_h(w2, w0, m % 16); }, |
| BCLRI_BSETI_BNEGI_DF(kMSALanesHalf, UINT16_MAX, |
| (~(1ull << (m % elem_size)) & elem))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ bclri_w(w2, w0, m % 32); }, |
| BCLRI_BSETI_BNEGI_DF(kMSALanesWord, UINT32_MAX, |
| (~(1ull << (m % elem_size)) & elem))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ bclri_d(w2, w0, m % 64); }, |
| BCLRI_BSETI_BNEGI_DF(kMSALanesDword, UINT64_MAX, |
| (~(1ull << (m % elem_size)) & elem))); |
| |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ bseti_b(w2, w0, m % 8); }, |
| BCLRI_BSETI_BNEGI_DF(kMSALanesByte, UINT8_MAX, |
| ((1ull << (m % elem_size)) | elem))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ bseti_h(w2, w0, m % 16); }, |
| BCLRI_BSETI_BNEGI_DF(kMSALanesHalf, UINT16_MAX, |
| ((1ull << (m % elem_size)) | elem))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ bseti_w(w2, w0, m % 32); }, |
| BCLRI_BSETI_BNEGI_DF(kMSALanesWord, UINT32_MAX, |
| ((1ull << (m % elem_size)) | elem))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ bseti_d(w2, w0, m % 64); }, |
| BCLRI_BSETI_BNEGI_DF(kMSALanesDword, UINT64_MAX, |
| ((1ull << (m % elem_size)) | elem))); |
| |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ bnegi_b(w2, w0, m % 8); }, |
| BCLRI_BSETI_BNEGI_DF(kMSALanesByte, UINT8_MAX, |
| ((1ull << (m % elem_size)) ^ elem))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ bnegi_h(w2, w0, m % 16); }, |
| BCLRI_BSETI_BNEGI_DF(kMSALanesHalf, UINT16_MAX, |
| ((1ull << (m % elem_size)) ^ elem))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ bnegi_w(w2, w0, m % 32); }, |
| BCLRI_BSETI_BNEGI_DF(kMSALanesWord, UINT32_MAX, |
| ((1ull << (m % elem_size)) ^ elem))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ bnegi_d(w2, w0, m % 64); }, |
| BCLRI_BSETI_BNEGI_DF(kMSALanesDword, UINT64_MAX, |
| ((1ull << (m % elem_size)) ^ elem))); |
| } |
| #undef BCLRI_BSETI_BNEGI_DF |
| } |
| |
| TEST(MSA_binsli_binsri) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| struct TestCaseMsaBit tc[] = {// wd_lo, wd_hi, ws_lo, ws_hi, m |
| {0x53F4457553BBD5B4, 0x5FB8250EACC296B2, |
| 0xF35862E13E38F8B0, 0x4F41FFDEF2BFE636, 3}, |
| {0xF61BFDB0F312E6FC, 0xC9437568DD1EA925, |
| 0x64BE4F6DBE9CAA51, 0x6B23DE1A687D9CB9, 5}, |
| {0x53F4457553BBD5B4, 0x5FB8250EACC296B2, |
| 0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 9}, |
| {0xF61BFDB0F312E6FC, 0xC9437568DD1EA925, |
| 0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 13}, |
| {0x53F4457553BBD5B4, 0x5FB8250EACC296B2, |
| 0x566BE7BA4365B70A, 0x01EBBC1937D76CB4, 21}, |
| {0xF61BFDB0F312E6FC, 0xC9437568DD1EA925, |
| 0x380E2DEB9D3F8AAE, 0x017E0DE0BCC6CA42, 30}, |
| {0x53F4457553BBD5B4, 0x5FB8250EACC296B2, |
| 0xA46A3A9BCB43F4E5, 0x1C62C8473BDFCFFB, 45}, |
| {0xF61BFDB0F312E6FC, 0xC9437568DD1EA925, |
| 0xF6759D85F23B5A2B, 0x5C042AE42C6D12C1, 61}}; |
| |
| #define BINSLI_BINSRI_DF(lanes, mask, func) \ |
| [](uint64_t wd, uint64_t ws, uint32_t m) { \ |
| uint64_t res = 0; \ |
| int elem_size = kMSARegSize / lanes; \ |
| int bits = m % elem_size + 1; \ |
| for (int i = 0; i < lanes / 2; ++i) { \ |
| int shift = elem_size * i; \ |
| uint64_t ws_elem = (ws >> shift) & mask; \ |
| if (bits == elem_size) { \ |
| res |= (ws_elem & mask) << shift; \ |
| } else { \ |
| uint64_t r_mask = (1ull << bits) - 1; \ |
| uint64_t l_mask = r_mask << (elem_size - bits); \ |
| USE(l_mask); \ |
| uint64_t wd_elem = (wd >> shift) & mask; \ |
| res |= ((func)&mask) << shift; \ |
| } \ |
| } \ |
| return res; \ |
| } |
| |
| for (size_t i = 0; i < sizeof(tc) / sizeof(TestCaseMsaBit); ++i) { |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ binsli_b(w2, w0, m % 8); }, |
| BINSLI_BINSRI_DF(kMSALanesByte, UINT8_MAX, |
| ((ws_elem & l_mask) | (wd_elem & ~l_mask)))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ binsli_h(w2, w0, m % 16); }, |
| BINSLI_BINSRI_DF(kMSALanesHalf, UINT16_MAX, |
| ((ws_elem & l_mask) | (wd_elem & ~l_mask)))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ binsli_w(w2, w0, m % 32); }, |
| BINSLI_BINSRI_DF(kMSALanesWord, UINT32_MAX, |
| ((ws_elem & l_mask) | (wd_elem & ~l_mask)))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ binsli_d(w2, w0, m % 64); }, |
| BINSLI_BINSRI_DF(kMSALanesDword, UINT64_MAX, |
| ((ws_elem & l_mask) | (wd_elem & ~l_mask)))); |
| |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ binsri_b(w2, w0, m % 8); }, |
| BINSLI_BINSRI_DF(kMSALanesByte, UINT8_MAX, |
| ((ws_elem & r_mask) | (wd_elem & ~r_mask)))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ binsri_h(w2, w0, m % 16); }, |
| BINSLI_BINSRI_DF(kMSALanesHalf, UINT16_MAX, |
| ((ws_elem & r_mask) | (wd_elem & ~r_mask)))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ binsri_w(w2, w0, m % 32); }, |
| BINSLI_BINSRI_DF(kMSALanesWord, UINT32_MAX, |
| ((ws_elem & r_mask) | (wd_elem & ~r_mask)))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ binsri_d(w2, w0, m % 64); }, |
| BINSLI_BINSRI_DF(kMSALanesDword, UINT64_MAX, |
| ((ws_elem & r_mask) | (wd_elem & ~r_mask)))); |
| } |
| #undef BINSLI_BINSRI_DF |
| } |
| |
| TEST(MSA_sat_s_sat_u) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| struct TestCaseMsaBit tc[] = { |
| // wd_lo, wd_hi, ws_lo, ws_hi, m |
| {0, 0, 0xF35862E13E3808B0, 0x4F41FFDEF2BFE636, 3}, |
| {0, 0, 0x64BE4F6DBE9CAA51, 0x6B23DE1A687D9CB9, 5}, |
| {0, 0, 0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 9}, |
| {0, 0, 0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 13}, |
| {0, 0, 0x566BE7BA4365B70A, 0x01EBBC1937D76CB4, 21}, |
| {0, 0, 0x380E2DEB9D3F8AAE, 0x017E0DE0BCC6CA42, 30}, |
| {0, 0, 0xA46A3A9BCB43F4E5, 0x1C62C8473BDFCFFB, 45}, |
| {0, 0, 0xF6759D85F23B5A2B, 0x5C042AE42C6D12C1, 61}}; |
| |
| #define SAT_DF(lanes, mask, func) \ |
| [](uint64_t wd, uint64_t ws, uint32_t m) { \ |
| uint64_t res = 0; \ |
| int elem_size = kMSARegSize / lanes; \ |
| m %= elem_size; \ |
| for (int i = 0; i < lanes / 2; ++i) { \ |
| int shift = elem_size * i; \ |
| uint64_t elem_u64 = (ws >> shift) & mask; \ |
| int64_t elem_i64 = static_cast<int64_t>(elem_u64 << (64 - elem_size)) >> \ |
| (64 - elem_size); \ |
| USE(elem_i64); \ |
| res |= ((func)&mask) << shift; \ |
| } \ |
| return res; \ |
| } |
| |
| #define M_MAX_INT(x) static_cast<int64_t>((1LL << ((x)-1)) - 1) |
| #define M_MIN_INT(x) static_cast<int64_t>(-(1LL << ((x)-1))) |
| #define M_MAX_UINT(x) static_cast<uint64_t>(-1ULL >> (64 - (x))) |
| |
| for (size_t i = 0; i < sizeof(tc) / sizeof(TestCaseMsaBit); ++i) { |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ sat_u_b(w2, w0, m % 8); }, |
| SAT_DF(kMSALanesByte, UINT8_MAX, |
| (elem_u64 < M_MAX_UINT(m + 1) ? elem_u64 : M_MAX_UINT(m + 1)))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ sat_u_h(w2, w0, m % 16); }, |
| SAT_DF(kMSALanesHalf, UINT16_MAX, |
| (elem_u64 < M_MAX_UINT(m + 1) ? elem_u64 : M_MAX_UINT(m + 1)))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ sat_u_w(w2, w0, m % 32); }, |
| SAT_DF(kMSALanesWord, UINT32_MAX, |
| (elem_u64 < M_MAX_UINT(m + 1) ? elem_u64 : M_MAX_UINT(m + 1)))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ sat_u_d(w2, w0, m % 64); }, |
| SAT_DF(kMSALanesDword, UINT64_MAX, |
| (elem_u64 < M_MAX_UINT(m + 1) ? elem_u64 : M_MAX_UINT(m + 1)))); |
| |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ sat_s_b(w2, w0, m % 8); }, |
| SAT_DF( |
| kMSALanesByte, UINT8_MAX, |
| (elem_i64 < M_MIN_INT(m + 1) |
| ? M_MIN_INT(m + 1) |
| : elem_i64 > M_MAX_INT(m + 1) ? M_MAX_INT(m + 1) : elem_i64))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ sat_s_h(w2, w0, m % 16); }, |
| SAT_DF( |
| kMSALanesHalf, UINT16_MAX, |
| (elem_i64 < M_MIN_INT(m + 1) |
| ? M_MIN_INT(m + 1) |
| : elem_i64 > M_MAX_INT(m + 1) ? M_MAX_INT(m + 1) : elem_i64))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ sat_s_w(w2, w0, m % 32); }, |
| SAT_DF( |
| kMSALanesWord, UINT32_MAX, |
| (elem_i64 < M_MIN_INT(m + 1) |
| ? M_MIN_INT(m + 1) |
| : elem_i64 > M_MAX_INT(m + 1) ? M_MAX_INT(m + 1) : elem_i64))); |
| run_msa_bit( |
| &tc[i], |
| [](MacroAssembler& assm, uint32_t m) { __ sat_s_d(w2, w0, m % 64); }, |
| SAT_DF( |
| kMSALanesDword, UINT64_MAX, |
| (elem_i64 < M_MIN_INT(m + 1) |
| ? M_MIN_INT(m + 1) |
| : elem_i64 > M_MAX_INT(m + 1) ? M_MAX_INT(m + 1) : elem_i64))); |
| } |
| |
| #undef SAT_DF |
| #undef M_MAX_INT |
| #undef M_MIN_INT |
| #undef M_MAX_UINT |
| } |
| |
| template <typename InstFunc, typename OperFunc> |
| void run_msa_i10(int32_t input, InstFunc GenerateVectorInstructionFunc, |
| OperFunc GenerateOperationFunc) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| CpuFeatureScope fscope(&assm, MIPS_SIMD); |
| msa_reg_t res; |
| |
| GenerateVectorInstructionFunc(assm, input); |
| |
| store_elements_of_vector(assm, w0, a0); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| #ifdef OBJECT_PRINT |
| code->Print(std::cout); |
| #endif |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| |
| (f.Call(&res, 0, 0, 0, 0)); |
| |
| CHECK_EQ(GenerateOperationFunc(input), res.d[0]); |
| CHECK_EQ(GenerateOperationFunc(input), res.d[1]); |
| } |
| |
| TEST(MSA_ldi) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| // signed 10bit integers: -512 .. 511 |
| int32_t tc[] = {0, -1, 1, 256, -256, -178, 352, -512, 511}; |
| |
| #define LDI_DF(lanes, mask) \ |
| [](int32_t s10) { \ |
| uint64_t res = 0; \ |
| int elem_size = kMSARegSize / lanes; \ |
| int64_t s10_64 = \ |
| ArithmeticShiftRight(static_cast<int64_t>(s10) << 54, 54); \ |
| for (int i = 0; i < lanes / 2; ++i) { \ |
| int shift = elem_size * i; \ |
| res |= static_cast<uint64_t>(s10_64 & mask) << shift; \ |
| } \ |
| return res; \ |
| } |
| |
| for (size_t i = 0; i < sizeof(tc) / sizeof(int32_t); ++i) { |
| run_msa_i10(tc[i], |
| [](MacroAssembler& assm, int32_t s10) { __ ldi_b(w0, s10); }, |
| LDI_DF(kMSALanesByte, UINT8_MAX)); |
| run_msa_i10(tc[i], |
| [](MacroAssembler& assm, int32_t s10) { __ ldi_h(w0, s10); }, |
| LDI_DF(kMSALanesHalf, UINT16_MAX)); |
| run_msa_i10(tc[i], |
| [](MacroAssembler& assm, int32_t s10) { __ ldi_w(w0, s10); }, |
| LDI_DF(kMSALanesWord, UINT32_MAX)); |
| run_msa_i10(tc[i], |
| [](MacroAssembler& assm, int32_t s10) { __ ldi_d(w0, s10); }, |
| LDI_DF(kMSALanesDword, UINT64_MAX)); |
| } |
| #undef LDI_DF |
| } |
| |
| template <typename T, typename InstFunc> |
| void run_msa_mi10(InstFunc GenerateVectorInstructionFunc) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| CpuFeatureScope fscope(&assm, MIPS_SIMD); |
| T in_test_vector[1024]; |
| T out_test_vector[1024]; |
| |
| T* in_array_middle = in_test_vector + arraysize(in_test_vector) / 2; |
| T* out_array_middle = out_test_vector + arraysize(out_test_vector) / 2; |
| |
| v8::base::RandomNumberGenerator rand_gen(FLAG_random_seed); |
| for (unsigned int i = 0; i < arraysize(in_test_vector); i++) { |
| in_test_vector[i] = static_cast<T>(rand_gen.NextInt()); |
| out_test_vector[i] = 0; |
| } |
| |
| GenerateVectorInstructionFunc(assm); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| #ifdef OBJECT_PRINT |
| code->Print(std::cout); |
| #endif |
| auto f = GeneratedCode<F4>::FromCode(*code); |
| |
| (f.Call(in_array_middle, out_array_middle, 0, 0, 0)); |
| |
| CHECK_EQ(memcmp(in_test_vector, out_test_vector, arraysize(in_test_vector)), |
| 0); |
| } |
| |
| TEST(MSA_load_store_vector) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| run_msa_mi10<uint8_t>([](MacroAssembler& assm) { |
| for (int i = -512; i < 512; i += 16) { |
| __ ld_b(w0, MemOperand(a0, i)); |
| __ st_b(w0, MemOperand(a1, i)); |
| } |
| }); |
| run_msa_mi10<uint16_t>([](MacroAssembler& assm) { |
| for (int i = -512; i < 512; i += 8) { |
| __ ld_h(w0, MemOperand(a0, i)); |
| __ st_h(w0, MemOperand(a1, i)); |
| } |
| }); |
| run_msa_mi10<uint32_t>([](MacroAssembler& assm) { |
| for (int i = -512; i < 512; i += 4) { |
| __ ld_w(w0, MemOperand(a0, i)); |
| __ st_w(w0, MemOperand(a1, i)); |
| } |
| }); |
| run_msa_mi10<uint64_t>([](MacroAssembler& assm) { |
| for (int i = -512; i < 512; i += 2) { |
| __ ld_d(w0, MemOperand(a0, i)); |
| __ st_d(w0, MemOperand(a1, i)); |
| } |
| }); |
| } |
| |
| struct TestCaseMsa3R { |
| uint64_t ws_lo; |
| uint64_t ws_hi; |
| uint64_t wt_lo; |
| uint64_t wt_hi; |
| uint64_t wd_lo; |
| uint64_t wd_hi; |
| }; |
| |
| static const uint64_t Unpredictable = 0x312014017725ll; |
| |
| template <typename InstFunc, typename OperFunc> |
| void run_msa_3r(struct TestCaseMsa3R* input, InstFunc GenerateI5InstructionFunc, |
| OperFunc GenerateOperationFunc) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| CpuFeatureScope fscope(&assm, MIPS_SIMD); |
| msa_reg_t res; |
| |
| load_elements_of_vector(assm, &(input->wt_lo), w0, t0, t1); |
| load_elements_of_vector(assm, &(input->ws_lo), w1, t0, t1); |
| load_elements_of_vector(assm, &(input->wd_lo), w2, t0, t1); |
| |
| GenerateI5InstructionFunc(assm); |
| |
| store_elements_of_vector(assm, w2, a0); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| #ifdef OBJECT_PRINT |
| code->Print(std::cout); |
| #endif |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| |
| (f.Call(&res, 0, 0, 0, 0)); |
| |
| GenerateOperationFunc(&input->ws_lo, &input->wt_lo, &input->wd_lo); |
| if (input->wd_lo != Unpredictable) { |
| CHECK_EQ(input->wd_lo, res.d[0]); |
| } |
| if (input->wd_hi != Unpredictable) { |
| CHECK_EQ(input->wd_hi, res.d[1]); |
| } |
| } |
| |
| TEST(MSA_3R_instructions) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| struct TestCaseMsa3R tc[] = { |
| {0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0x1169751BB9A7D9C3, |
| 0xF7A594AEC8EF8A9C, 0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C}, |
| {0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0x2B665362C4E812DF, |
| 0x3A0D80D68B3F8BC8, 0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8}, |
| {0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C, 0x1169751BB9A7D9C3, |
| 0xF7A594AEC8EF8A9C, 0x1169751BB9A7D9C3, 0xF7A594AEC8EF8A9C}, |
| {0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8, 0x2B665362C4E812DF, |
| 0x3A0D80D68B3F8BC8, 0x2B665362C4E812DF, 0x3A0D80D68B3F8BC8}, |
| {0xFFAB807F807FFFCD, 0x7F23FF80FF567F80, 0xFFAB807F807FFFCD, |
| 0x7F23FF80FF567F80, 0xFFAB807F807FFFCD, 0x7F23FF80FF567F80}, |
| {0x80FFEFFF7F12807F, 0x807F80FF7FDEFF78, 0x80FFEFFF7F12807F, |
| 0x807F80FF7FDEFF78, 0x80FFEFFF7F12807F, 0x807F80FF7FDEFF78}, |
| {0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF, |
| 0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF}, |
| {0x0000000000000000, 0xFFFFFFFFFFFFFFFF, 0xFFFFFFFFFFFFFFFF, |
| 0x0000000000000000, 0x0000000000000000, 0xFFFFFFFFFFFFFFFF}, |
| {0xFFFF0000FFFF0000, 0xFFFF0000FFFF0000, 0xFFFF0000FFFF0000, |
| 0xFFFF0000FFFF0000, 0xFFFF0000FFFF0000, 0xFFFF0000FFFF0000}, |
| {0xFF00FF00FF00FF00, 0xFF00FF00FF00FF00, 0xFF00FF00FF00FF00, |
| 0xFF00FF00FF00FF00, 0xFF00FF00FF00FF00, 0xFF00FF00FF00FF00}, |
| {0xF0F0F0F0F0F0F0F0, 0xF0F0F0F0F0F0F0F0, 0xF0F0F0F0F0F0F0F0, |
| 0xF0F0F0F0F0F0F0F0, 0xF0F0F0F0F0F0F0F0, 0xF0F0F0F0F0F0F0F0}, |
| {0xFF0000FFFF0000FF, 0xFF0000FFFF0000FF, 0xFF0000FFFF0000FF, |
| 0xFF0000FFFF0000FF, 0xFF0000FFFF0000FF, 0xFF0000FFFF0000FF}, |
| {0xFFFF00000000FFFF, 0xFFFF00000000FFFF, 0xFFFF00000000FFFF, |
| 0xFFFF00000000FFFF, 0xFFFF00000000FFFF, 0xFFFF00000000FFFF}}; |
| |
| #define SLL_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T src_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T shift_op = static_cast<T>((wt[i] >> shift) & mask) % size_in_bits; \ |
| res |= (static_cast<uint64_t>(src_op << shift_op) & mask) << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define SRA_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T src_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T shift_op = ((wt[i] >> shift) & mask) % size_in_bits; \ |
| res |= (static_cast<uint64_t>(ArithmeticShiftRight(src_op, shift_op) & \ |
| mask)) \ |
| << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define SRL_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T src_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T shift_op = static_cast<T>(((wt[i] >> shift) & mask) % size_in_bits); \ |
| res |= (static_cast<uint64_t>(src_op >> shift_op) & mask) << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define BCRL_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T src_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T shift_op = static_cast<T>(((wt[i] >> shift) & mask) % size_in_bits); \ |
| T r = (static_cast<T>(~(1ull << shift_op)) & src_op) & mask; \ |
| res |= static_cast<uint64_t>(r) << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define BSET_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T src_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T shift_op = static_cast<T>(((wt[i] >> shift) & mask) % size_in_bits); \ |
| T r = (static_cast<T>(1ull << shift_op) | src_op) & mask; \ |
| res |= static_cast<uint64_t>(r) << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define BNEG_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T src_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T shift_op = static_cast<T>(((wt[i] >> shift) & mask) % size_in_bits); \ |
| T r = (static_cast<T>(1ull << shift_op) ^ src_op) & mask; \ |
| res |= static_cast<uint64_t>(r) << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define BINSL_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T ws_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T wd_op = static_cast<T>((wd[i] >> shift) & mask); \ |
| T shift_op = static_cast<T>(((wt[i] >> shift) & mask) % size_in_bits); \ |
| int bits = shift_op + 1; \ |
| T r; \ |
| if (bits == size_in_bits) { \ |
| r = static_cast<T>(ws_op); \ |
| } else { \ |
| uint64_t mask2 = ((1ull << bits) - 1) << (size_in_bits - bits); \ |
| r = static_cast<T>((static_cast<T>(mask2) & ws_op) | \ |
| (static_cast<T>(~mask2) & wd_op)); \ |
| } \ |
| res |= static_cast<uint64_t>(r) << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define BINSR_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T ws_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T wd_op = static_cast<T>((wd[i] >> shift) & mask); \ |
| T shift_op = static_cast<T>(((wt[i] >> shift) & mask) % size_in_bits); \ |
| int bits = shift_op + 1; \ |
| T r; \ |
| if (bits == size_in_bits) { \ |
| r = static_cast<T>(ws_op); \ |
| } else { \ |
| uint64_t mask2 = (1ull << bits) - 1; \ |
| r = static_cast<T>((static_cast<T>(mask2) & ws_op) | \ |
| (static_cast<T>(~mask2) & wd_op)); \ |
| } \ |
| res |= static_cast<uint64_t>(r) << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define ADDV_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T ws_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T wt_op = static_cast<T>((wt[i] >> shift) & mask); \ |
| res |= (static_cast<uint64_t>(ws_op + wt_op) & mask) << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define SUBV_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T ws_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T wt_op = static_cast<T>((wt[i] >> shift) & mask); \ |
| res |= (static_cast<uint64_t>(ws_op - wt_op) & mask) << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define MAX_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T ws_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T wt_op = static_cast<T>((wt[i] >> shift) & mask); \ |
| res |= (static_cast<uint64_t>(Max<T>(ws_op, wt_op)) & mask) << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define MIN_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T ws_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T wt_op = static_cast<T>((wt[i] >> shift) & mask); \ |
| res |= (static_cast<uint64_t>(Min<T>(ws_op, wt_op)) & mask) << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define MAXA_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T ws_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T wt_op = static_cast<T>((wt[i] >> shift) & mask); \ |
| res |= \ |
| (static_cast<uint64_t>(Nabs(ws_op) < Nabs(wt_op) ? ws_op : wt_op) & \ |
| mask) \ |
| << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define MINA_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T ws_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T wt_op = static_cast<T>((wt[i] >> shift) & mask); \ |
| res |= \ |
| (static_cast<uint64_t>(Nabs(ws_op) > Nabs(wt_op) ? ws_op : wt_op) & \ |
| mask) \ |
| << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define CEQ_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T ws_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T wt_op = static_cast<T>((wt[i] >> shift) & mask); \ |
| res |= (static_cast<uint64_t>(!Compare(ws_op, wt_op) ? -1ull : 0ull) & \ |
| mask) \ |
| << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define CLT_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T ws_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T wt_op = static_cast<T>((wt[i] >> shift) & mask); \ |
| res |= (static_cast<uint64_t>((Compare(ws_op, wt_op) == -1) ? -1ull \ |
| : 0ull) & \ |
| mask) \ |
| << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define CLE_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T ws_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T wt_op = static_cast<T>((wt[i] >> shift) & mask); \ |
| res |= (static_cast<uint64_t>((Compare(ws_op, wt_op) != 1) ? -1ull \ |
| : 0ull) & \ |
| mask) \ |
| << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define ADD_A_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T ws_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T wt_op = static_cast<T>((wt[i] >> shift) & mask); \ |
| res |= (static_cast<uint64_t>(Abs(ws_op) + Abs(wt_op)) & mask) << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define ADDS_A_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T ws_op = Nabs(static_cast<T>((ws[i] >> shift) & mask)); \ |
| T wt_op = Nabs(static_cast<T>((wt[i] >> shift) & mask)); \ |
| T r; \ |
| if (ws_op < -std::numeric_limits<T>::max() - wt_op) { \ |
| r = std::numeric_limits<T>::max(); \ |
| } else { \ |
| r = -(ws_op + wt_op); \ |
| } \ |
| res |= (static_cast<uint64_t>(r) & mask) << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define ADDS_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T ws_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T wt_op = static_cast<T>((wt[i] >> shift) & mask); \ |
| res |= (static_cast<uint64_t>(SaturateAdd(ws_op, wt_op)) & mask) \ |
| << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define AVE_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T ws_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T wt_op = static_cast<T>((wt[i] >> shift) & mask); \ |
| res |= (static_cast<uint64_t>( \ |
| ((wt_op & ws_op) + ((ws_op ^ wt_op) >> 1)) & mask)) \ |
| << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define AVER_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T ws_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T wt_op = static_cast<T>((wt[i] >> shift) & mask); \ |
| res |= (static_cast<uint64_t>( \ |
| ((wt_op | ws_op) - ((ws_op ^ wt_op) >> 1)) & mask)) \ |
| << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define SUBS_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T ws_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T wt_op = static_cast<T>((wt[i] >> shift) & mask); \ |
| res |= (static_cast<uint64_t>(SaturateSub(ws_op, wt_op)) & mask) \ |
| << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define SUBSUS_U_DF(T, lanes, mask) \ |
| using uT = typename std::make_unsigned<T>::type; \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| uT ws_op = static_cast<uT>((ws[i] >> shift) & mask); \ |
| T wt_op = static_cast<T>((wt[i] >> shift) & mask); \ |
| T r; \ |
| if (wt_op > 0) { \ |
| uT wtu = static_cast<uT>(wt_op); \ |
| if (wtu > ws_op) { \ |
| r = 0; \ |
| } else { \ |
| r = static_cast<T>(ws_op - wtu); \ |
| } \ |
| } else { \ |
| if (ws_op > std::numeric_limits<uT>::max() + wt_op) { \ |
| r = static_cast<T>(std::numeric_limits<uT>::max()); \ |
| } else { \ |
| r = static_cast<T>(ws_op - wt_op); \ |
| } \ |
| } \ |
| res |= (static_cast<uint64_t>(r) & mask) << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define SUBSUU_S_DF(T, lanes, mask) \ |
| using uT = typename std::make_unsigned<T>::type; \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| uT ws_op = static_cast<uT>((ws[i] >> shift) & mask); \ |
| uT wt_op = static_cast<uT>((wt[i] >> shift) & mask); \ |
| uT wdu; \ |
| T r; \ |
| if (ws_op > wt_op) { \ |
| wdu = ws_op - wt_op; \ |
| if (wdu > std::numeric_limits<T>::max()) { \ |
| r = std::numeric_limits<T>::max(); \ |
| } else { \ |
| r = static_cast<T>(wdu); \ |
| } \ |
| } else { \ |
| wdu = wt_op - ws_op; \ |
| CHECK(-std::numeric_limits<T>::max() == \ |
| std::numeric_limits<T>::min() + 1); \ |
| if (wdu <= std::numeric_limits<T>::max()) { \ |
| r = -static_cast<T>(wdu); \ |
| } else { \ |
| r = std::numeric_limits<T>::min(); \ |
| } \ |
| } \ |
| res |= (static_cast<uint64_t>(r) & mask) << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define ASUB_S_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T ws_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T wt_op = static_cast<T>((wt[i] >> shift) & mask); \ |
| res |= (static_cast<uint64_t>(Abs(ws_op - wt_op)) & mask) << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define ASUB_U_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T ws_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T wt_op = static_cast<T>((wt[i] >> shift) & mask); \ |
| res |= (static_cast<uint64_t>(ws_op > wt_op ? ws_op - wt_op \ |
| : wt_op - ws_op) & \ |
| mask) \ |
| << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define MULV_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T ws_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T wt_op = static_cast<T>((wt[i] >> shift) & mask); \ |
| res |= (static_cast<uint64_t>(ws_op * wt_op) & mask) << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define MADDV_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T ws_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T wt_op = static_cast<T>((wt[i] >> shift) & mask); \ |
| T wd_op = static_cast<T>((wd[i] >> shift) & mask); \ |
| res |= (static_cast<uint64_t>(wd_op + ws_op * wt_op) & mask) << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define MSUBV_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T ws_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T wt_op = static_cast<T>((wt[i] >> shift) & mask); \ |
| T wd_op = static_cast<T>((wd[i] >> shift) & mask); \ |
| res |= (static_cast<uint64_t>(wd_op - ws_op * wt_op) & mask) << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define DIV_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T ws_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T wt_op = static_cast<T>((wt[i] >> shift) & mask); \ |
| if (wt_op == 0) { \ |
| res = Unpredictable; \ |
| break; \ |
| } \ |
| res |= (static_cast<uint64_t>(ws_op / wt_op) & mask) << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define MOD_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T ws_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T wt_op = static_cast<T>((wt[i] >> shift) & mask); \ |
| if (wt_op == 0) { \ |
| res = Unpredictable; \ |
| break; \ |
| } \ |
| res |= (static_cast<uint64_t>(wt_op != 0 ? ws_op % wt_op : 0) & mask) \ |
| << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define SRAR_DF(T, lanes, mask) \ |
| int size_in_bits = kMSARegSize / lanes; \ |
| for (int i = 0; i < 2; i++) { \ |
| uint64_t res = 0; \ |
| for (int j = 0; j < lanes / 2; ++j) { \ |
| uint64_t shift = size_in_bits * j; \ |
| T src_op = static_cast<T>((ws[i] >> shift) & mask); \ |
| T shift_op = ((wt[i] >> shift) & mask) % size_in_bits; \ |
| uint32_t bit = shift_op == 0 ? 0 : src_op >> (shift_op - 1) & 1; \ |
| res |= (static_cast<uint64_t>(ArithmeticShiftRight(src_op, shift_op) + \ |
| bit) & \ |
| mask) \ |
| << shift; \ |
| } \ |
| wd[i] = res; \ |
| } |
| |
| #define PCKEV_DF(T, lanes, mask) \ |
| T* ws_p = reinterpret_cast<T*>(ws); \ |
| T* wt_p = reinterpret_cast<T*>(wt); \ |
| T* wd_p = reinterpret_cast<T*>(wd); \ |
| for (int i = 0; i < lanes / 2; ++i) { \ |
| wd_p[i] = wt_p[2 * i]; \ |
| wd_p[i + lanes / 2] = ws_p[2 * i]; \ |
| } |
| |
| #define PCKOD_DF(T, lanes, mask) \ |
| T* ws_p = reinterpret_cast<T*>(ws); \ |
| T* wt_p = reinterpret_cast<T*>(wt); \ |
| T* wd_p = reinterpret_cast<T*>(wd); \ |
| for (int i = 0; i < lanes / 2; ++i) { \ |
| wd_p[i] = wt_p[2 * i + 1]; \ |
| wd_p[i + lanes / 2] = ws_p[2 * i + 1]; \ |
| } |
| |
| #define ILVL_DF(T, lanes, mask) \ |
| T* ws_p = reinterpret_cast<T*>(ws); \ |
| T* wt_p = reinterpret_cast<T*>(wt); \ |
| T* wd_p = reinterpret_cast<T*>(wd); \ |
| for (int i = 0; i < lanes / 2; ++i) { \ |
| wd_p[2 * i] = wt_p[i + lanes / 2]; \ |
| wd_p[2 * i + 1] = ws_p[i + lanes / 2]; \ |
| } |
| |
| #define ILVR_DF(T, lanes, mask) \ |
| T* ws_p = reinterpret_cast<T*>(ws); \ |
| T* wt_p = reinterpret_cast<T*>(wt); \ |
| T* wd_p = reinterpret_cast<T*>(wd); \ |
| for (int i = 0; i < lanes / 2; ++i) { \ |
| wd_p[2 * i] = wt_p[i]; \ |
| wd_p[2 * i + 1] = ws_p[i]; \ |
| } |
| |
| #define ILVEV_DF(T, lanes, mask) \ |
| T* ws_p = reinterpret_cast<T*>(ws); \ |
| T* wt_p = reinterpret_cast<T*>(wt); \ |
| T* wd_p = reinterpret_cast<T*>(wd); \ |
| for (int i = 0; i < lanes / 2; ++i) { \ |
| wd_p[2 * i] = wt_p[2 * i]; \ |
| wd_p[2 * i + 1] = ws_p[2 * i]; \ |
| } |
| |
| #define ILVOD_DF(T, lanes, mask) \ |
| T* ws_p = reinterpret_cast<T*>(ws); \ |
| T* wt_p = reinterpret_cast<T*>(wt); \ |
| T* wd_p = reinterpret_cast<T*>(wd); \ |
| for (int i = 0; i < lanes / 2; ++i) { \ |
| wd_p[2 * i] = wt_p[2 * i + 1]; \ |
| wd_p[2 * i + 1] = ws_p[2 * i + 1]; \ |
| } |
| |
| #define VSHF_DF(T, lanes, mask) \ |
| T* ws_p = reinterpret_cast<T*>(ws); \ |
| T* wt_p = reinterpret_cast<T*>(wt); \ |
| T* wd_p = reinterpret_cast<T*>(wd); \ |
| const int mask_not_valid = 0xC0; \ |
| const int mask_6bits = 0x3F; \ |
| for (int i = 0; i < lanes; ++i) { \ |
| if ((wd_p[i] & mask_not_valid)) { \ |
| wd_p[i] = 0; \ |
| } else { \ |
| int k = (wd_p[i] & mask_6bits) % (lanes * 2); \ |
| wd_p[i] = k > lanes ? ws_p[k - lanes] : wt_p[k]; \ |
| } \ |
| } |
| |
| #define HADD_DF(T, T_small, lanes) \ |
| T_small* ws_p = reinterpret_cast<T_small*>(ws); \ |
| T_small* wt_p = reinterpret_cast<T_small*>(wt); \ |
| T* wd_p = reinterpret_cast<T*>(wd); \ |
| for (int i = 0; i < lanes; ++i) { \ |
| wd_p[i] = static_cast<T>(ws_p[2 * i + 1]) + static_cast<T>(wt_p[2 * i]); \ |
| } |
| |
| #define HSUB_DF(T, T_small, lanes) \ |
| T_small* ws_p = reinterpret_cast<T_small*>(ws); \ |
| T_small* wt_p = reinterpret_cast<T_small*>(wt); \ |
| T* wd_p = reinterpret_cast<T*>(wd); \ |
| for (int i = 0; i < lanes; ++i) { \ |
| wd_p[i] = static_cast<T>(ws_p[2 * i + 1]) - static_cast<T>(wt_p[2 * i]); \ |
| } |
| |
| #define TEST_CASE(V) \ |
| V(sll_b, SLL_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(sll_h, SLL_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(sll_w, SLL_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(sll_d, SLL_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(srl_b, SRL_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(srl_h, SRL_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(srl_w, SRL_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(srl_d, SRL_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(bclr_b, BCRL_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(bclr_h, BCRL_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(bclr_w, BCRL_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(bclr_d, BCRL_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(bset_b, BSET_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(bset_h, BSET_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(bset_w, BSET_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(bset_d, BSET_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(bneg_b, BNEG_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(bneg_h, BNEG_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(bneg_w, BNEG_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(bneg_d, BNEG_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(binsl_b, BINSL_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(binsl_h, BINSL_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(binsl_w, BINSL_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(binsl_d, BINSL_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(binsr_b, BINSR_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(binsr_h, BINSR_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(binsr_w, BINSR_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(binsr_d, BINSR_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(addv_b, ADDV_DF, int8_t, kMSALanesByte, UINT8_MAX) \ |
| V(addv_h, ADDV_DF, int16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(addv_w, ADDV_DF, int32_t, kMSALanesWord, UINT32_MAX) \ |
| V(addv_d, ADDV_DF, int64_t, kMSALanesDword, UINT64_MAX) \ |
| V(subv_b, SUBV_DF, int8_t, kMSALanesByte, UINT8_MAX) \ |
| V(subv_h, SUBV_DF, int16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(subv_w, SUBV_DF, int32_t, kMSALanesWord, UINT32_MAX) \ |
| V(subv_d, SUBV_DF, int64_t, kMSALanesDword, UINT64_MAX) \ |
| V(max_s_b, MAX_DF, int8_t, kMSALanesByte, UINT8_MAX) \ |
| V(max_s_h, MAX_DF, int16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(max_s_w, MAX_DF, int32_t, kMSALanesWord, UINT32_MAX) \ |
| V(max_s_d, MAX_DF, int64_t, kMSALanesDword, UINT64_MAX) \ |
| V(max_u_b, MAX_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(max_u_h, MAX_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(max_u_w, MAX_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(max_u_d, MAX_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(min_s_b, MIN_DF, int8_t, kMSALanesByte, UINT8_MAX) \ |
| V(min_s_h, MIN_DF, int16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(min_s_w, MIN_DF, int32_t, kMSALanesWord, UINT32_MAX) \ |
| V(min_s_d, MIN_DF, int64_t, kMSALanesDword, UINT64_MAX) \ |
| V(min_u_b, MIN_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(min_u_h, MIN_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(min_u_w, MIN_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(min_u_d, MIN_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(max_a_b, MAXA_DF, int8_t, kMSALanesByte, UINT8_MAX) \ |
| V(max_a_h, MAXA_DF, int16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(max_a_w, MAXA_DF, int32_t, kMSALanesWord, UINT32_MAX) \ |
| V(max_a_d, MAXA_DF, int64_t, kMSALanesDword, UINT64_MAX) \ |
| V(min_a_b, MINA_DF, int8_t, kMSALanesByte, UINT8_MAX) \ |
| V(min_a_h, MINA_DF, int16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(min_a_w, MINA_DF, int32_t, kMSALanesWord, UINT32_MAX) \ |
| V(min_a_d, MINA_DF, int64_t, kMSALanesDword, UINT64_MAX) \ |
| V(ceq_b, CEQ_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(ceq_h, CEQ_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(ceq_w, CEQ_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(ceq_d, CEQ_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(clt_s_b, CLT_DF, int8_t, kMSALanesByte, UINT8_MAX) \ |
| V(clt_s_h, CLT_DF, int16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(clt_s_w, CLT_DF, int32_t, kMSALanesWord, UINT32_MAX) \ |
| V(clt_s_d, CLT_DF, int64_t, kMSALanesDword, UINT64_MAX) \ |
| V(clt_u_b, CLT_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(clt_u_h, CLT_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(clt_u_w, CLT_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(clt_u_d, CLT_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(cle_s_b, CLE_DF, int8_t, kMSALanesByte, UINT8_MAX) \ |
| V(cle_s_h, CLE_DF, int16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(cle_s_w, CLE_DF, int32_t, kMSALanesWord, UINT32_MAX) \ |
| V(cle_s_d, CLE_DF, int64_t, kMSALanesDword, UINT64_MAX) \ |
| V(cle_u_b, CLE_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(cle_u_h, CLE_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(cle_u_w, CLE_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(cle_u_d, CLE_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(add_a_b, ADD_A_DF, int8_t, kMSALanesByte, UINT8_MAX) \ |
| V(add_a_h, ADD_A_DF, int16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(add_a_w, ADD_A_DF, int32_t, kMSALanesWord, UINT32_MAX) \ |
| V(add_a_d, ADD_A_DF, int64_t, kMSALanesDword, UINT64_MAX) \ |
| V(adds_a_b, ADDS_A_DF, int8_t, kMSALanesByte, UINT8_MAX) \ |
| V(adds_a_h, ADDS_A_DF, int16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(adds_a_w, ADDS_A_DF, int32_t, kMSALanesWord, UINT32_MAX) \ |
| V(adds_a_d, ADDS_A_DF, int64_t, kMSALanesDword, UINT64_MAX) \ |
| V(adds_s_b, ADDS_DF, int8_t, kMSALanesByte, UINT8_MAX) \ |
| V(adds_s_h, ADDS_DF, int16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(adds_s_w, ADDS_DF, int32_t, kMSALanesWord, UINT32_MAX) \ |
| V(adds_s_d, ADDS_DF, int64_t, kMSALanesDword, UINT64_MAX) \ |
| V(adds_u_b, ADDS_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(adds_u_h, ADDS_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(adds_u_w, ADDS_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(adds_u_d, ADDS_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(ave_s_b, AVE_DF, int8_t, kMSALanesByte, UINT8_MAX) \ |
| V(ave_s_h, AVE_DF, int16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(ave_s_w, AVE_DF, int32_t, kMSALanesWord, UINT32_MAX) \ |
| V(ave_s_d, AVE_DF, int64_t, kMSALanesDword, UINT64_MAX) \ |
| V(ave_u_b, AVE_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(ave_u_h, AVE_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(ave_u_w, AVE_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(ave_u_d, AVE_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(aver_s_b, AVER_DF, int8_t, kMSALanesByte, UINT8_MAX) \ |
| V(aver_s_h, AVER_DF, int16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(aver_s_w, AVER_DF, int32_t, kMSALanesWord, UINT32_MAX) \ |
| V(aver_s_d, AVER_DF, int64_t, kMSALanesDword, UINT64_MAX) \ |
| V(aver_u_b, AVER_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(aver_u_h, AVER_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(aver_u_w, AVER_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(aver_u_d, AVER_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(subs_s_b, SUBS_DF, int8_t, kMSALanesByte, UINT8_MAX) \ |
| V(subs_s_h, SUBS_DF, int16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(subs_s_w, SUBS_DF, int32_t, kMSALanesWord, UINT32_MAX) \ |
| V(subs_s_d, SUBS_DF, int64_t, kMSALanesDword, UINT64_MAX) \ |
| V(subs_u_b, SUBS_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(subs_u_h, SUBS_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(subs_u_w, SUBS_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(subs_u_d, SUBS_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(subsus_u_b, SUBSUS_U_DF, int8_t, kMSALanesByte, UINT8_MAX) \ |
| V(subsus_u_h, SUBSUS_U_DF, int16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(subsus_u_w, SUBSUS_U_DF, int32_t, kMSALanesWord, UINT32_MAX) \ |
| V(subsus_u_d, SUBSUS_U_DF, int64_t, kMSALanesDword, UINT64_MAX) \ |
| V(subsuu_s_b, SUBSUU_S_DF, int8_t, kMSALanesByte, UINT8_MAX) \ |
| V(subsuu_s_h, SUBSUU_S_DF, int16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(subsuu_s_w, SUBSUU_S_DF, int32_t, kMSALanesWord, UINT32_MAX) \ |
| V(subsuu_s_d, SUBSUU_S_DF, int64_t, kMSALanesDword, UINT64_MAX) \ |
| V(asub_s_b, ASUB_S_DF, int8_t, kMSALanesByte, UINT8_MAX) \ |
| V(asub_s_h, ASUB_S_DF, int16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(asub_s_w, ASUB_S_DF, int32_t, kMSALanesWord, UINT32_MAX) \ |
| V(asub_s_d, ASUB_S_DF, int64_t, kMSALanesDword, UINT64_MAX) \ |
| V(asub_u_b, ASUB_U_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(asub_u_h, ASUB_U_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(asub_u_w, ASUB_U_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(asub_u_d, ASUB_U_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(mulv_b, MULV_DF, int8_t, kMSALanesByte, UINT8_MAX) \ |
| V(mulv_h, MULV_DF, int16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(mulv_w, MULV_DF, int32_t, kMSALanesWord, UINT32_MAX) \ |
| V(mulv_d, MULV_DF, int64_t, kMSALanesDword, UINT64_MAX) \ |
| V(maddv_b, MADDV_DF, int8_t, kMSALanesByte, UINT8_MAX) \ |
| V(maddv_h, MADDV_DF, int16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(maddv_w, MADDV_DF, int32_t, kMSALanesWord, UINT32_MAX) \ |
| V(maddv_d, MADDV_DF, int64_t, kMSALanesDword, UINT64_MAX) \ |
| V(msubv_b, MSUBV_DF, int8_t, kMSALanesByte, UINT8_MAX) \ |
| V(msubv_h, MSUBV_DF, int16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(msubv_w, MSUBV_DF, int32_t, kMSALanesWord, UINT32_MAX) \ |
| V(msubv_d, MSUBV_DF, int64_t, kMSALanesDword, UINT64_MAX) \ |
| V(div_s_b, DIV_DF, int8_t, kMSALanesByte, UINT8_MAX) \ |
| V(div_s_h, DIV_DF, int16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(div_s_w, DIV_DF, int32_t, kMSALanesWord, UINT32_MAX) \ |
| V(div_s_d, DIV_DF, int64_t, kMSALanesDword, UINT64_MAX) \ |
| V(div_u_b, DIV_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(div_u_h, DIV_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(div_u_w, DIV_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(div_u_d, DIV_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(mod_s_b, MOD_DF, int8_t, kMSALanesByte, UINT8_MAX) \ |
| V(mod_s_h, MOD_DF, int16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(mod_s_w, MOD_DF, int32_t, kMSALanesWord, UINT32_MAX) \ |
| V(mod_s_d, MOD_DF, int64_t, kMSALanesDword, UINT64_MAX) \ |
| V(mod_u_b, MOD_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(mod_u_h, MOD_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(mod_u_w, MOD_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(mod_u_d, MOD_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(srlr_b, SRAR_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(srlr_h, SRAR_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(srlr_w, SRAR_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(srlr_d, SRAR_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(pckev_b, PCKEV_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(pckev_h, PCKEV_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(pckev_w, PCKEV_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(pckev_d, PCKEV_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(pckod_b, PCKOD_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(pckod_h, PCKOD_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(pckod_w, PCKOD_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(pckod_d, PCKOD_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(ilvl_b, ILVL_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(ilvl_h, ILVL_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(ilvl_w, ILVL_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(ilvl_d, ILVL_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(ilvr_b, ILVR_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(ilvr_h, ILVR_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(ilvr_w, ILVR_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(ilvr_d, ILVR_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(ilvev_b, ILVEV_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(ilvev_h, ILVEV_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(ilvev_w, ILVEV_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(ilvev_d, ILVEV_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(ilvod_b, ILVOD_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(ilvod_h, ILVOD_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(ilvod_w, ILVOD_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(ilvod_d, ILVOD_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(vshf_b, VSHF_DF, uint8_t, kMSALanesByte, UINT8_MAX) \ |
| V(vshf_h, VSHF_DF, uint16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(vshf_w, VSHF_DF, uint32_t, kMSALanesWord, UINT32_MAX) \ |
| V(vshf_d, VSHF_DF, uint64_t, kMSALanesDword, UINT64_MAX) \ |
| V(hadd_s_h, HADD_DF, int16_t, int8_t, kMSALanesHalf) \ |
| V(hadd_s_w, HADD_DF, int32_t, int16_t, kMSALanesWord) \ |
| V(hadd_s_d, HADD_DF, int64_t, int32_t, kMSALanesDword) \ |
| V(hadd_u_h, HADD_DF, uint16_t, uint8_t, kMSALanesHalf) \ |
| V(hadd_u_w, HADD_DF, uint32_t, uint16_t, kMSALanesWord) \ |
| V(hadd_u_d, HADD_DF, uint64_t, uint32_t, kMSALanesDword) \ |
| V(hsub_s_h, HSUB_DF, int16_t, int8_t, kMSALanesHalf) \ |
| V(hsub_s_w, HSUB_DF, int32_t, int16_t, kMSALanesWord) \ |
| V(hsub_s_d, HSUB_DF, int64_t, int32_t, kMSALanesDword) \ |
| V(hsub_u_h, HSUB_DF, uint16_t, uint8_t, kMSALanesHalf) \ |
| V(hsub_u_w, HSUB_DF, uint32_t, uint16_t, kMSALanesWord) \ |
| V(hsub_u_d, HSUB_DF, uint64_t, uint32_t, kMSALanesDword) |
| |
| #define RUN_TEST(instr, verify, type, lanes, mask) \ |
| run_msa_3r(&tc[i], [](MacroAssembler& assm) { __ instr(w2, w1, w0); }, \ |
| [](uint64_t* ws, uint64_t* wt, uint64_t* wd) { \ |
| verify(type, lanes, mask); \ |
| }); |
| |
| for (size_t i = 0; i < arraysize(tc); ++i) { |
| TEST_CASE(RUN_TEST) |
| } |
| |
| #define RUN_TEST2(instr, verify, type, lanes, mask) \ |
| for (unsigned i = 0; i < arraysize(tc); i++) { \ |
| for (unsigned j = 0; j < 3; j++) { \ |
| for (unsigned k = 0; k < lanes; k++) { \ |
| type* element = reinterpret_cast<type*>(&tc[i]); \ |
| element[k + j * lanes] &= std::numeric_limits<type>::max(); \ |
| } \ |
| } \ |
| } \ |
| run_msa_3r(&tc[i], [](MacroAssembler& assm) { __ instr(w2, w1, w0); }, \ |
| [](uint64_t* ws, uint64_t* wt, uint64_t* wd) { \ |
| verify(type, lanes, mask); \ |
| }); |
| |
| #define TEST_CASE2(V) \ |
| V(sra_b, SRA_DF, int8_t, kMSALanesByte, UINT8_MAX) \ |
| V(sra_h, SRA_DF, int16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(sra_w, SRA_DF, int32_t, kMSALanesWord, UINT32_MAX) \ |
| V(sra_d, SRA_DF, int64_t, kMSALanesDword, UINT64_MAX) \ |
| V(srar_b, SRAR_DF, int8_t, kMSALanesByte, UINT8_MAX) \ |
| V(srar_h, SRAR_DF, int16_t, kMSALanesHalf, UINT16_MAX) \ |
| V(srar_w, SRAR_DF, int32_t, kMSALanesWord, UINT32_MAX) \ |
| V(srar_d, SRAR_DF, int64_t, kMSALanesDword, UINT64_MAX) |
| |
| for (size_t i = 0; i < arraysize(tc); ++i) { |
| TEST_CASE2(RUN_TEST2) |
| } |
| |
| #undef TEST_CASE |
| #undef TEST_CASE2 |
| #undef RUN_TEST |
| #undef RUN_TEST2 |
| #undef SLL_DF |
| #undef SRL_DF |
| #undef SRA_DF |
| #undef BCRL_DF |
| #undef BSET_DF |
| #undef BNEG_DF |
| #undef BINSL_DF |
| #undef BINSR_DF |
| #undef ADDV_DF |
| #undef SUBV_DF |
| #undef MAX_DF |
| #undef MIN_DF |
| #undef MAXA_DF |
| #undef MINA_DF |
| #undef CEQ_DF |
| #undef CLT_DF |
| #undef CLE_DF |
| #undef ADD_A_DF |
| #undef ADDS_A_DF |
| #undef ADDS_DF |
| #undef AVE_DF |
| #undef AVER_DF |
| #undef SUBS_DF |
| #undef SUBSUS_U_DF |
| #undef SUBSUU_S_DF |
| #undef ASUB_S_DF |
| #undef ASUB_U_DF |
| #undef MULV_DF |
| #undef MADDV_DF |
| #undef MSUBV_DF |
| #undef DIV_DF |
| #undef MOD_DF |
| #undef SRAR_DF |
| #undef PCKEV_DF |
| #undef PCKOD_DF |
| #undef ILVL_DF |
| #undef ILVR_DF |
| #undef ILVEV_DF |
| #undef ILVOD_DF |
| #undef VSHF_DF |
| #undef HADD_DF |
| #undef HSUB_DF |
| } // namespace internal |
| |
| struct TestCaseMsa3RF { |
| uint64_t ws_lo; |
| uint64_t ws_hi; |
| uint64_t wt_lo; |
| uint64_t wt_hi; |
| uint64_t wd_lo; |
| uint64_t wd_hi; |
| }; |
| |
| struct ExpectedResult_MSA3RF { |
| uint64_t exp_res_lo; |
| uint64_t exp_res_hi; |
| }; |
| |
| template <typename Func> |
| void run_msa_3rf(const struct TestCaseMsa3RF* input, |
| const struct ExpectedResult_MSA3RF* output, |
| Func Generate2RInstructionFunc) { |
| Isolate* isolate = CcTest::i_isolate(); |
| HandleScope scope(isolate); |
| |
| MacroAssembler assm(isolate, v8::internal::CodeObjectRequired::kYes); |
| CpuFeatureScope fscope(&assm, MIPS_SIMD); |
| msa_reg_t res; |
| |
| load_elements_of_vector( |
| assm, reinterpret_cast<const uint64_t*>(&input->ws_lo), w0, t0, t1); |
| load_elements_of_vector( |
| assm, reinterpret_cast<const uint64_t*>(&input->wt_lo), w1, t0, t1); |
| load_elements_of_vector( |
| assm, reinterpret_cast<const uint64_t*>(&input->wd_lo), w2, t0, t1); |
| Generate2RInstructionFunc(assm); |
| store_elements_of_vector(assm, w2, a0); |
| |
| __ jr(ra); |
| __ nop(); |
| |
| CodeDesc desc; |
| assm.GetCode(isolate, &desc); |
| Handle<Code> code = Factory::CodeBuilder(isolate, desc, Code::STUB).Build(); |
| #ifdef OBJECT_PRINT |
| code->Print(std::cout); |
| #endif |
| auto f = GeneratedCode<F3>::FromCode(*code); |
| |
| (f.Call(&res, 0, 0, 0, 0)); |
| |
| CHECK_EQ(output->exp_res_lo, res.d[0]); |
| CHECK_EQ(output->exp_res_hi, res.d[1]); |
| } |
| |
| struct TestCaseMsa3RF_F { |
| float ws_1, ws_2, ws_3, ws_4; |
| float wt_1, wt_2, wt_3, wt_4; |
| float wd_1, wd_2, wd_3, wd_4; |
| }; |
| struct ExpRes_32I { |
| int32_t exp_res_1; |
| int32_t exp_res_2; |
| int32_t exp_res_3; |
| int32_t exp_res_4; |
| }; |
| |
| struct TestCaseMsa3RF_D { |
| double ws_lo, ws_hi; |
| double wt_lo, wt_hi; |
| double wd_lo, wd_hi; |
| }; |
| struct ExpRes_64I { |
| int64_t exp_res_lo; |
| int64_t exp_res_hi; |
| }; |
| |
| TEST(MSA_floating_point_quiet_compare) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| const float qnan_f = std::numeric_limits<float>::quiet_NaN(); |
| const double qnan_d = std::numeric_limits<double>::quiet_NaN(); |
| const float inf_f = std::numeric_limits<float>::infinity(); |
| const double inf_d = std::numeric_limits<double>::infinity(); |
| const int32_t ones = -1; |
| |
| const struct TestCaseMsa3RF_F tc_w[]{ |
| {qnan_f, -qnan_f, inf_f, 2.14e9f, // ws |
| qnan_f, 0.f, qnan_f, -2.14e9f, // wt |
| 0, 0, 0, 0}, // wd |
| {inf_f, -inf_f, -3.4e38f, 1.5e-45f, -inf_f, -inf_f, -inf_f, inf_f, 0, 0, |
| 0, 0}, |
| {0.f, 19.871e24f, -1.5e-45f, -1.5e-45f, -19.871e24f, 19.871e24f, 1.5e-45f, |
| -1.5e-45f, 0, 0, 0, 0}}; |
| |
| const struct TestCaseMsa3RF_D tc_d[]{ |
| // ws_lo, ws_hi, wt_lo, wt_hi, wd_lo, wd_hi |
| {qnan_d, -qnan_d, qnan_f, 0., 0, 0}, |
| {inf_d, 9.22e18, qnan_d, -9.22e18, 0, 0}, |
| {inf_d, inf_d, -inf_d, inf_d, 0, 0}, |
| {-2.3e-308, 5e-324, -inf_d, inf_d, 0, 0}, |
| {0., 24.1e87, -1.6e308, 24.1e87, 0, 0}, |
| {-5e-324, -5e-324, 5e-324, -5e-324, 0, 0}}; |
| |
| const struct ExpectedResult_MSA3RF exp_res_fcaf = {0, 0}; |
| const struct ExpRes_32I exp_res_fcun_w[] = { |
| {ones, ones, ones, 0}, {0, 0, 0, 0}, {0, 0, 0, 0}}; |
| const struct ExpRes_64I exp_res_fcun_d[] = {{ones, ones}, {ones, 0}, {0, 0}, |
| {0, 0}, {0, 0}, {0, 0}}; |
| const struct ExpRes_32I exp_res_fceq_w[] = { |
| {0, 0, 0, 0}, {0, ones, 0, 0}, {0, ones, 0, ones}}; |
| const struct ExpRes_64I exp_res_fceq_d[] = {{0, 0}, {0, 0}, {0, ones}, |
| {0, 0}, {0, ones}, {0, ones}}; |
| const struct ExpRes_32I exp_res_fcueq_w[] = { |
| {ones, ones, ones, 0}, {0, ones, 0, 0}, {0, ones, 0, ones}}; |
| const struct ExpRes_64I exp_res_fcueq_d[] = { |
| {ones, ones}, {ones, 0}, {0, ones}, {0, 0}, {0, ones}, {0, ones}}; |
| const struct ExpRes_32I exp_res_fclt_w[] = { |
| {0, 0, 0, 0}, {0, 0, 0, ones}, {0, 0, ones, 0}}; |
| const struct ExpRes_64I exp_res_fclt_d[] = {{0, 0}, {0, 0}, {0, 0}, |
| {0, ones}, {0, 0}, {ones, 0}}; |
| const struct ExpRes_32I exp_res_fcult_w[] = { |
| {ones, ones, ones, 0}, {0, 0, 0, ones}, {0, 0, ones, 0}}; |
| const struct ExpRes_64I exp_res_fcult_d[] = { |
| {ones, ones}, {ones, 0}, {0, 0}, {0, ones}, {0, 0}, {ones, 0}}; |
| const struct ExpRes_32I exp_res_fcle_w[] = { |
| {0, 0, 0, 0}, {0, ones, 0, ones}, {0, ones, ones, ones}}; |
| const struct ExpRes_64I exp_res_fcle_d[] = { |
| {0, 0}, {0, 0}, {0, ones}, {0, ones}, {0, ones}, {ones, ones}}; |
| const struct ExpRes_32I exp_res_fcule_w[] = { |
| {ones, ones, ones, 0}, {0, ones, 0, ones}, {0, ones, ones, ones}}; |
| const struct ExpRes_64I exp_res_fcule_d[] = { |
| {ones, ones}, {ones, 0}, {0, ones}, {0, ones}, {0, ones}, {ones, ones}}; |
| const struct ExpRes_32I exp_res_fcor_w[] = { |
| {0, 0, 0, ones}, {ones, ones, ones, ones}, {ones, ones, ones, ones}}; |
| const struct ExpRes_64I exp_res_fcor_d[] = {{0, 0}, {0, ones}, |
| {ones, ones}, {ones, ones}, |
| {ones, ones}, {ones, ones}}; |
| const struct ExpRes_32I exp_res_fcune_w[] = { |
| {ones, ones, ones, ones}, {ones, 0, ones, ones}, {ones, 0, ones, 0}}; |
| const struct ExpRes_64I exp_res_fcune_d[] = {{ones, ones}, {ones, ones}, |
| {ones, 0}, {ones, ones}, |
| {ones, 0}, {ones, 0}}; |
| const struct ExpRes_32I exp_res_fcne_w[] = { |
| {0, 0, 0, ones}, {ones, 0, ones, ones}, {ones, 0, ones, 0}}; |
| const struct ExpRes_64I exp_res_fcne_d[] = { |
| {0, 0}, {0, ones}, {ones, 0}, {ones, ones}, {ones, 0}, {ones, 0}}; |
| |
| #define TEST_FP_QUIET_COMPARE_W(instruction, src, exp_res) \ |
| run_msa_3rf(reinterpret_cast<const struct TestCaseMsa3RF*>(src), \ |
| reinterpret_cast<const struct ExpectedResult_MSA3RF*>(exp_res), \ |
| [](MacroAssembler& assm) { __ instruction(w2, w0, w1); }); |
| #define TEST_FP_QUIET_COMPARE_D(instruction, src, exp_res) \ |
| run_msa_3rf(reinterpret_cast<const struct TestCaseMsa3RF*>(src), \ |
| reinterpret_cast<const struct ExpectedResult_MSA3RF*>(exp_res), \ |
| [](MacroAssembler& assm) { __ instruction(w2, w0, w1); }); |
| |
| for (uint64_t i = 0; i < arraysize(tc_w); i++) { |
| TEST_FP_QUIET_COMPARE_W(fcaf_w, &tc_w[i], &exp_res_fcaf) |
| TEST_FP_QUIET_COMPARE_W(fcun_w, &tc_w[i], &exp_res_fcun_w[i]) |
| TEST_FP_QUIET_COMPARE_W(fceq_w, &tc_w[i], &exp_res_fceq_w[i]) |
| TEST_FP_QUIET_COMPARE_W(fcueq_w, &tc_w[i], &exp_res_fcueq_w[i]) |
| TEST_FP_QUIET_COMPARE_W(fclt_w, &tc_w[i], &exp_res_fclt_w[i]) |
| TEST_FP_QUIET_COMPARE_W(fcult_w, &tc_w[i], &exp_res_fcult_w[i]) |
| TEST_FP_QUIET_COMPARE_W(fcle_w, &tc_w[i], &exp_res_fcle_w[i]) |
| TEST_FP_QUIET_COMPARE_W(fcule_w, &tc_w[i], &exp_res_fcule_w[i]) |
| TEST_FP_QUIET_COMPARE_W(fcor_w, &tc_w[i], &exp_res_fcor_w[i]) |
| TEST_FP_QUIET_COMPARE_W(fcune_w, &tc_w[i], &exp_res_fcune_w[i]) |
| TEST_FP_QUIET_COMPARE_W(fcne_w, &tc_w[i], &exp_res_fcne_w[i]) |
| } |
| for (uint64_t i = 0; i < arraysize(tc_d); i++) { |
| TEST_FP_QUIET_COMPARE_D(fcaf_d, &tc_d[i], &exp_res_fcaf) |
| TEST_FP_QUIET_COMPARE_D(fcun_d, &tc_d[i], &exp_res_fcun_d[i]) |
| TEST_FP_QUIET_COMPARE_D(fceq_d, &tc_d[i], &exp_res_fceq_d[i]) |
| TEST_FP_QUIET_COMPARE_D(fcueq_d, &tc_d[i], &exp_res_fcueq_d[i]) |
| TEST_FP_QUIET_COMPARE_D(fclt_d, &tc_d[i], &exp_res_fclt_d[i]) |
| TEST_FP_QUIET_COMPARE_D(fcult_d, &tc_d[i], &exp_res_fcult_d[i]) |
| TEST_FP_QUIET_COMPARE_D(fcle_d, &tc_d[i], &exp_res_fcle_d[i]) |
| TEST_FP_QUIET_COMPARE_D(fcule_d, &tc_d[i], &exp_res_fcule_d[i]) |
| TEST_FP_QUIET_COMPARE_D(fcor_d, &tc_d[i], &exp_res_fcor_d[i]) |
| TEST_FP_QUIET_COMPARE_D(fcune_d, &tc_d[i], &exp_res_fcune_d[i]) |
| TEST_FP_QUIET_COMPARE_D(fcne_d, &tc_d[i], &exp_res_fcne_d[i]) |
| } |
| #undef TEST_FP_QUIET_COMPARE_W |
| #undef TEST_FP_QUIET_COMPARE_D |
| } |
| |
| template <typename T> |
| inline const T* fadd_function(const T* src1, const T* src2, const T* src3, |
| T* dst) { |
| for (uint64_t i = 0; i < kMSALanesByte / sizeof(T); i++) { |
| dst[i] = src1[i] + src2[i]; |
| } |
| return dst; |
| } |
| template <typename T> |
| inline const T* fsub_function(const T* src1, const T* src2, const T* src3, |
| T* dst) { |
| for (uint64_t i = 0; i < kMSALanesByte / sizeof(T); i++) { |
| dst[i] = src1[i] - src2[i]; |
| } |
| return dst; |
| } |
| template <typename T> |
| inline const T* fmul_function(const T* src1, const T* src2, const T* src3, |
| T* dst) { |
| for (uint64_t i = 0; i < kMSALanesByte / sizeof(T); i++) { |
| dst[i] = src1[i] * src2[i]; |
| } |
| return dst; |
| } |
| template <typename T> |
| inline const T* fdiv_function(const T* src1, const T* src2, const T* src3, |
| T* dst) { |
| for (uint64_t i = 0; i < kMSALanesByte / sizeof(T); i++) { |
| dst[i] = src1[i] / src2[i]; |
| } |
| return dst; |
| } |
| template <typename T> |
| inline const T* fmadd_function(const T* src1, const T* src2, const T* src3, |
| T* dst) { |
| for (uint64_t i = 0; i < kMSALanesByte / sizeof(T); i++) { |
| dst[i] = std::fma(src1[i], src2[i], src3[i]); |
| } |
| return dst; |
| } |
| template <typename T> |
| inline const T* fmsub_function(const T* src1, const T* src2, const T* src3, |
| T* dst) { |
| for (uint64_t i = 0; i < kMSALanesByte / sizeof(T); i++) { |
| dst[i] = std::fma(src1[i], -src2[i], src3[i]); |
| } |
| return dst; |
| } |
| |
| TEST(MSA_floating_point_arithmetic) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| const float inf_f = std::numeric_limits<float>::infinity(); |
| const double inf_d = std::numeric_limits<double>::infinity(); |
| |
| const struct TestCaseMsa3RF_F tc_w[] = { |
| {0.3, -2.14e13f, inf_f, 0.f, // ws |
| -inf_f, std::sqrt(8.e-26f), -23.e34, -2.14e9f, // wt |
| -1e30f, 4.6e12f, 0, 2.14e9f}, // wd |
| {3.4e38f, -1.2e-38f, 1e19f, -1e19f, 3.4e38f, 1.2e-38f, -1e19f, -1e-19f, |
| 3.4e38f, 1.2e-38f * 3, 3.4e38f, -4e19f}, |
| {-3e-31f, 3e10f, 1e25f, 123.f, 1e-14f, 1e-34f, 4e25f, 321.f, 3e-17f, |
| 2e-24f, 2.f, -123456.f}}; |
| |
| const struct TestCaseMsa3RF_D tc_d[] = { |
| // ws_lo, ws_hi, wt_lo, wt_hi, wd_lo, wd_hi |
| {0.3, -2.14e103, -inf_d, std::sqrt(8.e-206), -1e30, 4.6e102}, |
| {inf_d, 0., -23.e304, -2.104e9, 0, 2.104e9}, |
| {3.4e307, -1.2e-307, 3.4e307, 1.2e-307, 3.4e307, 1.2e-307 * 3}, |
| {1e154, -1e154, -1e154, -1e-154, 2.9e38, -4e19}, |
| {-3e-301, 3e100, 1e-104, 1e-304, 3e-107, 2e-204}, |
| {1e205, 123., 4e205, 321., 2., -123456.}}; |
| |
| struct ExpectedResult_MSA3RF dst_container; |
| |
| #define FP_ARITHMETIC_DF_W(instr, function, src1, src2, src3) \ |
| run_msa_3rf( \ |
| reinterpret_cast<const struct TestCaseMsa3RF*>(src1), \ |
| reinterpret_cast<const struct ExpectedResult_MSA3RF*>(function( \ |
| src1, src2, src3, reinterpret_cast<float*>(&dst_container))), \ |
| [](MacroAssembler& assm) { __ instr(w2, w0, w1); }); |
| |
| #define FP_ARITHMETIC_DF_D(instr, function, src1, src2, src3) \ |
| run_msa_3rf( \ |
| reinterpret_cast<const struct TestCaseMsa3RF*>(src1), \ |
| reinterpret_cast<const struct ExpectedResult_MSA3RF*>(function( \ |
| src1, src2, src3, reinterpret_cast<double*>(&dst_container))), \ |
| [](MacroAssembler& assm) { __ instr(w2, w0, w1); }); |
| |
| for (uint64_t i = 0; i < arraysize(tc_w); i++) { |
| FP_ARITHMETIC_DF_W(fadd_w, fadd_function, &tc_w[i].ws_1, &tc_w[i].wt_1, |
| &tc_w[i].wd_1) |
| FP_ARITHMETIC_DF_W(fsub_w, fsub_function, &tc_w[i].ws_1, &tc_w[i].wt_1, |
| &tc_w[i].wd_1) |
| FP_ARITHMETIC_DF_W(fmul_w, fmul_function, &tc_w[i].ws_1, &tc_w[i].wt_1, |
| &tc_w[i].wd_1) |
| FP_ARITHMETIC_DF_W(fdiv_w, fdiv_function, &tc_w[i].ws_1, &tc_w[i].wt_1, |
| &tc_w[i].wd_1) |
| FP_ARITHMETIC_DF_W(fmadd_w, fmadd_function, &tc_w[i].ws_1, &tc_w[i].wt_1, |
| &tc_w[i].wd_1) |
| FP_ARITHMETIC_DF_W(fmsub_w, fmsub_function, &tc_w[i].ws_1, &tc_w[i].wt_1, |
| &tc_w[i].wd_1) |
| } |
| for (uint64_t i = 0; i < arraysize(tc_d); i++) { |
| FP_ARITHMETIC_DF_D(fadd_d, fadd_function, &tc_d[i].ws_lo, &tc_d[i].wt_lo, |
| &tc_d[i].wd_lo) |
| FP_ARITHMETIC_DF_D(fsub_d, fsub_function, &tc_d[i].ws_lo, &tc_d[i].wt_lo, |
| &tc_d[i].wd_lo) |
| FP_ARITHMETIC_DF_D(fmul_d, fmul_function, &tc_d[i].ws_lo, &tc_d[i].wt_lo, |
| &tc_d[i].wd_lo) |
| FP_ARITHMETIC_DF_D(fdiv_d, fdiv_function, &tc_d[i].ws_lo, &tc_d[i].wt_lo, |
| &tc_d[i].wd_lo) |
| FP_ARITHMETIC_DF_D(fmadd_d, fmadd_function, &tc_d[i].ws_lo, &tc_d[i].wt_lo, |
| &tc_d[i].wd_lo) |
| FP_ARITHMETIC_DF_D(fmsub_d, fmsub_function, &tc_d[i].ws_lo, &tc_d[i].wt_lo, |
| &tc_d[i].wd_lo) |
| } |
| #undef FP_ARITHMETIC_DF_W |
| #undef FP_ARITHMETIC_DF_D |
| } |
| |
| struct ExpRes_F { |
| float exp_res_1; |
| float exp_res_2; |
| float exp_res_3; |
| float exp_res_4; |
| }; |
| |
| struct ExpRes_D { |
| double exp_res_1; |
| double exp_res_2; |
| }; |
| |
| TEST(MSA_fmin_fmin_a_fmax_fmax_a) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| const float inf_f = std::numeric_limits<float>::infinity(); |
| const double inf_d = std::numeric_limits<double>::infinity(); |
| |
| const struct TestCaseMsa3RF_F tc_w[] = { |
| {0.3f, -2.14e13f, inf_f, -0.f, // ws |
| -inf_f, -std::sqrt(8.e26f), -23.e34f, -2.14e9f, // wt |
| 0, 0, 0, 0}, // wd |
| {3.4e38f, 1.2e-41f, 1e19f, 1e19f, // ws |
| 3.4e38f, -1.1e-41f, -1e-42f, -1e29f, // wt |
| 0, 0, 0, 0}}; // wd |
| |
| const struct TestCaseMsa3RF_D tc_d[] = { |
| // ws_lo, ws_hi, wt_lo, wt_hi, wd_lo, wd_hi |
| {0.3, -2.14e103, -inf_d, -std::sqrt(8e206), 0, 0}, |
| {inf_d, -0., -23e304, -2.14e90, 0, 0}, |
| {3.4e307, 1.2e-320, 3.4e307, -1.1e-320, 0, 0}, |
| {1e154, 1e154, -1e-321, -1e174, 0, 0}}; |
| |
| const struct ExpRes_F exp_res_fmax_w[] = {{0.3f, -2.14e13f, inf_f, -0.f}, |
| {3.4e38f, 1.2e-41f, 1e19f, 1e19f}}; |
| const struct ExpRes_F exp_res_fmax_a_w[] = { |
| {-inf_f, -std::sqrt(8e26f), inf_f, -2.14e9f}, |
| {3.4e38f, 1.2e-41f, 1e19f, -1e29f}}; |
| const struct ExpRes_F exp_res_fmin_w[] = { |
| {-inf_f, -std::sqrt(8.e26f), -23e34f, -2.14e9f}, |
| {3.4e38f, -1.1e-41f, -1e-42f, -1e29f}}; |
| const struct ExpRes_F exp_res_fmin_a_w[] = { |
| {0.3, -2.14e13f, -23.e34f, -0.f}, {3.4e38f, -1.1e-41f, -1e-42f, 1e19f}}; |
| |
| const struct ExpRes_D exp_res_fmax_d[] = { |
| {0.3, -2.14e103}, {inf_d, -0.}, {3.4e307, 1.2e-320}, {1e154, 1e154}}; |
| const struct ExpRes_D exp_res_fmax_a_d[] = {{-inf_d, -std::sqrt(8e206)}, |
| {inf_d, -2.14e90}, |
| {3.4e307, 1.2e-320}, |
| {1e154, -1e174}}; |
| const struct ExpRes_D exp_res_fmin_d[] = {{-inf_d, -std::sqrt(8e206)}, |
| {-23e304, -2.14e90}, |
| {3.4e307, -1.1e-320}, |
| {-1e-321, -1e174}}; |
| const struct ExpRes_D exp_res_fmin_a_d[] = { |
| {0.3, -2.14e103}, {-23e304, -0.}, {3.4e307, -1.1e-320}, {-1e-321, 1e154}}; |
| |
| #define TEST_FP_MIN_MAX_W(instruction, src, exp_res) \ |
| run_msa_3rf(reinterpret_cast<const struct TestCaseMsa3RF*>(src), \ |
| reinterpret_cast<const struct ExpectedResult_MSA3RF*>(exp_res), \ |
| [](MacroAssembler& assm) { __ instruction(w2, w0, w1); }); |
| |
| #define TEST_FP_MIN_MAX_D(instruction, src, exp_res) \ |
| run_msa_3rf(reinterpret_cast<const struct TestCaseMsa3RF*>(src), \ |
| reinterpret_cast<const struct ExpectedResult_MSA3RF*>(exp_res), \ |
| [](MacroAssembler& assm) { __ instruction(w2, w0, w1); }); |
| |
| for (uint64_t i = 0; i < arraysize(tc_w); i++) { |
| TEST_FP_MIN_MAX_W(fmax_w, &tc_w[i], &exp_res_fmax_w[i]) |
| TEST_FP_MIN_MAX_W(fmax_a_w, &tc_w[i], &exp_res_fmax_a_w[i]) |
| TEST_FP_MIN_MAX_W(fmin_w, &tc_w[i], &exp_res_fmin_w[i]) |
| TEST_FP_MIN_MAX_W(fmin_a_w, &tc_w[i], &exp_res_fmin_a_w[i]) |
| } |
| |
| for (uint64_t i = 0; i < arraysize(tc_d); i++) { |
| TEST_FP_MIN_MAX_D(fmax_d, &tc_d[i], &exp_res_fmax_d[i]) |
| TEST_FP_MIN_MAX_D(fmax_a_d, &tc_d[i], &exp_res_fmax_a_d[i]) |
| TEST_FP_MIN_MAX_D(fmin_d, &tc_d[i], &exp_res_fmin_d[i]) |
| TEST_FP_MIN_MAX_D(fmin_a_d, &tc_d[i], &exp_res_fmin_a_d[i]) |
| } |
| #undef TEST_FP_MIN_MAX_W |
| #undef TEST_FP_MIN_MAX_D |
| } |
| |
| struct TestCaseMsa3RF_16I { |
| int16_t ws_1, ws_2, ws_3, ws_4, ws_5, ws_6, ws_7, ws_8; |
| int16_t wt_1, wt_2, wt_3, wt_4, wt_5, wt_6, wt_7, wt_8; |
| int16_t wd_1, wd_2, wd_3, wd_4, wd_5, wd_6, wd_7, wd_8; |
| }; |
| struct ExpRes_16I { |
| int16_t exp_res_1; |
| int16_t exp_res_2; |
| int16_t exp_res_3; |
| int16_t exp_res_4; |
| int16_t exp_res_5; |
| int16_t exp_res_6; |
| int16_t exp_res_7; |
| int16_t exp_res_8; |
| }; |
| |
| struct TestCaseMsa3RF_32I { |
| int32_t ws_1, ws_2, ws_3, ws_4; |
| int32_t wt_1, wt_2, wt_3, wt_4; |
| int32_t wd_1, wd_2, wd_3, wd_4; |
| }; |
| |
| TEST(MSA_fixed_point_arithmetic) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| const struct TestCaseMsa3RF tc_h[]{ |
| {0x800080007FFF7FFF, 0xE1ED8000FAD3863A, 0x80007FFF00AF7FFF, |
| 0x800015A77FFFA0EB, 0x7FFF800080007FFF, 0x80007FFF1F207364}, |
| {0x800080007FFF006A, 0x002AFFC4329AD87B, 0x80007FFF7FFF00F3, |
| 0xFFECFFB4D0D7F429, 0x80007FFF80007C33, 0x54AC6BBCE53B8C91}}; |
| |
| const struct TestCaseMsa3RF tc_w[]{ |
| {0x8000000080000000, 0x7FFFFFFF7FFFFFFF, 0x800000007FFFFFFF, |
| 0x00001FF37FFFFFFF, 0x7FFFFFFF80000000, 0x800000007FFFFFFF}, |
| {0xE1ED035580000000, 0xFAD3863AED462C0B, 0x8000000015A70AEC, |
| 0x7FFFFFFFA0EBD354, 0x800000007FFFFFFF, 0xD0D7F4291F207364}, |
| {0x8000000080000000, 0x7FFFFFFF0000DA1F, 0x800000007FFFFFFF, |
| 0x7FFFFFFF00F39C3B, 0x800000007FFFFFFF, 0x800000007C33F2FD}, |
| {0x0000AC33FFFF329A, 0x54AC6BBCE53BD87B, 0xFFFFE2B4D0D7F429, |
| 0x0355ED462C0B1FF3, 0xB5DEB625939DD3F9, 0xE642ADFA69519596}}; |
| |
| const struct ExpectedResult_MSA3RF exp_res_mul_q_h[] = { |
| {0x7FFF800100AE7FFE, 0x1E13EA59FAD35A74}, |
| {0x7FFF80017FFE0000, 0xFFFF0000ED5B03A7}}; |
| const struct ExpectedResult_MSA3RF exp_res_madd_q_h[] = { |
| {0x7FFF800080AE7FFF, 0x9E136A5819F37FFF}, |
| {0x00000000FFFE7C33, 0x54AB6BBCD2969038}}; |
| const struct ExpectedResult_MSA3RF exp_res_msub_q_h[] = { |
| {0xFFFFFFFF80000000, 0x80007FFF244C18EF}, |
| {0x80007FFF80007C32, 0x54AC6BBBF7DF88E9}}; |
| const struct ExpectedResult_MSA3RF exp_res_mulr_q_h[] = { |
| {0x7FFF800100AF7FFE, 0x1E13EA59FAD35A75}, |
| {0x7FFF80017FFE0001, 0x00000000ED5B03A8}}; |
| const struct ExpectedResult_MSA3RF exp_res_maddr_q_h[] = { |
| {0x7FFF800080AF7FFF, 0x9E136A5819F37FFF}, |
| {0x00000000FFFE7C34, 0x54AC6BBCD2969039}}; |
| const struct ExpectedResult_MSA3RF exp_res_msubr_q_h[] = { |
| {0xFFFFFFFF80000001, 0x80007FFF244D18EF}, |
| {0x80007FFF80007C32, 0x54AC6BBCF7E088E9}}; |
| |
| const struct ExpectedResult_MSA3RF exp_res_mul_q_w[] = { |
| {0x7FFFFFFF80000001, 0x00001FF27FFFFFFE}, |
| {0x1E12FCABEA58F514, 0xFAD3863A0DE8DEE1}, |
| {0x7FFFFFFF80000001, 0x7FFFFFFE0000019F}, |
| {0xFFFFFFFF00004BAB, 0x0234E1FBF6CA3EE0}}; |
| const struct ExpectedResult_MSA3RF exp_res_madd_q_w[] = { |
| {0x7FFFFFFF80000000, 0x80001FF27FFFFFFF}, |
| {0x9E12FCAB6A58F513, 0xCBAB7A632D095245}, |
| {0x0000000000000000, 0xFFFFFFFE7C33F49C}, |
| {0xB5DEB624939E1FA4, 0xE8778FF5601BD476}}; |
| const struct ExpectedResult_MSA3RF exp_res_msub_q_w[] = { |
| {0xFFFFFFFFFFFFFFFF, 0x8000000000000000}, |
| {0x800000007FFFFFFF, 0xD6046DEE11379482}, |
| {0x800000007FFFFFFF, 0x800000007C33F15D}, |
| {0xB5DEB625939D884D, 0xE40DCBFE728756B5}}; |
| const struct ExpectedResult_MSA3RF exp_res_mulr_q_w[] = { |
| {0x7FFFFFFF80000001, 0x00001FF37FFFFFFE}, |
| {0x1E12FCABEA58F514, 0xFAD3863A0DE8DEE2}, |
| {0x7FFFFFFF80000001, 0x7FFFFFFE0000019F}, |
| {0x0000000000004BAC, 0x0234E1FCF6CA3EE1}}; |
| const struct ExpectedResult_MSA3RF exp_res_maddr_q_w[] = { |
| {0x7FFFFFFF80000000, 0x80001FF37FFFFFFF}, |
| {0x9E12FCAB6A58F513, 0xCBAB7A632D095246}, |
| {0x0000000000000000, 0xFFFFFFFE7C33F49C}, |
| {0xB5DEB625939E1FA5, 0xE8778FF6601BD477}}; |
| const struct ExpectedResult_MSA3RF exp_res_msubr_q_w[] = { |
| {0xFFFFFFFFFFFFFFFF, 0x8000000000000001}, |
| {0x800000007FFFFFFF, 0xD6046DEF11379482}, |
| {0x800000007FFFFFFF, 0x800000007C33F15E}, |
| {0xB5DEB625939D884D, 0xE40DCBFE728756B5}}; |
| |
| #define TEST_FIXED_POINT_DF_H(instruction, src, exp_res) \ |
| run_msa_3rf((src), (exp_res), \ |
| [](MacroAssembler& assm) { __ instruction(w2, w0, w1); }); |
| |
| #define TEST_FIXED_POINT_DF_W(instruction, src, exp_res) \ |
| run_msa_3rf((src), (exp_res), \ |
| [](MacroAssembler& assm) { __ instruction(w2, w0, w1); }); |
| |
| for (uint64_t i = 0; i < arraysize(tc_h); i++) { |
| TEST_FIXED_POINT_DF_H(mul_q_h, &tc_h[i], &exp_res_mul_q_h[i]) |
| TEST_FIXED_POINT_DF_H(madd_q_h, &tc_h[i], &exp_res_madd_q_h[i]) |
| TEST_FIXED_POINT_DF_H(msub_q_h, &tc_h[i], &exp_res_msub_q_h[i]) |
| TEST_FIXED_POINT_DF_H(mulr_q_h, &tc_h[i], &exp_res_mulr_q_h[i]) |
| TEST_FIXED_POINT_DF_H(maddr_q_h, &tc_h[i], &exp_res_maddr_q_h[i]) |
| TEST_FIXED_POINT_DF_H(msubr_q_h, &tc_h[i], &exp_res_msubr_q_h[i]) |
| } |
| |
| for (uint64_t i = 0; i < arraysize(tc_w); i++) { |
| TEST_FIXED_POINT_DF_W(mul_q_w, &tc_w[i], &exp_res_mul_q_w[i]) |
| TEST_FIXED_POINT_DF_W(madd_q_w, &tc_w[i], &exp_res_madd_q_w[i]) |
| TEST_FIXED_POINT_DF_W(msub_q_w, &tc_w[i], &exp_res_msub_q_w[i]) |
| TEST_FIXED_POINT_DF_W(mulr_q_w, &tc_w[i], &exp_res_mulr_q_w[i]) |
| TEST_FIXED_POINT_DF_W(maddr_q_w, &tc_w[i], &exp_res_maddr_q_w[i]) |
| TEST_FIXED_POINT_DF_W(msubr_q_w, &tc_w[i], &exp_res_msubr_q_w[i]) |
| } |
| #undef TEST_FIXED_POINT_DF_H |
| #undef TEST_FIXED_POINT_DF_W |
| } |
| |
| TEST(MSA_fexdo) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| const float inf_float = std::numeric_limits<float>::infinity(); |
| const float nan_float = std::numeric_limits<float>::quiet_NaN(); |
| const double inf_double = std::numeric_limits<double>::infinity(); |
| |
| const struct TestCaseMsa3RF_F tc_w[] = { |
| // ws_1, ws_2, ws_3, ws_4, wt_1, wt_2, wt_3, wt_4, wd_1, wd_2, wd_3, wd_4 |
| {inf_float, nan_float, 66505.f, 65504.f, 6.2e-5f, 5e-5f, -32.42f, |
| -inf_float, 0, 0, 0, 0}, |
| {-0.f, 0.f, 123.567f, -765.321f, -6e-8f, 5.9e-8f, 1e-7f, -1e-20f, 0, 0, 0, |
| 0}, |
| {1e-36f, 1e20f, -1e20f, 2e-20f, 6e-8f, -2.9e-8f, -66505.f, -65504.f, 0, 0, |
| 0, 0}}; |
| |
| const struct TestCaseMsa3RF_D tc_d[] = { |
| // ws_lo, ws_hi, wt_lo, wt_hi, wd_lo, wd_hi |
| {inf_double, -1234., 4e38, 3.4e38, 0, 0}, |
| {1.2e-38, 1.1e-39, -38.92f, -inf_double, 0, 0}, |
| {-0., 0., 123.567e31, -765.321e33, 0, 0}, |
| {-1.5e-45, 1.3e-45, 1e-42, -1e-200, 0, 0}, |
| {1e-202, 1e158, -1e159, 1e14, 0, 0}, |
| {1.5e-42, 1.3e-46, -123.567e31, 765.321e33, 0, 0}}; |
| |
| const struct ExpRes_16I exp_res_fexdo_w[] = { |
| {static_cast<int16_t>(0x0410), static_cast<int16_t>(0x0347), |
| static_cast<int16_t>(0xD00D), static_cast<int16_t>(0xFC00), |
| static_cast<int16_t>(0x7C00), static_cast<int16_t>(0x7DFF), |
| static_cast<int16_t>(0x7C00), static_cast<int16_t>(0x7BFF)}, |
| {static_cast<int16_t>(0x8001), static_cast<int16_t>(0x0001), |
| static_cast<int16_t>(0x0002), static_cast<int16_t>(0x8000), |
| static_cast<int16_t>(0x8000), static_cast<int16_t>(0x0000), |
| static_cast<int16_t>(0x57B9), static_cast<int16_t>(0xE1FB)}, |
| {static_cast<int16_t>(0x0001), static_cast<int16_t>(0x8000), |
| static_cast<int16_t>(0xFC00), static_cast<int16_t>(0xFBFF), |
| static_cast<int16_t>(0x0000), static_cast<int16_t>(0x7C00), |
| static_cast<int16_t>(0xFC00), static_cast<int16_t>(0x0000)}}; |
| |
| const struct ExpRes_32I exp_res_fexdo_d[] = { |
| {bit_cast<int32_t>(0x7F800000), bit_cast<int32_t>(0x7F7FC99E), |
| bit_cast<int32_t>(0x7F800000), bit_cast<int32_t>(0xC49A4000)}, |
| {bit_cast<int32_t>(0xC21BAE14), bit_cast<int32_t>(0xFF800000), |
| bit_cast<int32_t>(0x0082AB1E), bit_cast<int32_t>(0x000BFA5A)}, |
| {bit_cast<int32_t>(0x7673B164), bit_cast<int32_t>(0xFB13653D), |
| bit_cast<int32_t>(0x80000000), bit_cast<int32_t>(0x00000000)}, |
| {bit_cast<int32_t>(0x000002CA), bit_cast<int32_t>(0x80000000), |
| bit_cast<int32_t>(0x80000001), bit_cast<int32_t>(0x00000001)}, |
| {bit_cast<int32_t>(0xFF800000), bit_cast<int32_t>(0x56B5E621), |
| bit_cast<int32_t>(0x00000000), bit_cast<int32_t>(0x7F800000)}, |
| {bit_cast<int32_t>(0xF673B164), bit_cast<int32_t>(0x7B13653D), |
| bit_cast<int32_t>(0x0000042E), bit_cast<int32_t>(0x00000000)}}; |
| |
| #define TEST_FEXDO_H(instruction, src, exp_res) \ |
| run_msa_3rf(reinterpret_cast<const struct TestCaseMsa3RF*>(src), \ |
| reinterpret_cast<const struct ExpectedResult_MSA3RF*>(exp_res), \ |
| [](MacroAssembler& assm) { __ instruction(w2, w0, w1); }); |
| |
| #define TEST_FEXDO_W(instruction, src, exp_res) \ |
| run_msa_3rf(reinterpret_cast<const struct TestCaseMsa3RF*>(src), \ |
| reinterpret_cast<const struct ExpectedResult_MSA3RF*>(exp_res), \ |
| [](MacroAssembler& assm) { __ instruction(w2, w0, w1); }); |
| |
| for (uint64_t i = 0; i < arraysize(tc_w); i++) { |
| TEST_FEXDO_H(fexdo_h, &tc_w[i], &exp_res_fexdo_w[i]) |
| } |
| |
| for (uint64_t i = 0; i < arraysize(tc_d); i++) { |
| TEST_FEXDO_W(fexdo_w, &tc_d[i], &exp_res_fexdo_d[i]) |
| } |
| |
| #undef TEST_FEXDO_H |
| #undef TEST_FEXDO_W |
| } |
| |
| TEST(MSA_ftq) { |
| if (!IsMipsArchVariant(kMips32r6) || !CpuFeatures::IsSupported(MIPS_SIMD)) |
| return; |
| |
| CcTest::InitializeVM(); |
| |
| const float nan_float = std::numeric_limits<float>::quiet_NaN(); |
| const float inf_float = std::numeric_limits<float>::infinity(); |
| const double nan_double = std::numeric_limits<double>::quiet_NaN(); |
| const double inf_double = std::numeric_limits<double>::infinity(); |
| |
| const struct TestCaseMsa3RF_F tc_w[] = { |
| {1.f, -0.999f, 1.5f, -31e-6, 1e-7, -0.598, 0.0023, -0.f, 0, 0, 0, 0}, |
| {100.f, -102.f, -1.1f, 1.3f, 0.f, -1.f, 0.9999f, -0.000322, 0, 0, 0, 0}, |
| {nan_float, inf_float, -inf_float, -nan_float, -1e-40, 3e-44, 8.3e36, |
| -0.00003, 0, 0, 0, 0}}; |
| |
| const struct TestCaseMsa3RF_D tc_d[] = { |
| {1., -0.999, 1.5, -31e-6, 0, 0}, |
| {1e-7, -0.598, 0.0023, -0.f, 0, 0}, |
| {100.f, -102.f, -1.1f, 1.3f, 0, 0}, |
| {0.f, -1.f, 0.9999f, -0.000322, 0, 0}, |
| {nan_double, inf_double, -inf_double, -nan_double, 0, 0}, |
| {-3e306, 2e-307, 9e307, 2e-307, 0, 0}}; |
| |
| const struct ExpRes_16I exp_res_ftq_w[] = { |
| {static_cast<int16_t>(0x0000), static_cast<int16_t>(0xB375), |
| static_cast<int16_t>(0x004B), static_cast<int16_t>(0x0000), |
| static_cast<int16_t>(0x7FFF), static_cast<int16_t>(0x8021), |
| static_cast<int16_t>(0x7FFF), static_cast<int16_t>(0xFFFF)}, |
| {static_cast<int16_t>(0x0000), static_cast<int16_t>(0x8000), |
| static_cast<int16_t>(0x7FFD), static_cast<int16_t>(0xFFF5), |
| static_cast<int16_t>(0x7FFF), static_cast<int16_t>(0x8000), |
| static_cast<int16_t>(0x8000), static_cast<int16_t>(0x7FFF)}, |
| {static_cast<int16_t>(0x0000), static_cast<int16_t>(0x0000), |
| static_cast<int16_t>(0x7FFF), static_cast<int16_t>(0xFFFF), |
| static_cast<int16_t>(0x0000), static_cast<int16_t>(0x7FFF), |
| static_cast<int16_t>(0x8000), static_cast<int16_t>(0x0000)}}; |
| |
| const struct ExpRes_32I exp_res_ftq_d[] = { |
| {bit_cast<int32_t>(0x7FFFFFFF), bit_cast<int32_t>(0xFFFEFBF4), |
| bit_cast<int32_t>(0x7FFFFFFF), bit_cast<int32_t>(0x8020C49C)}, |
| {bit_cast<int32_t>(0x004B5DCC), bit_cast<int32_t>(0x00000000), |
| bit_cast<int32_t>(0x000000D7), bit_cast<int32_t>(0xB374BC6A)}, |
| {bit_cast<int32_t>(0x80000000), bit_cast<int32_t>(0x7FFFFFFF), |
| bit_cast<int32_t>(0x7FFFFFFF), bit_cast<int32_t>(0x80000000)}, |
| {bit_cast<int32_t>(0x7FFCB900), bit_cast<int32_t>(0xFFF572DE), |
| bit_cast<int32_t>(0x00000000), bit_cast<int32_t>(0x80000000)}, |
| {bit_cast<int32_t>(0x80000000), bit_cast<int32_t>(0x00000000), |
| bit_cast<int32_t>(0x00000000), bit_cast<int32_t>(0x7FFFFFFF)}, |
| {bit_cast<int32_t>(0x7FFFFFFF), bit_cast<int32_t>(0x00000000), |
| bit_cast<int32_t>(0x80000000), bit_cast<int32_t>(0x00000000)}}; |
| |
| #define TEST_FTQ_H(instruction, src, exp_res) \ |
| run_msa_3rf(reinterpret_cast<const struct TestCaseMsa3RF*>(src), \ |
| reinterpret_cast<const struct ExpectedResult_MSA3RF*>(exp_res), \ |
| [](MacroAssembler& assm) { __ instruction(w2, w0, w1); }); |
| |
| #define TEST_FTQ_W(instruction, src, exp_res) \ |
| run_msa_3rf(reinterpret_cast<const struct TestCaseMsa3RF*>(src), \ |
| reinterpret_cast<const struct ExpectedResult_MSA3RF*>(exp_res), \ |
| [](MacroAssembler& assm) { __ instruction(w2, w0, w1); }); |
| |
| for (uint64_t i = 0; i < arraysize(tc_w); i++) { |
| TEST_FTQ_H(ftq_h, &tc_w[i], &exp_res_ftq_w[i]) |
| } |
| |
| for (uint64_t i = 0; i < arraysize(tc_d); i++) { |
| TEST_FTQ_W(ftq_w, &tc_d[i], &exp_res_ftq_d[i]) |
| } |
| |
| #undef TEST_FTQ_H |
| #undef TEST_FTQ_W |
| } |
| |
| #undef __ |
| |
| } // namespace internal |
| } // namespace v8 |