| /* |
| * Copyright (C) 2008, 2009 Apple Inc. All rights reserved. |
| * Copyright (C) 2008 Cameron Zwarich <cwzwarich@uwaterloo.ca> |
| * Copyright (C) Research In Motion Limited 2010, 2011. All rights reserved. |
| * |
| * Redistribution and use in source and binary forms, with or without |
| * modification, are permitted provided that the following conditions |
| * are met: |
| * |
| * 1. Redistributions of source code must retain the above copyright |
| * notice, this list of conditions and the following disclaimer. |
| * 2. Redistributions in binary form must reproduce the above copyright |
| * notice, this list of conditions and the following disclaimer in the |
| * documentation and/or other materials provided with the distribution. |
| * 3. Neither the name of Apple Computer, Inc. ("Apple") nor the names of |
| * its contributors may be used to endorse or promote products derived |
| * from this software without specific prior written permission. |
| * |
| * THIS SOFTWARE IS PROVIDED BY APPLE AND ITS CONTRIBUTORS "AS IS" AND ANY |
| * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED |
| * WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE |
| * DISCLAIMED. IN NO EVENT SHALL APPLE OR ITS CONTRIBUTORS BE LIABLE FOR ANY |
| * DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES |
| * (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; |
| * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND |
| * ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
| * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF |
| * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| */ |
| |
| #include "config.h" |
| |
| #if ENABLE(JIT) |
| #include "JITStubs.h" |
| |
| #include "CommonSlowPaths.h" |
| #include "Arguments.h" |
| #include "ArrayConstructor.h" |
| #include "CallFrame.h" |
| #include "CodeBlock.h" |
| #include "CodeProfiling.h" |
| #include "DFGOSREntry.h" |
| #include "Debugger.h" |
| #include "ExceptionHelpers.h" |
| #include "GetterSetter.h" |
| #include "Heap.h" |
| #include <wtf/InlineASM.h> |
| #include "JIT.h" |
| #include "JITExceptions.h" |
| #include "JSActivation.h" |
| #include "JSArray.h" |
| #include "JSFunction.h" |
| #include "JSGlobalObjectFunctions.h" |
| #include "JSNameScope.h" |
| #include "JSNotAnObject.h" |
| #include "JSPropertyNameIterator.h" |
| #include "JSString.h" |
| #include "JSWithScope.h" |
| #include "LegacyProfiler.h" |
| #include "NameInstance.h" |
| #include "ObjectPrototype.h" |
| #include "Operations.h" |
| #include "Parser.h" |
| #include "RegExpObject.h" |
| #include "RegExpPrototype.h" |
| #include "Register.h" |
| #include "RepatchBuffer.h" |
| #include "SamplingTool.h" |
| #include "Strong.h" |
| #include <wtf/StdLibExtras.h> |
| #include <stdarg.h> |
| #include <stdio.h> |
| |
| using namespace std; |
| |
| namespace JSC { |
| |
| #if USE(JSVALUE32_64) |
| |
| #if COMPILER(GCC) && CPU(X86) |
| |
| // These ASSERTs remind you that, if you change the layout of JITStackFrame, you |
| // need to change the assembly trampolines below to match. |
| COMPILE_ASSERT(offsetof(struct JITStackFrame, code) % 16 == 0x0, JITStackFrame_maintains_16byte_stack_alignment); |
| COMPILE_ASSERT(offsetof(struct JITStackFrame, savedEBX) == 0x3c, JITStackFrame_stub_argument_space_matches_ctiTrampoline); |
| COMPILE_ASSERT(offsetof(struct JITStackFrame, callFrame) == 0x58, JITStackFrame_callFrame_offset_matches_ctiTrampoline); |
| COMPILE_ASSERT(offsetof(struct JITStackFrame, code) == 0x50, JITStackFrame_code_offset_matches_ctiTrampoline); |
| |
| asm ( |
| ".text\n" |
| ".globl " SYMBOL_STRING(ctiTrampoline) "\n" |
| HIDE_SYMBOL(ctiTrampoline) "\n" |
| SYMBOL_STRING(ctiTrampoline) ":" "\n" |
| "pushl %ebp" "\n" |
| "movl %esp, %ebp" "\n" |
| "pushl %esi" "\n" |
| "pushl %edi" "\n" |
| "pushl %ebx" "\n" |
| "subl $0x3c, %esp" "\n" |
| "movl 0x58(%esp), %edi" "\n" |
| "call *0x50(%esp)" "\n" |
| "addl $0x3c, %esp" "\n" |
| "popl %ebx" "\n" |
| "popl %edi" "\n" |
| "popl %esi" "\n" |
| "popl %ebp" "\n" |
| "ret" "\n" |
| ".globl " SYMBOL_STRING(ctiTrampolineEnd) "\n" |
| HIDE_SYMBOL(ctiTrampolineEnd) "\n" |
| SYMBOL_STRING(ctiTrampolineEnd) ":" "\n" |
| ); |
| |
| asm ( |
| ".globl " SYMBOL_STRING(ctiVMThrowTrampoline) "\n" |
| HIDE_SYMBOL(ctiVMThrowTrampoline) "\n" |
| SYMBOL_STRING(ctiVMThrowTrampoline) ":" "\n" |
| "movl %esp, %ecx" "\n" |
| "call " LOCAL_REFERENCE(cti_vm_throw) "\n" |
| "int3" "\n" |
| ); |
| |
| asm ( |
| ".globl " SYMBOL_STRING(ctiOpThrowNotCaught) "\n" |
| HIDE_SYMBOL(ctiOpThrowNotCaught) "\n" |
| SYMBOL_STRING(ctiOpThrowNotCaught) ":" "\n" |
| "addl $0x3c, %esp" "\n" |
| "popl %ebx" "\n" |
| "popl %edi" "\n" |
| "popl %esi" "\n" |
| "popl %ebp" "\n" |
| "ret" "\n" |
| ); |
| |
| #elif COMPILER(GCC) && CPU(X86_64) |
| |
| // These ASSERTs remind you that, if you change the layout of JITStackFrame, you |
| // need to change the assembly trampolines below to match. |
| COMPILE_ASSERT(offsetof(struct JITStackFrame, code) % 32 == 0x0, JITStackFrame_maintains_32byte_stack_alignment); |
| COMPILE_ASSERT(offsetof(struct JITStackFrame, savedRBX) == 0x48, JITStackFrame_stub_argument_space_matches_ctiTrampoline); |
| COMPILE_ASSERT(offsetof(struct JITStackFrame, callFrame) == 0x90, JITStackFrame_callFrame_offset_matches_ctiTrampoline); |
| COMPILE_ASSERT(offsetof(struct JITStackFrame, code) == 0x80, JITStackFrame_code_offset_matches_ctiTrampoline); |
| |
| asm ( |
| ".globl " SYMBOL_STRING(ctiTrampoline) "\n" |
| HIDE_SYMBOL(ctiTrampoline) "\n" |
| SYMBOL_STRING(ctiTrampoline) ":" "\n" |
| "pushq %rbp" "\n" |
| "movq %rsp, %rbp" "\n" |
| "pushq %r12" "\n" |
| "pushq %r13" "\n" |
| "pushq %r14" "\n" |
| "pushq %r15" "\n" |
| "pushq %rbx" "\n" |
| "subq $0x48, %rsp" "\n" |
| "movq $512, %r12" "\n" |
| "movq $0xFFFF000000000000, %r14" "\n" |
| "movq $0xFFFF000000000002, %r15" "\n" |
| "movq 0x90(%rsp), %r13" "\n" |
| "call *0x80(%rsp)" "\n" |
| "addq $0x48, %rsp" "\n" |
| "popq %rbx" "\n" |
| "popq %r15" "\n" |
| "popq %r14" "\n" |
| "popq %r13" "\n" |
| "popq %r12" "\n" |
| "popq %rbp" "\n" |
| "ret" "\n" |
| ".globl " SYMBOL_STRING(ctiTrampolineEnd) "\n" |
| HIDE_SYMBOL(ctiTrampolineEnd) "\n" |
| SYMBOL_STRING(ctiTrampolineEnd) ":" "\n" |
| ); |
| |
| asm ( |
| ".globl " SYMBOL_STRING(ctiVMThrowTrampoline) "\n" |
| HIDE_SYMBOL(ctiVMThrowTrampoline) "\n" |
| SYMBOL_STRING(ctiVMThrowTrampoline) ":" "\n" |
| "movq %rsp, %rdi" "\n" |
| "call " LOCAL_REFERENCE(cti_vm_throw) "\n" |
| "int3" "\n" |
| ); |
| |
| asm ( |
| ".globl " SYMBOL_STRING(ctiOpThrowNotCaught) "\n" |
| HIDE_SYMBOL(ctiOpThrowNotCaught) "\n" |
| SYMBOL_STRING(ctiOpThrowNotCaught) ":" "\n" |
| "addq $0x48, %rsp" "\n" |
| "popq %rbx" "\n" |
| "popq %r15" "\n" |
| "popq %r14" "\n" |
| "popq %r13" "\n" |
| "popq %r12" "\n" |
| "popq %rbp" "\n" |
| "ret" "\n" |
| ); |
| |
| #elif (COMPILER(GCC) || COMPILER(RVCT)) && CPU(ARM_THUMB2) |
| |
| #define THUNK_RETURN_ADDRESS_OFFSET 0x38 |
| #define PRESERVED_RETURN_ADDRESS_OFFSET 0x3C |
| #define PRESERVED_R4_OFFSET 0x40 |
| #define PRESERVED_R5_OFFSET 0x44 |
| #define PRESERVED_R6_OFFSET 0x48 |
| #define PRESERVED_R7_OFFSET 0x4C |
| #define PRESERVED_R8_OFFSET 0x50 |
| #define PRESERVED_R9_OFFSET 0x54 |
| #define PRESERVED_R10_OFFSET 0x58 |
| #define PRESERVED_R11_OFFSET 0x5C |
| #define REGISTER_FILE_OFFSET 0x60 |
| #define FIRST_STACK_ARGUMENT 0x68 |
| |
| #elif (COMPILER(GCC) || COMPILER(MSVC) || COMPILER(RVCT)) && CPU(ARM_TRADITIONAL) |
| |
| // Also update the MSVC section (defined at DEFINE_STUB_FUNCTION) |
| // when changing one of the following values. |
| #define THUNK_RETURN_ADDRESS_OFFSET 64 |
| #define PRESERVEDR4_OFFSET 68 |
| |
| #elif COMPILER(MSVC) && CPU(X86) |
| |
| // These ASSERTs remind you that, if you change the layout of JITStackFrame, you |
| // need to change the assembly trampolines below to match. |
| COMPILE_ASSERT(offsetof(struct JITStackFrame, code) % 16 == 0x0, JITStackFrame_maintains_16byte_stack_alignment); |
| COMPILE_ASSERT(offsetof(struct JITStackFrame, savedEBX) == 0x3c, JITStackFrame_stub_argument_space_matches_ctiTrampoline); |
| COMPILE_ASSERT(offsetof(struct JITStackFrame, callFrame) == 0x58, JITStackFrame_callFrame_offset_matches_ctiTrampoline); |
| COMPILE_ASSERT(offsetof(struct JITStackFrame, code) == 0x50, JITStackFrame_code_offset_matches_ctiTrampoline); |
| |
| extern "C" { |
| |
| __declspec(naked) EncodedJSValue ctiTrampoline(void* code, JSStack*, CallFrame*, void* /*unused1*/, void* /*unused2*/, JSGlobalData*) |
| { |
| __asm { |
| push ebp; |
| mov ebp, esp; |
| push esi; |
| push edi; |
| push ebx; |
| sub esp, 0x3c; |
| mov ecx, esp; |
| mov edi, [esp + 0x58]; |
| call [esp + 0x50]; |
| add esp, 0x3c; |
| pop ebx; |
| pop edi; |
| pop esi; |
| pop ebp; |
| ret; |
| } |
| } |
| |
| __declspec(naked) void ctiVMThrowTrampoline() |
| { |
| __asm { |
| mov ecx, esp; |
| call cti_vm_throw; |
| add esp, 0x3c; |
| pop ebx; |
| pop edi; |
| pop esi; |
| pop ebp; |
| ret; |
| } |
| } |
| |
| __declspec(naked) void ctiOpThrowNotCaught() |
| { |
| __asm { |
| add esp, 0x3c; |
| pop ebx; |
| pop edi; |
| pop esi; |
| pop ebp; |
| ret; |
| } |
| } |
| } |
| |
| #elif CPU(MIPS) |
| |
| #define PRESERVED_GP_OFFSET 60 |
| #define PRESERVED_S0_OFFSET 64 |
| #define PRESERVED_S1_OFFSET 68 |
| #define PRESERVED_S2_OFFSET 72 |
| #define PRESERVED_RETURN_ADDRESS_OFFSET 76 |
| #define THUNK_RETURN_ADDRESS_OFFSET 80 |
| #define REGISTER_FILE_OFFSET 84 |
| #define GLOBAL_DATA_OFFSET 100 |
| #define STACK_LENGTH 104 |
| #elif CPU(SH4) |
| #define SYMBOL_STRING(name) #name |
| /* code (r4), JSStack* (r5), CallFrame* (r6), void* unused1 (r7), void* unused2(sp), JSGlobalData (sp)*/ |
| |
| asm volatile ( |
| ".text\n" |
| ".globl " SYMBOL_STRING(ctiTrampoline) "\n" |
| HIDE_SYMBOL(ctiTrampoline) "\n" |
| SYMBOL_STRING(ctiTrampoline) ":" "\n" |
| "mov.l r7, @-r15" "\n" |
| "mov.l r6, @-r15" "\n" |
| "mov.l r5, @-r15" "\n" |
| "mov.l r8, @-r15" "\n" |
| "mov #127, r8" "\n" |
| "mov.l r14, @-r15" "\n" |
| "sts.l pr, @-r15" "\n" |
| "mov.l r13, @-r15" "\n" |
| "mov.l r11, @-r15" "\n" |
| "mov.l r10, @-r15" "\n" |
| "add #-60, r15" "\n" |
| "mov r6, r14" "\n" |
| "jsr @r4" "\n" |
| "nop" "\n" |
| "add #60, r15" "\n" |
| "mov.l @r15+,r10" "\n" |
| "mov.l @r15+,r11" "\n" |
| "mov.l @r15+,r13" "\n" |
| "lds.l @r15+,pr" "\n" |
| "mov.l @r15+,r14" "\n" |
| "mov.l @r15+,r8" "\n" |
| "add #12, r15" "\n" |
| "rts" "\n" |
| "nop" "\n" |
| ); |
| |
| asm volatile ( |
| ".globl " SYMBOL_STRING(ctiVMThrowTrampoline) "\n" |
| HIDE_SYMBOL(ctiVMThrowTrampoline) "\n" |
| SYMBOL_STRING(ctiVMThrowTrampoline) ":" "\n" |
| "mov.l .L2"SYMBOL_STRING(cti_vm_throw)",r0" "\n" |
| "mov r15, r4" "\n" |
| "mov.l @(r0,r12),r11" "\n" |
| "jsr @r11" "\n" |
| "nop" "\n" |
| "add #60, r15" "\n" |
| "mov.l @r15+,r10" "\n" |
| "mov.l @r15+,r11" "\n" |
| "mov.l @r15+,r13" "\n" |
| "lds.l @r15+,pr" "\n" |
| "mov.l @r15+,r14" "\n" |
| "mov.l @r15+,r8" "\n" |
| "add #12, r15" "\n" |
| "rts" "\n" |
| "nop" "\n" |
| ".align 2" "\n" |
| ".L2"SYMBOL_STRING(cti_vm_throw)":.long " SYMBOL_STRING(cti_vm_throw)"@GOT \n" |
| ); |
| |
| asm volatile ( |
| ".globl " SYMBOL_STRING(ctiOpThrowNotCaught) "\n" |
| HIDE_SYMBOL(ctiOpThrowNotCaught) "\n" |
| SYMBOL_STRING(ctiOpThrowNotCaught) ":" "\n" |
| "add #60, r15" "\n" |
| "mov.l @r15+,r10" "\n" |
| "mov.l @r15+,r11" "\n" |
| "mov.l @r15+,r13" "\n" |
| "lds.l @r15+,pr" "\n" |
| "mov.l @r15+,r14" "\n" |
| "mov.l @r15+,r8" "\n" |
| "add #12, r15" "\n" |
| "rts" "\n" |
| "nop" "\n" |
| ); |
| #else |
| #error "JIT not supported on this platform." |
| #endif |
| |
| #else // USE(JSVALUE32_64) |
| |
| #if COMPILER(GCC) && CPU(X86_64) |
| |
| // These ASSERTs remind you that, if you change the layout of JITStackFrame, you |
| // need to change the assembly trampolines below to match. |
| COMPILE_ASSERT(offsetof(struct JITStackFrame, callFrame) == 0x58, JITStackFrame_callFrame_offset_matches_ctiTrampoline); |
| COMPILE_ASSERT(offsetof(struct JITStackFrame, code) == 0x48, JITStackFrame_code_offset_matches_ctiTrampoline); |
| COMPILE_ASSERT(offsetof(struct JITStackFrame, savedRBX) == 0x78, JITStackFrame_stub_argument_space_matches_ctiTrampoline); |
| |
| asm ( |
| ".text\n" |
| ".globl " SYMBOL_STRING(ctiTrampoline) "\n" |
| HIDE_SYMBOL(ctiTrampoline) "\n" |
| SYMBOL_STRING(ctiTrampoline) ":" "\n" |
| "pushq %rbp" "\n" |
| "movq %rsp, %rbp" "\n" |
| "pushq %r12" "\n" |
| "pushq %r13" "\n" |
| "pushq %r14" "\n" |
| "pushq %r15" "\n" |
| "pushq %rbx" "\n" |
| // Form the JIT stubs area |
| "pushq %r9" "\n" |
| "pushq %r8" "\n" |
| "pushq %rcx" "\n" |
| "pushq %rdx" "\n" |
| "pushq %rsi" "\n" |
| "pushq %rdi" "\n" |
| "subq $0x48, %rsp" "\n" |
| "movq $512, %r12" "\n" |
| "movq $0xFFFF000000000000, %r14" "\n" |
| "movq $0xFFFF000000000002, %r15" "\n" |
| "movq %rdx, %r13" "\n" |
| "call *%rdi" "\n" |
| "addq $0x78, %rsp" "\n" |
| "popq %rbx" "\n" |
| "popq %r15" "\n" |
| "popq %r14" "\n" |
| "popq %r13" "\n" |
| "popq %r12" "\n" |
| "popq %rbp" "\n" |
| "ret" "\n" |
| ".globl " SYMBOL_STRING(ctiTrampolineEnd) "\n" |
| HIDE_SYMBOL(ctiTrampolineEnd) "\n" |
| SYMBOL_STRING(ctiTrampolineEnd) ":" "\n" |
| ); |
| |
| asm ( |
| ".globl " SYMBOL_STRING(ctiVMThrowTrampoline) "\n" |
| HIDE_SYMBOL(ctiVMThrowTrampoline) "\n" |
| SYMBOL_STRING(ctiVMThrowTrampoline) ":" "\n" |
| "movq %rsp, %rdi" "\n" |
| "call " LOCAL_REFERENCE(cti_vm_throw) "\n" |
| "int3" "\n" |
| ); |
| |
| asm ( |
| ".globl " SYMBOL_STRING(ctiOpThrowNotCaught) "\n" |
| HIDE_SYMBOL(ctiOpThrowNotCaught) "\n" |
| SYMBOL_STRING(ctiOpThrowNotCaught) ":" "\n" |
| "addq $0x78, %rsp" "\n" |
| "popq %rbx" "\n" |
| "popq %r15" "\n" |
| "popq %r14" "\n" |
| "popq %r13" "\n" |
| "popq %r12" "\n" |
| "popq %rbp" "\n" |
| "ret" "\n" |
| ); |
| |
| #else |
| #error "JIT not supported on this platform." |
| #endif |
| |
| #endif // USE(JSVALUE32_64) |
| |
| #if CPU(MIPS) |
| asm ( |
| ".text" "\n" |
| ".align 2" "\n" |
| ".set noreorder" "\n" |
| ".set nomacro" "\n" |
| ".set nomips16" "\n" |
| ".globl " SYMBOL_STRING(ctiTrampoline) "\n" |
| ".ent " SYMBOL_STRING(ctiTrampoline) "\n" |
| SYMBOL_STRING(ctiTrampoline) ":" "\n" |
| "addiu $29,$29,-" STRINGIZE_VALUE_OF(STACK_LENGTH) "\n" |
| "sw $31," STRINGIZE_VALUE_OF(PRESERVED_RETURN_ADDRESS_OFFSET) "($29)" "\n" |
| "sw $18," STRINGIZE_VALUE_OF(PRESERVED_S2_OFFSET) "($29)" "\n" |
| "sw $17," STRINGIZE_VALUE_OF(PRESERVED_S1_OFFSET) "($29)" "\n" |
| "sw $16," STRINGIZE_VALUE_OF(PRESERVED_S0_OFFSET) "($29)" "\n" |
| #if WTF_MIPS_PIC |
| "sw $28," STRINGIZE_VALUE_OF(PRESERVED_GP_OFFSET) "($29)" "\n" |
| #endif |
| "move $16,$6 # set callFrameRegister" "\n" |
| "li $17,512 # set timeoutCheckRegister" "\n" |
| "move $25,$4 # move executableAddress to t9" "\n" |
| "sw $5," STRINGIZE_VALUE_OF(REGISTER_FILE_OFFSET) "($29) # store JSStack to current stack" "\n" |
| "lw $9," STRINGIZE_VALUE_OF(STACK_LENGTH + 20) "($29) # load globalData from previous stack" "\n" |
| "jalr $25" "\n" |
| "sw $9," STRINGIZE_VALUE_OF(GLOBAL_DATA_OFFSET) "($29) # store globalData to current stack" "\n" |
| "lw $16," STRINGIZE_VALUE_OF(PRESERVED_S0_OFFSET) "($29)" "\n" |
| "lw $17," STRINGIZE_VALUE_OF(PRESERVED_S1_OFFSET) "($29)" "\n" |
| "lw $18," STRINGIZE_VALUE_OF(PRESERVED_S2_OFFSET) "($29)" "\n" |
| "lw $31," STRINGIZE_VALUE_OF(PRESERVED_RETURN_ADDRESS_OFFSET) "($29)" "\n" |
| "jr $31" "\n" |
| "addiu $29,$29," STRINGIZE_VALUE_OF(STACK_LENGTH) "\n" |
| ".set reorder" "\n" |
| ".set macro" "\n" |
| ".end " SYMBOL_STRING(ctiTrampoline) "\n" |
| ); |
| |
| asm ( |
| ".text" "\n" |
| ".align 2" "\n" |
| ".set noreorder" "\n" |
| ".set nomacro" "\n" |
| ".set nomips16" "\n" |
| ".globl " SYMBOL_STRING(ctiVMThrowTrampoline) "\n" |
| ".ent " SYMBOL_STRING(ctiVMThrowTrampoline) "\n" |
| SYMBOL_STRING(ctiVMThrowTrampoline) ":" "\n" |
| #if WTF_MIPS_PIC |
| "lw $28," STRINGIZE_VALUE_OF(PRESERVED_GP_OFFSET) "($29)" "\n" |
| ".set macro" "\n" |
| "la $25," SYMBOL_STRING(cti_vm_throw) "\n" |
| ".set nomacro" "\n" |
| "bal " SYMBOL_STRING(cti_vm_throw) "\n" |
| "move $4,$29" "\n" |
| #else |
| "jal " SYMBOL_STRING(cti_vm_throw) "\n" |
| "move $4,$29" "\n" |
| #endif |
| "lw $16," STRINGIZE_VALUE_OF(PRESERVED_S0_OFFSET) "($29)" "\n" |
| "lw $17," STRINGIZE_VALUE_OF(PRESERVED_S1_OFFSET) "($29)" "\n" |
| "lw $18," STRINGIZE_VALUE_OF(PRESERVED_S2_OFFSET) "($29)" "\n" |
| "lw $31," STRINGIZE_VALUE_OF(PRESERVED_RETURN_ADDRESS_OFFSET) "($29)" "\n" |
| "jr $31" "\n" |
| "addiu $29,$29," STRINGIZE_VALUE_OF(STACK_LENGTH) "\n" |
| ".set reorder" "\n" |
| ".set macro" "\n" |
| ".end " SYMBOL_STRING(ctiVMThrowTrampoline) "\n" |
| ); |
| |
| asm ( |
| ".text" "\n" |
| ".align 2" "\n" |
| ".set noreorder" "\n" |
| ".set nomacro" "\n" |
| ".set nomips16" "\n" |
| ".globl " SYMBOL_STRING(ctiOpThrowNotCaught) "\n" |
| ".ent " SYMBOL_STRING(ctiOpThrowNotCaught) "\n" |
| SYMBOL_STRING(ctiOpThrowNotCaught) ":" "\n" |
| "lw $16," STRINGIZE_VALUE_OF(PRESERVED_S0_OFFSET) "($29)" "\n" |
| "lw $17," STRINGIZE_VALUE_OF(PRESERVED_S1_OFFSET) "($29)" "\n" |
| "lw $18," STRINGIZE_VALUE_OF(PRESERVED_S2_OFFSET) "($29)" "\n" |
| "lw $31," STRINGIZE_VALUE_OF(PRESERVED_RETURN_ADDRESS_OFFSET) "($29)" "\n" |
| "jr $31" "\n" |
| "addiu $29,$29," STRINGIZE_VALUE_OF(STACK_LENGTH) "\n" |
| ".set reorder" "\n" |
| ".set macro" "\n" |
| ".end " SYMBOL_STRING(ctiOpThrowNotCaught) "\n" |
| ); |
| #endif |
| |
| #if COMPILER(GCC) && CPU(ARM_THUMB2) |
| |
| asm ( |
| ".text" "\n" |
| ".align 2" "\n" |
| ".globl " SYMBOL_STRING(ctiTrampoline) "\n" |
| HIDE_SYMBOL(ctiTrampoline) "\n" |
| ".thumb" "\n" |
| ".thumb_func " THUMB_FUNC_PARAM(ctiTrampoline) "\n" |
| SYMBOL_STRING(ctiTrampoline) ":" "\n" |
| "sub sp, sp, #" STRINGIZE_VALUE_OF(FIRST_STACK_ARGUMENT) "\n" |
| "str lr, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_RETURN_ADDRESS_OFFSET) "]" "\n" |
| "str r4, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R4_OFFSET) "]" "\n" |
| "str r5, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R5_OFFSET) "]" "\n" |
| "str r6, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R6_OFFSET) "]" "\n" |
| "str r7, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R7_OFFSET) "]" "\n" |
| "str r8, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R8_OFFSET) "]" "\n" |
| "str r9, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R9_OFFSET) "]" "\n" |
| "str r10, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R10_OFFSET) "]" "\n" |
| "str r11, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R11_OFFSET) "]" "\n" |
| "str r1, [sp, #" STRINGIZE_VALUE_OF(REGISTER_FILE_OFFSET) "]" "\n" |
| "mov r5, r2" "\n" |
| "mov r6, #512" "\n" |
| "blx r0" "\n" |
| "ldr r11, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R11_OFFSET) "]" "\n" |
| "ldr r10, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R10_OFFSET) "]" "\n" |
| "ldr r9, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R9_OFFSET) "]" "\n" |
| "ldr r8, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R8_OFFSET) "]" "\n" |
| "ldr r7, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R7_OFFSET) "]" "\n" |
| "ldr r6, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R6_OFFSET) "]" "\n" |
| "ldr r5, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R5_OFFSET) "]" "\n" |
| "ldr r4, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R4_OFFSET) "]" "\n" |
| "ldr lr, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_RETURN_ADDRESS_OFFSET) "]" "\n" |
| "add sp, sp, #" STRINGIZE_VALUE_OF(FIRST_STACK_ARGUMENT) "\n" |
| "bx lr" "\n" |
| ".align 2" "\n" |
| ".globl " SYMBOL_STRING(ctiTrampolineEnd) "\n" |
| HIDE_SYMBOL(ctiTrampolineEnd) "\n" |
| ".thumb" "\n" |
| ".thumb_func " THUMB_FUNC_PARAM(ctiTrampolineEnd) "\n" |
| SYMBOL_STRING(ctiTrampolineEnd) ":" "\n" |
| ); |
| |
| asm ( |
| ".text" "\n" |
| ".align 2" "\n" |
| ".globl " SYMBOL_STRING(ctiVMThrowTrampoline) "\n" |
| HIDE_SYMBOL(ctiVMThrowTrampoline) "\n" |
| ".thumb" "\n" |
| ".thumb_func " THUMB_FUNC_PARAM(ctiVMThrowTrampoline) "\n" |
| SYMBOL_STRING(ctiVMThrowTrampoline) ":" "\n" |
| "mov r0, sp" "\n" |
| "bl " LOCAL_REFERENCE(cti_vm_throw) "\n" |
| "ldr r11, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R11_OFFSET) "]" "\n" |
| "ldr r10, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R10_OFFSET) "]" "\n" |
| "ldr r9, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R9_OFFSET) "]" "\n" |
| "ldr r8, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R8_OFFSET) "]" "\n" |
| "ldr r7, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R7_OFFSET) "]" "\n" |
| "ldr r6, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R6_OFFSET) "]" "\n" |
| "ldr r5, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R5_OFFSET) "]" "\n" |
| "ldr r4, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R4_OFFSET) "]" "\n" |
| "ldr lr, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_RETURN_ADDRESS_OFFSET) "]" "\n" |
| "add sp, sp, #" STRINGIZE_VALUE_OF(FIRST_STACK_ARGUMENT) "\n" |
| "bx lr" "\n" |
| ); |
| |
| asm ( |
| ".text" "\n" |
| ".align 2" "\n" |
| ".globl " SYMBOL_STRING(ctiOpThrowNotCaught) "\n" |
| HIDE_SYMBOL(ctiOpThrowNotCaught) "\n" |
| ".thumb" "\n" |
| ".thumb_func " THUMB_FUNC_PARAM(ctiOpThrowNotCaught) "\n" |
| SYMBOL_STRING(ctiOpThrowNotCaught) ":" "\n" |
| "ldr r11, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R11_OFFSET) "]" "\n" |
| "ldr r10, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R10_OFFSET) "]" "\n" |
| "ldr r9, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R9_OFFSET) "]" "\n" |
| "ldr r8, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R8_OFFSET) "]" "\n" |
| "ldr r7, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R7_OFFSET) "]" "\n" |
| "ldr r6, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R6_OFFSET) "]" "\n" |
| "ldr r5, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R5_OFFSET) "]" "\n" |
| "ldr r4, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_R4_OFFSET) "]" "\n" |
| "ldr lr, [sp, #" STRINGIZE_VALUE_OF(PRESERVED_RETURN_ADDRESS_OFFSET) "]" "\n" |
| "add sp, sp, #" STRINGIZE_VALUE_OF(FIRST_STACK_ARGUMENT) "\n" |
| "bx lr" "\n" |
| ); |
| |
| #elif COMPILER(GCC) && CPU(ARM_TRADITIONAL) |
| |
| asm ( |
| ".text" "\n" |
| ".globl " SYMBOL_STRING(ctiTrampoline) "\n" |
| HIDE_SYMBOL(ctiTrampoline) "\n" |
| INLINE_ARM_FUNCTION(ctiTrampoline) |
| SYMBOL_STRING(ctiTrampoline) ":" "\n" |
| "stmdb sp!, {r1-r3}" "\n" |
| "stmdb sp!, {r4-r6, r8-r11, lr}" "\n" |
| "sub sp, sp, #" STRINGIZE_VALUE_OF(PRESERVEDR4_OFFSET) "\n" |
| "mov r5, r2" "\n" |
| "mov r6, #512" "\n" |
| // r0 contains the code |
| "blx r0" "\n" |
| "add sp, sp, #" STRINGIZE_VALUE_OF(PRESERVEDR4_OFFSET) "\n" |
| "ldmia sp!, {r4-r6, r8-r11, lr}" "\n" |
| "add sp, sp, #12" "\n" |
| "bx lr" "\n" |
| ".globl " SYMBOL_STRING(ctiTrampolineEnd) "\n" |
| HIDE_SYMBOL(ctiTrampolineEnd) "\n" |
| SYMBOL_STRING(ctiTrampolineEnd) ":" "\n" |
| ); |
| |
| asm ( |
| ".text" "\n" |
| ".globl " SYMBOL_STRING(ctiVMThrowTrampoline) "\n" |
| HIDE_SYMBOL(ctiVMThrowTrampoline) "\n" |
| INLINE_ARM_FUNCTION(ctiVMThrowTrampoline) |
| SYMBOL_STRING(ctiVMThrowTrampoline) ":" "\n" |
| "mov r0, sp" "\n" |
| "bl " SYMBOL_STRING(cti_vm_throw) "\n" |
| |
| // Both has the same return sequence |
| ".text" "\n" |
| ".globl " SYMBOL_STRING(ctiOpThrowNotCaught) "\n" |
| HIDE_SYMBOL(ctiOpThrowNotCaught) "\n" |
| INLINE_ARM_FUNCTION(ctiOpThrowNotCaught) |
| SYMBOL_STRING(ctiOpThrowNotCaught) ":" "\n" |
| "add sp, sp, #" STRINGIZE_VALUE_OF(PRESERVEDR4_OFFSET) "\n" |
| "ldmia sp!, {r4-r6, r8-r11, lr}" "\n" |
| "add sp, sp, #12" "\n" |
| "bx lr" "\n" |
| ); |
| |
| #elif COMPILER(RVCT) && CPU(ARM_THUMB2) |
| |
| __asm EncodedJSValue ctiTrampoline(void*, JSStack*, CallFrame*, void* /*unused1*/, void* /*unused2*/, JSGlobalData*) |
| { |
| PRESERVE8 |
| sub sp, sp, # FIRST_STACK_ARGUMENT |
| str lr, [sp, # PRESERVED_RETURN_ADDRESS_OFFSET ] |
| str r4, [sp, # PRESERVED_R4_OFFSET ] |
| str r5, [sp, # PRESERVED_R5_OFFSET ] |
| str r6, [sp, # PRESERVED_R6_OFFSET ] |
| str r7, [sp, # PRESERVED_R7_OFFSET ] |
| str r8, [sp, # PRESERVED_R8_OFFSET ] |
| str r9, [sp, # PRESERVED_R9_OFFSET ] |
| str r10, [sp, # PRESERVED_R10_OFFSET ] |
| str r11, [sp, # PRESERVED_R11_OFFSET ] |
| str r1, [sp, # REGISTER_FILE_OFFSET ] |
| mov r5, r2 |
| mov r6, #512 |
| blx r0 |
| ldr r11, [sp, # PRESERVED_R11_OFFSET ] |
| ldr r10, [sp, # PRESERVED_R10_OFFSET ] |
| ldr r9, [sp, # PRESERVED_R9_OFFSET ] |
| ldr r8, [sp, # PRESERVED_R8_OFFSET ] |
| ldr r7, [sp, # PRESERVED_R7_OFFSET ] |
| ldr r6, [sp, # PRESERVED_R6_OFFSET ] |
| ldr r5, [sp, # PRESERVED_R5_OFFSET ] |
| ldr r4, [sp, # PRESERVED_R4_OFFSET ] |
| ldr lr, [sp, # PRESERVED_RETURN_ADDRESS_OFFSET ] |
| add sp, sp, # FIRST_STACK_ARGUMENT |
| bx lr |
| } |
| |
| __asm void ctiVMThrowTrampoline() |
| { |
| PRESERVE8 |
| mov r0, sp |
| bl cti_vm_throw |
| ldr r11, [sp, # PRESERVED_R11_OFFSET ] |
| ldr r10, [sp, # PRESERVED_R10_OFFSET ] |
| ldr r9, [sp, # PRESERVED_R9_OFFSET ] |
| ldr r8, [sp, # PRESERVED_R8_OFFSET ] |
| ldr r7, [sp, # PRESERVED_R7_OFFSET ] |
| ldr r6, [sp, # PRESERVED_R6_OFFSET ] |
| ldr r6, [sp, # PRESERVED_R6_OFFSET ] |
| ldr r5, [sp, # PRESERVED_R5_OFFSET ] |
| ldr r4, [sp, # PRESERVED_R4_OFFSET ] |
| ldr lr, [sp, # PRESERVED_RETURN_ADDRESS_OFFSET ] |
| add sp, sp, # FIRST_STACK_ARGUMENT |
| bx lr |
| } |
| |
| __asm void ctiOpThrowNotCaught() |
| { |
| PRESERVE8 |
| ldr r11, [sp, # PRESERVED_R11_OFFSET ] |
| ldr r10, [sp, # PRESERVED_R10_OFFSET ] |
| ldr r9, [sp, # PRESERVED_R9_OFFSET ] |
| ldr r8, [sp, # PRESERVED_R8_OFFSET ] |
| ldr r7, [sp, # PRESERVED_R7_OFFSET ] |
| ldr r6, [sp, # PRESERVED_R6_OFFSET ] |
| ldr r6, [sp, # PRESERVED_R6_OFFSET ] |
| ldr r5, [sp, # PRESERVED_R5_OFFSET ] |
| ldr r4, [sp, # PRESERVED_R4_OFFSET ] |
| ldr lr, [sp, # PRESERVED_RETURN_ADDRESS_OFFSET ] |
| add sp, sp, # FIRST_STACK_ARGUMENT |
| bx lr |
| } |
| |
| #elif COMPILER(RVCT) && CPU(ARM_TRADITIONAL) |
| |
| __asm EncodedJSValue ctiTrampoline(void*, JSStack*, CallFrame*, void* /*unused1*/, void* /*unused2*/, JSGlobalData*) |
| { |
| ARM |
| stmdb sp!, {r1-r3} |
| stmdb sp!, {r4-r6, r8-r11, lr} |
| sub sp, sp, # PRESERVEDR4_OFFSET |
| mov r5, r2 |
| mov r6, #512 |
| mov lr, pc |
| bx r0 |
| add sp, sp, # PRESERVEDR4_OFFSET |
| ldmia sp!, {r4-r6, r8-r11, lr} |
| add sp, sp, #12 |
| bx lr |
| } |
| __asm void ctiTrampolineEnd() |
| { |
| } |
| |
| __asm void ctiVMThrowTrampoline() |
| { |
| ARM |
| PRESERVE8 |
| mov r0, sp |
| bl cti_vm_throw |
| add sp, sp, # PRESERVEDR4_OFFSET |
| ldmia sp!, {r4-r6, r8-r11, lr} |
| add sp, sp, #12 |
| bx lr |
| } |
| |
| __asm void ctiOpThrowNotCaught() |
| { |
| ARM |
| add sp, sp, # PRESERVEDR4_OFFSET |
| ldmia sp!, {r4-r8, lr} |
| add sp, sp, #12 |
| bx lr |
| } |
| #endif |
| |
| #if ENABLE(OPCODE_SAMPLING) |
| #define CTI_SAMPLER stackFrame.globalData->interpreter->sampler() |
| #else |
| #define CTI_SAMPLER 0 |
| #endif |
| |
| JITThunks::JITThunks(JSGlobalData* globalData) |
| : m_hostFunctionStubMap(adoptPtr(new HostFunctionStubMap)) |
| { |
| if (!globalData->canUseJIT()) |
| return; |
| |
| m_executableMemory = JIT::compileCTIMachineTrampolines(globalData, &m_trampolineStructure); |
| ASSERT(!!m_executableMemory); |
| #if CPU(ARM_THUMB2) |
| // Unfortunate the arm compiler does not like the use of offsetof on JITStackFrame (since it contains non POD types), |
| // and the OBJECT_OFFSETOF macro does not appear constantish enough for it to be happy with its use in COMPILE_ASSERT |
| // macros. |
| ASSERT(OBJECT_OFFSETOF(struct JITStackFrame, preservedReturnAddress) == PRESERVED_RETURN_ADDRESS_OFFSET); |
| ASSERT(OBJECT_OFFSETOF(struct JITStackFrame, preservedR4) == PRESERVED_R4_OFFSET); |
| ASSERT(OBJECT_OFFSETOF(struct JITStackFrame, preservedR5) == PRESERVED_R5_OFFSET); |
| ASSERT(OBJECT_OFFSETOF(struct JITStackFrame, preservedR6) == PRESERVED_R6_OFFSET); |
| ASSERT(OBJECT_OFFSETOF(struct JITStackFrame, preservedR7) == PRESERVED_R7_OFFSET); |
| ASSERT(OBJECT_OFFSETOF(struct JITStackFrame, preservedR8) == PRESERVED_R8_OFFSET); |
| ASSERT(OBJECT_OFFSETOF(struct JITStackFrame, preservedR9) == PRESERVED_R9_OFFSET); |
| ASSERT(OBJECT_OFFSETOF(struct JITStackFrame, preservedR10) == PRESERVED_R10_OFFSET); |
| ASSERT(OBJECT_OFFSETOF(struct JITStackFrame, preservedR11) == PRESERVED_R11_OFFSET); |
| |
| ASSERT(OBJECT_OFFSETOF(struct JITStackFrame, stack) == REGISTER_FILE_OFFSET); |
| // The fifth argument is the first item already on the stack. |
| ASSERT(OBJECT_OFFSETOF(struct JITStackFrame, unused1) == FIRST_STACK_ARGUMENT); |
| |
| ASSERT(OBJECT_OFFSETOF(struct JITStackFrame, thunkReturnAddress) == THUNK_RETURN_ADDRESS_OFFSET); |
| |
| #elif CPU(ARM_TRADITIONAL) |
| |
| ASSERT(OBJECT_OFFSETOF(struct JITStackFrame, thunkReturnAddress) == THUNK_RETURN_ADDRESS_OFFSET); |
| ASSERT(OBJECT_OFFSETOF(struct JITStackFrame, preservedR4) == PRESERVEDR4_OFFSET); |
| |
| |
| #elif CPU(MIPS) |
| ASSERT(OBJECT_OFFSETOF(struct JITStackFrame, preservedGP) == PRESERVED_GP_OFFSET); |
| ASSERT(OBJECT_OFFSETOF(struct JITStackFrame, preservedS0) == PRESERVED_S0_OFFSET); |
| ASSERT(OBJECT_OFFSETOF(struct JITStackFrame, preservedS1) == PRESERVED_S1_OFFSET); |
| ASSERT(OBJECT_OFFSETOF(struct JITStackFrame, preservedS2) == PRESERVED_S2_OFFSET); |
| ASSERT(OBJECT_OFFSETOF(struct JITStackFrame, preservedReturnAddress) == PRESERVED_RETURN_ADDRESS_OFFSET); |
| ASSERT(OBJECT_OFFSETOF(struct JITStackFrame, thunkReturnAddress) == THUNK_RETURN_ADDRESS_OFFSET); |
| ASSERT(OBJECT_OFFSETOF(struct JITStackFrame, stack) == REGISTER_FILE_OFFSET); |
| ASSERT(OBJECT_OFFSETOF(struct JITStackFrame, globalData) == GLOBAL_DATA_OFFSET); |
| |
| #endif |
| } |
| |
| JITThunks::~JITThunks() |
| { |
| } |
| |
| NEVER_INLINE void JITThunks::tryCachePutByID(CallFrame* callFrame, CodeBlock* codeBlock, ReturnAddressPtr returnAddress, JSValue baseValue, const PutPropertySlot& slot, StructureStubInfo* stubInfo, bool direct) |
| { |
| // The interpreter checks for recursion here; I do not believe this can occur in CTI. |
| |
| if (!baseValue.isCell()) |
| return; |
| |
| // Uncacheable: give up. |
| if (!slot.isCacheable()) { |
| ctiPatchCallByReturnAddress(codeBlock, returnAddress, FunctionPtr(direct ? cti_op_put_by_id_direct_generic : cti_op_put_by_id_generic)); |
| return; |
| } |
| |
| JSCell* baseCell = baseValue.asCell(); |
| Structure* structure = baseCell->structure(); |
| |
| if (structure->isUncacheableDictionary() || structure->typeInfo().prohibitsPropertyCaching()) { |
| ctiPatchCallByReturnAddress(codeBlock, returnAddress, FunctionPtr(direct ? cti_op_put_by_id_direct_generic : cti_op_put_by_id_generic)); |
| return; |
| } |
| |
| // If baseCell != base, then baseCell must be a proxy for another object. |
| if (baseCell != slot.base()) { |
| ctiPatchCallByReturnAddress(codeBlock, returnAddress, FunctionPtr(direct ? cti_op_put_by_id_direct_generic : cti_op_put_by_id_generic)); |
| return; |
| } |
| |
| // Cache hit: Specialize instruction and ref Structures. |
| |
| // Structure transition, cache transition info |
| if (slot.type() == PutPropertySlot::NewProperty) { |
| if (structure->isDictionary()) { |
| ctiPatchCallByReturnAddress(codeBlock, returnAddress, FunctionPtr(direct ? cti_op_put_by_id_direct_generic : cti_op_put_by_id_generic)); |
| return; |
| } |
| |
| // put_by_id_transition checks the prototype chain for setters. |
| if (normalizePrototypeChain(callFrame, baseCell) == InvalidPrototypeChain) { |
| ctiPatchCallByReturnAddress(codeBlock, returnAddress, FunctionPtr(direct ? cti_op_put_by_id_direct_generic : cti_op_put_by_id_generic)); |
| return; |
| } |
| |
| StructureChain* prototypeChain = structure->prototypeChain(callFrame); |
| ASSERT(structure->previousID()->transitionWatchpointSetHasBeenInvalidated()); |
| stubInfo->initPutByIdTransition(callFrame->globalData(), codeBlock->ownerExecutable(), structure->previousID(), structure, prototypeChain, direct); |
| JIT::compilePutByIdTransition(callFrame->scope()->globalData(), codeBlock, stubInfo, structure->previousID(), structure, slot.cachedOffset(), prototypeChain, returnAddress, direct); |
| return; |
| } |
| |
| stubInfo->initPutByIdReplace(callFrame->globalData(), codeBlock->ownerExecutable(), structure); |
| |
| JIT::patchPutByIdReplace(codeBlock, stubInfo, structure, slot.cachedOffset(), returnAddress, direct); |
| } |
| |
| NEVER_INLINE void JITThunks::tryCacheGetByID(CallFrame* callFrame, CodeBlock* codeBlock, ReturnAddressPtr returnAddress, JSValue baseValue, const Identifier& propertyName, const PropertySlot& slot, StructureStubInfo* stubInfo) |
| { |
| // FIXME: Write a test that proves we need to check for recursion here just |
| // like the interpreter does, then add a check for recursion. |
| |
| // FIXME: Cache property access for immediates. |
| if (!baseValue.isCell()) { |
| ctiPatchCallByReturnAddress(codeBlock, returnAddress, FunctionPtr(cti_op_get_by_id_generic)); |
| return; |
| } |
| |
| JSGlobalData* globalData = &callFrame->globalData(); |
| |
| if (isJSArray(baseValue) && propertyName == callFrame->propertyNames().length) { |
| JIT::compilePatchGetArrayLength(callFrame->scope()->globalData(), codeBlock, returnAddress); |
| return; |
| } |
| |
| if (isJSString(baseValue) && propertyName == callFrame->propertyNames().length) { |
| // The tradeoff of compiling an patched inline string length access routine does not seem |
| // to pay off, so we currently only do this for arrays. |
| ctiPatchCallByReturnAddress(codeBlock, returnAddress, globalData->jitStubs->ctiStringLengthTrampoline()); |
| return; |
| } |
| |
| // Uncacheable: give up. |
| if (!slot.isCacheable()) { |
| stubInfo->accessType = access_get_by_id_generic; |
| ctiPatchCallByReturnAddress(codeBlock, returnAddress, FunctionPtr(cti_op_get_by_id_generic)); |
| return; |
| } |
| |
| JSCell* baseCell = baseValue.asCell(); |
| Structure* structure = baseCell->structure(); |
| |
| if (structure->isUncacheableDictionary() || structure->typeInfo().prohibitsPropertyCaching()) { |
| stubInfo->accessType = access_get_by_id_generic; |
| ctiPatchCallByReturnAddress(codeBlock, returnAddress, FunctionPtr(cti_op_get_by_id_generic)); |
| return; |
| } |
| |
| // Cache hit: Specialize instruction and ref Structures. |
| |
| if (slot.slotBase() == baseValue) { |
| // set this up, so derefStructures can do it's job. |
| stubInfo->initGetByIdSelf(callFrame->globalData(), codeBlock->ownerExecutable(), structure); |
| if ((slot.cachedPropertyType() != PropertySlot::Value) |
| || !MacroAssembler::isCompactPtrAlignedAddressOffset(offsetRelativeToPatchedStorage(slot.cachedOffset()))) |
| ctiPatchCallByReturnAddress(codeBlock, returnAddress, FunctionPtr(cti_op_get_by_id_self_fail)); |
| else |
| JIT::patchGetByIdSelf(codeBlock, stubInfo, structure, slot.cachedOffset(), returnAddress); |
| return; |
| } |
| |
| if (structure->isDictionary()) { |
| stubInfo->accessType = access_get_by_id_generic; |
| ctiPatchCallByReturnAddress(codeBlock, returnAddress, FunctionPtr(cti_op_get_by_id_generic)); |
| return; |
| } |
| |
| if (slot.slotBase() == structure->prototypeForLookup(callFrame)) { |
| ASSERT(slot.slotBase().isObject()); |
| |
| JSObject* slotBaseObject = asObject(slot.slotBase()); |
| size_t offset = slot.cachedOffset(); |
| |
| if (structure->typeInfo().hasImpureGetOwnPropertySlot()) { |
| stubInfo->accessType = access_get_by_id_generic; |
| ctiPatchCallByReturnAddress(codeBlock, returnAddress, FunctionPtr(cti_op_get_by_id_generic)); |
| return; |
| } |
| |
| // Since we're accessing a prototype in a loop, it's a good bet that it |
| // should not be treated as a dictionary. |
| if (slotBaseObject->structure()->isDictionary()) { |
| slotBaseObject->flattenDictionaryObject(callFrame->globalData()); |
| offset = slotBaseObject->structure()->get(callFrame->globalData(), propertyName); |
| } |
| |
| stubInfo->initGetByIdProto(callFrame->globalData(), codeBlock->ownerExecutable(), structure, slotBaseObject->structure(), slot.cachedPropertyType() == PropertySlot::Value); |
| |
| ASSERT(!structure->isDictionary()); |
| ASSERT(!slotBaseObject->structure()->isDictionary()); |
| JIT::compileGetByIdProto(callFrame->scope()->globalData(), callFrame, codeBlock, stubInfo, structure, slotBaseObject->structure(), propertyName, slot, offset, returnAddress); |
| return; |
| } |
| |
| PropertyOffset offset = slot.cachedOffset(); |
| size_t count = normalizePrototypeChainForChainAccess(callFrame, baseValue, slot.slotBase(), propertyName, offset); |
| if (count == InvalidPrototypeChain) { |
| stubInfo->accessType = access_get_by_id_generic; |
| ctiPatchCallByReturnAddress(codeBlock, returnAddress, FunctionPtr(cti_op_get_by_id_generic)); |
| return; |
| } |
| |
| StructureChain* prototypeChain = structure->prototypeChain(callFrame); |
| stubInfo->initGetByIdChain(callFrame->globalData(), codeBlock->ownerExecutable(), structure, prototypeChain, count, slot.cachedPropertyType() == PropertySlot::Value); |
| JIT::compileGetByIdChain(callFrame->scope()->globalData(), callFrame, codeBlock, stubInfo, structure, prototypeChain, count, propertyName, slot, offset, returnAddress); |
| } |
| |
| #if !defined(NDEBUG) |
| |
| extern "C" { |
| |
| static void jscGeneratedNativeCode() |
| { |
| // When executing a JIT stub function (which might do an allocation), we hack the return address |
| // to pretend to be executing this function, to keep stack logging tools from blowing out |
| // memory. |
| } |
| |
| } |
| |
| struct StackHack { |
| ALWAYS_INLINE StackHack(JITStackFrame& stackFrame) |
| : stackFrame(stackFrame) |
| , savedReturnAddress(*stackFrame.returnAddressSlot()) |
| { |
| if (!CodeProfiling::enabled()) |
| *stackFrame.returnAddressSlot() = ReturnAddressPtr(FunctionPtr(jscGeneratedNativeCode)); |
| } |
| |
| ALWAYS_INLINE ~StackHack() |
| { |
| *stackFrame.returnAddressSlot() = savedReturnAddress; |
| } |
| |
| JITStackFrame& stackFrame; |
| ReturnAddressPtr savedReturnAddress; |
| }; |
| |
| #define STUB_INIT_STACK_FRAME(stackFrame) JITStackFrame& stackFrame = *reinterpret_cast_ptr<JITStackFrame*>(STUB_ARGS); StackHack stackHack(stackFrame) |
| #define STUB_SET_RETURN_ADDRESS(returnAddress) stackHack.savedReturnAddress = ReturnAddressPtr(returnAddress) |
| #define STUB_RETURN_ADDRESS stackHack.savedReturnAddress |
| |
| #else |
| |
| #define STUB_INIT_STACK_FRAME(stackFrame) JITStackFrame& stackFrame = *reinterpret_cast_ptr<JITStackFrame*>(STUB_ARGS) |
| #define STUB_SET_RETURN_ADDRESS(returnAddress) *stackFrame.returnAddressSlot() = ReturnAddressPtr(returnAddress) |
| #define STUB_RETURN_ADDRESS *stackFrame.returnAddressSlot() |
| |
| #endif |
| |
| // The reason this is not inlined is to avoid having to do a PIC branch |
| // to get the address of the ctiVMThrowTrampoline function. It's also |
| // good to keep the code size down by leaving as much of the exception |
| // handling code out of line as possible. |
| static NEVER_INLINE void returnToThrowTrampoline(JSGlobalData* globalData, ReturnAddressPtr exceptionLocation, ReturnAddressPtr& returnAddressSlot) |
| { |
| ASSERT(globalData->exception); |
| globalData->exceptionLocation = exceptionLocation; |
| returnAddressSlot = ReturnAddressPtr(FunctionPtr(ctiVMThrowTrampoline)); |
| } |
| |
| #define VM_THROW_EXCEPTION() \ |
| do { \ |
| VM_THROW_EXCEPTION_AT_END(); \ |
| return 0; \ |
| } while (0) |
| #define VM_THROW_EXCEPTION_AT_END() \ |
| do {\ |
| returnToThrowTrampoline(stackFrame.globalData, STUB_RETURN_ADDRESS, STUB_RETURN_ADDRESS);\ |
| } while (0) |
| |
| #define CHECK_FOR_EXCEPTION() \ |
| do { \ |
| if (UNLIKELY(stackFrame.globalData->exception)) \ |
| VM_THROW_EXCEPTION(); \ |
| } while (0) |
| #define CHECK_FOR_EXCEPTION_AT_END() \ |
| do { \ |
| if (UNLIKELY(stackFrame.globalData->exception)) \ |
| VM_THROW_EXCEPTION_AT_END(); \ |
| } while (0) |
| #define CHECK_FOR_EXCEPTION_VOID() \ |
| do { \ |
| if (UNLIKELY(stackFrame.globalData->exception)) { \ |
| VM_THROW_EXCEPTION_AT_END(); \ |
| return; \ |
| } \ |
| } while (0) |
| |
| // Helper function for JIT stubs that may throw an exception in the middle of |
| // processing a function call. This function rolls back the stack to |
| // our caller, so exception processing can proceed from a valid state. |
| template<typename T> static T throwExceptionFromOpCall(JITStackFrame& jitStackFrame, CallFrame* newCallFrame, ReturnAddressPtr& returnAddressSlot) |
| { |
| CallFrame* callFrame = newCallFrame->callerFrame(); |
| ASSERT(callFrame->globalData().exception); |
| jitStackFrame.callFrame = callFrame; |
| callFrame->globalData().topCallFrame = callFrame; |
| returnToThrowTrampoline(&callFrame->globalData(), ReturnAddressPtr(newCallFrame->returnPC()), returnAddressSlot); |
| return T(); |
| } |
| |
| template<typename T> static T throwExceptionFromOpCall(JITStackFrame& jitStackFrame, CallFrame* newCallFrame, ReturnAddressPtr& returnAddressSlot, JSValue exception) |
| { |
| newCallFrame->callerFrame()->globalData().exception = exception; |
| return throwExceptionFromOpCall<T>(jitStackFrame, newCallFrame, returnAddressSlot); |
| } |
| |
| #if CPU(ARM_THUMB2) && COMPILER(GCC) |
| |
| #define DEFINE_STUB_FUNCTION(rtype, op) \ |
| extern "C" { \ |
| rtype JITStubThunked_##op(STUB_ARGS_DECLARATION); \ |
| }; \ |
| asm ( \ |
| ".text" "\n" \ |
| ".align 2" "\n" \ |
| ".globl " SYMBOL_STRING(cti_##op) "\n" \ |
| HIDE_SYMBOL(cti_##op) "\n" \ |
| ".thumb" "\n" \ |
| ".thumb_func " THUMB_FUNC_PARAM(cti_##op) "\n" \ |
| SYMBOL_STRING(cti_##op) ":" "\n" \ |
| "str lr, [sp, #" STRINGIZE_VALUE_OF(THUNK_RETURN_ADDRESS_OFFSET) "]" "\n" \ |
| "bl " SYMBOL_STRING(JITStubThunked_##op) "\n" \ |
| "ldr lr, [sp, #" STRINGIZE_VALUE_OF(THUNK_RETURN_ADDRESS_OFFSET) "]" "\n" \ |
| "bx lr" "\n" \ |
| ); \ |
| rtype JITStubThunked_##op(STUB_ARGS_DECLARATION) \ |
| |
| #elif CPU(MIPS) |
| #if WTF_MIPS_PIC |
| #define DEFINE_STUB_FUNCTION(rtype, op) \ |
| extern "C" { \ |
| rtype JITStubThunked_##op(STUB_ARGS_DECLARATION); \ |
| }; \ |
| asm ( \ |
| ".text" "\n" \ |
| ".align 2" "\n" \ |
| ".set noreorder" "\n" \ |
| ".set nomacro" "\n" \ |
| ".set nomips16" "\n" \ |
| ".globl " SYMBOL_STRING(cti_##op) "\n" \ |
| ".ent " SYMBOL_STRING(cti_##op) "\n" \ |
| SYMBOL_STRING(cti_##op) ":" "\n" \ |
| "lw $28," STRINGIZE_VALUE_OF(PRESERVED_GP_OFFSET) "($29)" "\n" \ |
| "sw $31," STRINGIZE_VALUE_OF(THUNK_RETURN_ADDRESS_OFFSET) "($29)" "\n" \ |
| ".set macro" "\n" \ |
| "la $25," SYMBOL_STRING(JITStubThunked_##op) "\n" \ |
| ".set nomacro" "\n" \ |
| ".reloc 1f,R_MIPS_JALR," SYMBOL_STRING(JITStubThunked_##op) "\n" \ |
| "1: jalr $25" "\n" \ |
| "nop" "\n" \ |
| "lw $31," STRINGIZE_VALUE_OF(THUNK_RETURN_ADDRESS_OFFSET) "($29)" "\n" \ |
| "jr $31" "\n" \ |
| "nop" "\n" \ |
| ".set reorder" "\n" \ |
| ".set macro" "\n" \ |
| ".end " SYMBOL_STRING(cti_##op) "\n" \ |
| ); \ |
| rtype JITStubThunked_##op(STUB_ARGS_DECLARATION) |
| |
| #else // WTF_MIPS_PIC |
| #define DEFINE_STUB_FUNCTION(rtype, op) \ |
| extern "C" { \ |
| rtype JITStubThunked_##op(STUB_ARGS_DECLARATION); \ |
| }; \ |
| asm ( \ |
| ".text" "\n" \ |
| ".align 2" "\n" \ |
| ".set noreorder" "\n" \ |
| ".set nomacro" "\n" \ |
| ".set nomips16" "\n" \ |
| ".globl " SYMBOL_STRING(cti_##op) "\n" \ |
| ".ent " SYMBOL_STRING(cti_##op) "\n" \ |
| SYMBOL_STRING(cti_##op) ":" "\n" \ |
| "sw $31," STRINGIZE_VALUE_OF(THUNK_RETURN_ADDRESS_OFFSET) "($29)" "\n" \ |
| "jal " SYMBOL_STRING(JITStubThunked_##op) "\n" \ |
| "nop" "\n" \ |
| "lw $31," STRINGIZE_VALUE_OF(THUNK_RETURN_ADDRESS_OFFSET) "($29)" "\n" \ |
| "jr $31" "\n" \ |
| "nop" "\n" \ |
| ".set reorder" "\n" \ |
| ".set macro" "\n" \ |
| ".end " SYMBOL_STRING(cti_##op) "\n" \ |
| ); \ |
| rtype JITStubThunked_##op(STUB_ARGS_DECLARATION) |
| |
| #endif |
| |
| #elif CPU(ARM_TRADITIONAL) && COMPILER(GCC) |
| |
| #define DEFINE_STUB_FUNCTION(rtype, op) \ |
| extern "C" { \ |
| rtype JITStubThunked_##op(STUB_ARGS_DECLARATION); \ |
| }; \ |
| asm ( \ |
| ".globl " SYMBOL_STRING(cti_##op) "\n" \ |
| INLINE_ARM_FUNCTION(cti_##op) \ |
| SYMBOL_STRING(cti_##op) ":" "\n" \ |
| "str lr, [sp, #" STRINGIZE_VALUE_OF(THUNK_RETURN_ADDRESS_OFFSET) "]" "\n" \ |
| "bl " SYMBOL_STRING(JITStubThunked_##op) "\n" \ |
| "ldr lr, [sp, #" STRINGIZE_VALUE_OF(THUNK_RETURN_ADDRESS_OFFSET) "]" "\n" \ |
| "bx lr" "\n" \ |
| ); \ |
| rtype JITStubThunked_##op(STUB_ARGS_DECLARATION) |
| |
| #elif (CPU(ARM_THUMB2) || CPU(ARM_TRADITIONAL)) && COMPILER(RVCT) |
| |
| #define DEFINE_STUB_FUNCTION(rtype, op) rtype JITStubThunked_##op(STUB_ARGS_DECLARATION) |
| |
| /* The following is a workaround for RVCT toolchain; precompiler macros are not expanded before the code is passed to the assembler */ |
| |
| /* The following section is a template to generate code for GeneratedJITStubs_RVCT.h */ |
| /* The pattern "#xxx#" will be replaced with "xxx" */ |
| |
| /* |
| RVCT(extern "C" #rtype# JITStubThunked_#op#(STUB_ARGS_DECLARATION);) |
| RVCT(__asm #rtype# cti_#op#(STUB_ARGS_DECLARATION)) |
| RVCT({) |
| RVCT( PRESERVE8) |
| RVCT( IMPORT JITStubThunked_#op#) |
| RVCT( str lr, [sp, # THUNK_RETURN_ADDRESS_OFFSET]) |
| RVCT( bl JITStubThunked_#op#) |
| RVCT( ldr lr, [sp, # THUNK_RETURN_ADDRESS_OFFSET]) |
| RVCT( bx lr) |
| RVCT(}) |
| RVCT() |
| */ |
| |
| /* Include the generated file */ |
| #include "GeneratedJITStubs_RVCT.h" |
| |
| #elif CPU(ARM_TRADITIONAL) && COMPILER(MSVC) |
| |
| #define DEFINE_STUB_FUNCTION(rtype, op) extern "C" rtype JITStubThunked_##op(STUB_ARGS_DECLARATION) |
| |
| /* The following is a workaround for MSVC toolchain; inline assembler is not supported */ |
| |
| /* The following section is a template to generate code for GeneratedJITStubs_MSVC.asm */ |
| /* The pattern "#xxx#" will be replaced with "xxx" */ |
| |
| /* |
| MSVC_BEGIN( AREA Trampoline, CODE) |
| MSVC_BEGIN() |
| MSVC_BEGIN( EXPORT ctiTrampoline) |
| MSVC_BEGIN( EXPORT ctiTrampolineEnd) |
| MSVC_BEGIN( EXPORT ctiVMThrowTrampoline) |
| MSVC_BEGIN( EXPORT ctiOpThrowNotCaught) |
| MSVC_BEGIN() |
| MSVC_BEGIN(ctiTrampoline PROC) |
| MSVC_BEGIN( stmdb sp!, {r1-r3}) |
| MSVC_BEGIN( stmdb sp!, {r4-r6, r8-r11, lr}) |
| MSVC_BEGIN( sub sp, sp, #68 ; sync with PRESERVEDR4_OFFSET) |
| MSVC_BEGIN( mov r5, r2) |
| MSVC_BEGIN( mov r6, #512) |
| MSVC_BEGIN( ; r0 contains the code) |
| MSVC_BEGIN( mov lr, pc) |
| MSVC_BEGIN( bx r0) |
| MSVC_BEGIN( add sp, sp, #68 ; sync with PRESERVEDR4_OFFSET) |
| MSVC_BEGIN( ldmia sp!, {r4-r6, r8-r11, lr}) |
| MSVC_BEGIN( add sp, sp, #12) |
| MSVC_BEGIN( bx lr) |
| MSVC_BEGIN(ctiTrampolineEnd) |
| MSVC_BEGIN(ctiTrampoline ENDP) |
| MSVC_BEGIN() |
| MSVC_BEGIN(ctiVMThrowTrampoline PROC) |
| MSVC_BEGIN( mov r0, sp) |
| MSVC_BEGIN( bl cti_vm_throw) |
| MSVC_BEGIN(ctiOpThrowNotCaught) |
| MSVC_BEGIN( add sp, sp, #68 ; sync with PRESERVEDR4_OFFSET) |
| MSVC_BEGIN( ldmia sp!, {r4-r6, r8-r11, lr}) |
| MSVC_BEGIN( add sp, sp, #12) |
| MSVC_BEGIN( bx lr) |
| MSVC_BEGIN(ctiVMThrowTrampoline ENDP) |
| MSVC_BEGIN() |
| |
| MSVC( EXPORT cti_#op#) |
| MSVC( IMPORT JITStubThunked_#op#) |
| MSVC(cti_#op# PROC) |
| MSVC( str lr, [sp, #64] ; sync with THUNK_RETURN_ADDRESS_OFFSET) |
| MSVC( bl JITStubThunked_#op#) |
| MSVC( ldr lr, [sp, #64] ; sync with THUNK_RETURN_ADDRESS_OFFSET) |
| MSVC( bx lr) |
| MSVC(cti_#op# ENDP) |
| MSVC() |
| |
| MSVC_END( END) |
| */ |
| |
| #elif CPU(SH4) |
| #define DEFINE_STUB_FUNCTION(rtype, op) \ |
| extern "C" { \ |
| rtype JITStubThunked_##op(STUB_ARGS_DECLARATION); \ |
| }; \ |
| asm volatile( \ |
| ".align 2" "\n" \ |
| ".globl " SYMBOL_STRING(cti_##op) "\n" \ |
| SYMBOL_STRING(cti_##op) ":" "\n" \ |
| "sts pr, r11" "\n" \ |
| "mov.l r11, @(0x38, r15)" "\n" \ |
| "mov.l .L2"SYMBOL_STRING(JITStubThunked_##op)",r0" "\n" \ |
| "mov.l @(r0,r12),r11" "\n" \ |
| "jsr @r11" "\n" \ |
| "nop" "\n" \ |
| "mov.l @(0x38, r15), r11 " "\n" \ |
| "lds r11, pr " "\n" \ |
| "rts" "\n" \ |
| "nop" "\n" \ |
| ".align 2" "\n" \ |
| ".L2"SYMBOL_STRING(JITStubThunked_##op)":.long " SYMBOL_STRING(JITStubThunked_##op)"@GOT \n" \ |
| ); \ |
| rtype JITStubThunked_##op(STUB_ARGS_DECLARATION) |
| #else |
| #define DEFINE_STUB_FUNCTION(rtype, op) rtype JIT_STUB cti_##op(STUB_ARGS_DECLARATION) |
| #endif |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_create_this) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| CallFrame* callFrame = stackFrame.callFrame; |
| |
| JSFunction* constructor = jsCast<JSFunction*>(callFrame->callee()); |
| #if !ASSERT_DISABLED |
| ConstructData constructData; |
| ASSERT(constructor->methodTable()->getConstructData(constructor, constructData) == ConstructTypeJS); |
| #endif |
| |
| Structure* structure = constructor->cachedInheritorID(callFrame); |
| JSValue result = constructEmptyObject(callFrame, structure); |
| |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_convert_this) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| JSValue v1 = stackFrame.args[0].jsValue(); |
| CallFrame* callFrame = stackFrame.callFrame; |
| |
| ASSERT(v1.isPrimitive()); |
| |
| JSObject* result = v1.toThisObject(callFrame); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_add) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| JSValue v1 = stackFrame.args[0].jsValue(); |
| JSValue v2 = stackFrame.args[1].jsValue(); |
| CallFrame* callFrame = stackFrame.callFrame; |
| |
| if (v1.isString() && !v2.isObject()) { |
| JSValue result = jsString(callFrame, asString(v1), v2.toString(callFrame)); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return JSValue::encode(result); |
| } |
| |
| if (v1.isNumber() && v2.isNumber()) |
| return JSValue::encode(jsNumber(v1.asNumber() + v2.asNumber())); |
| |
| // All other cases are pretty uncommon |
| JSValue result = jsAddSlowCase(callFrame, v1, v2); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_pre_inc) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| JSValue v = stackFrame.args[0].jsValue(); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| JSValue result = jsNumber(v.toNumber(callFrame) + 1); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(int, timeout_check) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| JSGlobalData* globalData = stackFrame.globalData; |
| TimeoutChecker& timeoutChecker = globalData->timeoutChecker; |
| |
| if (globalData->terminator.shouldTerminate()) { |
| globalData->exception = createTerminatedExecutionException(globalData); |
| VM_THROW_EXCEPTION_AT_END(); |
| } else if (timeoutChecker.didTimeOut(stackFrame.callFrame)) { |
| globalData->exception = createInterruptedExecutionException(globalData); |
| VM_THROW_EXCEPTION_AT_END(); |
| } |
| |
| return timeoutChecker.ticksUntilNextCheck(); |
| } |
| |
| DEFINE_STUB_FUNCTION(void*, stack_check) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| CallFrame* callFrame = stackFrame.callFrame; |
| |
| if (UNLIKELY(!stackFrame.stack->grow(&callFrame->registers()[callFrame->codeBlock()->m_numCalleeRegisters]))) |
| return throwExceptionFromOpCall<void*>(stackFrame, callFrame, STUB_RETURN_ADDRESS, createStackOverflowError(callFrame->callerFrame())); |
| |
| return callFrame; |
| } |
| |
| DEFINE_STUB_FUNCTION(JSObject*, op_new_object) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| return constructEmptyObject(stackFrame.callFrame); |
| } |
| |
| DEFINE_STUB_FUNCTION(void, op_put_by_id_generic) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| PutPropertySlot slot(stackFrame.callFrame->codeBlock()->isStrictMode()); |
| stackFrame.args[0].jsValue().put(stackFrame.callFrame, stackFrame.args[1].identifier(), stackFrame.args[2].jsValue(), slot); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| } |
| |
| DEFINE_STUB_FUNCTION(void, op_put_by_id_direct_generic) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| PutPropertySlot slot(stackFrame.callFrame->codeBlock()->isStrictMode()); |
| JSValue baseValue = stackFrame.args[0].jsValue(); |
| ASSERT(baseValue.isObject()); |
| asObject(baseValue)->putDirect(stackFrame.callFrame->globalData(), stackFrame.args[1].identifier(), stackFrame.args[2].jsValue(), slot); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_get_by_id_generic) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| Identifier& ident = stackFrame.args[1].identifier(); |
| |
| JSValue baseValue = stackFrame.args[0].jsValue(); |
| PropertySlot slot(baseValue); |
| JSValue result = baseValue.get(callFrame, ident, slot); |
| |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(void, op_put_by_id) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| CallFrame* callFrame = stackFrame.callFrame; |
| Identifier& ident = stackFrame.args[1].identifier(); |
| |
| CodeBlock* codeBlock = stackFrame.callFrame->codeBlock(); |
| StructureStubInfo* stubInfo = &codeBlock->getStubInfo(STUB_RETURN_ADDRESS); |
| AccessType accessType = static_cast<AccessType>(stubInfo->accessType); |
| |
| PutPropertySlot slot(callFrame->codeBlock()->isStrictMode()); |
| stackFrame.args[0].jsValue().put(callFrame, ident, stackFrame.args[2].jsValue(), slot); |
| |
| if (accessType == static_cast<AccessType>(stubInfo->accessType)) { |
| if (!stubInfo->seenOnce()) |
| stubInfo->setSeen(); |
| else |
| JITThunks::tryCachePutByID(callFrame, codeBlock, STUB_RETURN_ADDRESS, stackFrame.args[0].jsValue(), slot, stubInfo, false); |
| } |
| |
| CHECK_FOR_EXCEPTION_AT_END(); |
| } |
| |
| DEFINE_STUB_FUNCTION(void, op_put_by_id_direct) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| CallFrame* callFrame = stackFrame.callFrame; |
| Identifier& ident = stackFrame.args[1].identifier(); |
| |
| CodeBlock* codeBlock = stackFrame.callFrame->codeBlock(); |
| StructureStubInfo* stubInfo = &codeBlock->getStubInfo(STUB_RETURN_ADDRESS); |
| AccessType accessType = static_cast<AccessType>(stubInfo->accessType); |
| |
| PutPropertySlot slot(callFrame->codeBlock()->isStrictMode()); |
| JSValue baseValue = stackFrame.args[0].jsValue(); |
| ASSERT(baseValue.isObject()); |
| |
| asObject(baseValue)->putDirect(callFrame->globalData(), ident, stackFrame.args[2].jsValue(), slot); |
| |
| if (accessType == static_cast<AccessType>(stubInfo->accessType)) { |
| if (!stubInfo->seenOnce()) |
| stubInfo->setSeen(); |
| else |
| JITThunks::tryCachePutByID(callFrame, codeBlock, STUB_RETURN_ADDRESS, stackFrame.args[0].jsValue(), slot, stubInfo, true); |
| } |
| |
| CHECK_FOR_EXCEPTION_AT_END(); |
| } |
| |
| DEFINE_STUB_FUNCTION(void, op_put_by_id_fail) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| Identifier& ident = stackFrame.args[1].identifier(); |
| |
| PutPropertySlot slot(callFrame->codeBlock()->isStrictMode()); |
| stackFrame.args[0].jsValue().put(callFrame, ident, stackFrame.args[2].jsValue(), slot); |
| |
| CHECK_FOR_EXCEPTION_AT_END(); |
| } |
| |
| DEFINE_STUB_FUNCTION(void, op_put_by_id_direct_fail) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| Identifier& ident = stackFrame.args[1].identifier(); |
| |
| PutPropertySlot slot(callFrame->codeBlock()->isStrictMode()); |
| JSValue baseValue = stackFrame.args[0].jsValue(); |
| ASSERT(baseValue.isObject()); |
| asObject(baseValue)->putDirect(callFrame->globalData(), ident, stackFrame.args[2].jsValue(), slot); |
| |
| CHECK_FOR_EXCEPTION_AT_END(); |
| } |
| |
| DEFINE_STUB_FUNCTION(JSObject*, op_put_by_id_transition_realloc) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| JSValue baseValue = stackFrame.args[0].jsValue(); |
| int32_t oldSize = stackFrame.args[3].int32(); |
| Structure* newStructure = stackFrame.args[4].structure(); |
| int32_t newSize = newStructure->outOfLineCapacity(); |
| |
| ASSERT(oldSize >= 0); |
| ASSERT(newSize > oldSize); |
| |
| ASSERT(baseValue.isObject()); |
| JSObject* base = asObject(baseValue); |
| JSGlobalData& globalData = *stackFrame.globalData; |
| Butterfly* butterfly = base->growOutOfLineStorage(globalData, oldSize, newSize); |
| base->setButterfly(globalData, butterfly, newStructure); |
| |
| return base; |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_get_by_id) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| CallFrame* callFrame = stackFrame.callFrame; |
| Identifier& ident = stackFrame.args[1].identifier(); |
| |
| CodeBlock* codeBlock = stackFrame.callFrame->codeBlock(); |
| StructureStubInfo* stubInfo = &codeBlock->getStubInfo(STUB_RETURN_ADDRESS); |
| AccessType accessType = static_cast<AccessType>(stubInfo->accessType); |
| |
| JSValue baseValue = stackFrame.args[0].jsValue(); |
| PropertySlot slot(baseValue); |
| JSValue result = baseValue.get(callFrame, ident, slot); |
| |
| if (accessType != static_cast<AccessType>(stubInfo->accessType)) |
| return JSValue::encode(result); |
| |
| if (!stubInfo->seenOnce()) |
| stubInfo->setSeen(); |
| else |
| JITThunks::tryCacheGetByID(callFrame, codeBlock, STUB_RETURN_ADDRESS, baseValue, ident, slot, stubInfo); |
| |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_get_by_id_self_fail) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| Identifier& ident = stackFrame.args[1].identifier(); |
| |
| CodeBlock* codeBlock = callFrame->codeBlock(); |
| StructureStubInfo* stubInfo = &codeBlock->getStubInfo(STUB_RETURN_ADDRESS); |
| AccessType accessType = static_cast<AccessType>(stubInfo->accessType); |
| |
| JSValue baseValue = stackFrame.args[0].jsValue(); |
| PropertySlot slot(baseValue); |
| JSValue result = baseValue.get(callFrame, ident, slot); |
| |
| if (accessType != static_cast<AccessType>(stubInfo->accessType)) |
| return JSValue::encode(result); |
| |
| CHECK_FOR_EXCEPTION(); |
| |
| if (baseValue.isCell() |
| && slot.isCacheable() |
| && !baseValue.asCell()->structure()->isUncacheableDictionary() |
| && slot.slotBase() == baseValue) { |
| |
| ASSERT(slot.slotBase().isObject()); |
| |
| PolymorphicAccessStructureList* polymorphicStructureList; |
| int listIndex = 1; |
| |
| if (stubInfo->accessType == access_get_by_id_self) { |
| ASSERT(!stubInfo->stubRoutine); |
| polymorphicStructureList = new PolymorphicAccessStructureList(callFrame->globalData(), codeBlock->ownerExecutable(), 0, stubInfo->u.getByIdSelf.baseObjectStructure.get(), true); |
| stubInfo->initGetByIdSelfList(polymorphicStructureList, 1); |
| } else { |
| polymorphicStructureList = stubInfo->u.getByIdSelfList.structureList; |
| listIndex = stubInfo->u.getByIdSelfList.listSize; |
| } |
| if (listIndex < POLYMORPHIC_LIST_CACHE_SIZE) { |
| stubInfo->u.getByIdSelfList.listSize++; |
| JIT::compileGetByIdSelfList(callFrame->scope()->globalData(), codeBlock, stubInfo, polymorphicStructureList, listIndex, baseValue.asCell()->structure(), ident, slot, slot.cachedOffset()); |
| |
| if (listIndex == (POLYMORPHIC_LIST_CACHE_SIZE - 1)) |
| ctiPatchCallByReturnAddress(codeBlock, STUB_RETURN_ADDRESS, FunctionPtr(cti_op_get_by_id_generic)); |
| } |
| } else |
| ctiPatchCallByReturnAddress(callFrame->codeBlock(), STUB_RETURN_ADDRESS, FunctionPtr(cti_op_get_by_id_generic)); |
| return JSValue::encode(result); |
| } |
| |
| static PolymorphicAccessStructureList* getPolymorphicAccessStructureListSlot(JSGlobalData& globalData, ScriptExecutable* owner, StructureStubInfo* stubInfo, int& listIndex) |
| { |
| PolymorphicAccessStructureList* prototypeStructureList = 0; |
| listIndex = 1; |
| |
| switch (stubInfo->accessType) { |
| case access_get_by_id_proto: |
| prototypeStructureList = new PolymorphicAccessStructureList(globalData, owner, stubInfo->stubRoutine, stubInfo->u.getByIdProto.baseObjectStructure.get(), stubInfo->u.getByIdProto.prototypeStructure.get(), true); |
| stubInfo->stubRoutine.clear(); |
| stubInfo->initGetByIdProtoList(prototypeStructureList, 2); |
| break; |
| case access_get_by_id_chain: |
| prototypeStructureList = new PolymorphicAccessStructureList(globalData, owner, stubInfo->stubRoutine, stubInfo->u.getByIdChain.baseObjectStructure.get(), stubInfo->u.getByIdChain.chain.get(), true); |
| stubInfo->stubRoutine.clear(); |
| stubInfo->initGetByIdProtoList(prototypeStructureList, 2); |
| break; |
| case access_get_by_id_proto_list: |
| prototypeStructureList = stubInfo->u.getByIdProtoList.structureList; |
| listIndex = stubInfo->u.getByIdProtoList.listSize; |
| if (listIndex < POLYMORPHIC_LIST_CACHE_SIZE) |
| stubInfo->u.getByIdProtoList.listSize++; |
| break; |
| default: |
| ASSERT_NOT_REACHED(); |
| } |
| |
| ASSERT(listIndex <= POLYMORPHIC_LIST_CACHE_SIZE); |
| return prototypeStructureList; |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_get_by_id_getter_stub) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| CallFrame* callFrame = stackFrame.callFrame; |
| GetterSetter* getterSetter = asGetterSetter(stackFrame.args[0].jsObject()); |
| if (!getterSetter->getter()) |
| return JSValue::encode(jsUndefined()); |
| JSObject* getter = asObject(getterSetter->getter()); |
| CallData callData; |
| CallType callType = getter->methodTable()->getCallData(getter, callData); |
| JSValue result = call(callFrame, getter, callType, callData, stackFrame.args[1].jsObject(), ArgList()); |
| if (callFrame->hadException()) |
| returnToThrowTrampoline(&callFrame->globalData(), stackFrame.args[2].returnAddress(), STUB_RETURN_ADDRESS); |
| |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_get_by_id_custom_stub) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| CallFrame* callFrame = stackFrame.callFrame; |
| JSObject* slotBase = stackFrame.args[0].jsObject(); |
| PropertySlot::GetValueFunc getter = reinterpret_cast<PropertySlot::GetValueFunc>(stackFrame.args[1].asPointer); |
| const Identifier& ident = stackFrame.args[2].identifier(); |
| JSValue result = getter(callFrame, slotBase, ident); |
| if (callFrame->hadException()) |
| returnToThrowTrampoline(&callFrame->globalData(), stackFrame.args[3].returnAddress(), STUB_RETURN_ADDRESS); |
| |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_get_by_id_proto_list) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| const Identifier& propertyName = stackFrame.args[1].identifier(); |
| |
| CodeBlock* codeBlock = callFrame->codeBlock(); |
| StructureStubInfo* stubInfo = &codeBlock->getStubInfo(STUB_RETURN_ADDRESS); |
| AccessType accessType = static_cast<AccessType>(stubInfo->accessType); |
| |
| JSValue baseValue = stackFrame.args[0].jsValue(); |
| PropertySlot slot(baseValue); |
| JSValue result = baseValue.get(callFrame, propertyName, slot); |
| |
| CHECK_FOR_EXCEPTION(); |
| |
| if (accessType != static_cast<AccessType>(stubInfo->accessType) |
| || !baseValue.isCell() |
| || !slot.isCacheable() |
| || baseValue.asCell()->structure()->isDictionary() |
| || baseValue.asCell()->structure()->typeInfo().prohibitsPropertyCaching()) { |
| ctiPatchCallByReturnAddress(callFrame->codeBlock(), STUB_RETURN_ADDRESS, FunctionPtr(cti_op_get_by_id_proto_fail)); |
| return JSValue::encode(result); |
| } |
| |
| Structure* structure = baseValue.asCell()->structure(); |
| |
| ASSERT(slot.slotBase().isObject()); |
| JSObject* slotBaseObject = asObject(slot.slotBase()); |
| |
| PropertyOffset offset = slot.cachedOffset(); |
| |
| if (slot.slotBase() == baseValue) |
| ctiPatchCallByReturnAddress(codeBlock, STUB_RETURN_ADDRESS, FunctionPtr(cti_op_get_by_id_proto_fail)); |
| else if (slot.slotBase() == baseValue.asCell()->structure()->prototypeForLookup(callFrame)) { |
| ASSERT(!baseValue.asCell()->structure()->isDictionary()); |
| |
| if (baseValue.asCell()->structure()->typeInfo().hasImpureGetOwnPropertySlot()) { |
| ctiPatchCallByReturnAddress(codeBlock, STUB_RETURN_ADDRESS, FunctionPtr(cti_op_get_by_id_proto_fail)); |
| return JSValue::encode(result); |
| } |
| |
| // Since we're accessing a prototype in a loop, it's a good bet that it |
| // should not be treated as a dictionary. |
| if (slotBaseObject->structure()->isDictionary()) { |
| slotBaseObject->flattenDictionaryObject(callFrame->globalData()); |
| offset = slotBaseObject->structure()->get(callFrame->globalData(), propertyName); |
| } |
| |
| int listIndex; |
| PolymorphicAccessStructureList* prototypeStructureList = getPolymorphicAccessStructureListSlot(callFrame->globalData(), codeBlock->ownerExecutable(), stubInfo, listIndex); |
| if (listIndex < POLYMORPHIC_LIST_CACHE_SIZE) { |
| JIT::compileGetByIdProtoList(callFrame->scope()->globalData(), callFrame, codeBlock, stubInfo, prototypeStructureList, listIndex, structure, slotBaseObject->structure(), propertyName, slot, offset); |
| |
| if (listIndex == (POLYMORPHIC_LIST_CACHE_SIZE - 1)) |
| ctiPatchCallByReturnAddress(codeBlock, STUB_RETURN_ADDRESS, FunctionPtr(cti_op_get_by_id_proto_list_full)); |
| } |
| } else { |
| size_t count = normalizePrototypeChainForChainAccess(callFrame, baseValue, slot.slotBase(), propertyName, offset); |
| if (count == InvalidPrototypeChain) { |
| ctiPatchCallByReturnAddress(codeBlock, STUB_RETURN_ADDRESS, FunctionPtr(cti_op_get_by_id_proto_fail)); |
| return JSValue::encode(result); |
| } |
| |
| ASSERT(!baseValue.asCell()->structure()->isDictionary()); |
| int listIndex; |
| PolymorphicAccessStructureList* prototypeStructureList = getPolymorphicAccessStructureListSlot(callFrame->globalData(), codeBlock->ownerExecutable(), stubInfo, listIndex); |
| |
| if (listIndex < POLYMORPHIC_LIST_CACHE_SIZE) { |
| StructureChain* protoChain = structure->prototypeChain(callFrame); |
| JIT::compileGetByIdChainList(callFrame->scope()->globalData(), callFrame, codeBlock, stubInfo, prototypeStructureList, listIndex, structure, protoChain, count, propertyName, slot, offset); |
| |
| if (listIndex == (POLYMORPHIC_LIST_CACHE_SIZE - 1)) |
| ctiPatchCallByReturnAddress(codeBlock, STUB_RETURN_ADDRESS, FunctionPtr(cti_op_get_by_id_proto_list_full)); |
| } |
| } |
| |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_get_by_id_proto_list_full) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| JSValue baseValue = stackFrame.args[0].jsValue(); |
| PropertySlot slot(baseValue); |
| JSValue result = baseValue.get(stackFrame.callFrame, stackFrame.args[1].identifier(), slot); |
| |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_get_by_id_proto_fail) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| JSValue baseValue = stackFrame.args[0].jsValue(); |
| PropertySlot slot(baseValue); |
| JSValue result = baseValue.get(stackFrame.callFrame, stackFrame.args[1].identifier(), slot); |
| |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_get_by_id_array_fail) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| JSValue baseValue = stackFrame.args[0].jsValue(); |
| PropertySlot slot(baseValue); |
| JSValue result = baseValue.get(stackFrame.callFrame, stackFrame.args[1].identifier(), slot); |
| |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_get_by_id_string_fail) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| JSValue baseValue = stackFrame.args[0].jsValue(); |
| PropertySlot slot(baseValue); |
| JSValue result = baseValue.get(stackFrame.callFrame, stackFrame.args[1].identifier(), slot); |
| |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_check_has_instance) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| JSValue value = stackFrame.args[0].jsValue(); |
| JSValue baseVal = stackFrame.args[1].jsValue(); |
| |
| if (baseVal.isObject()) { |
| JSObject* baseObject = asObject(baseVal); |
| ASSERT(!baseObject->structure()->typeInfo().implementsDefaultHasInstance()); |
| if (baseObject->structure()->typeInfo().implementsHasInstance()) { |
| bool result = baseObject->methodTable()->customHasInstance(baseObject, callFrame, value); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return JSValue::encode(jsBoolean(result)); |
| } |
| } |
| |
| stackFrame.globalData->exception = createInvalidParamError(callFrame, "instanceof", baseVal); |
| VM_THROW_EXCEPTION_AT_END(); |
| return JSValue::encode(JSValue()); |
| } |
| |
| #if ENABLE(DFG_JIT) |
| DEFINE_STUB_FUNCTION(void, optimize) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| CodeBlock* codeBlock = callFrame->codeBlock(); |
| unsigned bytecodeIndex = stackFrame.args[0].int32(); |
| |
| #if ENABLE(JIT_VERBOSE_OSR) |
| dataLog( |
| *codeBlock, ": Entered optimize with bytecodeIndex = ", bytecodeIndex, |
| ", executeCounter = ", codeBlock->jitExecuteCounter(), |
| ", optimizationDelayCounter = ", codeBlock->reoptimizationRetryCounter(), |
| ", exitCounter = "); |
| if (codeBlock->hasOptimizedReplacement()) |
| dataLog(codeBlock->replacement()->osrExitCounter()); |
| else |
| dataLog("N/A"); |
| dataLog("\n"); |
| #endif |
| |
| if (!codeBlock->checkIfOptimizationThresholdReached()) { |
| codeBlock->updateAllPredictions(); |
| #if ENABLE(JIT_VERBOSE_OSR) |
| dataLog("Choosing not to optimize ", *codeBlock, " yet.\n"); |
| #endif |
| return; |
| } |
| |
| if (codeBlock->hasOptimizedReplacement()) { |
| #if ENABLE(JIT_VERBOSE_OSR) |
| dataLog("Considering OSR ", *codeBlock, " -> ", *codeBlock->replacement(), ".\n"); |
| #endif |
| // If we have an optimized replacement, then it must be the case that we entered |
| // cti_optimize from a loop. That's because is there's an optimized replacement, |
| // then all calls to this function will be relinked to the replacement and so |
| // the prologue OSR will never fire. |
| |
| // This is an interesting threshold check. Consider that a function OSR exits |
| // in the middle of a loop, while having a relatively low exit count. The exit |
| // will reset the execution counter to some target threshold, meaning that this |
| // code won't be reached until that loop heats up for >=1000 executions. But then |
| // we do a second check here, to see if we should either reoptimize, or just |
| // attempt OSR entry. Hence it might even be correct for |
| // shouldReoptimizeFromLoopNow() to always return true. But we make it do some |
| // additional checking anyway, to reduce the amount of recompilation thrashing. |
| if (codeBlock->replacement()->shouldReoptimizeFromLoopNow()) { |
| #if ENABLE(JIT_VERBOSE_OSR) |
| dataLog("Triggering reoptimization of ", *codeBlock, "(", *codeBlock->replacement(), ") (in loop).\n"); |
| #endif |
| codeBlock->reoptimize(); |
| return; |
| } |
| } else { |
| if (!codeBlock->shouldOptimizeNow()) { |
| #if ENABLE(JIT_VERBOSE_OSR) |
| dataLog("Delaying optimization for ", *codeBlock, " (in loop) because of insufficient profiling.\n"); |
| #endif |
| return; |
| } |
| |
| #if ENABLE(JIT_VERBOSE_OSR) |
| dataLog("Triggering optimized compilation of ", *codeBlock, "\n"); |
| #endif |
| |
| JSScope* scope = callFrame->scope(); |
| JSObject* error = codeBlock->compileOptimized(callFrame, scope, bytecodeIndex); |
| #if ENABLE(JIT_VERBOSE_OSR) |
| if (error) |
| dataLog("WARNING: optimized compilation failed.\n"); |
| #else |
| UNUSED_PARAM(error); |
| #endif |
| |
| if (codeBlock->replacement() == codeBlock) { |
| #if ENABLE(JIT_VERBOSE_OSR) |
| dataLog("Optimizing ", *codeBlock, " failed.\n"); |
| #endif |
| |
| ASSERT(codeBlock->getJITType() == JITCode::BaselineJIT); |
| codeBlock->dontOptimizeAnytimeSoon(); |
| return; |
| } |
| } |
| |
| CodeBlock* optimizedCodeBlock = codeBlock->replacement(); |
| ASSERT(optimizedCodeBlock->getJITType() == JITCode::DFGJIT); |
| |
| if (void* address = DFG::prepareOSREntry(callFrame, optimizedCodeBlock, bytecodeIndex)) { |
| if (Options::showDFGDisassembly()) { |
| dataLog( |
| "Performing OSR ", *codeBlock, " -> ", *optimizedCodeBlock, ", address ", |
| RawPointer((STUB_RETURN_ADDRESS).value()), " -> ", RawPointer(address), ".\n"); |
| } |
| #if ENABLE(JIT_VERBOSE_OSR) |
| dataLog("Optimizing ", *codeBlock, " succeeded, performing OSR after a delay of ", codeBlock->optimizationDelayCounter(), ".\n"); |
| #endif |
| |
| codeBlock->optimizeSoon(); |
| STUB_SET_RETURN_ADDRESS(address); |
| return; |
| } |
| |
| #if ENABLE(JIT_VERBOSE_OSR) |
| dataLog("Optimizing ", *codeBlock, " succeeded, OSR failed, after a delay of ", codeBlock->optimizationDelayCounter(), ".\n"); |
| #endif |
| |
| // Count the OSR failure as a speculation failure. If this happens a lot, then |
| // reoptimize. |
| optimizedCodeBlock->countOSRExit(); |
| |
| #if ENABLE(JIT_VERBOSE_OSR) |
| dataLog("Encountered OSR failure ", *codeBlock, " -> ", *codeBlock->replacement(), ".\n"); |
| #endif |
| |
| // We are a lot more conservative about triggering reoptimization after OSR failure than |
| // before it. If we enter the optimize_from_loop trigger with a bucket full of fail |
| // already, then we really would like to reoptimize immediately. But this case covers |
| // something else: there weren't many (or any) speculation failures before, but we just |
| // failed to enter the speculative code because some variable had the wrong value or |
| // because the OSR code decided for any spurious reason that it did not want to OSR |
| // right now. So, we only trigger reoptimization only upon the more conservative (non-loop) |
| // reoptimization trigger. |
| if (optimizedCodeBlock->shouldReoptimizeNow()) { |
| #if ENABLE(JIT_VERBOSE_OSR) |
| dataLog("Triggering reoptimization of ", *codeBlock, " -> ", *codeBlock->replacement(), " (after OSR fail).\n"); |
| #endif |
| codeBlock->reoptimize(); |
| return; |
| } |
| |
| // OSR failed this time, but it might succeed next time! Let the code run a bit |
| // longer and then try again. |
| codeBlock->optimizeAfterWarmUp(); |
| } |
| #endif // ENABLE(DFG_JIT) |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_instanceof) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| JSValue value = stackFrame.args[0].jsValue(); |
| JSValue proto = stackFrame.args[1].jsValue(); |
| |
| ASSERT(!value.isObject() || !proto.isObject()); |
| |
| bool result = JSObject::defaultHasInstance(callFrame, value, proto); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return JSValue::encode(jsBoolean(result)); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_del_by_id) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| |
| JSObject* baseObj = stackFrame.args[0].jsValue().toObject(callFrame); |
| |
| bool couldDelete = baseObj->methodTable()->deleteProperty(baseObj, callFrame, stackFrame.args[1].identifier()); |
| JSValue result = jsBoolean(couldDelete); |
| if (!couldDelete && callFrame->codeBlock()->isStrictMode()) |
| stackFrame.globalData->exception = createTypeError(stackFrame.callFrame, "Unable to delete property."); |
| |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_mul) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| JSValue src1 = stackFrame.args[0].jsValue(); |
| JSValue src2 = stackFrame.args[1].jsValue(); |
| |
| if (src1.isNumber() && src2.isNumber()) |
| return JSValue::encode(jsNumber(src1.asNumber() * src2.asNumber())); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| JSValue result = jsNumber(src1.toNumber(callFrame) * src2.toNumber(callFrame)); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(JSObject*, op_new_func) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| ASSERT(stackFrame.callFrame->codeBlock()->codeType() != FunctionCode || !stackFrame.callFrame->codeBlock()->needsFullScopeChain() || stackFrame.callFrame->uncheckedR(stackFrame.callFrame->codeBlock()->activationRegister()).jsValue()); |
| return JSFunction::create(stackFrame.callFrame, stackFrame.args[0].function(), stackFrame.callFrame->scope()); |
| } |
| |
| inline void* jitCompileFor(CallFrame* callFrame, CodeSpecializationKind kind) |
| { |
| // This function is called by cti_op_call_jitCompile() and |
| // cti_op_construct_jitCompile() JIT glue trampolines to compile the |
| // callee function that we want to call. Both cti glue trampolines are |
| // called by JIT'ed code which has pushed a frame and initialized most of |
| // the frame content except for the codeBlock. |
| // |
| // Normally, the prologue of the callee is supposed to set the frame's cb |
| // pointer to the cb of the callee. But in this case, the callee code does |
| // not exist yet until it is compiled below. The compilation process will |
| // allocate memory which may trigger a GC. The GC, in turn, will scan the |
| // JSStack, and will expect the frame's cb to either be valid or 0. If |
| // we don't initialize it, the GC will be accessing invalid memory and may |
| // crash. |
| // |
| // Hence, we should nullify it here before proceeding with the compilation. |
| callFrame->setCodeBlock(0); |
| |
| JSFunction* function = jsCast<JSFunction*>(callFrame->callee()); |
| ASSERT(!function->isHostFunction()); |
| FunctionExecutable* executable = function->jsExecutable(); |
| JSScope* callDataScopeChain = function->scope(); |
| JSObject* error = executable->compileFor(callFrame, callDataScopeChain, kind); |
| if (!error) |
| return function; |
| callFrame->globalData().exception = error; |
| return 0; |
| } |
| |
| DEFINE_STUB_FUNCTION(void*, op_call_jitCompile) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| #if !ASSERT_DISABLED |
| CallData callData; |
| ASSERT(stackFrame.callFrame->callee()->methodTable()->getCallData(stackFrame.callFrame->callee(), callData) == CallTypeJS); |
| #endif |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| void* result = jitCompileFor(callFrame, CodeForCall); |
| if (!result) |
| return throwExceptionFromOpCall<void*>(stackFrame, callFrame, STUB_RETURN_ADDRESS); |
| |
| return result; |
| } |
| |
| DEFINE_STUB_FUNCTION(void*, op_construct_jitCompile) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| #if !ASSERT_DISABLED |
| ConstructData constructData; |
| ASSERT(jsCast<JSFunction*>(stackFrame.callFrame->callee())->methodTable()->getConstructData(stackFrame.callFrame->callee(), constructData) == ConstructTypeJS); |
| #endif |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| void* result = jitCompileFor(callFrame, CodeForConstruct); |
| if (!result) |
| return throwExceptionFromOpCall<void*>(stackFrame, callFrame, STUB_RETURN_ADDRESS); |
| |
| return result; |
| } |
| |
| DEFINE_STUB_FUNCTION(void*, op_call_arityCheck) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| |
| CallFrame* newCallFrame = CommonSlowPaths::arityCheckFor(callFrame, stackFrame.stack, CodeForCall); |
| if (!newCallFrame) |
| return throwExceptionFromOpCall<void*>(stackFrame, callFrame, STUB_RETURN_ADDRESS, createStackOverflowError(callFrame->callerFrame())); |
| |
| return newCallFrame; |
| } |
| |
| DEFINE_STUB_FUNCTION(void*, op_construct_arityCheck) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| |
| CallFrame* newCallFrame = CommonSlowPaths::arityCheckFor(callFrame, stackFrame.stack, CodeForConstruct); |
| if (!newCallFrame) |
| return throwExceptionFromOpCall<void*>(stackFrame, callFrame, STUB_RETURN_ADDRESS, createStackOverflowError(callFrame->callerFrame())); |
| |
| return newCallFrame; |
| } |
| |
| inline void* lazyLinkFor(CallFrame* callFrame, CodeSpecializationKind kind) |
| { |
| JSFunction* callee = jsCast<JSFunction*>(callFrame->callee()); |
| ExecutableBase* executable = callee->executable(); |
| |
| MacroAssemblerCodePtr codePtr; |
| CodeBlock* codeBlock = 0; |
| CallLinkInfo* callLinkInfo = &callFrame->callerFrame()->codeBlock()->getCallLinkInfo(callFrame->returnPC()); |
| |
| // This function is called by cti_vm_lazyLinkCall() and |
| // cti_lazyLinkConstruct JIT glue trampolines to link the callee function |
| // that we want to call. Both cti glue trampolines are called by JIT'ed |
| // code which has pushed a frame and initialized most of the frame content |
| // except for the codeBlock. |
| // |
| // Normally, the prologue of the callee is supposed to set the frame's cb |
| // field to the cb of the callee. But in this case, the callee may not |
| // exist yet, and if not, it will be generated in the compilation below. |
| // The compilation will allocate memory which may trigger a GC. The GC, in |
| // turn, will scan the JSStack, and will expect the frame's cb to be valid |
| // or 0. If we don't initialize it, the GC will be accessing invalid |
| // memory and may crash. |
| // |
| // Hence, we should nullify it here before proceeding with the compilation. |
| callFrame->setCodeBlock(0); |
| |
| if (executable->isHostFunction()) |
| codePtr = executable->generatedJITCodeFor(kind).addressForCall(); |
| else { |
| FunctionExecutable* functionExecutable = static_cast<FunctionExecutable*>(executable); |
| if (JSObject* error = functionExecutable->compileFor(callFrame, callee->scope(), kind)) { |
| callFrame->globalData().exception = error; |
| return 0; |
| } |
| codeBlock = &functionExecutable->generatedBytecodeFor(kind); |
| if (callFrame->argumentCountIncludingThis() < static_cast<size_t>(codeBlock->numParameters()) |
| || callLinkInfo->callType == CallLinkInfo::CallVarargs) |
| codePtr = functionExecutable->generatedJITCodeWithArityCheckFor(kind); |
| else |
| codePtr = functionExecutable->generatedJITCodeFor(kind).addressForCall(); |
| } |
| |
| if (!callLinkInfo->seenOnce()) |
| callLinkInfo->setSeen(); |
| else |
| JIT::linkFor(callee, callFrame->callerFrame()->codeBlock(), codeBlock, codePtr, callLinkInfo, &callFrame->globalData(), kind); |
| |
| return codePtr.executableAddress(); |
| } |
| |
| DEFINE_STUB_FUNCTION(void*, vm_lazyLinkCall) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| void* result = lazyLinkFor(callFrame, CodeForCall); |
| if (!result) |
| return throwExceptionFromOpCall<void*>(stackFrame, callFrame, STUB_RETURN_ADDRESS); |
| |
| return result; |
| } |
| |
| DEFINE_STUB_FUNCTION(void*, vm_lazyLinkConstruct) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| void* result = lazyLinkFor(callFrame, CodeForConstruct); |
| if (!result) |
| return throwExceptionFromOpCall<void*>(stackFrame, callFrame, STUB_RETURN_ADDRESS); |
| |
| return result; |
| } |
| |
| DEFINE_STUB_FUNCTION(JSObject*, op_push_activation) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| JSActivation* activation = JSActivation::create(stackFrame.callFrame->globalData(), stackFrame.callFrame, stackFrame.callFrame->codeBlock()); |
| stackFrame.callFrame->setScope(activation); |
| return activation; |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_call_NotJSFunction) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| |
| JSValue callee = callFrame->calleeAsValue(); |
| |
| CallData callData; |
| CallType callType = getCallData(callee, callData); |
| |
| ASSERT(callType != CallTypeJS); |
| if (callType != CallTypeHost) { |
| ASSERT(callType == CallTypeNone); |
| return throwExceptionFromOpCall<EncodedJSValue>(stackFrame, callFrame, STUB_RETURN_ADDRESS, createNotAFunctionError(callFrame->callerFrame(), callee)); |
| } |
| |
| EncodedJSValue returnValue; |
| { |
| SamplingTool::CallRecord callRecord(CTI_SAMPLER, true); |
| returnValue = callData.native.function(callFrame); |
| } |
| |
| if (stackFrame.globalData->exception) |
| return throwExceptionFromOpCall<EncodedJSValue>(stackFrame, callFrame, STUB_RETURN_ADDRESS); |
| |
| return returnValue; |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_create_arguments) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| Arguments* arguments = Arguments::create(*stackFrame.globalData, stackFrame.callFrame); |
| return JSValue::encode(JSValue(arguments)); |
| } |
| |
| DEFINE_STUB_FUNCTION(void, op_tear_off_activation) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| ASSERT(stackFrame.callFrame->codeBlock()->needsFullScopeChain()); |
| jsCast<JSActivation*>(stackFrame.args[0].jsValue())->tearOff(*stackFrame.globalData); |
| } |
| |
| DEFINE_STUB_FUNCTION(void, op_tear_off_arguments) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| ASSERT(callFrame->codeBlock()->usesArguments()); |
| Arguments* arguments = jsCast<Arguments*>(stackFrame.args[0].jsValue()); |
| if (JSValue activationValue = stackFrame.args[1].jsValue()) { |
| arguments->didTearOffActivation(callFrame, jsCast<JSActivation*>(activationValue)); |
| return; |
| } |
| arguments->tearOff(callFrame); |
| } |
| |
| DEFINE_STUB_FUNCTION(void, op_profile_will_call) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| if (LegacyProfiler* profiler = stackFrame.globalData->enabledProfiler()) |
| profiler->willExecute(stackFrame.callFrame, stackFrame.args[0].jsValue()); |
| } |
| |
| DEFINE_STUB_FUNCTION(void, op_profile_did_call) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| if (LegacyProfiler* profiler = stackFrame.globalData->enabledProfiler()) |
| profiler->didExecute(stackFrame.callFrame, stackFrame.args[0].jsValue()); |
| } |
| |
| DEFINE_STUB_FUNCTION(JSObject*, op_new_array) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| return constructArray(stackFrame.callFrame, stackFrame.args[2].arrayAllocationProfile(), reinterpret_cast<JSValue*>(&stackFrame.callFrame->registers()[stackFrame.args[0].int32()]), stackFrame.args[1].int32()); |
| } |
| |
| DEFINE_STUB_FUNCTION(JSObject*, op_new_array_with_size) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| return constructArrayWithSizeQuirk(stackFrame.callFrame, stackFrame.args[1].arrayAllocationProfile(), stackFrame.callFrame->lexicalGlobalObject(), stackFrame.args[0].jsValue()); |
| } |
| |
| DEFINE_STUB_FUNCTION(JSObject*, op_new_array_buffer) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| return constructArray(stackFrame.callFrame, stackFrame.args[2].arrayAllocationProfile(), stackFrame.callFrame->codeBlock()->constantBuffer(stackFrame.args[0].int32()), stackFrame.args[1].int32()); |
| } |
| |
| DEFINE_STUB_FUNCTION(void, op_init_global_const_check) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| CodeBlock* codeBlock = callFrame->codeBlock(); |
| symbolTablePut(codeBlock->globalObject(), callFrame, codeBlock->identifier(stackFrame.args[1].int32()), stackFrame.args[0].jsValue(), true); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_resolve) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| |
| JSValue result = JSScope::resolve(callFrame, stackFrame.args[0].identifier(), stackFrame.args[1].resolveOperations()); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(void, op_put_to_base) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| JSValue base = callFrame->r(stackFrame.args[0].int32()).jsValue(); |
| JSValue value = callFrame->r(stackFrame.args[2].int32()).jsValue(); |
| JSScope::resolvePut(callFrame, base, stackFrame.args[1].identifier(), value, stackFrame.args[3].putToBaseOperation()); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_construct_NotJSConstruct) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| JSValue callee = callFrame->calleeAsValue(); |
| |
| ConstructData constructData; |
| ConstructType constructType = getConstructData(callee, constructData); |
| |
| ASSERT(constructType != ConstructTypeJS); |
| if (constructType != ConstructTypeHost) { |
| ASSERT(constructType == ConstructTypeNone); |
| return throwExceptionFromOpCall<EncodedJSValue>(stackFrame, callFrame, STUB_RETURN_ADDRESS, createNotAConstructorError(callFrame->callerFrame(), callee)); |
| } |
| |
| EncodedJSValue returnValue; |
| { |
| SamplingTool::CallRecord callRecord(CTI_SAMPLER, true); |
| returnValue = constructData.native.function(callFrame); |
| } |
| |
| if (stackFrame.globalData->exception) |
| return throwExceptionFromOpCall<EncodedJSValue>(stackFrame, callFrame, STUB_RETURN_ADDRESS); |
| |
| return returnValue; |
| } |
| |
| static JSValue getByVal( |
| CallFrame* callFrame, JSValue baseValue, JSValue subscript, ReturnAddressPtr returnAddress) |
| { |
| if (LIKELY(baseValue.isCell() && subscript.isString())) { |
| if (JSValue result = baseValue.asCell()->fastGetOwnProperty(callFrame, asString(subscript)->value(callFrame))) |
| return result; |
| } |
| |
| if (subscript.isUInt32()) { |
| uint32_t i = subscript.asUInt32(); |
| if (isJSString(baseValue) && asString(baseValue)->canGetIndex(i)) { |
| ctiPatchCallByReturnAddress(callFrame->codeBlock(), returnAddress, FunctionPtr(cti_op_get_by_val_string)); |
| return asString(baseValue)->getIndex(callFrame, i); |
| } |
| return baseValue.get(callFrame, i); |
| } |
| |
| if (isName(subscript)) |
| return baseValue.get(callFrame, jsCast<NameInstance*>(subscript.asCell())->privateName()); |
| |
| Identifier property(callFrame, subscript.toString(callFrame)->value(callFrame)); |
| return baseValue.get(callFrame, property); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_get_by_val) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| |
| JSValue baseValue = stackFrame.args[0].jsValue(); |
| JSValue subscript = stackFrame.args[1].jsValue(); |
| |
| if (baseValue.isObject() && subscript.isInt32()) { |
| // See if it's worth optimizing this at all. |
| JSObject* object = asObject(baseValue); |
| bool didOptimize = false; |
| |
| unsigned bytecodeOffset = callFrame->bytecodeOffsetForNonDFGCode(); |
| ASSERT(bytecodeOffset); |
| ByValInfo& byValInfo = callFrame->codeBlock()->getByValInfo(bytecodeOffset - 1); |
| ASSERT(!byValInfo.stubRoutine); |
| |
| if (hasOptimizableIndexing(object->structure())) { |
| // Attempt to optimize. |
| JITArrayMode arrayMode = jitArrayModeForStructure(object->structure()); |
| if (arrayMode != byValInfo.arrayMode) { |
| JIT::compileGetByVal(&callFrame->globalData(), callFrame->codeBlock(), &byValInfo, STUB_RETURN_ADDRESS, arrayMode); |
| didOptimize = true; |
| } |
| } |
| |
| if (!didOptimize) { |
| // If we take slow path more than 10 times without patching then make sure we |
| // never make that mistake again. Or, if we failed to patch and we have some object |
| // that intercepts indexed get, then don't even wait until 10 times. For cases |
| // where we see non-index-intercepting objects, this gives 10 iterations worth of |
| // opportunity for us to observe that the get_by_val may be polymorphic. |
| if (++byValInfo.slowPathCount >= 10 |
| || object->structure()->typeInfo().interceptsGetOwnPropertySlotByIndexEvenWhenLengthIsNotZero()) { |
| // Don't ever try to optimize. |
| RepatchBuffer repatchBuffer(callFrame->codeBlock()); |
| repatchBuffer.relinkCallerToFunction(STUB_RETURN_ADDRESS, FunctionPtr(cti_op_get_by_val_generic)); |
| } |
| } |
| } |
| |
| JSValue result = getByVal(callFrame, baseValue, subscript, STUB_RETURN_ADDRESS); |
| CHECK_FOR_EXCEPTION(); |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_get_by_val_generic) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| |
| JSValue baseValue = stackFrame.args[0].jsValue(); |
| JSValue subscript = stackFrame.args[1].jsValue(); |
| |
| JSValue result = getByVal(callFrame, baseValue, subscript, STUB_RETURN_ADDRESS); |
| CHECK_FOR_EXCEPTION(); |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_get_by_val_string) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| |
| JSValue baseValue = stackFrame.args[0].jsValue(); |
| JSValue subscript = stackFrame.args[1].jsValue(); |
| |
| JSValue result; |
| |
| if (LIKELY(subscript.isUInt32())) { |
| uint32_t i = subscript.asUInt32(); |
| if (isJSString(baseValue) && asString(baseValue)->canGetIndex(i)) |
| result = asString(baseValue)->getIndex(callFrame, i); |
| else { |
| result = baseValue.get(callFrame, i); |
| if (!isJSString(baseValue)) |
| ctiPatchCallByReturnAddress(callFrame->codeBlock(), STUB_RETURN_ADDRESS, FunctionPtr(cti_op_get_by_val)); |
| } |
| } else if (isName(subscript)) |
| result = baseValue.get(callFrame, jsCast<NameInstance*>(subscript.asCell())->privateName()); |
| else { |
| Identifier property(callFrame, subscript.toString(callFrame)->value(callFrame)); |
| result = baseValue.get(callFrame, property); |
| } |
| |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_sub) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| JSValue src1 = stackFrame.args[0].jsValue(); |
| JSValue src2 = stackFrame.args[1].jsValue(); |
| |
| if (src1.isNumber() && src2.isNumber()) |
| return JSValue::encode(jsNumber(src1.asNumber() - src2.asNumber())); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| JSValue result = jsNumber(src1.toNumber(callFrame) - src2.toNumber(callFrame)); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return JSValue::encode(result); |
| } |
| |
| static void putByVal(CallFrame* callFrame, JSValue baseValue, JSValue subscript, JSValue value) |
| { |
| if (LIKELY(subscript.isUInt32())) { |
| uint32_t i = subscript.asUInt32(); |
| if (baseValue.isObject()) { |
| JSObject* object = asObject(baseValue); |
| if (object->canSetIndexQuickly(i)) |
| object->setIndexQuickly(callFrame->globalData(), i, value); |
| else |
| object->methodTable()->putByIndex(object, callFrame, i, value, callFrame->codeBlock()->isStrictMode()); |
| } else |
| baseValue.putByIndex(callFrame, i, value, callFrame->codeBlock()->isStrictMode()); |
| } else if (isName(subscript)) { |
| PutPropertySlot slot(callFrame->codeBlock()->isStrictMode()); |
| baseValue.put(callFrame, jsCast<NameInstance*>(subscript.asCell())->privateName(), value, slot); |
| } else { |
| Identifier property(callFrame, subscript.toString(callFrame)->value(callFrame)); |
| if (!callFrame->globalData().exception) { // Don't put to an object if toString threw an exception. |
| PutPropertySlot slot(callFrame->codeBlock()->isStrictMode()); |
| baseValue.put(callFrame, property, value, slot); |
| } |
| } |
| } |
| |
| DEFINE_STUB_FUNCTION(void, op_put_by_val) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| |
| JSValue baseValue = stackFrame.args[0].jsValue(); |
| JSValue subscript = stackFrame.args[1].jsValue(); |
| JSValue value = stackFrame.args[2].jsValue(); |
| |
| if (baseValue.isObject() && subscript.isInt32()) { |
| // See if it's worth optimizing at all. |
| JSObject* object = asObject(baseValue); |
| bool didOptimize = false; |
| |
| unsigned bytecodeOffset = callFrame->bytecodeOffsetForNonDFGCode(); |
| ASSERT(bytecodeOffset); |
| ByValInfo& byValInfo = callFrame->codeBlock()->getByValInfo(bytecodeOffset - 1); |
| ASSERT(!byValInfo.stubRoutine); |
| |
| if (hasOptimizableIndexing(object->structure())) { |
| // Attempt to optimize. |
| JITArrayMode arrayMode = jitArrayModeForStructure(object->structure()); |
| if (arrayMode != byValInfo.arrayMode) { |
| JIT::compilePutByVal(&callFrame->globalData(), callFrame->codeBlock(), &byValInfo, STUB_RETURN_ADDRESS, arrayMode); |
| didOptimize = true; |
| } |
| } |
| |
| if (!didOptimize) { |
| // If we take slow path more than 10 times without patching then make sure we |
| // never make that mistake again. Or, if we failed to patch and we have some object |
| // that intercepts indexed get, then don't even wait until 10 times. For cases |
| // where we see non-index-intercepting objects, this gives 10 iterations worth of |
| // opportunity for us to observe that the get_by_val may be polymorphic. |
| if (++byValInfo.slowPathCount >= 10 |
| || object->structure()->typeInfo().interceptsGetOwnPropertySlotByIndexEvenWhenLengthIsNotZero()) { |
| // Don't ever try to optimize. |
| RepatchBuffer repatchBuffer(callFrame->codeBlock()); |
| repatchBuffer.relinkCallerToFunction(STUB_RETURN_ADDRESS, FunctionPtr(cti_op_put_by_val_generic)); |
| } |
| } |
| } |
| |
| putByVal(callFrame, baseValue, subscript, value); |
| |
| CHECK_FOR_EXCEPTION_AT_END(); |
| } |
| |
| DEFINE_STUB_FUNCTION(void, op_put_by_val_generic) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| |
| JSValue baseValue = stackFrame.args[0].jsValue(); |
| JSValue subscript = stackFrame.args[1].jsValue(); |
| JSValue value = stackFrame.args[2].jsValue(); |
| |
| putByVal(callFrame, baseValue, subscript, value); |
| |
| CHECK_FOR_EXCEPTION_AT_END(); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_less) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| JSValue result = jsBoolean(jsLess<true>(callFrame, stackFrame.args[0].jsValue(), stackFrame.args[1].jsValue())); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_lesseq) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| JSValue result = jsBoolean(jsLessEq<true>(callFrame, stackFrame.args[0].jsValue(), stackFrame.args[1].jsValue())); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_greater) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| JSValue result = jsBoolean(jsLess<false>(callFrame, stackFrame.args[1].jsValue(), stackFrame.args[0].jsValue())); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_greatereq) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| JSValue result = jsBoolean(jsLessEq<false>(callFrame, stackFrame.args[1].jsValue(), stackFrame.args[0].jsValue())); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(void*, op_load_varargs) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| JSStack* stack = stackFrame.stack; |
| JSValue thisValue = stackFrame.args[0].jsValue(); |
| JSValue arguments = stackFrame.args[1].jsValue(); |
| int firstFreeRegister = stackFrame.args[2].int32(); |
| |
| CallFrame* newCallFrame = loadVarargs(callFrame, stack, thisValue, arguments, firstFreeRegister); |
| if (!newCallFrame) |
| VM_THROW_EXCEPTION(); |
| return newCallFrame; |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_negate) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| JSValue src = stackFrame.args[0].jsValue(); |
| |
| if (src.isNumber()) |
| return JSValue::encode(jsNumber(-src.asNumber())); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| JSValue result = jsNumber(-src.toNumber(callFrame)); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_resolve_base) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| return JSValue::encode(JSScope::resolveBase(stackFrame.callFrame, stackFrame.args[0].identifier(), false, stackFrame.args[1].resolveOperations(), stackFrame.args[2].putToBaseOperation())); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_resolve_base_strict_put) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| if (JSValue result = JSScope::resolveBase(stackFrame.callFrame, stackFrame.args[0].identifier(), true, stackFrame.args[1].resolveOperations(), stackFrame.args[2].putToBaseOperation())) |
| return JSValue::encode(result); |
| VM_THROW_EXCEPTION(); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_ensure_property_exists) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| JSValue base = stackFrame.callFrame->r(stackFrame.args[0].int32()).jsValue(); |
| JSObject* object = asObject(base); |
| PropertySlot slot(object); |
| ASSERT(stackFrame.callFrame->codeBlock()->isStrictMode()); |
| if (!object->getPropertySlot(stackFrame.callFrame, stackFrame.args[1].identifier(), slot)) { |
| stackFrame.globalData->exception = createErrorForInvalidGlobalAssignment(stackFrame.callFrame, stackFrame.args[1].identifier().string()); |
| VM_THROW_EXCEPTION(); |
| } |
| |
| return JSValue::encode(base); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_div) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| JSValue src1 = stackFrame.args[0].jsValue(); |
| JSValue src2 = stackFrame.args[1].jsValue(); |
| |
| if (src1.isNumber() && src2.isNumber()) |
| return JSValue::encode(jsNumber(src1.asNumber() / src2.asNumber())); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| JSValue result = jsNumber(src1.toNumber(callFrame) / src2.toNumber(callFrame)); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_pre_dec) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| JSValue v = stackFrame.args[0].jsValue(); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| JSValue result = jsNumber(v.toNumber(callFrame) - 1); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(int, op_jless) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| JSValue src1 = stackFrame.args[0].jsValue(); |
| JSValue src2 = stackFrame.args[1].jsValue(); |
| CallFrame* callFrame = stackFrame.callFrame; |
| |
| bool result = jsLess<true>(callFrame, src1, src2); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return result; |
| } |
| |
| DEFINE_STUB_FUNCTION(int, op_jlesseq) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| JSValue src1 = stackFrame.args[0].jsValue(); |
| JSValue src2 = stackFrame.args[1].jsValue(); |
| CallFrame* callFrame = stackFrame.callFrame; |
| |
| bool result = jsLessEq<true>(callFrame, src1, src2); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return result; |
| } |
| |
| DEFINE_STUB_FUNCTION(int, op_jgreater) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| JSValue src1 = stackFrame.args[0].jsValue(); |
| JSValue src2 = stackFrame.args[1].jsValue(); |
| CallFrame* callFrame = stackFrame.callFrame; |
| |
| bool result = jsLess<false>(callFrame, src2, src1); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return result; |
| } |
| |
| DEFINE_STUB_FUNCTION(int, op_jgreatereq) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| JSValue src1 = stackFrame.args[0].jsValue(); |
| JSValue src2 = stackFrame.args[1].jsValue(); |
| CallFrame* callFrame = stackFrame.callFrame; |
| |
| bool result = jsLessEq<false>(callFrame, src2, src1); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return result; |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_not) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| JSValue src = stackFrame.args[0].jsValue(); |
| |
| JSValue result = jsBoolean(!src.toBoolean(stackFrame.callFrame)); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return JSValue::encode(result); |
| } |
| |
| DEFINE_STUB_FUNCTION(int, op_jtrue) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| JSValue src1 = stackFrame.args[0].jsValue(); |
| |
| bool result = src1.toBoolean(stackFrame.callFrame); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| return result; |
| } |
| |
| DEFINE_STUB_FUNCTION(EncodedJSValue, op_post_inc) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| JSValue v = stackFrame.args[0].jsValue(); |
| |
| CallFrame* callFrame = stackFrame.callFrame; |
| |
| double number = v.toNumber(callFrame); |
| CHECK_FOR_EXCEPTION_AT_END(); |
| |
| callFrame->registers()[stackFrame.args[1].int32()] = jsNumber(number + 1); |
| return JSValue::encode(jsNumber(number)); |
| } |
| |
| DEFINE_STUB_FUNCTION(int, op_eq) |
| { |
| STUB_INIT_STACK_FRAME(stackFrame); |
| |
| JSValue src1 = stackFrame.args[0].jsValue(); |
| JSValue src2 = stackFrame.args[1].jsValue(); |
| |
| #if USE(JSVALUE32_64) |
| start: |
| if (src2.isUndefined()) { |
| return src1.isNull() || |
| (src1.isCell() && src1.asCell()->structure()->masqueradesAsUndefined(stackFrame.callFrame->lexicalGlobalObject())) |
| || src1.isUndefined(); |
| } |
| |
| if (src2.isNull()) { |
| return src1.isUndefined() || |
| (src1.isCell() && src1.asCell()->structure()->masqueradesAsUndefined(stackFrame.callFrame->lexicalGlobalObject())) |
| || src1.isNull(); |
| } |
| |
| if (src1.isInt32()) { |
| if (src2.isDouble()) |
| return src1.asInt32() == src2.asDouble(); |
| double d = src2.toNumber(stackFrame.callFrame); |
| CHECK_FOR_EXCEPTION(); |
| return src1.asInt32() == d; |
| } |
| |
| if (src1.isDouble()) { |
| if (src2.isInt32()) |
| return src1.asDouble() == src2.asInt32(); |
| double d = src2.toNumber(stackFrame.callFrame); |
| CHECK_FOR_EXCEPTION(); |
| return src1.asDouble() == d; |
| } |
| |
| if (src1.isTrue()) { |
| if (src2.isFalse()) |
| return false; |
| double d = src2.toNumber(stackFrame.callFrame); |
| CHECK_FOR_EXCEPTION(); |
| return d == 1.0; |
| } |
| |
| if (src1.isFalse()) { |
| if (src2.isTrue()) |
| return false; |
| double d = src2.toNumber(stackFrame.callFrame); |
| CHECK_FOR_EXCEPTION(); |
| return d == 0.0; |
| } |
| |
| if (src1.isUndefined()) |
| return src2.isCell() && src2.asCell()->structure()->masqueradesAsUndefined(stackFrame.callFrame->lexicalGlobalObject()); |
| |
| if (src1.isNull()) |
| return src2.isCell() && src2.asCell()->structure()->masqueradesAsUndefined(stackFrame.callFrame->lexicalGlobalObject()); |
| |
| JSCell* cell1 = src1.asCell(); |
| |
| if (cell1->isString()) { |
| if (src2.isInt32()) |
| return jsToNumber(jsCast<JSString*>(cell1)->value(stackFrame.callFrame)) == src2.asInt32(); |
| |
| if (src2.isDouble()) |
| return jsToNumber(jsCast<JSString*>(cell1)->value(stackFrame.callFrame)) == src2.asDouble(); |
| |
| |