| /* |
| * Copyright (C) 2011-2018 Apple Inc. All rights reserved. |
| * |
| * Redistribution and use in source and binary forms, with or without |
| * modification, are permitted provided that the following conditions |
| * are met: |
| * 1. Redistributions of source code must retain the above copyright |
| * notice, this list of conditions and the following disclaimer. |
| * 2. Redistributions in binary form must reproduce the above copyright |
| * notice, this list of conditions and the following disclaimer in the |
| * documentation and/or other materials provided with the distribution. |
| * |
| * THIS SOFTWARE IS PROVIDED BY APPLE INC. ``AS IS'' AND ANY |
| * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE |
| * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR |
| * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR |
| * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, |
| * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, |
| * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR |
| * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY |
| * OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
| * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE |
| * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| */ |
| |
| #include "config.h" |
| #include "LLIntData.h" |
| |
| #include "ArithProfile.h" |
| #include "BytecodeConventions.h" |
| #include "CodeBlock.h" |
| #include "CodeType.h" |
| #include "Instruction.h" |
| #include "JSScope.h" |
| #include "LLIntCLoop.h" |
| #include "MaxFrameExtentForSlowPathCall.h" |
| #include "Opcode.h" |
| #include "PropertyOffset.h" |
| #include "ShadowChicken.h" |
| #include "WriteBarrier.h" |
| |
| #define STATIC_ASSERT(cond) static_assert(cond, "LLInt assumes " #cond) |
| |
| namespace JSC { namespace LLInt { |
| |
| Instruction Data::s_exceptionInstructions[maxOpcodeLength + 1] = { }; |
| Opcode Data::s_opcodeMap[numOpcodeIDs] = { }; |
| |
| #if ENABLE(JIT) |
| extern "C" void llint_entry(void*); |
| #endif |
| |
| void initialize() |
| { |
| #if !ENABLE(JIT) |
| CLoop::initialize(); |
| |
| #else // ENABLE(JIT) |
| llint_entry(&Data::s_opcodeMap); |
| |
| for (int i = 0; i < NUMBER_OF_BYTECODE_IDS; ++i) { |
| PtrTag tag = (i == op_catch) ? ExceptionHandlerPtrTag : BytecodePtrTag; |
| Data::s_opcodeMap[i] = tagCodePtr(Data::s_opcodeMap[i], tag); |
| } |
| |
| if (VM::canUseJIT()) { |
| for (int i = NUMBER_OF_BYTECODE_IDS; i < NUMBER_OF_BYTECODE_IDS + NUMBER_OF_BYTECODE_HELPER_IDS; ++i) |
| Data::s_opcodeMap[i] = tagCodePtr(Data::s_opcodeMap[i], ptrTag(BytecodeHelperPtrTag, i)); |
| } else { |
| static const PtrTag tagsForOpcode[] = { |
| CodeEntryPtrTag, // llint_program_prologue |
| CodeEntryPtrTag, // llint_eval_prologue |
| CodeEntryPtrTag, // llint_module_program_prologue |
| CodeEntryPtrTag, // llint_function_for_call_prologue |
| CodeEntryPtrTag, // llint_function_for_construct_prologue |
| CodeEntryWithArityCheckPtrTag, // llint_function_for_call_arity_check |
| CodeEntryWithArityCheckPtrTag, // llint_function_for_construct_arity_check |
| CodeEntryPtrTag, // llint_generic_return_point |
| BytecodePtrTag, // llint_throw_from_slow_path_trampoline |
| ExceptionHandlerPtrTag, // llint_throw_during_call_trampoline |
| NativeCodePtrTag, // llint_native_call_trampoline |
| NativeCodePtrTag, // llint_native_construct_trampoline |
| NativeCodePtrTag, // llint_internal_function_call_trampoline |
| NativeCodePtrTag, // llint_internal_function_construct_trampoline |
| ExceptionHandlerPtrTag, // handleUncaughtException |
| }; |
| |
| static_assert(sizeof(tagsForOpcode) / sizeof(tagsForOpcode[0]) == NUMBER_OF_BYTECODE_HELPER_IDS, ""); |
| static_assert(static_cast<uintptr_t>(llint_program_prologue) == NUMBER_OF_BYTECODE_IDS, ""); |
| |
| for (int i = 0; i < NUMBER_OF_BYTECODE_HELPER_IDS; ++i) { |
| int opcodeID = i + NUMBER_OF_BYTECODE_IDS; |
| Data::s_opcodeMap[opcodeID] = tagCodePtr(Data::s_opcodeMap[opcodeID], tagsForOpcode[i]); |
| } |
| } |
| |
| void* handler = LLInt::getCodePtr(llint_throw_from_slow_path_trampoline); |
| for (int i = 0; i < maxOpcodeLength + 1; ++i) |
| Data::s_exceptionInstructions[i].u.pointer = handler; |
| |
| #endif // ENABLE(JIT) |
| } |
| |
| #if COMPILER(CLANG) |
| #pragma clang diagnostic push |
| #pragma clang diagnostic ignored "-Wmissing-noreturn" |
| #endif |
| void Data::performAssertions(VM& vm) |
| { |
| UNUSED_PARAM(vm); |
| |
| // Assertions to match LowLevelInterpreter.asm. If you change any of this code, be |
| // prepared to change LowLevelInterpreter.asm as well!! |
| |
| #if USE(JSVALUE64) |
| const ptrdiff_t PtrSize = 8; |
| const ptrdiff_t CallFrameHeaderSlots = 5; |
| #else // USE(JSVALUE64) // i.e. 32-bit version |
| const ptrdiff_t PtrSize = 4; |
| const ptrdiff_t CallFrameHeaderSlots = 4; |
| #endif |
| const ptrdiff_t SlotSize = 8; |
| |
| STATIC_ASSERT(sizeof(void*) == PtrSize); |
| STATIC_ASSERT(sizeof(Register) == SlotSize); |
| STATIC_ASSERT(CallFrame::headerSizeInRegisters == CallFrameHeaderSlots); |
| |
| ASSERT(!CallFrame::callerFrameOffset()); |
| STATIC_ASSERT(CallerFrameAndPC::sizeInRegisters == (PtrSize * 2) / SlotSize); |
| ASSERT(CallFrame::returnPCOffset() == CallFrame::callerFrameOffset() + PtrSize); |
| ASSERT(CallFrameSlot::codeBlock * sizeof(Register) == CallFrame::returnPCOffset() + PtrSize); |
| STATIC_ASSERT(CallFrameSlot::callee * sizeof(Register) == CallFrameSlot::codeBlock * sizeof(Register) + SlotSize); |
| STATIC_ASSERT(CallFrameSlot::argumentCount * sizeof(Register) == CallFrameSlot::callee * sizeof(Register) + SlotSize); |
| STATIC_ASSERT(CallFrameSlot::thisArgument * sizeof(Register) == CallFrameSlot::argumentCount * sizeof(Register) + SlotSize); |
| STATIC_ASSERT(CallFrame::headerSizeInRegisters == CallFrameSlot::thisArgument); |
| |
| ASSERT(CallFrame::argumentOffsetIncludingThis(0) == CallFrameSlot::thisArgument); |
| |
| #if CPU(BIG_ENDIAN) |
| ASSERT(OBJECT_OFFSETOF(EncodedValueDescriptor, asBits.tag) == 0); |
| ASSERT(OBJECT_OFFSETOF(EncodedValueDescriptor, asBits.payload) == 4); |
| #else |
| ASSERT(OBJECT_OFFSETOF(EncodedValueDescriptor, asBits.tag) == 4); |
| ASSERT(OBJECT_OFFSETOF(EncodedValueDescriptor, asBits.payload) == 0); |
| #endif |
| #if USE(JSVALUE32_64) |
| STATIC_ASSERT(JSValue::Int32Tag == static_cast<unsigned>(-1)); |
| STATIC_ASSERT(JSValue::BooleanTag == static_cast<unsigned>(-2)); |
| STATIC_ASSERT(JSValue::NullTag == static_cast<unsigned>(-3)); |
| STATIC_ASSERT(JSValue::UndefinedTag == static_cast<unsigned>(-4)); |
| STATIC_ASSERT(JSValue::CellTag == static_cast<unsigned>(-5)); |
| STATIC_ASSERT(JSValue::EmptyValueTag == static_cast<unsigned>(-6)); |
| STATIC_ASSERT(JSValue::DeletedValueTag == static_cast<unsigned>(-7)); |
| STATIC_ASSERT(JSValue::LowestTag == static_cast<unsigned>(-7)); |
| #else |
| STATIC_ASSERT(TagBitTypeOther == 0x2); |
| STATIC_ASSERT(TagBitBool == 0x4); |
| STATIC_ASSERT(TagBitUndefined == 0x8); |
| STATIC_ASSERT(ValueEmpty == 0x0); |
| STATIC_ASSERT(ValueFalse == (TagBitTypeOther | TagBitBool)); |
| STATIC_ASSERT(ValueTrue == (TagBitTypeOther | TagBitBool | 1)); |
| STATIC_ASSERT(ValueUndefined == (TagBitTypeOther | TagBitUndefined)); |
| STATIC_ASSERT(ValueNull == TagBitTypeOther); |
| #endif |
| #if (CPU(X86_64) && !OS(WINDOWS)) || CPU(ARM64) || !ENABLE(JIT) |
| STATIC_ASSERT(!maxFrameExtentForSlowPathCall); |
| #elif CPU(ARM) |
| STATIC_ASSERT(maxFrameExtentForSlowPathCall == 24); |
| #elif CPU(X86) || CPU(MIPS) |
| STATIC_ASSERT(maxFrameExtentForSlowPathCall == 40); |
| #elif CPU(X86_64) && OS(WINDOWS) |
| STATIC_ASSERT(maxFrameExtentForSlowPathCall == 64); |
| #endif |
| |
| #if !ENABLE(JIT) || USE(JSVALUE32_64) |
| ASSERT(!CodeBlock::llintBaselineCalleeSaveSpaceAsVirtualRegisters()); |
| #elif (CPU(X86_64) && !OS(WINDOWS)) || CPU(ARM64) |
| ASSERT(CodeBlock::llintBaselineCalleeSaveSpaceAsVirtualRegisters() == 3); |
| #elif (CPU(X86_64) && OS(WINDOWS)) |
| ASSERT(CodeBlock::llintBaselineCalleeSaveSpaceAsVirtualRegisters() == 3); |
| #endif |
| |
| ASSERT(!(reinterpret_cast<ptrdiff_t>((reinterpret_cast<WriteBarrier<JSCell>*>(0x4000)->slot())) - 0x4000)); |
| |
| // FIXME: make these assertions less horrible. |
| #if !ASSERT_DISABLED |
| Vector<int> testVector; |
| testVector.resize(42); |
| ASSERT(bitwise_cast<uint32_t*>(&testVector)[sizeof(void*)/sizeof(uint32_t) + 1] == 42); |
| ASSERT(bitwise_cast<int**>(&testVector)[0] == testVector.begin()); |
| #endif |
| |
| ASSERT(StringImpl::s_hashFlag8BitBuffer == 8); |
| |
| { |
| uint32_t bits = 0x480000; |
| UNUSED_PARAM(bits); |
| ArithProfile arithProfile; |
| arithProfile.lhsSawInt32(); |
| arithProfile.rhsSawInt32(); |
| ASSERT(arithProfile.bits() == bits); |
| ASSERT(ArithProfile::fromInt(bits).lhsObservedType().isOnlyInt32()); |
| ASSERT(ArithProfile::fromInt(bits).rhsObservedType().isOnlyInt32()); |
| } |
| { |
| uint32_t bits = 0x880000; |
| UNUSED_PARAM(bits); |
| ArithProfile arithProfile; |
| arithProfile.lhsSawNumber(); |
| arithProfile.rhsSawInt32(); |
| ASSERT(arithProfile.bits() == bits); |
| ASSERT(ArithProfile::fromInt(bits).lhsObservedType().isOnlyNumber()); |
| ASSERT(ArithProfile::fromInt(bits).rhsObservedType().isOnlyInt32()); |
| } |
| { |
| uint32_t bits = 0x900000; |
| UNUSED_PARAM(bits); |
| ArithProfile arithProfile; |
| arithProfile.lhsSawNumber(); |
| arithProfile.rhsSawNumber(); |
| ASSERT(arithProfile.bits() == bits); |
| ASSERT(ArithProfile::fromInt(bits).lhsObservedType().isOnlyNumber()); |
| ASSERT(ArithProfile::fromInt(bits).rhsObservedType().isOnlyNumber()); |
| } |
| { |
| uint32_t bits = 0x500000; |
| UNUSED_PARAM(bits); |
| ArithProfile arithProfile; |
| arithProfile.lhsSawInt32(); |
| arithProfile.rhsSawNumber(); |
| ASSERT(arithProfile.bits() == bits); |
| ASSERT(ArithProfile::fromInt(bits).lhsObservedType().isOnlyInt32()); |
| ASSERT(ArithProfile::fromInt(bits).rhsObservedType().isOnlyNumber()); |
| } |
| } |
| #if COMPILER(CLANG) |
| #pragma clang diagnostic pop |
| #endif |
| |
| } } // namespace JSC::LLInt |