blob: 5929b12bf5885a2dd08a03d7047a308353d569d1 [file] [log] [blame]
/*
* Copyright (C) 2009-2021 Apple Inc. All rights reserved.
* Copyright (C) 2019 the V8 project authors. All rights reserved.
*
* Redistribution and use in source and binary forms, with or without
* modification, are permitted provided that the following conditions
* are met:
* 1. Redistributions of source code must retain the above copyright
* notice, this list of conditions and the following disclaimer.
* 2. Redistributions in binary form must reproduce the above copyright
* notice, this list of conditions and the following disclaimer in the
* documentation and/or other materials provided with the distribution.
*
* THIS SOFTWARE IS PROVIDED BY APPLE INC. ``AS IS'' AND ANY
* EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR
* PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR
* CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL,
* EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO,
* PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR
* PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY
* OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
* (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
* OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
*/
#pragma once
#if ENABLE(YARR_JIT)
#include "MacroAssemblerCodeRef.h"
#include "MatchResult.h"
#include "VM.h"
#include "Yarr.h"
#include "YarrPattern.h"
#include <wtf/Bitmap.h>
#include <wtf/FixedVector.h>
#include <wtf/UniqueRef.h>
#define YARR_CALL
namespace JSC {
class VM;
class ExecutablePool;
namespace Yarr {
class MatchingContextHolder;
class YarrCodeBlock;
enum class JITFailureReason : uint8_t {
DecodeSurrogatePair,
BackReference,
ForwardReference,
VariableCountedParenthesisWithNonZeroMinimum,
ParenthesizedSubpattern,
FixedCountParenthesizedSubpattern,
ParenthesisNestedTooDeep,
ExecutableMemoryAllocationFailure,
};
class BoyerMooreBitmap {
WTF_MAKE_FAST_ALLOCATED(BoyerMooreBitmap);
public:
static constexpr unsigned mapSize = 128;
static constexpr unsigned mapMask = 128 - 1;
using Map = Bitmap<mapSize>;
BoyerMooreBitmap() = default;
unsigned count() const { return m_count; }
const Map& map() const { return m_map; }
bool isMaskEffective() const { return m_isMaskEffective; }
void add(UChar32 character)
{
unsigned position = character & mapMask;
if (position != static_cast<unsigned>(character))
m_isMaskEffective = true;
if (!m_map.get(position)) {
m_map.set(position);
++m_count;
}
}
private:
Map m_map { };
unsigned m_count { 0 };
bool m_isMaskEffective { false };
};
class BoyerMooreByteVector : public std::array<uint8_t, BoyerMooreBitmap::mapSize> {
WTF_MAKE_FAST_ALLOCATED;
public:
BoyerMooreByteVector(const BoyerMooreBitmap::Map& map)
{
fill(0);
map.forEachSetBit([&](unsigned index) {
(*this)[index] = 1;
});
}
};
#if CPU(ARM64E)
extern "C" EncodedMatchResult vmEntryToYarrJIT(const void* input, UCPURegister start, UCPURegister length, int* output, MatchingContextHolder* matchingContext, const void* codePtr);
extern "C" void vmEntryToYarrJITAfter(void);
#endif
class YarrCodeBlock {
WTF_MAKE_FAST_ALLOCATED;
WTF_MAKE_NONCOPYABLE(YarrCodeBlock);
using YarrJITCode8 = EncodedMatchResult (*)(const LChar* input, UCPURegister start, UCPURegister length, int* output, MatchingContextHolder& matchingContext) YARR_CALL;
using YarrJITCode16 = EncodedMatchResult (*)(const UChar* input, UCPURegister start, UCPURegister length, int* output, MatchingContextHolder& matchingContext) YARR_CALL;
using YarrJITCodeMatchOnly8 = EncodedMatchResult (*)(const LChar* input, UCPURegister start, UCPURegister length, void*, MatchingContextHolder& matchingContext) YARR_CALL;
using YarrJITCodeMatchOnly16 = EncodedMatchResult (*)(const UChar* input, UCPURegister start, UCPURegister length, void*, MatchingContextHolder& matchingContext) YARR_CALL;
public:
YarrCodeBlock() = default;
void setFallBackWithFailureReason(JITFailureReason failureReason) { m_failureReason = failureReason; }
std::optional<JITFailureReason> failureReason() { return m_failureReason; }
bool has8BitCode() { return m_ref8.size(); }
bool has16BitCode() { return m_ref16.size(); }
void set8BitCode(MacroAssemblerCodeRef<Yarr8BitPtrTag> ref, Vector<UniqueRef<BoyerMooreByteVector>> maps)
{
m_ref8 = ref;
m_maps.reserveCapacity(m_maps.size() + maps.size());
for (unsigned index = 0; index < maps.size(); ++index)
m_maps.uncheckedAppend(WTFMove(maps[index]));
}
void set16BitCode(MacroAssemblerCodeRef<Yarr16BitPtrTag> ref, Vector<UniqueRef<BoyerMooreByteVector>> maps)
{
m_ref16 = ref;
m_maps.reserveCapacity(m_maps.size() + maps.size());
for (unsigned index = 0; index < maps.size(); ++index)
m_maps.uncheckedAppend(WTFMove(maps[index]));
}
bool has8BitCodeMatchOnly() { return m_matchOnly8.size(); }
bool has16BitCodeMatchOnly() { return m_matchOnly16.size(); }
void set8BitCodeMatchOnly(MacroAssemblerCodeRef<YarrMatchOnly8BitPtrTag> matchOnly, Vector<UniqueRef<BoyerMooreByteVector>> maps)
{
m_matchOnly8 = matchOnly;
m_maps.reserveCapacity(m_maps.size() + maps.size());
for (unsigned index = 0; index < maps.size(); ++index)
m_maps.uncheckedAppend(WTFMove(maps[index]));
}
void set16BitCodeMatchOnly(MacroAssemblerCodeRef<YarrMatchOnly16BitPtrTag> matchOnly, Vector<UniqueRef<BoyerMooreByteVector>> maps)
{
m_matchOnly16 = matchOnly;
m_maps.reserveCapacity(m_maps.size() + maps.size());
for (unsigned index = 0; index < maps.size(); ++index)
m_maps.uncheckedAppend(WTFMove(maps[index]));
}
bool usesPatternContextBuffer() { return m_usesPatternContextBuffer; }
#if ENABLE(YARR_JIT_ALL_PARENS_EXPRESSIONS)
void setUsesPatternContextBuffer() { m_usesPatternContextBuffer = true; }
#endif
MatchResult execute(const LChar* input, unsigned start, unsigned length, int* output, MatchingContextHolder& matchingContext)
{
ASSERT(has8BitCode());
#if CPU(ARM64E)
if (Options::useJITCage())
return MatchResult(vmEntryToYarrJIT(input, start, length, output, &matchingContext, retagCodePtr<Yarr8BitPtrTag, YarrEntryPtrTag>(m_ref8.code().executableAddress())));
#endif
return MatchResult(untagCFunctionPtr<YarrJITCode8, Yarr8BitPtrTag>(m_ref8.code().executableAddress())(input, start, length, output, matchingContext));
}
MatchResult execute(const UChar* input, unsigned start, unsigned length, int* output, MatchingContextHolder& matchingContext)
{
ASSERT(has16BitCode());
#if CPU(ARM64E)
if (Options::useJITCage())
return MatchResult(vmEntryToYarrJIT(input, start, length, output, &matchingContext, retagCodePtr<Yarr16BitPtrTag, YarrEntryPtrTag>(m_ref16.code().executableAddress())));
#endif
return MatchResult(untagCFunctionPtr<YarrJITCode16, Yarr16BitPtrTag>(m_ref16.code().executableAddress())(input, start, length, output, matchingContext));
}
MatchResult execute(const LChar* input, unsigned start, unsigned length, MatchingContextHolder& matchingContext)
{
ASSERT(has8BitCodeMatchOnly());
#if CPU(ARM64E)
if (Options::useJITCage())
return MatchResult(vmEntryToYarrJIT(input, start, length, nullptr, &matchingContext, retagCodePtr<YarrMatchOnly8BitPtrTag, YarrEntryPtrTag>(m_matchOnly8.code().executableAddress())));
#endif
return MatchResult(untagCFunctionPtr<YarrJITCodeMatchOnly8, YarrMatchOnly8BitPtrTag>(m_matchOnly8.code().executableAddress())(input, start, length, nullptr, matchingContext));
}
MatchResult execute(const UChar* input, unsigned start, unsigned length, MatchingContextHolder& matchingContext)
{
ASSERT(has16BitCodeMatchOnly());
#if CPU(ARM64E)
if (Options::useJITCage())
return MatchResult(vmEntryToYarrJIT(input, start, length, nullptr, &matchingContext, retagCodePtr<YarrMatchOnly16BitPtrTag, YarrEntryPtrTag>(m_matchOnly16.code().executableAddress())));
#endif
return MatchResult(untagCFunctionPtr<YarrJITCodeMatchOnly16, YarrMatchOnly16BitPtrTag>(m_matchOnly16.code().executableAddress())(input, start, length, nullptr, matchingContext));
}
#if ENABLE(REGEXP_TRACING)
void *get8BitMatchOnlyAddr()
{
if (!has8BitCodeMatchOnly())
return 0;
return m_matchOnly8.code().executableAddress();
}
void *get16BitMatchOnlyAddr()
{
if (!has16BitCodeMatchOnly())
return 0;
return m_matchOnly16.code().executableAddress();
}
void *get8BitMatchAddr()
{
if (!has8BitCode())
return 0;
return m_ref8.code().executableAddress();
}
void *get16BitMatchAddr()
{
if (!has16BitCode())
return 0;
return m_ref16.code().executableAddress();
}
#endif
size_t size() const
{
return m_ref8.size() + m_ref16.size() + m_matchOnly8.size() + m_matchOnly16.size();
}
void clear(const AbstractLocker&)
{
m_ref8 = MacroAssemblerCodeRef<Yarr8BitPtrTag>();
m_ref16 = MacroAssemblerCodeRef<Yarr16BitPtrTag>();
m_matchOnly8 = MacroAssemblerCodeRef<YarrMatchOnly8BitPtrTag>();
m_matchOnly16 = MacroAssemblerCodeRef<YarrMatchOnly16BitPtrTag>();
m_maps.clear();
m_failureReason = std::nullopt;
}
const uint8_t* tryReuseBoyerMooreByteVector(unsigned index, BoyerMooreByteVector& vector) const
{
if (index < m_maps.size()) {
if (m_maps[index].get() == vector)
return m_maps[index]->data();
}
return nullptr;
}
private:
MacroAssemblerCodeRef<Yarr8BitPtrTag> m_ref8;
MacroAssemblerCodeRef<Yarr16BitPtrTag> m_ref16;
MacroAssemblerCodeRef<YarrMatchOnly8BitPtrTag> m_matchOnly8;
MacroAssemblerCodeRef<YarrMatchOnly16BitPtrTag> m_matchOnly16;
Vector<UniqueRef<BoyerMooreByteVector>> m_maps;
bool m_usesPatternContextBuffer { false };
std::optional<JITFailureReason> m_failureReason;
};
enum class JITCompileMode : uint8_t {
MatchOnly,
IncludeSubpatterns
};
void jitCompile(YarrPattern&, String& patternString, CharSize, VM*, YarrCodeBlock& jitObject, JITCompileMode);
} } // namespace JSC::Yarr
#endif