/* * Copyright (C) 2013 Apple Inc. All rights reserved. * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice, this list of conditions and the following disclaimer. * 2. Redistributions in binary form must reproduce the above copyright * notice, this list of conditions and the following disclaimer in the * documentation and/or other materials provided with the distribution. * * THIS SOFTWARE IS PROVIDED BY APPLE INC. ``AS IS'' AND ANY * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY * OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. */ #include "config.h" #include "DFGJITCode.h" #if ENABLE(DFG_JIT) #include "CodeBlock.h" namespace JSC { namespace DFG { JITCode::JITCode() : DirectJITCode(DFGJIT) { } JITCode::~JITCode() { } CommonData* JITCode::dfgCommon() { return &common; } JITCode* JITCode::dfg() { return this; } void JITCode::shrinkToFit() { common.shrinkToFit(); osrEntry.shrinkToFit(); osrExit.shrinkToFit(); speculationRecovery.shrinkToFit(); minifiedDFG.prepareAndShrink(); variableEventStream.shrinkToFit(); } void JITCode::reconstruct( CodeBlock* codeBlock, CodeOrigin codeOrigin, unsigned streamIndex, Operands& result) { variableEventStream.reconstruct( codeBlock, codeOrigin, minifiedDFG, streamIndex, result); } void JITCode::reconstruct( ExecState* exec, CodeBlock* codeBlock, CodeOrigin codeOrigin, unsigned streamIndex, Operands& result) { Operands recoveries; reconstruct(codeBlock, codeOrigin, streamIndex, recoveries); result = Operands(OperandsLike, recoveries); for (size_t i = result.size(); i--;) { int operand = result.operandForIndex(i); if (operandIsArgument(operand) && !VirtualRegister(operand).toArgument() && codeBlock->codeType() == FunctionCode && codeBlock->specializationKind() == CodeForConstruct) { // Ugh. If we're in a constructor, the 'this' argument may hold garbage. It will // also never be used. It doesn't matter what we put into the value for this, // but it has to be an actual value that can be grokked by subsequent DFG passes, // so we sanitize it here by turning it into Undefined. result[i] = jsUndefined(); continue; } result[i] = recoveries[i].recover(exec); } } #if ENABLE(FTL_JIT) bool JITCode::checkIfOptimizationThresholdReached(CodeBlock* codeBlock) { ASSERT(codeBlock->jitType() == JITCode::DFGJIT); return tierUpCounter.checkIfThresholdCrossedAndSet(codeBlock->baselineVersion()); } void JITCode::optimizeNextInvocation(CodeBlock* codeBlock) { ASSERT(codeBlock->jitType() == JITCode::DFGJIT); if (Options::verboseOSR()) dataLog(*codeBlock, ": FTL-optimizing next invocation.\n"); tierUpCounter.setNewThreshold(0, codeBlock->baselineVersion()); } void JITCode::dontOptimizeAnytimeSoon(CodeBlock* codeBlock) { ASSERT(codeBlock->jitType() == JITCode::DFGJIT); if (Options::verboseOSR()) dataLog(*codeBlock, ": Not FTL-optimizing anytime soon.\n"); tierUpCounter.deferIndefinitely(); } void JITCode::optimizeAfterWarmUp(CodeBlock* codeBlock) { ASSERT(codeBlock->jitType() == JITCode::DFGJIT); if (Options::verboseOSR()) dataLog(*codeBlock, ": FTL-optimizing after warm-up.\n"); CodeBlock* baseline = codeBlock->baselineVersion(); tierUpCounter.setNewThreshold( baseline->adjustedCounterValue(Options::thresholdForFTLOptimizeAfterWarmUp()), baseline); } void JITCode::optimizeSoon(CodeBlock* codeBlock) { ASSERT(codeBlock->jitType() == JITCode::DFGJIT); if (Options::verboseOSR()) dataLog(*codeBlock, ": FTL-optimizing soon.\n"); CodeBlock* baseline = codeBlock->baselineVersion(); tierUpCounter.setNewThreshold( baseline->adjustedCounterValue(Options::thresholdForFTLOptimizeSoon()), baseline); } void JITCode::forceOptimizationSlowPathConcurrently(CodeBlock* codeBlock) { ASSERT(codeBlock->jitType() == JITCode::DFGJIT); if (Options::verboseOSR()) dataLog(*codeBlock, ": Forcing slow path concurrently for FTL entry.\n"); tierUpCounter.forceSlowPathConcurrently(); } void JITCode::setOptimizationThresholdBasedOnCompilationResult( CodeBlock* codeBlock, CompilationResult result) { ASSERT(codeBlock->jitType() == JITCode::DFGJIT); switch (result) { case CompilationSuccessful: optimizeNextInvocation(codeBlock); return; case CompilationFailed: dontOptimizeAnytimeSoon(codeBlock); codeBlock->baselineVersion()->m_didFailFTLCompilation = true; return; case CompilationDeferred: optimizeAfterWarmUp(codeBlock); return; case CompilationInvalidated: // This is weird - it will only happen in cases when the DFG code block (i.e. // the code block that this JITCode belongs to) is also invalidated. So it // doesn't really matter what we do. But, we do the right thing anyway. Note // that us counting the reoptimization actually means that we might count it // twice. But that's generally OK. It's better to overcount reoptimizations // than it is to undercount them. codeBlock->baselineVersion()->countReoptimization(); optimizeAfterWarmUp(codeBlock); return; } RELEASE_ASSERT_NOT_REACHED(); } #endif // ENABLE(FTL_JIT) } } // namespace JSC::DFG #endif // ENABLE(DFG_JIT)