diff options
author | Lorry Tar Creator <lorry-tar-importer@lorry> | 2017-06-27 06:07:23 +0000 |
---|---|---|
committer | Lorry Tar Creator <lorry-tar-importer@lorry> | 2017-06-27 06:07:23 +0000 |
commit | 1bf1084f2b10c3b47fd1a588d85d21ed0eb41d0c (patch) | |
tree | 46dcd36c86e7fbc6e5df36deb463b33e9967a6f7 /Source/JavaScriptCore/runtime/SamplingProfiler.cpp | |
parent | 32761a6cee1d0dee366b885b7b9c777e67885688 (diff) | |
download | WebKitGtk-tarball-master.tar.gz |
webkitgtk-2.16.5HEADwebkitgtk-2.16.5master
Diffstat (limited to 'Source/JavaScriptCore/runtime/SamplingProfiler.cpp')
-rw-r--r-- | Source/JavaScriptCore/runtime/SamplingProfiler.cpp | 1096 |
1 files changed, 1096 insertions, 0 deletions
diff --git a/Source/JavaScriptCore/runtime/SamplingProfiler.cpp b/Source/JavaScriptCore/runtime/SamplingProfiler.cpp new file mode 100644 index 000000000..9326d7a0f --- /dev/null +++ b/Source/JavaScriptCore/runtime/SamplingProfiler.cpp @@ -0,0 +1,1096 @@ +/* + * Copyright (C) 2016 Apple Inc. All rights reserved. + * + * Redistribution and use in source and binary forms, with or without + * modification, are permitted provided that the following conditions + * are met: + * 1. Redistributions of source code must retain the above copyright + * notice, this list of conditions and the following disclaimer. + * 2. Redistributions in binary form must reproduce the above copyright + * notice, this list of conditions and the following disclaimer in the + * documentation and/or other materials provided with the distribution. + * + * THIS SOFTWARE IS PROVIDED BY APPLE INC. ``AS IS'' AND ANY + * EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE + * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR + * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL APPLE INC. OR + * CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, + * EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, + * PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR + * PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY + * OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT + * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE + * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. + */ + +#include "config.h" +#include "SamplingProfiler.h" + +#if ENABLE(SAMPLING_PROFILER) + +#include "CallFrame.h" +#include "CodeBlock.h" +#include "CodeBlockSet.h" +#include "HeapIterationScope.h" +#include "HeapUtil.h" +#include "InlineCallFrame.h" +#include "Interpreter.h" +#include "JSCInlines.h" +#include "JSFunction.h" +#include "LLIntPCRanges.h" +#include "MarkedBlock.h" +#include "MarkedBlockSet.h" +#include "MarkedSpaceInlines.h" +#include "NativeExecutable.h" +#include "PCToCodeOriginMap.h" +#include "SlotVisitor.h" +#include "StrongInlines.h" +#include "VM.h" +#include <wtf/HashSet.h> +#include <wtf/RandomNumber.h> +#include <wtf/RefPtr.h> +#include <wtf/text/StringBuilder.h> + +#if OS(DARWIN) +#include <cxxabi.h> +#include <dlfcn.h> +#endif + +namespace JSC { + +static double sNumTotalStackTraces = 0; +static double sNumTotalWalks = 0; +static double sNumFailedWalks = 0; +static const uint32_t sNumWalkReportingFrequency = 50; +static const double sWalkErrorPercentage = .05; +static const bool sReportStatsOnlyWhenTheyreAboveThreshold = false; +static const bool sReportStats = false; + +using FrameType = SamplingProfiler::FrameType; +using UnprocessedStackFrame = SamplingProfiler::UnprocessedStackFrame; + +ALWAYS_INLINE static void reportStats() +{ + if (sReportStats && sNumTotalWalks && static_cast<uint64_t>(sNumTotalWalks) % sNumWalkReportingFrequency == 0) { + if (!sReportStatsOnlyWhenTheyreAboveThreshold || (sNumFailedWalks / sNumTotalWalks > sWalkErrorPercentage)) { + dataLogF("Num total walks: %llu. Failed walks percent: %lf\n", + static_cast<unsigned long long>(sNumTotalWalks), sNumFailedWalks / sNumTotalWalks); + } + } +} + +class FrameWalker { +public: + FrameWalker(VM& vm, ExecState* callFrame, const LockHolder& codeBlockSetLocker, const LockHolder& machineThreadsLocker) + : m_vm(vm) + , m_callFrame(callFrame) + , m_vmEntryFrame(vm.topVMEntryFrame) + , m_codeBlockSetLocker(codeBlockSetLocker) + , m_machineThreadsLocker(machineThreadsLocker) + { + } + + SUPPRESS_ASAN + size_t walk(Vector<UnprocessedStackFrame>& stackTrace, bool& didRunOutOfSpace) + { + if (sReportStats) + sNumTotalWalks++; + resetAtMachineFrame(); + size_t maxStackTraceSize = stackTrace.size(); + while (!isAtTop() && !m_bailingOut && m_depth < maxStackTraceSize) { + recordJSFrame(stackTrace); + advanceToParentFrame(); + resetAtMachineFrame(); + } + didRunOutOfSpace = m_depth >= maxStackTraceSize && !isAtTop(); + reportStats(); + return m_depth; + } + + bool wasValidWalk() const + { + return !m_bailingOut; + } + +protected: + + SUPPRESS_ASAN + void recordJSFrame(Vector<UnprocessedStackFrame>& stackTrace) + { + CallSiteIndex callSiteIndex; + JSValue unsafeCallee = m_callFrame->unsafeCallee(); + CodeBlock* codeBlock = m_callFrame->unsafeCodeBlock(); + if (codeBlock) { + ASSERT(isValidCodeBlock(codeBlock)); + callSiteIndex = m_callFrame->unsafeCallSiteIndex(); + } + stackTrace[m_depth] = UnprocessedStackFrame(codeBlock, JSValue::encode(unsafeCallee), callSiteIndex); + m_depth++; + } + + SUPPRESS_ASAN + void advanceToParentFrame() + { + m_callFrame = m_callFrame->unsafeCallerFrame(m_vmEntryFrame); + } + + bool isAtTop() const + { + return !m_callFrame; + } + + SUPPRESS_ASAN + void resetAtMachineFrame() + { + if (isAtTop()) + return; + + if (!isValidFramePointer(m_callFrame)) { + // Guard against pausing the process at weird program points. + m_bailingOut = true; + if (sReportStats) + sNumFailedWalks++; + return; + } + + CodeBlock* codeBlock = m_callFrame->unsafeCodeBlock(); + if (!codeBlock) + return; + + if (!isValidCodeBlock(codeBlock)) { + m_bailingOut = true; + if (sReportStats) + sNumFailedWalks++; + return; + } + } + + bool isValidFramePointer(void* exec) + { + uint8_t* fpCast = bitwise_cast<uint8_t*>(exec); + for (MachineThreads::Thread* thread = m_vm.heap.machineThreads().threadsListHead(m_machineThreadsLocker); thread; thread = thread->next) { + uint8_t* stackBase = static_cast<uint8_t*>(thread->stackBase()); + uint8_t* stackLimit = static_cast<uint8_t*>(thread->stackEnd()); + RELEASE_ASSERT(stackBase); + RELEASE_ASSERT(stackLimit); + if (fpCast <= stackBase && fpCast >= stackLimit) + return true; + } + return false; + } + + bool isValidCodeBlock(CodeBlock* codeBlock) + { + if (!codeBlock) + return false; + bool result = m_vm.heap.codeBlockSet().contains(m_codeBlockSetLocker, codeBlock); + return result; + } + + VM& m_vm; + ExecState* m_callFrame; + VMEntryFrame* m_vmEntryFrame; + const LockHolder& m_codeBlockSetLocker; + const LockHolder& m_machineThreadsLocker; + bool m_bailingOut { false }; + size_t m_depth { 0 }; +}; + +class CFrameWalker : public FrameWalker { +public: + typedef FrameWalker Base; + + CFrameWalker(VM& vm, void* machineFrame, ExecState* callFrame, const LockHolder& codeBlockSetLocker, const LockHolder& machineThreadsLocker) + : Base(vm, callFrame, codeBlockSetLocker, machineThreadsLocker) + , m_machineFrame(machineFrame) + { + } + + size_t walk(Vector<UnprocessedStackFrame>& stackTrace, bool& didRunOutOfSpace) + { + if (sReportStats) + sNumTotalWalks++; + resetAtMachineFrame(); + size_t maxStackTraceSize = stackTrace.size(); + // The way the C walker decides if a frame it is about to trace is C or JS is by + // ensuring m_callFrame points to some frame above the machineFrame. + if (!isAtTop() && !m_bailingOut && m_machineFrame == m_callFrame) { + recordJSFrame(stackTrace); + Base::advanceToParentFrame(); + resetAtMachineFrame(); + } + + while (!isAtTop() && !m_bailingOut && m_depth < maxStackTraceSize) { + if (m_machineFrame >= m_callFrame) { + // If we get to this state we probably have an invalid trace. + m_bailingOut = true; + break; + } + + if (isCFrame()) { + RELEASE_ASSERT(!LLInt::isLLIntPC(frame()->callerFrame)); + stackTrace[m_depth] = UnprocessedStackFrame(frame()->pc); + m_depth++; + } else + recordJSFrame(stackTrace); + advanceToParentFrame(); + resetAtMachineFrame(); + } + didRunOutOfSpace = m_depth >= maxStackTraceSize && !isAtTop(); + reportStats(); + return m_depth; + } + +private: + + bool isCFrame() + { + return frame()->callerFrame != m_callFrame; + } + + void advanceToParentFrame() + { + if (!isCFrame()) + Base::advanceToParentFrame(); + m_machineFrame = frame()->callerFrame; + } + + void resetAtMachineFrame() + { + if (!isValidFramePointer(m_machineFrame)) { + // Guard against pausing the process at weird program points. + m_bailingOut = true; + if (sReportStats) + sNumFailedWalks++; + return; + } + Base::resetAtMachineFrame(); + } + + CallerFrameAndPC* frame() + { + return reinterpret_cast<CallerFrameAndPC*>(m_machineFrame); + } + + void* m_machineFrame; +}; + +SamplingProfiler::SamplingProfiler(VM& vm, RefPtr<Stopwatch>&& stopwatch) + : m_vm(vm) + , m_stopwatch(WTFMove(stopwatch)) + , m_timingInterval(std::chrono::microseconds(Options::sampleInterval())) + , m_threadIdentifier(0) + , m_jscExecutionThread(nullptr) + , m_isPaused(false) + , m_isShutDown(false) +{ + if (sReportStats) { + sNumTotalWalks = 0; + sNumFailedWalks = 0; + } + + m_currentFrames.grow(256); +} + +SamplingProfiler::~SamplingProfiler() +{ +} + +void SamplingProfiler::createThreadIfNecessary(const LockHolder&) +{ + ASSERT(m_lock.isLocked()); + + if (m_threadIdentifier) + return; + + RefPtr<SamplingProfiler> profiler = this; + m_threadIdentifier = createThread("jsc.sampling-profiler.thread", [profiler] { + profiler->timerLoop(); + }); +} + +void SamplingProfiler::timerLoop() +{ + while (true) { + std::chrono::microseconds stackTraceProcessingTime = std::chrono::microseconds(0); + { + LockHolder locker(m_lock); + if (UNLIKELY(m_isShutDown)) + return; + + if (!m_isPaused && m_jscExecutionThread) + takeSample(locker, stackTraceProcessingTime); + + m_lastTime = m_stopwatch->elapsedTime(); + } + + // Read section 6.2 of this paper for more elaboration of why we add a random + // fluctuation here. The main idea is to prevent our timer from being in sync + // with some system process such as a scheduled context switch. + // http://plv.colorado.edu/papers/mytkowicz-pldi10.pdf + double randomSignedNumber = (randomNumber() * 2.0) - 1.0; // A random number between [-1, 1). + std::chrono::microseconds randomFluctuation = std::chrono::microseconds(static_cast<uint64_t>(randomSignedNumber * static_cast<double>(m_timingInterval.count()) * 0.20l)); + std::this_thread::sleep_for(m_timingInterval - std::min(m_timingInterval, stackTraceProcessingTime) + randomFluctuation); + } +} + +void SamplingProfiler::takeSample(const LockHolder&, std::chrono::microseconds& stackTraceProcessingTime) +{ + ASSERT(m_lock.isLocked()); + if (m_vm.entryScope) { + double nowTime = m_stopwatch->elapsedTime(); + + LockHolder machineThreadsLocker(m_vm.heap.machineThreads().getLock()); + LockHolder codeBlockSetLocker(m_vm.heap.codeBlockSet().getLock()); + LockHolder executableAllocatorLocker(m_vm.executableAllocator.getLock()); + + bool didSuspend = m_jscExecutionThread->suspend(); + if (didSuspend) { + // While the JSC thread is suspended, we can't do things like malloc because the JSC thread + // may be holding the malloc lock. + void* machineFrame; + ExecState* callFrame; + void* machinePC; + bool topFrameIsLLInt = false; + void* llintPC; + { + MachineThreads::Thread::Registers registers; + m_jscExecutionThread->getRegisters(registers); + machineFrame = registers.framePointer(); + callFrame = static_cast<ExecState*>(machineFrame); + machinePC = registers.instructionPointer(); + llintPC = registers.llintPC(); + m_jscExecutionThread->freeRegisters(registers); + } + // FIXME: Lets have a way of detecting when we're parsing code. + // https://bugs.webkit.org/show_bug.cgi?id=152761 + if (m_vm.executableAllocator.isValidExecutableMemory(executableAllocatorLocker, machinePC)) { + if (m_vm.isExecutingInRegExpJIT) { + // FIXME: We're executing a regexp. Lets gather more intersting data. + // https://bugs.webkit.org/show_bug.cgi?id=152729 + callFrame = m_vm.topCallFrame; // We need to do this or else we'd fail our backtrace validation b/c this isn't a JS frame. + } + } else if (LLInt::isLLIntPC(machinePC)) { + topFrameIsLLInt = true; + // We're okay to take a normal stack trace when the PC + // is in LLInt code. + } else { + // We resort to topCallFrame to see if we can get anything + // useful. We usually get here when we're executing C code. + callFrame = m_vm.topCallFrame; + } + + size_t walkSize; + bool wasValidWalk; + bool didRunOutOfVectorSpace; + if (Options::sampleCCode()) { + CFrameWalker walker(m_vm, machineFrame, callFrame, codeBlockSetLocker, machineThreadsLocker); + walkSize = walker.walk(m_currentFrames, didRunOutOfVectorSpace); + wasValidWalk = walker.wasValidWalk(); + } else { + FrameWalker walker(m_vm, callFrame, codeBlockSetLocker, machineThreadsLocker); + walkSize = walker.walk(m_currentFrames, didRunOutOfVectorSpace); + wasValidWalk = walker.wasValidWalk(); + } + + m_jscExecutionThread->resume(); + + auto startTime = std::chrono::steady_clock::now(); + // We can now use data structures that malloc, and do other interesting things, again. + + // FIXME: It'd be interesting to take data about the program's state when + // we fail to take a stack trace: https://bugs.webkit.org/show_bug.cgi?id=152758 + if (wasValidWalk && walkSize) { + if (sReportStats) + sNumTotalStackTraces++; + Vector<UnprocessedStackFrame> stackTrace; + stackTrace.reserveInitialCapacity(walkSize); + for (size_t i = 0; i < walkSize; i++) { + UnprocessedStackFrame frame = m_currentFrames[i]; + stackTrace.uncheckedAppend(frame); + } + + m_unprocessedStackTraces.append(UnprocessedStackTrace { nowTime, machinePC, topFrameIsLLInt, llintPC, WTFMove(stackTrace) }); + + if (didRunOutOfVectorSpace) + m_currentFrames.grow(m_currentFrames.size() * 1.25); + } + + auto endTime = std::chrono::steady_clock::now(); + stackTraceProcessingTime = std::chrono::duration_cast<std::chrono::microseconds>(endTime - startTime); + } + } +} + +static ALWAYS_INLINE unsigned tryGetBytecodeIndex(unsigned llintPC, CodeBlock* codeBlock, bool& isValid) +{ +#if ENABLE(DFG_JIT) + RELEASE_ASSERT(!codeBlock->hasCodeOrigins()); +#endif + +#if USE(JSVALUE64) + unsigned bytecodeIndex = llintPC; + if (bytecodeIndex < codeBlock->instructionCount()) { + isValid = true; + return bytecodeIndex; + } + isValid = false; + return 0; +#else + Instruction* instruction = bitwise_cast<Instruction*>(llintPC); + if (instruction >= codeBlock->instructions().begin() && instruction < codeBlock->instructions().begin() + codeBlock->instructionCount()) { + isValid = true; + unsigned bytecodeIndex = instruction - codeBlock->instructions().begin(); + return bytecodeIndex; + } + isValid = false; + return 0; +#endif +} + +void SamplingProfiler::processUnverifiedStackTraces() +{ + // This function needs to be called from the JSC execution thread. + RELEASE_ASSERT(m_lock.isLocked()); + + TinyBloomFilter filter = m_vm.heap.objectSpace().blocks().filter(); + + for (UnprocessedStackTrace& unprocessedStackTrace : m_unprocessedStackTraces) { + m_stackTraces.append(StackTrace()); + StackTrace& stackTrace = m_stackTraces.last(); + stackTrace.timestamp = unprocessedStackTrace.timestamp; + + auto populateCodeLocation = [] (CodeBlock* codeBlock, unsigned bytecodeIndex, StackFrame::CodeLocation& location) { + if (bytecodeIndex < codeBlock->instructionCount()) { + int divot; + int startOffset; + int endOffset; + codeBlock->expressionRangeForBytecodeOffset(bytecodeIndex, divot, startOffset, endOffset, + location.lineNumber, location.columnNumber); + location.bytecodeIndex = bytecodeIndex; + } + if (Options::collectSamplingProfilerDataForJSCShell()) { + location.codeBlockHash = codeBlock->hash(); + location.jitType = codeBlock->jitType(); + } + }; + + auto appendCodeBlock = [&] (CodeBlock* codeBlock, unsigned bytecodeIndex) { + stackTrace.frames.append(StackFrame(codeBlock->ownerExecutable())); + m_liveCellPointers.add(codeBlock->ownerExecutable()); + populateCodeLocation(codeBlock, bytecodeIndex, stackTrace.frames.last().semanticLocation); + }; + + auto appendEmptyFrame = [&] { + stackTrace.frames.append(StackFrame()); + }; + + auto storeCalleeIntoLastFrame = [&] (EncodedJSValue encodedCallee) { + // Set the callee if it's a valid GC object. + JSValue callee = JSValue::decode(encodedCallee); + StackFrame& stackFrame = stackTrace.frames.last(); + bool alreadyHasExecutable = !!stackFrame.executable; + if (!HeapUtil::isValueGCObject(m_vm.heap, filter, callee)) { + if (!alreadyHasExecutable) + stackFrame.frameType = FrameType::Unknown; + return; + } + + JSCell* calleeCell = callee.asCell(); + auto setFallbackFrameType = [&] { + ASSERT(!alreadyHasExecutable); + FrameType result = FrameType::Unknown; + CallData callData; + CallType callType; + callType = getCallData(calleeCell, callData); + if (callType == CallType::Host) + result = FrameType::Host; + + stackFrame.frameType = result; + }; + + auto addCallee = [&] (JSObject* callee) { + stackFrame.callee = callee; + m_liveCellPointers.add(callee); + }; + + if (calleeCell->type() != JSFunctionType) { + if (JSObject* object = jsDynamicCast<JSObject*>(*calleeCell->vm(), calleeCell)) + addCallee(object); + + if (!alreadyHasExecutable) + setFallbackFrameType(); + + return; + } + + addCallee(jsCast<JSFunction*>(calleeCell)); + + if (alreadyHasExecutable) + return; + + ExecutableBase* executable = jsCast<JSFunction*>(calleeCell)->executable(); + if (!executable) { + setFallbackFrameType(); + return; + } + + RELEASE_ASSERT(HeapUtil::isPointerGCObjectJSCell(m_vm.heap, filter, executable)); + stackFrame.frameType = FrameType::Executable; + stackFrame.executable = executable; + m_liveCellPointers.add(executable); + }; + + auto appendCodeOrigin = [&] (CodeBlock* machineCodeBlock, CodeOrigin origin) { + size_t startIndex = stackTrace.frames.size(); // We want to change stack traces that we're about to append. + + CodeOrigin machineOrigin; + origin.walkUpInlineStack([&] (const CodeOrigin& codeOrigin) { + machineOrigin = codeOrigin; + appendCodeBlock(codeOrigin.inlineCallFrame ? codeOrigin.inlineCallFrame->baselineCodeBlock.get() : machineCodeBlock, codeOrigin.bytecodeIndex); + }); + + if (Options::collectSamplingProfilerDataForJSCShell()) { + RELEASE_ASSERT(machineOrigin.isSet()); + RELEASE_ASSERT(!machineOrigin.inlineCallFrame); + + StackFrame::CodeLocation machineLocation = stackTrace.frames.last().semanticLocation; + + // We want to tell each inlined frame about the machine frame + // they were inlined into. Currently, we only use this for dumping + // output on the command line, but we could extend it to the web + // inspector in the future if we find a need for it there. + RELEASE_ASSERT(stackTrace.frames.size()); + for (size_t i = startIndex; i < stackTrace.frames.size() - 1; i++) + stackTrace.frames[i].machineLocation = std::make_pair(machineLocation, Strong<CodeBlock>(m_vm, machineCodeBlock)); + } + }; + + // Prepend the top-most inlined frame if needed and gather + // location information about where the top frame is executing. + size_t startIndex = 0; + if (unprocessedStackTrace.frames.size() && !!unprocessedStackTrace.frames[0].verifiedCodeBlock) { + CodeBlock* topCodeBlock = unprocessedStackTrace.frames[0].verifiedCodeBlock; + if (unprocessedStackTrace.topFrameIsLLInt) { + // We reuse LLInt CodeBlocks for the baseline JIT, so we need to check for both jit types. + // This might also be false for various reasons (known and unknown), even though + // it's super unlikely. One reason that this can be false is when we throw from a DFG frame, + // and we end up having to unwind past a VMEntryFrame, we will end up executing + // inside the LLInt's handleUncaughtException. So we just protect against this + // by ignoring it. + unsigned bytecodeIndex = 0; + if (topCodeBlock->jitType() == JITCode::InterpreterThunk || topCodeBlock->jitType() == JITCode::BaselineJIT) { + bool isValidPC; + unsigned bits; +#if USE(JSVALUE64) + bits = static_cast<unsigned>(bitwise_cast<uintptr_t>(unprocessedStackTrace.llintPC)); +#else + bits = bitwise_cast<unsigned>(unprocessedStackTrace.llintPC); +#endif + bytecodeIndex = tryGetBytecodeIndex(bits, topCodeBlock, isValidPC); + + UNUSED_PARAM(isValidPC); // FIXME: do something with this info for the web inspector: https://bugs.webkit.org/show_bug.cgi?id=153455 + + appendCodeBlock(topCodeBlock, bytecodeIndex); + storeCalleeIntoLastFrame(unprocessedStackTrace.frames[0].unverifiedCallee); + startIndex = 1; + } + } else if (std::optional<CodeOrigin> codeOrigin = topCodeBlock->findPC(unprocessedStackTrace.topPC)) { + appendCodeOrigin(topCodeBlock, *codeOrigin); + storeCalleeIntoLastFrame(unprocessedStackTrace.frames[0].unverifiedCallee); + startIndex = 1; + } + } + + for (size_t i = startIndex; i < unprocessedStackTrace.frames.size(); i++) { + UnprocessedStackFrame& unprocessedStackFrame = unprocessedStackTrace.frames[i]; + if (CodeBlock* codeBlock = unprocessedStackFrame.verifiedCodeBlock) { + CallSiteIndex callSiteIndex = unprocessedStackFrame.callSiteIndex; + + auto appendCodeBlockNoInlining = [&] { + bool isValidPC; + appendCodeBlock(codeBlock, tryGetBytecodeIndex(callSiteIndex.bits(), codeBlock, isValidPC)); + }; + +#if ENABLE(DFG_JIT) + if (codeBlock->hasCodeOrigins()) { + if (codeBlock->canGetCodeOrigin(callSiteIndex)) + appendCodeOrigin(codeBlock, codeBlock->codeOrigin(callSiteIndex)); + else + appendCodeBlock(codeBlock, std::numeric_limits<unsigned>::max()); + } else + appendCodeBlockNoInlining(); +#else + appendCodeBlockNoInlining(); +#endif + } else if (unprocessedStackFrame.cCodePC) { + appendEmptyFrame(); + stackTrace.frames.last().cCodePC = unprocessedStackFrame.cCodePC; + stackTrace.frames.last().frameType = FrameType::C; + } else + appendEmptyFrame(); + + // Note that this is okay to do if we walked the inline stack because + // the machine frame will be at the top of the processed stack trace. + if (!unprocessedStackFrame.cCodePC) + storeCalleeIntoLastFrame(unprocessedStackFrame.unverifiedCallee); + } + } + + m_unprocessedStackTraces.clear(); +} + +void SamplingProfiler::visit(SlotVisitor& slotVisitor) +{ + RELEASE_ASSERT(m_lock.isLocked()); + for (JSCell* cell : m_liveCellPointers) + slotVisitor.appendUnbarriered(cell); +} + +void SamplingProfiler::shutdown() +{ + LockHolder locker(m_lock); + m_isShutDown = true; +} + +void SamplingProfiler::start() +{ + LockHolder locker(m_lock); + start(locker); +} + +void SamplingProfiler::start(const LockHolder& locker) +{ + ASSERT(m_lock.isLocked()); + m_isPaused = false; + createThreadIfNecessary(locker); +} + +void SamplingProfiler::pause(const LockHolder&) +{ + ASSERT(m_lock.isLocked()); + m_isPaused = true; + reportStats(); +} + +void SamplingProfiler::noticeCurrentThreadAsJSCExecutionThread(const LockHolder&) +{ + ASSERT(m_lock.isLocked()); + m_jscExecutionThread = m_vm.heap.machineThreads().machineThreadForCurrentThread(); +} + +void SamplingProfiler::noticeCurrentThreadAsJSCExecutionThread() +{ + LockHolder locker(m_lock); + noticeCurrentThreadAsJSCExecutionThread(locker); +} + +void SamplingProfiler::noticeJSLockAcquisition() +{ + LockHolder locker(m_lock); + noticeCurrentThreadAsJSCExecutionThread(locker); +} + +void SamplingProfiler::noticeVMEntry() +{ + LockHolder locker(m_lock); + ASSERT(m_vm.entryScope); + noticeCurrentThreadAsJSCExecutionThread(locker); + m_lastTime = m_stopwatch->elapsedTime(); + createThreadIfNecessary(locker); +} + +void SamplingProfiler::clearData(const LockHolder&) +{ + ASSERT(m_lock.isLocked()); + m_stackTraces.clear(); + m_liveCellPointers.clear(); + m_unprocessedStackTraces.clear(); +} + +String SamplingProfiler::StackFrame::nameFromCallee(VM& vm) +{ + if (!callee) + return String(); + + auto scope = DECLARE_CATCH_SCOPE(vm); + ExecState* exec = callee->globalObject()->globalExec(); + auto getPropertyIfPureOperation = [&] (const Identifier& ident) -> String { + PropertySlot slot(callee, PropertySlot::InternalMethodType::VMInquiry); + PropertyName propertyName(ident); + bool hasProperty = callee->getPropertySlot(exec, propertyName, slot); + ASSERT_UNUSED(scope, !scope.exception()); + if (hasProperty) { + if (slot.isValue()) { + JSValue nameValue = slot.getValue(exec, propertyName); + if (isJSString(nameValue)) + return asString(nameValue)->tryGetValue(); + } + } + return String(); + }; + + String name = getPropertyIfPureOperation(vm.propertyNames->displayName); + if (!name.isEmpty()) + return name; + + return getPropertyIfPureOperation(vm.propertyNames->name); +} + +String SamplingProfiler::StackFrame::displayName(VM& vm) +{ + { + String name = nameFromCallee(vm); + if (!name.isEmpty()) + return name; + } + + if (frameType == FrameType::Unknown || frameType == FrameType::C) { +#if OS(DARWIN) + if (frameType == FrameType::C) { + const char* mangledName = nullptr; + const char* cxaDemangled = nullptr; + Dl_info info; + if (dladdr(cCodePC, &info) && info.dli_sname) + mangledName = info.dli_sname; + if (mangledName) { + cxaDemangled = abi::__cxa_demangle(mangledName, 0, 0, 0); + return String(cxaDemangled ? cxaDemangled : mangledName); + } + WTF::dataLog("couldn't get a name"); + } +#endif + return ASCIILiteral("(unknown)"); + } + if (frameType == FrameType::Host) + return ASCIILiteral("(host)"); + + if (executable->isHostFunction()) + return static_cast<NativeExecutable*>(executable)->name(); + + if (executable->isFunctionExecutable()) + return static_cast<FunctionExecutable*>(executable)->inferredName().string(); + if (executable->isProgramExecutable() || executable->isEvalExecutable()) + return ASCIILiteral("(program)"); + if (executable->isModuleProgramExecutable()) + return ASCIILiteral("(module)"); + + RELEASE_ASSERT_NOT_REACHED(); + return String(); +} + +String SamplingProfiler::StackFrame::displayNameForJSONTests(VM& vm) +{ + { + String name = nameFromCallee(vm); + if (!name.isEmpty()) + return name; + } + + if (frameType == FrameType::Unknown || frameType == FrameType::C) + return ASCIILiteral("(unknown)"); + if (frameType == FrameType::Host) + return ASCIILiteral("(host)"); + + if (executable->isHostFunction()) + return static_cast<NativeExecutable*>(executable)->name(); + + if (executable->isFunctionExecutable()) { + String result = static_cast<FunctionExecutable*>(executable)->inferredName().string(); + if (result.isEmpty()) + return ASCIILiteral("(anonymous function)"); + return result; + } + if (executable->isEvalExecutable()) + return ASCIILiteral("(eval)"); + if (executable->isProgramExecutable()) + return ASCIILiteral("(program)"); + if (executable->isModuleProgramExecutable()) + return ASCIILiteral("(module)"); + + RELEASE_ASSERT_NOT_REACHED(); + return String(); +} + +int SamplingProfiler::StackFrame::functionStartLine() +{ + if (frameType == FrameType::Unknown || frameType == FrameType::Host || frameType == FrameType::C) + return -1; + + if (executable->isHostFunction()) + return -1; + return static_cast<ScriptExecutable*>(executable)->firstLine(); +} + +unsigned SamplingProfiler::StackFrame::functionStartColumn() +{ + if (frameType == FrameType::Unknown || frameType == FrameType::Host || frameType == FrameType::C) + return std::numeric_limits<unsigned>::max(); + + if (executable->isHostFunction()) + return std::numeric_limits<unsigned>::max(); + + return static_cast<ScriptExecutable*>(executable)->startColumn(); +} + +intptr_t SamplingProfiler::StackFrame::sourceID() +{ + if (frameType == FrameType::Unknown || frameType == FrameType::Host || frameType == FrameType::C) + return -1; + + if (executable->isHostFunction()) + return -1; + + return static_cast<ScriptExecutable*>(executable)->sourceID(); +} + +String SamplingProfiler::StackFrame::url() +{ + if (frameType == FrameType::Unknown || frameType == FrameType::Host || frameType == FrameType::C) + return emptyString(); + + if (executable->isHostFunction()) + return emptyString(); + + String url = static_cast<ScriptExecutable*>(executable)->sourceURL(); + if (url.isEmpty()) + return static_cast<ScriptExecutable*>(executable)->source().provider()->sourceURL(); // Fall back to sourceURL directive. + return url; +} + +Vector<SamplingProfiler::StackTrace> SamplingProfiler::releaseStackTraces(const LockHolder& locker) +{ + ASSERT(m_lock.isLocked()); + { + HeapIterationScope heapIterationScope(m_vm.heap); + processUnverifiedStackTraces(); + } + + Vector<StackTrace> result(WTFMove(m_stackTraces)); + clearData(locker); + return result; +} + +String SamplingProfiler::stackTracesAsJSON() +{ + DeferGC deferGC(m_vm.heap); + LockHolder locker(m_lock); + + { + HeapIterationScope heapIterationScope(m_vm.heap); + processUnverifiedStackTraces(); + } + + StringBuilder json; + json.append('['); + + bool loopedOnce = false; + auto comma = [&] { + if (loopedOnce) + json.append(','); + }; + for (StackTrace& stackTrace : m_stackTraces) { + comma(); + json.append('['); + loopedOnce = false; + for (StackFrame& stackFrame : stackTrace.frames) { + comma(); + json.append('"'); + json.append(stackFrame.displayNameForJSONTests(m_vm)); + json.append('"'); + loopedOnce = true; + } + json.append(']'); + loopedOnce = true; + } + + json.append(']'); + + clearData(locker); + + return json.toString(); +} + +void SamplingProfiler::registerForReportAtExit() +{ + static StaticLock registrationLock; + static HashSet<RefPtr<SamplingProfiler>>* profilesToReport; + + LockHolder holder(registrationLock); + + if (!profilesToReport) { + profilesToReport = new HashSet<RefPtr<SamplingProfiler>>(); + atexit([]() { + for (auto profile : *profilesToReport) + profile->reportDataToOptionFile(); + }); + } + + profilesToReport->add(adoptRef(this)); + m_needsReportAtExit = true; +} + +void SamplingProfiler::reportDataToOptionFile() +{ + if (m_needsReportAtExit) { + m_needsReportAtExit = false; + const char* path = Options::samplingProfilerPath(); + StringPrintStream pathOut; + pathOut.print(path, "/"); + pathOut.print("JSCSampilingProfile-", reinterpret_cast<uintptr_t>(this), ".txt"); + auto out = FilePrintStream::open(pathOut.toCString().data(), "w"); + reportTopFunctions(*out); + reportTopBytecodes(*out); + } +} + +void SamplingProfiler::reportTopFunctions() +{ + reportTopFunctions(WTF::dataFile()); +} + +void SamplingProfiler::reportTopFunctions(PrintStream& out) +{ + LockHolder locker(m_lock); + + { + HeapIterationScope heapIterationScope(m_vm.heap); + processUnverifiedStackTraces(); + } + + + HashMap<String, size_t> functionCounts; + for (StackTrace& stackTrace : m_stackTraces) { + if (!stackTrace.frames.size()) + continue; + + StackFrame& frame = stackTrace.frames.first(); + String frameDescription = makeString(frame.displayName(m_vm), ":", String::number(frame.sourceID())); + functionCounts.add(frameDescription, 0).iterator->value++; + } + + auto takeMax = [&] () -> std::pair<String, size_t> { + String maxFrameDescription; + size_t maxFrameCount = 0; + for (auto entry : functionCounts) { + if (entry.value > maxFrameCount) { + maxFrameCount = entry.value; + maxFrameDescription = entry.key; + } + } + if (!maxFrameDescription.isEmpty()) + functionCounts.remove(maxFrameDescription); + return std::make_pair(maxFrameDescription, maxFrameCount); + }; + + if (Options::samplingProfilerTopFunctionsCount()) { + out.print("\n\nSampling rate: ", m_timingInterval.count(), " microseconds\n"); + out.print("Top functions as <numSamples 'functionName:sourceID'>\n"); + for (size_t i = 0; i < Options::samplingProfilerTopFunctionsCount(); i++) { + auto pair = takeMax(); + if (pair.first.isEmpty()) + break; + out.printf("%6zu ", pair.second); + out.print(" '", pair.first, "'\n"); + } + } +} + +void SamplingProfiler::reportTopBytecodes() +{ + reportTopBytecodes(WTF::dataFile()); +} + +void SamplingProfiler::reportTopBytecodes(PrintStream& out) +{ + LockHolder locker(m_lock); + + { + HeapIterationScope heapIterationScope(m_vm.heap); + processUnverifiedStackTraces(); + } + + HashMap<String, size_t> bytecodeCounts; + for (StackTrace& stackTrace : m_stackTraces) { + if (!stackTrace.frames.size()) + continue; + + auto descriptionForLocation = [&] (StackFrame::CodeLocation location) -> String { + String bytecodeIndex; + String codeBlockHash; + if (location.hasBytecodeIndex()) + bytecodeIndex = String::number(location.bytecodeIndex); + else + bytecodeIndex = "<nil>"; + + if (location.hasCodeBlockHash()) { + StringPrintStream stream; + location.codeBlockHash.dump(stream); + codeBlockHash = stream.toString(); + } else + codeBlockHash = "<nil>"; + + return makeString("#", codeBlockHash, ":", JITCode::typeName(location.jitType), ":", bytecodeIndex); + }; + + StackFrame& frame = stackTrace.frames.first(); + String frameDescription = makeString(frame.displayName(m_vm), descriptionForLocation(frame.semanticLocation)); + if (std::optional<std::pair<StackFrame::CodeLocation, Strong<CodeBlock>>> machineLocation = frame.machineLocation) { + frameDescription = makeString(frameDescription, " <-- ", + machineLocation->second->inferredName().data(), descriptionForLocation(machineLocation->first)); + } + bytecodeCounts.add(frameDescription, 0).iterator->value++; + } + + auto takeMax = [&] () -> std::pair<String, size_t> { + String maxFrameDescription; + size_t maxFrameCount = 0; + for (auto entry : bytecodeCounts) { + if (entry.value > maxFrameCount) { + maxFrameCount = entry.value; + maxFrameDescription = entry.key; + } + } + if (!maxFrameDescription.isEmpty()) + bytecodeCounts.remove(maxFrameDescription); + return std::make_pair(maxFrameDescription, maxFrameCount); + }; + + if (Options::samplingProfilerTopBytecodesCount()) { + out.print("\n\nSampling rate: ", m_timingInterval.count(), " microseconds\n"); + out.print("Hottest bytecodes as <numSamples 'functionName#hash:JITType:bytecodeIndex'>\n"); + for (size_t i = 0; i < Options::samplingProfilerTopBytecodesCount(); i++) { + auto pair = takeMax(); + if (pair.first.isEmpty()) + break; + out.printf("%6zu ", pair.second); + out.print(" '", pair.first, "'\n"); + } + } +} + +} // namespace JSC + +namespace WTF { + +using namespace JSC; + +void printInternal(PrintStream& out, SamplingProfiler::FrameType frameType) +{ + switch (frameType) { + case SamplingProfiler::FrameType::Executable: + out.print("Executable"); + break; + case SamplingProfiler::FrameType::Host: + out.print("Host"); + break; + case SamplingProfiler::FrameType::C: + case SamplingProfiler::FrameType::Unknown: + out.print("Unknown"); + break; + } +} + +} // namespace WTF + +#endif // ENABLE(SAMPLING_PROFILER) |