diff options
author | Matt A. Tobin <mattatobin@localhost.localdomain> | 2018-02-02 04:16:08 -0500 |
---|---|---|
committer | Matt A. Tobin <mattatobin@localhost.localdomain> | 2018-02-02 04:16:08 -0500 |
commit | 5f8de423f190bbb79a62f804151bc24824fa32d8 (patch) | |
tree | 10027f336435511475e392454359edea8e25895d /js/src/jit/IonBuilder.cpp | |
parent | 49ee0794b5d912db1f95dce6eb52d781dc210db5 (diff) | |
download | UXP-5f8de423f190bbb79a62f804151bc24824fa32d8.tar UXP-5f8de423f190bbb79a62f804151bc24824fa32d8.tar.gz UXP-5f8de423f190bbb79a62f804151bc24824fa32d8.tar.lz UXP-5f8de423f190bbb79a62f804151bc24824fa32d8.tar.xz UXP-5f8de423f190bbb79a62f804151bc24824fa32d8.zip |
Add m-esr52 at 52.6.0
Diffstat (limited to 'js/src/jit/IonBuilder.cpp')
-rw-r--r-- | js/src/jit/IonBuilder.cpp | 14696 |
1 files changed, 14696 insertions, 0 deletions
diff --git a/js/src/jit/IonBuilder.cpp b/js/src/jit/IonBuilder.cpp new file mode 100644 index 000000000..1488d7d34 --- /dev/null +++ b/js/src/jit/IonBuilder.cpp @@ -0,0 +1,14696 @@ +/* -*- Mode: C++; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 4 -*- + * vim: set ts=8 sts=4 et sw=4 tw=99: + * This Source Code Form is subject to the terms of the Mozilla Public + * License, v. 2.0. If a copy of the MPL was not distributed with this + * file, You can obtain one at http://mozilla.org/MPL/2.0/. */ + +#include "jit/IonBuilder.h" + +#include "mozilla/DebugOnly.h" +#include "mozilla/SizePrintfMacros.h" + +#include "builtin/Eval.h" +#include "builtin/TypedObject.h" +#include "frontend/SourceNotes.h" +#include "jit/BaselineFrame.h" +#include "jit/BaselineInspector.h" +#include "jit/Ion.h" +#include "jit/IonOptimizationLevels.h" +#include "jit/JitSpewer.h" +#include "jit/Lowering.h" +#include "jit/MIRGraph.h" +#include "vm/ArgumentsObject.h" +#include "vm/Opcodes.h" +#include "vm/RegExpStatics.h" +#include "vm/TraceLogging.h" + +#include "jsopcodeinlines.h" +#include "jsscriptinlines.h" + +#include "jit/CompileInfo-inl.h" +#include "jit/shared/Lowering-shared-inl.h" +#include "vm/EnvironmentObject-inl.h" +#include "vm/NativeObject-inl.h" +#include "vm/ObjectGroup-inl.h" +#include "vm/UnboxedObject-inl.h" + +using namespace js; +using namespace js::jit; + +using mozilla::AssertedCast; +using mozilla::DebugOnly; +using mozilla::Maybe; + +using JS::TrackedStrategy; +using JS::TrackedOutcome; +using JS::TrackedTypeSite; + +class jit::BaselineFrameInspector +{ + public: + TypeSet::Type thisType; + JSObject* singletonEnvChain; + + Vector<TypeSet::Type, 4, JitAllocPolicy> argTypes; + Vector<TypeSet::Type, 4, JitAllocPolicy> varTypes; + + explicit BaselineFrameInspector(TempAllocator* temp) + : thisType(TypeSet::UndefinedType()), + singletonEnvChain(nullptr), + argTypes(*temp), + varTypes(*temp) + {} +}; + +BaselineFrameInspector* +jit::NewBaselineFrameInspector(TempAllocator* temp, BaselineFrame* frame, CompileInfo* info) +{ + MOZ_ASSERT(frame); + + BaselineFrameInspector* inspector = temp->lifoAlloc()->new_<BaselineFrameInspector>(temp); + if (!inspector) + return nullptr; + + // Note: copying the actual values into a temporary structure for use + // during compilation could capture nursery pointers, so the values' types + // are recorded instead. + + if (frame->isFunctionFrame()) + inspector->thisType = TypeSet::GetMaybeUntrackedValueType(frame->thisArgument()); + + if (frame->environmentChain()->isSingleton()) + inspector->singletonEnvChain = frame->environmentChain(); + + JSScript* script = frame->script(); + + if (script->functionNonDelazifying()) { + if (!inspector->argTypes.reserve(frame->numFormalArgs())) + return nullptr; + for (size_t i = 0; i < frame->numFormalArgs(); i++) { + if (script->formalIsAliased(i)) { + inspector->argTypes.infallibleAppend(TypeSet::UndefinedType()); + } else if (!script->argsObjAliasesFormals()) { + TypeSet::Type type = + TypeSet::GetMaybeUntrackedValueType(frame->unaliasedFormal(i)); + inspector->argTypes.infallibleAppend(type); + } else if (frame->hasArgsObj()) { + TypeSet::Type type = + TypeSet::GetMaybeUntrackedValueType(frame->argsObj().arg(i)); + inspector->argTypes.infallibleAppend(type); + } else { + inspector->argTypes.infallibleAppend(TypeSet::UndefinedType()); + } + } + } + + if (!inspector->varTypes.reserve(frame->numValueSlots())) + return nullptr; + for (size_t i = 0; i < frame->numValueSlots(); i++) { + TypeSet::Type type = TypeSet::GetMaybeUntrackedValueType(*frame->valueSlot(i)); + inspector->varTypes.infallibleAppend(type); + } + + return inspector; +} + +IonBuilder::IonBuilder(JSContext* analysisContext, CompileCompartment* comp, + const JitCompileOptions& options, TempAllocator* temp, + MIRGraph* graph, CompilerConstraintList* constraints, + BaselineInspector* inspector, CompileInfo* info, + const OptimizationInfo* optimizationInfo, + BaselineFrameInspector* baselineFrame, size_t inliningDepth, + uint32_t loopDepth) + : MIRGenerator(comp, options, temp, graph, info, optimizationInfo), + backgroundCodegen_(nullptr), + actionableAbortScript_(nullptr), + actionableAbortPc_(nullptr), + actionableAbortMessage_(nullptr), + rootList_(nullptr), + analysisContext(analysisContext), + baselineFrame_(baselineFrame), + constraints_(constraints), + analysis_(*temp, info->script()), + thisTypes(nullptr), + argTypes(nullptr), + typeArray(nullptr), + typeArrayHint(0), + bytecodeTypeMap(nullptr), + loopDepth_(loopDepth), + trackedOptimizationSites_(*temp), + lexicalCheck_(nullptr), + callerResumePoint_(nullptr), + callerBuilder_(nullptr), + cfgStack_(*temp), + loops_(*temp), + switches_(*temp), + labels_(*temp), + iterators_(*temp), + loopHeaders_(*temp), + inspector(inspector), + inliningDepth_(inliningDepth), + inlinedBytecodeLength_(0), + numLoopRestarts_(0), + failedBoundsCheck_(info->script()->failedBoundsCheck()), + failedShapeGuard_(info->script()->failedShapeGuard()), + failedLexicalCheck_(info->script()->failedLexicalCheck()), + nonStringIteration_(false), + lazyArguments_(nullptr), + inlineCallInfo_(nullptr), + maybeFallbackFunctionGetter_(nullptr) +{ + script_ = info->script(); + scriptHasIonScript_ = script_->hasIonScript(); + pc = info->startPC(); + abortReason_ = AbortReason_Disable; + + MOZ_ASSERT(script()->hasBaselineScript() == (info->analysisMode() != Analysis_ArgumentsUsage)); + MOZ_ASSERT(!!analysisContext == (info->analysisMode() == Analysis_DefiniteProperties)); + MOZ_ASSERT(script_->nTypeSets() < UINT16_MAX); + + if (!info->isAnalysis()) + script()->baselineScript()->setIonCompiledOrInlined(); +} + +void +IonBuilder::clearForBackEnd() +{ + MOZ_ASSERT(!analysisContext); + baselineFrame_ = nullptr; + + // The caches below allocate data from the malloc heap. Release this before + // later phases of compilation to avoid leaks, as the top level IonBuilder + // is not explicitly destroyed. Note that builders for inner scripts are + // constructed on the stack and will release this memory on destruction. + gsn.purge(); + envCoordinateNameCache.purge(); +} + +bool +IonBuilder::abort(const char* message, ...) +{ + // Don't call PCToLineNumber in release builds. +#ifdef JS_JITSPEW + va_list ap; + va_start(ap, message); + abortFmt(message, ap); + va_end(ap); +# ifdef DEBUG + JitSpew(JitSpew_IonAbort, "aborted @ %s:%d", script()->filename(), PCToLineNumber(script(), pc)); +# else + JitSpew(JitSpew_IonAbort, "aborted @ %s", script()->filename()); +# endif +#endif + trackActionableAbort(message); + return false; +} + +IonBuilder* +IonBuilder::outermostBuilder() +{ + IonBuilder* builder = this; + while (builder->callerBuilder_) + builder = builder->callerBuilder_; + return builder; +} + +void +IonBuilder::trackActionableAbort(const char* message) +{ + if (!isOptimizationTrackingEnabled()) + return; + + IonBuilder* topBuilder = outermostBuilder(); + if (topBuilder->hadActionableAbort()) + return; + + topBuilder->actionableAbortScript_ = script(); + topBuilder->actionableAbortPc_ = pc; + topBuilder->actionableAbortMessage_ = message; +} + +void +IonBuilder::spew(const char* message) +{ + // Don't call PCToLineNumber in release builds. +#ifdef DEBUG + JitSpew(JitSpew_IonMIR, "%s @ %s:%d", message, script()->filename(), PCToLineNumber(script(), pc)); +#endif +} + +static inline int32_t +GetJumpOffset(jsbytecode* pc) +{ + MOZ_ASSERT(CodeSpec[JSOp(*pc)].type() == JOF_JUMP); + return GET_JUMP_OFFSET(pc); +} + +IonBuilder::CFGState +IonBuilder::CFGState::If(jsbytecode* join, MTest* test) +{ + CFGState state; + state.state = IF_TRUE; + state.stopAt = join; + state.branch.ifFalse = test->ifFalse(); + state.branch.test = test; + return state; +} + +IonBuilder::CFGState +IonBuilder::CFGState::IfElse(jsbytecode* trueEnd, jsbytecode* falseEnd, MTest* test) +{ + MBasicBlock* ifFalse = test->ifFalse(); + + CFGState state; + // If the end of the false path is the same as the start of the + // false path, then the "else" block is empty and we can devolve + // this to the IF_TRUE case. We handle this here because there is + // still an extra GOTO on the true path and we want stopAt to point + // there, whereas the IF_TRUE case does not have the GOTO. + state.state = (falseEnd == ifFalse->pc()) + ? IF_TRUE_EMPTY_ELSE + : IF_ELSE_TRUE; + state.stopAt = trueEnd; + state.branch.falseEnd = falseEnd; + state.branch.ifFalse = ifFalse; + state.branch.test = test; + return state; +} + +IonBuilder::CFGState +IonBuilder::CFGState::AndOr(jsbytecode* join, MBasicBlock* lhs) +{ + CFGState state; + state.state = AND_OR; + state.stopAt = join; + state.branch.ifFalse = lhs; + state.branch.test = nullptr; + return state; +} + +IonBuilder::CFGState +IonBuilder::CFGState::TableSwitch(jsbytecode* exitpc, MTableSwitch* ins) +{ + CFGState state; + state.state = TABLE_SWITCH; + state.stopAt = exitpc; + state.tableswitch.exitpc = exitpc; + state.tableswitch.breaks = nullptr; + state.tableswitch.ins = ins; + state.tableswitch.currentBlock = 0; + return state; +} + +JSFunction* +IonBuilder::getSingleCallTarget(TemporaryTypeSet* calleeTypes) +{ + if (!calleeTypes) + return nullptr; + + JSObject* obj = calleeTypes->maybeSingleton(); + if (!obj || !obj->is<JSFunction>()) + return nullptr; + + return &obj->as<JSFunction>(); +} + +bool +IonBuilder::getPolyCallTargets(TemporaryTypeSet* calleeTypes, bool constructing, + ObjectVector& targets, uint32_t maxTargets) +{ + MOZ_ASSERT(targets.empty()); + + if (!calleeTypes) + return true; + + if (calleeTypes->baseFlags() != 0) + return true; + + unsigned objCount = calleeTypes->getObjectCount(); + + if (objCount == 0 || objCount > maxTargets) + return true; + + if (!targets.reserve(objCount)) + return false; + for (unsigned i = 0; i < objCount; i++) { + JSObject* obj = calleeTypes->getSingleton(i); + if (obj) { + MOZ_ASSERT(obj->isSingleton()); + } else { + ObjectGroup* group = calleeTypes->getGroup(i); + if (!group) + continue; + + obj = group->maybeInterpretedFunction(); + if (!obj) { + targets.clear(); + return true; + } + + MOZ_ASSERT(!obj->isSingleton()); + } + + // Don't optimize if the callee is not callable or constructable per + // the manner it is being invoked, so that CallKnown does not have to + // handle these cases (they will always throw). + if (constructing ? !obj->isConstructor() : !obj->isCallable()) { + targets.clear(); + return true; + } + + targets.infallibleAppend(obj); + } + + return true; +} + +IonBuilder::InliningDecision +IonBuilder::DontInline(JSScript* targetScript, const char* reason) +{ + if (targetScript) { + JitSpew(JitSpew_Inlining, "Cannot inline %s:%" PRIuSIZE ": %s", + targetScript->filename(), targetScript->lineno(), reason); + } else { + JitSpew(JitSpew_Inlining, "Cannot inline: %s", reason); + } + + return InliningDecision_DontInline; +} + +/* + * |hasCommonInliningPath| determines whether the current inlining path has been + * seen before based on the sequence of scripts in the chain of |IonBuilder|s. + * + * An inlining path for a function |f| is the sequence of functions whose + * inlinings precede |f| up to any previous occurrences of |f|. + * So, if we have the chain of inlinings + * + * f1 -> f2 -> f -> f3 -> f4 -> f5 -> f + * -------- -------------- + * + * the inlining paths for |f| are [f2, f1] and [f5, f4, f3]. + * When attempting to inline |f|, we find all existing inlining paths for |f| + * and check whether they share a common prefix with the path created were |f| + * inlined. + * + * For example, given mutually recursive functions |f| and |g|, a possible + * inlining is + * + * +---- Inlining stopped here... + * | + * v + * a -> f -> g -> f \ -> g -> f -> g -> ... + * + * where the vertical bar denotes the termination of inlining. + * Inlining is terminated because we have already observed the inlining path + * [f] when inlining function |g|. Note that this will inline recursive + * functions such as |fib| only one level, as |fib| has a zero length inlining + * path which trivially prefixes all inlining paths. + * + */ +bool +IonBuilder::hasCommonInliningPath(const JSScript* scriptToInline) +{ + // Find all previous inlinings of the |scriptToInline| and check for common + // inlining paths with the top of the inlining stack. + for (IonBuilder* it = this->callerBuilder_; it; it = it->callerBuilder_) { + if (it->script() != scriptToInline) + continue; + + // This only needs to check the top of each stack for a match, + // as a match of length one ensures a common prefix. + IonBuilder* path = it->callerBuilder_; + if (!path || this->script() == path->script()) + return true; + } + + return false; +} + +IonBuilder::InliningDecision +IonBuilder::canInlineTarget(JSFunction* target, CallInfo& callInfo) +{ + if (!optimizationInfo().inlineInterpreted()) { + trackOptimizationOutcome(TrackedOutcome::CantInlineGeneric); + return InliningDecision_DontInline; + } + + if (TraceLogTextIdEnabled(TraceLogger_InlinedScripts)) { + return DontInline(nullptr, "Tracelogging of inlined scripts is enabled" + "but Tracelogger cannot do that yet."); + } + + if (!target->isInterpreted()) { + trackOptimizationOutcome(TrackedOutcome::CantInlineNotInterpreted); + return DontInline(nullptr, "Non-interpreted target"); + } + + if (info().analysisMode() != Analysis_DefiniteProperties) { + // If |this| or an argument has an empty resultTypeSet, don't bother + // inlining, as the call is currently unreachable due to incomplete type + // information. This does not apply to the definite properties analysis, + // in that case we want to inline anyway. + + if (callInfo.thisArg()->emptyResultTypeSet()) { + trackOptimizationOutcome(TrackedOutcome::CantInlineUnreachable); + return DontInline(nullptr, "Empty TypeSet for |this|"); + } + + for (size_t i = 0; i < callInfo.argc(); i++) { + if (callInfo.getArg(i)->emptyResultTypeSet()) { + trackOptimizationOutcome(TrackedOutcome::CantInlineUnreachable); + return DontInline(nullptr, "Empty TypeSet for argument"); + } + } + } + + // Allow constructing lazy scripts when performing the definite properties + // analysis, as baseline has not been used to warm the caller up yet. + if (target->isInterpreted() && info().analysisMode() == Analysis_DefiniteProperties) { + RootedScript script(analysisContext, target->getOrCreateScript(analysisContext)); + if (!script) + return InliningDecision_Error; + + if (!script->hasBaselineScript() && script->canBaselineCompile()) { + MethodStatus status = BaselineCompile(analysisContext, script); + if (status == Method_Error) + return InliningDecision_Error; + if (status != Method_Compiled) { + trackOptimizationOutcome(TrackedOutcome::CantInlineNoBaseline); + return InliningDecision_DontInline; + } + } + } + + if (!target->hasScript()) { + trackOptimizationOutcome(TrackedOutcome::CantInlineLazy); + return DontInline(nullptr, "Lazy script"); + } + + JSScript* inlineScript = target->nonLazyScript(); + if (callInfo.constructing() && !target->isConstructor()) { + trackOptimizationOutcome(TrackedOutcome::CantInlineNotConstructor); + return DontInline(inlineScript, "Callee is not a constructor"); + } + + if (!callInfo.constructing() && target->isClassConstructor()) { + trackOptimizationOutcome(TrackedOutcome::CantInlineClassConstructor); + return DontInline(inlineScript, "Not constructing class constructor"); + } + + AnalysisMode analysisMode = info().analysisMode(); + if (!CanIonCompile(inlineScript, analysisMode)) { + trackOptimizationOutcome(TrackedOutcome::CantInlineDisabledIon); + return DontInline(inlineScript, "Disabled Ion compilation"); + } + + // Don't inline functions which don't have baseline scripts. + if (!inlineScript->hasBaselineScript()) { + trackOptimizationOutcome(TrackedOutcome::CantInlineNoBaseline); + return DontInline(inlineScript, "No baseline jitcode"); + } + + if (TooManyFormalArguments(target->nargs())) { + trackOptimizationOutcome(TrackedOutcome::CantInlineTooManyArgs); + return DontInline(inlineScript, "Too many args"); + } + + // We check the number of actual arguments against the maximum number of + // formal arguments as we do not want to encode all actual arguments in the + // callerResumePoint. + if (TooManyFormalArguments(callInfo.argc())) { + trackOptimizationOutcome(TrackedOutcome::CantInlineTooManyArgs); + return DontInline(inlineScript, "Too many actual args"); + } + + if (hasCommonInliningPath(inlineScript)) { + trackOptimizationOutcome(TrackedOutcome::HasCommonInliningPath); + return DontInline(inlineScript, "Common inlining path"); + } + + if (inlineScript->uninlineable()) { + trackOptimizationOutcome(TrackedOutcome::CantInlineGeneric); + return DontInline(inlineScript, "Uninlineable script"); + } + + if (inlineScript->needsArgsObj()) { + trackOptimizationOutcome(TrackedOutcome::CantInlineNeedsArgsObj); + return DontInline(inlineScript, "Script that needs an arguments object"); + } + + if (inlineScript->isDebuggee()) { + trackOptimizationOutcome(TrackedOutcome::CantInlineDebuggee); + return DontInline(inlineScript, "Script is debuggee"); + } + + TypeSet::ObjectKey* targetKey = TypeSet::ObjectKey::get(target); + if (targetKey->unknownProperties()) { + trackOptimizationOutcome(TrackedOutcome::CantInlineUnknownProps); + return DontInline(inlineScript, "Target type has unknown properties"); + } + + return InliningDecision_Inline; +} + +void +IonBuilder::popCfgStack() +{ + if (cfgStack_.back().isLoop()) + loops_.popBack(); + if (cfgStack_.back().state == CFGState::LABEL) + labels_.popBack(); + cfgStack_.popBack(); +} + +bool +IonBuilder::analyzeNewLoopTypes(MBasicBlock* entry, jsbytecode* start, jsbytecode* end) +{ + // The phi inputs at the loop head only reflect types for variables that + // were present at the start of the loop. If the variable changes to a new + // type within the loop body, and that type is carried around to the loop + // head, then we need to know about the new type up front. + // + // Since SSA information hasn't been constructed for the loop body yet, we + // need a separate analysis to pick out the types that might flow around + // the loop header. This is a best-effort analysis that may either over- + // or under-approximate the set of such types. + // + // Over-approximating the types may lead to inefficient generated code, and + // under-approximating the types will cause the loop body to be analyzed + // multiple times as the correct types are deduced (see finishLoop). + + // If we restarted processing of an outer loop then get loop header types + // directly from the last time we have previously processed this loop. This + // both avoids repeated work from the bytecode traverse below, and will + // also pick up types discovered while previously building the loop body. + for (size_t i = 0; i < loopHeaders_.length(); i++) { + if (loopHeaders_[i].pc == start) { + MBasicBlock* oldEntry = loopHeaders_[i].header; + + // If this block has been discarded, its resume points will have + // already discarded their operands. + if (!oldEntry->isDead()) { + MResumePoint* oldEntryRp = oldEntry->entryResumePoint(); + size_t stackDepth = oldEntryRp->stackDepth(); + for (size_t slot = 0; slot < stackDepth; slot++) { + MDefinition* oldDef = oldEntryRp->getOperand(slot); + if (!oldDef->isPhi()) { + MOZ_ASSERT(oldDef->block()->id() < oldEntry->id()); + MOZ_ASSERT(oldDef == entry->getSlot(slot)); + continue; + } + MPhi* oldPhi = oldDef->toPhi(); + MPhi* newPhi = entry->getSlot(slot)->toPhi(); + if (!newPhi->addBackedgeType(alloc(), oldPhi->type(), oldPhi->resultTypeSet())) + return false; + } + } + + // Update the most recent header for this loop encountered, in case + // new types flow to the phis and the loop is processed at least + // three times. + loopHeaders_[i].header = entry; + return true; + } + } + if (!loopHeaders_.append(LoopHeader(start, entry))) + return false; + + jsbytecode* last = nullptr; + jsbytecode* earlier = nullptr; + for (jsbytecode* pc = start; pc != end; earlier = last, last = pc, pc += GetBytecodeLength(pc)) { + uint32_t slot; + if (*pc == JSOP_SETLOCAL) + slot = info().localSlot(GET_LOCALNO(pc)); + else if (*pc == JSOP_SETARG) + slot = info().argSlotUnchecked(GET_ARGNO(pc)); + else + continue; + if (slot >= info().firstStackSlot()) + continue; + if (!analysis().maybeInfo(pc)) + continue; + if (!last) + continue; + + MPhi* phi = entry->getSlot(slot)->toPhi(); + + if (*last == JSOP_POS) + last = earlier; + + if (CodeSpec[*last].format & JOF_TYPESET) { + TemporaryTypeSet* typeSet = bytecodeTypes(last); + if (!typeSet->empty()) { + MIRType type = typeSet->getKnownMIRType(); + if (!phi->addBackedgeType(alloc(), type, typeSet)) + return false; + } + } else if (*last == JSOP_GETLOCAL || *last == JSOP_GETARG) { + uint32_t slot = (*last == JSOP_GETLOCAL) + ? info().localSlot(GET_LOCALNO(last)) + : info().argSlotUnchecked(GET_ARGNO(last)); + if (slot < info().firstStackSlot()) { + MPhi* otherPhi = entry->getSlot(slot)->toPhi(); + if (otherPhi->hasBackedgeType()) { + if (!phi->addBackedgeType(alloc(), otherPhi->type(), otherPhi->resultTypeSet())) + return false; + } + } + } else { + MIRType type = MIRType::None; + switch (*last) { + case JSOP_VOID: + case JSOP_UNDEFINED: + type = MIRType::Undefined; + break; + case JSOP_GIMPLICITTHIS: + if (!script()->hasNonSyntacticScope()) + type = MIRType::Undefined; + break; + case JSOP_NULL: + type = MIRType::Null; + break; + case JSOP_ZERO: + case JSOP_ONE: + case JSOP_INT8: + case JSOP_INT32: + case JSOP_UINT16: + case JSOP_UINT24: + case JSOP_BITAND: + case JSOP_BITOR: + case JSOP_BITXOR: + case JSOP_BITNOT: + case JSOP_RSH: + case JSOP_LSH: + case JSOP_URSH: + type = MIRType::Int32; + break; + case JSOP_FALSE: + case JSOP_TRUE: + case JSOP_EQ: + case JSOP_NE: + case JSOP_LT: + case JSOP_LE: + case JSOP_GT: + case JSOP_GE: + case JSOP_NOT: + case JSOP_STRICTEQ: + case JSOP_STRICTNE: + case JSOP_IN: + case JSOP_INSTANCEOF: + type = MIRType::Boolean; + break; + case JSOP_DOUBLE: + type = MIRType::Double; + break; + case JSOP_STRING: + case JSOP_TOSTRING: + case JSOP_TYPEOF: + case JSOP_TYPEOFEXPR: + type = MIRType::String; + break; + case JSOP_SYMBOL: + type = MIRType::Symbol; + break; + case JSOP_ADD: + case JSOP_SUB: + case JSOP_MUL: + case JSOP_DIV: + case JSOP_MOD: + case JSOP_NEG: + type = inspector->expectedResultType(last); + break; + default: + break; + } + if (type != MIRType::None) { + if (!phi->addBackedgeType(alloc(), type, nullptr)) + return false; + } + } + } + return true; +} + +bool +IonBuilder::pushLoop(CFGState::State initial, jsbytecode* stopAt, MBasicBlock* entry, bool osr, + jsbytecode* loopHead, jsbytecode* initialPc, + jsbytecode* bodyStart, jsbytecode* bodyEnd, + jsbytecode* exitpc, jsbytecode* continuepc) +{ + ControlFlowInfo loop(cfgStack_.length(), continuepc); + if (!loops_.append(loop)) + return false; + + CFGState state; + state.state = initial; + state.stopAt = stopAt; + state.loop.bodyStart = bodyStart; + state.loop.bodyEnd = bodyEnd; + state.loop.exitpc = exitpc; + state.loop.continuepc = continuepc; + state.loop.entry = entry; + state.loop.osr = osr; + state.loop.successor = nullptr; + state.loop.breaks = nullptr; + state.loop.continues = nullptr; + state.loop.initialState = initial; + state.loop.initialPc = initialPc; + state.loop.initialStopAt = stopAt; + state.loop.loopHead = loopHead; + return cfgStack_.append(state); +} + +bool +IonBuilder::init() +{ + { + LifoAlloc::AutoFallibleScope fallibleAllocator(alloc().lifoAlloc()); + if (!TypeScript::FreezeTypeSets(constraints(), script(), &thisTypes, &argTypes, &typeArray)) + return false; + } + + if (!alloc().ensureBallast()) + return false; + + if (inlineCallInfo_) { + // If we're inlining, the actual this/argument types are not necessarily + // a subset of the script's observed types. |argTypes| is never accessed + // for inlined scripts, so we just null it. + thisTypes = inlineCallInfo_->thisArg()->resultTypeSet(); + argTypes = nullptr; + } + + if (!analysis().init(alloc(), gsn)) + return false; + + // The baseline script normally has the bytecode type map, but compute + // it ourselves if we do not have a baseline script. + if (script()->hasBaselineScript()) { + bytecodeTypeMap = script()->baselineScript()->bytecodeTypeMap(); + } else { + bytecodeTypeMap = alloc_->lifoAlloc()->newArrayUninitialized<uint32_t>(script()->nTypeSets()); + if (!bytecodeTypeMap) + return false; + FillBytecodeTypeMap(script(), bytecodeTypeMap); + } + + return true; +} + +bool +IonBuilder::build() +{ + if (!init()) + return false; + + if (script()->hasBaselineScript()) + script()->baselineScript()->resetMaxInliningDepth(); + + if (!setCurrentAndSpecializePhis(newBlock(pc))) + return false; + if (!current) + return false; + +#ifdef JS_JITSPEW + if (info().isAnalysis()) { + JitSpew(JitSpew_IonScripts, "Analyzing script %s:%" PRIuSIZE " (%p) %s", + script()->filename(), script()->lineno(), (void*)script(), + AnalysisModeString(info().analysisMode())); + } else { + JitSpew(JitSpew_IonScripts, "%sompiling script %s:%" PRIuSIZE " (%p) (warmup-counter=%" PRIu32 ", level=%s)", + (script()->hasIonScript() ? "Rec" : "C"), + script()->filename(), script()->lineno(), (void*)script(), + script()->getWarmUpCount(), OptimizationLevelString(optimizationInfo().level())); + } +#endif + + if (!initParameters()) + return false; + initLocals(); + + // Initialize something for the env chain. We can bail out before the + // start instruction, but the snapshot is encoded *at* the start + // instruction, which means generating any code that could load into + // registers is illegal. + MInstruction* env = MConstant::New(alloc(), UndefinedValue()); + current->add(env); + current->initSlot(info().environmentChainSlot(), env); + + // Initialize the return value. + MInstruction* returnValue = MConstant::New(alloc(), UndefinedValue()); + current->add(returnValue); + current->initSlot(info().returnValueSlot(), returnValue); + + // Initialize the arguments object slot to undefined if necessary. + if (info().hasArguments()) { + MInstruction* argsObj = MConstant::New(alloc(), UndefinedValue()); + current->add(argsObj); + current->initSlot(info().argsObjSlot(), argsObj); + } + + // Emit the start instruction, so we can begin real instructions. + current->add(MStart::New(alloc())); + + // Guard against over-recursion. Do this before we start unboxing, since + // this will create an OSI point that will read the incoming argument + // values, which is nice to do before their last real use, to minimize + // register/stack pressure. + MCheckOverRecursed* check = MCheckOverRecursed::New(alloc()); + current->add(check); + MResumePoint* entryRpCopy = MResumePoint::Copy(alloc(), current->entryResumePoint()); + if (!entryRpCopy) + return false; + check->setResumePoint(entryRpCopy); + + // Parameters have been checked to correspond to the typeset, now we unbox + // what we can in an infallible manner. + if (!rewriteParameters()) + return false; + + // Check for redeclaration errors for global scripts. + if (!info().funMaybeLazy() && !info().module() && + script()->bodyScope()->is<GlobalScope>() && + script()->bodyScope()->as<GlobalScope>().hasBindings()) + { + MGlobalNameConflictsCheck* redeclCheck = MGlobalNameConflictsCheck::New(alloc()); + current->add(redeclCheck); + MResumePoint* entryRpCopy = MResumePoint::Copy(alloc(), current->entryResumePoint()); + if (!entryRpCopy) + return false; + redeclCheck->setResumePoint(entryRpCopy); + } + + // It's safe to start emitting actual IR, so now build the env chain. + if (!initEnvironmentChain()) + return false; + + if (info().needsArgsObj() && !initArgumentsObject()) + return false; + + // The type analysis phase attempts to insert unbox operations near + // definitions of values. It also attempts to replace uses in resume points + // with the narrower, unboxed variants. However, we must prevent this + // replacement from happening on values in the entry snapshot. Otherwise we + // could get this: + // + // v0 = MParameter(0) + // v1 = MParameter(1) + // -- ResumePoint(v2, v3) + // v2 = Unbox(v0, INT32) + // v3 = Unbox(v1, INT32) + // + // So we attach the initial resume point to each parameter, which the type + // analysis explicitly checks (this is the same mechanism used for + // effectful operations). + for (uint32_t i = 0; i < info().endArgSlot(); i++) { + MInstruction* ins = current->getEntrySlot(i)->toInstruction(); + if (ins->type() != MIRType::Value) + continue; + + MResumePoint* entryRpCopy = MResumePoint::Copy(alloc(), current->entryResumePoint()); + if (!entryRpCopy) + return false; + ins->setResumePoint(entryRpCopy); + } + + // lazyArguments should never be accessed in |argsObjAliasesFormals| scripts. + if (info().hasArguments() && !info().argsObjAliasesFormals()) { + lazyArguments_ = MConstant::New(alloc(), MagicValue(JS_OPTIMIZED_ARGUMENTS)); + current->add(lazyArguments_); + } + + insertRecompileCheck(); + + if (!traverseBytecode()) + return false; + + // Discard unreferenced & pre-allocated resume points. + replaceMaybeFallbackFunctionGetter(nullptr); + + if (script_->hasBaselineScript() && + inlinedBytecodeLength_ > script_->baselineScript()->inlinedBytecodeLength()) + { + script_->baselineScript()->setInlinedBytecodeLength(inlinedBytecodeLength_); + } + + if (!maybeAddOsrTypeBarriers()) + return false; + + if (!processIterators()) + return false; + + if (!info().isAnalysis() && !abortedPreliminaryGroups().empty()) { + abortReason_ = AbortReason_PreliminaryObjects; + return false; + } + + if (shouldForceAbort()) { + abortReason_ = AbortReason_Disable; + return false; + } + + MOZ_ASSERT(loopDepth_ == 0); + abortReason_ = AbortReason_NoAbort; + return true; +} + +bool +IonBuilder::processIterators() +{ + // Find phis that must directly hold an iterator live. + Vector<MPhi*, 0, SystemAllocPolicy> worklist; + for (size_t i = 0; i < iterators_.length(); i++) { + MInstruction* ins = iterators_[i]; + for (MUseDefIterator iter(ins); iter; iter++) { + if (iter.def()->isPhi()) { + if (!worklist.append(iter.def()->toPhi())) + return false; + } + } + } + + // Propagate the iterator and live status of phis to all other connected + // phis. + while (!worklist.empty()) { + MPhi* phi = worklist.popCopy(); + phi->setIterator(); + phi->setImplicitlyUsedUnchecked(); + + for (MUseDefIterator iter(phi); iter; iter++) { + if (iter.def()->isPhi()) { + MPhi* other = iter.def()->toPhi(); + if (!other->isIterator() && !worklist.append(other)) + return false; + } + } + } + + return true; +} + +bool +IonBuilder::buildInline(IonBuilder* callerBuilder, MResumePoint* callerResumePoint, + CallInfo& callInfo) +{ + inlineCallInfo_ = &callInfo; + + if (!init()) + return false; + + JitSpew(JitSpew_IonScripts, "Inlining script %s:%" PRIuSIZE " (%p)", + script()->filename(), script()->lineno(), (void*)script()); + + callerBuilder_ = callerBuilder; + callerResumePoint_ = callerResumePoint; + + if (callerBuilder->failedBoundsCheck_) + failedBoundsCheck_ = true; + + if (callerBuilder->failedShapeGuard_) + failedShapeGuard_ = true; + + if (callerBuilder->failedLexicalCheck_) + failedLexicalCheck_ = true; + + safeForMinorGC_ = callerBuilder->safeForMinorGC_; + + // Generate single entrance block. + if (!setCurrentAndSpecializePhis(newBlock(pc))) + return false; + if (!current) + return false; + + current->setCallerResumePoint(callerResumePoint); + + // Connect the entrance block to the last block in the caller's graph. + MBasicBlock* predecessor = callerBuilder->current; + MOZ_ASSERT(predecessor == callerResumePoint->block()); + + predecessor->end(MGoto::New(alloc(), current)); + if (!current->addPredecessorWithoutPhis(predecessor)) + return false; + + // Initialize env chain slot to Undefined. It's set later by + // |initEnvironmentChain|. + MInstruction* env = MConstant::New(alloc(), UndefinedValue()); + current->add(env); + current->initSlot(info().environmentChainSlot(), env); + + // Initialize |return value| slot. + MInstruction* returnValue = MConstant::New(alloc(), UndefinedValue()); + current->add(returnValue); + current->initSlot(info().returnValueSlot(), returnValue); + + // Initialize |arguments| slot. + if (info().hasArguments()) { + MInstruction* argsObj = MConstant::New(alloc(), UndefinedValue()); + current->add(argsObj); + current->initSlot(info().argsObjSlot(), argsObj); + } + + // Initialize |this| slot. + current->initSlot(info().thisSlot(), callInfo.thisArg()); + + JitSpew(JitSpew_Inlining, "Initializing %u arg slots", info().nargs()); + + // NB: Ion does not inline functions which |needsArgsObj|. So using argSlot() + // instead of argSlotUnchecked() below is OK + MOZ_ASSERT(!info().needsArgsObj()); + + // Initialize actually set arguments. + uint32_t existing_args = Min<uint32_t>(callInfo.argc(), info().nargs()); + for (size_t i = 0; i < existing_args; ++i) { + MDefinition* arg = callInfo.getArg(i); + current->initSlot(info().argSlot(i), arg); + } + + // Pass Undefined for missing arguments + for (size_t i = callInfo.argc(); i < info().nargs(); ++i) { + MConstant* arg = MConstant::New(alloc(), UndefinedValue()); + current->add(arg); + current->initSlot(info().argSlot(i), arg); + } + + JitSpew(JitSpew_Inlining, "Initializing %u locals", info().nlocals()); + + initLocals(); + + JitSpew(JitSpew_Inlining, "Inline entry block MResumePoint %p, %u stack slots", + (void*) current->entryResumePoint(), current->entryResumePoint()->stackDepth()); + + // +2 for the env chain and |this|, maybe another +1 for arguments object slot. + MOZ_ASSERT(current->entryResumePoint()->stackDepth() == info().totalSlots()); + + if (script_->argumentsHasVarBinding()) { + lazyArguments_ = MConstant::New(alloc(), MagicValue(JS_OPTIMIZED_ARGUMENTS)); + current->add(lazyArguments_); + } + + insertRecompileCheck(); + + // Initialize the env chain now that all resume points operands are + // initialized. + if (!initEnvironmentChain(callInfo.fun())) + return false; + + if (!traverseBytecode()) + return false; + + // Discard unreferenced & pre-allocated resume points. + replaceMaybeFallbackFunctionGetter(nullptr); + + MOZ_ASSERT(iterators_.empty(), "Iterators should be added to outer builder"); + + if (!info().isAnalysis() && !abortedPreliminaryGroups().empty()) { + abortReason_ = AbortReason_PreliminaryObjects; + return false; + } + + if (shouldForceAbort()) { + abortReason_ = AbortReason_Disable; + return false; + } + + return true; +} + +void +IonBuilder::rewriteParameter(uint32_t slotIdx, MDefinition* param, int32_t argIndex) +{ + MOZ_ASSERT(param->isParameter() || param->isGetArgumentsObjectArg()); + + TemporaryTypeSet* types = param->resultTypeSet(); + MDefinition* actual = ensureDefiniteType(param, types->getKnownMIRType()); + if (actual == param) + return; + + // Careful! We leave the original MParameter in the entry resume point. The + // arguments still need to be checked unless proven otherwise at the call + // site, and these checks can bailout. We can end up: + // v0 = Parameter(0) + // v1 = Unbox(v0, INT32) + // -- ResumePoint(v0) + // + // As usual, it would be invalid for v1 to be captured in the initial + // resume point, rather than v0. + current->rewriteSlot(slotIdx, actual); +} + +// Apply Type Inference information to parameters early on, unboxing them if +// they have a definitive type. The actual guards will be emitted by the code +// generator, explicitly, as part of the function prologue. +bool +IonBuilder::rewriteParameters() +{ + MOZ_ASSERT(info().environmentChainSlot() == 0); + + if (!info().funMaybeLazy()) + return true; + + for (uint32_t i = info().startArgSlot(); i < info().endArgSlot(); i++) { + if (!alloc().ensureBallast()) + return false; + MDefinition* param = current->getSlot(i); + rewriteParameter(i, param, param->toParameter()->index()); + } + + return true; +} + +bool +IonBuilder::initParameters() +{ + if (!info().funMaybeLazy()) + return true; + + // If we are doing OSR on a frame which initially executed in the + // interpreter and didn't accumulate type information, try to use that OSR + // frame to determine possible initial types for 'this' and parameters. + + if (thisTypes->empty() && baselineFrame_) { + TypeSet::Type type = baselineFrame_->thisType; + if (type.isSingletonUnchecked()) + checkNurseryObject(type.singleton()); + thisTypes->addType(type, alloc_->lifoAlloc()); + } + + MParameter* param = MParameter::New(alloc(), MParameter::THIS_SLOT, thisTypes); + current->add(param); + current->initSlot(info().thisSlot(), param); + + for (uint32_t i = 0; i < info().nargs(); i++) { + TemporaryTypeSet* types = &argTypes[i]; + if (types->empty() && baselineFrame_ && + !script_->baselineScript()->modifiesArguments()) + { + TypeSet::Type type = baselineFrame_->argTypes[i]; + if (type.isSingletonUnchecked()) + checkNurseryObject(type.singleton()); + types->addType(type, alloc_->lifoAlloc()); + } + + param = MParameter::New(alloc().fallible(), i, types); + if (!param) + return false; + current->add(param); + current->initSlot(info().argSlotUnchecked(i), param); + } + + return true; +} + +void +IonBuilder::initLocals() +{ + // Initialize all frame slots to undefined. Lexical bindings are temporal + // dead zoned in bytecode. + + if (info().nlocals() == 0) + return; + + MConstant* undef = MConstant::New(alloc(), UndefinedValue()); + current->add(undef); + + for (uint32_t i = 0; i < info().nlocals(); i++) + current->initSlot(info().localSlot(i), undef); +} + +bool +IonBuilder::initEnvironmentChain(MDefinition* callee) +{ + MInstruction* env = nullptr; + + // If the script doesn't use the envchain, then it's already initialized + // from earlier. However, always make a env chain when |needsArgsObj| is true + // for the script, since arguments object construction requires the env chain + // to be passed in. + if (!info().needsArgsObj() && !analysis().usesEnvironmentChain()) + return true; + + // The env chain is only tracked in scripts that have NAME opcodes which + // will try to access the env. For other scripts, the env instructions + // will be held live by resume points and code will still be generated for + // them, so just use a constant undefined value. + + if (JSFunction* fun = info().funMaybeLazy()) { + if (!callee) { + MCallee* calleeIns = MCallee::New(alloc()); + current->add(calleeIns); + callee = calleeIns; + } + env = MFunctionEnvironment::New(alloc(), callee); + current->add(env); + + // This reproduce what is done in CallObject::createForFunction. Skip + // this for the arguments analysis, as the script might not have a + // baseline script with template objects yet. + if (fun->needsSomeEnvironmentObject() && + info().analysisMode() != Analysis_ArgumentsUsage) + { + if (fun->needsNamedLambdaEnvironment()) { + env = createNamedLambdaObject(callee, env); + if (!env) + return false; + } + + // TODO: Parameter expression-induced extra var environment not + // yet handled. + if (fun->needsExtraBodyVarEnvironment()) + return abort("Extra var environment unsupported"); + + if (fun->needsCallObject()) { + env = createCallObject(callee, env); + if (!env) + return false; + } + } + } else if (ModuleObject* module = info().module()) { + // Modules use a pre-created env object. + env = constant(ObjectValue(module->initialEnvironment())); + } else { + // For global scripts without a non-syntactic global scope, the env + // chain is the global lexical env. + MOZ_ASSERT(!script()->isForEval()); + MOZ_ASSERT(!script()->hasNonSyntacticScope()); + env = constant(ObjectValue(script()->global().lexicalEnvironment())); + } + + current->setEnvironmentChain(env); + return true; +} + +bool +IonBuilder::initArgumentsObject() +{ + JitSpew(JitSpew_IonMIR, "%s:%" PRIuSIZE " - Emitting code to initialize arguments object! block=%p", + script()->filename(), script()->lineno(), current); + MOZ_ASSERT(info().needsArgsObj()); + + bool mapped = script()->hasMappedArgsObj(); + ArgumentsObject* templateObj = script()->compartment()->maybeArgumentsTemplateObject(mapped); + + MCreateArgumentsObject* argsObj = + MCreateArgumentsObject::New(alloc(), current->environmentChain(), templateObj); + current->add(argsObj); + current->setArgumentsObject(argsObj); + return true; +} + +bool +IonBuilder::addOsrValueTypeBarrier(uint32_t slot, MInstruction** def_, + MIRType type, TemporaryTypeSet* typeSet) +{ + MInstruction*& def = *def_; + MBasicBlock* osrBlock = def->block(); + + // Clear bogus type information added in newOsrPreheader(). + def->setResultType(MIRType::Value); + def->setResultTypeSet(nullptr); + + if (typeSet && !typeSet->unknown()) { + MInstruction* barrier = MTypeBarrier::New(alloc(), def, typeSet); + osrBlock->insertBefore(osrBlock->lastIns(), barrier); + osrBlock->rewriteSlot(slot, barrier); + def = barrier; + + // If the TypeSet is more precise than |type|, adjust |type| for the + // code below. + if (type == MIRType::Value) + type = barrier->type(); + } else if (type == MIRType::Null || + type == MIRType::Undefined || + type == MIRType::MagicOptimizedArguments) + { + // No unbox instruction will be added below, so check the type by + // adding a type barrier for a singleton type set. + TypeSet::Type ntype = TypeSet::PrimitiveType(ValueTypeFromMIRType(type)); + LifoAlloc* lifoAlloc = alloc().lifoAlloc(); + typeSet = lifoAlloc->new_<TemporaryTypeSet>(lifoAlloc, ntype); + if (!typeSet) + return false; + MInstruction* barrier = MTypeBarrier::New(alloc(), def, typeSet); + osrBlock->insertBefore(osrBlock->lastIns(), barrier); + osrBlock->rewriteSlot(slot, barrier); + def = barrier; + } + + switch (type) { + case MIRType::Boolean: + case MIRType::Int32: + case MIRType::Double: + case MIRType::String: + case MIRType::Symbol: + case MIRType::Object: + if (type != def->type()) { + MUnbox* unbox = MUnbox::New(alloc(), def, type, MUnbox::Fallible); + osrBlock->insertBefore(osrBlock->lastIns(), unbox); + osrBlock->rewriteSlot(slot, unbox); + def = unbox; + } + break; + + case MIRType::Null: + { + MConstant* c = MConstant::New(alloc(), NullValue()); + osrBlock->insertBefore(osrBlock->lastIns(), c); + osrBlock->rewriteSlot(slot, c); + def = c; + break; + } + + case MIRType::Undefined: + { + MConstant* c = MConstant::New(alloc(), UndefinedValue()); + osrBlock->insertBefore(osrBlock->lastIns(), c); + osrBlock->rewriteSlot(slot, c); + def = c; + break; + } + + case MIRType::MagicOptimizedArguments: + MOZ_ASSERT(lazyArguments_); + osrBlock->rewriteSlot(slot, lazyArguments_); + def = lazyArguments_; + break; + + default: + break; + } + + MOZ_ASSERT(def == osrBlock->getSlot(slot)); + return true; +} + +bool +IonBuilder::maybeAddOsrTypeBarriers() +{ + if (!info().osrPc()) + return true; + + // The loop has successfully been processed, and the loop header phis + // have their final type. Add unboxes and type barriers in the OSR + // block to check that the values have the appropriate type, and update + // the types in the preheader. + + MBasicBlock* osrBlock = graph().osrBlock(); + if (!osrBlock) { + // Because IonBuilder does not compile catch blocks, it's possible to + // end up without an OSR block if the OSR pc is only reachable via a + // break-statement inside the catch block. For instance: + // + // for (;;) { + // try { + // throw 3; + // } catch(e) { + // break; + // } + // } + // while (..) { } // <= OSR here, only reachable via catch block. + // + // For now we just abort in this case. + MOZ_ASSERT(graph().hasTryBlock()); + return abort("OSR block only reachable through catch block"); + } + + MBasicBlock* preheader = osrBlock->getSuccessor(0); + MBasicBlock* header = preheader->getSuccessor(0); + static const size_t OSR_PHI_POSITION = 1; + MOZ_ASSERT(preheader->getPredecessor(OSR_PHI_POSITION) == osrBlock); + + MResumePoint* headerRp = header->entryResumePoint(); + size_t stackDepth = headerRp->stackDepth(); + MOZ_ASSERT(stackDepth == osrBlock->stackDepth()); + for (uint32_t slot = info().startArgSlot(); slot < stackDepth; slot++) { + // Aliased slots are never accessed, since they need to go through + // the callobject. The typebarriers are added there and can be + // discarded here. + if (info().isSlotAliased(slot)) + continue; + + if (!alloc().ensureBallast()) + return false; + + MInstruction* def = osrBlock->getSlot(slot)->toInstruction(); + MPhi* preheaderPhi = preheader->getSlot(slot)->toPhi(); + MPhi* headerPhi = headerRp->getOperand(slot)->toPhi(); + + MIRType type = headerPhi->type(); + TemporaryTypeSet* typeSet = headerPhi->resultTypeSet(); + + if (!addOsrValueTypeBarrier(slot, &def, type, typeSet)) + return false; + + preheaderPhi->replaceOperand(OSR_PHI_POSITION, def); + preheaderPhi->setResultType(type); + preheaderPhi->setResultTypeSet(typeSet); + } + + return true; +} + +// We try to build a control-flow graph in the order that it would be built as +// if traversing the AST. This leads to a nice ordering and lets us build SSA +// in one pass, since the bytecode is structured. +// +// We traverse the bytecode iteratively, maintaining a current basic block. +// Each basic block has a mapping of local slots to instructions, as well as a +// stack depth. As we encounter instructions we mutate this mapping in the +// current block. +// +// Things get interesting when we encounter a control structure. This can be +// either an IFEQ, downward GOTO, or a decompiler hint stashed away in source +// notes. Once we encounter such an opcode, we recover the structure of the +// control flow (its branches and bounds), and push it on a stack. +// +// As we continue traversing the bytecode, we look for points that would +// terminate the topmost control flow path pushed on the stack. These are: +// (1) The bounds of the current structure (end of a loop or join/edge of a +// branch). +// (2) A "return", "break", or "continue" statement. +// +// For (1), we expect that there is a current block in the progress of being +// built, and we complete the necessary edges in the CFG. For (2), we expect +// that there is no active block. +// +// For normal diamond join points, we construct Phi nodes as we add +// predecessors. For loops, care must be taken to propagate Phi nodes back +// through uses in the loop body. +bool +IonBuilder::traverseBytecode() +{ + for (;;) { + MOZ_ASSERT(pc < info().limitPC()); + + for (;;) { + if (!alloc().ensureBallast()) + return false; + + // Check if we've hit an expected join point or edge in the bytecode. + // Leaving one control structure could place us at the edge of another, + // thus |while| instead of |if| so we don't skip any opcodes. + MOZ_ASSERT_IF(!cfgStack_.empty(), cfgStack_.back().stopAt >= pc); + if (!cfgStack_.empty() && cfgStack_.back().stopAt == pc) { + ControlStatus status = processCfgStack(); + if (status == ControlStatus_Error) + return false; + if (status == ControlStatus_Abort) + return abort("Aborted while processing control flow"); + if (!current) + return true; + continue; + } + + // Some opcodes need to be handled early because they affect control + // flow, terminating the current basic block and/or instructing the + // traversal algorithm to continue from a new pc. + // + // (1) If the opcode does not affect control flow, then the opcode + // is inspected and transformed to IR. This is the process_opcode + // label. + // (2) A loop could be detected via a forward GOTO. In this case, + // we don't want to process the GOTO, but the following + // instruction. + // (3) A RETURN, STOP, BREAK, or CONTINUE may require processing the + // CFG stack to terminate open branches. + // + // Similar to above, snooping control flow could land us at another + // control flow point, so we iterate until it's time to inspect a real + // opcode. + ControlStatus status; + if ((status = snoopControlFlow(JSOp(*pc))) == ControlStatus_None) + break; + if (status == ControlStatus_Error) + return false; + if (status == ControlStatus_Abort) + return abort("Aborted while processing control flow"); + if (!current) + return true; + } + +#ifdef DEBUG + // In debug builds, after compiling this op, check that all values + // popped by this opcode either: + // + // (1) Have the ImplicitlyUsed flag set on them. + // (2) Have more uses than before compiling this op (the value is + // used as operand of a new MIR instruction). + // + // This is used to catch problems where IonBuilder pops a value without + // adding any SSA uses and doesn't call setImplicitlyUsedUnchecked on it. + Vector<MDefinition*, 4, JitAllocPolicy> popped(alloc()); + Vector<size_t, 4, JitAllocPolicy> poppedUses(alloc()); + unsigned nuses = GetUseCount(script_, script_->pcToOffset(pc)); + + for (unsigned i = 0; i < nuses; i++) { + MDefinition* def = current->peek(-int32_t(i + 1)); + if (!popped.append(def) || !poppedUses.append(def->defUseCount())) + return false; + } +#endif + + // Nothing in inspectOpcode() is allowed to advance the pc. + JSOp op = JSOp(*pc); + if (!inspectOpcode(op)) + return false; + +#ifdef DEBUG + for (size_t i = 0; i < popped.length(); i++) { + switch (op) { + case JSOP_POP: + case JSOP_POPN: + case JSOP_DUPAT: + case JSOP_DUP: + case JSOP_DUP2: + case JSOP_PICK: + case JSOP_SWAP: + case JSOP_SETARG: + case JSOP_SETLOCAL: + case JSOP_INITLEXICAL: + case JSOP_SETRVAL: + case JSOP_VOID: + // Don't require SSA uses for values popped by these ops. + break; + + case JSOP_POS: + case JSOP_TOID: + case JSOP_TOSTRING: + // These ops may leave their input on the stack without setting + // the ImplicitlyUsed flag. If this value will be popped immediately, + // we may replace it with |undefined|, but the difference is + // not observable. + MOZ_ASSERT(i == 0); + if (current->peek(-1) == popped[0]) + break; + MOZ_FALLTHROUGH; + + default: + MOZ_ASSERT(popped[i]->isImplicitlyUsed() || + + // MNewDerivedTypedObject instances are + // often dead unless they escape from the + // fn. See IonBuilder::loadTypedObjectData() + // for more details. + popped[i]->isNewDerivedTypedObject() || + + popped[i]->defUseCount() > poppedUses[i]); + break; + } + } +#endif + + pc += CodeSpec[op].length; + current->updateTrackedSite(bytecodeSite(pc)); + } + + return true; +} + +IonBuilder::ControlStatus +IonBuilder::snoopControlFlow(JSOp op) +{ + switch (op) { + case JSOP_NOP: + return maybeLoop(op, info().getNote(gsn, pc)); + + case JSOP_POP: + return maybeLoop(op, info().getNote(gsn, pc)); + + case JSOP_RETURN: + case JSOP_RETRVAL: + return processReturn(op); + + case JSOP_THROW: + return processThrow(); + + case JSOP_GOTO: + { + jssrcnote* sn = info().getNote(gsn, pc); + switch (sn ? SN_TYPE(sn) : SRC_NULL) { + case SRC_BREAK: + case SRC_BREAK2LABEL: + return processBreak(op, sn); + + case SRC_CONTINUE: + return processContinue(op); + + case SRC_SWITCHBREAK: + return processSwitchBreak(op); + + case SRC_WHILE: + case SRC_FOR_IN: + case SRC_FOR_OF: + // while (cond) { } + return whileOrForInLoop(sn); + + default: + // Hard assert for now - make an error later. + MOZ_CRASH("unknown goto case"); + } + break; + } + + case JSOP_TABLESWITCH: + return tableSwitch(op, info().getNote(gsn, pc)); + + case JSOP_IFNE: + // We should never reach an IFNE, it's a stopAt point, which will + // trigger closing the loop. + MOZ_CRASH("we should never reach an ifne!"); + + default: + break; + } + return ControlStatus_None; +} + +bool +IonBuilder::inspectOpcode(JSOp op) +{ + MOZ_ASSERT(analysis_.maybeInfo(pc), "Compiling unreachable op"); + + switch (op) { + case JSOP_NOP: + case JSOP_NOP_DESTRUCTURING: + case JSOP_LINENO: + case JSOP_LOOPENTRY: + case JSOP_JUMPTARGET: + return true; + + case JSOP_LABEL: + return jsop_label(); + + case JSOP_UNDEFINED: + // If this ever changes, change what JSOP_GIMPLICITTHIS does too. + pushConstant(UndefinedValue()); + return true; + + case JSOP_IFEQ: + return jsop_ifeq(JSOP_IFEQ); + + case JSOP_TRY: + return jsop_try(); + + case JSOP_CONDSWITCH: + return jsop_condswitch(); + + case JSOP_BITNOT: + return jsop_bitnot(); + + case JSOP_BITAND: + case JSOP_BITOR: + case JSOP_BITXOR: + case JSOP_LSH: + case JSOP_RSH: + case JSOP_URSH: + return jsop_bitop(op); + + case JSOP_ADD: + case JSOP_SUB: + case JSOP_MUL: + case JSOP_DIV: + case JSOP_MOD: + return jsop_binary_arith(op); + + case JSOP_POW: + return jsop_pow(); + + case JSOP_POS: + return jsop_pos(); + + case JSOP_NEG: + return jsop_neg(); + + case JSOP_TOSTRING: + return jsop_tostring(); + + case JSOP_AND: + case JSOP_OR: + return jsop_andor(op); + + case JSOP_DEFVAR: + return jsop_defvar(GET_UINT32_INDEX(pc)); + + case JSOP_DEFLET: + case JSOP_DEFCONST: + return jsop_deflexical(GET_UINT32_INDEX(pc)); + + case JSOP_DEFFUN: + return jsop_deffun(GET_UINT32_INDEX(pc)); + + case JSOP_EQ: + case JSOP_NE: + case JSOP_STRICTEQ: + case JSOP_STRICTNE: + case JSOP_LT: + case JSOP_LE: + case JSOP_GT: + case JSOP_GE: + return jsop_compare(op); + + case JSOP_DOUBLE: + pushConstant(info().getConst(pc)); + return true; + + case JSOP_STRING: + pushConstant(StringValue(info().getAtom(pc))); + return true; + + case JSOP_SYMBOL: { + unsigned which = GET_UINT8(pc); + JS::Symbol* sym = compartment->runtime()->wellKnownSymbols().get(which); + pushConstant(SymbolValue(sym)); + return true; + } + + case JSOP_ZERO: + pushConstant(Int32Value(0)); + return true; + + case JSOP_ONE: + pushConstant(Int32Value(1)); + return true; + + case JSOP_NULL: + pushConstant(NullValue()); + return true; + + case JSOP_VOID: + current->pop(); + pushConstant(UndefinedValue()); + return true; + + case JSOP_HOLE: + pushConstant(MagicValue(JS_ELEMENTS_HOLE)); + return true; + + case JSOP_FALSE: + pushConstant(BooleanValue(false)); + return true; + + case JSOP_TRUE: + pushConstant(BooleanValue(true)); + return true; + + case JSOP_ARGUMENTS: + return jsop_arguments(); + + case JSOP_RUNONCE: + return jsop_runonce(); + + case JSOP_REST: + return jsop_rest(); + + case JSOP_GETARG: + if (info().argsObjAliasesFormals()) { + MGetArgumentsObjectArg* getArg = MGetArgumentsObjectArg::New(alloc(), + current->argumentsObject(), + GET_ARGNO(pc)); + current->add(getArg); + current->push(getArg); + } else { + current->pushArg(GET_ARGNO(pc)); + } + return true; + + case JSOP_SETARG: + return jsop_setarg(GET_ARGNO(pc)); + + case JSOP_GETLOCAL: + current->pushLocal(GET_LOCALNO(pc)); + return true; + + case JSOP_SETLOCAL: + current->setLocal(GET_LOCALNO(pc)); + return true; + + case JSOP_THROWSETCONST: + case JSOP_THROWSETALIASEDCONST: + case JSOP_THROWSETCALLEE: + return jsop_throwsetconst(); + + case JSOP_CHECKLEXICAL: + return jsop_checklexical(); + + case JSOP_INITLEXICAL: + current->setLocal(GET_LOCALNO(pc)); + return true; + + case JSOP_INITGLEXICAL: { + MOZ_ASSERT(!script()->hasNonSyntacticScope()); + MDefinition* value = current->pop(); + current->push(constant(ObjectValue(script()->global().lexicalEnvironment()))); + current->push(value); + return jsop_setprop(info().getAtom(pc)->asPropertyName()); + } + + case JSOP_CHECKALIASEDLEXICAL: + return jsop_checkaliasedlexical(EnvironmentCoordinate(pc)); + + case JSOP_INITALIASEDLEXICAL: + return jsop_setaliasedvar(EnvironmentCoordinate(pc)); + + case JSOP_UNINITIALIZED: + pushConstant(MagicValue(JS_UNINITIALIZED_LEXICAL)); + return true; + + case JSOP_POP: { + MDefinition* def = current->pop(); + + // POP opcodes frequently appear where values are killed, e.g. after + // SET* opcodes. Place a resume point afterwards to avoid capturing + // the dead value in later snapshots, except in places where that + // resume point is obviously unnecessary. + if (pc[JSOP_POP_LENGTH] == JSOP_POP) + return true; + if (def->isConstant()) + return true; + return maybeInsertResume(); + } + + case JSOP_POPN: + for (uint32_t i = 0, n = GET_UINT16(pc); i < n; i++) + current->pop(); + return true; + + case JSOP_DUPAT: + current->pushSlot(current->stackDepth() - 1 - GET_UINT24(pc)); + return true; + + case JSOP_NEWINIT: + if (GET_UINT8(pc) == JSProto_Array) + return jsop_newarray(0); + return jsop_newobject(); + + case JSOP_NEWARRAY: + return jsop_newarray(GET_UINT32(pc)); + + case JSOP_NEWARRAY_COPYONWRITE: + return jsop_newarray_copyonwrite(); + + case JSOP_NEWOBJECT: + return jsop_newobject(); + + case JSOP_INITELEM: + case JSOP_INITHIDDENELEM: + return jsop_initelem(); + + case JSOP_INITELEM_ARRAY: + return jsop_initelem_array(); + + case JSOP_INITPROP: + case JSOP_INITLOCKEDPROP: + case JSOP_INITHIDDENPROP: + { + PropertyName* name = info().getAtom(pc)->asPropertyName(); + return jsop_initprop(name); + } + + case JSOP_MUTATEPROTO: + { + return jsop_mutateproto(); + } + + case JSOP_INITPROP_GETTER: + case JSOP_INITHIDDENPROP_GETTER: + case JSOP_INITPROP_SETTER: + case JSOP_INITHIDDENPROP_SETTER: { + PropertyName* name = info().getAtom(pc)->asPropertyName(); + return jsop_initprop_getter_setter(name); + } + + case JSOP_INITELEM_GETTER: + case JSOP_INITHIDDENELEM_GETTER: + case JSOP_INITELEM_SETTER: + case JSOP_INITHIDDENELEM_SETTER: + return jsop_initelem_getter_setter(); + + case JSOP_FUNCALL: + return jsop_funcall(GET_ARGC(pc)); + + case JSOP_FUNAPPLY: + return jsop_funapply(GET_ARGC(pc)); + + case JSOP_CALL: + case JSOP_CALLITER: + case JSOP_NEW: + case JSOP_SUPERCALL: + return jsop_call(GET_ARGC(pc), (JSOp)*pc == JSOP_NEW || (JSOp)*pc == JSOP_SUPERCALL); + + case JSOP_EVAL: + case JSOP_STRICTEVAL: + return jsop_eval(GET_ARGC(pc)); + + case JSOP_INT8: + pushConstant(Int32Value(GET_INT8(pc))); + return true; + + case JSOP_UINT16: + pushConstant(Int32Value(GET_UINT16(pc))); + return true; + + case JSOP_GETGNAME: + { + PropertyName* name = info().getAtom(pc)->asPropertyName(); + if (!script()->hasNonSyntacticScope()) + return jsop_getgname(name); + return jsop_getname(name); + } + + case JSOP_SETGNAME: + case JSOP_STRICTSETGNAME: + { + PropertyName* name = info().getAtom(pc)->asPropertyName(); + JSObject* obj = nullptr; + if (!script()->hasNonSyntacticScope()) + obj = testGlobalLexicalBinding(name); + if (obj) + return setStaticName(obj, name); + return jsop_setprop(name); + } + + case JSOP_GETNAME: + { + PropertyName* name = info().getAtom(pc)->asPropertyName(); + return jsop_getname(name); + } + + case JSOP_GETINTRINSIC: + { + PropertyName* name = info().getAtom(pc)->asPropertyName(); + return jsop_intrinsic(name); + } + + case JSOP_GETIMPORT: + { + PropertyName* name = info().getAtom(pc)->asPropertyName(); + return jsop_getimport(name); + } + + case JSOP_BINDGNAME: + if (!script()->hasNonSyntacticScope()) { + if (JSObject* env = testGlobalLexicalBinding(info().getName(pc))) { + pushConstant(ObjectValue(*env)); + return true; + } + } + // Fall through to JSOP_BINDNAME + MOZ_FALLTHROUGH; + case JSOP_BINDNAME: + return jsop_bindname(info().getName(pc)); + + case JSOP_BINDVAR: + return jsop_bindvar(); + + case JSOP_DUP: + current->pushSlot(current->stackDepth() - 1); + return true; + + case JSOP_DUP2: + return jsop_dup2(); + + case JSOP_SWAP: + current->swapAt(-1); + return true; + + case JSOP_PICK: + current->pick(-GET_INT8(pc)); + return true; + + case JSOP_GETALIASEDVAR: + return jsop_getaliasedvar(EnvironmentCoordinate(pc)); + + case JSOP_SETALIASEDVAR: + return jsop_setaliasedvar(EnvironmentCoordinate(pc)); + + case JSOP_UINT24: + pushConstant(Int32Value(GET_UINT24(pc))); + return true; + + case JSOP_INT32: + pushConstant(Int32Value(GET_INT32(pc))); + return true; + + case JSOP_LOOPHEAD: + // JSOP_LOOPHEAD is handled when processing the loop header. + MOZ_CRASH("JSOP_LOOPHEAD outside loop"); + + case JSOP_GETELEM: + case JSOP_CALLELEM: + if (!jsop_getelem()) + return false; + if (op == JSOP_CALLELEM && !improveThisTypesForCall()) + return false; + return true; + + case JSOP_SETELEM: + case JSOP_STRICTSETELEM: + return jsop_setelem(); + + case JSOP_LENGTH: + return jsop_length(); + + case JSOP_NOT: + return jsop_not(); + + case JSOP_FUNCTIONTHIS: + return jsop_functionthis(); + + case JSOP_GLOBALTHIS: + return jsop_globalthis(); + + case JSOP_CALLEE: { + MDefinition* callee = getCallee(); + current->push(callee); + return true; + } + + case JSOP_GETPROP: + case JSOP_CALLPROP: + { + PropertyName* name = info().getAtom(pc)->asPropertyName(); + if (!jsop_getprop(name)) + return false; + if (op == JSOP_CALLPROP && !improveThisTypesForCall()) + return false; + return true; + } + + case JSOP_SETPROP: + case JSOP_STRICTSETPROP: + case JSOP_SETNAME: + case JSOP_STRICTSETNAME: + { + PropertyName* name = info().getAtom(pc)->asPropertyName(); + return jsop_setprop(name); + } + + case JSOP_DELPROP: + case JSOP_STRICTDELPROP: + { + PropertyName* name = info().getAtom(pc)->asPropertyName(); + return jsop_delprop(name); + } + + case JSOP_DELELEM: + case JSOP_STRICTDELELEM: + return jsop_delelem(); + + case JSOP_REGEXP: + return jsop_regexp(info().getRegExp(pc)); + + case JSOP_CALLSITEOBJ: + pushConstant(ObjectValue(*(info().getObject(pc)))); + return true; + + case JSOP_OBJECT: + return jsop_object(info().getObject(pc)); + + case JSOP_TYPEOF: + case JSOP_TYPEOFEXPR: + return jsop_typeof(); + + case JSOP_TOASYNC: + return jsop_toasync(); + + case JSOP_TOID: + return jsop_toid(); + + case JSOP_LAMBDA: + return jsop_lambda(info().getFunction(pc)); + + case JSOP_LAMBDA_ARROW: + return jsop_lambda_arrow(info().getFunction(pc)); + + case JSOP_ITER: + return jsop_iter(GET_INT8(pc)); + + case JSOP_MOREITER: + return jsop_itermore(); + + case JSOP_ISNOITER: + return jsop_isnoiter(); + + case JSOP_ENDITER: + return jsop_iterend(); + + case JSOP_IN: + return jsop_in(); + + case JSOP_SETRVAL: + MOZ_ASSERT(!script()->noScriptRval()); + current->setSlot(info().returnValueSlot(), current->pop()); + return true; + + case JSOP_INSTANCEOF: + return jsop_instanceof(); + + case JSOP_DEBUGLEAVELEXICALENV: + return true; + + case JSOP_DEBUGGER: + return jsop_debugger(); + + case JSOP_GIMPLICITTHIS: + if (!script()->hasNonSyntacticScope()) { + pushConstant(UndefinedValue()); + return true; + } + + // Just fall through to the unsupported bytecode case. + break; + + case JSOP_NEWTARGET: + return jsop_newtarget(); + + case JSOP_CHECKISOBJ: + return jsop_checkisobj(GET_UINT8(pc)); + + case JSOP_CHECKOBJCOERCIBLE: + return jsop_checkobjcoercible(); + + case JSOP_DEBUGCHECKSELFHOSTED: + { +#ifdef DEBUG + MDebugCheckSelfHosted* check = MDebugCheckSelfHosted::New(alloc(), current->pop()); + current->add(check); + current->push(check); + if (!resumeAfter(check)) + return false; +#endif + return true; + } + + case JSOP_IS_CONSTRUCTING: + pushConstant(MagicValue(JS_IS_CONSTRUCTING)); + return true; + +#ifdef DEBUG + case JSOP_PUSHLEXICALENV: + case JSOP_FRESHENLEXICALENV: + case JSOP_RECREATELEXICALENV: + case JSOP_POPLEXICALENV: + // These opcodes are currently unhandled by Ion, but in principle + // there's no reason they couldn't be. Whenever this happens, OSR + // will have to consider that JSOP_{FRESHEN,RECREATE}LEXICALENV + // mutates the env chain -- right now MBasicBlock::environmentChain() + // caches the env chain. JSOP_{FRESHEN,RECREATE}LEXICALENV must + // update that stale value. +#endif + default: + break; + } + + // Track a simpler message, since the actionable abort message is a + // static string, and the internal opcode name isn't an actionable + // thing anyways. + trackActionableAbort("Unsupported bytecode"); +#ifdef DEBUG + return abort("Unsupported opcode: %s", CodeName[op]); +#else + return abort("Unsupported opcode: %d", op); +#endif +} + +// Given that the current control flow structure has ended forcefully, +// via a return, break, or continue (rather than joining), propagate the +// termination up. For example, a return nested 5 loops deep may terminate +// every outer loop at once, if there are no intervening conditionals: +// +// for (...) { +// for (...) { +// return x; +// } +// } +// +// If |current| is nullptr when this function returns, then there is no more +// control flow to be processed. +IonBuilder::ControlStatus +IonBuilder::processControlEnd() +{ + MOZ_ASSERT(!current); + + if (cfgStack_.empty()) { + // If there is no more control flow to process, then this is the + // last return in the function. + return ControlStatus_Ended; + } + + return processCfgStack(); +} + +// Processes the top of the CFG stack. This is used from two places: +// (1) processControlEnd(), whereby a break, continue, or return may interrupt +// an in-progress CFG structure before reaching its actual termination +// point in the bytecode. +// (2) traverseBytecode(), whereby we reach the last instruction in a CFG +// structure. +IonBuilder::ControlStatus +IonBuilder::processCfgStack() +{ + ControlStatus status = processCfgEntry(cfgStack_.back()); + + // If this terminated a CFG structure, act like processControlEnd() and + // keep propagating upward. + while (status == ControlStatus_Ended) { + popCfgStack(); + if (cfgStack_.empty()) + return status; + status = processCfgEntry(cfgStack_.back()); + } + + // If some join took place, the current structure is finished. + if (status == ControlStatus_Joined) + popCfgStack(); + + return status; +} + +IonBuilder::ControlStatus +IonBuilder::processCfgEntry(CFGState& state) +{ + switch (state.state) { + case CFGState::IF_TRUE: + case CFGState::IF_TRUE_EMPTY_ELSE: + return processIfEnd(state); + + case CFGState::IF_ELSE_TRUE: + return processIfElseTrueEnd(state); + + case CFGState::IF_ELSE_FALSE: + return processIfElseFalseEnd(state); + + case CFGState::DO_WHILE_LOOP_BODY: + return processDoWhileBodyEnd(state); + + case CFGState::DO_WHILE_LOOP_COND: + return processDoWhileCondEnd(state); + + case CFGState::WHILE_LOOP_COND: + return processWhileCondEnd(state); + + case CFGState::WHILE_LOOP_BODY: + return processWhileBodyEnd(state); + + case CFGState::FOR_LOOP_COND: + return processForCondEnd(state); + + case CFGState::FOR_LOOP_BODY: + return processForBodyEnd(state); + + case CFGState::FOR_LOOP_UPDATE: + return processForUpdateEnd(state); + + case CFGState::TABLE_SWITCH: + return processNextTableSwitchCase(state); + + case CFGState::COND_SWITCH_CASE: + return processCondSwitchCase(state); + + case CFGState::COND_SWITCH_BODY: + return processCondSwitchBody(state); + + case CFGState::AND_OR: + return processAndOrEnd(state); + + case CFGState::LABEL: + return processLabelEnd(state); + + case CFGState::TRY: + return processTryEnd(state); + + default: + MOZ_CRASH("unknown cfgstate"); + } +} + +IonBuilder::ControlStatus +IonBuilder::processIfEnd(CFGState& state) +{ + bool thenBranchTerminated = !current; + if (!thenBranchTerminated) { + // Here, the false block is the join point. Create an edge from the + // current block to the false block. Note that a RETURN opcode + // could have already ended the block. + current->end(MGoto::New(alloc(), state.branch.ifFalse)); + + if (!state.branch.ifFalse->addPredecessor(alloc(), current)) + return ControlStatus_Error; + } + + if (!setCurrentAndSpecializePhis(state.branch.ifFalse)) + return ControlStatus_Error; + graph().moveBlockToEnd(current); + pc = current->pc(); + + if (thenBranchTerminated) { + // If we can't reach here via the then-branch, we can filter the types + // after the if-statement based on the if-condition. + MTest* test = state.branch.test; + if (!improveTypesAtTest(test->getOperand(0), test->ifTrue() == current, test)) + return ControlStatus_Error; + } + + return ControlStatus_Joined; +} + +IonBuilder::ControlStatus +IonBuilder::processIfElseTrueEnd(CFGState& state) +{ + // We've reached the end of the true branch of an if-else. Don't + // create an edge yet, just transition to parsing the false branch. + state.state = CFGState::IF_ELSE_FALSE; + state.branch.ifTrue = current; + state.stopAt = state.branch.falseEnd; + pc = state.branch.ifFalse->pc(); + if (!setCurrentAndSpecializePhis(state.branch.ifFalse)) + return ControlStatus_Error; + graph().moveBlockToEnd(current); + + MTest* test = state.branch.test; + if (!improveTypesAtTest(test->getOperand(0), test->ifTrue() == current, test)) + return ControlStatus_Error; + + return ControlStatus_Jumped; +} + +IonBuilder::ControlStatus +IonBuilder::processIfElseFalseEnd(CFGState& state) +{ + // Update the state to have the latest block from the false path. + state.branch.ifFalse = current; + + // To create the join node, we need an incoming edge that has not been + // terminated yet. + MBasicBlock* pred = state.branch.ifTrue + ? state.branch.ifTrue + : state.branch.ifFalse; + MBasicBlock* other = (pred == state.branch.ifTrue) ? state.branch.ifFalse : state.branch.ifTrue; + + if (!pred) + return ControlStatus_Ended; + + // Create a new block to represent the join. + MBasicBlock* join = newBlock(pred, state.branch.falseEnd); + if (!join) + return ControlStatus_Error; + + // Create edges from the true and false blocks as needed. + pred->end(MGoto::New(alloc(), join)); + + if (other) { + other->end(MGoto::New(alloc(), join)); + if (!join->addPredecessor(alloc(), other)) + return ControlStatus_Error; + } + + // Ignore unreachable remainder of false block if existent. + if (!setCurrentAndSpecializePhis(join)) + return ControlStatus_Error; + pc = current->pc(); + return ControlStatus_Joined; +} + +IonBuilder::ControlStatus +IonBuilder::processBrokenLoop(CFGState& state) +{ + MOZ_ASSERT(!current); + + MOZ_ASSERT(loopDepth_); + loopDepth_--; + + // A broken loop is not a real loop (it has no header or backedge), so + // reset the loop depth. + for (MBasicBlockIterator i(graph().begin(state.loop.entry)); i != graph().end(); i++) { + if (i->loopDepth() > loopDepth_) + i->setLoopDepth(i->loopDepth() - 1); + } + + // If the loop started with a condition (while/for) then even if the + // structure never actually loops, the condition itself can still fail and + // thus we must resume at the successor, if one exists. + if (!setCurrentAndSpecializePhis(state.loop.successor)) + return ControlStatus_Error; + if (current) { + MOZ_ASSERT(current->loopDepth() == loopDepth_); + graph().moveBlockToEnd(current); + } + + // Join the breaks together and continue parsing. + if (state.loop.breaks) { + MBasicBlock* block = createBreakCatchBlock(state.loop.breaks, state.loop.exitpc); + if (!block) + return ControlStatus_Error; + + if (current) { + current->end(MGoto::New(alloc(), block)); + if (!block->addPredecessor(alloc(), current)) + return ControlStatus_Error; + } + + if (!setCurrentAndSpecializePhis(block)) + return ControlStatus_Error; + } + + // If the loop is not gated on a condition, and has only returns, we'll + // reach this case. For example: + // do { ... return; } while (); + if (!current) + return ControlStatus_Ended; + + // Otherwise, the loop is gated on a condition and/or has breaks so keep + // parsing at the successor. + pc = current->pc(); + return ControlStatus_Joined; +} + +IonBuilder::ControlStatus +IonBuilder::finishLoop(CFGState& state, MBasicBlock* successor) +{ + MOZ_ASSERT(current); + + MOZ_ASSERT(loopDepth_); + loopDepth_--; + MOZ_ASSERT_IF(successor, successor->loopDepth() == loopDepth_); + + // Compute phis in the loop header and propagate them throughout the loop, + // including the successor. + AbortReason r = state.loop.entry->setBackedge(alloc(), current); + if (r == AbortReason_Alloc) + return ControlStatus_Error; + if (r == AbortReason_Disable) { + // If there are types for variables on the backedge that were not + // present at the original loop header, then uses of the variables' + // phis may have generated incorrect nodes. The new types have been + // incorporated into the header phis, so remove all blocks for the + // loop body and restart with the new types. + return restartLoop(state); + } + + if (successor) { + graph().moveBlockToEnd(successor); + successor->inheritPhis(state.loop.entry); + } + + if (state.loop.breaks) { + // Propagate phis placed in the header to individual break exit points. + DeferredEdge* edge = state.loop.breaks; + while (edge) { + edge->block->inheritPhis(state.loop.entry); + edge = edge->next; + } + + // Create a catch block to join all break exits. + MBasicBlock* block = createBreakCatchBlock(state.loop.breaks, state.loop.exitpc); + if (!block) + return ControlStatus_Error; + + if (successor) { + // Finally, create an unconditional edge from the successor to the + // catch block. + successor->end(MGoto::New(alloc(), block)); + if (!block->addPredecessor(alloc(), successor)) + return ControlStatus_Error; + } + successor = block; + } + + if (!setCurrentAndSpecializePhis(successor)) + return ControlStatus_Error; + + // An infinite loop (for (;;) { }) will not have a successor. + if (!current) + return ControlStatus_Ended; + + pc = current->pc(); + return ControlStatus_Joined; +} + +IonBuilder::ControlStatus +IonBuilder::restartLoop(const CFGState& state) +{ + AutoTraceLog logCompile(traceLogger(), TraceLogger_IonBuilderRestartLoop); + + spew("New types at loop header, restarting loop body"); + + if (JitOptions.limitScriptSize) { + if (++numLoopRestarts_ >= MAX_LOOP_RESTARTS) + return ControlStatus_Abort; + } + + MBasicBlock* header = state.loop.entry; + + // Discard unreferenced & pre-allocated resume points. + replaceMaybeFallbackFunctionGetter(nullptr); + + // Remove all blocks in the loop body other than the header, which has phis + // of the appropriate type and incoming edges to preserve. + graph().removeBlocksAfter(header); + + // Remove all instructions from the header itself, and all resume points + // except the entry resume point. + header->discardAllInstructions(); + header->discardAllResumePoints(/* discardEntry = */ false); + header->setStackDepth(header->getPredecessor(0)->stackDepth()); + + popCfgStack(); + + loopDepth_++; + + // Keep a local copy for these pointers since state will be overwritten in + // pushLoop since state is a reference to cfgStack_.back() + jsbytecode* condpc = state.loop.condpc; + jsbytecode* updatepc = state.loop.updatepc; + jsbytecode* updateEnd = state.loop.updateEnd; + + if (!pushLoop(state.loop.initialState, state.loop.initialStopAt, header, state.loop.osr, + state.loop.loopHead, state.loop.initialPc, + state.loop.bodyStart, state.loop.bodyEnd, + state.loop.exitpc, state.loop.continuepc)) + { + return ControlStatus_Error; + } + + CFGState& nstate = cfgStack_.back(); + + nstate.loop.condpc = condpc; + nstate.loop.updatepc = updatepc; + nstate.loop.updateEnd = updateEnd; + + // Don't specializePhis(), as the header has been visited before and the + // phis have already had their type set. + setCurrent(header); + + if (!jsop_loophead(nstate.loop.loopHead)) + return ControlStatus_Error; + + pc = nstate.loop.initialPc; + return ControlStatus_Jumped; +} + +IonBuilder::ControlStatus +IonBuilder::processDoWhileBodyEnd(CFGState& state) +{ + if (!processDeferredContinues(state)) + return ControlStatus_Error; + + // No current means control flow cannot reach the condition, so this will + // never loop. + if (!current) + return processBrokenLoop(state); + + MBasicBlock* header = newBlock(current, state.loop.updatepc); + if (!header) + return ControlStatus_Error; + current->end(MGoto::New(alloc(), header)); + + state.state = CFGState::DO_WHILE_LOOP_COND; + state.stopAt = state.loop.updateEnd; + pc = state.loop.updatepc; + if (!setCurrentAndSpecializePhis(header)) + return ControlStatus_Error; + return ControlStatus_Jumped; +} + +IonBuilder::ControlStatus +IonBuilder::processDoWhileCondEnd(CFGState& state) +{ + MOZ_ASSERT(JSOp(*pc) == JSOP_IFNE); + + // We're guaranteed a |current|, it's impossible to break or return from + // inside the conditional expression. + MOZ_ASSERT(current); + + // Pop the last value, and create the successor block. + MDefinition* vins = current->pop(); + MBasicBlock* successor = newBlock(current, GetNextPc(pc), loopDepth_ - 1); + if (!successor) + return ControlStatus_Error; + + // Test for do {} while(false) and don't create a loop in that case. + if (MConstant* vinsConst = vins->maybeConstantValue()) { + bool b; + if (vinsConst->valueToBoolean(&b) && !b) { + current->end(MGoto::New(alloc(), successor)); + current = nullptr; + + state.loop.successor = successor; + return processBrokenLoop(state); + } + } + + // Create the test instruction and end the current block. + MTest* test = newTest(vins, state.loop.entry, successor); + current->end(test); + return finishLoop(state, successor); +} + +IonBuilder::ControlStatus +IonBuilder::processWhileCondEnd(CFGState& state) +{ + MOZ_ASSERT(JSOp(*pc) == JSOP_IFNE || JSOp(*pc) == JSOP_IFEQ); + + // Balance the stack past the IFNE. + MDefinition* ins = current->pop(); + + // Create the body and successor blocks. + MBasicBlock* body = newBlock(current, state.loop.bodyStart); + state.loop.successor = newBlock(current, state.loop.exitpc, loopDepth_ - 1); + if (!body || !state.loop.successor) + return ControlStatus_Error; + + MTest* test; + if (JSOp(*pc) == JSOP_IFNE) + test = newTest(ins, body, state.loop.successor); + else + test = newTest(ins, state.loop.successor, body); + current->end(test); + + state.state = CFGState::WHILE_LOOP_BODY; + state.stopAt = state.loop.bodyEnd; + pc = state.loop.bodyStart; + if (!setCurrentAndSpecializePhis(body)) + return ControlStatus_Error; + + // Filter the types in the loop body. + if (!improveTypesAtTest(test->getOperand(0), test->ifTrue() == current, test)) + return ControlStatus_Error; + + // If this is a for-in loop, unbox the current value as string if possible. + if (ins->isIsNoIter()) { + MIteratorMore* iterMore = ins->toIsNoIter()->input()->toIteratorMore(); + jsbytecode* iterMorePc = iterMore->resumePoint()->pc(); + MOZ_ASSERT(*iterMorePc == JSOP_MOREITER); + + if (!nonStringIteration_ && !inspector->hasSeenNonStringIterMore(iterMorePc)) { + MDefinition* val = current->peek(-1); + MOZ_ASSERT(val == iterMore); + MInstruction* ins = MUnbox::New(alloc(), val, MIRType::String, MUnbox::Fallible, + Bailout_NonStringInputInvalidate); + current->add(ins); + current->rewriteAtDepth(-1, ins); + } + } + + return ControlStatus_Jumped; +} + +IonBuilder::ControlStatus +IonBuilder::processWhileBodyEnd(CFGState& state) +{ + if (!processDeferredContinues(state)) + return ControlStatus_Error; + + if (!current) + return processBrokenLoop(state); + + current->end(MGoto::New(alloc(), state.loop.entry)); + return finishLoop(state, state.loop.successor); +} + +IonBuilder::ControlStatus +IonBuilder::processForCondEnd(CFGState& state) +{ + MOZ_ASSERT(JSOp(*pc) == JSOP_IFNE); + + // Balance the stack past the IFNE. + MDefinition* ins = current->pop(); + + // Create the body and successor blocks. + MBasicBlock* body = newBlock(current, state.loop.bodyStart); + state.loop.successor = newBlock(current, state.loop.exitpc, loopDepth_ - 1); + if (!body || !state.loop.successor) + return ControlStatus_Error; + + MTest* test = newTest(ins, body, state.loop.successor); + current->end(test); + + state.state = CFGState::FOR_LOOP_BODY; + state.stopAt = state.loop.bodyEnd; + pc = state.loop.bodyStart; + if (!setCurrentAndSpecializePhis(body)) + return ControlStatus_Error; + return ControlStatus_Jumped; +} + +IonBuilder::ControlStatus +IonBuilder::processForBodyEnd(CFGState& state) +{ + if (!processDeferredContinues(state)) + return ControlStatus_Error; + + // If there is no updatepc, just go right to processing what would be the + // end of the update clause. Otherwise, |current| might be nullptr; if this is + // the case, the udpate is unreachable anyway. + if (!state.loop.updatepc || !current) + return processForUpdateEnd(state); + + pc = state.loop.updatepc; + + state.state = CFGState::FOR_LOOP_UPDATE; + state.stopAt = state.loop.updateEnd; + return ControlStatus_Jumped; +} + +IonBuilder::ControlStatus +IonBuilder::processForUpdateEnd(CFGState& state) +{ + // If there is no current, we couldn't reach the loop edge and there was no + // update clause. + if (!current) + return processBrokenLoop(state); + + current->end(MGoto::New(alloc(), state.loop.entry)); + return finishLoop(state, state.loop.successor); +} + +IonBuilder::DeferredEdge* +IonBuilder::filterDeadDeferredEdges(DeferredEdge* edge) +{ + DeferredEdge* head = edge; + DeferredEdge* prev = nullptr; + + while (edge) { + if (edge->block->isDead()) { + if (prev) + prev->next = edge->next; + else + head = edge->next; + } else { + prev = edge; + } + edge = edge->next; + } + + // There must be at least one deferred edge from a block that was not + // deleted; blocks are deleted when restarting processing of a loop, and + // the final version of the loop body will have edges from live blocks. + MOZ_ASSERT(head); + + return head; +} + +bool +IonBuilder::processDeferredContinues(CFGState& state) +{ + // If there are any continues for this loop, and there is an update block, + // then we need to create a new basic block to house the update. + if (state.loop.continues) { + DeferredEdge* edge = filterDeadDeferredEdges(state.loop.continues); + + MBasicBlock* update = newBlock(edge->block, loops_.back().continuepc); + if (!update) + return false; + + if (current) { + current->end(MGoto::New(alloc(), update)); + if (!update->addPredecessor(alloc(), current)) + return false; + } + + // No need to use addPredecessor for first edge, + // because it is already predecessor. + edge->block->end(MGoto::New(alloc(), update)); + edge = edge->next; + + // Remaining edges + while (edge) { + edge->block->end(MGoto::New(alloc(), update)); + if (!update->addPredecessor(alloc(), edge->block)) + return false; + edge = edge->next; + } + state.loop.continues = nullptr; + + if (!setCurrentAndSpecializePhis(update)) + return ControlStatus_Error; + } + + return true; +} + +MBasicBlock* +IonBuilder::createBreakCatchBlock(DeferredEdge* edge, jsbytecode* pc) +{ + edge = filterDeadDeferredEdges(edge); + + // Create block, using the first break statement as predecessor + MBasicBlock* successor = newBlock(edge->block, pc); + if (!successor) + return nullptr; + + // No need to use addPredecessor for first edge, + // because it is already predecessor. + edge->block->end(MGoto::New(alloc(), successor)); + edge = edge->next; + + // Finish up remaining breaks. + while (edge) { + MGoto* brk = MGoto::New(alloc().fallible(), successor); + if (!brk) + return nullptr; + edge->block->end(brk); + if (!successor->addPredecessor(alloc(), edge->block)) + return nullptr; + edge = edge->next; + } + + return successor; +} + +IonBuilder::ControlStatus +IonBuilder::processNextTableSwitchCase(CFGState& state) +{ + MOZ_ASSERT(state.state == CFGState::TABLE_SWITCH); + + state.tableswitch.currentBlock++; + + // Test if there are still unprocessed successors (cases/default) + if (state.tableswitch.currentBlock >= state.tableswitch.ins->numBlocks()) + return processSwitchEnd(state.tableswitch.breaks, state.tableswitch.exitpc); + + // Get the next successor + MBasicBlock* successor = state.tableswitch.ins->getBlock(state.tableswitch.currentBlock); + + // Add current block as predecessor if available. + // This means the previous case didn't have a break statement. + // So flow will continue in this block. + if (current) { + current->end(MGoto::New(alloc(), successor)); + if (!successor->addPredecessor(alloc(), current)) + return ControlStatus_Error; + } else { + // If this is an actual case statement, optimize by replacing the + // input to the switch case with the actual number of the case. + // This constant has been emitted when creating the case blocks. + if (state.tableswitch.ins->getDefault() != successor) { + MConstant* constant = successor->begin()->toConstant(); + for (uint32_t j = 0; j < successor->stackDepth(); j++) { + MDefinition* ins = successor->getSlot(j); + if (ins != state.tableswitch.ins->getOperand(0)) + continue; + + constant->setDependency(state.tableswitch.ins); + successor->setSlot(j, constant); + } + } + } + + // Insert successor after the current block, to maintain RPO. + graph().moveBlockToEnd(successor); + + // If this is the last successor the block should stop at the end of the tableswitch + // Else it should stop at the start of the next successor + if (state.tableswitch.currentBlock+1 < state.tableswitch.ins->numBlocks()) + state.stopAt = state.tableswitch.ins->getBlock(state.tableswitch.currentBlock+1)->pc(); + else + state.stopAt = state.tableswitch.exitpc; + + if (!setCurrentAndSpecializePhis(successor)) + return ControlStatus_Error; + pc = current->pc(); + return ControlStatus_Jumped; +} + +IonBuilder::ControlStatus +IonBuilder::processAndOrEnd(CFGState& state) +{ + MOZ_ASSERT(current); + MBasicBlock* lhs = state.branch.ifFalse; + + // Create a new block to represent the join. + MBasicBlock* join = newBlock(current, state.stopAt); + if (!join) + return ControlStatus_Error; + + // End the rhs. + current->end(MGoto::New(alloc(), join)); + + // End the lhs. + lhs->end(MGoto::New(alloc(), join)); + if (!join->addPredecessor(alloc(), state.branch.ifFalse)) + return ControlStatus_Error; + + // Set the join path as current path. + if (!setCurrentAndSpecializePhis(join)) + return ControlStatus_Error; + pc = current->pc(); + return ControlStatus_Joined; +} + +IonBuilder::ControlStatus +IonBuilder::processLabelEnd(CFGState& state) +{ + MOZ_ASSERT(state.state == CFGState::LABEL); + + // If there are no breaks and no current, controlflow is terminated. + if (!state.label.breaks && !current) + return ControlStatus_Ended; + + // If there are no breaks to this label, there's nothing to do. + if (!state.label.breaks) + return ControlStatus_Joined; + + MBasicBlock* successor = createBreakCatchBlock(state.label.breaks, state.stopAt); + if (!successor) + return ControlStatus_Error; + + if (current) { + current->end(MGoto::New(alloc(), successor)); + if (!successor->addPredecessor(alloc(), current)) + return ControlStatus_Error; + } + + pc = state.stopAt; + if (!setCurrentAndSpecializePhis(successor)) + return ControlStatus_Error; + return ControlStatus_Joined; +} + +IonBuilder::ControlStatus +IonBuilder::processTryEnd(CFGState& state) +{ + MOZ_ASSERT(state.state == CFGState::TRY); + + if (!state.try_.successor) { + MOZ_ASSERT(!current); + return ControlStatus_Ended; + } + + if (current) { + current->end(MGoto::New(alloc(), state.try_.successor)); + + if (!state.try_.successor->addPredecessor(alloc(), current)) + return ControlStatus_Error; + } + + // Start parsing the code after this try-catch statement. + if (!setCurrentAndSpecializePhis(state.try_.successor)) + return ControlStatus_Error; + graph().moveBlockToEnd(current); + pc = current->pc(); + return ControlStatus_Joined; +} + +IonBuilder::ControlStatus +IonBuilder::processBreak(JSOp op, jssrcnote* sn) +{ + MOZ_ASSERT(op == JSOP_GOTO); + + MOZ_ASSERT(SN_TYPE(sn) == SRC_BREAK || + SN_TYPE(sn) == SRC_BREAK2LABEL); + + // Find the break target. + jsbytecode* target = pc + GetJumpOffset(pc); + DebugOnly<bool> found = false; + + if (SN_TYPE(sn) == SRC_BREAK2LABEL) { + for (size_t i = labels_.length() - 1; i < labels_.length(); i--) { + CFGState& cfg = cfgStack_[labels_[i].cfgEntry]; + MOZ_ASSERT(cfg.state == CFGState::LABEL); + if (cfg.stopAt == target) { + cfg.label.breaks = new(alloc()) DeferredEdge(current, cfg.label.breaks); + found = true; + break; + } + } + } else { + for (size_t i = loops_.length() - 1; i < loops_.length(); i--) { + CFGState& cfg = cfgStack_[loops_[i].cfgEntry]; + MOZ_ASSERT(cfg.isLoop()); + if (cfg.loop.exitpc == target) { + cfg.loop.breaks = new(alloc()) DeferredEdge(current, cfg.loop.breaks); + found = true; + break; + } + } + } + + MOZ_ASSERT(found); + + setCurrent(nullptr); + pc += CodeSpec[op].length; + return processControlEnd(); +} + +static inline jsbytecode* +EffectiveContinue(jsbytecode* pc) +{ + if (JSOp(*pc) == JSOP_GOTO) + return pc + GetJumpOffset(pc); + return pc; +} + +IonBuilder::ControlStatus +IonBuilder::processContinue(JSOp op) +{ + MOZ_ASSERT(op == JSOP_GOTO); + + // Find the target loop. + CFGState* found = nullptr; + jsbytecode* target = pc + GetJumpOffset(pc); + for (size_t i = loops_.length() - 1; i < loops_.length(); i--) { + // +1 to skip JSOP_JUMPTARGET. + if (loops_[i].continuepc == target + 1 || + EffectiveContinue(loops_[i].continuepc) == target) + { + found = &cfgStack_[loops_[i].cfgEntry]; + break; + } + } + + // There must always be a valid target loop structure. If not, there's + // probably an off-by-something error in which pc we track. + MOZ_ASSERT(found); + CFGState& state = *found; + + state.loop.continues = new(alloc()) DeferredEdge(current, state.loop.continues); + + setCurrent(nullptr); + pc += CodeSpec[op].length; + return processControlEnd(); +} + +IonBuilder::ControlStatus +IonBuilder::processSwitchBreak(JSOp op) +{ + MOZ_ASSERT(op == JSOP_GOTO); + + // Find the target switch. + CFGState* found = nullptr; + jsbytecode* target = pc + GetJumpOffset(pc); + for (size_t i = switches_.length() - 1; i < switches_.length(); i--) { + if (switches_[i].continuepc == target) { + found = &cfgStack_[switches_[i].cfgEntry]; + break; + } + } + + // There must always be a valid target loop structure. If not, there's + // probably an off-by-something error in which pc we track. + MOZ_ASSERT(found); + CFGState& state = *found; + + DeferredEdge** breaks = nullptr; + switch (state.state) { + case CFGState::TABLE_SWITCH: + breaks = &state.tableswitch.breaks; + break; + case CFGState::COND_SWITCH_BODY: + breaks = &state.condswitch.breaks; + break; + default: + MOZ_CRASH("Unexpected switch state."); + } + + *breaks = new(alloc()) DeferredEdge(current, *breaks); + + setCurrent(nullptr); + pc += CodeSpec[op].length; + return processControlEnd(); +} + +IonBuilder::ControlStatus +IonBuilder::processSwitchEnd(DeferredEdge* breaks, jsbytecode* exitpc) +{ + // No break statements, no current. + // This means that control flow is cut-off from this point + // (e.g. all cases have return statements). + if (!breaks && !current) + return ControlStatus_Ended; + + // Create successor block. + // If there are breaks, create block with breaks as predecessor + // Else create a block with current as predecessor + MBasicBlock* successor = nullptr; + if (breaks) + successor = createBreakCatchBlock(breaks, exitpc); + else + successor = newBlock(current, exitpc); + + if (!successor) + return ControlStatus_Error; + + // If there is current, the current block flows into this one. + // So current is also a predecessor to this block + if (current) { + current->end(MGoto::New(alloc(), successor)); + if (breaks) { + if (!successor->addPredecessor(alloc(), current)) + return ControlStatus_Error; + } + } + + pc = exitpc; + if (!setCurrentAndSpecializePhis(successor)) + return ControlStatus_Error; + return ControlStatus_Joined; +} + +IonBuilder::ControlStatus +IonBuilder::maybeLoop(JSOp op, jssrcnote* sn) +{ + // This function looks at the opcode and source note and tries to + // determine the structure of the loop. For some opcodes, like + // POP/NOP which are not explicitly control flow, this source note is + // optional. For opcodes with control flow, like GOTO, an unrecognized + // or not-present source note is a compilation failure. + switch (op) { + case JSOP_POP: + // for (init; ; update?) ... + if (sn && SN_TYPE(sn) == SRC_FOR) { + current->pop(); + return forLoop(op, sn); + } + break; + + case JSOP_NOP: + if (sn) { + // do { } while (cond) + if (SN_TYPE(sn) == SRC_WHILE) + return doWhileLoop(op, sn); + // Build a mapping such that given a basic block, whose successor + // has a phi + + // for (; ; update?) + if (SN_TYPE(sn) == SRC_FOR) + return forLoop(op, sn); + } + break; + + default: + MOZ_CRASH("unexpected opcode"); + } + + return ControlStatus_None; +} + +void +IonBuilder::assertValidLoopHeadOp(jsbytecode* pc) +{ +#ifdef DEBUG + MOZ_ASSERT(JSOp(*pc) == JSOP_LOOPHEAD); + + // Make sure this is the next opcode after the loop header, + // unless the for loop is unconditional. + CFGState& state = cfgStack_.back(); + MOZ_ASSERT_IF((JSOp)*(state.loop.entry->pc()) == JSOP_GOTO, + GetNextPc(state.loop.entry->pc()) == pc); + + // do-while loops have a source note. + jssrcnote* sn = info().getNote(gsn, pc); + if (sn) { + jsbytecode* ifne = pc + GetSrcNoteOffset(sn, 0); + + jsbytecode* expected_ifne; + switch (state.state) { + case CFGState::DO_WHILE_LOOP_BODY: + expected_ifne = state.loop.updateEnd; + break; + + default: + MOZ_CRASH("JSOP_LOOPHEAD unexpected source note"); + } + + // Make sure this loop goes to the same ifne as the loop header's + // source notes or GOTO. + MOZ_ASSERT(ifne == expected_ifne); + } else { + MOZ_ASSERT(state.state != CFGState::DO_WHILE_LOOP_BODY); + } +#endif +} + +IonBuilder::ControlStatus +IonBuilder::doWhileLoop(JSOp op, jssrcnote* sn) +{ + // do { } while() loops have the following structure: + // NOP ; SRC_WHILE (offset to COND) + // LOOPHEAD ; SRC_WHILE (offset to IFNE) + // LOOPENTRY + // ... ; body + // ... + // COND ; start of condition + // ... + // IFNE -> ; goes to LOOPHEAD + int condition_offset = GetSrcNoteOffset(sn, 0); + jsbytecode* conditionpc = pc + condition_offset; + + jssrcnote* sn2 = info().getNote(gsn, pc+1); + int offset = GetSrcNoteOffset(sn2, 0); + jsbytecode* ifne = pc + offset + 1; + MOZ_ASSERT(ifne > pc); + + // Verify that the IFNE goes back to a loophead op. + jsbytecode* loopHead = GetNextPc(pc); + MOZ_ASSERT(JSOp(*loopHead) == JSOP_LOOPHEAD); + MOZ_ASSERT(loopHead == ifne + GetJumpOffset(ifne)); + + jsbytecode* loopEntry = GetNextPc(loopHead); + bool canOsr = LoopEntryCanIonOsr(loopEntry); + bool osr = info().hasOsrAt(loopEntry); + + if (osr) { + MBasicBlock* preheader = newOsrPreheader(current, loopEntry, pc); + if (!preheader) + return ControlStatus_Error; + current->end(MGoto::New(alloc(), preheader)); + if (!setCurrentAndSpecializePhis(preheader)) + return ControlStatus_Error; + } + + unsigned stackPhiCount = 0; + MBasicBlock* header = newPendingLoopHeader(current, loopEntry, osr, canOsr, stackPhiCount); + if (!header) + return ControlStatus_Error; + current->end(MGoto::New(alloc(), header)); + + jsbytecode* loophead = GetNextPc(pc); + jsbytecode* bodyStart = GetNextPc(loophead); + jsbytecode* bodyEnd = conditionpc; + jsbytecode* exitpc = GetNextPc(ifne); + if (!analyzeNewLoopTypes(header, bodyStart, exitpc)) + return ControlStatus_Error; + if (!pushLoop(CFGState::DO_WHILE_LOOP_BODY, conditionpc, header, osr, + loopHead, bodyStart, bodyStart, bodyEnd, exitpc, conditionpc)) + { + return ControlStatus_Error; + } + + CFGState& state = cfgStack_.back(); + state.loop.updatepc = conditionpc; + state.loop.updateEnd = ifne; + + if (!setCurrentAndSpecializePhis(header)) + return ControlStatus_Error; + if (!jsop_loophead(loophead)) + return ControlStatus_Error; + + pc = bodyStart; + return ControlStatus_Jumped; +} + +IonBuilder::ControlStatus +IonBuilder::whileOrForInLoop(jssrcnote* sn) +{ + // while (cond) { } loops have the following structure: + // GOTO cond ; SRC_WHILE (offset to IFNE) + // LOOPHEAD + // ... + // cond: + // LOOPENTRY + // ... + // IFNE ; goes to LOOPHEAD + // for (x in y) { } loops are similar; the cond will be a MOREITER. + MOZ_ASSERT(SN_TYPE(sn) == SRC_FOR_OF || SN_TYPE(sn) == SRC_FOR_IN || SN_TYPE(sn) == SRC_WHILE); + int ifneOffset = GetSrcNoteOffset(sn, 0); + jsbytecode* ifne = pc + ifneOffset; + MOZ_ASSERT(ifne > pc); + + // Verify that the IFNE goes back to a loophead op. + MOZ_ASSERT(JSOp(*GetNextPc(pc)) == JSOP_LOOPHEAD); + MOZ_ASSERT(GetNextPc(pc) == ifne + GetJumpOffset(ifne)); + + jsbytecode* loopEntry = pc + GetJumpOffset(pc); + bool canOsr = LoopEntryCanIonOsr(loopEntry); + bool osr = info().hasOsrAt(loopEntry); + + if (osr) { + MBasicBlock* preheader = newOsrPreheader(current, loopEntry, pc); + if (!preheader) + return ControlStatus_Error; + current->end(MGoto::New(alloc(), preheader)); + if (!setCurrentAndSpecializePhis(preheader)) + return ControlStatus_Error; + } + + unsigned stackPhiCount; + if (SN_TYPE(sn) == SRC_FOR_OF) + stackPhiCount = 2; + else if (SN_TYPE(sn) == SRC_FOR_IN) + stackPhiCount = 1; + else + stackPhiCount = 0; + + MBasicBlock* header = newPendingLoopHeader(current, loopEntry, osr, canOsr, stackPhiCount); + if (!header) + return ControlStatus_Error; + current->end(MGoto::New(alloc(), header)); + + // Skip past the JSOP_LOOPHEAD for the body start. + jsbytecode* loopHead = GetNextPc(pc); + jsbytecode* bodyStart = GetNextPc(loopHead); + jsbytecode* bodyEnd = pc + GetJumpOffset(pc); + jsbytecode* exitpc = GetNextPc(ifne); + jsbytecode* continuepc = pc; + if (!analyzeNewLoopTypes(header, bodyStart, exitpc)) + return ControlStatus_Error; + if (!pushLoop(CFGState::WHILE_LOOP_COND, ifne, header, osr, + loopHead, bodyEnd, bodyStart, bodyEnd, exitpc, continuepc)) + { + return ControlStatus_Error; + } + + // Parse the condition first. + if (!setCurrentAndSpecializePhis(header)) + return ControlStatus_Error; + if (!jsop_loophead(loopHead)) + return ControlStatus_Error; + + pc = bodyEnd; + return ControlStatus_Jumped; +} + +IonBuilder::ControlStatus +IonBuilder::forLoop(JSOp op, jssrcnote* sn) +{ + // Skip the NOP. + MOZ_ASSERT(op == JSOP_NOP); + pc = GetNextPc(pc); + + jsbytecode* condpc = pc + GetSrcNoteOffset(sn, 0); + jsbytecode* updatepc = pc + GetSrcNoteOffset(sn, 1); + jsbytecode* ifne = pc + GetSrcNoteOffset(sn, 2); + jsbytecode* exitpc = GetNextPc(ifne); + + // for loops have the following structures: + // + // NOP or POP + // [GOTO cond | NOP] + // LOOPHEAD + // body: + // ; [body] + // [increment:] + // [{FRESHEN,RECREATE}LEXICALENV, if needed by a lexical env] + // ; [increment] + // [cond:] + // LOOPENTRY + // GOTO body + // + // If there is a condition (condpc != ifne), this acts similar to a while + // loop otherwise, it acts like a do-while loop. + // + // Note that currently Ion doesn't compile pushlexicalenv/poplexicalenv, + // necessary prerequisites to {freshen,recreate}lexicalenv. So the code + // below doesn't and needn't consider either op's implications. + jsbytecode* bodyStart = pc; + jsbytecode* bodyEnd = updatepc; + jsbytecode* loopEntry = condpc; + if (condpc != ifne) { + MOZ_ASSERT(JSOp(*bodyStart) == JSOP_GOTO); + MOZ_ASSERT(bodyStart + GetJumpOffset(bodyStart) == condpc); + bodyStart = GetNextPc(bodyStart); + } else { + // No loop condition, such as for(j = 0; ; j++) + if (op != JSOP_NOP) { + // If the loop starts with POP, we have to skip a NOP. + MOZ_ASSERT(JSOp(*bodyStart) == JSOP_NOP); + bodyStart = GetNextPc(bodyStart); + } + loopEntry = GetNextPc(bodyStart); + } + jsbytecode* loopHead = bodyStart; + MOZ_ASSERT(JSOp(*bodyStart) == JSOP_LOOPHEAD); + MOZ_ASSERT(ifne + GetJumpOffset(ifne) == bodyStart); + bodyStart = GetNextPc(bodyStart); + + bool osr = info().hasOsrAt(loopEntry); + bool canOsr = LoopEntryCanIonOsr(loopEntry); + + if (osr) { + MBasicBlock* preheader = newOsrPreheader(current, loopEntry, pc); + if (!preheader) + return ControlStatus_Error; + current->end(MGoto::New(alloc(), preheader)); + if (!setCurrentAndSpecializePhis(preheader)) + return ControlStatus_Error; + } + + unsigned stackPhiCount = 0; + MBasicBlock* header = newPendingLoopHeader(current, loopEntry, osr, canOsr, stackPhiCount); + if (!header) + return ControlStatus_Error; + current->end(MGoto::New(alloc(), header)); + + // If there is no condition, we immediately parse the body. Otherwise, we + // parse the condition. + jsbytecode* stopAt; + CFGState::State initial; + if (condpc != ifne) { + pc = condpc; + stopAt = ifne; + initial = CFGState::FOR_LOOP_COND; + } else { + pc = bodyStart; + stopAt = bodyEnd; + initial = CFGState::FOR_LOOP_BODY; + } + + if (!analyzeNewLoopTypes(header, bodyStart, exitpc)) + return ControlStatus_Error; + if (!pushLoop(initial, stopAt, header, osr, + loopHead, pc, bodyStart, bodyEnd, exitpc, updatepc)) + { + return ControlStatus_Error; + } + + CFGState& state = cfgStack_.back(); + state.loop.condpc = (condpc != ifne) ? condpc : nullptr; + state.loop.updatepc = (updatepc != condpc) ? updatepc : nullptr; + if (state.loop.updatepc) + state.loop.updateEnd = condpc; + + if (!setCurrentAndSpecializePhis(header)) + return ControlStatus_Error; + if (!jsop_loophead(loopHead)) + return ControlStatus_Error; + + return ControlStatus_Jumped; +} + +int +IonBuilder::CmpSuccessors(const void* a, const void* b) +{ + const MBasicBlock* a0 = * (MBasicBlock * const*)a; + const MBasicBlock* b0 = * (MBasicBlock * const*)b; + if (a0->pc() == b0->pc()) + return 0; + + return (a0->pc() > b0->pc()) ? 1 : -1; +} + +IonBuilder::ControlStatus +IonBuilder::tableSwitch(JSOp op, jssrcnote* sn) +{ + // TableSwitch op contains the following data + // (length between data is JUMP_OFFSET_LEN) + // + // 0: Offset of default case + // 1: Lowest number in tableswitch + // 2: Highest number in tableswitch + // 3: Offset of case low + // 4: Offset of case low+1 + // .: ... + // .: Offset of case high + + MOZ_ASSERT(op == JSOP_TABLESWITCH); + MOZ_ASSERT(SN_TYPE(sn) == SRC_TABLESWITCH); + + // Pop input. + MDefinition* ins = current->pop(); + + // Get the default and exit pc + jsbytecode* exitpc = pc + GetSrcNoteOffset(sn, 0); + jsbytecode* defaultpc = pc + GET_JUMP_OFFSET(pc); + + MOZ_ASSERT(defaultpc > pc && defaultpc <= exitpc); + + // Get the low and high from the tableswitch + jsbytecode* pc2 = pc; + pc2 += JUMP_OFFSET_LEN; + int low = GET_JUMP_OFFSET(pc2); + pc2 += JUMP_OFFSET_LEN; + int high = GET_JUMP_OFFSET(pc2); + pc2 += JUMP_OFFSET_LEN; + + // Create MIR instruction + MTableSwitch* tableswitch = MTableSwitch::New(alloc(), ins, low, high); + + // Create default case + MBasicBlock* defaultcase = newBlock(current, defaultpc); + if (!defaultcase) + return ControlStatus_Error; + + if (!tableswitch->addDefault(defaultcase)) + return ControlStatus_Error; + + if (!tableswitch->addBlock(defaultcase)) + return ControlStatus_Error; + + // Create cases + jsbytecode* casepc = nullptr; + for (int i = 0; i < high-low+1; i++) { + casepc = pc + GET_JUMP_OFFSET(pc2); + + MOZ_ASSERT(casepc >= pc && casepc <= exitpc); + MBasicBlock* caseblock; + + if (casepc == pc) { + // If the casepc equals the current pc, it is not a written case, + // but a filled gap. That way we can use a tableswitch instead of + // condswitch, even if not all numbers are consecutive. + // In that case this block goes to the default case + caseblock = newBlock(current, defaultpc); + if (!caseblock) + return ControlStatus_Error; + caseblock->end(MGoto::New(alloc(), defaultcase)); + if (!defaultcase->addPredecessor(alloc(), caseblock)) + return ControlStatus_Error; + } else { + // If this is an actual case (not filled gap), + // add this block to the list that still needs to get processed. + caseblock = newBlock(current, casepc); + if (!caseblock) + return ControlStatus_Error; + + if (!tableswitch->addBlock(caseblock)) + return ControlStatus_Error; + + // Add constant to indicate which case this is for use by + // processNextTableSwitchCase. + MConstant* constant = MConstant::New(alloc(), Int32Value(i + low)); + caseblock->add(constant); + } + + size_t caseIndex; + if (!tableswitch->addSuccessor(caseblock, &caseIndex)) + return ControlStatus_Error; + + if (!tableswitch->addCase(caseIndex)) + return ControlStatus_Error; + + pc2 += JUMP_OFFSET_LEN; + } + + // Move defaultcase to the end, to maintain RPO. + graph().moveBlockToEnd(defaultcase); + + MOZ_ASSERT(tableswitch->numCases() == (uint32_t)(high - low + 1)); + MOZ_ASSERT(tableswitch->numSuccessors() > 0); + + // Sort the list of blocks that still needs to get processed by pc + qsort(tableswitch->blocks(), tableswitch->numBlocks(), + sizeof(MBasicBlock*), CmpSuccessors); + + // Create info + ControlFlowInfo switchinfo(cfgStack_.length(), exitpc); + if (!switches_.append(switchinfo)) + return ControlStatus_Error; + + // Use a state to retrieve some information + CFGState state = CFGState::TableSwitch(exitpc, tableswitch); + + // Save the MIR instruction as last instruction of this block. + current->end(tableswitch); + + // If there is only one successor the block should stop at the end of the switch + // Else it should stop at the start of the next successor + if (tableswitch->numBlocks() > 1) + state.stopAt = tableswitch->getBlock(1)->pc(); + if (!setCurrentAndSpecializePhis(tableswitch->getBlock(0))) + return ControlStatus_Error; + + if (!cfgStack_.append(state)) + return ControlStatus_Error; + + pc = current->pc(); + return ControlStatus_Jumped; +} + +bool +IonBuilder::replaceTypeSet(MDefinition* subject, TemporaryTypeSet* type, MTest* test) +{ + if (type->unknown()) + return true; + + // Don't emit MFilterTypeSet if it doesn't improve the typeset. + if (subject->resultTypeSet()) { + if (subject->resultTypeSet()->equals(type)) + return true; + } else { + TemporaryTypeSet oldTypes(alloc_->lifoAlloc(), subject->type()); + if (oldTypes.equals(type)) + return true; + } + + MInstruction* replace = nullptr; + MDefinition* ins; + + for (uint32_t i = 0; i < current->stackDepth(); i++) { + ins = current->getSlot(i); + + // Instead of creating a new MFilterTypeSet, try to update the old one. + if (ins->isFilterTypeSet() && ins->getOperand(0) == subject && + ins->dependency() == test) + { + TemporaryTypeSet* intersect = + TypeSet::intersectSets(ins->resultTypeSet(), type, alloc_->lifoAlloc()); + if (!intersect) + return false; + + ins->toFilterTypeSet()->setResultType(intersect->getKnownMIRType()); + ins->toFilterTypeSet()->setResultTypeSet(intersect); + + if (ins->type() == MIRType::Undefined) + current->setSlot(i, constant(UndefinedValue())); + else if (ins->type() == MIRType::Null) + current->setSlot(i, constant(NullValue())); + else if (ins->type() == MIRType::MagicOptimizedArguments) + current->setSlot(i, constant(MagicValue(JS_OPTIMIZED_ARGUMENTS))); + else + MOZ_ASSERT(!IsMagicType(ins->type())); + continue; + } + + if (ins == subject) { + if (!replace) { + replace = MFilterTypeSet::New(alloc(), subject, type); + if (!replace) + return false; + + current->add(replace); + + // Make sure we don't hoist it above the MTest, we can use the + // 'dependency' of an MInstruction. This is normally used by + // Alias Analysis, but won't get overwritten, since this + // instruction doesn't have an AliasSet. + replace->setDependency(test); + + if (replace->type() == MIRType::Undefined) + replace = constant(UndefinedValue()); + else if (replace->type() == MIRType::Null) + replace = constant(NullValue()); + else if (replace->type() == MIRType::MagicOptimizedArguments) + replace = constant(MagicValue(JS_OPTIMIZED_ARGUMENTS)); + else + MOZ_ASSERT(!IsMagicType(ins->type())); + } + current->setSlot(i, replace); + } + } + return true; +} + +bool +IonBuilder::detectAndOrStructure(MPhi* ins, bool* branchIsAnd) +{ + // Look for a triangle pattern: + // + // initialBlock + // / | + // branchBlock | + // \ | + // testBlock + // + // Where ins is a phi from testBlock which combines two values + // pushed onto the stack by initialBlock and branchBlock. + + if (ins->numOperands() != 2) + return false; + + MBasicBlock* testBlock = ins->block(); + MOZ_ASSERT(testBlock->numPredecessors() == 2); + + MBasicBlock* initialBlock; + MBasicBlock* branchBlock; + if (testBlock->getPredecessor(0)->lastIns()->isTest()) { + initialBlock = testBlock->getPredecessor(0); + branchBlock = testBlock->getPredecessor(1); + } else if (testBlock->getPredecessor(1)->lastIns()->isTest()) { + initialBlock = testBlock->getPredecessor(1); + branchBlock = testBlock->getPredecessor(0); + } else { + return false; + } + + if (branchBlock->numSuccessors() != 1) + return false; + + if (branchBlock->numPredecessors() != 1 || branchBlock->getPredecessor(0) != initialBlock) + return false; + + if (initialBlock->numSuccessors() != 2) + return false; + + MDefinition* branchResult = ins->getOperand(testBlock->indexForPredecessor(branchBlock)); + MDefinition* initialResult = ins->getOperand(testBlock->indexForPredecessor(initialBlock)); + + if (branchBlock->stackDepth() != initialBlock->stackDepth()) + return false; + if (branchBlock->stackDepth() != testBlock->stackDepth() + 1) + return false; + if (branchResult != branchBlock->peek(-1) || initialResult != initialBlock->peek(-1)) + return false; + + MTest* initialTest = initialBlock->lastIns()->toTest(); + bool branchIsTrue = branchBlock == initialTest->ifTrue(); + if (initialTest->input() == ins->getOperand(0)) + *branchIsAnd = branchIsTrue != (testBlock->getPredecessor(0) == branchBlock); + else if (initialTest->input() == ins->getOperand(1)) + *branchIsAnd = branchIsTrue != (testBlock->getPredecessor(1) == branchBlock); + else + return false; + + return true; +} + +bool +IonBuilder::improveTypesAtCompare(MCompare* ins, bool trueBranch, MTest* test) +{ + if (ins->compareType() == MCompare::Compare_Undefined || + ins->compareType() == MCompare::Compare_Null) + { + return improveTypesAtNullOrUndefinedCompare(ins, trueBranch, test); + } + + if ((ins->lhs()->isTypeOf() || ins->rhs()->isTypeOf()) && + (ins->lhs()->isConstant() || ins->rhs()->isConstant())) + { + return improveTypesAtTypeOfCompare(ins, trueBranch, test); + } + + return true; +} + +bool +IonBuilder::improveTypesAtTypeOfCompare(MCompare* ins, bool trueBranch, MTest* test) +{ + MTypeOf* typeOf = ins->lhs()->isTypeOf() ? ins->lhs()->toTypeOf() : ins->rhs()->toTypeOf(); + MConstant* constant = ins->lhs()->isConstant() ? ins->lhs()->toConstant() : ins->rhs()->toConstant(); + + if (constant->type() != MIRType::String) + return true; + + bool equal = ins->jsop() == JSOP_EQ || ins->jsop() == JSOP_STRICTEQ; + bool notEqual = ins->jsop() == JSOP_NE || ins->jsop() == JSOP_STRICTNE; + + if (notEqual) + trueBranch = !trueBranch; + + // Relational compares not supported. + if (!equal && !notEqual) + return true; + + MDefinition* subject = typeOf->input(); + TemporaryTypeSet* inputTypes = subject->resultTypeSet(); + + // Create temporary typeset equal to the type if there is no resultTypeSet. + TemporaryTypeSet tmp; + if (!inputTypes) { + if (subject->type() == MIRType::Value) + return true; + inputTypes = &tmp; + tmp.addType(TypeSet::PrimitiveType(ValueTypeFromMIRType(subject->type())), alloc_->lifoAlloc()); + } + + if (inputTypes->unknown()) + return true; + + // Note: we cannot remove the AnyObject type in the false branch, + // since there are multiple ways to get an object. That is the reason + // for the 'trueBranch' test. + TemporaryTypeSet filter; + const JSAtomState& names = GetJitContext()->runtime->names(); + if (constant->toString() == TypeName(JSTYPE_VOID, names)) { + filter.addType(TypeSet::UndefinedType(), alloc_->lifoAlloc()); + if (typeOf->inputMaybeCallableOrEmulatesUndefined() && trueBranch) + filter.addType(TypeSet::AnyObjectType(), alloc_->lifoAlloc()); + } else if (constant->toString() == TypeName(JSTYPE_BOOLEAN, names)) { + filter.addType(TypeSet::BooleanType(), alloc_->lifoAlloc()); + } else if (constant->toString() == TypeName(JSTYPE_NUMBER, names)) { + filter.addType(TypeSet::Int32Type(), alloc_->lifoAlloc()); + filter.addType(TypeSet::DoubleType(), alloc_->lifoAlloc()); + } else if (constant->toString() == TypeName(JSTYPE_STRING, names)) { + filter.addType(TypeSet::StringType(), alloc_->lifoAlloc()); + } else if (constant->toString() == TypeName(JSTYPE_SYMBOL, names)) { + filter.addType(TypeSet::SymbolType(), alloc_->lifoAlloc()); + } else if (constant->toString() == TypeName(JSTYPE_OBJECT, names)) { + filter.addType(TypeSet::NullType(), alloc_->lifoAlloc()); + if (trueBranch) + filter.addType(TypeSet::AnyObjectType(), alloc_->lifoAlloc()); + } else if (constant->toString() == TypeName(JSTYPE_FUNCTION, names)) { + if (typeOf->inputMaybeCallableOrEmulatesUndefined() && trueBranch) + filter.addType(TypeSet::AnyObjectType(), alloc_->lifoAlloc()); + } else { + return true; + } + + TemporaryTypeSet* type; + if (trueBranch) + type = TypeSet::intersectSets(&filter, inputTypes, alloc_->lifoAlloc()); + else + type = TypeSet::removeSet(inputTypes, &filter, alloc_->lifoAlloc()); + + if (!type) + return false; + + return replaceTypeSet(subject, type, test); +} + +bool +IonBuilder::improveTypesAtNullOrUndefinedCompare(MCompare* ins, bool trueBranch, MTest* test) +{ + MOZ_ASSERT(ins->compareType() == MCompare::Compare_Undefined || + ins->compareType() == MCompare::Compare_Null); + + // altersUndefined/Null represents if we can filter/set Undefined/Null. + bool altersUndefined, altersNull; + JSOp op = ins->jsop(); + + switch(op) { + case JSOP_STRICTNE: + case JSOP_STRICTEQ: + altersUndefined = ins->compareType() == MCompare::Compare_Undefined; + altersNull = ins->compareType() == MCompare::Compare_Null; + break; + case JSOP_NE: + case JSOP_EQ: + altersUndefined = altersNull = true; + break; + default: + MOZ_CRASH("Relational compares not supported"); + } + + MDefinition* subject = ins->lhs(); + TemporaryTypeSet* inputTypes = subject->resultTypeSet(); + + MOZ_ASSERT(IsNullOrUndefined(ins->rhs()->type())); + + // Create temporary typeset equal to the type if there is no resultTypeSet. + TemporaryTypeSet tmp; + if (!inputTypes) { + if (subject->type() == MIRType::Value) + return true; + inputTypes = &tmp; + tmp.addType(TypeSet::PrimitiveType(ValueTypeFromMIRType(subject->type())), alloc_->lifoAlloc()); + } + + if (inputTypes->unknown()) + return true; + + TemporaryTypeSet* type; + + // Decide if we need to filter the type or set it. + if ((op == JSOP_STRICTEQ || op == JSOP_EQ) ^ trueBranch) { + // Remove undefined/null + TemporaryTypeSet remove; + if (altersUndefined) + remove.addType(TypeSet::UndefinedType(), alloc_->lifoAlloc()); + if (altersNull) + remove.addType(TypeSet::NullType(), alloc_->lifoAlloc()); + + type = TypeSet::removeSet(inputTypes, &remove, alloc_->lifoAlloc()); + } else { + // Set undefined/null. + TemporaryTypeSet base; + if (altersUndefined) { + base.addType(TypeSet::UndefinedType(), alloc_->lifoAlloc()); + // If TypeSet emulates undefined, then we cannot filter the objects. + if (inputTypes->maybeEmulatesUndefined(constraints())) + base.addType(TypeSet::AnyObjectType(), alloc_->lifoAlloc()); + } + + if (altersNull) + base.addType(TypeSet::NullType(), alloc_->lifoAlloc()); + + type = TypeSet::intersectSets(&base, inputTypes, alloc_->lifoAlloc()); + } + + if (!type) + return false; + + return replaceTypeSet(subject, type, test); +} + +bool +IonBuilder::improveTypesAtTest(MDefinition* ins, bool trueBranch, MTest* test) +{ + // We explore the test condition to try and deduce as much type information + // as possible. + + // All branches of this switch that don't want to fall through to the + // default behavior must return. The default behavior assumes that a true + // test means the incoming ins is not null or undefined and that a false + // tests means it's one of null, undefined, false, 0, "", and objects + // emulating undefined + switch (ins->op()) { + case MDefinition::Op_Not: + return improveTypesAtTest(ins->toNot()->getOperand(0), !trueBranch, test); + case MDefinition::Op_IsObject: { + MDefinition* subject = ins->getOperand(0); + TemporaryTypeSet* oldType = subject->resultTypeSet(); + + // Create temporary typeset equal to the type if there is no resultTypeSet. + TemporaryTypeSet tmp; + if (!oldType) { + if (subject->type() == MIRType::Value) + return true; + oldType = &tmp; + tmp.addType(TypeSet::PrimitiveType(ValueTypeFromMIRType(subject->type())), alloc_->lifoAlloc()); + } + + if (oldType->unknown()) + return true; + + TemporaryTypeSet* type = nullptr; + if (trueBranch) + type = oldType->cloneObjectsOnly(alloc_->lifoAlloc()); + else + type = oldType->cloneWithoutObjects(alloc_->lifoAlloc()); + + if (!type) + return false; + + return replaceTypeSet(subject, type, test); + } + case MDefinition::Op_Phi: { + bool branchIsAnd = true; + if (!detectAndOrStructure(ins->toPhi(), &branchIsAnd)) { + // Just fall through to the default behavior. + break; + } + + // Now we have detected the triangular structure and determined if it + // was an AND or an OR. + if (branchIsAnd) { + if (trueBranch) { + if (!improveTypesAtTest(ins->toPhi()->getOperand(0), true, test)) + return false; + if (!improveTypesAtTest(ins->toPhi()->getOperand(1), true, test)) + return false; + } + } else { + /* + * if (a || b) { + * ... + * } else { + * ... + * } + * + * If we have a statements like the one described above, + * And we are in the else branch of it. It amounts to: + * if (!(a || b)) and being in the true branch. + * + * Simplifying, we have (!a && !b) + * In this case we can use the same logic we use for branchIsAnd + * + */ + if (!trueBranch) { + if (!improveTypesAtTest(ins->toPhi()->getOperand(0), false, test)) + return false; + if (!improveTypesAtTest(ins->toPhi()->getOperand(1), false, test)) + return false; + } + } + return true; + } + + case MDefinition::Op_Compare: + return improveTypesAtCompare(ins->toCompare(), trueBranch, test); + + default: + break; + } + + // By default MTest tests ToBoolean(input). As a result in the true branch we can filter + // undefined and null. In false branch we can only encounter undefined, null, false, 0, "" + // and objects that emulate undefined. + + TemporaryTypeSet* oldType = ins->resultTypeSet(); + TemporaryTypeSet* type; + + // Create temporary typeset equal to the type if there is no resultTypeSet. + TemporaryTypeSet tmp; + if (!oldType) { + if (ins->type() == MIRType::Value) + return true; + oldType = &tmp; + tmp.addType(TypeSet::PrimitiveType(ValueTypeFromMIRType(ins->type())), alloc_->lifoAlloc()); + } + + // If ins does not have a typeset we return as we cannot optimize. + if (oldType->unknown()) + return true; + + // Decide either to set or remove. + if (trueBranch) { + TemporaryTypeSet remove; + remove.addType(TypeSet::UndefinedType(), alloc_->lifoAlloc()); + remove.addType(TypeSet::NullType(), alloc_->lifoAlloc()); + type = TypeSet::removeSet(oldType, &remove, alloc_->lifoAlloc()); + } else { + TemporaryTypeSet base; + base.addType(TypeSet::UndefinedType(), alloc_->lifoAlloc()); // ToBoolean(undefined) == false + base.addType(TypeSet::NullType(), alloc_->lifoAlloc()); // ToBoolean(null) == false + base.addType(TypeSet::BooleanType(), alloc_->lifoAlloc()); // ToBoolean(false) == false + base.addType(TypeSet::Int32Type(), alloc_->lifoAlloc()); // ToBoolean(0) == false + base.addType(TypeSet::DoubleType(), alloc_->lifoAlloc()); // ToBoolean(0.0) == false + base.addType(TypeSet::StringType(), alloc_->lifoAlloc()); // ToBoolean("") == false + + // If the typeset does emulate undefined, then we cannot filter out + // objects. + if (oldType->maybeEmulatesUndefined(constraints())) + base.addType(TypeSet::AnyObjectType(), alloc_->lifoAlloc()); + + type = TypeSet::intersectSets(&base, oldType, alloc_->lifoAlloc()); + } + + return type && replaceTypeSet(ins, type, test); +} + +bool +IonBuilder::jsop_label() +{ + MOZ_ASSERT(JSOp(*pc) == JSOP_LABEL); + + jsbytecode* endpc = pc + GET_JUMP_OFFSET(pc); + MOZ_ASSERT(endpc > pc); + + ControlFlowInfo label(cfgStack_.length(), endpc); + if (!labels_.append(label)) + return false; + + return cfgStack_.append(CFGState::Label(endpc)); +} + +bool +IonBuilder::jsop_condswitch() +{ + // CondSwitch op looks as follows: + // condswitch [length +exit_pc; first case offset +next-case ] + // { + // { + // ... any code ... + // case (+jump) [pcdelta offset +next-case] + // }+ + // default (+jump) + // ... jump targets ... + // } + // + // The default case is always emitted even if there is no default case in + // the source. The last case statement pcdelta source note might have a 0 + // offset on the last case (not all the time). + // + // A conditional evaluate the condition of each case and compare it to the + // switch value with a strict equality. Cases conditions are iterated + // linearly until one is matching. If one case succeeds, the flow jumps into + // the corresponding body block. The body block might alias others and + // might continue in the next body block if the body is not terminated with + // a break. + // + // Algorithm: + // 1/ Loop over the case chain to reach the default target + // & Estimate the number of uniq bodies. + // 2/ Generate code for all cases (see processCondSwitchCase). + // 3/ Generate code for all bodies (see processCondSwitchBody). + + MOZ_ASSERT(JSOp(*pc) == JSOP_CONDSWITCH); + jssrcnote* sn = info().getNote(gsn, pc); + MOZ_ASSERT(SN_TYPE(sn) == SRC_CONDSWITCH); + + // Get the exit pc + jsbytecode* exitpc = pc + GetSrcNoteOffset(sn, 0); + jsbytecode* firstCase = pc + GetSrcNoteOffset(sn, 1); + + // Iterate all cases in the conditional switch. + // - Stop at the default case. (always emitted after the last case) + // - Estimate the number of uniq bodies. This estimation might be off by 1 + // if the default body alias a case body. + jsbytecode* curCase = firstCase; + jsbytecode* lastTarget = GetJumpOffset(curCase) + curCase; + size_t nbBodies = 2; // default target and the first body. + + MOZ_ASSERT(pc < curCase && curCase <= exitpc); + while (JSOp(*curCase) == JSOP_CASE) { + // Fetch the next case. + jssrcnote* caseSn = info().getNote(gsn, curCase); + MOZ_ASSERT(caseSn && SN_TYPE(caseSn) == SRC_NEXTCASE); + ptrdiff_t off = GetSrcNoteOffset(caseSn, 0); + MOZ_ASSERT_IF(off == 0, JSOp(*GetNextPc(curCase)) == JSOP_JUMPTARGET); + curCase = off ? curCase + off : GetNextPc(GetNextPc(curCase)); + MOZ_ASSERT(pc < curCase && curCase <= exitpc); + + // Count non-aliased cases. + jsbytecode* curTarget = GetJumpOffset(curCase) + curCase; + if (lastTarget < curTarget) + nbBodies++; + lastTarget = curTarget; + } + + // The current case now be the default case which jump to the body of the + // default case, which might be behind the last target. + MOZ_ASSERT(JSOp(*curCase) == JSOP_DEFAULT); + jsbytecode* defaultTarget = GetJumpOffset(curCase) + curCase; + MOZ_ASSERT(curCase < defaultTarget && defaultTarget <= exitpc); + + // Allocate the current graph state. + CFGState state = CFGState::CondSwitch(this, exitpc, defaultTarget); + if (!state.condswitch.bodies || !state.condswitch.bodies->init(alloc(), nbBodies)) + return ControlStatus_Error; + + // We loop on case conditions with processCondSwitchCase. + MOZ_ASSERT(JSOp(*firstCase) == JSOP_CASE); + state.stopAt = firstCase; + state.state = CFGState::COND_SWITCH_CASE; + + return cfgStack_.append(state); +} + +IonBuilder::CFGState +IonBuilder::CFGState::CondSwitch(IonBuilder* builder, jsbytecode* exitpc, jsbytecode* defaultTarget) +{ + CFGState state; + state.state = COND_SWITCH_CASE; + state.stopAt = nullptr; + state.condswitch.bodies = (FixedList<MBasicBlock*>*)builder->alloc_->allocate( + sizeof(FixedList<MBasicBlock*>)); + state.condswitch.currentIdx = 0; + state.condswitch.defaultTarget = defaultTarget; + state.condswitch.defaultIdx = uint32_t(-1); + state.condswitch.exitpc = exitpc; + state.condswitch.breaks = nullptr; + return state; +} + +IonBuilder::CFGState +IonBuilder::CFGState::Label(jsbytecode* exitpc) +{ + CFGState state; + state.state = LABEL; + state.stopAt = exitpc; + state.label.breaks = nullptr; + return state; +} + +IonBuilder::CFGState +IonBuilder::CFGState::Try(jsbytecode* exitpc, MBasicBlock* successor) +{ + CFGState state; + state.state = TRY; + state.stopAt = exitpc; + state.try_.successor = successor; + return state; +} + +IonBuilder::ControlStatus +IonBuilder::processCondSwitchCase(CFGState& state) +{ + MOZ_ASSERT(state.state == CFGState::COND_SWITCH_CASE); + MOZ_ASSERT(!state.condswitch.breaks); + MOZ_ASSERT(current); + MOZ_ASSERT(JSOp(*pc) == JSOP_CASE); + FixedList<MBasicBlock*>& bodies = *state.condswitch.bodies; + jsbytecode* defaultTarget = state.condswitch.defaultTarget; + uint32_t& currentIdx = state.condswitch.currentIdx; + jsbytecode* lastTarget = currentIdx ? bodies[currentIdx - 1]->pc() : nullptr; + + // Fetch the following case in which we will continue. + jssrcnote* sn = info().getNote(gsn, pc); + ptrdiff_t off = GetSrcNoteOffset(sn, 0); + MOZ_ASSERT_IF(off == 0, JSOp(*GetNextPc(pc)) == JSOP_JUMPTARGET); + jsbytecode* casePc = off ? pc + off : GetNextPc(GetNextPc(pc)); + bool caseIsDefault = JSOp(*casePc) == JSOP_DEFAULT; + MOZ_ASSERT(JSOp(*casePc) == JSOP_CASE || caseIsDefault); + + // Allocate the block of the matching case. + bool bodyIsNew = false; + MBasicBlock* bodyBlock = nullptr; + jsbytecode* bodyTarget = pc + GetJumpOffset(pc); + if (lastTarget < bodyTarget) { + // If the default body is in the middle or aliasing the current target. + if (lastTarget < defaultTarget && defaultTarget <= bodyTarget) { + MOZ_ASSERT(state.condswitch.defaultIdx == uint32_t(-1)); + state.condswitch.defaultIdx = currentIdx; + bodies[currentIdx] = nullptr; + // If the default body does not alias any and it would be allocated + // later and stored in the defaultIdx location. + if (defaultTarget < bodyTarget) + currentIdx++; + } + + bodyIsNew = true; + // Pop switch and case operands. + bodyBlock = newBlockPopN(current, bodyTarget, 2); + bodies[currentIdx++] = bodyBlock; + } else { + // This body alias the previous one. + MOZ_ASSERT(lastTarget == bodyTarget); + MOZ_ASSERT(currentIdx > 0); + bodyBlock = bodies[currentIdx - 1]; + } + + if (!bodyBlock) + return ControlStatus_Error; + + lastTarget = bodyTarget; + + // Allocate the block of the non-matching case. This can either be a normal + // case or the default case. + bool caseIsNew = false; + MBasicBlock* caseBlock = nullptr; + if (!caseIsDefault) { + caseIsNew = true; + // Pop the case operand. + caseBlock = newBlockPopN(current, GetNextPc(pc), 1); + } else { + // The non-matching case is the default case, which jump directly to its + // body. Skip the creation of a default case block and directly create + // the default body if it does not alias any previous body. + + if (state.condswitch.defaultIdx == uint32_t(-1)) { + // The default target is the last target. + MOZ_ASSERT(lastTarget < defaultTarget); + state.condswitch.defaultIdx = currentIdx++; + caseIsNew = true; + } else if (bodies[state.condswitch.defaultIdx] == nullptr) { + // The default target is in the middle and it does not alias any + // case target. + MOZ_ASSERT(defaultTarget < lastTarget); + caseIsNew = true; + } else { + // The default target is in the middle and it alias a case target. + MOZ_ASSERT(defaultTarget <= lastTarget); + caseBlock = bodies[state.condswitch.defaultIdx]; + } + + // Allocate and register the default body. + if (caseIsNew) { + // Pop the case & switch operands. + caseBlock = newBlockPopN(current, defaultTarget, 2); + bodies[state.condswitch.defaultIdx] = caseBlock; + } + } + + if (!caseBlock) + return ControlStatus_Error; + + // Terminate the last case condition block by emitting the code + // corresponding to JSOP_CASE bytecode. + if (bodyBlock != caseBlock) { + MDefinition* caseOperand = current->pop(); + MDefinition* switchOperand = current->peek(-1); + + if (!jsop_compare(JSOP_STRICTEQ, switchOperand, caseOperand)) + return ControlStatus_Error; + MInstruction* cmpResult = current->pop()->toInstruction(); + MOZ_ASSERT(!cmpResult->isEffectful()); + current->end(newTest(cmpResult, bodyBlock, caseBlock)); + + // Add last case as predecessor of the body if the body is aliasing + // the previous case body. + if (!bodyIsNew && !bodyBlock->addPredecessorPopN(alloc(), current, 1)) + return ControlStatus_Error; + + // Add last case as predecessor of the non-matching case if the + // non-matching case is an aliased default case. We need to pop the + // switch operand as we skip the default case block and use the default + // body block directly. + MOZ_ASSERT_IF(!caseIsNew, caseIsDefault); + if (!caseIsNew && !caseBlock->addPredecessorPopN(alloc(), current, 1)) + return ControlStatus_Error; + } else { + // The default case alias the last case body. + MOZ_ASSERT(caseIsDefault); + current->pop(); // Case operand + current->pop(); // Switch operand + current->end(MGoto::New(alloc(), bodyBlock)); + if (!bodyIsNew && !bodyBlock->addPredecessor(alloc(), current)) + return ControlStatus_Error; + } + + if (caseIsDefault) { + // The last case condition is finished. Loop in processCondSwitchBody, + // with potential stops in processSwitchBreak. Check that the bodies + // fixed list is over-estimate by at most 1, and shrink the size such as + // length can be used as an upper bound while iterating bodies. + MOZ_ASSERT(currentIdx == bodies.length() || currentIdx + 1 == bodies.length()); + bodies.shrink(bodies.length() - currentIdx); + + // Handle break statements in processSwitchBreak while processing + // bodies. + ControlFlowInfo breakInfo(cfgStack_.length() - 1, state.condswitch.exitpc); + if (!switches_.append(breakInfo)) + return ControlStatus_Error; + + // Jump into the first body. + currentIdx = 0; + setCurrent(nullptr); + state.state = CFGState::COND_SWITCH_BODY; + return processCondSwitchBody(state); + } + + // Continue until the case condition. + if (!setCurrentAndSpecializePhis(caseBlock)) + return ControlStatus_Error; + pc = current->pc(); + state.stopAt = casePc; + return ControlStatus_Jumped; +} + +IonBuilder::ControlStatus +IonBuilder::processCondSwitchBody(CFGState& state) +{ + MOZ_ASSERT(state.state == CFGState::COND_SWITCH_BODY); + MOZ_ASSERT(pc <= state.condswitch.exitpc); + FixedList<MBasicBlock*>& bodies = *state.condswitch.bodies; + uint32_t& currentIdx = state.condswitch.currentIdx; + + MOZ_ASSERT(currentIdx <= bodies.length()); + if (currentIdx == bodies.length()) { + MOZ_ASSERT_IF(current, pc == state.condswitch.exitpc); + return processSwitchEnd(state.condswitch.breaks, state.condswitch.exitpc); + } + + // Get the next body + MBasicBlock* nextBody = bodies[currentIdx++]; + MOZ_ASSERT_IF(current, pc == nextBody->pc()); + + // Fix the reverse post-order iteration. + graph().moveBlockToEnd(nextBody); + + // The last body continue into the new one. + if (current) { + current->end(MGoto::New(alloc(), nextBody)); + if (!nextBody->addPredecessor(alloc(), current)) + return ControlStatus_Error; + } + + // Continue in the next body. + if (!setCurrentAndSpecializePhis(nextBody)) + return ControlStatus_Error; + pc = current->pc(); + + if (currentIdx < bodies.length()) + state.stopAt = bodies[currentIdx]->pc(); + else + state.stopAt = state.condswitch.exitpc; + return ControlStatus_Jumped; +} + +bool +IonBuilder::jsop_andor(JSOp op) +{ + MOZ_ASSERT(op == JSOP_AND || op == JSOP_OR); + + jsbytecode* rhsStart = pc + CodeSpec[op].length; + jsbytecode* joinStart = pc + GetJumpOffset(pc); + MOZ_ASSERT(joinStart > pc); + + // We have to leave the LHS on the stack. + MDefinition* lhs = current->peek(-1); + + MBasicBlock* evalLhs = newBlock(current, joinStart); + MBasicBlock* evalRhs = newBlock(current, rhsStart); + if (!evalLhs || !evalRhs) + return false; + + MTest* test = (op == JSOP_AND) + ? newTest(lhs, evalRhs, evalLhs) + : newTest(lhs, evalLhs, evalRhs); + current->end(test); + + // Create the lhs block and specialize. + if (!setCurrentAndSpecializePhis(evalLhs)) + return false; + + if (!improveTypesAtTest(test->getOperand(0), test->ifTrue() == current, test)) + return false; + + // Create the rhs block. + if (!cfgStack_.append(CFGState::AndOr(joinStart, evalLhs))) + return false; + + if (!setCurrentAndSpecializePhis(evalRhs)) + return false; + + if (!improveTypesAtTest(test->getOperand(0), test->ifTrue() == current, test)) + return false; + + return true; +} + +bool +IonBuilder::jsop_dup2() +{ + uint32_t lhsSlot = current->stackDepth() - 2; + uint32_t rhsSlot = current->stackDepth() - 1; + current->pushSlot(lhsSlot); + current->pushSlot(rhsSlot); + return true; +} + +bool +IonBuilder::jsop_loophead(jsbytecode* pc) +{ + assertValidLoopHeadOp(pc); + + current->add(MInterruptCheck::New(alloc())); + insertRecompileCheck(); + + return true; +} + +bool +IonBuilder::jsop_ifeq(JSOp op) +{ + // IFEQ always has a forward offset. + jsbytecode* trueStart = pc + CodeSpec[op].length; + jsbytecode* falseStart = pc + GetJumpOffset(pc); + MOZ_ASSERT(falseStart > pc); + + // We only handle cases that emit source notes. + jssrcnote* sn = info().getNote(gsn, pc); + if (!sn) + return abort("expected sourcenote"); + + MDefinition* ins = current->pop(); + + // Create true and false branches. + MBasicBlock* ifTrue = newBlock(current, trueStart); + MBasicBlock* ifFalse = newBlock(current, falseStart); + if (!ifTrue || !ifFalse) + return false; + + MTest* test = newTest(ins, ifTrue, ifFalse); + current->end(test); + + // The bytecode for if/ternary gets emitted either like this: + // + // IFEQ X ; src note (IF_ELSE, COND) points to the GOTO + // ... + // GOTO Z + // X: ... ; else/else if + // ... + // Z: ; join + // + // Or like this: + // + // IFEQ X ; src note (IF) has no offset + // ... + // Z: ... ; join + // + // We want to parse the bytecode as if we were parsing the AST, so for the + // IF_ELSE/COND cases, we use the source note and follow the GOTO. For the + // IF case, the IFEQ offset is the join point. + switch (SN_TYPE(sn)) { + case SRC_IF: + if (!cfgStack_.append(CFGState::If(falseStart, test))) + return false; + break; + + case SRC_IF_ELSE: + case SRC_COND: + { + // Infer the join point from the JSOP_GOTO[X] sitting here, then + // assert as we much we can that this is the right GOTO. + jsbytecode* trueEnd = pc + GetSrcNoteOffset(sn, 0); + MOZ_ASSERT(trueEnd > pc); + MOZ_ASSERT(trueEnd < falseStart); + MOZ_ASSERT(JSOp(*trueEnd) == JSOP_GOTO); + MOZ_ASSERT(!info().getNote(gsn, trueEnd)); + + jsbytecode* falseEnd = trueEnd + GetJumpOffset(trueEnd); + MOZ_ASSERT(falseEnd > trueEnd); + MOZ_ASSERT(falseEnd >= falseStart); + + if (!cfgStack_.append(CFGState::IfElse(trueEnd, falseEnd, test))) + return false; + break; + } + + default: + MOZ_CRASH("unexpected source note type"); + } + + // Switch to parsing the true branch. Note that no PC update is needed, + // it's the next instruction. + if (!setCurrentAndSpecializePhis(ifTrue)) + return false; + + // Filter the types in the true branch. + if (!improveTypesAtTest(test->getOperand(0), test->ifTrue() == current, test)) + return false; + + return true; +} + +bool +IonBuilder::jsop_try() +{ + MOZ_ASSERT(JSOp(*pc) == JSOP_TRY); + + // Try-finally is not yet supported. + if (analysis().hasTryFinally()) + return abort("Has try-finally"); + + // Try-catch within inline frames is not yet supported. + MOZ_ASSERT(!isInlineBuilder()); + + // Try-catch during the arguments usage analysis is not yet supported. Code + // accessing the arguments within the 'catch' block is not accounted for. + if (info().analysisMode() == Analysis_ArgumentsUsage) + return abort("Try-catch during arguments usage analysis"); + + graph().setHasTryBlock(); + + jssrcnote* sn = info().getNote(gsn, pc); + MOZ_ASSERT(SN_TYPE(sn) == SRC_TRY); + + // Get the pc of the last instruction in the try block. It's a JSOP_GOTO to + // jump over the catch block. + jsbytecode* endpc = pc + GetSrcNoteOffset(sn, 0); + MOZ_ASSERT(JSOp(*endpc) == JSOP_GOTO); + MOZ_ASSERT(GetJumpOffset(endpc) > 0); + + jsbytecode* afterTry = endpc + GetJumpOffset(endpc); + + // If controlflow in the try body is terminated (by a return or throw + // statement), the code after the try-statement may still be reachable + // via the catch block (which we don't compile) and OSR can enter it. + // For example: + // + // try { + // throw 3; + // } catch(e) { } + // + // for (var i=0; i<1000; i++) {} + // + // To handle this, we create two blocks: one for the try block and one + // for the code following the try-catch statement. Both blocks are + // connected to the graph with an MGotoWithFake instruction that always + // jumps to the try block. This ensures the successor block always has a + // predecessor. + // + // If the code after the try block is unreachable (control flow in both the + // try and catch blocks is terminated), only create the try block, to avoid + // parsing unreachable code. + + MBasicBlock* tryBlock = newBlock(current, GetNextPc(pc)); + if (!tryBlock) + return false; + + MBasicBlock* successor; + if (analysis().maybeInfo(afterTry)) { + successor = newBlock(current, afterTry); + if (!successor) + return false; + + current->end(MGotoWithFake::New(alloc(), tryBlock, successor)); + } else { + successor = nullptr; + current->end(MGoto::New(alloc(), tryBlock)); + } + + if (!cfgStack_.append(CFGState::Try(endpc, successor))) + return false; + + // The baseline compiler should not attempt to enter the catch block + // via OSR. + MOZ_ASSERT(info().osrPc() < endpc || info().osrPc() >= afterTry); + + // Start parsing the try block. + return setCurrentAndSpecializePhis(tryBlock); +} + +IonBuilder::ControlStatus +IonBuilder::processReturn(JSOp op) +{ + MDefinition* def; + switch (op) { + case JSOP_RETURN: + // Return the last instruction. + def = current->pop(); + break; + + case JSOP_RETRVAL: + // Return undefined eagerly if script doesn't use return value. + if (script()->noScriptRval()) { + MInstruction* ins = MConstant::New(alloc(), UndefinedValue()); + current->add(ins); + def = ins; + break; + } + + def = current->getSlot(info().returnValueSlot()); + break; + + default: + def = nullptr; + MOZ_CRASH("unknown return op"); + } + + MReturn* ret = MReturn::New(alloc(), def); + current->end(ret); + + if (!graph().addReturn(current)) + return ControlStatus_Error; + + // Make sure no one tries to use this block now. + setCurrent(nullptr); + return processControlEnd(); +} + +IonBuilder::ControlStatus +IonBuilder::processThrow() +{ + MDefinition* def = current->pop(); + + // MThrow is not marked as effectful. This means when it throws and we + // are inside a try block, we could use an earlier resume point and this + // resume point may not be up-to-date, for example: + // + // (function() { + // try { + // var x = 1; + // foo(); // resume point + // x = 2; + // throw foo; + // } catch(e) { + // print(x); + // } + // ])(); + // + // If we use the resume point after the call, this will print 1 instead + // of 2. To fix this, we create a resume point right before the MThrow. + // + // Note that this is not a problem for instructions other than MThrow + // because they are either marked as effectful (have their own resume + // point) or cannot throw a catchable exception. + // + // We always install this resume point (instead of only when the function + // has a try block) in order to handle the Debugger onExceptionUnwind + // hook. When we need to handle the hook, we bail out to baseline right + // after the throw and propagate the exception when debug mode is on. This + // is opposed to the normal behavior of resuming directly in the + // associated catch block. + MNop* nop = MNop::New(alloc()); + current->add(nop); + + if (!resumeAfter(nop)) + return ControlStatus_Error; + + MThrow* ins = MThrow::New(alloc(), def); + current->end(ins); + + // Make sure no one tries to use this block now. + setCurrent(nullptr); + return processControlEnd(); +} + +void +IonBuilder::pushConstant(const Value& v) +{ + current->push(constant(v)); +} + +bool +IonBuilder::bitnotTrySpecialized(bool* emitted, MDefinition* input) +{ + MOZ_ASSERT(*emitted == false); + + // Try to emit a specialized bitnot instruction based on the input type + // of the operand. + + if (input->mightBeType(MIRType::Object) || input->mightBeType(MIRType::Symbol)) + return true; + + MBitNot* ins = MBitNot::New(alloc(), input); + ins->setSpecialization(MIRType::Int32); + + current->add(ins); + current->push(ins); + + *emitted = true; + return true; +} + +bool +IonBuilder::jsop_bitnot() +{ + bool emitted = false; + + MDefinition* input = current->pop(); + + if (!forceInlineCaches()) { + if (!bitnotTrySpecialized(&emitted, input) || emitted) + return emitted; + } + + if (!arithTrySharedStub(&emitted, JSOP_BITNOT, nullptr, input) || emitted) + return emitted; + + // Not possible to optimize. Do a slow vm call. + MBitNot* ins = MBitNot::New(alloc(), input); + + current->add(ins); + current->push(ins); + MOZ_ASSERT(ins->isEffectful()); + return resumeAfter(ins); +} + +bool +IonBuilder::jsop_bitop(JSOp op) +{ + // Pop inputs. + MDefinition* right = current->pop(); + MDefinition* left = current->pop(); + + MBinaryBitwiseInstruction* ins; + switch (op) { + case JSOP_BITAND: + ins = MBitAnd::New(alloc(), left, right); + break; + + case JSOP_BITOR: + ins = MBitOr::New(alloc(), left, right); + break; + + case JSOP_BITXOR: + ins = MBitXor::New(alloc(), left, right); + break; + + case JSOP_LSH: + ins = MLsh::New(alloc(), left, right); + break; + + case JSOP_RSH: + ins = MRsh::New(alloc(), left, right); + break; + + case JSOP_URSH: + ins = MUrsh::New(alloc(), left, right); + break; + + default: + MOZ_CRASH("unexpected bitop"); + } + + current->add(ins); + ins->infer(inspector, pc); + + current->push(ins); + if (ins->isEffectful() && !resumeAfter(ins)) + return false; + + return true; +} + +MDefinition::Opcode +JSOpToMDefinition(JSOp op) +{ + switch (op) { + case JSOP_ADD: + return MDefinition::Op_Add; + case JSOP_SUB: + return MDefinition::Op_Sub; + case JSOP_MUL: + return MDefinition::Op_Mul; + case JSOP_DIV: + return MDefinition::Op_Div; + case JSOP_MOD: + return MDefinition::Op_Mod; + default: + MOZ_CRASH("unexpected binary opcode"); + } +} + +bool +IonBuilder::binaryArithTryConcat(bool* emitted, JSOp op, MDefinition* left, MDefinition* right) +{ + MOZ_ASSERT(*emitted == false); + + // Try to convert an addition into a concat operation if the inputs + // indicate this might be a concatenation. + + // Only try to replace this with concat when we have an addition. + if (op != JSOP_ADD) + return true; + + trackOptimizationAttempt(TrackedStrategy::BinaryArith_Concat); + + // Make sure one of the inputs is a string. + if (left->type() != MIRType::String && right->type() != MIRType::String) { + trackOptimizationOutcome(TrackedOutcome::OperandNotString); + return true; + } + + // The none-string input (if present) should be atleast a numerical type. + // Which we can easily coerce to string. + if (right->type() != MIRType::String && !IsNumberType(right->type())) { + trackOptimizationOutcome(TrackedOutcome::OperandNotStringOrNumber); + return true; + } + if (left->type() != MIRType::String && !IsNumberType(left->type())) { + trackOptimizationOutcome(TrackedOutcome::OperandNotStringOrNumber); + return true; + } + + MConcat* ins = MConcat::New(alloc(), left, right); + current->add(ins); + current->push(ins); + + if (!maybeInsertResume()) + return false; + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +bool +IonBuilder::powTrySpecialized(bool* emitted, MDefinition* base, MDefinition* power, + MIRType outputType) +{ + // Typechecking. + MDefinition* output = nullptr; + MIRType baseType = base->type(); + MIRType powerType = power->type(); + + if (outputType != MIRType::Int32 && outputType != MIRType::Double) + return true; + if (!IsNumberType(baseType)) + return true; + if (!IsNumberType(powerType)) + return true; + + if (powerType == MIRType::Float32) + powerType = MIRType::Double; + + MPow* pow = MPow::New(alloc(), base, power, powerType); + current->add(pow); + output = pow; + + // Cast to the right type + if (outputType == MIRType::Int32 && output->type() != MIRType::Int32) { + MToInt32* toInt = MToInt32::New(alloc(), output); + current->add(toInt); + output = toInt; + } + if (outputType == MIRType::Double && output->type() != MIRType::Double) { + MToDouble* toDouble = MToDouble::New(alloc(), output); + current->add(toDouble); + output = toDouble; + } + + current->push(output); + *emitted = true; + return true; +} + +static inline bool +SimpleArithOperand(MDefinition* op) +{ + return !op->mightBeType(MIRType::Object) + && !op->mightBeType(MIRType::String) + && !op->mightBeType(MIRType::Symbol) + && !op->mightBeType(MIRType::MagicOptimizedArguments) + && !op->mightBeType(MIRType::MagicHole) + && !op->mightBeType(MIRType::MagicIsConstructing); +} + +bool +IonBuilder::binaryArithTrySpecialized(bool* emitted, JSOp op, MDefinition* left, MDefinition* right) +{ + MOZ_ASSERT(*emitted == false); + + // Try to emit a specialized binary instruction based on the input types + // of the operands. + + trackOptimizationAttempt(TrackedStrategy::BinaryArith_SpecializedTypes); + + // Anything complex - strings, symbols, and objects - are not specialized + if (!SimpleArithOperand(left) || !SimpleArithOperand(right)) { + trackOptimizationOutcome(TrackedOutcome::OperandNotSimpleArith); + return true; + } + + // One of the inputs need to be a number. + if (!IsNumberType(left->type()) && !IsNumberType(right->type())) { + trackOptimizationOutcome(TrackedOutcome::OperandNotNumber); + return true; + } + + MDefinition::Opcode defOp = JSOpToMDefinition(op); + MBinaryArithInstruction* ins = MBinaryArithInstruction::New(alloc(), defOp, left, right); + ins->setNumberSpecialization(alloc(), inspector, pc); + + if (op == JSOP_ADD || op == JSOP_MUL) + ins->setCommutative(); + + current->add(ins); + current->push(ins); + + MOZ_ASSERT(!ins->isEffectful()); + if (!maybeInsertResume()) + return false; + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +bool +IonBuilder::binaryArithTrySpecializedOnBaselineInspector(bool* emitted, JSOp op, + MDefinition* left, MDefinition* right) +{ + MOZ_ASSERT(*emitted == false); + + // Try to emit a specialized binary instruction speculating the + // type using the baseline caches. + + trackOptimizationAttempt(TrackedStrategy::BinaryArith_SpecializedOnBaselineTypes); + + MIRType specialization = inspector->expectedBinaryArithSpecialization(pc); + if (specialization == MIRType::None) { + trackOptimizationOutcome(TrackedOutcome::NoTypeInfo); + return true; + } + + MDefinition::Opcode def_op = JSOpToMDefinition(op); + MBinaryArithInstruction* ins = MBinaryArithInstruction::New(alloc(), def_op, left, right); + ins->setSpecialization(specialization); + + current->add(ins); + current->push(ins); + + MOZ_ASSERT(!ins->isEffectful()); + if (!maybeInsertResume()) + return false; + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +bool +IonBuilder::arithTrySharedStub(bool* emitted, JSOp op, + MDefinition* left, MDefinition* right) +{ + MOZ_ASSERT(*emitted == false); + JSOp actualOp = JSOp(*pc); + + // Try to emit a shared stub cache. + + if (JitOptions.disableSharedStubs) + return true; + + // The actual jsop 'jsop_pos' is not supported yet. + if (actualOp == JSOP_POS) + return true; + + // FIXME: The JSOP_BITNOT path doesn't track optimizations yet. + if (actualOp != JSOP_BITNOT) { + trackOptimizationAttempt(TrackedStrategy::BinaryArith_SharedCache); + trackOptimizationSuccess(); + } + + MInstruction* stub = nullptr; + switch (actualOp) { + case JSOP_NEG: + case JSOP_BITNOT: + MOZ_ASSERT_IF(op == JSOP_MUL, + left->maybeConstantValue() && left->maybeConstantValue()->toInt32() == -1); + MOZ_ASSERT_IF(op != JSOP_MUL, !left); + + stub = MUnarySharedStub::New(alloc(), right); + break; + case JSOP_ADD: + case JSOP_SUB: + case JSOP_MUL: + case JSOP_DIV: + case JSOP_MOD: + case JSOP_POW: + stub = MBinarySharedStub::New(alloc(), left, right); + break; + default: + MOZ_CRASH("unsupported arith"); + } + + current->add(stub); + current->push(stub); + + // Decrease type from 'any type' to 'empty type' when one of the operands + // is 'empty typed'. + maybeMarkEmpty(stub); + + if (!resumeAfter(stub)) + return false; + + *emitted = true; + return true; +} + +bool +IonBuilder::jsop_binary_arith(JSOp op, MDefinition* left, MDefinition* right) +{ + bool emitted = false; + + startTrackingOptimizations(); + + trackTypeInfo(TrackedTypeSite::Operand, left->type(), left->resultTypeSet()); + trackTypeInfo(TrackedTypeSite::Operand, right->type(), right->resultTypeSet()); + + if (!forceInlineCaches()) { + if (!binaryArithTryConcat(&emitted, op, left, right) || emitted) + return emitted; + + if (!binaryArithTrySpecialized(&emitted, op, left, right) || emitted) + return emitted; + + if (!binaryArithTrySpecializedOnBaselineInspector(&emitted, op, left, right) || emitted) + return emitted; + } + + if (!arithTrySharedStub(&emitted, op, left, right) || emitted) + return emitted; + + // Not possible to optimize. Do a slow vm call. + trackOptimizationAttempt(TrackedStrategy::BinaryArith_Call); + trackOptimizationSuccess(); + + MDefinition::Opcode def_op = JSOpToMDefinition(op); + MBinaryArithInstruction* ins = MBinaryArithInstruction::New(alloc(), def_op, left, right); + + // Decrease type from 'any type' to 'empty type' when one of the operands + // is 'empty typed'. + maybeMarkEmpty(ins); + + current->add(ins); + current->push(ins); + MOZ_ASSERT(ins->isEffectful()); + return resumeAfter(ins); +} + +bool +IonBuilder::jsop_binary_arith(JSOp op) +{ + MDefinition* right = current->pop(); + MDefinition* left = current->pop(); + + return jsop_binary_arith(op, left, right); +} + + +bool +IonBuilder::jsop_pow() +{ + MDefinition* exponent = current->pop(); + MDefinition* base = current->pop(); + + bool emitted = false; + + if (!forceInlineCaches()) { + if (!powTrySpecialized(&emitted, base, exponent, MIRType::Double) || emitted) + return emitted; + } + + if (!arithTrySharedStub(&emitted, JSOP_POW, base, exponent) || emitted) + return emitted; + + // For now, use MIRType::Double, as a safe cover-all. See bug 1188079. + MPow* pow = MPow::New(alloc(), base, exponent, MIRType::Double); + current->add(pow); + current->push(pow); + return true; +} + +bool +IonBuilder::jsop_pos() +{ + if (IsNumberType(current->peek(-1)->type())) { + // Already int32 or double. Set the operand as implicitly used so it + // doesn't get optimized out if it has no other uses, as we could bail + // out. + current->peek(-1)->setImplicitlyUsedUnchecked(); + return true; + } + + // Compile +x as x * 1. + MDefinition* value = current->pop(); + MConstant* one = MConstant::New(alloc(), Int32Value(1)); + current->add(one); + + return jsop_binary_arith(JSOP_MUL, value, one); +} + +bool +IonBuilder::jsop_neg() +{ + // Since JSOP_NEG does not use a slot, we cannot push the MConstant. + // The MConstant is therefore passed to JSOP_MUL without slot traffic. + MConstant* negator = MConstant::New(alloc(), Int32Value(-1)); + current->add(negator); + + MDefinition* right = current->pop(); + + return jsop_binary_arith(JSOP_MUL, negator, right); +} + +bool +IonBuilder::jsop_tostring() +{ + if (current->peek(-1)->type() == MIRType::String) + return true; + + MDefinition* value = current->pop(); + MToString* ins = MToString::New(alloc(), value); + current->add(ins); + current->push(ins); + MOZ_ASSERT(!ins->isEffectful()); + return true; +} + +class AutoAccumulateReturns +{ + MIRGraph& graph_; + MIRGraphReturns* prev_; + + public: + AutoAccumulateReturns(MIRGraph& graph, MIRGraphReturns& returns) + : graph_(graph) + { + prev_ = graph_.returnAccumulator(); + graph_.setReturnAccumulator(&returns); + } + ~AutoAccumulateReturns() { + graph_.setReturnAccumulator(prev_); + } +}; + +IonBuilder::InliningStatus +IonBuilder::inlineScriptedCall(CallInfo& callInfo, JSFunction* target) +{ + MOZ_ASSERT(target->hasScript()); + MOZ_ASSERT(IsIonInlinablePC(pc)); + + MBasicBlock::BackupPoint backup(current); + if (!backup.init(alloc())) + return InliningStatus_Error; + + callInfo.setImplicitlyUsedUnchecked(); + + // Ensure sufficient space in the slots: needed for inlining from FUNAPPLY. + uint32_t depth = current->stackDepth() + callInfo.numFormals(); + if (depth > current->nslots()) { + if (!current->increaseSlots(depth - current->nslots())) + return InliningStatus_Error; + } + + // Create new |this| on the caller-side for inlined constructors. + if (callInfo.constructing()) { + MDefinition* thisDefn = createThis(target, callInfo.fun(), callInfo.getNewTarget()); + if (!thisDefn) + return InliningStatus_Error; + callInfo.setThis(thisDefn); + } + + // Capture formals in the outer resume point. + callInfo.pushFormals(current); + + MResumePoint* outerResumePoint = + MResumePoint::New(alloc(), current, pc, MResumePoint::Outer); + if (!outerResumePoint) + return InliningStatus_Error; + current->setOuterResumePoint(outerResumePoint); + + // Pop formals again, except leave |fun| on stack for duration of call. + callInfo.popFormals(current); + current->push(callInfo.fun()); + + JSScript* calleeScript = target->nonLazyScript(); + BaselineInspector inspector(calleeScript); + + // Improve type information of |this| when not set. + if (callInfo.constructing() && + !callInfo.thisArg()->resultTypeSet()) + { + StackTypeSet* types = TypeScript::ThisTypes(calleeScript); + if (types && !types->unknown()) { + TemporaryTypeSet* clonedTypes = types->clone(alloc_->lifoAlloc()); + if (!clonedTypes) + return InliningStatus_Error; + MTypeBarrier* barrier = MTypeBarrier::New(alloc(), callInfo.thisArg(), clonedTypes); + current->add(barrier); + if (barrier->type() == MIRType::Undefined) + callInfo.setThis(constant(UndefinedValue())); + else if (barrier->type() == MIRType::Null) + callInfo.setThis(constant(NullValue())); + else + callInfo.setThis(barrier); + } + } + + // Start inlining. + LifoAlloc* lifoAlloc = alloc_->lifoAlloc(); + InlineScriptTree* inlineScriptTree = + info().inlineScriptTree()->addCallee(alloc_, pc, calleeScript); + if (!inlineScriptTree) + return InliningStatus_Error; + CompileInfo* info = lifoAlloc->new_<CompileInfo>(calleeScript, target, + (jsbytecode*)nullptr, + this->info().analysisMode(), + /* needsArgsObj = */ false, + inlineScriptTree); + if (!info) + return InliningStatus_Error; + + MIRGraphReturns returns(alloc()); + AutoAccumulateReturns aar(graph(), returns); + + // Build the graph. + IonBuilder inlineBuilder(analysisContext, compartment, options, &alloc(), &graph(), constraints(), + &inspector, info, &optimizationInfo(), nullptr, inliningDepth_ + 1, + loopDepth_); + if (!inlineBuilder.buildInline(this, outerResumePoint, callInfo)) { + if (analysisContext && analysisContext->isExceptionPending()) { + JitSpew(JitSpew_IonAbort, "Inline builder raised exception."); + abortReason_ = AbortReason_Error; + return InliningStatus_Error; + } + + // Inlining the callee failed. Mark the callee as uninlineable only if + // the inlining was aborted for a non-exception reason. + if (inlineBuilder.abortReason_ == AbortReason_Disable) { + calleeScript->setUninlineable(); + if (!JitOptions.disableInlineBacktracking) { + current = backup.restore(); + return InliningStatus_NotInlined; + } + abortReason_ = AbortReason_Inlining; + } else if (inlineBuilder.abortReason_ == AbortReason_Inlining) { + abortReason_ = AbortReason_Inlining; + } else if (inlineBuilder.abortReason_ == AbortReason_Alloc) { + abortReason_ = AbortReason_Alloc; + } else if (inlineBuilder.abortReason_ == AbortReason_PreliminaryObjects) { + const ObjectGroupVector& groups = inlineBuilder.abortedPreliminaryGroups(); + MOZ_ASSERT(!groups.empty()); + for (size_t i = 0; i < groups.length(); i++) + addAbortedPreliminaryGroup(groups[i]); + abortReason_ = AbortReason_PreliminaryObjects; + } + + return InliningStatus_Error; + } + + // Create return block. + jsbytecode* postCall = GetNextPc(pc); + MBasicBlock* returnBlock = newBlock(nullptr, postCall); + if (!returnBlock) + return InliningStatus_Error; + returnBlock->setCallerResumePoint(callerResumePoint_); + + // Inherit the slots from current and pop |fun|. + returnBlock->inheritSlots(current); + returnBlock->pop(); + + // Accumulate return values. + if (returns.empty()) { + // Inlining of functions that have no exit is not supported. + calleeScript->setUninlineable(); + if (!JitOptions.disableInlineBacktracking) { + current = backup.restore(); + return InliningStatus_NotInlined; + } + abortReason_ = AbortReason_Inlining; + return InliningStatus_Error; + } + MDefinition* retvalDefn = patchInlinedReturns(callInfo, returns, returnBlock); + if (!retvalDefn) + return InliningStatus_Error; + returnBlock->push(retvalDefn); + + // Initialize entry slots now that the stack has been fixed up. + if (!returnBlock->initEntrySlots(alloc())) + return InliningStatus_Error; + + if (!setCurrentAndSpecializePhis(returnBlock)) + return InliningStatus_Error; + + return InliningStatus_Inlined; +} + +MDefinition* +IonBuilder::patchInlinedReturn(CallInfo& callInfo, MBasicBlock* exit, MBasicBlock* bottom) +{ + // Replaces the MReturn in the exit block with an MGoto. + MDefinition* rdef = exit->lastIns()->toReturn()->input(); + exit->discardLastIns(); + + // Constructors must be patched by the caller to always return an object. + if (callInfo.constructing()) { + if (rdef->type() == MIRType::Value) { + // Unknown return: dynamically detect objects. + MReturnFromCtor* filter = MReturnFromCtor::New(alloc(), rdef, callInfo.thisArg()); + exit->add(filter); + rdef = filter; + } else if (rdef->type() != MIRType::Object) { + // Known non-object return: force |this|. + rdef = callInfo.thisArg(); + } + } else if (callInfo.isSetter()) { + // Setters return their argument, not whatever value is returned. + rdef = callInfo.getArg(0); + } + + if (!callInfo.isSetter()) + rdef = specializeInlinedReturn(rdef, exit); + + MGoto* replacement = MGoto::New(alloc(), bottom); + exit->end(replacement); + if (!bottom->addPredecessorWithoutPhis(exit)) + return nullptr; + + return rdef; +} + +MDefinition* +IonBuilder::specializeInlinedReturn(MDefinition* rdef, MBasicBlock* exit) +{ + // Remove types from the return definition that weren't observed. + TemporaryTypeSet* types = bytecodeTypes(pc); + + // The observed typeset doesn't contain extra information. + if (types->empty() || types->unknown()) + return rdef; + + // Decide if specializing is needed using the result typeset if available, + // else use the result type. + + if (rdef->resultTypeSet()) { + // Don't specialize if return typeset is a subset of the + // observed typeset. The return typeset is already more specific. + if (rdef->resultTypeSet()->isSubset(types)) + return rdef; + } else { + MIRType observedType = types->getKnownMIRType(); + + // Don't specialize if type is MIRType::Float32 and TI reports + // MIRType::Double. Float is more specific than double. + if (observedType == MIRType::Double && rdef->type() == MIRType::Float32) + return rdef; + + // Don't specialize if types are inaccordance, except for MIRType::Value + // and MIRType::Object (when not unknown object), since the typeset + // contains more specific information. + if (observedType == rdef->type() && + observedType != MIRType::Value && + (observedType != MIRType::Object || types->unknownObject())) + { + return rdef; + } + } + + setCurrent(exit); + + MTypeBarrier* barrier = nullptr; + rdef = addTypeBarrier(rdef, types, BarrierKind::TypeSet, &barrier); + if (barrier) + barrier->setNotMovable(); + + return rdef; +} + +MDefinition* +IonBuilder::patchInlinedReturns(CallInfo& callInfo, MIRGraphReturns& returns, MBasicBlock* bottom) +{ + // Replaces MReturns with MGotos, returning the MDefinition + // representing the return value, or nullptr. + MOZ_ASSERT(returns.length() > 0); + + if (returns.length() == 1) + return patchInlinedReturn(callInfo, returns[0], bottom); + + // Accumulate multiple returns with a phi. + MPhi* phi = MPhi::New(alloc()); + if (!phi->reserveLength(returns.length())) + return nullptr; + + for (size_t i = 0; i < returns.length(); i++) { + MDefinition* rdef = patchInlinedReturn(callInfo, returns[i], bottom); + if (!rdef) + return nullptr; + phi->addInput(rdef); + } + + bottom->addPhi(phi); + return phi; +} + +IonBuilder::InliningDecision +IonBuilder::makeInliningDecision(JSObject* targetArg, CallInfo& callInfo) +{ + // When there is no target, inlining is impossible. + if (targetArg == nullptr) { + trackOptimizationOutcome(TrackedOutcome::CantInlineNoTarget); + return InliningDecision_DontInline; + } + + // Inlining non-function targets is handled by inlineNonFunctionCall(). + if (!targetArg->is<JSFunction>()) + return InliningDecision_Inline; + + JSFunction* target = &targetArg->as<JSFunction>(); + + // Never inline during the arguments usage analysis. + if (info().analysisMode() == Analysis_ArgumentsUsage) + return InliningDecision_DontInline; + + // Native functions provide their own detection in inlineNativeCall(). + if (target->isNative()) + return InliningDecision_Inline; + + // Determine whether inlining is possible at callee site + InliningDecision decision = canInlineTarget(target, callInfo); + if (decision != InliningDecision_Inline) + return decision; + + // Heuristics! + JSScript* targetScript = target->nonLazyScript(); + + // Callee must not be excessively large. + // This heuristic also applies to the callsite as a whole. + bool offThread = options.offThreadCompilationAvailable(); + if (targetScript->length() > optimizationInfo().inlineMaxBytecodePerCallSite(offThread)) { + trackOptimizationOutcome(TrackedOutcome::CantInlineBigCallee); + return DontInline(targetScript, "Vetoed: callee excessively large"); + } + + // Callee must have been called a few times to have somewhat stable + // type information, except for definite properties analysis, + // as the caller has not run yet. + if (targetScript->getWarmUpCount() < optimizationInfo().inliningWarmUpThreshold() && + !targetScript->baselineScript()->ionCompiledOrInlined() && + info().analysisMode() != Analysis_DefiniteProperties) + { + trackOptimizationOutcome(TrackedOutcome::CantInlineNotHot); + JitSpew(JitSpew_Inlining, "Cannot inline %s:%" PRIuSIZE ": callee is insufficiently hot.", + targetScript->filename(), targetScript->lineno()); + return InliningDecision_WarmUpCountTooLow; + } + + // Don't inline if the callee is known to inline a lot of code, to avoid + // huge MIR graphs. + uint32_t inlinedBytecodeLength = targetScript->baselineScript()->inlinedBytecodeLength(); + if (inlinedBytecodeLength > optimizationInfo().inlineMaxCalleeInlinedBytecodeLength()) { + trackOptimizationOutcome(TrackedOutcome::CantInlineBigCalleeInlinedBytecodeLength); + return DontInline(targetScript, "Vetoed: callee inlinedBytecodeLength is too big"); + } + + IonBuilder* outerBuilder = outermostBuilder(); + + // Cap the total bytecode length we inline under a single script, to avoid + // excessive inlining in pathological cases. + size_t totalBytecodeLength = outerBuilder->inlinedBytecodeLength_ + targetScript->length(); + if (totalBytecodeLength > optimizationInfo().inlineMaxTotalBytecodeLength()) { + trackOptimizationOutcome(TrackedOutcome::CantInlineExceededTotalBytecodeLength); + return DontInline(targetScript, "Vetoed: exceeding max total bytecode length"); + } + + // Cap the inlining depth. + + uint32_t maxInlineDepth; + if (JitOptions.isSmallFunction(targetScript)) { + maxInlineDepth = optimizationInfo().smallFunctionMaxInlineDepth(); + } else { + maxInlineDepth = optimizationInfo().maxInlineDepth(); + + // Caller must not be excessively large. + if (script()->length() >= optimizationInfo().inliningMaxCallerBytecodeLength()) { + trackOptimizationOutcome(TrackedOutcome::CantInlineBigCaller); + return DontInline(targetScript, "Vetoed: caller excessively large"); + } + } + + BaselineScript* outerBaseline = outermostBuilder()->script()->baselineScript(); + if (inliningDepth_ >= maxInlineDepth) { + // We hit the depth limit and won't inline this function. Give the + // outermost script a max inlining depth of 0, so that it won't be + // inlined in other scripts. This heuristic is currently only used + // when we're inlining scripts with loops, see the comment below. + outerBaseline->setMaxInliningDepth(0); + + trackOptimizationOutcome(TrackedOutcome::CantInlineExceededDepth); + return DontInline(targetScript, "Vetoed: exceeding allowed inline depth"); + } + + // Inlining functions with loops can be complicated. For instance, if we're + // close to the inlining depth limit and we inline the function f below, we + // can no longer inline the call to g: + // + // function f() { + // while (cond) { + // g(); + // } + // } + // + // If the loop has many iterations, it's more efficient to call f and inline + // g in f. + // + // To avoid this problem, we record a separate max inlining depth for each + // script, indicating at which depth we won't be able to inline all functions + // we inlined this time. This solves the issue above, because we will only + // inline f if it means we can also inline g. + if (targetScript->hasLoops() && + inliningDepth_ >= targetScript->baselineScript()->maxInliningDepth()) + { + trackOptimizationOutcome(TrackedOutcome::CantInlineExceededDepth); + return DontInline(targetScript, "Vetoed: exceeding allowed script inline depth"); + } + + // Update the max depth at which we can inline the outer script. + MOZ_ASSERT(maxInlineDepth > inliningDepth_); + uint32_t scriptInlineDepth = maxInlineDepth - inliningDepth_ - 1; + if (scriptInlineDepth < outerBaseline->maxInliningDepth()) + outerBaseline->setMaxInliningDepth(scriptInlineDepth); + + // End of heuristics, we will inline this function. + + // TI calls ObjectStateChange to trigger invalidation of the caller. + TypeSet::ObjectKey* targetKey = TypeSet::ObjectKey::get(target); + targetKey->watchStateChangeForInlinedCall(constraints()); + + outerBuilder->inlinedBytecodeLength_ += targetScript->length(); + + return InliningDecision_Inline; +} + +bool +IonBuilder::selectInliningTargets(const ObjectVector& targets, CallInfo& callInfo, BoolVector& choiceSet, + uint32_t* numInlineable) +{ + *numInlineable = 0; + uint32_t totalSize = 0; + + // For each target, ask whether it may be inlined. + if (!choiceSet.reserve(targets.length())) + return false; + + // Don't inline polymorphic sites during the definite properties analysis. + // AddClearDefiniteFunctionUsesInScript depends on this for correctness. + if (info().analysisMode() == Analysis_DefiniteProperties && targets.length() > 1) + return true; + + for (size_t i = 0; i < targets.length(); i++) { + JSObject* target = targets[i]; + + trackOptimizationAttempt(TrackedStrategy::Call_Inline); + trackTypeInfo(TrackedTypeSite::Call_Target, target); + + bool inlineable; + InliningDecision decision = makeInliningDecision(target, callInfo); + switch (decision) { + case InliningDecision_Error: + return false; + case InliningDecision_DontInline: + case InliningDecision_WarmUpCountTooLow: + inlineable = false; + break; + case InliningDecision_Inline: + inlineable = true; + break; + default: + MOZ_CRASH("Unhandled InliningDecision value!"); + } + + if (target->is<JSFunction>()) { + // Enforce a maximum inlined bytecode limit at the callsite. + if (inlineable && target->as<JSFunction>().isInterpreted()) { + totalSize += target->as<JSFunction>().nonLazyScript()->length(); + bool offThread = options.offThreadCompilationAvailable(); + if (totalSize > optimizationInfo().inlineMaxBytecodePerCallSite(offThread)) + inlineable = false; + } + } else { + // Non-function targets are not supported by polymorphic inlining. + inlineable = false; + } + + choiceSet.infallibleAppend(inlineable); + if (inlineable) + *numInlineable += 1; + } + + // If optimization tracking is turned on and one of the inlineable targets + // is a native, track the type info of the call. Most native inlinings + // depend on the types of the arguments and the return value. + if (isOptimizationTrackingEnabled()) { + for (size_t i = 0; i < targets.length(); i++) { + if (choiceSet[i] && targets[i]->as<JSFunction>().isNative()) { + trackTypeInfo(callInfo); + break; + } + } + } + + MOZ_ASSERT(choiceSet.length() == targets.length()); + return true; +} + +static bool +CanInlineGetPropertyCache(MGetPropertyCache* cache, MDefinition* thisDef) +{ + MOZ_ASSERT(cache->object()->type() == MIRType::Object); + if (cache->object() != thisDef) + return false; + + InlinePropertyTable* table = cache->propTable(); + if (!table) + return false; + if (table->numEntries() == 0) + return false; + return true; +} + +class WrapMGetPropertyCache +{ + MGetPropertyCache* cache_; + + private: + void discardPriorResumePoint() { + if (!cache_) + return; + + InlinePropertyTable* propTable = cache_->propTable(); + if (!propTable) + return; + MResumePoint* rp = propTable->takePriorResumePoint(); + if (!rp) + return; + cache_->block()->discardPreAllocatedResumePoint(rp); + } + + public: + explicit WrapMGetPropertyCache(MGetPropertyCache* cache) + : cache_(cache) + { } + + ~WrapMGetPropertyCache() { + discardPriorResumePoint(); + } + + MGetPropertyCache* get() { + return cache_; + } + MGetPropertyCache* operator->() { + return get(); + } + + // This function returns the cache given to the constructor if the + // GetPropertyCache can be moved into the ObjectGroup fallback path. + MGetPropertyCache* moveableCache(bool hasTypeBarrier, MDefinition* thisDef) { + // If we have unhandled uses of the MGetPropertyCache, then we cannot + // move it to the ObjectGroup fallback path. + if (!hasTypeBarrier) { + if (cache_->hasUses()) + return nullptr; + } else { + // There is the TypeBarrier consumer, so we check that this is the + // only consumer. + MOZ_ASSERT(cache_->hasUses()); + if (!cache_->hasOneUse()) + return nullptr; + } + + // If the this-object is not identical to the object of the + // MGetPropertyCache, then we cannot use the InlinePropertyTable, or if + // we do not yet have enough information from the ObjectGroup. + if (!CanInlineGetPropertyCache(cache_, thisDef)) + return nullptr; + + MGetPropertyCache* ret = cache_; + cache_ = nullptr; + return ret; + } +}; + +MGetPropertyCache* +IonBuilder::getInlineableGetPropertyCache(CallInfo& callInfo) +{ + if (callInfo.constructing()) + return nullptr; + + MDefinition* thisDef = callInfo.thisArg(); + if (thisDef->type() != MIRType::Object) + return nullptr; + + MDefinition* funcDef = callInfo.fun(); + if (funcDef->type() != MIRType::Object) + return nullptr; + + // MGetPropertyCache with no uses may be optimized away. + if (funcDef->isGetPropertyCache()) { + WrapMGetPropertyCache cache(funcDef->toGetPropertyCache()); + return cache.moveableCache(/* hasTypeBarrier = */ false, thisDef); + } + + // Optimize away the following common pattern: + // MTypeBarrier[MIRType::Object] <- MGetPropertyCache + if (funcDef->isTypeBarrier()) { + MTypeBarrier* barrier = funcDef->toTypeBarrier(); + if (barrier->hasUses()) + return nullptr; + if (barrier->type() != MIRType::Object) + return nullptr; + if (!barrier->input()->isGetPropertyCache()) + return nullptr; + + WrapMGetPropertyCache cache(barrier->input()->toGetPropertyCache()); + return cache.moveableCache(/* hasTypeBarrier = */ true, thisDef); + } + + return nullptr; +} + +IonBuilder::InliningStatus +IonBuilder::inlineSingleCall(CallInfo& callInfo, JSObject* targetArg) +{ + if (!targetArg->is<JSFunction>()) { + InliningStatus status = inlineNonFunctionCall(callInfo, targetArg); + trackInlineSuccess(status); + return status; + } + + JSFunction* target = &targetArg->as<JSFunction>(); + if (target->isNative()) { + InliningStatus status = inlineNativeCall(callInfo, target); + trackInlineSuccess(status); + return status; + } + + // Track success now, as inlining a scripted call makes a new return block + // which has a different pc than the current call pc. + trackInlineSuccess(); + return inlineScriptedCall(callInfo, target); +} + +IonBuilder::InliningStatus +IonBuilder::inlineCallsite(const ObjectVector& targets, CallInfo& callInfo) +{ + if (targets.empty()) { + trackOptimizationAttempt(TrackedStrategy::Call_Inline); + trackOptimizationOutcome(TrackedOutcome::CantInlineNoTarget); + return InliningStatus_NotInlined; + } + + // Is the function provided by an MGetPropertyCache? + // If so, the cache may be movable to a fallback path, with a dispatch + // instruction guarding on the incoming ObjectGroup. + WrapMGetPropertyCache propCache(getInlineableGetPropertyCache(callInfo)); + keepFallbackFunctionGetter(propCache.get()); + + // Inline single targets -- unless they derive from a cache, in which case + // avoiding the cache and guarding is still faster. + if (!propCache.get() && targets.length() == 1) { + JSObject* target = targets[0]; + + trackOptimizationAttempt(TrackedStrategy::Call_Inline); + trackTypeInfo(TrackedTypeSite::Call_Target, target); + + InliningDecision decision = makeInliningDecision(target, callInfo); + switch (decision) { + case InliningDecision_Error: + return InliningStatus_Error; + case InliningDecision_DontInline: + return InliningStatus_NotInlined; + case InliningDecision_WarmUpCountTooLow: + return InliningStatus_WarmUpCountTooLow; + case InliningDecision_Inline: + break; + } + + // Inlining will elminate uses of the original callee, but it needs to + // be preserved in phis if we bail out. Mark the old callee definition as + // implicitly used to ensure this happens. + callInfo.fun()->setImplicitlyUsedUnchecked(); + + // If the callee is not going to be a lambda (which may vary across + // different invocations), then the callee definition can be replaced by a + // constant. + if (target->isSingleton()) { + // Replace the function with an MConstant. + MConstant* constFun = constant(ObjectValue(*target)); + if (callInfo.constructing() && callInfo.getNewTarget() == callInfo.fun()) + callInfo.setNewTarget(constFun); + callInfo.setFun(constFun); + } + + return inlineSingleCall(callInfo, target); + } + + // Choose a subset of the targets for polymorphic inlining. + BoolVector choiceSet(alloc()); + uint32_t numInlined; + if (!selectInliningTargets(targets, callInfo, choiceSet, &numInlined)) + return InliningStatus_Error; + if (numInlined == 0) + return InliningStatus_NotInlined; + + // Perform a polymorphic dispatch. + if (!inlineCalls(callInfo, targets, choiceSet, propCache.get())) + return InliningStatus_Error; + + return InliningStatus_Inlined; +} + +bool +IonBuilder::inlineGenericFallback(JSFunction* target, CallInfo& callInfo, MBasicBlock* dispatchBlock) +{ + // Generate a new block with all arguments on-stack. + MBasicBlock* fallbackBlock = newBlock(dispatchBlock, pc); + if (!fallbackBlock) + return false; + + // Create a new CallInfo to track modified state within this block. + CallInfo fallbackInfo(alloc(), callInfo.constructing()); + if (!fallbackInfo.init(callInfo)) + return false; + fallbackInfo.popFormals(fallbackBlock); + + // Generate an MCall, which uses stateful |current|. + if (!setCurrentAndSpecializePhis(fallbackBlock)) + return false; + if (!makeCall(target, fallbackInfo)) + return false; + + // Pass return block to caller as |current|. + return true; +} + +bool +IonBuilder::inlineObjectGroupFallback(CallInfo& callInfo, MBasicBlock* dispatchBlock, + MObjectGroupDispatch* dispatch, MGetPropertyCache* cache, + MBasicBlock** fallbackTarget) +{ + // Getting here implies the following: + // 1. The call function is an MGetPropertyCache, or an MGetPropertyCache + // followed by an MTypeBarrier. + MOZ_ASSERT(callInfo.fun()->isGetPropertyCache() || callInfo.fun()->isTypeBarrier()); + + // 2. The MGetPropertyCache has inlineable cases by guarding on the ObjectGroup. + MOZ_ASSERT(dispatch->numCases() > 0); + + // 3. The MGetPropertyCache (and, if applicable, MTypeBarrier) only + // have at most a single use. + MOZ_ASSERT_IF(callInfo.fun()->isGetPropertyCache(), !cache->hasUses()); + MOZ_ASSERT_IF(callInfo.fun()->isTypeBarrier(), cache->hasOneUse()); + + // This means that no resume points yet capture the MGetPropertyCache, + // so everything from the MGetPropertyCache up until the call is movable. + // We now move the MGetPropertyCache and friends into a fallback path. + MOZ_ASSERT(cache->idempotent()); + + // Create a new CallInfo to track modified state within the fallback path. + CallInfo fallbackInfo(alloc(), callInfo.constructing()); + if (!fallbackInfo.init(callInfo)) + return false; + + // Capture stack prior to the call operation. This captures the function. + MResumePoint* preCallResumePoint = + MResumePoint::New(alloc(), dispatchBlock, pc, MResumePoint::ResumeAt); + if (!preCallResumePoint) + return false; + + DebugOnly<size_t> preCallFuncIndex = preCallResumePoint->stackDepth() - callInfo.numFormals(); + MOZ_ASSERT(preCallResumePoint->getOperand(preCallFuncIndex) == fallbackInfo.fun()); + + // In the dispatch block, replace the function's slot entry with Undefined. + MConstant* undefined = MConstant::New(alloc(), UndefinedValue()); + dispatchBlock->add(undefined); + dispatchBlock->rewriteAtDepth(-int(callInfo.numFormals()), undefined); + + // Construct a block that does nothing but remove formals from the stack. + // This is effectively changing the entry resume point of the later fallback block. + MBasicBlock* prepBlock = newBlock(dispatchBlock, pc); + if (!prepBlock) + return false; + fallbackInfo.popFormals(prepBlock); + + // Construct a block into which the MGetPropertyCache can be moved. + // This is subtle: the pc and resume point are those of the MGetPropertyCache! + InlinePropertyTable* propTable = cache->propTable(); + MResumePoint* priorResumePoint = propTable->takePriorResumePoint(); + MOZ_ASSERT(propTable->pc() != nullptr); + MOZ_ASSERT(priorResumePoint != nullptr); + MBasicBlock* getPropBlock = newBlock(prepBlock, propTable->pc(), priorResumePoint); + if (!getPropBlock) + return false; + + prepBlock->end(MGoto::New(alloc(), getPropBlock)); + + // Since the getPropBlock inherited the stack from right before the MGetPropertyCache, + // the target of the MGetPropertyCache is still on the stack. + DebugOnly<MDefinition*> checkObject = getPropBlock->pop(); + MOZ_ASSERT(checkObject == cache->object()); + + // Move the MGetPropertyCache and friends into the getPropBlock. + if (fallbackInfo.fun()->isGetPropertyCache()) { + MOZ_ASSERT(fallbackInfo.fun()->toGetPropertyCache() == cache); + getPropBlock->addFromElsewhere(cache); + getPropBlock->push(cache); + } else { + MTypeBarrier* barrier = callInfo.fun()->toTypeBarrier(); + MOZ_ASSERT(barrier->type() == MIRType::Object); + MOZ_ASSERT(barrier->input()->isGetPropertyCache()); + MOZ_ASSERT(barrier->input()->toGetPropertyCache() == cache); + + getPropBlock->addFromElsewhere(cache); + getPropBlock->addFromElsewhere(barrier); + getPropBlock->push(barrier); + } + + // Construct an end block with the correct resume point. + MBasicBlock* preCallBlock = newBlock(getPropBlock, pc, preCallResumePoint); + if (!preCallBlock) + return false; + getPropBlock->end(MGoto::New(alloc(), preCallBlock)); + + // Now inline the MCallGeneric, using preCallBlock as the dispatch point. + if (!inlineGenericFallback(nullptr, fallbackInfo, preCallBlock)) + return false; + + // inlineGenericFallback() set the return block as |current|. + preCallBlock->end(MGoto::New(alloc(), current)); + *fallbackTarget = prepBlock; + return true; +} + +bool +IonBuilder::inlineCalls(CallInfo& callInfo, const ObjectVector& targets, BoolVector& choiceSet, + MGetPropertyCache* maybeCache) +{ + // Only handle polymorphic inlining. + MOZ_ASSERT(IsIonInlinablePC(pc)); + MOZ_ASSERT(choiceSet.length() == targets.length()); + MOZ_ASSERT_IF(!maybeCache, targets.length() >= 2); + MOZ_ASSERT_IF(maybeCache, targets.length() >= 1); + + MBasicBlock* dispatchBlock = current; + callInfo.setImplicitlyUsedUnchecked(); + callInfo.pushFormals(dispatchBlock); + + // Patch any InlinePropertyTable to only contain functions that are + // inlineable. The InlinePropertyTable will also be patched at the end to + // exclude native functions that vetoed inlining. + if (maybeCache) { + InlinePropertyTable* propTable = maybeCache->propTable(); + propTable->trimToTargets(targets); + if (propTable->numEntries() == 0) + maybeCache = nullptr; + } + + // Generate a dispatch based on guard kind. + MDispatchInstruction* dispatch; + if (maybeCache) { + dispatch = MObjectGroupDispatch::New(alloc(), maybeCache->object(), maybeCache->propTable()); + callInfo.fun()->setImplicitlyUsedUnchecked(); + } else { + dispatch = MFunctionDispatch::New(alloc(), callInfo.fun()); + } + + // Generate a return block to host the rval-collecting MPhi. + jsbytecode* postCall = GetNextPc(pc); + MBasicBlock* returnBlock = newBlock(nullptr, postCall); + if (!returnBlock) + return false; + returnBlock->setCallerResumePoint(callerResumePoint_); + + // Set up stack, used to manually create a post-call resume point. + returnBlock->inheritSlots(dispatchBlock); + callInfo.popFormals(returnBlock); + + MPhi* retPhi = MPhi::New(alloc()); + returnBlock->addPhi(retPhi); + returnBlock->push(retPhi); + + // Create a resume point from current stack state. + if (!returnBlock->initEntrySlots(alloc())) + return false; + + // Reserve the capacity for the phi. + // Note: this is an upperbound. Unreachable targets and uninlineable natives are also counted. + uint32_t count = 1; // Possible fallback block. + for (uint32_t i = 0; i < targets.length(); i++) { + if (choiceSet[i]) + count++; + } + if (!retPhi->reserveLength(count)) + return false; + + // Inline each of the inlineable targets. + for (uint32_t i = 0; i < targets.length(); i++) { + // Target must be inlineable. + if (!choiceSet[i]) + continue; + + // Even though we made one round of inline decisions already, we may + // be amending them below. + amendOptimizationAttempt(i); + + // Target must be reachable by the MDispatchInstruction. + JSFunction* target = &targets[i]->as<JSFunction>(); + if (maybeCache && !maybeCache->propTable()->hasFunction(target)) { + choiceSet[i] = false; + trackOptimizationOutcome(TrackedOutcome::CantInlineNotInDispatch); + continue; + } + + MBasicBlock* inlineBlock = newBlock(dispatchBlock, pc); + if (!inlineBlock) + return false; + + // Create a function MConstant to use in the entry ResumePoint. If we + // can't use a constant, add a no-op MPolyInlineGuard, to prevent + // hoisting env chain gets above the dispatch instruction. + MInstruction* funcDef; + if (target->isSingleton()) + funcDef = MConstant::New(alloc(), ObjectValue(*target), constraints()); + else + funcDef = MPolyInlineGuard::New(alloc(), callInfo.fun()); + + funcDef->setImplicitlyUsedUnchecked(); + dispatchBlock->add(funcDef); + + // Use the inlined callee in the inline resume point and on stack. + int funIndex = inlineBlock->entryResumePoint()->stackDepth() - callInfo.numFormals(); + inlineBlock->entryResumePoint()->replaceOperand(funIndex, funcDef); + inlineBlock->rewriteSlot(funIndex, funcDef); + + // Create a new CallInfo to track modified state within the inline block. + CallInfo inlineInfo(alloc(), callInfo.constructing()); + if (!inlineInfo.init(callInfo)) + return false; + inlineInfo.popFormals(inlineBlock); + inlineInfo.setFun(funcDef); + + if (maybeCache) { + // Assign the 'this' value a TypeSet specialized to the groups that + // can generate this inlining target. + MOZ_ASSERT(callInfo.thisArg() == maybeCache->object()); + TemporaryTypeSet* thisTypes = maybeCache->propTable()->buildTypeSetForFunction(target); + if (!thisTypes) + return false; + + MFilterTypeSet* filter = MFilterTypeSet::New(alloc(), inlineInfo.thisArg(), thisTypes); + inlineBlock->add(filter); + inlineInfo.setThis(filter); + } + + // Inline the call into the inlineBlock. + if (!setCurrentAndSpecializePhis(inlineBlock)) + return false; + InliningStatus status = inlineSingleCall(inlineInfo, target); + if (status == InliningStatus_Error) + return false; + + // Natives may veto inlining. + if (status == InliningStatus_NotInlined) { + MOZ_ASSERT(current == inlineBlock); + graph().removeBlock(inlineBlock); + choiceSet[i] = false; + continue; + } + + // inlineSingleCall() changed |current| to the inline return block. + MBasicBlock* inlineReturnBlock = current; + setCurrent(dispatchBlock); + + // Connect the inline path to the returnBlock. + ObjectGroup* funcGroup = target->isSingleton() ? nullptr : target->group(); + if (!dispatch->addCase(target, funcGroup, inlineBlock)) + return false; + + MDefinition* retVal = inlineReturnBlock->peek(-1); + retPhi->addInput(retVal); + inlineReturnBlock->end(MGoto::New(alloc(), returnBlock)); + if (!returnBlock->addPredecessorWithoutPhis(inlineReturnBlock)) + return false; + } + + // Patch the InlinePropertyTable to not dispatch to vetoed paths. + bool useFallback; + if (maybeCache) { + InlinePropertyTable* propTable = maybeCache->propTable(); + propTable->trimTo(targets, choiceSet); + + if (propTable->numEntries() == 0 || !propTable->hasPriorResumePoint()) { + // Output a generic fallback path. + MOZ_ASSERT_IF(propTable->numEntries() == 0, dispatch->numCases() == 0); + maybeCache = nullptr; + useFallback = true; + } else { + // We need a fallback path if the ObjectGroup dispatch does not + // handle all incoming objects. + useFallback = false; + TemporaryTypeSet* objectTypes = maybeCache->object()->resultTypeSet(); + for (uint32_t i = 0; i < objectTypes->getObjectCount(); i++) { + TypeSet::ObjectKey* obj = objectTypes->getObject(i); + if (!obj) + continue; + + if (!obj->isGroup()) { + useFallback = true; + break; + } + + if (!propTable->hasObjectGroup(obj->group())) { + useFallback = true; + break; + } + } + + if (!useFallback) { + // The object group dispatch handles all possible incoming + // objects, so the cache and barrier will not be reached and + // can be eliminated. + if (callInfo.fun()->isGetPropertyCache()) { + MOZ_ASSERT(callInfo.fun() == maybeCache); + } else { + MTypeBarrier* barrier = callInfo.fun()->toTypeBarrier(); + MOZ_ASSERT(!barrier->hasUses()); + MOZ_ASSERT(barrier->type() == MIRType::Object); + MOZ_ASSERT(barrier->input()->isGetPropertyCache()); + MOZ_ASSERT(barrier->input()->toGetPropertyCache() == maybeCache); + barrier->block()->discard(barrier); + } + + MOZ_ASSERT(!maybeCache->hasUses()); + maybeCache->block()->discard(maybeCache); + } + } + } else { + useFallback = dispatch->numCases() < targets.length(); + } + + // If necessary, generate a fallback path. + if (useFallback) { + // Generate fallback blocks, and set |current| to the fallback return block. + if (maybeCache) { + MBasicBlock* fallbackTarget; + if (!inlineObjectGroupFallback(callInfo, dispatchBlock, + dispatch->toObjectGroupDispatch(), + maybeCache, &fallbackTarget)) + { + return false; + } + dispatch->addFallback(fallbackTarget); + } else { + JSFunction* remaining = nullptr; + + // If there is only 1 remaining case, we can annotate the fallback call + // with the target information. + if (dispatch->numCases() + 1 == targets.length()) { + for (uint32_t i = 0; i < targets.length(); i++) { + if (choiceSet[i]) + continue; + + MOZ_ASSERT(!remaining); + if (targets[i]->is<JSFunction>() && targets[i]->as<JSFunction>().isSingleton()) + remaining = &targets[i]->as<JSFunction>(); + break; + } + } + + if (!inlineGenericFallback(remaining, callInfo, dispatchBlock)) + return false; + dispatch->addFallback(current); + } + + MBasicBlock* fallbackReturnBlock = current; + + // Connect fallback case to return infrastructure. + MDefinition* retVal = fallbackReturnBlock->peek(-1); + retPhi->addInput(retVal); + fallbackReturnBlock->end(MGoto::New(alloc(), returnBlock)); + if (!returnBlock->addPredecessorWithoutPhis(fallbackReturnBlock)) + return false; + } + + // Finally add the dispatch instruction. + // This must be done at the end so that add() may be called above. + dispatchBlock->end(dispatch); + + // Check the depth change: +1 for retval + MOZ_ASSERT(returnBlock->stackDepth() == dispatchBlock->stackDepth() - callInfo.numFormals() + 1); + + graph().moveBlockToEnd(returnBlock); + return setCurrentAndSpecializePhis(returnBlock); +} + +MInstruction* +IonBuilder::createNamedLambdaObject(MDefinition* callee, MDefinition* env) +{ + // Get a template CallObject that we'll use to generate inline object + // creation. + LexicalEnvironmentObject* templateObj = inspector->templateNamedLambdaObject(); + + // One field is added to the function to handle its name. This cannot be a + // dynamic slot because there is still plenty of room on the NamedLambda object. + MOZ_ASSERT(!templateObj->hasDynamicSlots()); + + // Allocate the actual object. It is important that no intervening + // instructions could potentially bailout, thus leaking the dynamic slots + // pointer. + MInstruction* declEnvObj = MNewNamedLambdaObject::New(alloc(), templateObj); + current->add(declEnvObj); + + // Initialize the object's reserved slots. No post barrier is needed here: + // the object will be allocated in the nursery if possible, and if the + // tenured heap is used instead, a minor collection will have been performed + // that moved env/callee to the tenured heap. + current->add(MStoreFixedSlot::New(alloc(), declEnvObj, + NamedLambdaObject::enclosingEnvironmentSlot(), env)); + current->add(MStoreFixedSlot::New(alloc(), declEnvObj, + NamedLambdaObject::lambdaSlot(), callee)); + + return declEnvObj; +} + +MInstruction* +IonBuilder::createCallObject(MDefinition* callee, MDefinition* env) +{ + // Get a template CallObject that we'll use to generate inline object + // creation. + CallObject* templateObj = inspector->templateCallObject(); + + // Allocate the object. Run-once scripts need a singleton type, so always do + // a VM call in such cases. + MNewCallObjectBase* callObj; + if (script()->treatAsRunOnce() || templateObj->isSingleton()) + callObj = MNewSingletonCallObject::New(alloc(), templateObj); + else + callObj = MNewCallObject::New(alloc(), templateObj); + current->add(callObj); + + // Initialize the object's reserved slots. No post barrier is needed here, + // for the same reason as in createNamedLambdaObject. + current->add(MStoreFixedSlot::New(alloc(), callObj, CallObject::enclosingEnvironmentSlot(), env)); + current->add(MStoreFixedSlot::New(alloc(), callObj, CallObject::calleeSlot(), callee)); + + //if (!script()->functionHasParameterExprs()) { + + // Copy closed-over argument slots if there aren't parameter expressions. + MSlots* slots = nullptr; + for (PositionalFormalParameterIter fi(script()); fi; fi++) { + if (!fi.closedOver()) + continue; + + if (!alloc().ensureBallast()) + return nullptr; + + unsigned slot = fi.location().slot(); + unsigned formal = fi.argumentSlot(); + unsigned numFixedSlots = templateObj->numFixedSlots(); + MDefinition* param; + if (script()->functionHasParameterExprs()) + param = constant(MagicValue(JS_UNINITIALIZED_LEXICAL)); + else + param = current->getSlot(info().argSlotUnchecked(formal)); + if (slot >= numFixedSlots) { + if (!slots) { + slots = MSlots::New(alloc(), callObj); + current->add(slots); + } + current->add(MStoreSlot::New(alloc(), slots, slot - numFixedSlots, param)); + } else { + current->add(MStoreFixedSlot::New(alloc(), callObj, slot, param)); + } + } + + return callObj; +} + +MDefinition* +IonBuilder::createThisScripted(MDefinition* callee, MDefinition* newTarget) +{ + // Get callee.prototype. + // + // This instruction MUST be idempotent: since it does not correspond to an + // explicit operation in the bytecode, we cannot use resumeAfter(). + // Getters may not override |prototype| fetching, so this operation is indeed idempotent. + // - First try an idempotent property cache. + // - Upon failing idempotent property cache, we can't use a non-idempotent cache, + // therefore we fallback to CallGetProperty + // + // Note: both CallGetProperty and GetPropertyCache can trigger a GC, + // and thus invalidation. + MInstruction* getProto; + if (!invalidatedIdempotentCache()) { + MConstant* id = constant(StringValue(names().prototype)); + MGetPropertyCache* getPropCache = MGetPropertyCache::New(alloc(), newTarget, id, + /* monitored = */ false); + getPropCache->setIdempotent(); + getProto = getPropCache; + } else { + MCallGetProperty* callGetProp = MCallGetProperty::New(alloc(), newTarget, names().prototype); + callGetProp->setIdempotent(); + getProto = callGetProp; + } + current->add(getProto); + + // Create this from prototype + MCreateThisWithProto* createThis = MCreateThisWithProto::New(alloc(), callee, newTarget, getProto); + current->add(createThis); + + return createThis; +} + +JSObject* +IonBuilder::getSingletonPrototype(JSFunction* target) +{ + TypeSet::ObjectKey* targetKey = TypeSet::ObjectKey::get(target); + if (targetKey->unknownProperties()) + return nullptr; + + jsid protoid = NameToId(names().prototype); + HeapTypeSetKey protoProperty = targetKey->property(protoid); + + return protoProperty.singleton(constraints()); +} + +MDefinition* +IonBuilder::createThisScriptedSingleton(JSFunction* target, MDefinition* callee) +{ + if (!target->hasScript()) + return nullptr; + + // Get the singleton prototype (if exists) + JSObject* proto = getSingletonPrototype(target); + if (!proto) + return nullptr; + + JSObject* templateObject = inspector->getTemplateObject(pc); + if (!templateObject) + return nullptr; + if (!templateObject->is<PlainObject>() && !templateObject->is<UnboxedPlainObject>()) + return nullptr; + if (templateObject->staticPrototype() != proto) + return nullptr; + + TypeSet::ObjectKey* templateObjectKey = TypeSet::ObjectKey::get(templateObject->group()); + if (templateObjectKey->hasFlags(constraints(), OBJECT_FLAG_NEW_SCRIPT_CLEARED)) + return nullptr; + + StackTypeSet* thisTypes = TypeScript::ThisTypes(target->nonLazyScript()); + if (!thisTypes || !thisTypes->hasType(TypeSet::ObjectType(templateObject))) + return nullptr; + + // Generate an inline path to create a new |this| object with + // the given singleton prototype. + MConstant* templateConst = MConstant::NewConstraintlessObject(alloc(), templateObject); + MCreateThisWithTemplate* createThis = + MCreateThisWithTemplate::New(alloc(), constraints(), templateConst, + templateObject->group()->initialHeap(constraints())); + current->add(templateConst); + current->add(createThis); + + return createThis; +} + +MDefinition* +IonBuilder::createThisScriptedBaseline(MDefinition* callee) +{ + // Try to inline |this| creation based on Baseline feedback. + + JSFunction* target = inspector->getSingleCallee(pc); + if (!target || !target->hasScript()) + return nullptr; + + JSObject* templateObject = inspector->getTemplateObject(pc); + if (!templateObject) + return nullptr; + if (!templateObject->is<PlainObject>() && !templateObject->is<UnboxedPlainObject>()) + return nullptr; + + Shape* shape = target->lookupPure(compartment->runtime()->names().prototype); + if (!shape || !shape->hasDefaultGetter() || !shape->hasSlot()) + return nullptr; + + Value protov = target->getSlot(shape->slot()); + if (!protov.isObject()) + return nullptr; + + JSObject* proto = checkNurseryObject(&protov.toObject()); + if (proto != templateObject->staticPrototype()) + return nullptr; + + TypeSet::ObjectKey* templateObjectKey = TypeSet::ObjectKey::get(templateObject->group()); + if (templateObjectKey->hasFlags(constraints(), OBJECT_FLAG_NEW_SCRIPT_CLEARED)) + return nullptr; + + StackTypeSet* thisTypes = TypeScript::ThisTypes(target->nonLazyScript()); + if (!thisTypes || !thisTypes->hasType(TypeSet::ObjectType(templateObject))) + return nullptr; + + // Shape guard. + callee = addShapeGuard(callee, target->lastProperty(), Bailout_ShapeGuard); + + // Guard callee.prototype == proto. + MOZ_ASSERT(shape->numFixedSlots() == 0, "Must be a dynamic slot"); + MSlots* slots = MSlots::New(alloc(), callee); + current->add(slots); + MLoadSlot* prototype = MLoadSlot::New(alloc(), slots, shape->slot()); + current->add(prototype); + MDefinition* protoConst = constant(ObjectValue(*proto)); + MGuardObjectIdentity* guard = MGuardObjectIdentity::New(alloc(), prototype, protoConst, + /* bailOnEquality = */ false); + current->add(guard); + + // Generate an inline path to create a new |this| object with + // the given prototype. + MConstant* templateConst = MConstant::NewConstraintlessObject(alloc(), templateObject); + MCreateThisWithTemplate* createThis = + MCreateThisWithTemplate::New(alloc(), constraints(), templateConst, + templateObject->group()->initialHeap(constraints())); + current->add(templateConst); + current->add(createThis); + + return createThis; +} + +MDefinition* +IonBuilder::createThis(JSFunction* target, MDefinition* callee, MDefinition* newTarget) +{ + // Create |this| for unknown target. + if (!target) { + if (MDefinition* createThis = createThisScriptedBaseline(callee)) + return createThis; + + MCreateThis* createThis = MCreateThis::New(alloc(), callee, newTarget); + current->add(createThis); + return createThis; + } + + // Native constructors build the new Object themselves. + if (target->isNative()) { + if (!target->isConstructor()) + return nullptr; + + MConstant* magic = MConstant::New(alloc(), MagicValue(JS_IS_CONSTRUCTING)); + current->add(magic); + return magic; + } + + if (target->isBoundFunction()) + return constant(MagicValue(JS_UNINITIALIZED_LEXICAL)); + + if (target->isDerivedClassConstructor()) { + MOZ_ASSERT(target->isClassConstructor()); + return constant(MagicValue(JS_UNINITIALIZED_LEXICAL)); + } + + // Try baking in the prototype. + if (MDefinition* createThis = createThisScriptedSingleton(target, callee)) + return createThis; + + if (MDefinition* createThis = createThisScriptedBaseline(callee)) + return createThis; + + return createThisScripted(callee, newTarget); +} + +bool +IonBuilder::jsop_funcall(uint32_t argc) +{ + // Stack for JSOP_FUNCALL: + // 1: arg0 + // ... + // argc: argN + // argc+1: JSFunction*, the 'f' in |f.call()|, in |this| position. + // argc+2: The native 'call' function. + + int calleeDepth = -((int)argc + 2); + int funcDepth = -((int)argc + 1); + + // If |Function.prototype.call| may be overridden, don't optimize callsite. + TemporaryTypeSet* calleeTypes = current->peek(calleeDepth)->resultTypeSet(); + JSFunction* native = getSingleCallTarget(calleeTypes); + if (!native || !native->isNative() || native->native() != &fun_call) { + CallInfo callInfo(alloc(), false); + if (!callInfo.init(current, argc)) + return false; + return makeCall(native, callInfo); + } + current->peek(calleeDepth)->setImplicitlyUsedUnchecked(); + + // Extract call target. + TemporaryTypeSet* funTypes = current->peek(funcDepth)->resultTypeSet(); + JSFunction* target = getSingleCallTarget(funTypes); + + // Shimmy the slots down to remove the native 'call' function. + current->shimmySlots(funcDepth - 1); + + bool zeroArguments = (argc == 0); + + // If no |this| argument was provided, explicitly pass Undefined. + // Pushing is safe here, since one stack slot has been removed. + if (zeroArguments) { + pushConstant(UndefinedValue()); + } else { + // |this| becomes implicit in the call. + argc -= 1; + } + + CallInfo callInfo(alloc(), false); + if (!callInfo.init(current, argc)) + return false; + + // Try to inline the call. + if (!zeroArguments) { + InliningDecision decision = makeInliningDecision(target, callInfo); + switch (decision) { + case InliningDecision_Error: + return false; + case InliningDecision_DontInline: + case InliningDecision_WarmUpCountTooLow: + break; + case InliningDecision_Inline: + if (target->isInterpreted()) { + InliningStatus status = inlineScriptedCall(callInfo, target); + if (status == InliningStatus_Inlined) + return true; + if (status == InliningStatus_Error) + return false; + } + break; + } + } + + // Call without inlining. + return makeCall(target, callInfo); +} + +bool +IonBuilder::jsop_funapply(uint32_t argc) +{ + int calleeDepth = -((int)argc + 2); + + TemporaryTypeSet* calleeTypes = current->peek(calleeDepth)->resultTypeSet(); + JSFunction* native = getSingleCallTarget(calleeTypes); + if (argc != 2 || info().analysisMode() == Analysis_ArgumentsUsage) { + CallInfo callInfo(alloc(), false); + if (!callInfo.init(current, argc)) + return false; + return makeCall(native, callInfo); + } + + // Disable compilation if the second argument to |apply| cannot be guaranteed + // to be either definitely |arguments| or definitely not |arguments|. + MDefinition* argument = current->peek(-1); + if (script()->argumentsHasVarBinding() && + argument->mightBeType(MIRType::MagicOptimizedArguments) && + argument->type() != MIRType::MagicOptimizedArguments) + { + return abort("fun.apply with MaybeArguments"); + } + + // Fallback to regular call if arg 2 is not definitely |arguments|. + if (argument->type() != MIRType::MagicOptimizedArguments) { + // Optimize fun.apply(self, array) if the length is sane and there are no holes. + TemporaryTypeSet* objTypes = argument->resultTypeSet(); + if (native && native->isNative() && native->native() == fun_apply && + objTypes && + objTypes->getKnownClass(constraints()) == &ArrayObject::class_ && + !objTypes->hasObjectFlags(constraints(), OBJECT_FLAG_LENGTH_OVERFLOW) && + ElementAccessIsPacked(constraints(), argument)) + { + return jsop_funapplyarray(argc); + } + + CallInfo callInfo(alloc(), false); + if (!callInfo.init(current, argc)) + return false; + return makeCall(native, callInfo); + } + + if ((!native || !native->isNative() || + native->native() != fun_apply) && + info().analysisMode() != Analysis_DefiniteProperties) + { + return abort("fun.apply speculation failed"); + } + + // Use funapply that definitely uses |arguments| + return jsop_funapplyarguments(argc); +} + +bool +IonBuilder::jsop_funapplyarray(uint32_t argc) +{ + MOZ_ASSERT(argc == 2); + + int funcDepth = -((int)argc + 1); + + // Extract call target. + TemporaryTypeSet* funTypes = current->peek(funcDepth)->resultTypeSet(); + JSFunction* target = getSingleCallTarget(funTypes); + + // Pop the array agument + MDefinition* argObj = current->pop(); + + MElements* elements = MElements::New(alloc(), argObj); + current->add(elements); + + // Pop the |this| argument. + MDefinition* argThis = current->pop(); + + // Unwrap the (JSFunction *) parameter. + MDefinition* argFunc = current->pop(); + + // Pop apply function. + MDefinition* nativeFunc = current->pop(); + nativeFunc->setImplicitlyUsedUnchecked(); + + WrappedFunction* wrappedTarget = target ? new(alloc()) WrappedFunction(target) : nullptr; + MApplyArray* apply = MApplyArray::New(alloc(), wrappedTarget, argFunc, elements, argThis); + current->add(apply); + current->push(apply); + if (!resumeAfter(apply)) + return false; + + TemporaryTypeSet* types = bytecodeTypes(pc); + return pushTypeBarrier(apply, types, BarrierKind::TypeSet); +} + +bool +IonBuilder::jsop_funapplyarguments(uint32_t argc) +{ + // Stack for JSOP_FUNAPPLY: + // 1: Vp + // 2: This + // argc+1: JSFunction*, the 'f' in |f.call()|, in |this| position. + // argc+2: The native 'apply' function. + + int funcDepth = -((int)argc + 1); + + // Extract call target. + TemporaryTypeSet* funTypes = current->peek(funcDepth)->resultTypeSet(); + JSFunction* target = getSingleCallTarget(funTypes); + + // When this script isn't inlined, use MApplyArgs, + // to copy the arguments from the stack and call the function + if (inliningDepth_ == 0 && info().analysisMode() != Analysis_DefiniteProperties) { + // The array argument corresponds to the arguments object. As the JIT + // is implicitly reading the arguments object in the next instruction, + // we need to prevent the deletion of the arguments object from resume + // points, so that Baseline will behave correctly after a bailout. + MDefinition* vp = current->pop(); + vp->setImplicitlyUsedUnchecked(); + + MDefinition* argThis = current->pop(); + + // Unwrap the (JSFunction*) parameter. + MDefinition* argFunc = current->pop(); + + // Pop apply function. + MDefinition* nativeFunc = current->pop(); + nativeFunc->setImplicitlyUsedUnchecked(); + + MArgumentsLength* numArgs = MArgumentsLength::New(alloc()); + current->add(numArgs); + + WrappedFunction* wrappedTarget = target ? new(alloc()) WrappedFunction(target) : nullptr; + MApplyArgs* apply = MApplyArgs::New(alloc(), wrappedTarget, argFunc, numArgs, argThis); + current->add(apply); + current->push(apply); + if (!resumeAfter(apply)) + return false; + + TemporaryTypeSet* types = bytecodeTypes(pc); + return pushTypeBarrier(apply, types, BarrierKind::TypeSet); + } + + // When inlining we have the arguments the function gets called with + // and can optimize even more, by just calling the functions with the args. + // We also try this path when doing the definite properties analysis, as we + // can inline the apply() target and don't care about the actual arguments + // that were passed in. + + CallInfo callInfo(alloc(), false); + + // Vp + MDefinition* vp = current->pop(); + vp->setImplicitlyUsedUnchecked(); + + // Arguments + if (inliningDepth_) { + if (!callInfo.setArgs(inlineCallInfo_->argv())) + return false; + } + + // This + MDefinition* argThis = current->pop(); + callInfo.setThis(argThis); + + // Pop function parameter. + MDefinition* argFunc = current->pop(); + callInfo.setFun(argFunc); + + // Pop apply function. + MDefinition* nativeFunc = current->pop(); + nativeFunc->setImplicitlyUsedUnchecked(); + + // Try to inline the call. + InliningDecision decision = makeInliningDecision(target, callInfo); + switch (decision) { + case InliningDecision_Error: + return false; + case InliningDecision_DontInline: + case InliningDecision_WarmUpCountTooLow: + break; + case InliningDecision_Inline: + if (target->isInterpreted()) { + InliningStatus status = inlineScriptedCall(callInfo, target); + if (status == InliningStatus_Inlined) + return true; + if (status == InliningStatus_Error) + return false; + } + } + + return makeCall(target, callInfo); +} + +bool +IonBuilder::jsop_call(uint32_t argc, bool constructing) +{ + startTrackingOptimizations(); + + // If this call has never executed, try to seed the observed type set + // based on how the call result is used. + TemporaryTypeSet* observed = bytecodeTypes(pc); + if (observed->empty()) { + if (BytecodeFlowsToBitop(pc)) { + observed->addType(TypeSet::Int32Type(), alloc_->lifoAlloc()); + } else if (*GetNextPc(pc) == JSOP_POS) { + // Note: this is lame, overspecialized on the code patterns used + // by asm.js and should be replaced by a more general mechanism. + // See bug 870847. + observed->addType(TypeSet::DoubleType(), alloc_->lifoAlloc()); + } + } + + int calleeDepth = -((int)argc + 2 + constructing); + + // Acquire known call target if existent. + ObjectVector targets(alloc()); + TemporaryTypeSet* calleeTypes = current->peek(calleeDepth)->resultTypeSet(); + if (calleeTypes && !getPolyCallTargets(calleeTypes, constructing, targets, 4)) + return false; + + CallInfo callInfo(alloc(), constructing); + if (!callInfo.init(current, argc)) + return false; + + // Try inlining + InliningStatus status = inlineCallsite(targets, callInfo); + if (status == InliningStatus_Inlined) + return true; + if (status == InliningStatus_Error) + return false; + + // Discard unreferenced & pre-allocated resume points. + replaceMaybeFallbackFunctionGetter(nullptr); + + // No inline, just make the call. + JSFunction* target = nullptr; + if (targets.length() == 1 && targets[0]->is<JSFunction>()) + target = &targets[0]->as<JSFunction>(); + + if (target && status == InliningStatus_WarmUpCountTooLow) { + MRecompileCheck* check = + MRecompileCheck::New(alloc(), target->nonLazyScript(), + optimizationInfo().inliningRecompileThreshold(), + MRecompileCheck::RecompileCheck_Inlining); + current->add(check); + } + + return makeCall(target, callInfo); +} + +bool +IonBuilder::testShouldDOMCall(TypeSet* inTypes, JSFunction* func, JSJitInfo::OpType opType) +{ + if (!func->isNative() || !func->jitInfo()) + return false; + + // If all the DOM objects flowing through are legal with this + // property, we can bake in a call to the bottom half of the DOM + // accessor + DOMInstanceClassHasProtoAtDepth instanceChecker = + compartment->runtime()->DOMcallbacks()->instanceClassMatchesProto; + + const JSJitInfo* jinfo = func->jitInfo(); + if (jinfo->type() != opType) + return false; + + for (unsigned i = 0; i < inTypes->getObjectCount(); i++) { + TypeSet::ObjectKey* key = inTypes->getObject(i); + if (!key) + continue; + + if (!key->hasStableClassAndProto(constraints())) + return false; + + if (!instanceChecker(key->clasp(), jinfo->protoID, jinfo->depth)) + return false; + } + + return true; +} + +static bool +ArgumentTypesMatch(MDefinition* def, StackTypeSet* calleeTypes) +{ + if (!calleeTypes) + return false; + + if (def->resultTypeSet()) { + MOZ_ASSERT(def->type() == MIRType::Value || def->mightBeType(def->type())); + return def->resultTypeSet()->isSubset(calleeTypes); + } + + if (def->type() == MIRType::Value) + return false; + + if (def->type() == MIRType::Object) + return calleeTypes->unknownObject(); + + return calleeTypes->mightBeMIRType(def->type()); +} + +bool +IonBuilder::testNeedsArgumentCheck(JSFunction* target, CallInfo& callInfo) +{ + // If we have a known target, check if the caller arg types are a subset of callee. + // Since typeset accumulates and can't decrease that means we don't need to check + // the arguments anymore. + if (!target->hasScript()) + return true; + + JSScript* targetScript = target->nonLazyScript(); + + if (!ArgumentTypesMatch(callInfo.thisArg(), TypeScript::ThisTypes(targetScript))) + return true; + uint32_t expected_args = Min<uint32_t>(callInfo.argc(), target->nargs()); + for (size_t i = 0; i < expected_args; i++) { + if (!ArgumentTypesMatch(callInfo.getArg(i), TypeScript::ArgTypes(targetScript, i))) + return true; + } + for (size_t i = callInfo.argc(); i < target->nargs(); i++) { + if (!TypeScript::ArgTypes(targetScript, i)->mightBeMIRType(MIRType::Undefined)) + return true; + } + + return false; +} + +MCall* +IonBuilder::makeCallHelper(JSFunction* target, CallInfo& callInfo) +{ + // This function may be called with mutated stack. + // Querying TI for popped types is invalid. + + uint32_t targetArgs = callInfo.argc(); + + // Collect number of missing arguments provided that the target is + // scripted. Native functions are passed an explicit 'argc' parameter. + if (target && !target->isNative()) + targetArgs = Max<uint32_t>(target->nargs(), callInfo.argc()); + + bool isDOMCall = false; + if (target && !callInfo.constructing()) { + // We know we have a single call target. Check whether the "this" types + // are DOM types and our function a DOM function, and if so flag the + // MCall accordingly. + TemporaryTypeSet* thisTypes = callInfo.thisArg()->resultTypeSet(); + if (thisTypes && + thisTypes->getKnownMIRType() == MIRType::Object && + thisTypes->isDOMClass(constraints()) && + testShouldDOMCall(thisTypes, target, JSJitInfo::Method)) + { + isDOMCall = true; + } + } + + MCall* call = MCall::New(alloc(), target, targetArgs + 1 + callInfo.constructing(), + callInfo.argc(), callInfo.constructing(), isDOMCall); + if (!call) + return nullptr; + + if (callInfo.constructing()) + call->addArg(targetArgs + 1, callInfo.getNewTarget()); + + // Explicitly pad any missing arguments with |undefined|. + // This permits skipping the argumentsRectifier. + for (int i = targetArgs; i > (int)callInfo.argc(); i--) { + MOZ_ASSERT_IF(target, !target->isNative()); + MConstant* undef = constant(UndefinedValue()); + if (!alloc().ensureBallast()) + return nullptr; + call->addArg(i, undef); + } + + // Add explicit arguments. + // Skip addArg(0) because it is reserved for this + for (int32_t i = callInfo.argc() - 1; i >= 0; i--) + call->addArg(i + 1, callInfo.getArg(i)); + + // Now that we've told it about all the args, compute whether it's movable + call->computeMovable(); + + // Inline the constructor on the caller-side. + if (callInfo.constructing()) { + MDefinition* create = createThis(target, callInfo.fun(), callInfo.getNewTarget()); + if (!create) { + abort("Failure inlining constructor for call."); + return nullptr; + } + + callInfo.thisArg()->setImplicitlyUsedUnchecked(); + callInfo.setThis(create); + } + + // Pass |this| and function. + MDefinition* thisArg = callInfo.thisArg(); + call->addArg(0, thisArg); + + if (target && !testNeedsArgumentCheck(target, callInfo)) + call->disableArgCheck(); + + call->initFunction(callInfo.fun()); + + current->add(call); + return call; +} + +static bool +DOMCallNeedsBarrier(const JSJitInfo* jitinfo, TemporaryTypeSet* types) +{ + MOZ_ASSERT(jitinfo->type() != JSJitInfo::InlinableNative); + + // If the return type of our DOM native is in "types" already, we don't + // actually need a barrier. + if (jitinfo->returnType() == JSVAL_TYPE_UNKNOWN) + return true; + + // JSVAL_TYPE_OBJECT doesn't tell us much; we still have to barrier on the + // actual type of the object. + if (jitinfo->returnType() == JSVAL_TYPE_OBJECT) + return true; + + // No need for a barrier if we're already expecting the type we'll produce. + return MIRTypeFromValueType(jitinfo->returnType()) != types->getKnownMIRType(); +} + +bool +IonBuilder::makeCall(JSFunction* target, CallInfo& callInfo) +{ + // Constructor calls to non-constructors should throw. We don't want to use + // CallKnown in this case. + MOZ_ASSERT_IF(callInfo.constructing() && target, target->isConstructor()); + + MCall* call = makeCallHelper(target, callInfo); + if (!call) + return false; + + current->push(call); + if (call->isEffectful() && !resumeAfter(call)) + return false; + + TemporaryTypeSet* types = bytecodeTypes(pc); + + if (call->isCallDOMNative()) + return pushDOMTypeBarrier(call, types, call->getSingleTarget()->rawJSFunction()); + + return pushTypeBarrier(call, types, BarrierKind::TypeSet); +} + +bool +IonBuilder::jsop_eval(uint32_t argc) +{ + int calleeDepth = -((int)argc + 2); + TemporaryTypeSet* calleeTypes = current->peek(calleeDepth)->resultTypeSet(); + + // Emit a normal call if the eval has never executed. This keeps us from + // disabling compilation for the script when testing with --ion-eager. + if (calleeTypes && calleeTypes->empty()) + return jsop_call(argc, /* constructing = */ false); + + JSFunction* singleton = getSingleCallTarget(calleeTypes); + if (!singleton) + return abort("No singleton callee for eval()"); + + if (script()->global().valueIsEval(ObjectValue(*singleton))) { + if (argc != 1) + return abort("Direct eval with more than one argument"); + + if (!info().funMaybeLazy()) + return abort("Direct eval in global code"); + + if (info().funMaybeLazy()->isArrow()) + return abort("Direct eval from arrow function"); + + CallInfo callInfo(alloc(), /* constructing = */ false); + if (!callInfo.init(current, argc)) + return false; + callInfo.setImplicitlyUsedUnchecked(); + + callInfo.fun()->setImplicitlyUsedUnchecked(); + + MDefinition* envChain = current->environmentChain(); + MDefinition* string = callInfo.getArg(0); + + // Direct eval acts as identity on non-string types according to + // ES5 15.1.2.1 step 1. + if (!string->mightBeType(MIRType::String)) { + current->push(string); + TemporaryTypeSet* types = bytecodeTypes(pc); + return pushTypeBarrier(string, types, BarrierKind::TypeSet); + } + + if (!jsop_newtarget()) + return false; + MDefinition* newTargetValue = current->pop(); + + // Try to pattern match 'eval(v + "()")'. In this case v is likely a + // name on the env chain and the eval is performing a call on that + // value. Use an env chain lookup rather than a full eval. + if (string->isConcat() && + string->getOperand(1)->type() == MIRType::String && + string->getOperand(1)->maybeConstantValue()) + { + JSAtom* atom = &string->getOperand(1)->maybeConstantValue()->toString()->asAtom(); + + if (StringEqualsAscii(atom, "()")) { + MDefinition* name = string->getOperand(0); + MInstruction* dynamicName = MGetDynamicName::New(alloc(), envChain, name); + current->add(dynamicName); + + current->push(dynamicName); + current->push(constant(UndefinedValue())); // thisv + + CallInfo evalCallInfo(alloc(), /* constructing = */ false); + if (!evalCallInfo.init(current, /* argc = */ 0)) + return false; + + return makeCall(nullptr, evalCallInfo); + } + } + + MInstruction* ins = MCallDirectEval::New(alloc(), envChain, string, + newTargetValue, pc); + current->add(ins); + current->push(ins); + + TemporaryTypeSet* types = bytecodeTypes(pc); + return resumeAfter(ins) && pushTypeBarrier(ins, types, BarrierKind::TypeSet); + } + + return jsop_call(argc, /* constructing = */ false); +} + +bool +IonBuilder::jsop_compare(JSOp op) +{ + MDefinition* right = current->pop(); + MDefinition* left = current->pop(); + + return jsop_compare(op, left, right); +} + +bool +IonBuilder::jsop_compare(JSOp op, MDefinition* left, MDefinition* right) +{ + bool emitted = false; + + if (!forceInlineCaches()) { + if (!compareTrySpecialized(&emitted, op, left, right) || emitted) + return emitted; + if (!compareTryBitwise(&emitted, op, left, right) || emitted) + return emitted; + if (!compareTrySpecializedOnBaselineInspector(&emitted, op, left, right) || emitted) + return emitted; + } + + if (!compareTrySharedStub(&emitted, op, left, right) || emitted) + return emitted; + + // Not possible to optimize. Do a slow vm call. + MCompare* ins = MCompare::New(alloc(), left, right, op); + ins->cacheOperandMightEmulateUndefined(constraints()); + + current->add(ins); + current->push(ins); + if (ins->isEffectful() && !resumeAfter(ins)) + return false; + return true; +} + +static bool +ObjectOrSimplePrimitive(MDefinition* op) +{ + // Return true if op is either undefined/null/boolean/int32 or an object. + return !op->mightBeType(MIRType::String) + && !op->mightBeType(MIRType::Symbol) + && !op->mightBeType(MIRType::Double) + && !op->mightBeType(MIRType::Float32) + && !op->mightBeType(MIRType::MagicOptimizedArguments) + && !op->mightBeType(MIRType::MagicHole) + && !op->mightBeType(MIRType::MagicIsConstructing); +} + +bool +IonBuilder::compareTrySpecialized(bool* emitted, JSOp op, MDefinition* left, MDefinition* right) +{ + MOZ_ASSERT(*emitted == false); + + // Try to emit an compare based on the input types. + + MCompare::CompareType type = MCompare::determineCompareType(op, left, right); + if (type == MCompare::Compare_Unknown) + return true; + + MCompare* ins = MCompare::New(alloc(), left, right, op); + ins->setCompareType(type); + ins->cacheOperandMightEmulateUndefined(constraints()); + + // Some compare types need to have the specific type in the rhs. + // Swap operands if that is not the case. + if (type == MCompare::Compare_StrictString && right->type() != MIRType::String) + ins->swapOperands(); + else if (type == MCompare::Compare_Null && right->type() != MIRType::Null) + ins->swapOperands(); + else if (type == MCompare::Compare_Undefined && right->type() != MIRType::Undefined) + ins->swapOperands(); + else if (type == MCompare::Compare_Boolean && right->type() != MIRType::Boolean) + ins->swapOperands(); + + // Replace inputs with unsigned variants if needed. + if (type == MCompare::Compare_UInt32) + ins->replaceWithUnsignedOperands(); + + current->add(ins); + current->push(ins); + + MOZ_ASSERT(!ins->isEffectful()); + *emitted = true; + return true; +} + +bool +IonBuilder::compareTryBitwise(bool* emitted, JSOp op, MDefinition* left, MDefinition* right) +{ + MOZ_ASSERT(*emitted == false); + + // Try to emit a bitwise compare. Check if a bitwise compare equals the wanted + // result for all observed operand types. + + // Onlye allow loose and strict equality. + if (op != JSOP_EQ && op != JSOP_NE && op != JSOP_STRICTEQ && op != JSOP_STRICTNE) + return true; + + // Only primitive (not double/string) or objects are supported. + // I.e. Undefined/Null/Boolean/Int32 and Object + if (!ObjectOrSimplePrimitive(left) || !ObjectOrSimplePrimitive(right)) + return true; + + // Objects that emulate undefined are not supported. + if (left->maybeEmulatesUndefined(constraints()) || right->maybeEmulatesUndefined(constraints())) + return true; + + // In the loose comparison more values could be the same, + // but value comparison reporting otherwise. + if (op == JSOP_EQ || op == JSOP_NE) { + + // Undefined compared loosy to Null is not supported, + // because tag is different, but value can be the same (undefined == null). + if ((left->mightBeType(MIRType::Undefined) && right->mightBeType(MIRType::Null)) || + (left->mightBeType(MIRType::Null) && right->mightBeType(MIRType::Undefined))) + { + return true; + } + + // Int32 compared loosy to Boolean is not supported, + // because tag is different, but value can be the same (1 == true). + if ((left->mightBeType(MIRType::Int32) && right->mightBeType(MIRType::Boolean)) || + (left->mightBeType(MIRType::Boolean) && right->mightBeType(MIRType::Int32))) + { + return true; + } + + // For loosy comparison of an object with a Boolean/Number/String + // the valueOf the object is taken. Therefore not supported. + bool simpleLHS = left->mightBeType(MIRType::Boolean) || left->mightBeType(MIRType::Int32); + bool simpleRHS = right->mightBeType(MIRType::Boolean) || right->mightBeType(MIRType::Int32); + if ((left->mightBeType(MIRType::Object) && simpleRHS) || + (right->mightBeType(MIRType::Object) && simpleLHS)) + { + return true; + } + } + + MCompare* ins = MCompare::New(alloc(), left, right, op); + ins->setCompareType(MCompare::Compare_Bitwise); + ins->cacheOperandMightEmulateUndefined(constraints()); + + current->add(ins); + current->push(ins); + + MOZ_ASSERT(!ins->isEffectful()); + *emitted = true; + return true; +} + +bool +IonBuilder::compareTrySpecializedOnBaselineInspector(bool* emitted, JSOp op, MDefinition* left, + MDefinition* right) +{ + MOZ_ASSERT(*emitted == false); + + // Try to specialize based on any baseline caches that have been generated + // for the opcode. These will cause the instruction's type policy to insert + // fallible unboxes to the appropriate input types. + + // Strict equality isn't supported. + if (op == JSOP_STRICTEQ || op == JSOP_STRICTNE) + return true; + + MCompare::CompareType type = inspector->expectedCompareType(pc); + if (type == MCompare::Compare_Unknown) + return true; + + MCompare* ins = MCompare::New(alloc(), left, right, op); + ins->setCompareType(type); + ins->cacheOperandMightEmulateUndefined(constraints()); + + current->add(ins); + current->push(ins); + + MOZ_ASSERT(!ins->isEffectful()); + *emitted = true; + return true; +} + +bool +IonBuilder::compareTrySharedStub(bool* emitted, JSOp op, MDefinition* left, MDefinition* right) +{ + MOZ_ASSERT(*emitted == false); + + // Try to emit a shared stub cache. + + if (JitOptions.disableSharedStubs) + return true; + + if (JSOp(*pc) == JSOP_CASE) + return true; + + MBinarySharedStub* stub = MBinarySharedStub::New(alloc(), left, right); + current->add(stub); + current->push(stub); + if (!resumeAfter(stub)) + return false; + + MUnbox* unbox = MUnbox::New(alloc(), current->pop(), MIRType::Boolean, MUnbox::Infallible); + current->add(unbox); + current->push(unbox); + + *emitted = true; + return true; +} + +bool +IonBuilder::newArrayTryTemplateObject(bool* emitted, JSObject* templateObject, uint32_t length) +{ + MOZ_ASSERT(*emitted == false); + + if (!templateObject) + return true; + + if (templateObject->is<UnboxedArrayObject>()) { + MOZ_ASSERT(templateObject->as<UnboxedArrayObject>().capacity() >= length); + if (!templateObject->as<UnboxedArrayObject>().hasInlineElements()) + return true; + } + + MOZ_ASSERT(length <= NativeObject::MAX_DENSE_ELEMENTS_COUNT); + + size_t arraySlots = + gc::GetGCKindSlots(templateObject->asTenured().getAllocKind()) - ObjectElements::VALUES_PER_HEADER; + + if (length > arraySlots) + return true; + + // Emit fastpath. + + gc::InitialHeap heap = templateObject->group()->initialHeap(constraints()); + MConstant* templateConst = MConstant::NewConstraintlessObject(alloc(), templateObject); + current->add(templateConst); + + MNewArray* ins = MNewArray::New(alloc(), constraints(), length, templateConst, heap, pc); + current->add(ins); + current->push(ins); + + *emitted = true; + return true; +} + +bool +IonBuilder::newArrayTrySharedStub(bool* emitted) +{ + MOZ_ASSERT(*emitted == false); + + // Try to emit a shared stub cache. + + if (JitOptions.disableSharedStubs) + return true; + + if (*pc != JSOP_NEWINIT && *pc != JSOP_NEWARRAY) + return true; + + MInstruction* stub = MNullarySharedStub::New(alloc()); + current->add(stub); + current->push(stub); + + if (!resumeAfter(stub)) + return false; + + *emitted = true; + return true; +} + +bool +IonBuilder::newArrayTryVM(bool* emitted, JSObject* templateObject, uint32_t length) +{ + MOZ_ASSERT(*emitted == false); + + // Emit a VM call. + + gc::InitialHeap heap = gc::DefaultHeap; + MConstant* templateConst = MConstant::New(alloc(), NullValue()); + + if (templateObject) { + heap = templateObject->group()->initialHeap(constraints()); + templateConst = MConstant::NewConstraintlessObject(alloc(), templateObject); + } + + current->add(templateConst); + + MNewArray* ins = MNewArray::NewVM(alloc(), constraints(), length, templateConst, heap, pc); + current->add(ins); + current->push(ins); + + *emitted = true; + return true; +} + +bool +IonBuilder::jsop_newarray(uint32_t length) +{ + JSObject* templateObject = inspector->getTemplateObject(pc); + if (!jsop_newarray(templateObject, length)) + return false; + + // Improve resulting typeset. + ObjectGroup* templateGroup = inspector->getTemplateObjectGroup(pc); + if (templateGroup) { + TemporaryTypeSet* types = MakeSingletonTypeSet(constraints(), templateGroup); + current->peek(-1)->setResultTypeSet(types); + } + + return true; +} + +bool +IonBuilder::jsop_newarray(JSObject* templateObject, uint32_t length) +{ + bool emitted = false; + + if (!forceInlineCaches()) { + if (!newArrayTryTemplateObject(&emitted, templateObject, length) || emitted) + return emitted; + } + + if (!newArrayTrySharedStub(&emitted) || emitted) + return emitted; + + if (!newArrayTryVM(&emitted, templateObject, length) || emitted) + return emitted; + + MOZ_CRASH("newarray should have been emited"); +} + +bool +IonBuilder::jsop_newarray_copyonwrite() +{ + ArrayObject* templateObject = ObjectGroup::getCopyOnWriteObject(script(), pc); + + // The baseline compiler should have ensured the template object has a type + // with the copy on write flag set already. During the arguments usage + // analysis the baseline compiler hasn't run yet, however, though in this + // case the template object's type doesn't matter. + MOZ_ASSERT_IF(info().analysisMode() != Analysis_ArgumentsUsage, + templateObject->group()->hasAnyFlags(OBJECT_FLAG_COPY_ON_WRITE)); + + MNewArrayCopyOnWrite* ins = + MNewArrayCopyOnWrite::New(alloc(), constraints(), templateObject, + templateObject->group()->initialHeap(constraints())); + + current->add(ins); + current->push(ins); + + return true; +} + +bool +IonBuilder::newObjectTryTemplateObject(bool* emitted, JSObject* templateObject) +{ + MOZ_ASSERT(*emitted == false); + + if (!templateObject) + return true; + + if (templateObject->is<PlainObject>() && templateObject->as<PlainObject>().hasDynamicSlots()) + return true; + + // Emit fastpath. + + MNewObject::Mode mode; + if (JSOp(*pc) == JSOP_NEWOBJECT || JSOp(*pc) == JSOP_NEWINIT) + mode = MNewObject::ObjectLiteral; + else + mode = MNewObject::ObjectCreate; + + gc::InitialHeap heap = templateObject->group()->initialHeap(constraints()); + MConstant* templateConst = MConstant::NewConstraintlessObject(alloc(), templateObject); + current->add(templateConst); + + MNewObject* ins = MNewObject::New(alloc(), constraints(), templateConst, heap, mode); + current->add(ins); + current->push(ins); + + if (!resumeAfter(ins)) + return false; + + *emitted = true; + return true; +} + +bool +IonBuilder::newObjectTrySharedStub(bool* emitted) +{ + MOZ_ASSERT(*emitted == false); + + // Try to emit a shared stub cache. + + if (JitOptions.disableSharedStubs) + return true; + + MInstruction* stub = MNullarySharedStub::New(alloc()); + current->add(stub); + current->push(stub); + + if (!resumeAfter(stub)) + return false; + + *emitted = true; + return true; +} + +bool +IonBuilder::newObjectTryVM(bool* emitted, JSObject* templateObject) +{ + // Emit a VM call. + MOZ_ASSERT(JSOp(*pc) == JSOP_NEWOBJECT || JSOp(*pc) == JSOP_NEWINIT); + + gc::InitialHeap heap = gc::DefaultHeap; + MConstant* templateConst = MConstant::New(alloc(), NullValue()); + + if (templateObject) { + heap = templateObject->group()->initialHeap(constraints()); + templateConst = MConstant::NewConstraintlessObject(alloc(), templateObject); + } + + current->add(templateConst); + + MNewObject* ins = MNewObject::NewVM(alloc(), constraints(), templateConst, heap, + MNewObject::ObjectLiteral); + current->add(ins); + current->push(ins); + + if (!resumeAfter(ins)) + return false; + + *emitted = true; + return true; +} + +bool +IonBuilder::jsop_newobject() +{ + bool emitted = false; + + JSObject* templateObject = inspector->getTemplateObject(pc); + + if (!forceInlineCaches()) { + if (!newObjectTryTemplateObject(&emitted, templateObject) || emitted) + return emitted; + } + if (!newObjectTrySharedStub(&emitted) || emitted) + return emitted; + + if (!newObjectTryVM(&emitted, templateObject) || emitted) + return emitted; + + MOZ_CRASH("newobject should have been emited"); +} + +bool +IonBuilder::jsop_initelem() +{ + MDefinition* value = current->pop(); + MDefinition* id = current->pop(); + MDefinition* obj = current->peek(-1); + + MInitElem* initElem = MInitElem::New(alloc(), obj, id, value); + current->add(initElem); + + return resumeAfter(initElem); +} + +bool +IonBuilder::jsop_initelem_array() +{ + MDefinition* value = current->pop(); + MDefinition* obj = current->peek(-1); + + // Make sure that arrays have the type being written to them by the + // intializer, and that arrays are marked as non-packed when writing holes + // to them during initialization. + bool needStub = false; + JSValueType unboxedType = JSVAL_TYPE_MAGIC; + if (shouldAbortOnPreliminaryGroups(obj)) { + needStub = true; + } else if (!obj->resultTypeSet() || + obj->resultTypeSet()->unknownObject() || + obj->resultTypeSet()->getObjectCount() != 1) + { + needStub = true; + } else { + MOZ_ASSERT(obj->resultTypeSet()->getObjectCount() == 1); + TypeSet::ObjectKey* initializer = obj->resultTypeSet()->getObject(0); + if (initializer->clasp() == &UnboxedArrayObject::class_) { + if (initializer->group()->unboxedLayout().nativeGroup()) + needStub = true; + else + unboxedType = initializer->group()->unboxedLayout().elementType(); + } + if (value->type() == MIRType::MagicHole) { + if (!initializer->hasFlags(constraints(), OBJECT_FLAG_NON_PACKED)) + needStub = true; + } else if (!initializer->unknownProperties()) { + HeapTypeSetKey elemTypes = initializer->property(JSID_VOID); + if (!TypeSetIncludes(elemTypes.maybeTypes(), value->type(), value->resultTypeSet())) { + elemTypes.freeze(constraints()); + needStub = true; + } + } + } + + uint32_t index = GET_UINT32(pc); + if (needStub) { + MCallInitElementArray* store = MCallInitElementArray::New(alloc(), obj, index, value); + current->add(store); + return resumeAfter(store); + } + + return initializeArrayElement(obj, index, value, unboxedType, /* addResumePoint = */ true); +} + +bool +IonBuilder::initializeArrayElement(MDefinition* obj, size_t index, MDefinition* value, + JSValueType unboxedType, + bool addResumePointAndIncrementInitializedLength) +{ + MConstant* id = MConstant::New(alloc(), Int32Value(index)); + current->add(id); + + // Get the elements vector. + MElements* elements = MElements::New(alloc(), obj, unboxedType != JSVAL_TYPE_MAGIC); + current->add(elements); + + if (unboxedType != JSVAL_TYPE_MAGIC) { + // Note: storeUnboxedValue takes care of any post barriers on the value. + storeUnboxedValue(obj, elements, 0, id, unboxedType, value, /* preBarrier = */ false); + + if (addResumePointAndIncrementInitializedLength) { + MInstruction* increment = MIncrementUnboxedArrayInitializedLength::New(alloc(), obj); + current->add(increment); + + if (!resumeAfter(increment)) + return false; + } + } else { + if (NeedsPostBarrier(value)) + current->add(MPostWriteBarrier::New(alloc(), obj, value)); + + if ((obj->isNewArray() && obj->toNewArray()->convertDoubleElements()) || + (obj->isNullarySharedStub() && + obj->resultTypeSet()->convertDoubleElements(constraints()) == TemporaryTypeSet::AlwaysConvertToDoubles)) + { + MInstruction* valueDouble = MToDouble::New(alloc(), value); + current->add(valueDouble); + value = valueDouble; + } + + // Store the value. + MStoreElement* store = MStoreElement::New(alloc(), elements, id, value, + /* needsHoleCheck = */ false); + current->add(store); + + if (addResumePointAndIncrementInitializedLength) { + // Update the initialized length. (The template object for this + // array has the array's ultimate length, so the length field is + // already correct: no updating needed.) + MSetInitializedLength* initLength = MSetInitializedLength::New(alloc(), elements, id); + current->add(initLength); + + if (!resumeAfter(initLength)) + return false; + } + } + + return true; +} + +bool +IonBuilder::jsop_mutateproto() +{ + MDefinition* value = current->pop(); + MDefinition* obj = current->peek(-1); + + MMutateProto* mutate = MMutateProto::New(alloc(), obj, value); + current->add(mutate); + return resumeAfter(mutate); +} + +bool +IonBuilder::jsop_initprop(PropertyName* name) +{ + bool useSlowPath = false; + + MDefinition* value = current->peek(-1); + MDefinition* obj = current->peek(-2); + if (obj->isLambda()) { + useSlowPath = true; + } else if (obj->isNewObject()) { + if (JSObject* templateObject = obj->toNewObject()->templateObject()) { + if (templateObject->is<PlainObject>()) { + if (!templateObject->as<PlainObject>().containsPure(name)) + useSlowPath = true; + } else { + MOZ_ASSERT(templateObject->as<UnboxedPlainObject>().layout().lookup(name)); + } + } else { + useSlowPath = true; + } + } else { + MOZ_ASSERT(obj->isNullarySharedStub()); + useSlowPath = true; + } + + if (useSlowPath) { + current->pop(); + MInitProp* init = MInitProp::New(alloc(), obj, name, value); + current->add(init); + return resumeAfter(init); + } + + MInstruction* last = *current->rbegin(); + + // This is definitely initializing an 'own' property of the object, treat + // it as an assignment. + if (!jsop_setprop(name)) + return false; + + // SETPROP pushed the value, instead of the object. Fix this on the stack, + // and check the most recent resume point to see if it needs updating too. + current->pop(); + current->push(obj); + for (MInstructionReverseIterator riter = current->rbegin(); *riter != last; riter++) { + if (MResumePoint* resumePoint = riter->resumePoint()) { + MOZ_ASSERT(resumePoint->pc() == pc); + if (resumePoint->mode() == MResumePoint::ResumeAfter) { + size_t index = resumePoint->numOperands() - 1; + resumePoint->replaceOperand(index, obj); + } + break; + } + } + + return true; +} + +bool +IonBuilder::jsop_initprop_getter_setter(PropertyName* name) +{ + MDefinition* value = current->pop(); + MDefinition* obj = current->peek(-1); + + MInitPropGetterSetter* init = MInitPropGetterSetter::New(alloc(), obj, name, value); + current->add(init); + return resumeAfter(init); +} + +bool +IonBuilder::jsop_initelem_getter_setter() +{ + MDefinition* value = current->pop(); + MDefinition* id = current->pop(); + MDefinition* obj = current->peek(-1); + + MInitElemGetterSetter* init = MInitElemGetterSetter::New(alloc(), obj, id, value); + current->add(init); + return resumeAfter(init); +} + +MBasicBlock* +IonBuilder::addBlock(MBasicBlock* block, uint32_t loopDepth) +{ + if (!block) + return nullptr; + if (block->pc() && script()->hasScriptCounts()) + block->setHitCount(script()->getHitCount(block->pc())); + graph().addBlock(block); + block->setLoopDepth(loopDepth); + return block; +} + +MBasicBlock* +IonBuilder::newBlock(MBasicBlock* predecessor, jsbytecode* pc) +{ + MBasicBlock* block = MBasicBlock::New(graph(), &analysis(), info(), predecessor, + bytecodeSite(pc), MBasicBlock::NORMAL); + return addBlock(block, loopDepth_); +} + +MBasicBlock* +IonBuilder::newBlock(MBasicBlock* predecessor, jsbytecode* pc, MResumePoint* priorResumePoint) +{ + MBasicBlock* block = MBasicBlock::NewWithResumePoint(graph(), info(), predecessor, + bytecodeSite(pc), priorResumePoint); + return addBlock(block, loopDepth_); +} + +MBasicBlock* +IonBuilder::newBlockPopN(MBasicBlock* predecessor, jsbytecode* pc, uint32_t popped) +{ + MBasicBlock* block = MBasicBlock::NewPopN(graph(), info(), predecessor, bytecodeSite(pc), + MBasicBlock::NORMAL, popped); + return addBlock(block, loopDepth_); +} + +MBasicBlock* +IonBuilder::newBlockAfter(MBasicBlock* at, MBasicBlock* predecessor, jsbytecode* pc) +{ + MBasicBlock* block = MBasicBlock::New(graph(), &analysis(), info(), predecessor, + bytecodeSite(pc), MBasicBlock::NORMAL); + if (!block) + return nullptr; + block->setHitCount(0); // osr block + graph().insertBlockAfter(at, block); + return block; +} + +MBasicBlock* +IonBuilder::newBlock(MBasicBlock* predecessor, jsbytecode* pc, uint32_t loopDepth) +{ + MBasicBlock* block = MBasicBlock::New(graph(), &analysis(), info(), predecessor, + bytecodeSite(pc), MBasicBlock::NORMAL); + return addBlock(block, loopDepth); +} + +MBasicBlock* +IonBuilder::newOsrPreheader(MBasicBlock* predecessor, jsbytecode* loopEntry, jsbytecode* beforeLoopEntry) +{ + MOZ_ASSERT(LoopEntryCanIonOsr(loopEntry)); + MOZ_ASSERT(loopEntry == info().osrPc()); + + // Create two blocks: one for the OSR entry with no predecessors, one for + // the preheader, which has the OSR entry block as a predecessor. The + // OSR block is always the second block (with id 1). + MBasicBlock* osrBlock = newBlockAfter(*graph().begin(), loopEntry); + MBasicBlock* preheader = newBlock(predecessor, loopEntry); + if (!osrBlock || !preheader) + return nullptr; + + // Give the pre-header the same hit count as the code before the loop. + if (script()->hasScriptCounts()) + preheader->setHitCount(script()->getHitCount(beforeLoopEntry)); + + MOsrEntry* entry = MOsrEntry::New(alloc()); + osrBlock->add(entry); + + // Initialize |envChain|. + { + uint32_t slot = info().environmentChainSlot(); + + MInstruction* envv; + if (analysis().usesEnvironmentChain()) { + envv = MOsrEnvironmentChain::New(alloc(), entry); + } else { + // Use an undefined value if the script does not need its env + // chain, to match the type that is already being tracked for the + // slot. + envv = MConstant::New(alloc(), UndefinedValue()); + } + + osrBlock->add(envv); + osrBlock->initSlot(slot, envv); + } + // Initialize |return value| + { + MInstruction* returnValue; + if (!script()->noScriptRval()) + returnValue = MOsrReturnValue::New(alloc(), entry); + else + returnValue = MConstant::New(alloc(), UndefinedValue()); + osrBlock->add(returnValue); + osrBlock->initSlot(info().returnValueSlot(), returnValue); + } + + // Initialize arguments object. + bool needsArgsObj = info().needsArgsObj(); + MInstruction* argsObj = nullptr; + if (info().hasArguments()) { + if (needsArgsObj) + argsObj = MOsrArgumentsObject::New(alloc(), entry); + else + argsObj = MConstant::New(alloc(), UndefinedValue()); + osrBlock->add(argsObj); + osrBlock->initSlot(info().argsObjSlot(), argsObj); + } + + if (info().funMaybeLazy()) { + // Initialize |this| parameter. + MParameter* thisv = MParameter::New(alloc(), MParameter::THIS_SLOT, nullptr); + osrBlock->add(thisv); + osrBlock->initSlot(info().thisSlot(), thisv); + + // Initialize arguments. + for (uint32_t i = 0; i < info().nargs(); i++) { + uint32_t slot = needsArgsObj ? info().argSlotUnchecked(i) : info().argSlot(i); + + // Only grab arguments from the arguments object if the arguments object + // aliases formals. If the argsobj does not alias formals, then the + // formals may have been assigned to during interpretation, and that change + // will not be reflected in the argsobj. + if (needsArgsObj && info().argsObjAliasesFormals()) { + MOZ_ASSERT(argsObj && argsObj->isOsrArgumentsObject()); + // If this is an aliased formal, then the arguments object + // contains a hole at this index. Any references to this + // variable in the jitcode will come from JSOP_*ALIASEDVAR + // opcodes, so the slot itself can be set to undefined. If + // it's not aliased, it must be retrieved from the arguments + // object. + MInstruction* osrv; + if (script()->formalIsAliased(i)) + osrv = MConstant::New(alloc(), UndefinedValue()); + else + osrv = MGetArgumentsObjectArg::New(alloc(), argsObj, i); + + osrBlock->add(osrv); + osrBlock->initSlot(slot, osrv); + } else { + MParameter* arg = MParameter::New(alloc(), i, nullptr); + osrBlock->add(arg); + osrBlock->initSlot(slot, arg); + } + } + } + + // Initialize locals. + for (uint32_t i = 0; i < info().nlocals(); i++) { + uint32_t slot = info().localSlot(i); + ptrdiff_t offset = BaselineFrame::reverseOffsetOfLocal(i); + + MOsrValue* osrv = MOsrValue::New(alloc().fallible(), entry, offset); + if (!osrv) + return nullptr; + osrBlock->add(osrv); + osrBlock->initSlot(slot, osrv); + } + + // Initialize stack. + uint32_t numStackSlots = preheader->stackDepth() - info().firstStackSlot(); + for (uint32_t i = 0; i < numStackSlots; i++) { + uint32_t slot = info().stackSlot(i); + ptrdiff_t offset = BaselineFrame::reverseOffsetOfLocal(info().nlocals() + i); + + MOsrValue* osrv = MOsrValue::New(alloc().fallible(), entry, offset); + if (!osrv) + return nullptr; + osrBlock->add(osrv); + osrBlock->initSlot(slot, osrv); + } + + // Create an MStart to hold the first valid MResumePoint. + MStart* start = MStart::New(alloc()); + osrBlock->add(start); + + // MOsrValue instructions are infallible, so the first MResumePoint must + // occur after they execute, at the point of the MStart. + if (!resumeAt(start, loopEntry)) + return nullptr; + + // Link the same MResumePoint from the MStart to each MOsrValue. + // This causes logic in ShouldSpecializeInput() to not replace Uses with + // Unboxes in the MResumePiont, so that the MStart always sees Values. + if (!osrBlock->linkOsrValues(start)) + return nullptr; + + // Clone types of the other predecessor of the pre-header to the osr block, + // such as pre-header phi's won't discard specialized type of the + // predecessor. + MOZ_ASSERT(predecessor->stackDepth() == osrBlock->stackDepth()); + MOZ_ASSERT(info().environmentChainSlot() == 0); + + // Treat the OSR values as having the same type as the existing values + // coming in to the loop. These will be fixed up with appropriate + // unboxing and type barriers in finishLoop, once the possible types + // at the loop header are known. + for (uint32_t i = info().startArgSlot(); i < osrBlock->stackDepth(); i++) { + MDefinition* existing = current->getSlot(i); + MDefinition* def = osrBlock->getSlot(i); + MOZ_ASSERT_IF(!needsArgsObj || !info().isSlotAliased(i), def->type() == MIRType::Value); + + // Aliased slots are never accessed, since they need to go through + // the callobject. No need to type them here. + if (info().isSlotAliased(i)) + continue; + + def->setResultType(existing->type()); + def->setResultTypeSet(existing->resultTypeSet()); + } + + // Finish the osrBlock. + osrBlock->end(MGoto::New(alloc(), preheader)); + if (!preheader->addPredecessor(alloc(), osrBlock)) + return nullptr; + graph().setOsrBlock(osrBlock); + + return preheader; +} + +MBasicBlock* +IonBuilder::newPendingLoopHeader(MBasicBlock* predecessor, jsbytecode* pc, bool osr, bool canOsr, + unsigned stackPhiCount) +{ + loopDepth_++; + // If this site can OSR, all values on the expression stack are part of the loop. + if (canOsr) + stackPhiCount = predecessor->stackDepth() - info().firstStackSlot(); + MBasicBlock* block = MBasicBlock::NewPendingLoopHeader(graph(), info(), predecessor, + bytecodeSite(pc), stackPhiCount); + if (!addBlock(block, loopDepth_)) + return nullptr; + + if (osr) { + // Incorporate type information from the OSR frame into the loop + // header. The OSR frame may have unexpected types due to type changes + // within the loop body or due to incomplete profiling information, + // in which case this may avoid restarts of loop analysis or bailouts + // during the OSR itself. + + MOZ_ASSERT(info().firstLocalSlot() - info().firstArgSlot() == + baselineFrame_->argTypes.length()); + MOZ_ASSERT(block->stackDepth() - info().firstLocalSlot() == + baselineFrame_->varTypes.length()); + + // Unbox the MOsrValue if it is known to be unboxable. + for (uint32_t i = info().startArgSlot(); i < block->stackDepth(); i++) { + + // The value of aliased args and slots are in the callobject. So we can't + // the value from the baseline frame. + if (info().isSlotAliased(i)) + continue; + + MPhi* phi = block->getSlot(i)->toPhi(); + + // Get the type from the baseline frame. + TypeSet::Type existingType = TypeSet::UndefinedType(); + uint32_t arg = i - info().firstArgSlot(); + uint32_t var = i - info().firstLocalSlot(); + if (info().funMaybeLazy() && i == info().thisSlot()) + existingType = baselineFrame_->thisType; + else if (arg < info().nargs()) + existingType = baselineFrame_->argTypes[arg]; + else + existingType = baselineFrame_->varTypes[var]; + + if (existingType.isSingletonUnchecked()) + checkNurseryObject(existingType.singleton()); + + // Extract typeset from value. + LifoAlloc* lifoAlloc = alloc().lifoAlloc(); + TemporaryTypeSet* typeSet = + lifoAlloc->new_<TemporaryTypeSet>(lifoAlloc, existingType); + if (!typeSet) + return nullptr; + MIRType type = typeSet->getKnownMIRType(); + if (!phi->addBackedgeType(alloc(), type, typeSet)) + return nullptr; + } + } + + return block; +} + +MTest* +IonBuilder::newTest(MDefinition* ins, MBasicBlock* ifTrue, MBasicBlock* ifFalse) +{ + MTest* test = MTest::New(alloc(), ins, ifTrue, ifFalse); + test->cacheOperandMightEmulateUndefined(constraints()); + return test; +} + +// A resume point is a mapping of stack slots to MDefinitions. It is used to +// capture the environment such that if a guard fails, and IonMonkey needs +// to exit back to the interpreter, the interpreter state can be +// reconstructed. +// +// We capture stack state at critical points: +// * (1) At the beginning of every basic block. +// * (2) After every effectful operation. +// +// As long as these two properties are maintained, instructions can +// be moved, hoisted, or, eliminated without problems, and ops without side +// effects do not need to worry about capturing state at precisely the +// right point in time. +// +// Effectful instructions, of course, need to capture state after completion, +// where the interpreter will not attempt to repeat the operation. For this, +// ResumeAfter must be used. The state is attached directly to the effectful +// instruction to ensure that no intermediate instructions could be injected +// in between by a future analysis pass. +// +// During LIR construction, if an instruction can bail back to the interpreter, +// we create an LSnapshot, which uses the last known resume point to request +// register/stack assignments for every live value. +bool +IonBuilder::resume(MInstruction* ins, jsbytecode* pc, MResumePoint::Mode mode) +{ + MOZ_ASSERT(ins->isEffectful() || !ins->isMovable()); + + MResumePoint* resumePoint = MResumePoint::New(alloc(), ins->block(), pc, + mode); + if (!resumePoint) { + abortReason_ = AbortReason_Alloc; + return false; + } + ins->setResumePoint(resumePoint); + return true; +} + +bool +IonBuilder::resumeAt(MInstruction* ins, jsbytecode* pc) +{ + return resume(ins, pc, MResumePoint::ResumeAt); +} + +bool +IonBuilder::resumeAfter(MInstruction* ins) +{ + return resume(ins, pc, MResumePoint::ResumeAfter); +} + +bool +IonBuilder::maybeInsertResume() +{ + // Create a resume point at the current position, without an existing + // effectful instruction. This resume point is not necessary for correct + // behavior (see above), but is added to avoid holding any values from the + // previous resume point which are now dead. This shortens the live ranges + // of such values and improves register allocation. + // + // This optimization is not performed outside of loop bodies, where good + // register allocation is not as critical, in order to avoid creating + // excessive resume points. + + if (loopDepth_ == 0) + return true; + + MNop* ins = MNop::New(alloc()); + current->add(ins); + + return resumeAfter(ins); +} + +void +IonBuilder::maybeMarkEmpty(MDefinition* ins) +{ + MOZ_ASSERT(ins->type() == MIRType::Value); + + // When one of the operands has no type information, mark the output + // as having no possible types too. This is to avoid degrading + // subsequent analysis. + for (size_t i = 0; i < ins->numOperands(); i++) { + if (!ins->emptyResultTypeSet()) + continue; + + TemporaryTypeSet* types = alloc().lifoAlloc()->new_<TemporaryTypeSet>(); + if (types) + ins->setResultTypeSet(types); + } +} + +// Return whether property lookups can be performed effectlessly on clasp. +static bool +ClassHasEffectlessLookup(const Class* clasp) +{ + return (clasp == &UnboxedPlainObject::class_) || + (clasp == &UnboxedArrayObject::class_) || + IsTypedObjectClass(clasp) || + (clasp->isNative() && !clasp->getOpsLookupProperty()); +} + +// Return whether an object might have a property for name which is not +// accounted for by type information. +static bool +ObjectHasExtraOwnProperty(CompileCompartment* comp, TypeSet::ObjectKey* object, jsid id) +{ + // Some typed object properties are not reflected in type information. + if (object->isGroup() && object->group()->maybeTypeDescr()) + return object->group()->typeDescr().hasProperty(comp->runtime()->names(), id); + + const Class* clasp = object->clasp(); + + // Array |length| properties are not reflected in type information. + if (clasp == &ArrayObject::class_) + return JSID_IS_ATOM(id, comp->runtime()->names().length); + + // Resolve hooks can install new properties on objects on demand. + JSObject* singleton = object->isSingleton() ? object->singleton() : nullptr; + return ClassMayResolveId(comp->runtime()->names(), clasp, id, singleton); +} + +void +IonBuilder::insertRecompileCheck() +{ + // No need for recompile checks if this is the highest optimization level. + OptimizationLevel curLevel = optimizationInfo().level(); + if (IonOptimizations.isLastLevel(curLevel)) + return; + + // Add recompile check. + + // Get the topmost builder. The topmost script will get recompiled when + // warm-up counter is high enough to justify a higher optimization level. + IonBuilder* topBuilder = outermostBuilder(); + + // Add recompile check to recompile when the warm-up count reaches the + // threshold of the next optimization level. + OptimizationLevel nextLevel = IonOptimizations.nextLevel(curLevel); + const OptimizationInfo* info = IonOptimizations.get(nextLevel); + uint32_t warmUpThreshold = info->compilerWarmUpThreshold(topBuilder->script()); + MRecompileCheck* check = MRecompileCheck::New(alloc(), topBuilder->script(), warmUpThreshold, + MRecompileCheck::RecompileCheck_OptimizationLevel); + current->add(check); +} + +JSObject* +IonBuilder::testSingletonProperty(JSObject* obj, jsid id) +{ + // We would like to completely no-op property/global accesses which can + // produce only a particular JSObject. When indicating the access result is + // definitely an object, type inference does not account for the + // possibility that the property is entirely missing from the input object + // and its prototypes (if this happens, a semantic trigger would be hit and + // the pushed types updated, even if there is no type barrier). + // + // If the access definitely goes through obj, either directly or on the + // prototype chain, and the object has singleton type, then the type + // information for that property reflects the value that will definitely be + // read on accesses to the object. If the property is later deleted or + // reconfigured as a getter/setter then the type information for the + // property will change and trigger invalidation. + + while (obj) { + if (!ClassHasEffectlessLookup(obj->getClass())) + return nullptr; + + TypeSet::ObjectKey* objKey = TypeSet::ObjectKey::get(obj); + if (analysisContext) + objKey->ensureTrackedProperty(analysisContext, id); + + if (objKey->unknownProperties()) + return nullptr; + + HeapTypeSetKey property = objKey->property(id); + if (property.isOwnProperty(constraints())) { + if (obj->isSingleton()) + return property.singleton(constraints()); + return nullptr; + } + + if (ObjectHasExtraOwnProperty(compartment, objKey, id)) + return nullptr; + + obj = checkNurseryObject(obj->staticPrototype()); + } + + return nullptr; +} + +JSObject* +IonBuilder::testSingletonPropertyTypes(MDefinition* obj, jsid id) +{ + // As for TestSingletonProperty, but the input is any value in a type set + // rather than a specific object. + + TemporaryTypeSet* types = obj->resultTypeSet(); + if (types && types->unknownObject()) + return nullptr; + + JSObject* objectSingleton = types ? types->maybeSingleton() : nullptr; + if (objectSingleton) + return testSingletonProperty(objectSingleton, id); + + MIRType objType = obj->type(); + if (objType == MIRType::Value && types) + objType = types->getKnownMIRType(); + + JSProtoKey key; + switch (objType) { + case MIRType::String: + key = JSProto_String; + break; + + case MIRType::Symbol: + key = JSProto_Symbol; + break; + + case MIRType::Int32: + case MIRType::Double: + key = JSProto_Number; + break; + + case MIRType::Boolean: + key = JSProto_Boolean; + break; + + case MIRType::Object: { + if (!types) + return nullptr; + + // For property accesses which may be on many objects, we just need to + // find a prototype common to all the objects; if that prototype + // has the singleton property, the access will not be on a missing property. + JSObject* singleton = nullptr; + for (unsigned i = 0; i < types->getObjectCount(); i++) { + TypeSet::ObjectKey* key = types->getObject(i); + if (!key) + continue; + if (analysisContext) + key->ensureTrackedProperty(analysisContext, id); + + const Class* clasp = key->clasp(); + if (!ClassHasEffectlessLookup(clasp) || ObjectHasExtraOwnProperty(compartment, key, id)) + return nullptr; + if (key->unknownProperties()) + return nullptr; + HeapTypeSetKey property = key->property(id); + if (property.isOwnProperty(constraints())) + return nullptr; + + if (JSObject* proto = checkNurseryObject(key->proto().toObjectOrNull())) { + // Test this type. + JSObject* thisSingleton = testSingletonProperty(proto, id); + if (!thisSingleton) + return nullptr; + if (singleton) { + if (thisSingleton != singleton) + return nullptr; + } else { + singleton = thisSingleton; + } + } else { + // Can't be on the prototype chain with no prototypes... + return nullptr; + } + } + return singleton; + } + default: + return nullptr; + } + + JSObject* proto = GetBuiltinPrototypePure(&script()->global(), key); + if (proto) + return testSingletonProperty(proto, id); + + return nullptr; +} + +ResultWithOOM<bool> +IonBuilder::testNotDefinedProperty(MDefinition* obj, jsid id) +{ + TemporaryTypeSet* types = obj->resultTypeSet(); + if (!types || types->unknownObject() || types->getKnownMIRType() != MIRType::Object) + return ResultWithOOM<bool>::ok(false); + + for (unsigned i = 0, count = types->getObjectCount(); i < count; i++) { + TypeSet::ObjectKey* key = types->getObject(i); + if (!key) + continue; + + while (true) { + if (!alloc().ensureBallast()) + return ResultWithOOM<bool>::fail(); + + if (!key->hasStableClassAndProto(constraints()) || key->unknownProperties()) + return ResultWithOOM<bool>::ok(false); + + const Class* clasp = key->clasp(); + if (!ClassHasEffectlessLookup(clasp) || ObjectHasExtraOwnProperty(compartment, key, id)) + return ResultWithOOM<bool>::ok(false); + + // If the object is a singleton, we can do a lookup now to avoid + // unnecessary invalidations later on, in case the property types + // have not yet been instantiated. + if (key->isSingleton() && + key->singleton()->is<NativeObject>() && + key->singleton()->as<NativeObject>().lookupPure(id)) + { + return ResultWithOOM<bool>::ok(false); + } + + HeapTypeSetKey property = key->property(id); + if (property.isOwnProperty(constraints())) + return ResultWithOOM<bool>::ok(false); + + JSObject* proto = checkNurseryObject(key->proto().toObjectOrNull()); + if (!proto) + break; + key = TypeSet::ObjectKey::get(proto); + } + } + + return ResultWithOOM<bool>::ok(true); +} + +bool +IonBuilder::pushTypeBarrier(MDefinition* def, TemporaryTypeSet* observed, BarrierKind kind) +{ + MOZ_ASSERT(def == current->peek(-1)); + + MDefinition* replace = addTypeBarrier(current->pop(), observed, kind); + if (!replace) + return false; + + current->push(replace); + return true; +} + +// Given an observed type set, annotates the IR as much as possible: +// (1) If no type information is provided, the given value is returned. +// (2) If a single type definitely exists, and no type barrier is needed, +// then an infallible unbox instruction is returned. +// (3) If a type barrier is needed, but has an unknown type set, the given +// value is returned. +// (4) Lastly, a type barrier instruction is added and returned. +MDefinition* +IonBuilder::addTypeBarrier(MDefinition* def, TemporaryTypeSet* observed, BarrierKind kind, + MTypeBarrier** pbarrier) +{ + // Barriers are never needed for instructions whose result will not be used. + if (BytecodeIsPopped(pc)) + return def; + + // If the instruction has no side effects, we'll resume the entire operation. + // The actual type barrier will occur in the interpreter. If the + // instruction is effectful, even if it has a singleton type, there + // must be a resume point capturing the original def, and resuming + // to that point will explicitly monitor the new type. + if (kind == BarrierKind::NoBarrier) { + MDefinition* replace = ensureDefiniteType(def, observed->getKnownMIRType()); + replace->setResultTypeSet(observed); + return replace; + } + + if (observed->unknown()) + return def; + + MTypeBarrier* barrier = MTypeBarrier::New(alloc(), def, observed, kind); + current->add(barrier); + + if (pbarrier) + *pbarrier = barrier; + + if (barrier->type() == MIRType::Undefined) + return constant(UndefinedValue()); + if (barrier->type() == MIRType::Null) + return constant(NullValue()); + + return barrier; +} + +bool +IonBuilder::pushDOMTypeBarrier(MInstruction* ins, TemporaryTypeSet* observed, JSFunction* func) +{ + MOZ_ASSERT(func && func->isNative() && func->jitInfo()); + + const JSJitInfo* jitinfo = func->jitInfo(); + bool barrier = DOMCallNeedsBarrier(jitinfo, observed); + // Need to be a bit careful: if jitinfo->returnType is JSVAL_TYPE_DOUBLE but + // types->getKnownMIRType() is MIRType::Int32, then don't unconditionally + // unbox as a double. Instead, go ahead and barrier on having an int type, + // since we know we need a barrier anyway due to the type mismatch. This is + // the only situation in which TI actually has more information about the + // JSValueType than codegen can, short of jitinfo->returnType just being + // JSVAL_TYPE_UNKNOWN. + MDefinition* replace = ins; + if (jitinfo->returnType() != JSVAL_TYPE_DOUBLE || + observed->getKnownMIRType() != MIRType::Int32) { + replace = ensureDefiniteType(ins, MIRTypeFromValueType(jitinfo->returnType())); + if (replace != ins) { + current->pop(); + current->push(replace); + } + } else { + MOZ_ASSERT(barrier); + } + + return pushTypeBarrier(replace, observed, + barrier ? BarrierKind::TypeSet : BarrierKind::NoBarrier); +} + +MDefinition* +IonBuilder::ensureDefiniteType(MDefinition* def, MIRType definiteType) +{ + MInstruction* replace; + switch (definiteType) { + case MIRType::Undefined: + def->setImplicitlyUsedUnchecked(); + replace = MConstant::New(alloc(), UndefinedValue()); + break; + + case MIRType::Null: + def->setImplicitlyUsedUnchecked(); + replace = MConstant::New(alloc(), NullValue()); + break; + + case MIRType::Value: + return def; + + default: { + if (def->type() != MIRType::Value) { + if (def->type() == MIRType::Int32 && definiteType == MIRType::Double) { + replace = MToDouble::New(alloc(), def); + break; + } + MOZ_ASSERT(def->type() == definiteType); + return def; + } + replace = MUnbox::New(alloc(), def, definiteType, MUnbox::Infallible); + break; + } + } + + current->add(replace); + return replace; +} + +MDefinition* +IonBuilder::ensureDefiniteTypeSet(MDefinition* def, TemporaryTypeSet* types) +{ + // We cannot arbitrarily add a typeset to a definition. It can be shared + // in another path. So we always need to create a new MIR. + + // Use ensureDefiniteType to do unboxing. If that happened the type can + // be added on the newly created unbox operation. + MDefinition* replace = ensureDefiniteType(def, types->getKnownMIRType()); + if (replace != def) { + replace->setResultTypeSet(types); + return replace; + } + + // Don't replace if input type is more accurate than given typeset. + if (def->type() != types->getKnownMIRType()) { + MOZ_ASSERT(types->getKnownMIRType() == MIRType::Value); + return def; + } + + // Create a NOP mir instruction to filter the typeset. + MFilterTypeSet* filter = MFilterTypeSet::New(alloc(), def, types); + current->add(filter); + return filter; +} + +static size_t +NumFixedSlots(JSObject* object) +{ + // Note: we can't use object->numFixedSlots() here, as this will read the + // shape and can race with the main thread if we are building off thread. + // The allocation kind and object class (which goes through the type) can + // be read freely, however. + gc::AllocKind kind = object->asTenured().getAllocKind(); + return gc::GetGCKindSlots(kind, object->getClass()); +} + +static bool +IsUninitializedGlobalLexicalSlot(JSObject* obj, PropertyName* name) +{ + LexicalEnvironmentObject &globalLexical = obj->as<LexicalEnvironmentObject>(); + MOZ_ASSERT(globalLexical.isGlobal()); + Shape* shape = globalLexical.lookupPure(name); + if (!shape) + return false; + return globalLexical.getSlot(shape->slot()).isMagic(JS_UNINITIALIZED_LEXICAL); +} + +bool +IonBuilder::getStaticName(JSObject* staticObject, PropertyName* name, bool* psucceeded, + MDefinition* lexicalCheck) +{ + MOZ_ASSERT(*psucceeded == false); + + jsid id = NameToId(name); + + bool isGlobalLexical = staticObject->is<LexicalEnvironmentObject>() && + staticObject->as<LexicalEnvironmentObject>().isGlobal(); + MOZ_ASSERT(isGlobalLexical || + staticObject->is<GlobalObject>() || + staticObject->is<CallObject>() || + staticObject->is<ModuleEnvironmentObject>()); + MOZ_ASSERT(staticObject->isSingleton()); + + *psucceeded = true; + + // Always emit the lexical check. This could be optimized, but is + // currently not for simplicity's sake. + if (lexicalCheck) { + *psucceeded = false; + return true; + } + + TypeSet::ObjectKey* staticKey = TypeSet::ObjectKey::get(staticObject); + if (analysisContext) + staticKey->ensureTrackedProperty(analysisContext, NameToId(name)); + + if (staticKey->unknownProperties()) { + *psucceeded = false; + return true; + } + + HeapTypeSetKey property = staticKey->property(id); + if (!property.maybeTypes() || + !property.maybeTypes()->definiteProperty() || + property.nonData(constraints())) + { + // The property has been reconfigured as non-configurable, non-enumerable + // or non-writable. + *psucceeded = false; + return true; + } + + // Don't optimize global lexical bindings if they aren't initialized at + // compile time. + if (isGlobalLexical && IsUninitializedGlobalLexicalSlot(staticObject, name)) { + *psucceeded = false; + return true; + } + + TemporaryTypeSet* types = bytecodeTypes(pc); + BarrierKind barrier = PropertyReadNeedsTypeBarrier(analysisContext, constraints(), staticKey, + name, types, /* updateObserved = */ true); + + if (barrier == BarrierKind::NoBarrier) { + // Try to inline properties holding a known constant object. + JSObject* singleton = types->maybeSingleton(); + if (singleton) { + if (testSingletonProperty(staticObject, id) == singleton) { + pushConstant(ObjectValue(*singleton)); + return true; + } + } + + // Try to inline properties that have never been overwritten. + Value constantValue; + if (property.constant(constraints(), &constantValue)) { + pushConstant(constantValue); + return true; + } + } + + if (!loadStaticSlot(staticObject, barrier, types, property.maybeTypes()->definiteSlot())) { + *psucceeded = false; + return false; + } + + return true; +} + +bool +IonBuilder::loadStaticSlot(JSObject* staticObject, BarrierKind barrier, TemporaryTypeSet* types, + uint32_t slot) +{ + if (barrier == BarrierKind::NoBarrier) { + // Try to inline properties that can only have one value. + MIRType knownType = types->getKnownMIRType(); + if (knownType == MIRType::Undefined) { + pushConstant(UndefinedValue()); + return true; + } + if (knownType == MIRType::Null) { + pushConstant(NullValue()); + return true; + } + } + + MInstruction* obj = constant(ObjectValue(*staticObject)); + + MIRType rvalType = types->getKnownMIRType(); + if (barrier != BarrierKind::NoBarrier) + rvalType = MIRType::Value; + + return loadSlot(obj, slot, NumFixedSlots(staticObject), rvalType, barrier, types); +} + +// Whether a write of the given value may need a post-write barrier for GC purposes. +bool +jit::NeedsPostBarrier(MDefinition* value) +{ + if (!GetJitContext()->runtime->gcNursery().exists()) + return false; + return value->mightBeType(MIRType::Object); +} + +bool +IonBuilder::setStaticName(JSObject* staticObject, PropertyName* name) +{ + jsid id = NameToId(name); + + bool isGlobalLexical = staticObject->is<LexicalEnvironmentObject>() && + staticObject->as<LexicalEnvironmentObject>().isGlobal(); + MOZ_ASSERT(isGlobalLexical || + staticObject->is<GlobalObject>() || + staticObject->is<CallObject>()); + + MDefinition* value = current->peek(-1); + + TypeSet::ObjectKey* staticKey = TypeSet::ObjectKey::get(staticObject); + if (staticKey->unknownProperties()) + return jsop_setprop(name); + + HeapTypeSetKey property = staticKey->property(id); + if (!property.maybeTypes() || + !property.maybeTypes()->definiteProperty() || + property.nonData(constraints()) || + property.nonWritable(constraints())) + { + // The property has been reconfigured as non-configurable, non-enumerable + // or non-writable. + return jsop_setprop(name); + } + + if (!CanWriteProperty(alloc(), constraints(), property, value)) + return jsop_setprop(name); + + // Don't optimize global lexical bindings if they aren't initialized at + // compile time. + if (isGlobalLexical && IsUninitializedGlobalLexicalSlot(staticObject, name)) + return jsop_setprop(name); + + current->pop(); + + // Pop the bound object on the stack. + MDefinition* obj = current->pop(); + MOZ_ASSERT(&obj->toConstant()->toObject() == staticObject); + + if (NeedsPostBarrier(value)) + current->add(MPostWriteBarrier::New(alloc(), obj, value)); + + // If the property has a known type, we may be able to optimize typed stores by not + // storing the type tag. + MIRType slotType = MIRType::None; + MIRType knownType = property.knownMIRType(constraints()); + if (knownType != MIRType::Value) + slotType = knownType; + + bool needsBarrier = property.needsBarrier(constraints()); + return storeSlot(obj, property.maybeTypes()->definiteSlot(), NumFixedSlots(staticObject), + value, needsBarrier, slotType); +} + +JSObject* +IonBuilder::testGlobalLexicalBinding(PropertyName* name) +{ + MOZ_ASSERT(JSOp(*pc) == JSOP_BINDGNAME || + JSOp(*pc) == JSOP_GETGNAME || + JSOp(*pc) == JSOP_SETGNAME || + JSOp(*pc) == JSOP_STRICTSETGNAME); + + // The global isn't the global lexical env's prototype, but its enclosing + // env. Test for the existence of |name| manually on the global lexical + // env. If it is not found, look for it on the global itself. + + NativeObject* obj = &script()->global().lexicalEnvironment(); + TypeSet::ObjectKey* lexicalKey = TypeSet::ObjectKey::get(obj); + jsid id = NameToId(name); + if (analysisContext) + lexicalKey->ensureTrackedProperty(analysisContext, id); + + // If the property is not found on the global lexical env but it is found + // on the global and is configurable, try to freeze the typeset for its + // non-existence. If we don't have type information then fail. + // + // In the case that it is found on the global but is non-configurable, + // the binding cannot be shadowed by a global lexical binding. + Maybe<HeapTypeSetKey> lexicalProperty; + if (!lexicalKey->unknownProperties()) + lexicalProperty.emplace(lexicalKey->property(id)); + Shape* shape = obj->lookupPure(name); + if (shape) { + if ((JSOp(*pc) != JSOP_GETGNAME && !shape->writable()) || + obj->getSlot(shape->slot()).isMagic(JS_UNINITIALIZED_LEXICAL)) + { + return nullptr; + } + } else { + shape = script()->global().lookupPure(name); + if (!shape || shape->configurable()) { + if (lexicalProperty.isSome()) + MOZ_ALWAYS_FALSE(lexicalProperty->isOwnProperty(constraints())); + else + return nullptr; + } + obj = &script()->global(); + } + + return obj; +} + +bool +IonBuilder::jsop_getgname(PropertyName* name) +{ + // Optimize undefined/NaN/Infinity first. We must ensure we handle these + // cases *exactly* like Baseline, because it's invalid to add an Ion IC or + // VM call (that might trigger invalidation) if there's no Baseline IC for + // this op. + if (name == names().undefined) { + pushConstant(UndefinedValue()); + return true; + } + if (name == names().NaN) { + pushConstant(compartment->runtime()->NaNValue()); + return true; + } + if (name == names().Infinity) { + pushConstant(compartment->runtime()->positiveInfinityValue()); + return true; + } + + if (JSObject* obj = testGlobalLexicalBinding(name)) { + bool emitted = false; + if (!getStaticName(obj, name, &emitted) || emitted) + return emitted; + + if (!forceInlineCaches() && obj->is<GlobalObject>()) { + TemporaryTypeSet* types = bytecodeTypes(pc); + MDefinition* globalObj = constant(ObjectValue(*obj)); + if (!getPropTryCommonGetter(&emitted, globalObj, name, types) || emitted) + return emitted; + } + } + + return jsop_getname(name); +} + +bool +IonBuilder::jsop_getname(PropertyName* name) +{ + MDefinition* object; + if (IsGlobalOp(JSOp(*pc)) && !script()->hasNonSyntacticScope()) { + MInstruction* global = constant(ObjectValue(script()->global().lexicalEnvironment())); + object = global; + } else { + current->push(current->environmentChain()); + object = current->pop(); + } + + MGetNameCache* ins; + if (JSOp(*GetNextPc(pc)) == JSOP_TYPEOF) + ins = MGetNameCache::New(alloc(), object, name, MGetNameCache::NAMETYPEOF); + else + ins = MGetNameCache::New(alloc(), object, name, MGetNameCache::NAME); + + current->add(ins); + current->push(ins); + + if (!resumeAfter(ins)) + return false; + + TemporaryTypeSet* types = bytecodeTypes(pc); + return pushTypeBarrier(ins, types, BarrierKind::TypeSet); +} + +bool +IonBuilder::jsop_intrinsic(PropertyName* name) +{ + TemporaryTypeSet* types = bytecodeTypes(pc); + + Value vp = UndefinedValue(); + // If the intrinsic value doesn't yet exist, we haven't executed this + // opcode yet, so we need to get it and monitor the result. + if (!script()->global().maybeExistingIntrinsicValue(name, &vp)) { + MCallGetIntrinsicValue* ins = MCallGetIntrinsicValue::New(alloc(), name); + + current->add(ins); + current->push(ins); + + if (!resumeAfter(ins)) + return false; + + return pushTypeBarrier(ins, types, BarrierKind::TypeSet); + } + + if (types->empty()) + types->addType(TypeSet::GetValueType(vp), alloc().lifoAlloc()); + + // Bake in the intrinsic, guaranteed to exist because a non-empty typeset + // means the intrinsic was successfully gotten in the VM call above. + // Assert that TI agrees with us on the type. + MOZ_ASSERT(types->hasType(TypeSet::GetValueType(vp))); + + pushConstant(vp); + return true; +} + +bool +IonBuilder::jsop_getimport(PropertyName* name) +{ + ModuleEnvironmentObject* env = GetModuleEnvironmentForScript(script()); + MOZ_ASSERT(env); + + Shape* shape; + ModuleEnvironmentObject* targetEnv; + MOZ_ALWAYS_TRUE(env->lookupImport(NameToId(name), &targetEnv, &shape)); + + PropertyName* localName = JSID_TO_STRING(shape->propid())->asAtom().asPropertyName(); + bool emitted = false; + if (!getStaticName(targetEnv, localName, &emitted)) + return false; + + if (!emitted) { + // This can happen if we don't have type information. + TypeSet::ObjectKey* staticKey = TypeSet::ObjectKey::get(targetEnv); + TemporaryTypeSet* types = bytecodeTypes(pc); + BarrierKind barrier = PropertyReadNeedsTypeBarrier(analysisContext, constraints(), staticKey, + name, types, /* updateObserved = */ true); + + if (!loadStaticSlot(targetEnv, barrier, types, shape->slot())) + return false; + } + + // In the rare case where this import hasn't been initialized already (we + // have an import cycle where modules reference each other's imports), emit + // a check. + if (targetEnv->getSlot(shape->slot()).isMagic(JS_UNINITIALIZED_LEXICAL)) { + MDefinition* checked = addLexicalCheck(current->pop()); + current->push(checked); + } + + return true; +} + +bool +IonBuilder::jsop_bindname(PropertyName* name) +{ + MDefinition* envChain; + if (analysis().usesEnvironmentChain()) { + envChain = current->environmentChain(); + } else { + // We take the slow path when trying to BINDGNAME a name that resolves + // to a 'const' or an uninitialized binding. + MOZ_ASSERT(JSOp(*pc) == JSOP_BINDGNAME); + envChain = constant(ObjectValue(script()->global().lexicalEnvironment())); + } + MBindNameCache* ins = MBindNameCache::New(alloc(), envChain, name, script(), pc); + + current->add(ins); + current->push(ins); + + return resumeAfter(ins); +} + +bool +IonBuilder::jsop_bindvar() +{ + MOZ_ASSERT(analysis().usesEnvironmentChain()); + MCallBindVar* ins = MCallBindVar::New(alloc(), current->environmentChain()); + current->add(ins); + current->push(ins); + return true; +} + +static MIRType +GetElemKnownType(bool needsHoleCheck, TemporaryTypeSet* types) +{ + MIRType knownType = types->getKnownMIRType(); + + // Null and undefined have no payload so they can't be specialized. + // Since folding null/undefined while building SSA is not safe (see the + // comment in IsPhiObservable), we just add an untyped load instruction + // and rely on pushTypeBarrier and DCE to replace it with a null/undefined + // constant. + if (knownType == MIRType::Undefined || knownType == MIRType::Null) + knownType = MIRType::Value; + + // Different architectures may want typed element reads which require + // hole checks to be done as either value or typed reads. + if (needsHoleCheck && !LIRGenerator::allowTypedElementHoleCheck()) + knownType = MIRType::Value; + + return knownType; +} + +bool +IonBuilder::jsop_getelem() +{ + startTrackingOptimizations(); + + MDefinition* index = current->pop(); + MDefinition* obj = current->pop(); + + trackTypeInfo(TrackedTypeSite::Receiver, obj->type(), obj->resultTypeSet()); + trackTypeInfo(TrackedTypeSite::Index, index->type(), index->resultTypeSet()); + + // Always use a call if we are performing analysis and not actually + // emitting code, to simplify later analysis. + if (info().isAnalysis() || shouldAbortOnPreliminaryGroups(obj)) { + MInstruction* ins = MCallGetElement::New(alloc(), obj, index); + + current->add(ins); + current->push(ins); + + if (!resumeAfter(ins)) + return false; + + TemporaryTypeSet* types = bytecodeTypes(pc); + return pushTypeBarrier(ins, types, BarrierKind::TypeSet); + } + + obj = maybeUnboxForPropertyAccess(obj); + if (obj->type() == MIRType::Object) + obj = convertUnboxedObjects(obj); + + bool emitted = false; + + if (!forceInlineCaches()) { + trackOptimizationAttempt(TrackedStrategy::GetElem_TypedObject); + if (!getElemTryTypedObject(&emitted, obj, index) || emitted) + return emitted; + + // Note: no trackOptimizationAttempt call is needed, getElemTryGetProp + // will call it. + if (!getElemTryGetProp(&emitted, obj, index) || emitted) + return emitted; + + trackOptimizationAttempt(TrackedStrategy::GetElem_Dense); + if (!getElemTryDense(&emitted, obj, index) || emitted) + return emitted; + + trackOptimizationAttempt(TrackedStrategy::GetElem_TypedStatic); + if (!getElemTryTypedStatic(&emitted, obj, index) || emitted) + return emitted; + + trackOptimizationAttempt(TrackedStrategy::GetElem_TypedArray); + if (!getElemTryTypedArray(&emitted, obj, index) || emitted) + return emitted; + + trackOptimizationAttempt(TrackedStrategy::GetElem_String); + if (!getElemTryString(&emitted, obj, index) || emitted) + return emitted; + + trackOptimizationAttempt(TrackedStrategy::GetElem_Arguments); + if (!getElemTryArguments(&emitted, obj, index) || emitted) + return emitted; + + trackOptimizationAttempt(TrackedStrategy::GetElem_ArgumentsInlined); + if (!getElemTryArgumentsInlined(&emitted, obj, index) || emitted) + return emitted; + } + + if (script()->argumentsHasVarBinding() && obj->mightBeType(MIRType::MagicOptimizedArguments)) + return abort("Type is not definitely lazy arguments."); + + trackOptimizationAttempt(TrackedStrategy::GetElem_InlineCache); + if (!getElemTryCache(&emitted, obj, index) || emitted) + return emitted; + + // Emit call. + MInstruction* ins = MCallGetElement::New(alloc(), obj, index); + + current->add(ins); + current->push(ins); + + if (!resumeAfter(ins)) + return false; + + if (*pc == JSOP_CALLELEM && IsNullOrUndefined(obj->type())) { + // Due to inlining, it's possible the observed TypeSet is non-empty, + // even though we know |obj| is null/undefined and the MCallGetElement + // will throw. Don't push a TypeBarrier in this case, to avoid + // inlining the following (unreachable) JSOP_CALL. + return true; + } + + TemporaryTypeSet* types = bytecodeTypes(pc); + return pushTypeBarrier(ins, types, BarrierKind::TypeSet); +} + +bool +IonBuilder::getElemTryTypedObject(bool* emitted, MDefinition* obj, MDefinition* index) +{ + MOZ_ASSERT(*emitted == false); + + // The next several failures are all due to types not predicting that we + // are definitely doing a getelem access on a typed object. + trackOptimizationOutcome(TrackedOutcome::AccessNotTypedObject); + + TypedObjectPrediction objPrediction = typedObjectPrediction(obj); + if (objPrediction.isUseless()) + return true; + + if (!objPrediction.ofArrayKind()) + return true; + + TypedObjectPrediction elemPrediction = objPrediction.arrayElementType(); + if (elemPrediction.isUseless()) + return true; + + uint32_t elemSize; + if (!elemPrediction.hasKnownSize(&elemSize)) + return true; + + switch (elemPrediction.kind()) { + case type::Simd: + // FIXME (bug 894105): load into a MIRType::float32x4 etc + trackOptimizationOutcome(TrackedOutcome::GenericFailure); + return true; + + case type::Struct: + case type::Array: + return getElemTryComplexElemOfTypedObject(emitted, + obj, + index, + objPrediction, + elemPrediction, + elemSize); + case type::Scalar: + return getElemTryScalarElemOfTypedObject(emitted, + obj, + index, + objPrediction, + elemPrediction, + elemSize); + + case type::Reference: + return getElemTryReferenceElemOfTypedObject(emitted, + obj, + index, + objPrediction, + elemPrediction); + } + + MOZ_CRASH("Bad kind"); +} + +bool +IonBuilder::checkTypedObjectIndexInBounds(uint32_t elemSize, + MDefinition* obj, + MDefinition* index, + TypedObjectPrediction objPrediction, + LinearSum* indexAsByteOffset) +{ + // Ensure index is an integer. + MInstruction* idInt32 = MToInt32::New(alloc(), index); + current->add(idInt32); + + // If we know the length statically from the type, just embed it. + // Otherwise, load it from the appropriate reserved slot on the + // typed object. We know it's an int32, so we can convert from + // Value to int32 using truncation. + int32_t lenOfAll; + MDefinition* length; + if (objPrediction.hasKnownArrayLength(&lenOfAll)) { + length = constantInt(lenOfAll); + + // If we are not loading the length from the object itself, only + // optimize if the array buffer can never be a detached array buffer. + TypeSet::ObjectKey* globalKey = TypeSet::ObjectKey::get(&script()->global()); + if (globalKey->hasFlags(constraints(), OBJECT_FLAG_TYPED_OBJECT_HAS_DETACHED_BUFFER)) { + trackOptimizationOutcome(TrackedOutcome::TypedObjectHasDetachedBuffer); + return false; + } + } else { + trackOptimizationOutcome(TrackedOutcome::TypedObjectArrayRange); + return false; + } + + index = addBoundsCheck(idInt32, length); + + return indexAsByteOffset->add(index, AssertedCast<int32_t>(elemSize)); +} + +bool +IonBuilder::getElemTryScalarElemOfTypedObject(bool* emitted, + MDefinition* obj, + MDefinition* index, + TypedObjectPrediction objPrediction, + TypedObjectPrediction elemPrediction, + uint32_t elemSize) +{ + MOZ_ASSERT(objPrediction.ofArrayKind()); + + // Must always be loading the same scalar type + ScalarTypeDescr::Type elemType = elemPrediction.scalarType(); + MOZ_ASSERT(elemSize == ScalarTypeDescr::alignment(elemType)); + + LinearSum indexAsByteOffset(alloc()); + if (!checkTypedObjectIndexInBounds(elemSize, obj, index, objPrediction, &indexAsByteOffset)) + return true; + + trackOptimizationSuccess(); + *emitted = true; + + return pushScalarLoadFromTypedObject(obj, indexAsByteOffset, elemType); +} + +bool +IonBuilder::getElemTryReferenceElemOfTypedObject(bool* emitted, + MDefinition* obj, + MDefinition* index, + TypedObjectPrediction objPrediction, + TypedObjectPrediction elemPrediction) +{ + MOZ_ASSERT(objPrediction.ofArrayKind()); + + ReferenceTypeDescr::Type elemType = elemPrediction.referenceType(); + uint32_t elemSize = ReferenceTypeDescr::size(elemType); + + LinearSum indexAsByteOffset(alloc()); + if (!checkTypedObjectIndexInBounds(elemSize, obj, index, objPrediction, &indexAsByteOffset)) + return true; + + trackOptimizationSuccess(); + *emitted = true; + + return pushReferenceLoadFromTypedObject(obj, indexAsByteOffset, elemType, nullptr); +} + +bool +IonBuilder::pushScalarLoadFromTypedObject(MDefinition* obj, + const LinearSum& byteOffset, + ScalarTypeDescr::Type elemType) +{ + uint32_t size = ScalarTypeDescr::size(elemType); + MOZ_ASSERT(size == ScalarTypeDescr::alignment(elemType)); + + // Find location within the owner object. + MDefinition* elements; + MDefinition* scaledOffset; + int32_t adjustment; + loadTypedObjectElements(obj, byteOffset, size, &elements, &scaledOffset, &adjustment); + + // Load the element. + MLoadUnboxedScalar* load = MLoadUnboxedScalar::New(alloc(), elements, scaledOffset, + elemType, + DoesNotRequireMemoryBarrier, + adjustment); + current->add(load); + current->push(load); + + // If we are reading in-bounds elements, we can use knowledge about + // the array type to determine the result type, even if the opcode has + // never executed. The known pushed type is only used to distinguish + // uint32 reads that may produce either doubles or integers. + TemporaryTypeSet* resultTypes = bytecodeTypes(pc); + bool allowDouble = resultTypes->hasType(TypeSet::DoubleType()); + + // Note: knownType is not necessarily in resultTypes; e.g. if we + // have only observed integers coming out of float array. + MIRType knownType = MIRTypeForTypedArrayRead(elemType, allowDouble); + + // Note: we can ignore the type barrier here, we know the type must + // be valid and unbarriered. Also, need not set resultTypeSet, + // because knownType is scalar and a resultTypeSet would provide + // no useful additional info. + load->setResultType(knownType); + + return true; +} + +bool +IonBuilder::pushReferenceLoadFromTypedObject(MDefinition* typedObj, + const LinearSum& byteOffset, + ReferenceTypeDescr::Type type, + PropertyName* name) +{ + // Find location within the owner object. + MDefinition* elements; + MDefinition* scaledOffset; + int32_t adjustment; + uint32_t alignment = ReferenceTypeDescr::alignment(type); + loadTypedObjectElements(typedObj, byteOffset, alignment, &elements, &scaledOffset, &adjustment); + + TemporaryTypeSet* observedTypes = bytecodeTypes(pc); + + MInstruction* load = nullptr; // initialize to silence GCC warning + BarrierKind barrier = PropertyReadNeedsTypeBarrier(analysisContext, constraints(), + typedObj, name, observedTypes); + + switch (type) { + case ReferenceTypeDescr::TYPE_ANY: { + // Make sure the barrier reflects the possibility of reading undefined. + bool bailOnUndefined = barrier == BarrierKind::NoBarrier && + !observedTypes->hasType(TypeSet::UndefinedType()); + if (bailOnUndefined) + barrier = BarrierKind::TypeTagOnly; + load = MLoadElement::New(alloc(), elements, scaledOffset, false, false, adjustment); + break; + } + case ReferenceTypeDescr::TYPE_OBJECT: { + // Make sure the barrier reflects the possibility of reading null. When + // there is no other barrier needed we include the null bailout with + // MLoadUnboxedObjectOrNull, which avoids the need to box the result + // for a type barrier instruction. + MLoadUnboxedObjectOrNull::NullBehavior nullBehavior; + if (barrier == BarrierKind::NoBarrier && !observedTypes->hasType(TypeSet::NullType())) + nullBehavior = MLoadUnboxedObjectOrNull::BailOnNull; + else + nullBehavior = MLoadUnboxedObjectOrNull::HandleNull; + load = MLoadUnboxedObjectOrNull::New(alloc(), elements, scaledOffset, nullBehavior, + adjustment); + break; + } + case ReferenceTypeDescr::TYPE_STRING: { + load = MLoadUnboxedString::New(alloc(), elements, scaledOffset, adjustment); + observedTypes->addType(TypeSet::StringType(), alloc().lifoAlloc()); + break; + } + } + + current->add(load); + current->push(load); + + return pushTypeBarrier(load, observedTypes, barrier); +} + +bool +IonBuilder::getElemTryComplexElemOfTypedObject(bool* emitted, + MDefinition* obj, + MDefinition* index, + TypedObjectPrediction objPrediction, + TypedObjectPrediction elemPrediction, + uint32_t elemSize) +{ + MOZ_ASSERT(objPrediction.ofArrayKind()); + + MDefinition* type = loadTypedObjectType(obj); + MDefinition* elemTypeObj = typeObjectForElementFromArrayStructType(type); + + LinearSum indexAsByteOffset(alloc()); + if (!checkTypedObjectIndexInBounds(elemSize, obj, index, objPrediction, &indexAsByteOffset)) + return true; + + return pushDerivedTypedObject(emitted, obj, indexAsByteOffset, + elemPrediction, elemTypeObj); +} + +bool +IonBuilder::pushDerivedTypedObject(bool* emitted, + MDefinition* obj, + const LinearSum& baseByteOffset, + TypedObjectPrediction derivedPrediction, + MDefinition* derivedTypeObj) +{ + // Find location within the owner object. + MDefinition* owner; + LinearSum ownerByteOffset(alloc()); + loadTypedObjectData(obj, &owner, &ownerByteOffset); + + if (!ownerByteOffset.add(baseByteOffset, 1)) + setForceAbort(); + + MDefinition* offset = ConvertLinearSum(alloc(), current, ownerByteOffset, + /* convertConstant = */ true); + + // Create the derived typed object. + MInstruction* derivedTypedObj = MNewDerivedTypedObject::New(alloc(), + derivedPrediction, + derivedTypeObj, + owner, + offset); + current->add(derivedTypedObj); + current->push(derivedTypedObj); + + // Determine (if possible) the class/proto that `derivedTypedObj` will + // have. For derived typed objects, the opacity will be the same as the + // incoming object from which the derived typed object is, well, derived. + // The prototype will be determined based on the type descriptor (and is + // immutable). + TemporaryTypeSet* objTypes = obj->resultTypeSet(); + const Class* expectedClass = nullptr; + if (const Class* objClass = objTypes ? objTypes->getKnownClass(constraints()) : nullptr) { + MOZ_ASSERT(IsTypedObjectClass(objClass)); + expectedClass = GetOutlineTypedObjectClass(IsOpaqueTypedObjectClass(objClass)); + } + const TypedProto* expectedProto = derivedPrediction.getKnownPrototype(); + MOZ_ASSERT_IF(expectedClass, IsTypedObjectClass(expectedClass)); + + // Determine (if possible) the class/proto that the observed type set + // describes. + TemporaryTypeSet* observedTypes = bytecodeTypes(pc); + const Class* observedClass = observedTypes->getKnownClass(constraints()); + + // If expectedClass/expectedProto are both non-null (and hence known), we + // can predict precisely what object group derivedTypedObj will have. + // Therefore, if we observe that this group is already contained in the set + // of observedTypes, we can skip the barrier. + // + // Barriers still wind up being needed in some relatively + // rare cases: + // + // - if multiple kinds of typed objects flow into this point, + // in which case we will not be able to predict expectedClass + // nor expectedProto. + // + // - if the code has never executed, in which case the set of + // observed types will be incomplete. + // + // Barriers are particularly expensive here because they prevent + // us from optimizing the MNewDerivedTypedObject away. + JSObject* observedProto; + if (observedTypes->getCommonPrototype(constraints(), &observedProto) && + observedClass && observedProto && observedClass == expectedClass && + observedProto == expectedProto) + { + derivedTypedObj->setResultTypeSet(observedTypes); + } else { + if (!pushTypeBarrier(derivedTypedObj, observedTypes, BarrierKind::TypeSet)) + return false; + } + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +bool +IonBuilder::getElemTryGetProp(bool* emitted, MDefinition* obj, MDefinition* index) +{ + // If index is a constant string or symbol, try to optimize this GETELEM + // as a GETPROP. + + MOZ_ASSERT(*emitted == false); + + MConstant* indexConst = index->maybeConstantValue(); + jsid id; + if (!indexConst || !ValueToIdPure(indexConst->toJSValue(), &id)) + return true; + + if (id != IdToTypeId(id)) + return true; + + TemporaryTypeSet* types = bytecodeTypes(pc); + + trackOptimizationAttempt(TrackedStrategy::GetProp_Constant); + if (!getPropTryConstant(emitted, obj, id, types) || *emitted) { + if (*emitted) + index->setImplicitlyUsedUnchecked(); + return *emitted; + } + + trackOptimizationAttempt(TrackedStrategy::GetProp_NotDefined); + if (!getPropTryNotDefined(emitted, obj, id, types) || *emitted) { + if (*emitted) + index->setImplicitlyUsedUnchecked(); + return *emitted; + } + + return true; +} + +bool +IonBuilder::getElemTryDense(bool* emitted, MDefinition* obj, MDefinition* index) +{ + MOZ_ASSERT(*emitted == false); + + JSValueType unboxedType = UnboxedArrayElementType(constraints(), obj, index); + if (unboxedType == JSVAL_TYPE_MAGIC) { + if (!ElementAccessIsDenseNative(constraints(), obj, index)) { + trackOptimizationOutcome(TrackedOutcome::AccessNotDense); + return true; + } + } + + // Don't generate a fast path if there have been bounds check failures + // and this access might be on a sparse property. + if (ElementAccessHasExtraIndexedProperty(this, obj) && failedBoundsCheck_) { + trackOptimizationOutcome(TrackedOutcome::ProtoIndexedProps); + return true; + } + + // Don't generate a fast path if this pc has seen negative indexes accessed, + // which will not appear to be extra indexed properties. + if (inspector->hasSeenNegativeIndexGetElement(pc)) { + trackOptimizationOutcome(TrackedOutcome::ArraySeenNegativeIndex); + return true; + } + + if (!jsop_getelem_dense(obj, index, unboxedType)) + return false; + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +JSObject* +IonBuilder::getStaticTypedArrayObject(MDefinition* obj, MDefinition* index) +{ + Scalar::Type arrayType; + if (!ElementAccessIsTypedArray(constraints(), obj, index, &arrayType)) { + trackOptimizationOutcome(TrackedOutcome::AccessNotTypedArray); + return nullptr; + } + + if (!LIRGenerator::allowStaticTypedArrayAccesses()) { + trackOptimizationOutcome(TrackedOutcome::Disabled); + return nullptr; + } + + if (ElementAccessHasExtraIndexedProperty(this, obj)) { + trackOptimizationOutcome(TrackedOutcome::ProtoIndexedProps); + return nullptr; + } + + if (!obj->resultTypeSet()) { + trackOptimizationOutcome(TrackedOutcome::NoTypeInfo); + return nullptr; + } + + JSObject* tarrObj = obj->resultTypeSet()->maybeSingleton(); + if (!tarrObj) { + trackOptimizationOutcome(TrackedOutcome::NotSingleton); + return nullptr; + } + + TypeSet::ObjectKey* tarrKey = TypeSet::ObjectKey::get(tarrObj); + if (tarrKey->unknownProperties()) { + trackOptimizationOutcome(TrackedOutcome::UnknownProperties); + return nullptr; + } + + return tarrObj; +} + +bool +IonBuilder::getElemTryTypedStatic(bool* emitted, MDefinition* obj, MDefinition* index) +{ + MOZ_ASSERT(*emitted == false); + + JSObject* tarrObj = getStaticTypedArrayObject(obj, index); + if (!tarrObj) + return true; + + // LoadTypedArrayElementStatic currently treats uint32 arrays as int32. + Scalar::Type viewType = tarrObj->as<TypedArrayObject>().type(); + if (viewType == Scalar::Uint32) { + trackOptimizationOutcome(TrackedOutcome::StaticTypedArrayUint32); + return true; + } + + MDefinition* ptr = convertShiftToMaskForStaticTypedArray(index, viewType); + if (!ptr) + return true; + + // Emit LoadTypedArrayElementStatic. + + if (tarrObj->is<TypedArrayObject>()) { + TypeSet::ObjectKey* tarrKey = TypeSet::ObjectKey::get(tarrObj); + tarrKey->watchStateChangeForTypedArrayData(constraints()); + } + + obj->setImplicitlyUsedUnchecked(); + index->setImplicitlyUsedUnchecked(); + + MLoadTypedArrayElementStatic* load = MLoadTypedArrayElementStatic::New(alloc(), tarrObj, ptr); + current->add(load); + current->push(load); + + // The load is infallible if an undefined result will be coerced to the + // appropriate numeric type if the read is out of bounds. The truncation + // analysis picks up some of these cases, but is incomplete with respect + // to others. For now, sniff the bytecode for simple patterns following + // the load which guarantee a truncation or numeric conversion. + if (viewType == Scalar::Float32 || viewType == Scalar::Float64) { + jsbytecode* next = pc + JSOP_GETELEM_LENGTH; + if (*next == JSOP_POS) + load->setInfallible(); + } else { + jsbytecode* next = pc + JSOP_GETELEM_LENGTH; + if (*next == JSOP_ZERO && *(next + JSOP_ZERO_LENGTH) == JSOP_BITOR) + load->setInfallible(); + } + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +bool +IonBuilder::getElemTryTypedArray(bool* emitted, MDefinition* obj, MDefinition* index) +{ + MOZ_ASSERT(*emitted == false); + + Scalar::Type arrayType; + if (!ElementAccessIsTypedArray(constraints(), obj, index, &arrayType)) { + trackOptimizationOutcome(TrackedOutcome::AccessNotTypedArray); + return true; + } + + // Emit typed getelem variant. + if (!jsop_getelem_typed(obj, index, arrayType)) + return false; + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +bool +IonBuilder::getElemTryString(bool* emitted, MDefinition* obj, MDefinition* index) +{ + MOZ_ASSERT(*emitted == false); + + if (obj->type() != MIRType::String || !IsNumberType(index->type())) { + trackOptimizationOutcome(TrackedOutcome::AccessNotString); + return true; + } + + // If the index is expected to be out-of-bounds, don't optimize to avoid + // frequent bailouts. + if (bytecodeTypes(pc)->hasType(TypeSet::UndefinedType())) { + trackOptimizationOutcome(TrackedOutcome::OutOfBounds); + return true; + } + + // Emit fast path for string[index]. + MInstruction* idInt32 = MToInt32::New(alloc(), index); + current->add(idInt32); + index = idInt32; + + MStringLength* length = MStringLength::New(alloc(), obj); + current->add(length); + + index = addBoundsCheck(index, length); + + MCharCodeAt* charCode = MCharCodeAt::New(alloc(), obj, index); + current->add(charCode); + + MFromCharCode* result = MFromCharCode::New(alloc(), charCode); + current->add(result); + current->push(result); + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +bool +IonBuilder::getElemTryArguments(bool* emitted, MDefinition* obj, MDefinition* index) +{ + MOZ_ASSERT(*emitted == false); + + if (inliningDepth_ > 0) + return true; + + if (obj->type() != MIRType::MagicOptimizedArguments) + return true; + + // Emit GetFrameArgument. + + MOZ_ASSERT(!info().argsObjAliasesFormals()); + + // Type Inference has guaranteed this is an optimized arguments object. + obj->setImplicitlyUsedUnchecked(); + + // To ensure that we are not looking above the number of actual arguments. + MArgumentsLength* length = MArgumentsLength::New(alloc()); + current->add(length); + + // Ensure index is an integer. + MInstruction* idInt32 = MToInt32::New(alloc(), index); + current->add(idInt32); + index = idInt32; + + // Bailouts if we read more than the number of actual arguments. + index = addBoundsCheck(index, length); + + // Load the argument from the actual arguments. + MGetFrameArgument* load = MGetFrameArgument::New(alloc(), index, analysis_.hasSetArg()); + current->add(load); + current->push(load); + + TemporaryTypeSet* types = bytecodeTypes(pc); + if (!pushTypeBarrier(load, types, BarrierKind::TypeSet)) + return false; + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +bool +IonBuilder::getElemTryArgumentsInlined(bool* emitted, MDefinition* obj, MDefinition* index) +{ + MOZ_ASSERT(*emitted == false); + + if (inliningDepth_ == 0) + return true; + + if (obj->type() != MIRType::MagicOptimizedArguments) + return true; + + // Emit inlined arguments. + obj->setImplicitlyUsedUnchecked(); + + MOZ_ASSERT(!info().argsObjAliasesFormals()); + + // When the id is constant, we can just return the corresponding inlined argument + MConstant* indexConst = index->maybeConstantValue(); + if (indexConst && indexConst->type() == MIRType::Int32) { + MOZ_ASSERT(inliningDepth_ > 0); + + int32_t id = indexConst->toInt32(); + index->setImplicitlyUsedUnchecked(); + + if (id < (int32_t)inlineCallInfo_->argc() && id >= 0) + current->push(inlineCallInfo_->getArg(id)); + else + pushConstant(UndefinedValue()); + + trackOptimizationSuccess(); + *emitted = true; + return true; + } + + // inlined not constant not supported, yet. + return abort("NYI inlined not constant get argument element"); +} + +bool +IonBuilder::getElemTryCache(bool* emitted, MDefinition* obj, MDefinition* index) +{ + MOZ_ASSERT(*emitted == false); + + // Make sure we have at least an object. + if (!obj->mightBeType(MIRType::Object)) { + trackOptimizationOutcome(TrackedOutcome::NotObject); + return true; + } + + // Don't cache for strings. + if (obj->mightBeType(MIRType::String)) { + trackOptimizationOutcome(TrackedOutcome::GetElemStringNotCached); + return true; + } + + // Index should be integer, string, or symbol + if (!index->mightBeType(MIRType::Int32) && + !index->mightBeType(MIRType::String) && + !index->mightBeType(MIRType::Symbol)) + { + trackOptimizationOutcome(TrackedOutcome::IndexType); + return true; + } + + // Turn off cacheing if the element is int32 and we've seen non-native objects as the target + // of this getelem. + bool nonNativeGetElement = inspector->hasSeenNonNativeGetElement(pc); + if (index->mightBeType(MIRType::Int32) && nonNativeGetElement) { + trackOptimizationOutcome(TrackedOutcome::NonNativeReceiver); + return true; + } + + // Emit GetElementCache. + + TemporaryTypeSet* types = bytecodeTypes(pc); + BarrierKind barrier = PropertyReadNeedsTypeBarrier(analysisContext, constraints(), obj, + nullptr, types); + + // Always add a barrier if the index might be a string or symbol, so that + // the cache can attach stubs for particular properties. + if (index->mightBeType(MIRType::String) || index->mightBeType(MIRType::Symbol)) + barrier = BarrierKind::TypeSet; + + MGetPropertyCache* ins = MGetPropertyCache::New(alloc(), obj, index, + barrier == BarrierKind::TypeSet); + current->add(ins); + current->push(ins); + + if (!resumeAfter(ins)) + return false; + + // Spice up type information. + if (index->type() == MIRType::Int32 && barrier == BarrierKind::NoBarrier) { + bool needHoleCheck = !ElementAccessIsPacked(constraints(), obj); + MIRType knownType = GetElemKnownType(needHoleCheck, types); + + if (knownType != MIRType::Value && knownType != MIRType::Double) + ins->setResultType(knownType); + } + + if (!pushTypeBarrier(ins, types, barrier)) + return false; + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +TemporaryTypeSet* +IonBuilder::computeHeapType(const TemporaryTypeSet* objTypes, const jsid id) +{ + if (objTypes->unknownObject() || objTypes->getObjectCount() == 0) + return nullptr; + + TemporaryTypeSet empty; + TemporaryTypeSet* acc = ∅ + LifoAlloc* lifoAlloc = alloc().lifoAlloc(); + + Vector<HeapTypeSetKey, 4, SystemAllocPolicy> properties; + if (!properties.reserve(objTypes->getObjectCount())) + return nullptr; + + for (unsigned i = 0; i < objTypes->getObjectCount(); i++) { + TypeSet::ObjectKey* key = objTypes->getObject(i); + + if (key->unknownProperties()) + return nullptr; + + HeapTypeSetKey property = key->property(id); + HeapTypeSet* currentSet = property.maybeTypes(); + + if (!currentSet || currentSet->unknown()) + return nullptr; + + properties.infallibleAppend(property); + acc = TypeSet::unionSets(acc, currentSet, lifoAlloc); + if (!acc) + return nullptr; + } + + // Freeze all the properties associated with the refined type set. + for (HeapTypeSetKey* i = properties.begin(); i != properties.end(); i++) + i->freeze(constraints()); + + return acc; +} + +bool +IonBuilder::jsop_getelem_dense(MDefinition* obj, MDefinition* index, JSValueType unboxedType) +{ + TemporaryTypeSet* types = bytecodeTypes(pc); + + MOZ_ASSERT(index->type() == MIRType::Int32 || index->type() == MIRType::Double); + if (JSOp(*pc) == JSOP_CALLELEM) { + // Indexed call on an element of an array. Populate the observed types + // with any objects that could be in the array, to avoid extraneous + // type barriers. + AddObjectsForPropertyRead(obj, nullptr, types); + } + + BarrierKind barrier = PropertyReadNeedsTypeBarrier(analysisContext, constraints(), obj, + nullptr, types); + bool needsHoleCheck = !ElementAccessIsPacked(constraints(), obj); + + // Reads which are on holes in the object do not have to bail out if + // undefined values have been observed at this access site and the access + // cannot hit another indexed property on the object or its prototypes. + bool readOutOfBounds = + types->hasType(TypeSet::UndefinedType()) && + !ElementAccessHasExtraIndexedProperty(this, obj); + + MIRType knownType = MIRType::Value; + if (unboxedType == JSVAL_TYPE_MAGIC && barrier == BarrierKind::NoBarrier) + knownType = GetElemKnownType(needsHoleCheck, types); + + // Ensure index is an integer. + MInstruction* idInt32 = MToInt32::New(alloc(), index); + current->add(idInt32); + index = idInt32; + + // Get the elements vector. + MInstruction* elements = MElements::New(alloc(), obj, unboxedType != JSVAL_TYPE_MAGIC); + current->add(elements); + + // Note: to help GVN, use the original MElements instruction and not + // MConvertElementsToDoubles as operand. This is fine because converting + // elements to double does not change the initialized length. + MInstruction* initLength = initializedLength(obj, elements, unboxedType); + + // If we can load the element as a definite double, make sure to check that + // the array has been converted to homogenous doubles first. + TemporaryTypeSet* objTypes = obj->resultTypeSet(); + bool inBounds = !readOutOfBounds && !needsHoleCheck; + + if (inBounds) { + TemporaryTypeSet* heapTypes = computeHeapType(objTypes, JSID_VOID); + if (heapTypes && heapTypes->isSubset(types)) { + knownType = heapTypes->getKnownMIRType(); + types = heapTypes; + } + } + + bool loadDouble = + unboxedType == JSVAL_TYPE_MAGIC && + barrier == BarrierKind::NoBarrier && + loopDepth_ && + inBounds && + knownType == MIRType::Double && + objTypes && + objTypes->convertDoubleElements(constraints()) == TemporaryTypeSet::AlwaysConvertToDoubles; + if (loadDouble) + elements = addConvertElementsToDoubles(elements); + + MInstruction* load; + + if (!readOutOfBounds) { + // This load should not return undefined, so likely we're reading + // in-bounds elements, and the array is packed or its holes are not + // read. This is the best case: we can separate the bounds check for + // hoisting. + index = addBoundsCheck(index, initLength); + + if (unboxedType != JSVAL_TYPE_MAGIC) { + load = loadUnboxedValue(elements, 0, index, unboxedType, barrier, types); + } else { + load = MLoadElement::New(alloc(), elements, index, needsHoleCheck, loadDouble); + current->add(load); + } + } else { + // This load may return undefined, so assume that we *can* read holes, + // or that we can read out-of-bounds accesses. In this case, the bounds + // check is part of the opcode. + load = MLoadElementHole::New(alloc(), elements, index, initLength, + unboxedType, needsHoleCheck); + current->add(load); + + // If maybeUndefined was true, the typeset must have undefined, and + // then either additional types or a barrier. This means we should + // never have a typed version of LoadElementHole. + MOZ_ASSERT(knownType == MIRType::Value); + } + + if (knownType != MIRType::Value) { + if (unboxedType == JSVAL_TYPE_MAGIC) + load->setResultType(knownType); + load->setResultTypeSet(types); + } + + current->push(load); + return pushTypeBarrier(load, types, barrier); +} + +MInstruction* +IonBuilder::addArrayBufferByteLength(MDefinition* obj) +{ + MLoadFixedSlot* ins = MLoadFixedSlot::New(alloc(), obj, size_t(ArrayBufferObject::BYTE_LENGTH_SLOT)); + current->add(ins); + ins->setResultType(MIRType::Int32); + return ins; +} + +void +IonBuilder::addTypedArrayLengthAndData(MDefinition* obj, + BoundsChecking checking, + MDefinition** index, + MInstruction** length, MInstruction** elements) +{ + MOZ_ASSERT((index != nullptr) == (elements != nullptr)); + + JSObject* tarr = nullptr; + + if (MConstant* objConst = obj->maybeConstantValue()) { + if (objConst->type() == MIRType::Object) + tarr = &objConst->toObject(); + } else if (TemporaryTypeSet* types = obj->resultTypeSet()) { + tarr = types->maybeSingleton(); + } + + if (tarr) { + SharedMem<void*> data = tarr->as<TypedArrayObject>().viewDataEither(); + // Bug 979449 - Optimistically embed the elements and use TI to + // invalidate if we move them. + bool isTenured = !tarr->runtimeFromMainThread()->gc.nursery.isInside(data); + if (isTenured && tarr->isSingleton()) { + // The 'data' pointer of TypedArrayObject can change in rare circumstances + // (ArrayBufferObject::changeContents). + TypeSet::ObjectKey* tarrKey = TypeSet::ObjectKey::get(tarr); + if (!tarrKey->unknownProperties()) { + if (tarr->is<TypedArrayObject>()) + tarrKey->watchStateChangeForTypedArrayData(constraints()); + + obj->setImplicitlyUsedUnchecked(); + + int32_t len = AssertedCast<int32_t>(tarr->as<TypedArrayObject>().length()); + *length = MConstant::New(alloc(), Int32Value(len)); + current->add(*length); + + if (index) { + if (checking == DoBoundsCheck) + *index = addBoundsCheck(*index, *length); + + *elements = MConstantElements::New(alloc(), data); + current->add(*elements); + } + return; + } + } + } + + *length = MTypedArrayLength::New(alloc(), obj); + current->add(*length); + + if (index) { + if (checking == DoBoundsCheck) + *index = addBoundsCheck(*index, *length); + + *elements = MTypedArrayElements::New(alloc(), obj); + current->add(*elements); + } +} + +MDefinition* +IonBuilder::convertShiftToMaskForStaticTypedArray(MDefinition* id, + Scalar::Type viewType) +{ + trackOptimizationOutcome(TrackedOutcome::StaticTypedArrayCantComputeMask); + + // No shifting is necessary if the typed array has single byte elements. + if (TypedArrayShift(viewType) == 0) + return id; + + // If the index is an already shifted constant, undo the shift to get the + // absolute offset being accessed. + if (MConstant* idConst = id->maybeConstantValue()) { + if (idConst->type() == MIRType::Int32) { + int32_t index = idConst->toInt32(); + MConstant* offset = MConstant::New(alloc(), Int32Value(index << TypedArrayShift(viewType))); + current->add(offset); + return offset; + } + } + + if (!id->isRsh() || id->isEffectful()) + return nullptr; + + MConstant* shiftAmount = id->toRsh()->rhs()->maybeConstantValue(); + if (!shiftAmount || shiftAmount->type() != MIRType::Int32) + return nullptr; + if (uint32_t(shiftAmount->toInt32()) != TypedArrayShift(viewType)) + return nullptr; + + // Instead of shifting, mask off the low bits of the index so that + // a non-scaled access on the typed array can be performed. + MConstant* mask = MConstant::New(alloc(), Int32Value(~((1 << shiftAmount->toInt32()) - 1))); + MBitAnd* ptr = MBitAnd::New(alloc(), id->getOperand(0), mask); + + ptr->infer(nullptr, nullptr); + MOZ_ASSERT(!ptr->isEffectful()); + + current->add(mask); + current->add(ptr); + + return ptr; +} + +bool +IonBuilder::jsop_getelem_typed(MDefinition* obj, MDefinition* index, + Scalar::Type arrayType) +{ + TemporaryTypeSet* types = bytecodeTypes(pc); + + bool maybeUndefined = types->hasType(TypeSet::UndefinedType()); + + // Reading from an Uint32Array will result in a double for values + // that don't fit in an int32. We have to bailout if this happens + // and the instruction is not known to return a double. + bool allowDouble = types->hasType(TypeSet::DoubleType()); + + // Ensure id is an integer. + MInstruction* idInt32 = MToInt32::New(alloc(), index); + current->add(idInt32); + index = idInt32; + + if (!maybeUndefined) { + // Assume the index is in range, so that we can hoist the length, + // elements vector and bounds check. + + // If we are reading in-bounds elements, we can use knowledge about + // the array type to determine the result type, even if the opcode has + // never executed. The known pushed type is only used to distinguish + // uint32 reads that may produce either doubles or integers. + MIRType knownType = MIRTypeForTypedArrayRead(arrayType, allowDouble); + + // Get length, bounds-check, then get elements, and add all instructions. + MInstruction* length; + MInstruction* elements; + addTypedArrayLengthAndData(obj, DoBoundsCheck, &index, &length, &elements); + + // Load the element. + MLoadUnboxedScalar* load = MLoadUnboxedScalar::New(alloc(), elements, index, arrayType); + current->add(load); + current->push(load); + + // Note: we can ignore the type barrier here, we know the type must + // be valid and unbarriered. + load->setResultType(knownType); + return true; + } else { + // We need a type barrier if the array's element type has never been + // observed (we've only read out-of-bounds values). Note that for + // Uint32Array, we only check for int32: if allowDouble is false we + // will bailout when we read a double. + BarrierKind barrier = BarrierKind::TypeSet; + switch (arrayType) { + case Scalar::Int8: + case Scalar::Uint8: + case Scalar::Uint8Clamped: + case Scalar::Int16: + case Scalar::Uint16: + case Scalar::Int32: + case Scalar::Uint32: + if (types->hasType(TypeSet::Int32Type())) + barrier = BarrierKind::NoBarrier; + break; + case Scalar::Float32: + case Scalar::Float64: + if (allowDouble) + barrier = BarrierKind::NoBarrier; + break; + default: + MOZ_CRASH("Unknown typed array type"); + } + + // Assume we will read out-of-bound values. In this case the + // bounds check will be part of the instruction, and the instruction + // will always return a Value. + MLoadTypedArrayElementHole* load = + MLoadTypedArrayElementHole::New(alloc(), obj, index, arrayType, allowDouble); + current->add(load); + current->push(load); + + return pushTypeBarrier(load, types, barrier); + } +} + +bool +IonBuilder::jsop_setelem() +{ + bool emitted = false; + startTrackingOptimizations(); + + MDefinition* value = current->pop(); + MDefinition* index = current->pop(); + MDefinition* object = convertUnboxedObjects(current->pop()); + + trackTypeInfo(TrackedTypeSite::Receiver, object->type(), object->resultTypeSet()); + trackTypeInfo(TrackedTypeSite::Index, index->type(), index->resultTypeSet()); + trackTypeInfo(TrackedTypeSite::Value, value->type(), value->resultTypeSet()); + + if (shouldAbortOnPreliminaryGroups(object)) { + MInstruction* ins = MCallSetElement::New(alloc(), object, index, value, IsStrictSetPC(pc)); + current->add(ins); + current->push(value); + return resumeAfter(ins); + } + + if (!forceInlineCaches()) { + trackOptimizationAttempt(TrackedStrategy::SetElem_TypedObject); + if (!setElemTryTypedObject(&emitted, object, index, value) || emitted) + return emitted; + + trackOptimizationAttempt(TrackedStrategy::SetElem_TypedStatic); + if (!setElemTryTypedStatic(&emitted, object, index, value) || emitted) + return emitted; + + trackOptimizationAttempt(TrackedStrategy::SetElem_TypedArray); + if (!setElemTryTypedArray(&emitted, object, index, value) || emitted) + return emitted; + + trackOptimizationAttempt(TrackedStrategy::SetElem_Dense); + SetElemICInspector icInspect(inspector->setElemICInspector(pc)); + bool writeHole = icInspect.sawOOBDenseWrite(); + if (!setElemTryDense(&emitted, object, index, value, writeHole) || emitted) + return emitted; + + trackOptimizationAttempt(TrackedStrategy::SetElem_Arguments); + if (!setElemTryArguments(&emitted, object, index, value) || emitted) + return emitted; + } + + if (script()->argumentsHasVarBinding() && + object->mightBeType(MIRType::MagicOptimizedArguments) && + info().analysisMode() != Analysis_ArgumentsUsage) + { + return abort("Type is not definitely lazy arguments."); + } + + trackOptimizationAttempt(TrackedStrategy::SetElem_InlineCache); + if (!setElemTryCache(&emitted, object, index, value) || emitted) + return emitted; + + // Emit call. + MInstruction* ins = MCallSetElement::New(alloc(), object, index, value, IsStrictSetPC(pc)); + current->add(ins); + current->push(value); + + return resumeAfter(ins); +} + +bool +IonBuilder::setElemTryTypedObject(bool* emitted, MDefinition* obj, + MDefinition* index, MDefinition* value) +{ + MOZ_ASSERT(*emitted == false); + + // The next several failures are all due to types not predicting that we + // are definitely doing a getelem access on a typed object. + trackOptimizationOutcome(TrackedOutcome::AccessNotTypedObject); + + TypedObjectPrediction objPrediction = typedObjectPrediction(obj); + if (objPrediction.isUseless()) + return true; + + if (!objPrediction.ofArrayKind()) + return true; + + TypedObjectPrediction elemPrediction = objPrediction.arrayElementType(); + if (elemPrediction.isUseless()) + return true; + + uint32_t elemSize; + if (!elemPrediction.hasKnownSize(&elemSize)) + return true; + + switch (elemPrediction.kind()) { + case type::Simd: + // FIXME (bug 894105): store a MIRType::float32x4 etc + trackOptimizationOutcome(TrackedOutcome::GenericFailure); + return true; + + case type::Reference: + return setElemTryReferenceElemOfTypedObject(emitted, obj, index, + objPrediction, value, elemPrediction); + + case type::Scalar: + return setElemTryScalarElemOfTypedObject(emitted, + obj, + index, + objPrediction, + value, + elemPrediction, + elemSize); + + case type::Struct: + case type::Array: + // Not yet optimized. + trackOptimizationOutcome(TrackedOutcome::GenericFailure); + return true; + } + + MOZ_CRASH("Bad kind"); +} + +bool +IonBuilder::setElemTryReferenceElemOfTypedObject(bool* emitted, + MDefinition* obj, + MDefinition* index, + TypedObjectPrediction objPrediction, + MDefinition* value, + TypedObjectPrediction elemPrediction) +{ + ReferenceTypeDescr::Type elemType = elemPrediction.referenceType(); + uint32_t elemSize = ReferenceTypeDescr::size(elemType); + + LinearSum indexAsByteOffset(alloc()); + if (!checkTypedObjectIndexInBounds(elemSize, obj, index, objPrediction, &indexAsByteOffset)) + return true; + + if (!storeReferenceTypedObjectValue(obj, indexAsByteOffset, elemType, value, nullptr)) + return true; + + current->push(value); + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +bool +IonBuilder::setElemTryScalarElemOfTypedObject(bool* emitted, + MDefinition* obj, + MDefinition* index, + TypedObjectPrediction objPrediction, + MDefinition* value, + TypedObjectPrediction elemPrediction, + uint32_t elemSize) +{ + // Must always be loading the same scalar type + ScalarTypeDescr::Type elemType = elemPrediction.scalarType(); + MOZ_ASSERT(elemSize == ScalarTypeDescr::alignment(elemType)); + + LinearSum indexAsByteOffset(alloc()); + if (!checkTypedObjectIndexInBounds(elemSize, obj, index, objPrediction, &indexAsByteOffset)) + return true; + + // Store the element + if (!storeScalarTypedObjectValue(obj, indexAsByteOffset, elemType, value)) + return false; + + current->push(value); + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +bool +IonBuilder::setElemTryTypedStatic(bool* emitted, MDefinition* object, + MDefinition* index, MDefinition* value) +{ + MOZ_ASSERT(*emitted == false); + + JSObject* tarrObj = getStaticTypedArrayObject(object, index); + if (!tarrObj) + return true; + + SharedMem<void*> viewData = tarrObj->as<TypedArrayObject>().viewDataEither(); + if (tarrObj->runtimeFromMainThread()->gc.nursery.isInside(viewData)) + return true; + + Scalar::Type viewType = tarrObj->as<TypedArrayObject>().type(); + MDefinition* ptr = convertShiftToMaskForStaticTypedArray(index, viewType); + if (!ptr) + return true; + + // Emit StoreTypedArrayElementStatic. + + if (tarrObj->is<TypedArrayObject>()) { + TypeSet::ObjectKey* tarrKey = TypeSet::ObjectKey::get(tarrObj); + tarrKey->watchStateChangeForTypedArrayData(constraints()); + } + + object->setImplicitlyUsedUnchecked(); + index->setImplicitlyUsedUnchecked(); + + // Clamp value to [0, 255] for Uint8ClampedArray. + MDefinition* toWrite = value; + if (viewType == Scalar::Uint8Clamped) { + toWrite = MClampToUint8::New(alloc(), value); + current->add(toWrite->toInstruction()); + } + + MInstruction* store = MStoreTypedArrayElementStatic::New(alloc(), tarrObj, ptr, toWrite); + current->add(store); + current->push(value); + + if (!resumeAfter(store)) + return false; + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +bool +IonBuilder::setElemTryTypedArray(bool* emitted, MDefinition* object, + MDefinition* index, MDefinition* value) +{ + MOZ_ASSERT(*emitted == false); + + Scalar::Type arrayType; + if (!ElementAccessIsTypedArray(constraints(), object, index, &arrayType)) { + trackOptimizationOutcome(TrackedOutcome::AccessNotTypedArray); + return true; + } + + // Emit typed setelem variant. + if (!jsop_setelem_typed(arrayType, object, index, value)) + return false; + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +bool +IonBuilder::setElemTryDense(bool* emitted, MDefinition* object, + MDefinition* index, MDefinition* value, bool writeHole) +{ + MOZ_ASSERT(*emitted == false); + + JSValueType unboxedType = UnboxedArrayElementType(constraints(), object, index); + if (unboxedType == JSVAL_TYPE_MAGIC) { + if (!ElementAccessIsDenseNative(constraints(), object, index)) { + trackOptimizationOutcome(TrackedOutcome::AccessNotDense); + return true; + } + } + + if (PropertyWriteNeedsTypeBarrier(alloc(), constraints(), current, + &object, nullptr, &value, /* canModify = */ true)) + { + trackOptimizationOutcome(TrackedOutcome::NeedsTypeBarrier); + return true; + } + + if (!object->resultTypeSet()) { + trackOptimizationOutcome(TrackedOutcome::NoTypeInfo); + return true; + } + + TemporaryTypeSet::DoubleConversion conversion = + object->resultTypeSet()->convertDoubleElements(constraints()); + + // If AmbiguousDoubleConversion, only handle int32 values for now. + if (conversion == TemporaryTypeSet::AmbiguousDoubleConversion && + value->type() != MIRType::Int32) + { + trackOptimizationOutcome(TrackedOutcome::ArrayDoubleConversion); + return true; + } + + // Don't generate a fast path if there have been bounds check failures + // and this access might be on a sparse property. + if (ElementAccessHasExtraIndexedProperty(this, object) && failedBoundsCheck_) { + trackOptimizationOutcome(TrackedOutcome::ProtoIndexedProps); + return true; + } + + // Emit dense setelem variant. + if (!jsop_setelem_dense(conversion, object, index, value, unboxedType, writeHole, emitted)) + return false; + + if (!*emitted) { + trackOptimizationOutcome(TrackedOutcome::NonWritableProperty); + return true; + } + + trackOptimizationSuccess(); + return true; +} + +bool +IonBuilder::setElemTryArguments(bool* emitted, MDefinition* object, + MDefinition* index, MDefinition* value) +{ + MOZ_ASSERT(*emitted == false); + + if (object->type() != MIRType::MagicOptimizedArguments) + return true; + + // Arguments are not supported yet. + return abort("NYI arguments[]="); +} + +bool +IonBuilder::setElemTryCache(bool* emitted, MDefinition* object, + MDefinition* index, MDefinition* value) +{ + MOZ_ASSERT(*emitted == false); + + if (!object->mightBeType(MIRType::Object)) { + trackOptimizationOutcome(TrackedOutcome::NotObject); + return true; + } + + if (!index->mightBeType(MIRType::Int32) && + !index->mightBeType(MIRType::String) && + !index->mightBeType(MIRType::Symbol)) + { + trackOptimizationOutcome(TrackedOutcome::IndexType); + return true; + } + + bool barrier = true; + bool indexIsInt32 = index->type() == MIRType::Int32; + + if (indexIsInt32 && + !PropertyWriteNeedsTypeBarrier(alloc(), constraints(), current, + &object, nullptr, &value, /* canModify = */ true)) + { + barrier = false; + } + + // We can avoid worrying about holes in the IC if we know a priori we are safe + // from them. If TI can guard that there are no indexed properties on the prototype + // chain, we know that we anen't missing any setters by overwriting the hole with + // another value. + bool guardHoles = ElementAccessHasExtraIndexedProperty(this, object); + + // Make sure the object being written to doesn't have copy on write elements. + const Class* clasp = object->resultTypeSet() ? object->resultTypeSet()->getKnownClass(constraints()) : nullptr; + bool checkNative = !clasp || !clasp->isNative(); + object = addMaybeCopyElementsForWrite(object, checkNative); + + if (NeedsPostBarrier(value)) { + if (indexIsInt32) + current->add(MPostWriteElementBarrier::New(alloc(), object, value, index)); + else + current->add(MPostWriteBarrier::New(alloc(), object, value)); + } + + // Emit SetPropertyCache. + bool strict = JSOp(*pc) == JSOP_STRICTSETELEM; + MSetPropertyCache* ins = + MSetPropertyCache::New(alloc(), object, index, value, strict, barrier, guardHoles); + current->add(ins); + current->push(value); + + if (!resumeAfter(ins)) + return false; + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +bool +IonBuilder::jsop_setelem_dense(TemporaryTypeSet::DoubleConversion conversion, + MDefinition* obj, MDefinition* id, MDefinition* value, + JSValueType unboxedType, bool writeHole, bool* emitted) +{ + MOZ_ASSERT(*emitted == false); + + MIRType elementType = MIRType::None; + if (unboxedType == JSVAL_TYPE_MAGIC) + elementType = DenseNativeElementType(constraints(), obj); + bool packed = ElementAccessIsPacked(constraints(), obj); + + // Writes which are on holes in the object do not have to bail out if they + // cannot hit another indexed property on the object or its prototypes. + bool hasNoExtraIndexedProperty = !ElementAccessHasExtraIndexedProperty(this, obj); + + bool mayBeFrozen = ElementAccessMightBeFrozen(constraints(), obj); + + if (mayBeFrozen && !hasNoExtraIndexedProperty) { + // FallibleStoreElement does not know how to deal with extra indexed + // properties on the prototype. This case should be rare so we fall back + // to an IC. + return true; + } + + *emitted = true; + + // Ensure id is an integer. + MInstruction* idInt32 = MToInt32::New(alloc(), id); + current->add(idInt32); + id = idInt32; + + if (NeedsPostBarrier(value)) + current->add(MPostWriteElementBarrier::New(alloc(), obj, value, id)); + + // Copy the elements vector if necessary. + obj = addMaybeCopyElementsForWrite(obj, /* checkNative = */ false); + + // Get the elements vector. + MElements* elements = MElements::New(alloc(), obj, unboxedType != JSVAL_TYPE_MAGIC); + current->add(elements); + + // Ensure the value is a double, if double conversion might be needed. + MDefinition* newValue = value; + switch (conversion) { + case TemporaryTypeSet::AlwaysConvertToDoubles: + case TemporaryTypeSet::MaybeConvertToDoubles: { + MInstruction* valueDouble = MToDouble::New(alloc(), value); + current->add(valueDouble); + newValue = valueDouble; + break; + } + + case TemporaryTypeSet::AmbiguousDoubleConversion: { + MOZ_ASSERT(value->type() == MIRType::Int32); + MInstruction* maybeDouble = MMaybeToDoubleElement::New(alloc(), elements, value); + current->add(maybeDouble); + newValue = maybeDouble; + break; + } + + case TemporaryTypeSet::DontConvertToDoubles: + break; + + default: + MOZ_CRASH("Unknown double conversion"); + } + + // Use MStoreElementHole if this SETELEM has written to out-of-bounds + // indexes in the past. Otherwise, use MStoreElement so that we can hoist + // the initialized length and bounds check. + // If an object may have been frozen, no previous expectation hold and we + // fallback to MFallibleStoreElement. + MInstruction* store; + MStoreElementCommon* common = nullptr; + if (writeHole && hasNoExtraIndexedProperty && !mayBeFrozen) { + MStoreElementHole* ins = MStoreElementHole::New(alloc(), obj, elements, id, newValue, unboxedType); + store = ins; + common = ins; + + current->add(ins); + current->push(value); + } else if (mayBeFrozen) { + MOZ_ASSERT(hasNoExtraIndexedProperty, + "FallibleStoreElement codegen assumes no extra indexed properties"); + + bool strict = IsStrictSetPC(pc); + MFallibleStoreElement* ins = MFallibleStoreElement::New(alloc(), obj, elements, id, + newValue, unboxedType, strict); + store = ins; + common = ins; + + current->add(ins); + current->push(value); + } else { + MInstruction* initLength = initializedLength(obj, elements, unboxedType); + + id = addBoundsCheck(id, initLength); + bool needsHoleCheck = !packed && !hasNoExtraIndexedProperty; + + if (unboxedType != JSVAL_TYPE_MAGIC) { + store = storeUnboxedValue(obj, elements, 0, id, unboxedType, newValue); + } else { + MStoreElement* ins = MStoreElement::New(alloc(), elements, id, newValue, needsHoleCheck); + store = ins; + common = ins; + + current->add(store); + } + + current->push(value); + } + + if (!resumeAfter(store)) + return false; + + if (common) { + // Determine whether a write barrier is required. + if (obj->resultTypeSet()->propertyNeedsBarrier(constraints(), JSID_VOID)) + common->setNeedsBarrier(); + + if (elementType != MIRType::None && packed) + common->setElementType(elementType); + } + + return true; +} + + +bool +IonBuilder::jsop_setelem_typed(Scalar::Type arrayType, + MDefinition* obj, MDefinition* id, MDefinition* value) +{ + SetElemICInspector icInspect(inspector->setElemICInspector(pc)); + bool expectOOB = icInspect.sawOOBTypedArrayWrite(); + + if (expectOOB) + spew("Emitting OOB TypedArray SetElem"); + + // Ensure id is an integer. + MInstruction* idInt32 = MToInt32::New(alloc(), id); + current->add(idInt32); + id = idInt32; + + // Get length, bounds-check, then get elements, and add all instructions. + MInstruction* length; + MInstruction* elements; + BoundsChecking checking = expectOOB ? SkipBoundsCheck : DoBoundsCheck; + addTypedArrayLengthAndData(obj, checking, &id, &length, &elements); + + // Clamp value to [0, 255] for Uint8ClampedArray. + MDefinition* toWrite = value; + if (arrayType == Scalar::Uint8Clamped) { + toWrite = MClampToUint8::New(alloc(), value); + current->add(toWrite->toInstruction()); + } + + // Store the value. + MInstruction* ins; + if (expectOOB) { + ins = MStoreTypedArrayElementHole::New(alloc(), elements, length, id, toWrite, arrayType); + } else { + MStoreUnboxedScalar* store = + MStoreUnboxedScalar::New(alloc(), elements, id, toWrite, arrayType, + MStoreUnboxedScalar::TruncateInput); + ins = store; + } + + current->add(ins); + current->push(value); + + return resumeAfter(ins); +} + +bool +IonBuilder::jsop_length() +{ + if (jsop_length_fastPath()) + return true; + + PropertyName* name = info().getAtom(pc)->asPropertyName(); + return jsop_getprop(name); +} + +bool +IonBuilder::jsop_length_fastPath() +{ + TemporaryTypeSet* types = bytecodeTypes(pc); + + if (types->getKnownMIRType() != MIRType::Int32) + return false; + + MDefinition* obj = current->peek(-1); + + if (shouldAbortOnPreliminaryGroups(obj)) + return false; + + if (obj->mightBeType(MIRType::String)) { + if (obj->mightBeType(MIRType::Object)) + return false; + current->pop(); + MStringLength* ins = MStringLength::New(alloc(), obj); + current->add(ins); + current->push(ins); + return true; + } + + if (obj->mightBeType(MIRType::Object)) { + TemporaryTypeSet* objTypes = obj->resultTypeSet(); + + // Compute the length for array objects. + if (objTypes && + objTypes->getKnownClass(constraints()) == &ArrayObject::class_ && + !objTypes->hasObjectFlags(constraints(), OBJECT_FLAG_LENGTH_OVERFLOW)) + { + current->pop(); + MElements* elements = MElements::New(alloc(), obj); + current->add(elements); + + // Read length. + MArrayLength* length = MArrayLength::New(alloc(), elements); + current->add(length); + current->push(length); + return true; + } + + // Compute the length for unboxed array objects. + if (UnboxedArrayElementType(constraints(), obj, nullptr) != JSVAL_TYPE_MAGIC && + !objTypes->hasObjectFlags(constraints(), OBJECT_FLAG_LENGTH_OVERFLOW)) + { + current->pop(); + + MUnboxedArrayLength* length = MUnboxedArrayLength::New(alloc(), obj); + current->add(length); + current->push(length); + return true; + } + + // Compute the length for array typed objects. + TypedObjectPrediction prediction = typedObjectPrediction(obj); + if (!prediction.isUseless()) { + TypeSet::ObjectKey* globalKey = TypeSet::ObjectKey::get(&script()->global()); + if (globalKey->hasFlags(constraints(), OBJECT_FLAG_TYPED_OBJECT_HAS_DETACHED_BUFFER)) + return false; + + MInstruction* length; + int32_t sizedLength; + if (prediction.hasKnownArrayLength(&sizedLength)) { + obj->setImplicitlyUsedUnchecked(); + length = MConstant::New(alloc(), Int32Value(sizedLength)); + } else { + return false; + } + + current->pop(); + current->add(length); + current->push(length); + return true; + } + } + + return false; +} + +bool +IonBuilder::jsop_arguments() +{ + if (info().needsArgsObj()) { + current->push(current->argumentsObject()); + return true; + } + MOZ_ASSERT(lazyArguments_); + current->push(lazyArguments_); + return true; +} + +bool +IonBuilder::jsop_newtarget() +{ + if (!info().funMaybeLazy()) { + MOZ_ASSERT(!info().script()->isForEval()); + pushConstant(NullValue()); + return true; + } + + MOZ_ASSERT(info().funMaybeLazy()); + + if (info().funMaybeLazy()->isArrow()) { + MArrowNewTarget* arrowNewTarget = MArrowNewTarget::New(alloc(), getCallee()); + current->add(arrowNewTarget); + current->push(arrowNewTarget); + return true; + } + + if (inliningDepth_ == 0) { + MNewTarget* newTarget = MNewTarget::New(alloc()); + current->add(newTarget); + current->push(newTarget); + return true; + } + + if (!inlineCallInfo_->constructing()) { + pushConstant(UndefinedValue()); + return true; + } + + current->push(inlineCallInfo_->getNewTarget()); + return true; +} + +bool +IonBuilder::jsop_rest() +{ + if (info().analysisMode() == Analysis_ArgumentsUsage) { + // There's no BaselineScript with the template object. Just push a + // dummy value, it does not affect the arguments analysis. + MUnknownValue* unknown = MUnknownValue::New(alloc()); + current->add(unknown); + current->push(unknown); + return true; + } + + ArrayObject* templateObject = &inspector->getTemplateObject(pc)->as<ArrayObject>(); + + if (inliningDepth_ == 0) { + // We don't know anything about the callee. + MArgumentsLength* numActuals = MArgumentsLength::New(alloc()); + current->add(numActuals); + + // Pass in the number of actual arguments, the number of formals (not + // including the rest parameter slot itself), and the template object. + MRest* rest = MRest::New(alloc(), constraints(), numActuals, info().nargs() - 1, + templateObject); + current->add(rest); + current->push(rest); + return true; + } + + // We know the exact number of arguments the callee pushed. + unsigned numActuals = inlineCallInfo_->argc(); + unsigned numFormals = info().nargs() - 1; + unsigned numRest = numActuals > numFormals ? numActuals - numFormals : 0; + + if (!jsop_newarray(numRest)) + return false; + + if (numRest == 0) { + // No more updating to do. (Note that in this one case the length from + // the template object is already correct.) + return true; + } + + MDefinition *array = current->peek(-1); + MElements* elements = MElements::New(alloc(), array); + current->add(elements); + + // Unroll the argument copy loop. We don't need to do any bounds or hole + // checking here. + MConstant* index = nullptr; + for (unsigned i = numFormals; i < numActuals; i++) { + index = MConstant::New(alloc(), Int32Value(i - numFormals)); + current->add(index); + + MDefinition* arg = inlineCallInfo_->argv()[i]; + MStoreElement* store = MStoreElement::New(alloc(), elements, index, arg, + /* needsHoleCheck = */ false); + current->add(store); + + if (NeedsPostBarrier(arg)) + current->add(MPostWriteBarrier::New(alloc(), array, arg)); + } + + // The array's length is incorrectly 0 now, from the template object + // created by BaselineCompiler::emit_JSOP_REST() before the actual argument + // count was known. Set the correct length now that we know that count. + MSetArrayLength* length = MSetArrayLength::New(alloc(), elements, index); + current->add(length); + + // Update the initialized length for all the (necessarily non-hole) + // elements added. + MSetInitializedLength* initLength = MSetInitializedLength::New(alloc(), elements, index); + current->add(initLength); + + return true; +} + +bool +IonBuilder::jsop_checkisobj(uint8_t kind) +{ + MDefinition* toCheck = current->peek(-1); + + if (toCheck->type() == MIRType::Object) { + toCheck->setImplicitlyUsedUnchecked(); + return true; + } + + MCheckIsObj* check = MCheckIsObj::New(alloc(), current->pop(), kind); + current->add(check); + current->push(check); + return true; +} + +bool +IonBuilder::jsop_checkobjcoercible() +{ + MDefinition* toCheck = current->peek(-1); + + if (!toCheck->mightBeType(MIRType::Undefined) && + !toCheck->mightBeType(MIRType::Null)) + { + toCheck->setImplicitlyUsedUnchecked(); + return true; + } + + MOZ_ASSERT(toCheck->type() == MIRType::Value || + toCheck->type() == MIRType::Null || + toCheck->type() == MIRType::Undefined); + + // If we want to squeeze more perf here, we can throw without checking, + // if IsNullOrUndefined(toCheck->type()). Since this is a failure case, + // it should be OK. + MCheckObjCoercible* check = MCheckObjCoercible::New(alloc(), current->pop()); + current->add(check); + current->push(check); + return resumeAfter(check); +} + +uint32_t +IonBuilder::getDefiniteSlot(TemporaryTypeSet* types, PropertyName* name, uint32_t* pnfixed) +{ + if (!types || types->unknownObject() || !types->objectOrSentinel()) { + trackOptimizationOutcome(TrackedOutcome::NoTypeInfo); + return UINT32_MAX; + } + + uint32_t slot = UINT32_MAX; + + for (size_t i = 0; i < types->getObjectCount(); i++) { + TypeSet::ObjectKey* key = types->getObject(i); + if (!key) + continue; + + if (key->unknownProperties()) { + trackOptimizationOutcome(TrackedOutcome::UnknownProperties); + return UINT32_MAX; + } + + if (key->isSingleton()) { + trackOptimizationOutcome(TrackedOutcome::Singleton); + return UINT32_MAX; + } + + HeapTypeSetKey property = key->property(NameToId(name)); + if (!property.maybeTypes() || + !property.maybeTypes()->definiteProperty() || + property.nonData(constraints())) + { + trackOptimizationOutcome(TrackedOutcome::NotFixedSlot); + return UINT32_MAX; + } + + // Definite slots will always be fixed slots when they are in the + // allowable range for fixed slots, except for objects which were + // converted from unboxed objects and have a smaller allocation size. + size_t nfixed = NativeObject::MAX_FIXED_SLOTS; + if (ObjectGroup* group = key->group()->maybeOriginalUnboxedGroup()) + nfixed = gc::GetGCKindSlots(group->unboxedLayout().getAllocKind()); + + uint32_t propertySlot = property.maybeTypes()->definiteSlot(); + if (slot == UINT32_MAX) { + slot = propertySlot; + *pnfixed = nfixed; + } else if (slot != propertySlot || nfixed != *pnfixed) { + trackOptimizationOutcome(TrackedOutcome::InconsistentFixedSlot); + return UINT32_MAX; + } + } + + return slot; +} + +uint32_t +IonBuilder::getUnboxedOffset(TemporaryTypeSet* types, PropertyName* name, JSValueType* punboxedType) +{ + if (!types || types->unknownObject() || !types->objectOrSentinel()) { + trackOptimizationOutcome(TrackedOutcome::NoTypeInfo); + return UINT32_MAX; + } + + uint32_t offset = UINT32_MAX; + + for (size_t i = 0; i < types->getObjectCount(); i++) { + TypeSet::ObjectKey* key = types->getObject(i); + if (!key) + continue; + + if (key->unknownProperties()) { + trackOptimizationOutcome(TrackedOutcome::UnknownProperties); + return UINT32_MAX; + } + + if (key->isSingleton()) { + trackOptimizationOutcome(TrackedOutcome::Singleton); + return UINT32_MAX; + } + + UnboxedLayout* layout = key->group()->maybeUnboxedLayout(); + if (!layout) { + trackOptimizationOutcome(TrackedOutcome::NotUnboxed); + return UINT32_MAX; + } + + const UnboxedLayout::Property* property = layout->lookup(name); + if (!property) { + trackOptimizationOutcome(TrackedOutcome::StructNoField); + return UINT32_MAX; + } + + if (layout->nativeGroup()) { + trackOptimizationOutcome(TrackedOutcome::UnboxedConvertedToNative); + return UINT32_MAX; + } + + key->watchStateChangeForUnboxedConvertedToNative(constraints()); + + if (offset == UINT32_MAX) { + offset = property->offset; + *punboxedType = property->type; + } else if (offset != property->offset) { + trackOptimizationOutcome(TrackedOutcome::InconsistentFieldOffset); + return UINT32_MAX; + } else if (*punboxedType != property->type) { + trackOptimizationOutcome(TrackedOutcome::InconsistentFieldType); + return UINT32_MAX; + } + } + + return offset; +} + +bool +IonBuilder::jsop_runonce() +{ + MRunOncePrologue* ins = MRunOncePrologue::New(alloc()); + current->add(ins); + return resumeAfter(ins); +} + +bool +IonBuilder::jsop_not() +{ + MDefinition* value = current->pop(); + + MNot* ins = MNot::New(alloc(), value, constraints()); + current->add(ins); + current->push(ins); + return true; +} + +bool +IonBuilder::objectsHaveCommonPrototype(TemporaryTypeSet* types, PropertyName* name, + bool isGetter, JSObject* foundProto, bool* guardGlobal) +{ + // With foundProto a prototype with a getter or setter for name, return + // whether looking up name on any object in |types| will go through + // foundProto, i.e. all the objects have foundProto on their prototype + // chain and do not have a property for name before reaching foundProto. + + // No sense looking if we don't know what's going on. + if (!types || types->unknownObject()) + return false; + *guardGlobal = false; + + for (unsigned i = 0; i < types->getObjectCount(); i++) { + if (types->getSingleton(i) == foundProto) + continue; + + TypeSet::ObjectKey* key = types->getObject(i); + if (!key) + continue; + + while (key) { + if (key->unknownProperties()) + return false; + + const Class* clasp = key->clasp(); + if (!ClassHasEffectlessLookup(clasp)) + return false; + JSObject* singleton = key->isSingleton() ? key->singleton() : nullptr; + if (ObjectHasExtraOwnProperty(compartment, key, NameToId(name))) { + if (!singleton || !singleton->is<GlobalObject>()) + return false; + *guardGlobal = true; + } + + // Look for a getter/setter on the class itself which may need + // to be called. + if (isGetter && clasp->getOpsGetProperty()) + return false; + if (!isGetter && clasp->getOpsSetProperty()) + return false; + + // Test for isOwnProperty() without freezing. If we end up + // optimizing, freezePropertiesForCommonPropFunc will freeze the + // property type sets later on. + HeapTypeSetKey property = key->property(NameToId(name)); + if (TypeSet* types = property.maybeTypes()) { + if (!types->empty() || types->nonDataProperty()) + return false; + } + if (singleton) { + if (CanHaveEmptyPropertyTypesForOwnProperty(singleton)) { + MOZ_ASSERT(singleton->is<GlobalObject>()); + *guardGlobal = true; + } + } + + JSObject* proto = checkNurseryObject(key->proto().toObjectOrNull()); + + if (proto == foundProto) + break; + if (!proto) { + // The foundProto being searched for did not show up on the + // object's prototype chain. + return false; + } + key = TypeSet::ObjectKey::get(proto); + } + } + + return true; +} + +void +IonBuilder::freezePropertiesForCommonPrototype(TemporaryTypeSet* types, PropertyName* name, + JSObject* foundProto, + bool allowEmptyTypesforGlobal/* = false*/) +{ + for (unsigned i = 0; i < types->getObjectCount(); i++) { + // If we found a Singleton object's own-property, there's nothing to + // freeze. + if (types->getSingleton(i) == foundProto) + continue; + + TypeSet::ObjectKey* key = types->getObject(i); + if (!key) + continue; + + while (true) { + HeapTypeSetKey property = key->property(NameToId(name)); + JS_ALWAYS_TRUE(!property.isOwnProperty(constraints(), allowEmptyTypesforGlobal)); + + // Don't mark the proto. It will be held down by the shape + // guard. This allows us to use properties found on prototypes + // with properties unknown to TI. + if (key->proto() == TaggedProto(foundProto)) + break; + key = TypeSet::ObjectKey::get(key->proto().toObjectOrNull()); + } + } +} + +bool +IonBuilder::testCommonGetterSetter(TemporaryTypeSet* types, PropertyName* name, + bool isGetter, JSObject* foundProto, Shape* lastProperty, + JSFunction* getterOrSetter, + MDefinition** guard, + Shape* globalShape/* = nullptr*/, + MDefinition** globalGuard/* = nullptr */) +{ + MOZ_ASSERT_IF(globalShape, globalGuard); + bool guardGlobal; + + // Check if all objects being accessed will lookup the name through foundProto. + if (!objectsHaveCommonPrototype(types, name, isGetter, foundProto, &guardGlobal) || + (guardGlobal && !globalShape)) + { + trackOptimizationOutcome(TrackedOutcome::MultiProtoPaths); + return false; + } + + // We can optimize the getter/setter, so freeze all involved properties to + // ensure there isn't a lower shadowing getter or setter installed in the + // future. + freezePropertiesForCommonPrototype(types, name, foundProto, guardGlobal); + + // Add a shape guard on the prototype we found the property on. The rest of + // the prototype chain is guarded by TI freezes, except when name is a global + // name. In this case, we also have to guard on the globals shape to be able + // to optimize, because the way global property sets are handled means + // freezing doesn't work for what we want here. Note that a shape guard is + // good enough here, even in the proxy case, because we have ensured there + // are no lookup hooks for this property. + if (guardGlobal) { + JSObject* obj = &script()->global(); + MDefinition* globalObj = constant(ObjectValue(*obj)); + *globalGuard = addShapeGuard(globalObj, globalShape, Bailout_ShapeGuard); + } + + if (foundProto->isNative()) { + NativeObject& nativeProto = foundProto->as<NativeObject>(); + if (nativeProto.lastProperty() == lastProperty) { + // The proto shape is the same as it was at the point when we + // created the baseline IC, so looking up the prop on the object as + // it is now should be safe. + Shape* propShape = nativeProto.lookupPure(name); + MOZ_ASSERT_IF(isGetter, propShape->getterObject() == getterOrSetter); + MOZ_ASSERT_IF(!isGetter, propShape->setterObject() == getterOrSetter); + if (propShape && !propShape->configurable()) + return true; + } + } + + MInstruction* wrapper = constant(ObjectValue(*foundProto)); + *guard = addShapeGuard(wrapper, lastProperty, Bailout_ShapeGuard); + return true; +} + +void +IonBuilder::replaceMaybeFallbackFunctionGetter(MGetPropertyCache* cache) +{ + // Discard the last prior resume point of the previous MGetPropertyCache. + WrapMGetPropertyCache rai(maybeFallbackFunctionGetter_); + maybeFallbackFunctionGetter_ = cache; +} + +bool +IonBuilder::annotateGetPropertyCache(MDefinition* obj, PropertyName* name, + MGetPropertyCache* getPropCache, TemporaryTypeSet* objTypes, + TemporaryTypeSet* pushedTypes) +{ + // Ensure every pushed value is a singleton. + if (pushedTypes->unknownObject() || pushedTypes->baseFlags() != 0) + return true; + + for (unsigned i = 0; i < pushedTypes->getObjectCount(); i++) { + if (pushedTypes->getGroup(i) != nullptr) + return true; + } + + // Object's typeset should be a proper object + if (!objTypes || objTypes->baseFlags() || objTypes->unknownObject()) + return true; + + unsigned int objCount = objTypes->getObjectCount(); + if (objCount == 0) + return true; + + InlinePropertyTable* inlinePropTable = getPropCache->initInlinePropertyTable(alloc(), pc); + if (!inlinePropTable) + return false; + + // Ensure that the relevant property typeset for each group is + // is a single-object typeset containing a JSFunction + for (unsigned int i = 0; i < objCount; i++) { + ObjectGroup* group = objTypes->getGroup(i); + if (!group) + continue; + TypeSet::ObjectKey* key = TypeSet::ObjectKey::get(group); + if (key->unknownProperties() || !key->proto().isObject()) + continue; + JSObject* proto = checkNurseryObject(key->proto().toObject()); + + const Class* clasp = key->clasp(); + if (!ClassHasEffectlessLookup(clasp) || ObjectHasExtraOwnProperty(compartment, key, NameToId(name))) + continue; + + HeapTypeSetKey ownTypes = key->property(NameToId(name)); + if (ownTypes.isOwnProperty(constraints())) + continue; + + JSObject* singleton = testSingletonProperty(proto, NameToId(name)); + if (!singleton || !singleton->is<JSFunction>()) + continue; + + // Don't add cases corresponding to non-observed pushes + if (!pushedTypes->hasType(TypeSet::ObjectType(singleton))) + continue; + + if (!inlinePropTable->addEntry(alloc(), group, &singleton->as<JSFunction>())) + return false; + } + + if (inlinePropTable->numEntries() == 0) { + getPropCache->clearInlinePropertyTable(); + return true; + } + +#ifdef JS_JITSPEW + if (inlinePropTable->numEntries() > 0) + JitSpew(JitSpew_Inlining, "Annotated GetPropertyCache with %d/%d inline cases", + (int) inlinePropTable->numEntries(), (int) objCount); +#endif + + // If we successfully annotated the GetPropertyCache and there are inline cases, + // then keep a resume point of the state right before this instruction for use + // later when we have to bail out to this point in the fallback case of a + // PolyInlineDispatch. + if (inlinePropTable->numEntries() > 0) { + // Push the object back onto the stack temporarily to capture the resume point. + current->push(obj); + MResumePoint* resumePoint = MResumePoint::New(alloc(), current, pc, + MResumePoint::ResumeAt); + if (!resumePoint) + return false; + inlinePropTable->setPriorResumePoint(resumePoint); + replaceMaybeFallbackFunctionGetter(getPropCache); + current->pop(); + } + return true; +} + +// Returns true if an idempotent cache has ever invalidated this script +// or an outer script. +bool +IonBuilder::invalidatedIdempotentCache() +{ + IonBuilder* builder = this; + do { + if (builder->script()->invalidatedIdempotentCache()) + return true; + builder = builder->callerBuilder_; + } while (builder); + + return false; +} + +bool +IonBuilder::loadSlot(MDefinition* obj, size_t slot, size_t nfixed, MIRType rvalType, + BarrierKind barrier, TemporaryTypeSet* types) +{ + if (slot < nfixed) { + MLoadFixedSlot* load = MLoadFixedSlot::New(alloc(), obj, slot); + current->add(load); + current->push(load); + + load->setResultType(rvalType); + return pushTypeBarrier(load, types, barrier); + } + + MSlots* slots = MSlots::New(alloc(), obj); + current->add(slots); + + MLoadSlot* load = MLoadSlot::New(alloc(), slots, slot - nfixed); + current->add(load); + current->push(load); + + load->setResultType(rvalType); + return pushTypeBarrier(load, types, barrier); +} + +bool +IonBuilder::loadSlot(MDefinition* obj, Shape* shape, MIRType rvalType, + BarrierKind barrier, TemporaryTypeSet* types) +{ + return loadSlot(obj, shape->slot(), shape->numFixedSlots(), rvalType, barrier, types); +} + +bool +IonBuilder::storeSlot(MDefinition* obj, size_t slot, size_t nfixed, + MDefinition* value, bool needsBarrier, + MIRType slotType /* = MIRType::None */) +{ + if (slot < nfixed) { + MStoreFixedSlot* store = MStoreFixedSlot::New(alloc(), obj, slot, value); + current->add(store); + current->push(value); + if (needsBarrier) + store->setNeedsBarrier(); + return resumeAfter(store); + } + + MSlots* slots = MSlots::New(alloc(), obj); + current->add(slots); + + MStoreSlot* store = MStoreSlot::New(alloc(), slots, slot - nfixed, value); + current->add(store); + current->push(value); + if (needsBarrier) + store->setNeedsBarrier(); + if (slotType != MIRType::None) + store->setSlotType(slotType); + return resumeAfter(store); +} + +bool +IonBuilder::storeSlot(MDefinition* obj, Shape* shape, MDefinition* value, bool needsBarrier, + MIRType slotType /* = MIRType::None */) +{ + MOZ_ASSERT(shape->writable()); + return storeSlot(obj, shape->slot(), shape->numFixedSlots(), value, needsBarrier, slotType); +} + +bool +IonBuilder::shouldAbortOnPreliminaryGroups(MDefinition *obj) +{ + // Watch for groups which still have preliminary object information and + // have not had the new script properties or unboxed layout analyses + // performed. Normally this is done after a small number of the objects + // have been created, but if only a few have been created we can still + // perform the analysis with a smaller object population. The analysis can + // have side effects so we will end up aborting compilation after building + // finishes and retrying later. + TemporaryTypeSet *types = obj->resultTypeSet(); + if (!types || types->unknownObject()) + return false; + + bool preliminary = false; + for (size_t i = 0; i < types->getObjectCount(); i++) { + TypeSet::ObjectKey* key = types->getObject(i); + if (!key) + continue; + + if (ObjectGroup* group = key->maybeGroup()) { + if (group->hasUnanalyzedPreliminaryObjects()) { + addAbortedPreliminaryGroup(group); + preliminary = true; + } + } + } + + return preliminary; +} + +MDefinition* +IonBuilder::maybeUnboxForPropertyAccess(MDefinition* def) +{ + if (def->type() != MIRType::Value) + return def; + + MIRType type = inspector->expectedPropertyAccessInputType(pc); + if (type == MIRType::Value || !def->mightBeType(type)) + return def; + + MUnbox* unbox = MUnbox::New(alloc(), def, type, MUnbox::Fallible); + current->add(unbox); + + // Fixup type information for a common case where a property call + // is converted to the following bytecodes + // + // a.foo() + // ================= Compiles to ================ + // LOAD "a" + // DUP + // CALLPROP "foo" + // SWAP + // CALL 0 + // + // If we have better type information to unbox the first copy going into + // the CALLPROP operation, we can replace the duplicated copy on the + // stack as well. + if (*pc == JSOP_CALLPROP || *pc == JSOP_CALLELEM) { + uint32_t idx = current->stackDepth() - 1; + MOZ_ASSERT(current->getSlot(idx) == def); + current->setSlot(idx, unbox); + } + + return unbox; +} + +bool +IonBuilder::jsop_getprop(PropertyName* name) +{ + bool emitted = false; + startTrackingOptimizations(); + + MDefinition* obj = current->pop(); + TemporaryTypeSet* types = bytecodeTypes(pc); + + trackTypeInfo(TrackedTypeSite::Receiver, obj->type(), obj->resultTypeSet()); + + if (!info().isAnalysis()) { + // The calls below can abort compilation, so we only try this if we're + // not analyzing. + // Try to optimize arguments.length. + trackOptimizationAttempt(TrackedStrategy::GetProp_ArgumentsLength); + if (!getPropTryArgumentsLength(&emitted, obj) || emitted) + return emitted; + + // Try to optimize arguments.callee. + trackOptimizationAttempt(TrackedStrategy::GetProp_ArgumentsCallee); + if (!getPropTryArgumentsCallee(&emitted, obj, name) || emitted) + return emitted; + } + + obj = maybeUnboxForPropertyAccess(obj); + if (obj->type() == MIRType::Object) + obj = convertUnboxedObjects(obj); + + BarrierKind barrier = PropertyReadNeedsTypeBarrier(analysisContext, constraints(), + obj, name, types); + + // Try to optimize to a specific constant. + trackOptimizationAttempt(TrackedStrategy::GetProp_InferredConstant); + if (barrier == BarrierKind::NoBarrier) { + if (!getPropTryInferredConstant(&emitted, obj, name, types) || emitted) + return emitted; + } else { + trackOptimizationOutcome(TrackedOutcome::NeedsTypeBarrier); + } + + // Always use a call if we are performing analysis and + // not actually emitting code, to simplify later analysis. Also skip deeper + // analysis if there are no known types for this operation, as it will + // always invalidate when executing. + if (info().isAnalysis() || types->empty() || shouldAbortOnPreliminaryGroups(obj)) { + if (types->empty()) { + // Since no further optimizations will be tried, use the IC + // strategy, which would have been the last one to be tried, as a + // sentinel value for why everything failed. + trackOptimizationAttempt(TrackedStrategy::GetProp_InlineCache); + trackOptimizationOutcome(TrackedOutcome::NoTypeInfo); + } + + MCallGetProperty* call = MCallGetProperty::New(alloc(), obj, name); + current->add(call); + + // During the definite properties analysis we can still try to bake in + // constants read off the prototype chain, to allow inlining later on. + // In this case we still need the getprop call so that the later + // analysis knows when the |this| value has been read from. + if (info().isAnalysis()) { + if (!getPropTryConstant(&emitted, obj, NameToId(name), types) || emitted) + return emitted; + } + + current->push(call); + return resumeAfter(call) && pushTypeBarrier(call, types, BarrierKind::TypeSet); + } + + // Try to optimize accesses on outer window proxies, for example window.foo. + // This needs to come before the various strategies getPropTryInnerize tries + // internally, since some of those strategies will "succeed" in silly ways + // even for an outer object. + trackOptimizationAttempt(TrackedStrategy::GetProp_Innerize); + if (!getPropTryInnerize(&emitted, obj, name, types) || emitted) + return emitted; + + if (!forceInlineCaches()) { + // Try to hardcode known constants. + trackOptimizationAttempt(TrackedStrategy::GetProp_Constant); + if (!getPropTryConstant(&emitted, obj, NameToId(name), types) || emitted) + return emitted; + + // Try to hardcode known not-defined + trackOptimizationAttempt(TrackedStrategy::GetProp_NotDefined); + if (!getPropTryNotDefined(&emitted, obj, NameToId(name), types) || emitted) + return emitted; + + // Try to emit loads from known binary data blocks + trackOptimizationAttempt(TrackedStrategy::GetProp_TypedObject); + if (!getPropTryTypedObject(&emitted, obj, name) || emitted) + return emitted; + + // Try to emit loads from definite slots. + trackOptimizationAttempt(TrackedStrategy::GetProp_DefiniteSlot); + if (!getPropTryDefiniteSlot(&emitted, obj, name, barrier, types) || emitted) + return emitted; + + // Try to emit loads from unboxed objects. + trackOptimizationAttempt(TrackedStrategy::GetProp_Unboxed); + if (!getPropTryUnboxed(&emitted, obj, name, barrier, types) || emitted) + return emitted; + + // Try to inline a common property getter, or make a call. + trackOptimizationAttempt(TrackedStrategy::GetProp_CommonGetter); + if (!getPropTryCommonGetter(&emitted, obj, name, types) || emitted) + return emitted; + + // Try to emit a monomorphic/polymorphic access based on baseline caches. + trackOptimizationAttempt(TrackedStrategy::GetProp_InlineAccess); + if (!getPropTryInlineAccess(&emitted, obj, name, barrier, types) || emitted) + return emitted; + + // Try to emit loads from a module namespace. + trackOptimizationAttempt(TrackedStrategy::GetProp_ModuleNamespace); + if (!getPropTryModuleNamespace(&emitted, obj, name, barrier, types) || emitted) + return emitted; + } + + // Try to emit a polymorphic cache. + trackOptimizationAttempt(TrackedStrategy::GetProp_InlineCache); + if (!getPropTryCache(&emitted, obj, name, barrier, types) || emitted) + return emitted; + + // Try to emit a shared stub. + trackOptimizationAttempt(TrackedStrategy::GetProp_SharedCache); + if (!getPropTrySharedStub(&emitted, obj, types) || emitted) + return emitted; + + // Emit a call. + MCallGetProperty* call = MCallGetProperty::New(alloc(), obj, name); + current->add(call); + current->push(call); + if (!resumeAfter(call)) + return false; + + if (*pc == JSOP_CALLPROP && IsNullOrUndefined(obj->type())) { + // Due to inlining, it's possible the observed TypeSet is non-empty, + // even though we know |obj| is null/undefined and the MCallGetProperty + // will throw. Don't push a TypeBarrier in this case, to avoid + // inlining the following (unreachable) JSOP_CALL. + return true; + } + + return pushTypeBarrier(call, types, BarrierKind::TypeSet); +} + +bool +IonBuilder::improveThisTypesForCall() +{ + // After a CALLPROP (or CALLELEM) for obj.prop(), the this-value and callee + // for the call are on top of the stack: + // + // ... [this: obj], [callee: obj.prop] + // + // If obj is null or undefined, obj.prop would have thrown an exception so + // at this point we can remove null and undefined from obj's TypeSet, to + // improve type information for the call that will follow. + + MOZ_ASSERT(*pc == JSOP_CALLPROP || *pc == JSOP_CALLELEM); + + // Ensure |this| has types {object, null/undefined}. + MDefinition* thisDef = current->peek(-2); + if (thisDef->type() != MIRType::Value || + !thisDef->mightBeType(MIRType::Object) || + !thisDef->resultTypeSet() || + !thisDef->resultTypeSet()->objectOrSentinel()) + { + return true; + } + + // Remove null/undefined from the TypeSet. + TemporaryTypeSet* types = thisDef->resultTypeSet()->cloneObjectsOnly(alloc_->lifoAlloc()); + if (!types) + return false; + + MFilterTypeSet* filter = MFilterTypeSet::New(alloc(), thisDef, types); + current->add(filter); + current->rewriteAtDepth(-2, filter); + + // FilterTypeSetPolicy::adjustInputs will insert an infallible Unbox(Object) + // for the input. Don't hoist this unbox above the getprop or getelem + // operation. + filter->setDependency(current->peek(-1)->toInstruction()); + return true; +} + +bool +IonBuilder::checkIsDefinitelyOptimizedArguments(MDefinition* obj, bool* isOptimizedArgs) +{ + if (obj->type() != MIRType::MagicOptimizedArguments) { + if (script()->argumentsHasVarBinding() && + obj->mightBeType(MIRType::MagicOptimizedArguments)) + { + return abort("Type is not definitely lazy arguments."); + } + + *isOptimizedArgs = false; + return true; + } + + *isOptimizedArgs = true; + return true; +} + +bool +IonBuilder::getPropTryInferredConstant(bool* emitted, MDefinition* obj, PropertyName* name, + TemporaryTypeSet* types) +{ + MOZ_ASSERT(*emitted == false); + + // Need a result typeset to optimize. + TemporaryTypeSet* objTypes = obj->resultTypeSet(); + if (!objTypes) { + trackOptimizationOutcome(TrackedOutcome::NoTypeInfo); + return true; + } + + JSObject* singleton = objTypes->maybeSingleton(); + if (!singleton) { + trackOptimizationOutcome(TrackedOutcome::NotSingleton); + return true; + } + + TypeSet::ObjectKey* key = TypeSet::ObjectKey::get(singleton); + if (key->unknownProperties()) { + trackOptimizationOutcome(TrackedOutcome::UnknownProperties); + return true; + } + + HeapTypeSetKey property = key->property(NameToId(name)); + + Value constantValue = UndefinedValue(); + if (property.constant(constraints(), &constantValue)) { + spew("Optimized constant property"); + obj->setImplicitlyUsedUnchecked(); + pushConstant(constantValue); + types->addType(TypeSet::GetValueType(constantValue), alloc_->lifoAlloc()); + trackOptimizationSuccess(); + *emitted = true; + } + + return true; +} + +bool +IonBuilder::getPropTryArgumentsLength(bool* emitted, MDefinition* obj) +{ + MOZ_ASSERT(*emitted == false); + + if (JSOp(*pc) != JSOP_LENGTH) + return true; + + bool isOptimizedArgs = false; + if (!checkIsDefinitelyOptimizedArguments(obj, &isOptimizedArgs)) + return false; + if (!isOptimizedArgs) + return true; + + trackOptimizationSuccess(); + *emitted = true; + + obj->setImplicitlyUsedUnchecked(); + + // We don't know anything from the callee + if (inliningDepth_ == 0) { + MInstruction* ins = MArgumentsLength::New(alloc()); + current->add(ins); + current->push(ins); + return true; + } + + // We are inlining and know the number of arguments the callee pushed + pushConstant(Int32Value(inlineCallInfo_->argv().length())); + return true; +} + +bool +IonBuilder::getPropTryArgumentsCallee(bool* emitted, MDefinition* obj, PropertyName* name) +{ + MOZ_ASSERT(*emitted == false); + + if (name != names().callee) + return true; + + bool isOptimizedArgs = false; + if (!checkIsDefinitelyOptimizedArguments(obj, &isOptimizedArgs)) + return false; + if (!isOptimizedArgs) + return true; + + MOZ_ASSERT(script()->hasMappedArgsObj()); + + obj->setImplicitlyUsedUnchecked(); + current->push(getCallee()); + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +bool +IonBuilder::getPropTryConstant(bool* emitted, MDefinition* obj, jsid id, TemporaryTypeSet* types) +{ + MOZ_ASSERT(*emitted == false); + + if (!types->mightBeMIRType(MIRType::Object)) { + // If we have not observed an object result here, don't look for a + // singleton constant. + trackOptimizationOutcome(TrackedOutcome::NotObject); + return true; + } + + JSObject* singleton = testSingletonPropertyTypes(obj, id); + if (!singleton) { + trackOptimizationOutcome(TrackedOutcome::NotSingleton); + return true; + } + + // Property access is a known constant -- safe to emit. + obj->setImplicitlyUsedUnchecked(); + + pushConstant(ObjectValue(*singleton)); + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +bool +IonBuilder::getPropTryNotDefined(bool* emitted, MDefinition* obj, jsid id, TemporaryTypeSet* types) +{ + MOZ_ASSERT(*emitted == false); + + if (!types->mightBeMIRType(MIRType::Undefined)) { + // Only optimize if we expect this property access to return undefined. + trackOptimizationOutcome(TrackedOutcome::NotUndefined); + return true; + } + + ResultWithOOM<bool> res = testNotDefinedProperty(obj, id); + if (res.oom) + return false; + if (!res.value) { + trackOptimizationOutcome(TrackedOutcome::GenericFailure); + return true; + } + + obj->setImplicitlyUsedUnchecked(); + pushConstant(UndefinedValue()); + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +bool +IonBuilder::getPropTryTypedObject(bool* emitted, + MDefinition* obj, + PropertyName* name) +{ + TypedObjectPrediction fieldPrediction; + size_t fieldOffset; + size_t fieldIndex; + if (!typedObjectHasField(obj, name, &fieldOffset, &fieldPrediction, &fieldIndex)) + return true; + + switch (fieldPrediction.kind()) { + case type::Simd: + // FIXME (bug 894104): load into a MIRType::float32x4 etc + return true; + + case type::Struct: + case type::Array: + return getPropTryComplexPropOfTypedObject(emitted, + obj, + fieldOffset, + fieldPrediction, + fieldIndex); + + case type::Reference: + return getPropTryReferencePropOfTypedObject(emitted, + obj, + fieldOffset, + fieldPrediction, + name); + + case type::Scalar: + return getPropTryScalarPropOfTypedObject(emitted, + obj, + fieldOffset, + fieldPrediction); + } + + MOZ_CRASH("Bad kind"); +} + +bool +IonBuilder::getPropTryScalarPropOfTypedObject(bool* emitted, MDefinition* typedObj, + int32_t fieldOffset, + TypedObjectPrediction fieldPrediction) +{ + // Must always be loading the same scalar type + Scalar::Type fieldType = fieldPrediction.scalarType(); + + // Don't optimize if the typed object's underlying buffer may be detached. + TypeSet::ObjectKey* globalKey = TypeSet::ObjectKey::get(&script()->global()); + if (globalKey->hasFlags(constraints(), OBJECT_FLAG_TYPED_OBJECT_HAS_DETACHED_BUFFER)) + return true; + + trackOptimizationSuccess(); + *emitted = true; + + LinearSum byteOffset(alloc()); + if (!byteOffset.add(fieldOffset)) + setForceAbort(); + + return pushScalarLoadFromTypedObject(typedObj, byteOffset, fieldType); +} + +bool +IonBuilder::getPropTryReferencePropOfTypedObject(bool* emitted, MDefinition* typedObj, + int32_t fieldOffset, + TypedObjectPrediction fieldPrediction, + PropertyName* name) +{ + ReferenceTypeDescr::Type fieldType = fieldPrediction.referenceType(); + + TypeSet::ObjectKey* globalKey = TypeSet::ObjectKey::get(&script()->global()); + if (globalKey->hasFlags(constraints(), OBJECT_FLAG_TYPED_OBJECT_HAS_DETACHED_BUFFER)) + return true; + + trackOptimizationSuccess(); + *emitted = true; + + LinearSum byteOffset(alloc()); + if (!byteOffset.add(fieldOffset)) + setForceAbort(); + + return pushReferenceLoadFromTypedObject(typedObj, byteOffset, fieldType, name); +} + +bool +IonBuilder::getPropTryComplexPropOfTypedObject(bool* emitted, + MDefinition* typedObj, + int32_t fieldOffset, + TypedObjectPrediction fieldPrediction, + size_t fieldIndex) +{ + // Don't optimize if the typed object's underlying buffer may be detached. + TypeSet::ObjectKey* globalKey = TypeSet::ObjectKey::get(&script()->global()); + if (globalKey->hasFlags(constraints(), OBJECT_FLAG_TYPED_OBJECT_HAS_DETACHED_BUFFER)) + return true; + + // OK, perform the optimization + + // Identify the type object for the field. + MDefinition* type = loadTypedObjectType(typedObj); + MDefinition* fieldTypeObj = typeObjectForFieldFromStructType(type, fieldIndex); + + LinearSum byteOffset(alloc()); + if (!byteOffset.add(fieldOffset)) + setForceAbort(); + + return pushDerivedTypedObject(emitted, typedObj, byteOffset, + fieldPrediction, fieldTypeObj); +} + +MDefinition* +IonBuilder::convertUnboxedObjects(MDefinition* obj) +{ + // If obj might be in any particular unboxed group which should be + // converted to a native representation, perform that conversion. This does + // not guarantee the object will not have such a group afterwards, if the + // object's possible groups are not precisely known. + TemporaryTypeSet* types = obj->resultTypeSet(); + if (!types || types->unknownObject() || !types->objectOrSentinel()) + return obj; + + BaselineInspector::ObjectGroupVector list(alloc()); + for (size_t i = 0; i < types->getObjectCount(); i++) { + TypeSet::ObjectKey* key = obj->resultTypeSet()->getObject(i); + if (!key || !key->isGroup()) + continue; + + if (UnboxedLayout* layout = key->group()->maybeUnboxedLayout()) { + AutoEnterOOMUnsafeRegion oomUnsafe; + if (layout->nativeGroup() && !list.append(key->group())) + oomUnsafe.crash("IonBuilder::convertUnboxedObjects"); + } + } + + return convertUnboxedObjects(obj, list); +} + +MDefinition* +IonBuilder::convertUnboxedObjects(MDefinition* obj, + const BaselineInspector::ObjectGroupVector& list) +{ + for (size_t i = 0; i < list.length(); i++) { + ObjectGroup* group = list[i]; + if (TemporaryTypeSet* types = obj->resultTypeSet()) { + if (!types->hasType(TypeSet::ObjectType(group))) + continue; + } + obj = MConvertUnboxedObjectToNative::New(alloc(), obj, group); + current->add(obj->toInstruction()); + } + return obj; +} + +bool +IonBuilder::getPropTryDefiniteSlot(bool* emitted, MDefinition* obj, PropertyName* name, + BarrierKind barrier, TemporaryTypeSet* types) +{ + MOZ_ASSERT(*emitted == false); + + uint32_t nfixed; + uint32_t slot = getDefiniteSlot(obj->resultTypeSet(), name, &nfixed); + if (slot == UINT32_MAX) + return true; + + if (obj->type() != MIRType::Object) { + MGuardObject* guard = MGuardObject::New(alloc(), obj); + current->add(guard); + obj = guard; + } + + MInstruction* load; + if (slot < nfixed) { + load = MLoadFixedSlot::New(alloc(), obj, slot); + } else { + MInstruction* slots = MSlots::New(alloc(), obj); + current->add(slots); + + load = MLoadSlot::New(alloc(), slots, slot - nfixed); + } + + if (barrier == BarrierKind::NoBarrier) + load->setResultType(types->getKnownMIRType()); + + current->add(load); + current->push(load); + + if (!pushTypeBarrier(load, types, barrier)) + return false; + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +bool +IonBuilder::getPropTryModuleNamespace(bool* emitted, MDefinition* obj, PropertyName* name, + BarrierKind barrier, TemporaryTypeSet* types) +{ + MOZ_ASSERT(*emitted == false); + + TemporaryTypeSet* objTypes = obj->resultTypeSet(); + if (!objTypes) { + trackOptimizationOutcome(TrackedOutcome::NoTypeInfo); + return true; + } + + JSObject* singleton = objTypes->maybeSingleton(); + if (!singleton) { + trackOptimizationOutcome(TrackedOutcome::NotSingleton); + return true; + } + + if (!singleton->is<ModuleNamespaceObject>()) { + trackOptimizationOutcome(TrackedOutcome::NotModuleNamespace); + return true; + } + + ModuleNamespaceObject* ns = &singleton->as<ModuleNamespaceObject>(); + ModuleEnvironmentObject* env; + Shape* shape; + if (!ns->bindings().lookup(NameToId(name), &env, &shape)) { + trackOptimizationOutcome(TrackedOutcome::UnknownProperty); + return true; + } + + obj->setImplicitlyUsedUnchecked(); + MConstant* envConst = constant(ObjectValue(*env)); + uint32_t slot = shape->slot(); + uint32_t nfixed = env->numFixedSlots(); + if (!loadSlot(envConst, slot, nfixed, types->getKnownMIRType(), barrier, types)) + return false; + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +MInstruction* +IonBuilder::loadUnboxedProperty(MDefinition* obj, size_t offset, JSValueType unboxedType, + BarrierKind barrier, TemporaryTypeSet* types) +{ + // loadUnboxedValue is designed to load any value as if it were contained in + // an array. Thus a property offset is converted to an index, when the + // object is reinterpreted as an array of properties of the same size. + size_t index = offset / UnboxedTypeSize(unboxedType); + MInstruction* indexConstant = MConstant::New(alloc(), Int32Value(index)); + current->add(indexConstant); + + return loadUnboxedValue(obj, UnboxedPlainObject::offsetOfData(), + indexConstant, unboxedType, barrier, types); +} + +MInstruction* +IonBuilder::loadUnboxedValue(MDefinition* elements, size_t elementsOffset, + MDefinition* index, JSValueType unboxedType, + BarrierKind barrier, TemporaryTypeSet* types) +{ + MInstruction* load; + switch (unboxedType) { + case JSVAL_TYPE_BOOLEAN: + load = MLoadUnboxedScalar::New(alloc(), elements, index, Scalar::Uint8, + DoesNotRequireMemoryBarrier, elementsOffset); + load->setResultType(MIRType::Boolean); + break; + + case JSVAL_TYPE_INT32: + load = MLoadUnboxedScalar::New(alloc(), elements, index, Scalar::Int32, + DoesNotRequireMemoryBarrier, elementsOffset); + load->setResultType(MIRType::Int32); + break; + + case JSVAL_TYPE_DOUBLE: + load = MLoadUnboxedScalar::New(alloc(), elements, index, Scalar::Float64, + DoesNotRequireMemoryBarrier, elementsOffset, + /* canonicalizeDoubles = */ false); + load->setResultType(MIRType::Double); + break; + + case JSVAL_TYPE_STRING: + load = MLoadUnboxedString::New(alloc(), elements, index, elementsOffset); + break; + + case JSVAL_TYPE_OBJECT: { + MLoadUnboxedObjectOrNull::NullBehavior nullBehavior; + if (types->hasType(TypeSet::NullType())) + nullBehavior = MLoadUnboxedObjectOrNull::HandleNull; + else if (barrier != BarrierKind::NoBarrier) + nullBehavior = MLoadUnboxedObjectOrNull::BailOnNull; + else + nullBehavior = MLoadUnboxedObjectOrNull::NullNotPossible; + load = MLoadUnboxedObjectOrNull::New(alloc(), elements, index, nullBehavior, + elementsOffset); + break; + } + + default: + MOZ_CRASH(); + } + + current->add(load); + return load; +} + +bool +IonBuilder::getPropTryUnboxed(bool* emitted, MDefinition* obj, PropertyName* name, + BarrierKind barrier, TemporaryTypeSet* types) +{ + MOZ_ASSERT(*emitted == false); + + JSValueType unboxedType; + uint32_t offset = getUnboxedOffset(obj->resultTypeSet(), name, &unboxedType); + if (offset == UINT32_MAX) + return true; + + if (obj->type() != MIRType::Object) { + MGuardObject* guard = MGuardObject::New(alloc(), obj); + current->add(guard); + obj = guard; + } + + MInstruction* load = loadUnboxedProperty(obj, offset, unboxedType, barrier, types); + current->push(load); + + if (!pushTypeBarrier(load, types, barrier)) + return false; + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +MDefinition* +IonBuilder::addShapeGuardsForGetterSetter(MDefinition* obj, JSObject* holder, Shape* holderShape, + const BaselineInspector::ReceiverVector& receivers, + const BaselineInspector::ObjectGroupVector& convertUnboxedGroups, + bool isOwnProperty) +{ + MOZ_ASSERT(holder); + MOZ_ASSERT(holderShape); + + obj = convertUnboxedObjects(obj, convertUnboxedGroups); + + if (isOwnProperty) { + MOZ_ASSERT(receivers.empty()); + return addShapeGuard(obj, holderShape, Bailout_ShapeGuard); + } + + MDefinition* holderDef = constant(ObjectValue(*holder)); + addShapeGuard(holderDef, holderShape, Bailout_ShapeGuard); + + return addGuardReceiverPolymorphic(obj, receivers); +} + +bool +IonBuilder::getPropTryCommonGetter(bool* emitted, MDefinition* obj, PropertyName* name, + TemporaryTypeSet* types) +{ + MOZ_ASSERT(*emitted == false); + + Shape* lastProperty = nullptr; + JSFunction* commonGetter = nullptr; + Shape* globalShape = nullptr; + JSObject* foundProto = nullptr; + bool isOwnProperty = false; + BaselineInspector::ReceiverVector receivers(alloc()); + BaselineInspector::ObjectGroupVector convertUnboxedGroups(alloc()); + if (!inspector->commonGetPropFunction(pc, &foundProto, &lastProperty, &commonGetter, + &globalShape, &isOwnProperty, + receivers, convertUnboxedGroups)) + { + return true; + } + + TemporaryTypeSet* objTypes = obj->resultTypeSet(); + MDefinition* guard = nullptr; + MDefinition* globalGuard = nullptr; + bool canUseTIForGetter = + testCommonGetterSetter(objTypes, name, /* isGetter = */ true, + foundProto, lastProperty, commonGetter, &guard, + globalShape, &globalGuard); + if (!canUseTIForGetter) { + // If type information is bad, we can still optimize the getter if we + // shape guard. + obj = addShapeGuardsForGetterSetter(obj, foundProto, lastProperty, + receivers, convertUnboxedGroups, + isOwnProperty); + if (!obj) + return false; + } + + bool isDOM = objTypes && objTypes->isDOMClass(constraints()); + + if (isDOM && testShouldDOMCall(objTypes, commonGetter, JSJitInfo::Getter)) { + const JSJitInfo* jitinfo = commonGetter->jitInfo(); + MInstruction* get; + if (jitinfo->isAlwaysInSlot) { + // If our object is a singleton and we know the property is + // constant (which is true if and only if the get doesn't alias + // anything), we can just read the slot here and use that constant. + JSObject* singleton = objTypes->maybeSingleton(); + if (singleton && jitinfo->aliasSet() == JSJitInfo::AliasNone) { + size_t slot = jitinfo->slotIndex; + *emitted = true; + pushConstant(GetReservedSlot(singleton, slot)); + return true; + } + + // We can't use MLoadFixedSlot here because it might not have the + // right aliasing behavior; we want to alias DOM setters as needed. + get = MGetDOMMember::New(alloc(), jitinfo, obj, guard, globalGuard); + } else { + get = MGetDOMProperty::New(alloc(), jitinfo, obj, guard, globalGuard); + } + if (!get) { + return false; + } + current->add(get); + current->push(get); + + if (get->isEffectful() && !resumeAfter(get)) + return false; + + if (!pushDOMTypeBarrier(get, types, commonGetter)) + return false; + + trackOptimizationOutcome(TrackedOutcome::DOM); + *emitted = true; + return true; + } + + // Don't call the getter with a primitive value. + if (obj->type() != MIRType::Object) { + MGuardObject* guardObj = MGuardObject::New(alloc(), obj); + current->add(guardObj); + obj = guardObj; + } + + // Spoof stack to expected state for call. + + // Make sure there's enough room + if (!current->ensureHasSlots(2)) + return false; + current->push(constant(ObjectValue(*commonGetter))); + + current->push(obj); + + CallInfo callInfo(alloc(), false); + if (!callInfo.init(current, 0)) + return false; + + if (commonGetter->isNative()) { + InliningStatus status = inlineNativeGetter(callInfo, commonGetter); + switch (status) { + case InliningStatus_Error: + return false; + case InliningStatus_WarmUpCountTooLow: + case InliningStatus_NotInlined: + break; + case InliningStatus_Inlined: + trackOptimizationOutcome(TrackedOutcome::Inlined); + *emitted = true; + return true; + } + } + + // Inline if we can, otherwise, forget it and just generate a call. + if (commonGetter->isInterpreted()) { + InliningDecision decision = makeInliningDecision(commonGetter, callInfo); + switch (decision) { + case InliningDecision_Error: + return false; + case InliningDecision_DontInline: + case InliningDecision_WarmUpCountTooLow: + break; + case InliningDecision_Inline: { + InliningStatus status = inlineScriptedCall(callInfo, commonGetter); + if (status == InliningStatus_Inlined) { + *emitted = true; + return true; + } + if (status == InliningStatus_Error) + return false; + break; + } + } + } + + if (!makeCall(commonGetter, callInfo)) + return false; + + // If the getter could have been inlined, don't track success. The call to + // makeInliningDecision above would have tracked a specific reason why we + // couldn't inline. + if (!commonGetter->isInterpreted()) + trackOptimizationSuccess(); + + *emitted = true; + return true; +} + +bool +IonBuilder::canInlinePropertyOpShapes(const BaselineInspector::ReceiverVector& receivers) +{ + if (receivers.empty()) { + trackOptimizationOutcome(TrackedOutcome::NoShapeInfo); + return false; + } + + for (size_t i = 0; i < receivers.length(); i++) { + // We inline the property access as long as the shape is not in + // dictionary mode. We cannot be sure that the shape is still a + // lastProperty, and calling Shape::search() on dictionary mode + // shapes that aren't lastProperty is invalid. + if (receivers[i].shape && receivers[i].shape->inDictionary()) { + trackOptimizationOutcome(TrackedOutcome::InDictionaryMode); + return false; + } + } + + return true; +} + +static Shape* +PropertyShapesHaveSameSlot(const BaselineInspector::ReceiverVector& receivers, jsid id) +{ + Shape* firstShape = nullptr; + for (size_t i = 0; i < receivers.length(); i++) { + if (receivers[i].group) + return nullptr; + + Shape* shape = receivers[i].shape->searchLinear(id); + MOZ_ASSERT(shape); + + if (i == 0) { + firstShape = shape; + } else if (shape->slot() != firstShape->slot() || + shape->numFixedSlots() != firstShape->numFixedSlots()) + { + return nullptr; + } + } + + return firstShape; +} + +bool +IonBuilder::getPropTryInlineAccess(bool* emitted, MDefinition* obj, PropertyName* name, + BarrierKind barrier, TemporaryTypeSet* types) +{ + MOZ_ASSERT(*emitted == false); + + BaselineInspector::ReceiverVector receivers(alloc()); + BaselineInspector::ObjectGroupVector convertUnboxedGroups(alloc()); + if (!inspector->maybeInfoForPropertyOp(pc, receivers, convertUnboxedGroups)) + return false; + + if (!canInlinePropertyOpShapes(receivers)) + return true; + + obj = convertUnboxedObjects(obj, convertUnboxedGroups); + + MIRType rvalType = types->getKnownMIRType(); + if (barrier != BarrierKind::NoBarrier || IsNullOrUndefined(rvalType)) + rvalType = MIRType::Value; + + if (receivers.length() == 1) { + if (!receivers[0].group) { + // Monomorphic load from a native object. + spew("Inlining monomorphic native GETPROP"); + + obj = addShapeGuard(obj, receivers[0].shape, Bailout_ShapeGuard); + + Shape* shape = receivers[0].shape->searchLinear(NameToId(name)); + MOZ_ASSERT(shape); + + if (!loadSlot(obj, shape, rvalType, barrier, types)) + return false; + + trackOptimizationOutcome(TrackedOutcome::Monomorphic); + *emitted = true; + return true; + } + + if (receivers[0].shape) { + // Monomorphic load from an unboxed object expando. + spew("Inlining monomorphic unboxed expando GETPROP"); + + obj = addGroupGuard(obj, receivers[0].group, Bailout_ShapeGuard); + obj = addUnboxedExpandoGuard(obj, /* hasExpando = */ true, Bailout_ShapeGuard); + + MInstruction* expando = MLoadUnboxedExpando::New(alloc(), obj); + current->add(expando); + + expando = addShapeGuard(expando, receivers[0].shape, Bailout_ShapeGuard); + + Shape* shape = receivers[0].shape->searchLinear(NameToId(name)); + MOZ_ASSERT(shape); + + if (!loadSlot(expando, shape, rvalType, barrier, types)) + return false; + + trackOptimizationOutcome(TrackedOutcome::Monomorphic); + *emitted = true; + return true; + } + + // Monomorphic load from an unboxed object. + ObjectGroup* group = receivers[0].group; + if (obj->resultTypeSet() && !obj->resultTypeSet()->hasType(TypeSet::ObjectType(group))) + return true; + + obj = addGroupGuard(obj, group, Bailout_ShapeGuard); + + const UnboxedLayout::Property* property = group->unboxedLayout().lookup(name); + MInstruction* load = loadUnboxedProperty(obj, property->offset, property->type, barrier, types); + current->push(load); + + if (!pushTypeBarrier(load, types, barrier)) + return false; + + trackOptimizationOutcome(TrackedOutcome::Monomorphic); + *emitted = true; + return true; + } + + MOZ_ASSERT(receivers.length() > 1); + spew("Inlining polymorphic GETPROP"); + + if (Shape* propShape = PropertyShapesHaveSameSlot(receivers, NameToId(name))) { + obj = addGuardReceiverPolymorphic(obj, receivers); + if (!obj) + return false; + + if (!loadSlot(obj, propShape, rvalType, barrier, types)) + return false; + + trackOptimizationOutcome(TrackedOutcome::Polymorphic); + *emitted = true; + return true; + } + + MGetPropertyPolymorphic* load = MGetPropertyPolymorphic::New(alloc(), obj, name); + current->add(load); + current->push(load); + + for (size_t i = 0; i < receivers.length(); i++) { + Shape* propShape = nullptr; + if (receivers[i].shape) { + propShape = receivers[i].shape->searchLinear(NameToId(name)); + MOZ_ASSERT(propShape); + } + if (!load->addReceiver(receivers[i], propShape)) + return false; + } + + if (failedShapeGuard_) + load->setNotMovable(); + + load->setResultType(rvalType); + if (!pushTypeBarrier(load, types, barrier)) + return false; + + trackOptimizationOutcome(TrackedOutcome::Polymorphic); + *emitted = true; + return true; +} + +bool +IonBuilder::getPropTryCache(bool* emitted, MDefinition* obj, PropertyName* name, + BarrierKind barrier, TemporaryTypeSet* types) +{ + MOZ_ASSERT(*emitted == false); + + // The input value must either be an object, or we should have strong suspicions + // that it can be safely unboxed to an object. + if (obj->type() != MIRType::Object) { + TemporaryTypeSet* types = obj->resultTypeSet(); + if (!types || !types->objectOrSentinel()) { + trackOptimizationOutcome(TrackedOutcome::NoTypeInfo); + return true; + } + } + + // Since getters have no guaranteed return values, we must barrier in order to be + // able to attach stubs for them. + if (inspector->hasSeenAccessedGetter(pc)) + barrier = BarrierKind::TypeSet; + + // Caches can read values from prototypes, so update the barrier to + // reflect such possible values. + if (barrier != BarrierKind::TypeSet) { + ResultWithOOM<BarrierKind> protoBarrier = + PropertyReadOnPrototypeNeedsTypeBarrier(this, obj, name, types); + if (protoBarrier.oom) + return false; + if (protoBarrier.value != BarrierKind::NoBarrier) { + MOZ_ASSERT(barrier <= protoBarrier.value); + barrier = protoBarrier.value; + } + } + + MConstant* id = constant(StringValue(name)); + MGetPropertyCache* load = MGetPropertyCache::New(alloc(), obj, id, + barrier == BarrierKind::TypeSet); + + // Try to mark the cache as idempotent. + if (obj->type() == MIRType::Object && !invalidatedIdempotentCache()) { + if (PropertyReadIsIdempotent(constraints(), obj, name)) + load->setIdempotent(); + } + + // When we are in the context of making a call from the value returned from + // a property, we query the typeObject for the given property name to fill + // the InlinePropertyTable of the GetPropertyCache. This information is + // then used in inlineCallsite and inlineCalls, if the "this" definition is + // matching the "object" definition of the GetPropertyCache (see + // CanInlineGetPropertyCache). + // + // If this GetPropertyCache is idempotent, then we can dispatch to the right + // function only by checking the typed object, instead of querying the value + // of the property. Thus this GetPropertyCache can be moved into the + // fallback path (see inlineObjectGroupFallback). Otherwise, we always have + // to do the GetPropertyCache, and we can dispatch based on the JSFunction + // value. + if (JSOp(*pc) == JSOP_CALLPROP && load->idempotent()) { + if (!annotateGetPropertyCache(obj, name, load, obj->resultTypeSet(), types)) + return false; + } + + current->add(load); + current->push(load); + + if (load->isEffectful() && !resumeAfter(load)) + return false; + + MIRType rvalType = types->getKnownMIRType(); + if (barrier != BarrierKind::NoBarrier || IsNullOrUndefined(rvalType)) + rvalType = MIRType::Value; + load->setResultType(rvalType); + + if (!pushTypeBarrier(load, types, barrier)) + return false; + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +bool +IonBuilder::getPropTrySharedStub(bool* emitted, MDefinition* obj, TemporaryTypeSet* types) +{ + MOZ_ASSERT(*emitted == false); + + // Try to emit a shared stub cache. + + if (JitOptions.disableSharedStubs) + return true; + + MInstruction* stub = MUnarySharedStub::New(alloc(), obj); + current->add(stub); + current->push(stub); + + if (!resumeAfter(stub)) + return false; + + // Due to inlining, it's possible the observed TypeSet is non-empty, + // even though we know |obj| is null/undefined and the MCallGetProperty + // will throw. Don't push a TypeBarrier in this case, to avoid + // inlining the following (unreachable) JSOP_CALL. + if (*pc != JSOP_CALLPROP || !IsNullOrUndefined(obj->type())) { + if (!pushTypeBarrier(stub, types, BarrierKind::TypeSet)) + return false; + } + + *emitted = true; + return true; +} + +MDefinition* +IonBuilder::tryInnerizeWindow(MDefinition* obj) +{ + // Try to optimize accesses on outer window proxies (window.foo, for + // example) to go directly to the inner window, the global. + // + // Callers should be careful not to pass the inner object to getters or + // setters that require outerization. + + if (obj->type() != MIRType::Object) + return obj; + + TemporaryTypeSet* types = obj->resultTypeSet(); + if (!types) + return obj; + + JSObject* singleton = types->maybeSingleton(); + if (!singleton) + return obj; + + if (!IsWindowProxy(singleton)) + return obj; + + // This must be a WindowProxy for the current Window/global. Else it'd be + // a cross-compartment wrapper and IsWindowProxy returns false for those. + MOZ_ASSERT(ToWindowIfWindowProxy(singleton) == &script()->global()); + + // When we navigate, the WindowProxy is brain transplanted and we'll mark + // its ObjectGroup as having unknown properties. The type constraint we add + // here will invalidate JIT code when this happens. + TypeSet::ObjectKey* key = TypeSet::ObjectKey::get(singleton); + if (key->hasFlags(constraints(), OBJECT_FLAG_UNKNOWN_PROPERTIES)) + return obj; + + obj->setImplicitlyUsedUnchecked(); + return constant(ObjectValue(script()->global())); +} + +bool +IonBuilder::getPropTryInnerize(bool* emitted, MDefinition* obj, PropertyName* name, + TemporaryTypeSet* types) +{ + // See the comment in tryInnerizeWindow for how this works. + + // Note that it's important that we do this _before_ we'd try to + // do the optimizations below on obj normally, since some of those + // optimizations have fallback paths that are slower than the path + // we'd produce here. + + MOZ_ASSERT(*emitted == false); + + MDefinition* inner = tryInnerizeWindow(obj); + if (inner == obj) + return true; + + if (!forceInlineCaches()) { + trackOptimizationAttempt(TrackedStrategy::GetProp_Constant); + if (!getPropTryConstant(emitted, inner, NameToId(name), types) || *emitted) + return *emitted; + + trackOptimizationAttempt(TrackedStrategy::GetProp_StaticName); + if (!getStaticName(&script()->global(), name, emitted) || *emitted) + return *emitted; + + trackOptimizationAttempt(TrackedStrategy::GetProp_CommonGetter); + if (!getPropTryCommonGetter(emitted, inner, name, types) || *emitted) + return *emitted; + } + + // Passing the inner object to GetProperty IC is safe, see the + // needsOuterizedThisObject check in IsCacheableGetPropCallNative. + BarrierKind barrier = PropertyReadNeedsTypeBarrier(analysisContext, constraints(), + inner, name, types); + trackOptimizationAttempt(TrackedStrategy::GetProp_InlineCache); + if (!getPropTryCache(emitted, inner, name, barrier, types) || *emitted) + return *emitted; + + MOZ_ASSERT(*emitted == false); + return true; +} + +bool +IonBuilder::jsop_setprop(PropertyName* name) +{ + MDefinition* value = current->pop(); + MDefinition* obj = convertUnboxedObjects(current->pop()); + + bool emitted = false; + startTrackingOptimizations(); + trackTypeInfo(TrackedTypeSite::Receiver, obj->type(), obj->resultTypeSet()); + trackTypeInfo(TrackedTypeSite::Value, value->type(), value->resultTypeSet()); + + // Always use a call if we are doing the definite properties analysis and + // not actually emitting code, to simplify later analysis. + if (info().isAnalysis() || shouldAbortOnPreliminaryGroups(obj)) { + bool strict = IsStrictSetPC(pc); + MInstruction* ins = MCallSetProperty::New(alloc(), obj, value, name, strict); + current->add(ins); + current->push(value); + return resumeAfter(ins); + } + + if (!forceInlineCaches()) { + // Try to inline a common property setter, or make a call. + trackOptimizationAttempt(TrackedStrategy::SetProp_CommonSetter); + if (!setPropTryCommonSetter(&emitted, obj, name, value) || emitted) + return emitted; + + // Try to emit stores to known binary data blocks + trackOptimizationAttempt(TrackedStrategy::SetProp_TypedObject); + if (!setPropTryTypedObject(&emitted, obj, name, value) || emitted) + return emitted; + } + + TemporaryTypeSet* objTypes = obj->resultTypeSet(); + bool barrier = PropertyWriteNeedsTypeBarrier(alloc(), constraints(), current, &obj, name, &value, + /* canModify = */ true); + + if (!forceInlineCaches()) { + // Try to emit stores to unboxed objects. + trackOptimizationAttempt(TrackedStrategy::SetProp_Unboxed); + if (!setPropTryUnboxed(&emitted, obj, name, value, barrier, objTypes) || emitted) + return emitted; + } + + // Add post barrier if needed. The instructions above manage any post + // barriers they need directly. + if (NeedsPostBarrier(value)) + current->add(MPostWriteBarrier::New(alloc(), obj, value)); + + if (!forceInlineCaches()) { + // Try to emit store from definite slots. + trackOptimizationAttempt(TrackedStrategy::SetProp_DefiniteSlot); + if (!setPropTryDefiniteSlot(&emitted, obj, name, value, barrier, objTypes) || emitted) + return emitted; + + // Try to emit a monomorphic/polymorphic store based on baseline caches. + trackOptimizationAttempt(TrackedStrategy::SetProp_InlineAccess); + if (!setPropTryInlineAccess(&emitted, obj, name, value, barrier, objTypes) || emitted) + return emitted; + } + + // Emit a polymorphic cache. + trackOptimizationAttempt(TrackedStrategy::SetProp_InlineCache); + return setPropTryCache(&emitted, obj, name, value, barrier, objTypes); +} + +bool +IonBuilder::setPropTryCommonSetter(bool* emitted, MDefinition* obj, + PropertyName* name, MDefinition* value) +{ + MOZ_ASSERT(*emitted == false); + + Shape* lastProperty = nullptr; + JSFunction* commonSetter = nullptr; + JSObject* foundProto = nullptr; + bool isOwnProperty; + BaselineInspector::ReceiverVector receivers(alloc()); + BaselineInspector::ObjectGroupVector convertUnboxedGroups(alloc()); + if (!inspector->commonSetPropFunction(pc, &foundProto, &lastProperty, &commonSetter, + &isOwnProperty, + receivers, convertUnboxedGroups)) + { + trackOptimizationOutcome(TrackedOutcome::NoProtoFound); + return true; + } + + TemporaryTypeSet* objTypes = obj->resultTypeSet(); + MDefinition* guard = nullptr; + bool canUseTIForSetter = + testCommonGetterSetter(objTypes, name, /* isGetter = */ false, + foundProto, lastProperty, commonSetter, &guard); + if (!canUseTIForSetter) { + // If type information is bad, we can still optimize the setter if we + // shape guard. + obj = addShapeGuardsForGetterSetter(obj, foundProto, lastProperty, + receivers, convertUnboxedGroups, + isOwnProperty); + if (!obj) + return false; + } + + // Emit common setter. + + // Setters can be called even if the property write needs a type + // barrier, as calling the setter does not actually write any data + // properties. + + // Try emitting dom call. + if (!setPropTryCommonDOMSetter(emitted, obj, value, commonSetter, objTypes)) + return false; + + if (*emitted) { + trackOptimizationOutcome(TrackedOutcome::DOM); + return true; + } + + // Don't call the setter with a primitive value. + if (obj->type() != MIRType::Object) { + MGuardObject* guardObj = MGuardObject::New(alloc(), obj); + current->add(guardObj); + obj = guardObj; + } + + // Dummy up the stack, as in getprop. We are pushing an extra value, so + // ensure there is enough space. + if (!current->ensureHasSlots(3)) + return false; + + current->push(constant(ObjectValue(*commonSetter))); + current->push(obj); + current->push(value); + + // Call the setter. Note that we have to push the original value, not + // the setter's return value. + CallInfo callInfo(alloc(), false); + if (!callInfo.init(current, 1)) + return false; + + // Ensure that we know we are calling a setter in case we inline it. + callInfo.markAsSetter(); + + // Inline the setter if we can. + if (commonSetter->isInterpreted()) { + InliningDecision decision = makeInliningDecision(commonSetter, callInfo); + switch (decision) { + case InliningDecision_Error: + return false; + case InliningDecision_DontInline: + case InliningDecision_WarmUpCountTooLow: + break; + case InliningDecision_Inline: { + InliningStatus status = inlineScriptedCall(callInfo, commonSetter); + if (status == InliningStatus_Inlined) { + *emitted = true; + return true; + } + if (status == InliningStatus_Error) + return false; + } + } + } + + MCall* call = makeCallHelper(commonSetter, callInfo); + if (!call) + return false; + + current->push(value); + if (!resumeAfter(call)) + return false; + + // If the setter could have been inlined, don't track success. The call to + // makeInliningDecision above would have tracked a specific reason why we + // couldn't inline. + if (!commonSetter->isInterpreted()) + trackOptimizationSuccess(); + + *emitted = true; + return true; +} + +bool +IonBuilder::setPropTryCommonDOMSetter(bool* emitted, MDefinition* obj, + MDefinition* value, JSFunction* setter, + TemporaryTypeSet* objTypes) +{ + MOZ_ASSERT(*emitted == false); + + if (!objTypes || !objTypes->isDOMClass(constraints())) + return true; + + if (!testShouldDOMCall(objTypes, setter, JSJitInfo::Setter)) + return true; + + // Emit SetDOMProperty. + MOZ_ASSERT(setter->jitInfo()->type() == JSJitInfo::Setter); + MSetDOMProperty* set = MSetDOMProperty::New(alloc(), setter->jitInfo()->setter, obj, value); + + current->add(set); + current->push(value); + + if (!resumeAfter(set)) + return false; + + *emitted = true; + return true; +} + +bool +IonBuilder::setPropTryTypedObject(bool* emitted, MDefinition* obj, + PropertyName* name, MDefinition* value) +{ + TypedObjectPrediction fieldPrediction; + size_t fieldOffset; + size_t fieldIndex; + if (!typedObjectHasField(obj, name, &fieldOffset, &fieldPrediction, &fieldIndex)) + return true; + + switch (fieldPrediction.kind()) { + case type::Simd: + // FIXME (bug 894104): store into a MIRType::float32x4 etc + return true; + + case type::Reference: + return setPropTryReferencePropOfTypedObject(emitted, obj, fieldOffset, + value, fieldPrediction, name); + + case type::Scalar: + return setPropTryScalarPropOfTypedObject(emitted, obj, fieldOffset, + value, fieldPrediction); + + case type::Struct: + case type::Array: + return true; + } + + MOZ_CRASH("Unknown kind"); +} + +bool +IonBuilder::setPropTryReferencePropOfTypedObject(bool* emitted, + MDefinition* obj, + int32_t fieldOffset, + MDefinition* value, + TypedObjectPrediction fieldPrediction, + PropertyName* name) +{ + ReferenceTypeDescr::Type fieldType = fieldPrediction.referenceType(); + + TypeSet::ObjectKey* globalKey = TypeSet::ObjectKey::get(&script()->global()); + if (globalKey->hasFlags(constraints(), OBJECT_FLAG_TYPED_OBJECT_HAS_DETACHED_BUFFER)) + return true; + + LinearSum byteOffset(alloc()); + if (!byteOffset.add(fieldOffset)) + setForceAbort(); + + if (!storeReferenceTypedObjectValue(obj, byteOffset, fieldType, value, name)) + return true; + + current->push(value); + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +bool +IonBuilder::setPropTryScalarPropOfTypedObject(bool* emitted, + MDefinition* obj, + int32_t fieldOffset, + MDefinition* value, + TypedObjectPrediction fieldPrediction) +{ + // Must always be loading the same scalar type + Scalar::Type fieldType = fieldPrediction.scalarType(); + + // Don't optimize if the typed object's underlying buffer may be detached. + TypeSet::ObjectKey* globalKey = TypeSet::ObjectKey::get(&script()->global()); + if (globalKey->hasFlags(constraints(), OBJECT_FLAG_TYPED_OBJECT_HAS_DETACHED_BUFFER)) + return true; + + LinearSum byteOffset(alloc()); + if (!byteOffset.add(fieldOffset)) + setForceAbort(); + + if (!storeScalarTypedObjectValue(obj, byteOffset, fieldType, value)) + return false; + + current->push(value); + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +bool +IonBuilder::setPropTryDefiniteSlot(bool* emitted, MDefinition* obj, + PropertyName* name, MDefinition* value, + bool barrier, TemporaryTypeSet* objTypes) +{ + MOZ_ASSERT(*emitted == false); + + if (barrier) { + trackOptimizationOutcome(TrackedOutcome::NeedsTypeBarrier); + return true; + } + + uint32_t nfixed; + uint32_t slot = getDefiniteSlot(obj->resultTypeSet(), name, &nfixed); + if (slot == UINT32_MAX) + return true; + + bool writeBarrier = false; + for (size_t i = 0; i < obj->resultTypeSet()->getObjectCount(); i++) { + TypeSet::ObjectKey* key = obj->resultTypeSet()->getObject(i); + if (!key) + continue; + + HeapTypeSetKey property = key->property(NameToId(name)); + if (property.nonWritable(constraints())) { + trackOptimizationOutcome(TrackedOutcome::NonWritableProperty); + return true; + } + writeBarrier |= property.needsBarrier(constraints()); + } + + MInstruction* store; + if (slot < nfixed) { + store = MStoreFixedSlot::New(alloc(), obj, slot, value); + if (writeBarrier) + store->toStoreFixedSlot()->setNeedsBarrier(); + } else { + MInstruction* slots = MSlots::New(alloc(), obj); + current->add(slots); + + store = MStoreSlot::New(alloc(), slots, slot - nfixed, value); + if (writeBarrier) + store->toStoreSlot()->setNeedsBarrier(); + } + + current->add(store); + current->push(value); + + if (!resumeAfter(store)) + return false; + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +MInstruction* +IonBuilder::storeUnboxedProperty(MDefinition* obj, size_t offset, JSValueType unboxedType, + MDefinition* value) +{ + size_t scaledOffsetConstant = offset / UnboxedTypeSize(unboxedType); + MInstruction* scaledOffset = MConstant::New(alloc(), Int32Value(scaledOffsetConstant)); + current->add(scaledOffset); + + return storeUnboxedValue(obj, obj, UnboxedPlainObject::offsetOfData(), + scaledOffset, unboxedType, value); +} + +MInstruction* +IonBuilder::storeUnboxedValue(MDefinition* obj, MDefinition* elements, int32_t elementsOffset, + MDefinition* scaledOffset, JSValueType unboxedType, + MDefinition* value, bool preBarrier /* = true */) +{ + MInstruction* store; + switch (unboxedType) { + case JSVAL_TYPE_BOOLEAN: + store = MStoreUnboxedScalar::New(alloc(), elements, scaledOffset, value, Scalar::Uint8, + MStoreUnboxedScalar::DontTruncateInput, + DoesNotRequireMemoryBarrier, elementsOffset); + break; + + case JSVAL_TYPE_INT32: + store = MStoreUnboxedScalar::New(alloc(), elements, scaledOffset, value, Scalar::Int32, + MStoreUnboxedScalar::DontTruncateInput, + DoesNotRequireMemoryBarrier, elementsOffset); + break; + + case JSVAL_TYPE_DOUBLE: + store = MStoreUnboxedScalar::New(alloc(), elements, scaledOffset, value, Scalar::Float64, + MStoreUnboxedScalar::DontTruncateInput, + DoesNotRequireMemoryBarrier, elementsOffset); + break; + + case JSVAL_TYPE_STRING: + store = MStoreUnboxedString::New(alloc(), elements, scaledOffset, value, + elementsOffset, preBarrier); + break; + + case JSVAL_TYPE_OBJECT: + MOZ_ASSERT(value->type() == MIRType::Object || + value->type() == MIRType::Null || + value->type() == MIRType::Value); + MOZ_ASSERT(!value->mightBeType(MIRType::Undefined), + "MToObjectOrNull slow path is invalid for unboxed objects"); + store = MStoreUnboxedObjectOrNull::New(alloc(), elements, scaledOffset, value, obj, + elementsOffset, preBarrier); + break; + + default: + MOZ_CRASH(); + } + + current->add(store); + return store; +} + +bool +IonBuilder::setPropTryUnboxed(bool* emitted, MDefinition* obj, + PropertyName* name, MDefinition* value, + bool barrier, TemporaryTypeSet* objTypes) +{ + MOZ_ASSERT(*emitted == false); + + if (barrier) { + trackOptimizationOutcome(TrackedOutcome::NeedsTypeBarrier); + return true; + } + + JSValueType unboxedType; + uint32_t offset = getUnboxedOffset(obj->resultTypeSet(), name, &unboxedType); + if (offset == UINT32_MAX) + return true; + + if (obj->type() != MIRType::Object) { + MGuardObject* guard = MGuardObject::New(alloc(), obj); + current->add(guard); + obj = guard; + } + + MInstruction* store = storeUnboxedProperty(obj, offset, unboxedType, value); + + current->push(value); + + if (!resumeAfter(store)) + return false; + + *emitted = true; + return true; +} + +bool +IonBuilder::setPropTryInlineAccess(bool* emitted, MDefinition* obj, + PropertyName* name, MDefinition* value, + bool barrier, TemporaryTypeSet* objTypes) +{ + MOZ_ASSERT(*emitted == false); + + if (barrier) { + trackOptimizationOutcome(TrackedOutcome::NeedsTypeBarrier); + return true; + } + + BaselineInspector::ReceiverVector receivers(alloc()); + BaselineInspector::ObjectGroupVector convertUnboxedGroups(alloc()); + if (!inspector->maybeInfoForPropertyOp(pc, receivers, convertUnboxedGroups)) + return false; + + if (!canInlinePropertyOpShapes(receivers)) + return true; + + obj = convertUnboxedObjects(obj, convertUnboxedGroups); + + if (receivers.length() == 1) { + if (!receivers[0].group) { + // Monomorphic store to a native object. + spew("Inlining monomorphic native SETPROP"); + + obj = addShapeGuard(obj, receivers[0].shape, Bailout_ShapeGuard); + + Shape* shape = receivers[0].shape->searchLinear(NameToId(name)); + MOZ_ASSERT(shape); + + bool needsBarrier = objTypes->propertyNeedsBarrier(constraints(), NameToId(name)); + if (!storeSlot(obj, shape, value, needsBarrier)) + return false; + + trackOptimizationOutcome(TrackedOutcome::Monomorphic); + *emitted = true; + return true; + } + + if (receivers[0].shape) { + // Monomorphic store to an unboxed object expando. + spew("Inlining monomorphic unboxed expando SETPROP"); + + obj = addGroupGuard(obj, receivers[0].group, Bailout_ShapeGuard); + obj = addUnboxedExpandoGuard(obj, /* hasExpando = */ true, Bailout_ShapeGuard); + + MInstruction* expando = MLoadUnboxedExpando::New(alloc(), obj); + current->add(expando); + + expando = addShapeGuard(expando, receivers[0].shape, Bailout_ShapeGuard); + + Shape* shape = receivers[0].shape->searchLinear(NameToId(name)); + MOZ_ASSERT(shape); + + bool needsBarrier = objTypes->propertyNeedsBarrier(constraints(), NameToId(name)); + if (!storeSlot(expando, shape, value, needsBarrier)) + return false; + + trackOptimizationOutcome(TrackedOutcome::Monomorphic); + *emitted = true; + return true; + } + + // Monomorphic store to an unboxed object. + spew("Inlining monomorphic unboxed SETPROP"); + + ObjectGroup* group = receivers[0].group; + if (!objTypes->hasType(TypeSet::ObjectType(group))) + return true; + + obj = addGroupGuard(obj, group, Bailout_ShapeGuard); + + const UnboxedLayout::Property* property = group->unboxedLayout().lookup(name); + storeUnboxedProperty(obj, property->offset, property->type, value); + + current->push(value); + + trackOptimizationOutcome(TrackedOutcome::Monomorphic); + *emitted = true; + return true; + } + + MOZ_ASSERT(receivers.length() > 1); + spew("Inlining polymorphic SETPROP"); + + if (Shape* propShape = PropertyShapesHaveSameSlot(receivers, NameToId(name))) { + obj = addGuardReceiverPolymorphic(obj, receivers); + if (!obj) + return false; + + bool needsBarrier = objTypes->propertyNeedsBarrier(constraints(), NameToId(name)); + if (!storeSlot(obj, propShape, value, needsBarrier)) + return false; + + trackOptimizationOutcome(TrackedOutcome::Polymorphic); + *emitted = true; + return true; + } + + MSetPropertyPolymorphic* ins = MSetPropertyPolymorphic::New(alloc(), obj, value, name); + current->add(ins); + current->push(value); + + for (size_t i = 0; i < receivers.length(); i++) { + Shape* propShape = nullptr; + if (receivers[i].shape) { + propShape = receivers[i].shape->searchLinear(NameToId(name)); + MOZ_ASSERT(propShape); + } + if (!ins->addReceiver(receivers[i], propShape)) + return false; + } + + if (objTypes->propertyNeedsBarrier(constraints(), NameToId(name))) + ins->setNeedsBarrier(); + + if (!resumeAfter(ins)) + return false; + + trackOptimizationOutcome(TrackedOutcome::Polymorphic); + *emitted = true; + return true; +} + +bool +IonBuilder::setPropTryCache(bool* emitted, MDefinition* obj, + PropertyName* name, MDefinition* value, + bool barrier, TemporaryTypeSet* objTypes) +{ + MOZ_ASSERT(*emitted == false); + + bool strict = IsStrictSetPC(pc); + + MConstant* id = constant(StringValue(name)); + MSetPropertyCache* ins = MSetPropertyCache::New(alloc(), obj, id, value, strict, barrier, + /* guardHoles = */ false); + current->add(ins); + current->push(value); + + if (!resumeAfter(ins)) + return false; + + trackOptimizationSuccess(); + *emitted = true; + return true; +} + +bool +IonBuilder::jsop_delprop(PropertyName* name) +{ + MDefinition* obj = current->pop(); + + bool strict = JSOp(*pc) == JSOP_STRICTDELPROP; + MInstruction* ins = MDeleteProperty::New(alloc(), obj, name, strict); + + current->add(ins); + current->push(ins); + + return resumeAfter(ins); +} + +bool +IonBuilder::jsop_delelem() +{ + MDefinition* index = current->pop(); + MDefinition* obj = current->pop(); + + bool strict = JSOp(*pc) == JSOP_STRICTDELELEM; + MDeleteElement* ins = MDeleteElement::New(alloc(), obj, index, strict); + current->add(ins); + current->push(ins); + + return resumeAfter(ins); +} + +bool +IonBuilder::jsop_regexp(RegExpObject* reobj) +{ + MRegExp* regexp = MRegExp::New(alloc(), constraints(), reobj); + current->add(regexp); + current->push(regexp); + + return true; +} + +bool +IonBuilder::jsop_object(JSObject* obj) +{ + if (options.cloneSingletons()) { + MCloneLiteral* clone = MCloneLiteral::New(alloc(), constant(ObjectValue(*obj))); + current->add(clone); + current->push(clone); + return resumeAfter(clone); + } + + compartment->setSingletonsAsValues(); + pushConstant(ObjectValue(*obj)); + return true; +} + +bool +IonBuilder::jsop_lambda(JSFunction* fun) +{ + MOZ_ASSERT(analysis().usesEnvironmentChain()); + MOZ_ASSERT(!fun->isArrow()); + + if (IsAsmJSModule(fun)) + return abort("asm.js module function"); + + MConstant* cst = MConstant::NewConstraintlessObject(alloc(), fun); + current->add(cst); + MLambda* ins = MLambda::New(alloc(), constraints(), current->environmentChain(), cst); + current->add(ins); + current->push(ins); + + return resumeAfter(ins); +} + +bool +IonBuilder::jsop_lambda_arrow(JSFunction* fun) +{ + MOZ_ASSERT(analysis().usesEnvironmentChain()); + MOZ_ASSERT(fun->isArrow()); + MOZ_ASSERT(!fun->isNative()); + + MDefinition* newTargetDef = current->pop(); + MLambdaArrow* ins = MLambdaArrow::New(alloc(), constraints(), current->environmentChain(), + newTargetDef, fun); + current->add(ins); + current->push(ins); + + return resumeAfter(ins); +} + +bool +IonBuilder::jsop_setarg(uint32_t arg) +{ + // To handle this case, we should spill the arguments to the space where + // actual arguments are stored. The tricky part is that if we add a MIR + // to wrap the spilling action, we don't want the spilling to be + // captured by the GETARG and by the resume point, only by + // MGetFrameArgument. + MOZ_ASSERT(analysis_.hasSetArg()); + MDefinition* val = current->peek(-1); + + // If an arguments object is in use, and it aliases formals, then all SETARGs + // must go through the arguments object. + if (info().argsObjAliasesFormals()) { + if (NeedsPostBarrier(val)) + current->add(MPostWriteBarrier::New(alloc(), current->argumentsObject(), val)); + current->add(MSetArgumentsObjectArg::New(alloc(), current->argumentsObject(), + GET_ARGNO(pc), val)); + return true; + } + + // :TODO: if hasArguments() is true, and the script has a JSOP_SETARG, then + // convert all arg accesses to go through the arguments object. (see Bug 957475) + if (info().hasArguments()) + return abort("NYI: arguments & setarg."); + + // Otherwise, if a magic arguments is in use, and it aliases formals, and there exist + // arguments[...] GETELEM expressions in the script, then SetFrameArgument must be used. + // If no arguments[...] GETELEM expressions are in the script, and an argsobj is not + // required, then it means that any aliased argument set can never be observed, and + // the frame does not actually need to be updated with the new arg value. + if (info().argumentsAliasesFormals()) { + // JSOP_SETARG with magic arguments within inline frames is not yet supported. + MOZ_ASSERT(script()->uninlineable() && !isInlineBuilder()); + + MSetFrameArgument* store = MSetFrameArgument::New(alloc(), arg, val); + modifiesFrameArguments_ = true; + current->add(store); + current->setArg(arg); + return true; + } + + // If this assignment is at the start of the function and is coercing + // the original value for the argument which was passed in, loosen + // the type information for that original argument if it is currently + // empty due to originally executing in the interpreter. + if (graph().numBlocks() == 1 && + (val->isBitOr() || val->isBitAnd() || val->isMul() /* for JSOP_POS */)) + { + for (size_t i = 0; i < val->numOperands(); i++) { + MDefinition* op = val->getOperand(i); + if (op->isParameter() && + op->toParameter()->index() == (int32_t)arg && + op->resultTypeSet() && + op->resultTypeSet()->empty()) + { + bool otherUses = false; + for (MUseDefIterator iter(op); iter; iter++) { + MDefinition* def = iter.def(); + if (def == val) + continue; + otherUses = true; + } + if (!otherUses) { + MOZ_ASSERT(op->resultTypeSet() == &argTypes[arg]); + argTypes[arg].addType(TypeSet::UnknownType(), alloc_->lifoAlloc()); + if (val->isMul()) { + val->setResultType(MIRType::Double); + val->toMul()->setSpecialization(MIRType::Double); + } else { + MOZ_ASSERT(val->type() == MIRType::Int32); + } + val->setResultTypeSet(nullptr); + } + } + } + } + + current->setArg(arg); + return true; +} + +bool +IonBuilder::jsop_defvar(uint32_t index) +{ + MOZ_ASSERT(JSOp(*pc) == JSOP_DEFVAR); + + PropertyName* name = script()->getName(index); + + // Bake in attrs. + unsigned attrs = JSPROP_ENUMERATE | JSPROP_PERMANENT; + MOZ_ASSERT(!script()->isForEval()); + + // Pass the EnvironmentChain. + MOZ_ASSERT(analysis().usesEnvironmentChain()); + + // Bake the name pointer into the MDefVar. + MDefVar* defvar = MDefVar::New(alloc(), name, attrs, current->environmentChain()); + current->add(defvar); + + return resumeAfter(defvar); +} + +bool +IonBuilder::jsop_deflexical(uint32_t index) +{ + MOZ_ASSERT(!script()->hasNonSyntacticScope()); + MOZ_ASSERT(JSOp(*pc) == JSOP_DEFLET || JSOp(*pc) == JSOP_DEFCONST); + + PropertyName* name = script()->getName(index); + unsigned attrs = JSPROP_ENUMERATE | JSPROP_PERMANENT; + if (JSOp(*pc) == JSOP_DEFCONST) + attrs |= JSPROP_READONLY; + + MDefLexical* deflex = MDefLexical::New(alloc(), name, attrs); + current->add(deflex); + + return resumeAfter(deflex); +} + +bool +IonBuilder::jsop_deffun(uint32_t index) +{ + MOZ_ASSERT(analysis().usesEnvironmentChain()); + + MDefFun* deffun = MDefFun::New(alloc(), current->pop(), current->environmentChain()); + current->add(deffun); + + return resumeAfter(deffun); +} + +bool +IonBuilder::jsop_throwsetconst() +{ + current->peek(-1)->setImplicitlyUsedUnchecked(); + MInstruction* lexicalError = MThrowRuntimeLexicalError::New(alloc(), JSMSG_BAD_CONST_ASSIGN); + current->add(lexicalError); + return resumeAfter(lexicalError); +} + +bool +IonBuilder::jsop_checklexical() +{ + uint32_t slot = info().localSlot(GET_LOCALNO(pc)); + MDefinition* lexical = addLexicalCheck(current->getSlot(slot)); + if (!lexical) + return false; + current->setSlot(slot, lexical); + return true; +} + +bool +IonBuilder::jsop_checkaliasedlexical(EnvironmentCoordinate ec) +{ + MDefinition* let = addLexicalCheck(getAliasedVar(ec)); + if (!let) + return false; + + jsbytecode* nextPc = pc + JSOP_CHECKALIASEDLEXICAL_LENGTH; + MOZ_ASSERT(JSOp(*nextPc) == JSOP_GETALIASEDVAR || + JSOp(*nextPc) == JSOP_SETALIASEDVAR || + JSOp(*nextPc) == JSOP_THROWSETALIASEDCONST); + MOZ_ASSERT(ec == EnvironmentCoordinate(nextPc)); + + // If we are checking for a load, push the checked let so that the load + // can use it. + if (JSOp(*nextPc) == JSOP_GETALIASEDVAR) + setLexicalCheck(let); + + return true; +} + +bool +IonBuilder::jsop_functionthis() +{ + MOZ_ASSERT(info().funMaybeLazy()); + MOZ_ASSERT(!info().funMaybeLazy()->isArrow()); + + if (script()->strict() || info().funMaybeLazy()->isSelfHostedBuiltin()) { + // No need to wrap primitive |this| in strict mode or self-hosted code. + current->pushSlot(info().thisSlot()); + return true; + } + + if (thisTypes && (thisTypes->getKnownMIRType() == MIRType::Object || + (thisTypes->empty() && baselineFrame_ && baselineFrame_->thisType.isSomeObject()))) + { + // This is safe, because if the entry type of |this| is an object, it + // will necessarily be an object throughout the entire function. OSR + // can introduce a phi, but this phi will be specialized. + current->pushSlot(info().thisSlot()); + return true; + } + + // If we are doing an analysis, we might not yet know the type of |this|. + // Instead of bailing out just push the |this| slot, as this code won't + // actually execute and it does not matter whether |this| is primitive. + if (info().isAnalysis()) { + current->pushSlot(info().thisSlot()); + return true; + } + + // Hard case: |this| may be a primitive we have to wrap. + MDefinition* def = current->getSlot(info().thisSlot()); + + if (def->type() == MIRType::Object) { + current->push(def); + return true; + } + + if (IsNullOrUndefined(def->type())) { + pushConstant(GetThisValue(&script()->global())); + return true; + } + + MComputeThis* thisObj = MComputeThis::New(alloc(), def); + current->add(thisObj); + current->push(thisObj); + + return resumeAfter(thisObj); +} + +bool +IonBuilder::jsop_globalthis() +{ + if (script()->hasNonSyntacticScope()) { + // Ion does not compile global scripts with a non-syntactic scope, but + // we can end up here when we're compiling an arrow function. + return abort("JSOP_GLOBALTHIS in script with non-syntactic scope"); + } + + LexicalEnvironmentObject* globalLexical = &script()->global().lexicalEnvironment(); + pushConstant(globalLexical->thisValue()); + return true; +} + +bool +IonBuilder::jsop_typeof() +{ + MDefinition* input = current->pop(); + MTypeOf* ins = MTypeOf::New(alloc(), input, input->type()); + + ins->cacheInputMaybeCallableOrEmulatesUndefined(constraints()); + + current->add(ins); + current->push(ins); + + return true; +} + +bool +IonBuilder::jsop_toasync() +{ + MDefinition* unwrapped = current->pop(); + MOZ_ASSERT(unwrapped->type() == MIRType::Object); + + MToAsync* ins = MToAsync::New(alloc(), unwrapped); + + current->add(ins); + current->push(ins); + + return resumeAfter(ins); +} + +bool +IonBuilder::jsop_toid() +{ + // No-op if the index is an integer. + if (current->peek(-1)->type() == MIRType::Int32) + return true; + + MDefinition* index = current->pop(); + MToId* ins = MToId::New(alloc(), index); + + current->add(ins); + current->push(ins); + + return resumeAfter(ins); +} + +bool +IonBuilder::jsop_iter(uint8_t flags) +{ + if (flags != JSITER_ENUMERATE) + nonStringIteration_ = true; + + MDefinition* obj = current->pop(); + MInstruction* ins = MIteratorStart::New(alloc(), obj, flags); + + if (!outermostBuilder()->iterators_.append(ins)) + return false; + + current->add(ins); + current->push(ins); + + return resumeAfter(ins); +} + +bool +IonBuilder::jsop_itermore() +{ + MDefinition* iter = current->peek(-1); + MInstruction* ins = MIteratorMore::New(alloc(), iter); + + current->add(ins); + current->push(ins); + + return resumeAfter(ins); +} + +bool +IonBuilder::jsop_isnoiter() +{ + MDefinition* def = current->peek(-1); + MOZ_ASSERT(def->isIteratorMore()); + + MInstruction* ins = MIsNoIter::New(alloc(), def); + current->add(ins); + current->push(ins); + + return true; +} + +bool +IonBuilder::jsop_iterend() +{ + MDefinition* iter = current->pop(); + MInstruction* ins = MIteratorEnd::New(alloc(), iter); + + current->add(ins); + + return resumeAfter(ins); +} + +MDefinition* +IonBuilder::walkEnvironmentChain(unsigned hops) +{ + MDefinition* env = current->getSlot(info().environmentChainSlot()); + + for (unsigned i = 0; i < hops; i++) { + MInstruction* ins = MEnclosingEnvironment::New(alloc(), env); + current->add(ins); + env = ins; + } + + return env; +} + +bool +IonBuilder::hasStaticEnvironmentObject(EnvironmentCoordinate ec, JSObject** pcall) +{ + JSScript* outerScript = EnvironmentCoordinateFunctionScript(script(), pc); + if (!outerScript || !outerScript->treatAsRunOnce()) + return false; + + TypeSet::ObjectKey* funKey = + TypeSet::ObjectKey::get(outerScript->functionNonDelazifying()); + if (funKey->hasFlags(constraints(), OBJECT_FLAG_RUNONCE_INVALIDATED)) + return false; + + // The script this aliased var operation is accessing will run only once, + // so there will be only one call object and the aliased var access can be + // compiled in the same manner as a global access. We still need to find + // the call object though. + + // Look for the call object on the current script's function's env chain. + // If the current script is inner to the outer script and the function has + // singleton type then it should show up here. + + MDefinition* envDef = current->getSlot(info().environmentChainSlot()); + envDef->setImplicitlyUsedUnchecked(); + + JSObject* environment = script()->functionNonDelazifying()->environment(); + while (environment && !environment->is<GlobalObject>()) { + if (environment->is<CallObject>() && + environment->as<CallObject>().callee().nonLazyScript() == outerScript) + { + MOZ_ASSERT(environment->isSingleton()); + *pcall = environment; + return true; + } + environment = environment->enclosingEnvironment(); + } + + // Look for the call object on the current frame, if we are compiling the + // outer script itself. Don't do this if we are at entry to the outer + // script, as the call object we see will not be the real one --- after + // entering the Ion code a different call object will be created. + + if (script() == outerScript && baselineFrame_ && info().osrPc()) { + JSObject* singletonScope = baselineFrame_->singletonEnvChain; + if (singletonScope && + singletonScope->is<CallObject>() && + singletonScope->as<CallObject>().callee().nonLazyScript() == outerScript) + { + MOZ_ASSERT(singletonScope->isSingleton()); + *pcall = singletonScope; + return true; + } + } + + return true; +} + +MDefinition* +IonBuilder::getAliasedVar(EnvironmentCoordinate ec) +{ + MDefinition* obj = walkEnvironmentChain(ec.hops()); + + Shape* shape = EnvironmentCoordinateToEnvironmentShape(script(), pc); + + MInstruction* load; + if (shape->numFixedSlots() <= ec.slot()) { + MInstruction* slots = MSlots::New(alloc(), obj); + current->add(slots); + + load = MLoadSlot::New(alloc(), slots, ec.slot() - shape->numFixedSlots()); + } else { + load = MLoadFixedSlot::New(alloc(), obj, ec.slot()); + } + + current->add(load); + return load; +} + +bool +IonBuilder::jsop_getaliasedvar(EnvironmentCoordinate ec) +{ + JSObject* call = nullptr; + if (hasStaticEnvironmentObject(ec, &call) && call) { + PropertyName* name = EnvironmentCoordinateName(envCoordinateNameCache, script(), pc); + bool emitted = false; + if (!getStaticName(call, name, &emitted, takeLexicalCheck()) || emitted) + return emitted; + } + + // See jsop_checkaliasedlexical. + MDefinition* load = takeLexicalCheck(); + if (!load) + load = getAliasedVar(ec); + current->push(load); + + TemporaryTypeSet* types = bytecodeTypes(pc); + return pushTypeBarrier(load, types, BarrierKind::TypeSet); +} + +bool +IonBuilder::jsop_setaliasedvar(EnvironmentCoordinate ec) +{ + JSObject* call = nullptr; + if (hasStaticEnvironmentObject(ec, &call)) { + uint32_t depth = current->stackDepth() + 1; + if (depth > current->nslots()) { + if (!current->increaseSlots(depth - current->nslots())) + return false; + } + MDefinition* value = current->pop(); + PropertyName* name = EnvironmentCoordinateName(envCoordinateNameCache, script(), pc); + + if (call) { + // Push the object on the stack to match the bound object expected in + // the global and property set cases. + pushConstant(ObjectValue(*call)); + current->push(value); + return setStaticName(call, name); + } + + // The call object has type information we need to respect but we + // couldn't find it. Just do a normal property assign. + MDefinition* obj = walkEnvironmentChain(ec.hops()); + current->push(obj); + current->push(value); + return jsop_setprop(name); + } + + MDefinition* rval = current->peek(-1); + MDefinition* obj = walkEnvironmentChain(ec.hops()); + + Shape* shape = EnvironmentCoordinateToEnvironmentShape(script(), pc); + + if (NeedsPostBarrier(rval)) + current->add(MPostWriteBarrier::New(alloc(), obj, rval)); + + MInstruction* store; + if (shape->numFixedSlots() <= ec.slot()) { + MInstruction* slots = MSlots::New(alloc(), obj); + current->add(slots); + + store = MStoreSlot::NewBarriered(alloc(), slots, ec.slot() - shape->numFixedSlots(), rval); + } else { + store = MStoreFixedSlot::NewBarriered(alloc(), obj, ec.slot(), rval); + } + + current->add(store); + return resumeAfter(store); +} + +bool +IonBuilder::jsop_in() +{ + MDefinition* obj = convertUnboxedObjects(current->pop()); + MDefinition* id = current->pop(); + + bool emitted = false; + + if (!inTryDense(&emitted, obj, id) || emitted) + return emitted; + + if (!inTryFold(&emitted, obj, id) || emitted) + return emitted; + + MIn* ins = MIn::New(alloc(), id, obj); + + current->add(ins); + current->push(ins); + + return resumeAfter(ins); +} + +bool +IonBuilder::inTryDense(bool* emitted, MDefinition* obj, MDefinition* id) +{ + MOZ_ASSERT(!*emitted); + + if (shouldAbortOnPreliminaryGroups(obj)) + return true; + + JSValueType unboxedType = UnboxedArrayElementType(constraints(), obj, id); + if (unboxedType == JSVAL_TYPE_MAGIC) { + if (!ElementAccessIsDenseNative(constraints(), obj, id)) + return true; + } + + if (ElementAccessHasExtraIndexedProperty(this, obj)) + return true; + + *emitted = true; + + bool needsHoleCheck = !ElementAccessIsPacked(constraints(), obj); + + // Ensure id is an integer. + MInstruction* idInt32 = MToInt32::New(alloc(), id); + current->add(idInt32); + id = idInt32; + + // Get the elements vector. + MElements* elements = MElements::New(alloc(), obj, unboxedType != JSVAL_TYPE_MAGIC); + current->add(elements); + + MInstruction* initLength = initializedLength(obj, elements, unboxedType); + + // If there are no holes, speculate the InArray check will not fail. + if (!needsHoleCheck && !failedBoundsCheck_) { + addBoundsCheck(idInt32, initLength); + pushConstant(BooleanValue(true)); + return true; + } + + // Check if id < initLength and elem[id] not a hole. + MInArray* ins = MInArray::New(alloc(), elements, id, initLength, obj, needsHoleCheck, + unboxedType); + + current->add(ins); + current->push(ins); + + return true; +} + +bool +IonBuilder::inTryFold(bool* emitted, MDefinition* obj, MDefinition* id) +{ + // Fold |id in obj| to |false|, if we know the object (or an object on its + // prototype chain) does not have this property. + + MOZ_ASSERT(!*emitted); + + MConstant* idConst = id->maybeConstantValue(); + jsid propId; + if (!idConst || !ValueToIdPure(idConst->toJSValue(), &propId)) + return true; + + if (propId != IdToTypeId(propId)) + return true; + + ResultWithOOM<bool> res = testNotDefinedProperty(obj, propId); + if (res.oom) + return false; + if (!res.value) + return true; + + *emitted = true; + + pushConstant(BooleanValue(false)); + obj->setImplicitlyUsedUnchecked(); + id->setImplicitlyUsedUnchecked(); + return true; +} + +bool +IonBuilder::hasOnProtoChain(TypeSet::ObjectKey* key, JSObject* protoObject, bool* hasOnProto) +{ + MOZ_ASSERT(protoObject); + + while (true) { + if (!key->hasStableClassAndProto(constraints()) || !key->clasp()->isNative()) + return false; + + JSObject* proto = checkNurseryObject(key->proto().toObjectOrNull()); + if (!proto) { + *hasOnProto = false; + return true; + } + + if (proto == protoObject) { + *hasOnProto = true; + return true; + } + + key = TypeSet::ObjectKey::get(proto); + } + + MOZ_CRASH("Unreachable"); +} + +bool +IonBuilder::tryFoldInstanceOf(MDefinition* lhs, JSObject* protoObject) +{ + // Try to fold the js::IsDelegate part of the instanceof operation. + + if (!lhs->mightBeType(MIRType::Object)) { + // If the lhs is a primitive, the result is false. + lhs->setImplicitlyUsedUnchecked(); + pushConstant(BooleanValue(false)); + return true; + } + + TemporaryTypeSet* lhsTypes = lhs->resultTypeSet(); + if (!lhsTypes || lhsTypes->unknownObject()) + return false; + + // We can fold if either all objects have protoObject on their proto chain + // or none have. + bool isFirst = true; + bool knownIsInstance = false; + + for (unsigned i = 0; i < lhsTypes->getObjectCount(); i++) { + TypeSet::ObjectKey* key = lhsTypes->getObject(i); + if (!key) + continue; + + bool isInstance; + if (!hasOnProtoChain(key, protoObject, &isInstance)) + return false; + + if (isFirst) { + knownIsInstance = isInstance; + isFirst = false; + } else if (knownIsInstance != isInstance) { + // Some of the objects have protoObject on their proto chain and + // others don't, so we can't optimize this. + return false; + } + } + + if (knownIsInstance && lhsTypes->getKnownMIRType() != MIRType::Object) { + // The result is true for all objects, but the lhs might be a primitive. + // We can't fold this completely but we can use a much faster IsObject + // test. + MIsObject* isObject = MIsObject::New(alloc(), lhs); + current->add(isObject); + current->push(isObject); + return true; + } + + lhs->setImplicitlyUsedUnchecked(); + pushConstant(BooleanValue(knownIsInstance)); + return true; +} + +bool +IonBuilder::jsop_instanceof() +{ + MDefinition* rhs = current->pop(); + MDefinition* obj = current->pop(); + + // If this is an 'x instanceof function' operation and we can determine the + // exact function and prototype object being tested for, use a typed path. + do { + TemporaryTypeSet* rhsTypes = rhs->resultTypeSet(); + JSObject* rhsObject = rhsTypes ? rhsTypes->maybeSingleton() : nullptr; + if (!rhsObject || !rhsObject->is<JSFunction>() || rhsObject->isBoundFunction()) + break; + + // Refuse to optimize anything whose [[Prototype]] isn't Function.prototype + // since we can't guarantee that it uses the default @@hasInstance method. + if (rhsObject->hasUncacheableProto() || !rhsObject->hasStaticPrototype()) + break; + + Value funProto = script()->global().getPrototype(JSProto_Function); + if (!funProto.isObject() || rhsObject->staticPrototype() != &funProto.toObject()) + break; + + // If the user has supplied their own @@hasInstance method we shouldn't + // clobber it. + JSFunction* fun = &rhsObject->as<JSFunction>(); + const WellKnownSymbols* symbols = &compartment->runtime()->wellKnownSymbols(); + if (!js::FunctionHasDefaultHasInstance(fun, *symbols)) + break; + + // Ensure that we will bail if the @@hasInstance property or [[Prototype]] + // change. + TypeSet::ObjectKey* rhsKey = TypeSet::ObjectKey::get(rhsObject); + if (!rhsKey->hasStableClassAndProto(constraints())) + break; + + if (rhsKey->unknownProperties()) + break; + + HeapTypeSetKey hasInstanceObject = + rhsKey->property(SYMBOL_TO_JSID(symbols->hasInstance)); + if (hasInstanceObject.isOwnProperty(constraints())) + break; + + HeapTypeSetKey protoProperty = + rhsKey->property(NameToId(names().prototype)); + JSObject* protoObject = protoProperty.singleton(constraints()); + if (!protoObject) + break; + + rhs->setImplicitlyUsedUnchecked(); + + if (tryFoldInstanceOf(obj, protoObject)) + return true; + + MInstanceOf* ins = MInstanceOf::New(alloc(), obj, protoObject); + + current->add(ins); + current->push(ins); + + return resumeAfter(ins); + } while (false); + + // Try to inline a fast path based on Baseline ICs. + do { + Shape* shape; + uint32_t slot; + JSObject* protoObject; + if (!inspector->instanceOfData(pc, &shape, &slot, &protoObject)) + break; + + // Shape guard. + rhs = addShapeGuard(rhs, shape, Bailout_ShapeGuard); + + // Guard .prototype == protoObject. + MOZ_ASSERT(shape->numFixedSlots() == 0, "Must be a dynamic slot"); + MSlots* slots = MSlots::New(alloc(), rhs); + current->add(slots); + MLoadSlot* prototype = MLoadSlot::New(alloc(), slots, slot); + current->add(prototype); + MConstant* protoConst = MConstant::NewConstraintlessObject(alloc(), protoObject); + current->add(protoConst); + MGuardObjectIdentity* guard = MGuardObjectIdentity::New(alloc(), prototype, protoConst, + /* bailOnEquality = */ false); + current->add(guard); + + if (tryFoldInstanceOf(obj, protoObject)) + return true; + + MInstanceOf* ins = MInstanceOf::New(alloc(), obj, protoObject); + current->add(ins); + current->push(ins); + return resumeAfter(ins); + } while (false); + + MCallInstanceOf* ins = MCallInstanceOf::New(alloc(), obj, rhs); + + current->add(ins); + current->push(ins); + + return resumeAfter(ins); +} + +bool +IonBuilder::jsop_debugger() +{ + MDebugger* debugger = MDebugger::New(alloc()); + current->add(debugger); + + // The |debugger;| statement will always bail out to baseline if + // cx->compartment()->isDebuggee(). Resume in-place and have baseline + // handle the details. + return resumeAt(debugger, pc); +} + +MInstruction* +IonBuilder::addConvertElementsToDoubles(MDefinition* elements) +{ + MInstruction* convert = MConvertElementsToDoubles::New(alloc(), elements); + current->add(convert); + return convert; +} + +MDefinition* +IonBuilder::addMaybeCopyElementsForWrite(MDefinition* object, bool checkNative) +{ + if (!ElementAccessMightBeCopyOnWrite(constraints(), object)) + return object; + MInstruction* copy = MMaybeCopyElementsForWrite::New(alloc(), object, checkNative); + current->add(copy); + return copy; +} + +MInstruction* +IonBuilder::addBoundsCheck(MDefinition* index, MDefinition* length) +{ + MInstruction* check = MBoundsCheck::New(alloc(), index, length); + current->add(check); + + // If a bounds check failed in the past, don't optimize bounds checks. + if (failedBoundsCheck_) + check->setNotMovable(); + + return check; +} + +MInstruction* +IonBuilder::addShapeGuard(MDefinition* obj, Shape* const shape, BailoutKind bailoutKind) +{ + MGuardShape* guard = MGuardShape::New(alloc(), obj, shape, bailoutKind); + current->add(guard); + + // If a shape guard failed in the past, don't optimize shape guard. + if (failedShapeGuard_) + guard->setNotMovable(); + + return guard; +} + +MInstruction* +IonBuilder::addGroupGuard(MDefinition* obj, ObjectGroup* group, BailoutKind bailoutKind) +{ + MGuardObjectGroup* guard = MGuardObjectGroup::New(alloc(), obj, group, + /* bailOnEquality = */ false, bailoutKind); + current->add(guard); + + // If a shape guard failed in the past, don't optimize group guards. + if (failedShapeGuard_) + guard->setNotMovable(); + + LifoAlloc* lifoAlloc = alloc().lifoAlloc(); + guard->setResultTypeSet(lifoAlloc->new_<TemporaryTypeSet>(lifoAlloc, + TypeSet::ObjectType(group))); + + return guard; +} + +MInstruction* +IonBuilder::addUnboxedExpandoGuard(MDefinition* obj, bool hasExpando, BailoutKind bailoutKind) +{ + MGuardUnboxedExpando* guard = MGuardUnboxedExpando::New(alloc(), obj, hasExpando, bailoutKind); + current->add(guard); + + // If a shape guard failed in the past, don't optimize group guards. + if (failedShapeGuard_) + guard->setNotMovable(); + + return guard; +} + +MInstruction* +IonBuilder::addGuardReceiverPolymorphic(MDefinition* obj, + const BaselineInspector::ReceiverVector& receivers) +{ + if (receivers.length() == 1) { + if (!receivers[0].group) { + // Monomorphic guard on a native object. + return addShapeGuard(obj, receivers[0].shape, Bailout_ShapeGuard); + } + + if (!receivers[0].shape) { + // Guard on an unboxed object that does not have an expando. + obj = addGroupGuard(obj, receivers[0].group, Bailout_ShapeGuard); + return addUnboxedExpandoGuard(obj, /* hasExpando = */ false, Bailout_ShapeGuard); + } + + // Monomorphic receiver guards are not yet supported when the receiver + // is an unboxed object with an expando. + } + + MGuardReceiverPolymorphic* guard = MGuardReceiverPolymorphic::New(alloc(), obj); + current->add(guard); + + if (failedShapeGuard_) + guard->setNotMovable(); + + for (size_t i = 0; i < receivers.length(); i++) { + if (!guard->addReceiver(receivers[i])) + return nullptr; + } + + return guard; +} + +MInstruction* +IonBuilder::addSharedTypedArrayGuard(MDefinition* obj) +{ + MGuardSharedTypedArray* guard = MGuardSharedTypedArray::New(alloc(), obj); + current->add(guard); + return guard; +} + +TemporaryTypeSet* +IonBuilder::bytecodeTypes(jsbytecode* pc) +{ + return TypeScript::BytecodeTypes(script(), pc, bytecodeTypeMap, &typeArrayHint, typeArray); +} + +TypedObjectPrediction +IonBuilder::typedObjectPrediction(MDefinition* typedObj) +{ + // Extract TypedObjectPrediction directly if we can + if (typedObj->isNewDerivedTypedObject()) { + return typedObj->toNewDerivedTypedObject()->prediction(); + } + + TemporaryTypeSet* types = typedObj->resultTypeSet(); + return typedObjectPrediction(types); +} + +TypedObjectPrediction +IonBuilder::typedObjectPrediction(TemporaryTypeSet* types) +{ + // Type set must be known to be an object. + if (!types || types->getKnownMIRType() != MIRType::Object) + return TypedObjectPrediction(); + + // And only known objects. + if (types->unknownObject()) + return TypedObjectPrediction(); + + TypedObjectPrediction out; + for (uint32_t i = 0; i < types->getObjectCount(); i++) { + ObjectGroup* group = types->getGroup(i); + if (!group || !TypeSet::ObjectKey::get(group)->hasStableClassAndProto(constraints())) + return TypedObjectPrediction(); + + if (!IsTypedObjectClass(group->clasp())) + return TypedObjectPrediction(); + + out.addDescr(group->typeDescr()); + } + + return out; +} + +MDefinition* +IonBuilder::loadTypedObjectType(MDefinition* typedObj) +{ + // Shortcircuit derived type objects, meaning the intermediate + // objects created to represent `a.b` in an expression like + // `a.b.c`. In that case, the type object can be simply pulled + // from the operands of that instruction. + if (typedObj->isNewDerivedTypedObject()) + return typedObj->toNewDerivedTypedObject()->type(); + + MInstruction* descr = MTypedObjectDescr::New(alloc(), typedObj); + current->add(descr); + + return descr; +} + +// Given a typed object `typedObj` and an offset `offset` into that +// object's data, returns another typed object and adusted offset +// where the data can be found. Often, these returned values are the +// same as the inputs, but in cases where intermediate derived type +// objects have been created, the return values will remove +// intermediate layers (often rendering those derived type objects +// into dead code). +void +IonBuilder::loadTypedObjectData(MDefinition* typedObj, + MDefinition** owner, + LinearSum* ownerOffset) +{ + MOZ_ASSERT(typedObj->type() == MIRType::Object); + + // Shortcircuit derived type objects, meaning the intermediate + // objects created to represent `a.b` in an expression like + // `a.b.c`. In that case, the owned and a base offset can be + // pulled from the operands of the instruction and combined with + // `offset`. + if (typedObj->isNewDerivedTypedObject()) { + MNewDerivedTypedObject* ins = typedObj->toNewDerivedTypedObject(); + + SimpleLinearSum base = ExtractLinearSum(ins->offset()); + if (!ownerOffset->add(base)) + setForceAbort(); + + *owner = ins->owner(); + return; + } + + *owner = typedObj; +} + +// Takes as input a typed object, an offset into that typed object's +// memory, and the type repr of the data found at that offset. Returns +// the elements pointer and a scaled offset. The scaled offset is +// expressed in units of `unit`; when working with typed array MIR, +// this is typically the alignment. +void +IonBuilder::loadTypedObjectElements(MDefinition* typedObj, + const LinearSum& baseByteOffset, + uint32_t scale, + MDefinition** ownerElements, + MDefinition** ownerScaledOffset, + int32_t* ownerByteAdjustment) +{ + MDefinition* owner; + LinearSum ownerByteOffset(alloc()); + loadTypedObjectData(typedObj, &owner, &ownerByteOffset); + + if (!ownerByteOffset.add(baseByteOffset)) + setForceAbort(); + + TemporaryTypeSet* ownerTypes = owner->resultTypeSet(); + const Class* clasp = ownerTypes ? ownerTypes->getKnownClass(constraints()) : nullptr; + if (clasp && IsInlineTypedObjectClass(clasp)) { + // Perform the load directly from the owner pointer. + if (!ownerByteOffset.add(InlineTypedObject::offsetOfDataStart())) + setForceAbort(); + *ownerElements = owner; + } else { + bool definitelyOutline = clasp && IsOutlineTypedObjectClass(clasp); + *ownerElements = MTypedObjectElements::New(alloc(), owner, definitelyOutline); + current->add((*ownerElements)->toInstruction()); + } + + // Extract the constant adjustment from the byte offset. + *ownerByteAdjustment = ownerByteOffset.constant(); + int32_t negativeAdjustment; + if (!SafeSub(0, *ownerByteAdjustment, &negativeAdjustment)) + setForceAbort(); + if (!ownerByteOffset.add(negativeAdjustment)) + setForceAbort(); + + // Scale the byte offset if required by the MIR node which will access the + // typed object. In principle we should always be able to cleanly divide + // the terms in this lienar sum due to alignment restrictions, but due to + // limitations of ExtractLinearSum when applied to the terms in derived + // typed objects this isn't always be possible. In these cases, fall back + // on an explicit division operation. + if (ownerByteOffset.divide(scale)) { + *ownerScaledOffset = ConvertLinearSum(alloc(), current, ownerByteOffset); + } else { + MDefinition* unscaledOffset = ConvertLinearSum(alloc(), current, ownerByteOffset); + *ownerScaledOffset = MDiv::New(alloc(), unscaledOffset, constantInt(scale), + MIRType::Int32, /* unsigned = */ false); + current->add((*ownerScaledOffset)->toInstruction()); + } +} + +// Looks up the offset/type-repr-set of the field `id`, given the type +// set `objTypes` of the field owner. If a field is found, returns true +// and sets *fieldOffset, *fieldPrediction, and *fieldIndex. Returns false +// otherwise. Infallible. +bool +IonBuilder::typedObjectHasField(MDefinition* typedObj, + PropertyName* name, + size_t* fieldOffset, + TypedObjectPrediction* fieldPrediction, + size_t* fieldIndex) +{ + TypedObjectPrediction objPrediction = typedObjectPrediction(typedObj); + if (objPrediction.isUseless()) { + trackOptimizationOutcome(TrackedOutcome::AccessNotTypedObject); + return false; + } + + // Must be accessing a struct. + if (objPrediction.kind() != type::Struct) { + trackOptimizationOutcome(TrackedOutcome::NotStruct); + return false; + } + + // Determine the type/offset of the field `name`, if any. + if (!objPrediction.hasFieldNamed(NameToId(name), fieldOffset, + fieldPrediction, fieldIndex)) + { + trackOptimizationOutcome(TrackedOutcome::StructNoField); + return false; + } + + return true; +} + +MDefinition* +IonBuilder::typeObjectForElementFromArrayStructType(MDefinition* typeObj) +{ + MInstruction* elemType = MLoadFixedSlot::New(alloc(), typeObj, JS_DESCR_SLOT_ARRAY_ELEM_TYPE); + current->add(elemType); + + MInstruction* unboxElemType = MUnbox::New(alloc(), elemType, MIRType::Object, MUnbox::Infallible); + current->add(unboxElemType); + + return unboxElemType; +} + +MDefinition* +IonBuilder::typeObjectForFieldFromStructType(MDefinition* typeObj, + size_t fieldIndex) +{ + // Load list of field type objects. + + MInstruction* fieldTypes = MLoadFixedSlot::New(alloc(), typeObj, JS_DESCR_SLOT_STRUCT_FIELD_TYPES); + current->add(fieldTypes); + + MInstruction* unboxFieldTypes = MUnbox::New(alloc(), fieldTypes, MIRType::Object, MUnbox::Infallible); + current->add(unboxFieldTypes); + + // Index into list with index of field. + + MInstruction* fieldTypesElements = MElements::New(alloc(), unboxFieldTypes); + current->add(fieldTypesElements); + + MConstant* fieldIndexDef = constantInt(fieldIndex); + + MInstruction* fieldType = MLoadElement::New(alloc(), fieldTypesElements, fieldIndexDef, false, false); + current->add(fieldType); + + MInstruction* unboxFieldType = MUnbox::New(alloc(), fieldType, MIRType::Object, MUnbox::Infallible); + current->add(unboxFieldType); + + return unboxFieldType; +} + +bool +IonBuilder::storeScalarTypedObjectValue(MDefinition* typedObj, + const LinearSum& byteOffset, + ScalarTypeDescr::Type type, + MDefinition* value) +{ + // Find location within the owner object. + MDefinition* elements; + MDefinition* scaledOffset; + int32_t adjustment; + uint32_t alignment = ScalarTypeDescr::alignment(type); + loadTypedObjectElements(typedObj, byteOffset, alignment, &elements, &scaledOffset, &adjustment); + + // Clamp value to [0, 255] when type is Uint8Clamped + MDefinition* toWrite = value; + if (type == Scalar::Uint8Clamped) { + toWrite = MClampToUint8::New(alloc(), value); + current->add(toWrite->toInstruction()); + } + + MStoreUnboxedScalar* store = + MStoreUnboxedScalar::New(alloc(), elements, scaledOffset, toWrite, + type, MStoreUnboxedScalar::TruncateInput, + DoesNotRequireMemoryBarrier, adjustment); + current->add(store); + + return true; +} + +bool +IonBuilder::storeReferenceTypedObjectValue(MDefinition* typedObj, + const LinearSum& byteOffset, + ReferenceTypeDescr::Type type, + MDefinition* value, + PropertyName* name) +{ + // Make sure we aren't adding new type information for writes of object and value + // references. + if (type != ReferenceTypeDescr::TYPE_STRING) { + MOZ_ASSERT(type == ReferenceTypeDescr::TYPE_ANY || + type == ReferenceTypeDescr::TYPE_OBJECT); + MIRType implicitType = + (type == ReferenceTypeDescr::TYPE_ANY) ? MIRType::Undefined : MIRType::Null; + + if (PropertyWriteNeedsTypeBarrier(alloc(), constraints(), current, &typedObj, name, &value, + /* canModify = */ true, implicitType)) + { + trackOptimizationOutcome(TrackedOutcome::NeedsTypeBarrier); + return false; + } + } + + // Find location within the owner object. + MDefinition* elements; + MDefinition* scaledOffset; + int32_t adjustment; + uint32_t alignment = ReferenceTypeDescr::alignment(type); + loadTypedObjectElements(typedObj, byteOffset, alignment, &elements, &scaledOffset, &adjustment); + + MInstruction* store = nullptr; // initialize to silence GCC warning + switch (type) { + case ReferenceTypeDescr::TYPE_ANY: + if (NeedsPostBarrier(value)) + current->add(MPostWriteBarrier::New(alloc(), typedObj, value)); + store = MStoreElement::New(alloc(), elements, scaledOffset, value, false, adjustment); + store->toStoreElement()->setNeedsBarrier(); + break; + case ReferenceTypeDescr::TYPE_OBJECT: + // Note: We cannot necessarily tell at this point whether a post + // barrier is needed, because the type policy may insert ToObjectOrNull + // instructions later, and those may require a post barrier. Therefore, + // defer the insertion of post barriers to the type policy. + store = MStoreUnboxedObjectOrNull::New(alloc(), elements, scaledOffset, value, typedObj, adjustment); + break; + case ReferenceTypeDescr::TYPE_STRING: + // Strings are not nursery allocated, so these writes do not need post + // barriers. + store = MStoreUnboxedString::New(alloc(), elements, scaledOffset, value, adjustment); + break; + } + + current->add(store); + return true; +} + +JSObject* +IonBuilder::checkNurseryObject(JSObject* obj) +{ + // If we try to use any nursery pointers during compilation, make sure that + // the main thread will cancel this compilation before performing a minor + // GC. All constants used during compilation should either go through this + // function or should come from a type set (which has a similar barrier). + if (obj && IsInsideNursery(obj)) { + compartment->runtime()->setMinorGCShouldCancelIonCompilations(); + IonBuilder* builder = this; + while (builder) { + builder->setNotSafeForMinorGC(); + builder = builder->callerBuilder_; + } + } + + return obj; +} + +MConstant* +IonBuilder::constant(const Value& v) +{ + MOZ_ASSERT(!v.isString() || v.toString()->isAtom(), + "Handle non-atomized strings outside IonBuilder."); + + if (v.isObject()) + checkNurseryObject(&v.toObject()); + + MConstant* c = MConstant::New(alloc(), v, constraints()); + current->add(c); + return c; +} + +MConstant* +IonBuilder::constantInt(int32_t i) +{ + return constant(Int32Value(i)); +} + +MInstruction* +IonBuilder::initializedLength(MDefinition* obj, MDefinition* elements, JSValueType unboxedType) +{ + MInstruction* res; + if (unboxedType != JSVAL_TYPE_MAGIC) + res = MUnboxedArrayInitializedLength::New(alloc(), obj); + else + res = MInitializedLength::New(alloc(), elements); + current->add(res); + return res; +} + +MInstruction* +IonBuilder::setInitializedLength(MDefinition* obj, JSValueType unboxedType, size_t count) +{ + MOZ_ASSERT(count); + + MInstruction* res; + if (unboxedType != JSVAL_TYPE_MAGIC) { + res = MSetUnboxedArrayInitializedLength::New(alloc(), obj, constant(Int32Value(count))); + } else { + // MSetInitializedLength takes the index of the last element, rather + // than the count itself. + MInstruction* elements = MElements::New(alloc(), obj, /* unboxed = */ false); + current->add(elements); + res = MSetInitializedLength::New(alloc(), elements, constant(Int32Value(count - 1))); + } + current->add(res); + return res; +} + +MDefinition* +IonBuilder::getCallee() +{ + if (inliningDepth_ == 0) { + MInstruction* callee = MCallee::New(alloc()); + current->add(callee); + return callee; + } + + return inlineCallInfo_->fun(); +} + +MDefinition* +IonBuilder::addLexicalCheck(MDefinition* input) +{ + MOZ_ASSERT(JSOp(*pc) == JSOP_CHECKLEXICAL || + JSOp(*pc) == JSOP_CHECKALIASEDLEXICAL || + JSOp(*pc) == JSOP_GETIMPORT); + + MInstruction* lexicalCheck; + + // If we're guaranteed to not be JS_UNINITIALIZED_LEXICAL, no need to check. + if (input->type() == MIRType::MagicUninitializedLexical) { + // Mark the input as implicitly used so the JS_UNINITIALIZED_LEXICAL + // magic value will be preserved on bailout. + input->setImplicitlyUsedUnchecked(); + lexicalCheck = MThrowRuntimeLexicalError::New(alloc(), JSMSG_UNINITIALIZED_LEXICAL); + current->add(lexicalCheck); + if (!resumeAfter(lexicalCheck)) + return nullptr; + return constant(UndefinedValue()); + } + + if (input->type() == MIRType::Value) { + lexicalCheck = MLexicalCheck::New(alloc(), input); + current->add(lexicalCheck); + if (failedLexicalCheck_) + lexicalCheck->setNotMovableUnchecked(); + return lexicalCheck; + } + + return input; +} + +MDefinition* +IonBuilder::convertToBoolean(MDefinition* input) +{ + // Convert to bool with the '!!' idiom + MNot* resultInverted = MNot::New(alloc(), input, constraints()); + current->add(resultInverted); + MNot* result = MNot::New(alloc(), resultInverted, constraints()); + current->add(result); + + return result; +} + +void +IonBuilder::trace(JSTracer* trc) +{ + if (!compartment->runtime()->runtimeMatches(trc->runtime())) + return; + + MOZ_ASSERT(rootList_); + rootList_->trace(trc); +} |