/* -*- Mode: C++; tab-width: 8; indent-tabs-mode: nil; c-basic-offset: 4 -*- * vim: set ts=8 sts=4 et sw=4 tw=99: * This Source Code Form is subject to the terms of the Mozilla Public * License, v. 2.0. If a copy of the MPL was not distributed with this * file, You can obtain one at http://mozilla.org/MPL/2.0/. */ #include "jit/IonBuilder.h" #include "mozilla/DebugOnly.h" #include "mozilla/SizePrintfMacros.h" #include "builtin/Eval.h" #include "builtin/TypedObject.h" #include "frontend/SourceNotes.h" #include "jit/BaselineFrame.h" #include "jit/BaselineInspector.h" #include "jit/Ion.h" #include "jit/IonOptimizationLevels.h" #include "jit/JitSpewer.h" #include "jit/Lowering.h" #include "jit/MIRGraph.h" #include "vm/ArgumentsObject.h" #include "vm/Opcodes.h" #include "vm/RegExpStatics.h" #include "vm/TraceLogging.h" #include "jsopcodeinlines.h" #include "jsscriptinlines.h" #include "jit/CompileInfo-inl.h" #include "jit/shared/Lowering-shared-inl.h" #include "vm/EnvironmentObject-inl.h" #include "vm/NativeObject-inl.h" #include "vm/ObjectGroup-inl.h" using namespace js; using namespace js::jit; using mozilla::AssertedCast; using mozilla::DebugOnly; using mozilla::Maybe; using JS::TrackedStrategy; using JS::TrackedOutcome; using JS::TrackedTypeSite; class jit::BaselineFrameInspector { public: TypeSet::Type thisType; JSObject* singletonEnvChain; Vector argTypes; Vector varTypes; explicit BaselineFrameInspector(TempAllocator* temp) : thisType(TypeSet::UndefinedType()), singletonEnvChain(nullptr), argTypes(*temp), varTypes(*temp) {} }; BaselineFrameInspector* jit::NewBaselineFrameInspector(TempAllocator* temp, BaselineFrame* frame, CompileInfo* info) { MOZ_ASSERT(frame); BaselineFrameInspector* inspector = temp->lifoAlloc()->new_(temp); if (!inspector) return nullptr; // Note: copying the actual values into a temporary structure for use // during compilation could capture nursery pointers, so the values' types // are recorded instead. if (frame->isFunctionFrame()) inspector->thisType = TypeSet::GetMaybeUntrackedValueType(frame->thisArgument()); if (frame->environmentChain()->isSingleton()) inspector->singletonEnvChain = frame->environmentChain(); JSScript* script = frame->script(); if (script->functionNonDelazifying()) { if (!inspector->argTypes.reserve(frame->numFormalArgs())) return nullptr; for (size_t i = 0; i < frame->numFormalArgs(); i++) { if (script->formalIsAliased(i)) { inspector->argTypes.infallibleAppend(TypeSet::UndefinedType()); } else if (!script->argsObjAliasesFormals()) { TypeSet::Type type = TypeSet::GetMaybeUntrackedValueType(frame->unaliasedFormal(i)); inspector->argTypes.infallibleAppend(type); } else if (frame->hasArgsObj()) { TypeSet::Type type = TypeSet::GetMaybeUntrackedValueType(frame->argsObj().arg(i)); inspector->argTypes.infallibleAppend(type); } else { inspector->argTypes.infallibleAppend(TypeSet::UndefinedType()); } } } if (!inspector->varTypes.reserve(frame->numValueSlots())) return nullptr; for (size_t i = 0; i < frame->numValueSlots(); i++) { TypeSet::Type type = TypeSet::GetMaybeUntrackedValueType(*frame->valueSlot(i)); inspector->varTypes.infallibleAppend(type); } return inspector; } IonBuilder::IonBuilder(JSContext* analysisContext, CompileCompartment* comp, const JitCompileOptions& options, TempAllocator* temp, MIRGraph* graph, CompilerConstraintList* constraints, BaselineInspector* inspector, CompileInfo* info, const OptimizationInfo* optimizationInfo, BaselineFrameInspector* baselineFrame, size_t inliningDepth, uint32_t loopDepth) : MIRGenerator(comp, options, temp, graph, info, optimizationInfo), backgroundCodegen_(nullptr), actionableAbortScript_(nullptr), actionableAbortPc_(nullptr), actionableAbortMessage_(nullptr), rootList_(nullptr), analysisContext(analysisContext), baselineFrame_(baselineFrame), constraints_(constraints), analysis_(*temp, info->script()), thisTypes(nullptr), argTypes(nullptr), typeArray(nullptr), typeArrayHint(0), bytecodeTypeMap(nullptr), loopDepth_(loopDepth), trackedOptimizationSites_(*temp), lexicalCheck_(nullptr), callerResumePoint_(nullptr), callerBuilder_(nullptr), cfgStack_(*temp), loops_(*temp), switches_(*temp), labels_(*temp), iterators_(*temp), loopHeaders_(*temp), inspector(inspector), inliningDepth_(inliningDepth), inlinedBytecodeLength_(0), numLoopRestarts_(0), failedBoundsCheck_(info->script()->failedBoundsCheck()), failedShapeGuard_(info->script()->failedShapeGuard()), failedLexicalCheck_(info->script()->failedLexicalCheck()), nonStringIteration_(false), lazyArguments_(nullptr), inlineCallInfo_(nullptr), maybeFallbackFunctionGetter_(nullptr) { script_ = info->script(); scriptHasIonScript_ = script_->hasIonScript(); pc = info->startPC(); abortReason_ = AbortReason_Disable; MOZ_ASSERT(script()->hasBaselineScript() == (info->analysisMode() != Analysis_ArgumentsUsage)); MOZ_ASSERT(!!analysisContext == (info->analysisMode() == Analysis_DefiniteProperties)); MOZ_ASSERT(script_->nTypeSets() < UINT16_MAX); if (!info->isAnalysis()) script()->baselineScript()->setIonCompiledOrInlined(); } void IonBuilder::clearForBackEnd() { MOZ_ASSERT(!analysisContext); baselineFrame_ = nullptr; // The caches below allocate data from the malloc heap. Release this before // later phases of compilation to avoid leaks, as the top level IonBuilder // is not explicitly destroyed. Note that builders for inner scripts are // constructed on the stack and will release this memory on destruction. gsn.purge(); envCoordinateNameCache.purge(); } bool IonBuilder::abort(const char* message, ...) { // Don't call PCToLineNumber in release builds. #ifdef JS_JITSPEW va_list ap; va_start(ap, message); abortFmt(message, ap); va_end(ap); # ifdef DEBUG JitSpew(JitSpew_IonAbort, "aborted @ %s:%d", script()->filename(), PCToLineNumber(script(), pc)); # else JitSpew(JitSpew_IonAbort, "aborted @ %s", script()->filename()); # endif #endif trackActionableAbort(message); return false; } IonBuilder* IonBuilder::outermostBuilder() { IonBuilder* builder = this; while (builder->callerBuilder_) builder = builder->callerBuilder_; return builder; } void IonBuilder::trackActionableAbort(const char* message) { if (!isOptimizationTrackingEnabled()) return; IonBuilder* topBuilder = outermostBuilder(); if (topBuilder->hadActionableAbort()) return; topBuilder->actionableAbortScript_ = script(); topBuilder->actionableAbortPc_ = pc; topBuilder->actionableAbortMessage_ = message; } void IonBuilder::spew(const char* message) { // Don't call PCToLineNumber in release builds. #ifdef DEBUG JitSpew(JitSpew_IonMIR, "%s @ %s:%d", message, script()->filename(), PCToLineNumber(script(), pc)); #endif } static inline int32_t GetJumpOffset(jsbytecode* pc) { MOZ_ASSERT(CodeSpec[JSOp(*pc)].type() == JOF_JUMP); return GET_JUMP_OFFSET(pc); } IonBuilder::CFGState IonBuilder::CFGState::If(jsbytecode* join, MTest* test) { CFGState state; state.state = IF_TRUE; state.stopAt = join; state.branch.ifFalse = test->ifFalse(); state.branch.test = test; return state; } IonBuilder::CFGState IonBuilder::CFGState::IfElse(jsbytecode* trueEnd, jsbytecode* falseEnd, MTest* test) { MBasicBlock* ifFalse = test->ifFalse(); CFGState state; // If the end of the false path is the same as the start of the // false path, then the "else" block is empty and we can devolve // this to the IF_TRUE case. We handle this here because there is // still an extra GOTO on the true path and we want stopAt to point // there, whereas the IF_TRUE case does not have the GOTO. state.state = (falseEnd == ifFalse->pc()) ? IF_TRUE_EMPTY_ELSE : IF_ELSE_TRUE; state.stopAt = trueEnd; state.branch.falseEnd = falseEnd; state.branch.ifFalse = ifFalse; state.branch.test = test; return state; } IonBuilder::CFGState IonBuilder::CFGState::AndOr(jsbytecode* join, MBasicBlock* lhs) { CFGState state; state.state = AND_OR; state.stopAt = join; state.branch.ifFalse = lhs; state.branch.test = nullptr; return state; } IonBuilder::CFGState IonBuilder::CFGState::TableSwitch(jsbytecode* exitpc, MTableSwitch* ins) { CFGState state; state.state = TABLE_SWITCH; state.stopAt = exitpc; state.tableswitch.exitpc = exitpc; state.tableswitch.breaks = nullptr; state.tableswitch.ins = ins; state.tableswitch.currentBlock = 0; return state; } JSFunction* IonBuilder::getSingleCallTarget(TemporaryTypeSet* calleeTypes) { if (!calleeTypes) return nullptr; JSObject* obj = calleeTypes->maybeSingleton(); if (!obj || !obj->is()) return nullptr; return &obj->as(); } bool IonBuilder::getPolyCallTargets(TemporaryTypeSet* calleeTypes, bool constructing, ObjectVector& targets, uint32_t maxTargets) { MOZ_ASSERT(targets.empty()); if (!calleeTypes) return true; if (calleeTypes->baseFlags() != 0) return true; unsigned objCount = calleeTypes->getObjectCount(); if (objCount == 0 || objCount > maxTargets) return true; if (!targets.reserve(objCount)) return false; for (unsigned i = 0; i < objCount; i++) { JSObject* obj = calleeTypes->getSingleton(i); if (obj) { MOZ_ASSERT(obj->isSingleton()); } else { ObjectGroup* group = calleeTypes->getGroup(i); if (!group) continue; obj = group->maybeInterpretedFunction(); if (!obj) { targets.clear(); return true; } MOZ_ASSERT(!obj->isSingleton()); } // Don't optimize if the callee is not callable or constructable per // the manner it is being invoked, so that CallKnown does not have to // handle these cases (they will always throw). if (constructing ? !obj->isConstructor() : !obj->isCallable()) { targets.clear(); return true; } targets.infallibleAppend(obj); } return true; } IonBuilder::InliningDecision IonBuilder::DontInline(JSScript* targetScript, const char* reason) { if (targetScript) { JitSpew(JitSpew_Inlining, "Cannot inline %s:%" PRIuSIZE ": %s", targetScript->filename(), targetScript->lineno(), reason); } else { JitSpew(JitSpew_Inlining, "Cannot inline: %s", reason); } return InliningDecision_DontInline; } /* * |hasCommonInliningPath| determines whether the current inlining path has been * seen before based on the sequence of scripts in the chain of |IonBuilder|s. * * An inlining path for a function |f| is the sequence of functions whose * inlinings precede |f| up to any previous occurrences of |f|. * So, if we have the chain of inlinings * * f1 -> f2 -> f -> f3 -> f4 -> f5 -> f * -------- -------------- * * the inlining paths for |f| are [f2, f1] and [f5, f4, f3]. * When attempting to inline |f|, we find all existing inlining paths for |f| * and check whether they share a common prefix with the path created were |f| * inlined. * * For example, given mutually recursive functions |f| and |g|, a possible * inlining is * * +---- Inlining stopped here... * | * v * a -> f -> g -> f \ -> g -> f -> g -> ... * * where the vertical bar denotes the termination of inlining. * Inlining is terminated because we have already observed the inlining path * [f] when inlining function |g|. Note that this will inline recursive * functions such as |fib| only one level, as |fib| has a zero length inlining * path which trivially prefixes all inlining paths. * */ bool IonBuilder::hasCommonInliningPath(const JSScript* scriptToInline) { // Find all previous inlinings of the |scriptToInline| and check for common // inlining paths with the top of the inlining stack. for (IonBuilder* it = this->callerBuilder_; it; it = it->callerBuilder_) { if (it->script() != scriptToInline) continue; // This only needs to check the top of each stack for a match, // as a match of length one ensures a common prefix. IonBuilder* path = it->callerBuilder_; if (!path || this->script() == path->script()) return true; } return false; } IonBuilder::InliningDecision IonBuilder::canInlineTarget(JSFunction* target, CallInfo& callInfo) { if (!optimizationInfo().inlineInterpreted()) { trackOptimizationOutcome(TrackedOutcome::CantInlineGeneric); return InliningDecision_DontInline; } if (TraceLogTextIdEnabled(TraceLogger_InlinedScripts)) { return DontInline(nullptr, "Tracelogging of inlined scripts is enabled" "but Tracelogger cannot do that yet."); } if (!target->isInterpreted()) { trackOptimizationOutcome(TrackedOutcome::CantInlineNotInterpreted); return DontInline(nullptr, "Non-interpreted target"); } if (info().analysisMode() != Analysis_DefiniteProperties) { // If |this| or an argument has an empty resultTypeSet, don't bother // inlining, as the call is currently unreachable due to incomplete type // information. This does not apply to the definite properties analysis, // in that case we want to inline anyway. if (callInfo.thisArg()->emptyResultTypeSet()) { trackOptimizationOutcome(TrackedOutcome::CantInlineUnreachable); return DontInline(nullptr, "Empty TypeSet for |this|"); } for (size_t i = 0; i < callInfo.argc(); i++) { if (callInfo.getArg(i)->emptyResultTypeSet()) { trackOptimizationOutcome(TrackedOutcome::CantInlineUnreachable); return DontInline(nullptr, "Empty TypeSet for argument"); } } } // Allow constructing lazy scripts when performing the definite properties // analysis, as baseline has not been used to warm the caller up yet. if (target->isInterpreted() && info().analysisMode() == Analysis_DefiniteProperties) { RootedFunction fun(analysisContext, target); RootedScript script(analysisContext, JSFunction::getOrCreateScript(analysisContext, fun)); if (!script) return InliningDecision_Error; if (!script->hasBaselineScript() && script->canBaselineCompile()) { MethodStatus status = BaselineCompile(analysisContext, script); if (status == Method_Error) return InliningDecision_Error; if (status != Method_Compiled) { trackOptimizationOutcome(TrackedOutcome::CantInlineNoBaseline); return InliningDecision_DontInline; } } } if (!target->hasScript()) { trackOptimizationOutcome(TrackedOutcome::CantInlineLazy); return DontInline(nullptr, "Lazy script"); } JSScript* inlineScript = target->nonLazyScript(); if (callInfo.constructing() && !target->isConstructor()) { trackOptimizationOutcome(TrackedOutcome::CantInlineNotConstructor); return DontInline(inlineScript, "Callee is not a constructor"); } if (!callInfo.constructing() && target->isClassConstructor()) { trackOptimizationOutcome(TrackedOutcome::CantInlineClassConstructor); return DontInline(inlineScript, "Not constructing class constructor"); } AnalysisMode analysisMode = info().analysisMode(); if (!CanIonCompile(inlineScript, analysisMode)) { trackOptimizationOutcome(TrackedOutcome::CantInlineDisabledIon); return DontInline(inlineScript, "Disabled Ion compilation"); } // Don't inline functions which don't have baseline scripts. if (!inlineScript->hasBaselineScript()) { trackOptimizationOutcome(TrackedOutcome::CantInlineNoBaseline); return DontInline(inlineScript, "No baseline jitcode"); } if (TooManyFormalArguments(target->nargs())) { trackOptimizationOutcome(TrackedOutcome::CantInlineTooManyArgs); return DontInline(inlineScript, "Too many args"); } // We check the number of actual arguments against the maximum number of // formal arguments as we do not want to encode all actual arguments in the // callerResumePoint. if (TooManyFormalArguments(callInfo.argc())) { trackOptimizationOutcome(TrackedOutcome::CantInlineTooManyArgs); return DontInline(inlineScript, "Too many actual args"); } if (hasCommonInliningPath(inlineScript)) { trackOptimizationOutcome(TrackedOutcome::HasCommonInliningPath); return DontInline(inlineScript, "Common inlining path"); } if (inlineScript->uninlineable()) { trackOptimizationOutcome(TrackedOutcome::CantInlineGeneric); return DontInline(inlineScript, "Uninlineable script"); } if (inlineScript->needsArgsObj()) { trackOptimizationOutcome(TrackedOutcome::CantInlineNeedsArgsObj); return DontInline(inlineScript, "Script that needs an arguments object"); } if (inlineScript->isDebuggee()) { trackOptimizationOutcome(TrackedOutcome::CantInlineDebuggee); return DontInline(inlineScript, "Script is debuggee"); } TypeSet::ObjectKey* targetKey = TypeSet::ObjectKey::get(target); if (targetKey->unknownProperties()) { trackOptimizationOutcome(TrackedOutcome::CantInlineUnknownProps); return DontInline(inlineScript, "Target type has unknown properties"); } return InliningDecision_Inline; } void IonBuilder::popCfgStack() { if (cfgStack_.back().isLoop()) loops_.popBack(); if (cfgStack_.back().state == CFGState::LABEL) labels_.popBack(); cfgStack_.popBack(); } bool IonBuilder::analyzeNewLoopTypes(MBasicBlock* entry, jsbytecode* start, jsbytecode* end) { // The phi inputs at the loop head only reflect types for variables that // were present at the start of the loop. If the variable changes to a new // type within the loop body, and that type is carried around to the loop // head, then we need to know about the new type up front. // // Since SSA information hasn't been constructed for the loop body yet, we // need a separate analysis to pick out the types that might flow around // the loop header. This is a best-effort analysis that may either over- // or under-approximate the set of such types. // // Over-approximating the types may lead to inefficient generated code, and // under-approximating the types will cause the loop body to be analyzed // multiple times as the correct types are deduced (see finishLoop). // If we restarted processing of an outer loop then get loop header types // directly from the last time we have previously processed this loop. This // both avoids repeated work from the bytecode traverse below, and will // also pick up types discovered while previously building the loop body. for (size_t i = 0; i < loopHeaders_.length(); i++) { if (loopHeaders_[i].pc == start) { MBasicBlock* oldEntry = loopHeaders_[i].header; // If this block has been discarded, its resume points will have // already discarded their operands. if (!oldEntry->isDead()) { MResumePoint* oldEntryRp = oldEntry->entryResumePoint(); size_t stackDepth = oldEntryRp->stackDepth(); for (size_t slot = 0; slot < stackDepth; slot++) { MDefinition* oldDef = oldEntryRp->getOperand(slot); if (!oldDef->isPhi()) { MOZ_ASSERT(oldDef->block()->id() < oldEntry->id()); MOZ_ASSERT(oldDef == entry->getSlot(slot)); continue; } MPhi* oldPhi = oldDef->toPhi(); MPhi* newPhi = entry->getSlot(slot)->toPhi(); if (!newPhi->addBackedgeType(alloc(), oldPhi->type(), oldPhi->resultTypeSet())) return false; } } // Update the most recent header for this loop encountered, in case // new types flow to the phis and the loop is processed at least // three times. loopHeaders_[i].header = entry; return true; } } if (!loopHeaders_.append(LoopHeader(start, entry))) return false; jsbytecode* last = nullptr; jsbytecode* earlier = nullptr; for (jsbytecode* pc = start; pc != end; earlier = last, last = pc, pc += GetBytecodeLength(pc)) { uint32_t slot; if (*pc == JSOP_SETLOCAL) slot = info().localSlot(GET_LOCALNO(pc)); else if (*pc == JSOP_SETARG) slot = info().argSlotUnchecked(GET_ARGNO(pc)); else continue; if (slot >= info().firstStackSlot()) continue; if (!analysis().maybeInfo(pc)) continue; if (!last) continue; MPhi* phi = entry->getSlot(slot)->toPhi(); if (*last == JSOP_POS) last = earlier; if (CodeSpec[*last].format & JOF_TYPESET) { TemporaryTypeSet* typeSet = bytecodeTypes(last); if (!typeSet->empty()) { MIRType type = typeSet->getKnownMIRType(); if (!phi->addBackedgeType(alloc(), type, typeSet)) return false; } } else if (*last == JSOP_GETLOCAL || *last == JSOP_GETARG) { uint32_t slot = (*last == JSOP_GETLOCAL) ? info().localSlot(GET_LOCALNO(last)) : info().argSlotUnchecked(GET_ARGNO(last)); if (slot < info().firstStackSlot()) { MPhi* otherPhi = entry->getSlot(slot)->toPhi(); if (otherPhi->hasBackedgeType()) { if (!phi->addBackedgeType(alloc(), otherPhi->type(), otherPhi->resultTypeSet())) return false; } } } else { MIRType type = MIRType::None; switch (*last) { case JSOP_VOID: case JSOP_UNDEFINED: type = MIRType::Undefined; break; case JSOP_GIMPLICITTHIS: if (!script()->hasNonSyntacticScope()) type = MIRType::Undefined; break; case JSOP_NULL: type = MIRType::Null; break; case JSOP_ZERO: case JSOP_ONE: case JSOP_INT8: case JSOP_INT32: case JSOP_UINT16: case JSOP_UINT24: case JSOP_BITAND: case JSOP_BITOR: case JSOP_BITXOR: case JSOP_BITNOT: case JSOP_RSH: case JSOP_LSH: case JSOP_URSH: type = MIRType::Int32; break; case JSOP_FALSE: case JSOP_TRUE: case JSOP_EQ: case JSOP_NE: case JSOP_LT: case JSOP_LE: case JSOP_GT: case JSOP_GE: case JSOP_NOT: case JSOP_STRICTEQ: case JSOP_STRICTNE: case JSOP_IN: case JSOP_INSTANCEOF: type = MIRType::Boolean; break; case JSOP_DOUBLE: type = MIRType::Double; break; case JSOP_STRING: case JSOP_TOSTRING: case JSOP_TYPEOF: case JSOP_TYPEOFEXPR: type = MIRType::String; break; case JSOP_SYMBOL: type = MIRType::Symbol; break; case JSOP_ADD: case JSOP_SUB: case JSOP_MUL: case JSOP_DIV: case JSOP_MOD: case JSOP_NEG: type = inspector->expectedResultType(last); break; default: break; } if (type != MIRType::None) { if (!phi->addBackedgeType(alloc(), type, nullptr)) return false; } } } return true; } bool IonBuilder::pushLoop(CFGState::State initial, jsbytecode* stopAt, MBasicBlock* entry, bool osr, jsbytecode* loopHead, jsbytecode* initialPc, jsbytecode* bodyStart, jsbytecode* bodyEnd, jsbytecode* exitpc, jsbytecode* continuepc) { ControlFlowInfo loop(cfgStack_.length(), continuepc); if (!loops_.append(loop)) return false; CFGState state; state.state = initial; state.stopAt = stopAt; state.loop.bodyStart = bodyStart; state.loop.bodyEnd = bodyEnd; state.loop.exitpc = exitpc; state.loop.continuepc = continuepc; state.loop.entry = entry; state.loop.osr = osr; state.loop.successor = nullptr; state.loop.breaks = nullptr; state.loop.continues = nullptr; state.loop.initialState = initial; state.loop.initialPc = initialPc; state.loop.initialStopAt = stopAt; state.loop.loopHead = loopHead; return cfgStack_.append(state); } bool IonBuilder::init() { { LifoAlloc::AutoFallibleScope fallibleAllocator(alloc().lifoAlloc()); if (!TypeScript::FreezeTypeSets(constraints(), script(), &thisTypes, &argTypes, &typeArray)) return false; } if (!alloc().ensureBallast()) return false; if (inlineCallInfo_) { // If we're inlining, the actual this/argument types are not necessarily // a subset of the script's observed types. |argTypes| is never accessed // for inlined scripts, so we just null it. thisTypes = inlineCallInfo_->thisArg()->resultTypeSet(); argTypes = nullptr; } if (!analysis().init(alloc(), gsn)) return false; // The baseline script normally has the bytecode type map, but compute // it ourselves if we do not have a baseline script. if (script()->hasBaselineScript()) { bytecodeTypeMap = script()->baselineScript()->bytecodeTypeMap(); } else { bytecodeTypeMap = alloc_->lifoAlloc()->newArrayUninitialized(script()->nTypeSets()); if (!bytecodeTypeMap) return false; FillBytecodeTypeMap(script(), bytecodeTypeMap); } return true; } bool IonBuilder::build() { if (!init()) return false; if (script()->hasBaselineScript()) script()->baselineScript()->resetMaxInliningDepth(); if (!setCurrentAndSpecializePhis(newBlock(pc))) return false; if (!current) return false; #ifdef JS_JITSPEW if (info().isAnalysis()) { JitSpew(JitSpew_IonScripts, "Analyzing script %s:%" PRIuSIZE " (%p) %s", script()->filename(), script()->lineno(), (void*)script(), AnalysisModeString(info().analysisMode())); } else { JitSpew(JitSpew_IonScripts, "%sompiling script %s:%" PRIuSIZE " (%p) (warmup-counter=%" PRIu32 ", level=%s)", (script()->hasIonScript() ? "Rec" : "C"), script()->filename(), script()->lineno(), (void*)script(), script()->getWarmUpCount(), OptimizationLevelString(optimizationInfo().level())); } #endif if (!initParameters()) return false; initLocals(); // Initialize something for the env chain. We can bail out before the // start instruction, but the snapshot is encoded *at* the start // instruction, which means generating any code that could load into // registers is illegal. MInstruction* env = MConstant::New(alloc(), UndefinedValue()); current->add(env); current->initSlot(info().environmentChainSlot(), env); // Initialize the return value. MInstruction* returnValue = MConstant::New(alloc(), UndefinedValue()); current->add(returnValue); current->initSlot(info().returnValueSlot(), returnValue); // Initialize the arguments object slot to undefined if necessary. if (info().hasArguments()) { MInstruction* argsObj = MConstant::New(alloc(), UndefinedValue()); current->add(argsObj); current->initSlot(info().argsObjSlot(), argsObj); } // Emit the start instruction, so we can begin real instructions. current->add(MStart::New(alloc())); // Guard against over-recursion. Do this before we start unboxing, since // this will create an OSI point that will read the incoming argument // values, which is nice to do before their last real use, to minimize // register/stack pressure. MCheckOverRecursed* check = MCheckOverRecursed::New(alloc()); current->add(check); MResumePoint* entryRpCopy = MResumePoint::Copy(alloc(), current->entryResumePoint()); if (!entryRpCopy) return false; check->setResumePoint(entryRpCopy); // Parameters have been checked to correspond to the typeset, now we unbox // what we can in an infallible manner. if (!rewriteParameters()) return false; // Check for redeclaration errors for global scripts. if (!info().funMaybeLazy() && !info().module() && script()->bodyScope()->is() && script()->bodyScope()->as().hasBindings()) { MGlobalNameConflictsCheck* redeclCheck = MGlobalNameConflictsCheck::New(alloc()); current->add(redeclCheck); MResumePoint* entryRpCopy = MResumePoint::Copy(alloc(), current->entryResumePoint()); if (!entryRpCopy) return false; redeclCheck->setResumePoint(entryRpCopy); } // It's safe to start emitting actual IR, so now build the env chain. if (!initEnvironmentChain()) return false; if (info().needsArgsObj() && !initArgumentsObject()) return false; // The type analysis phase attempts to insert unbox operations near // definitions of values. It also attempts to replace uses in resume points // with the narrower, unboxed variants. However, we must prevent this // replacement from happening on values in the entry snapshot. Otherwise we // could get this: // // v0 = MParameter(0) // v1 = MParameter(1) // -- ResumePoint(v2, v3) // v2 = Unbox(v0, INT32) // v3 = Unbox(v1, INT32) // // So we attach the initial resume point to each parameter, which the type // analysis explicitly checks (this is the same mechanism used for // effectful operations). for (uint32_t i = 0; i < info().endArgSlot(); i++) { MInstruction* ins = current->getEntrySlot(i)->toInstruction(); if (ins->type() != MIRType::Value) continue; MResumePoint* entryRpCopy = MResumePoint::Copy(alloc(), current->entryResumePoint()); if (!entryRpCopy) return false; ins->setResumePoint(entryRpCopy); } // lazyArguments should never be accessed in |argsObjAliasesFormals| scripts. if (info().hasArguments() && !info().argsObjAliasesFormals()) { lazyArguments_ = MConstant::New(alloc(), MagicValue(JS_OPTIMIZED_ARGUMENTS)); current->add(lazyArguments_); } insertRecompileCheck(); if (!traverseBytecode()) return false; // Discard unreferenced & pre-allocated resume points. replaceMaybeFallbackFunctionGetter(nullptr); if (script_->hasBaselineScript() && inlinedBytecodeLength_ > script_->baselineScript()->inlinedBytecodeLength()) { script_->baselineScript()->setInlinedBytecodeLength(inlinedBytecodeLength_); } if (!maybeAddOsrTypeBarriers()) return false; if (!processIterators()) return false; if (!info().isAnalysis() && !abortedPreliminaryGroups().empty()) { abortReason_ = AbortReason_PreliminaryObjects; return false; } if (shouldForceAbort()) { abortReason_ = AbortReason_Disable; return false; } MOZ_ASSERT(loopDepth_ == 0); abortReason_ = AbortReason_NoAbort; return true; } bool IonBuilder::processIterators() { // Find and mark phis that must transitively hold an iterator live. Vector worklist; for (size_t i = 0; i < iterators_.length(); i++) { MDefinition* iter = iterators_[i]; if (!iter->isInWorklist()) { if (!worklist.append(iter)) return false; iter->setInWorklist(); } } while (!worklist.empty()) { MDefinition* def = worklist.popCopy(); def->setNotInWorklist(); if (def->isPhi()) { MPhi* phi = def->toPhi(); phi->setIterator(); phi->setImplicitlyUsedUnchecked(); } for (MUseDefIterator iter(def); iter; iter++) { MDefinition* use = iter.def(); if (!use->isInWorklist() && (!use->isPhi() || !use->toPhi()->isIterator())) { if (!worklist.append(use)) return false; use->setInWorklist(); } } } return true; } bool IonBuilder::buildInline(IonBuilder* callerBuilder, MResumePoint* callerResumePoint, CallInfo& callInfo) { inlineCallInfo_ = &callInfo; if (!init()) return false; JitSpew(JitSpew_IonScripts, "Inlining script %s:%" PRIuSIZE " (%p)", script()->filename(), script()->lineno(), (void*)script()); callerBuilder_ = callerBuilder; callerResumePoint_ = callerResumePoint; if (callerBuilder->failedBoundsCheck_) failedBoundsCheck_ = true; if (callerBuilder->failedShapeGuard_) failedShapeGuard_ = true; if (callerBuilder->failedLexicalCheck_) failedLexicalCheck_ = true; safeForMinorGC_ = callerBuilder->safeForMinorGC_; // Generate single entrance block. if (!setCurrentAndSpecializePhis(newBlock(pc))) return false; if (!current) return false; current->setCallerResumePoint(callerResumePoint); // Connect the entrance block to the last block in the caller's graph. MBasicBlock* predecessor = callerBuilder->current; MOZ_ASSERT(predecessor == callerResumePoint->block()); predecessor->end(MGoto::New(alloc(), current)); if (!current->addPredecessorWithoutPhis(predecessor)) return false; // Initialize env chain slot to Undefined. It's set later by // |initEnvironmentChain|. MInstruction* env = MConstant::New(alloc(), UndefinedValue()); current->add(env); current->initSlot(info().environmentChainSlot(), env); // Initialize |return value| slot. MInstruction* returnValue = MConstant::New(alloc(), UndefinedValue()); current->add(returnValue); current->initSlot(info().returnValueSlot(), returnValue); // Initialize |arguments| slot. if (info().hasArguments()) { MInstruction* argsObj = MConstant::New(alloc(), UndefinedValue()); current->add(argsObj); current->initSlot(info().argsObjSlot(), argsObj); } // Initialize |this| slot. current->initSlot(info().thisSlot(), callInfo.thisArg()); JitSpew(JitSpew_Inlining, "Initializing %u arg slots", info().nargs()); // NB: Ion does not inline functions which |needsArgsObj|. So using argSlot() // instead of argSlotUnchecked() below is OK MOZ_ASSERT(!info().needsArgsObj()); // Initialize actually set arguments. uint32_t existing_args = Min(callInfo.argc(), info().nargs()); for (size_t i = 0; i < existing_args; ++i) { MDefinition* arg = callInfo.getArg(i); current->initSlot(info().argSlot(i), arg); } // Pass Undefined for missing arguments for (size_t i = callInfo.argc(); i < info().nargs(); ++i) { MConstant* arg = MConstant::New(alloc(), UndefinedValue()); current->add(arg); current->initSlot(info().argSlot(i), arg); } JitSpew(JitSpew_Inlining, "Initializing %u locals", info().nlocals()); initLocals(); JitSpew(JitSpew_Inlining, "Inline entry block MResumePoint %p, %u stack slots", (void*) current->entryResumePoint(), current->entryResumePoint()->stackDepth()); // +2 for the env chain and |this|, maybe another +1 for arguments object slot. MOZ_ASSERT(current->entryResumePoint()->stackDepth() == info().totalSlots()); if (script_->argumentsHasVarBinding()) { lazyArguments_ = MConstant::New(alloc(), MagicValue(JS_OPTIMIZED_ARGUMENTS)); current->add(lazyArguments_); } insertRecompileCheck(); // Initialize the env chain now that all resume points operands are // initialized. if (!initEnvironmentChain(callInfo.fun())) return false; if (!traverseBytecode()) return false; // Discard unreferenced & pre-allocated resume points. replaceMaybeFallbackFunctionGetter(nullptr); MOZ_ASSERT(iterators_.empty(), "Iterators should be added to outer builder"); if (!info().isAnalysis() && !abortedPreliminaryGroups().empty()) { abortReason_ = AbortReason_PreliminaryObjects; return false; } if (shouldForceAbort()) { abortReason_ = AbortReason_Disable; return false; } return true; } void IonBuilder::rewriteParameter(uint32_t slotIdx, MDefinition* param, int32_t argIndex) { MOZ_ASSERT(param->isParameter() || param->isGetArgumentsObjectArg()); TemporaryTypeSet* types = param->resultTypeSet(); MDefinition* actual = ensureDefiniteType(param, types->getKnownMIRType()); if (actual == param) return; // Careful! We leave the original MParameter in the entry resume point. The // arguments still need to be checked unless proven otherwise at the call // site, and these checks can bailout. We can end up: // v0 = Parameter(0) // v1 = Unbox(v0, INT32) // -- ResumePoint(v0) // // As usual, it would be invalid for v1 to be captured in the initial // resume point, rather than v0. current->rewriteSlot(slotIdx, actual); } // Apply Type Inference information to parameters early on, unboxing them if // they have a definitive type. The actual guards will be emitted by the code // generator, explicitly, as part of the function prologue. bool IonBuilder::rewriteParameters() { MOZ_ASSERT(info().environmentChainSlot() == 0); if (!info().funMaybeLazy()) return true; for (uint32_t i = info().startArgSlot(); i < info().endArgSlot(); i++) { if (!alloc().ensureBallast()) return false; MDefinition* param = current->getSlot(i); rewriteParameter(i, param, param->toParameter()->index()); } return true; } bool IonBuilder::initParameters() { if (!info().funMaybeLazy()) return true; // If we are doing OSR on a frame which initially executed in the // interpreter and didn't accumulate type information, try to use that OSR // frame to determine possible initial types for 'this' and parameters. if (thisTypes->empty() && baselineFrame_) { TypeSet::Type type = baselineFrame_->thisType; if (type.isSingletonUnchecked()) checkNurseryObject(type.singleton()); thisTypes->addType(type, alloc_->lifoAlloc()); } MParameter* param = MParameter::New(alloc(), MParameter::THIS_SLOT, thisTypes); current->add(param); current->initSlot(info().thisSlot(), param); for (uint32_t i = 0; i < info().nargs(); i++) { TemporaryTypeSet* types = &argTypes[i]; if (types->empty() && baselineFrame_ && !script_->baselineScript()->modifiesArguments()) { TypeSet::Type type = baselineFrame_->argTypes[i]; if (type.isSingletonUnchecked()) checkNurseryObject(type.singleton()); types->addType(type, alloc_->lifoAlloc()); } param = MParameter::New(alloc().fallible(), i, types); if (!param) return false; current->add(param); current->initSlot(info().argSlotUnchecked(i), param); } return true; } void IonBuilder::initLocals() { // Initialize all frame slots to undefined. Lexical bindings are temporal // dead zoned in bytecode. if (info().nlocals() == 0) return; MConstant* undef = MConstant::New(alloc(), UndefinedValue()); current->add(undef); for (uint32_t i = 0; i < info().nlocals(); i++) current->initSlot(info().localSlot(i), undef); } bool IonBuilder::initEnvironmentChain(MDefinition* callee) { MInstruction* env = nullptr; // If the script doesn't use the envchain, then it's already initialized // from earlier. However, always make a env chain when |needsArgsObj| is true // for the script, since arguments object construction requires the env chain // to be passed in. if (!info().needsArgsObj() && !analysis().usesEnvironmentChain()) return true; // The env chain is only tracked in scripts that have NAME opcodes which // will try to access the env. For other scripts, the env instructions // will be held live by resume points and code will still be generated for // them, so just use a constant undefined value. if (JSFunction* fun = info().funMaybeLazy()) { if (!callee) { MCallee* calleeIns = MCallee::New(alloc()); current->add(calleeIns); callee = calleeIns; } env = MFunctionEnvironment::New(alloc(), callee); current->add(env); // This reproduce what is done in CallObject::createForFunction. Skip // this for the arguments analysis, as the script might not have a // baseline script with template objects yet. if (fun->needsSomeEnvironmentObject() && info().analysisMode() != Analysis_ArgumentsUsage) { if (fun->needsNamedLambdaEnvironment()) { env = createNamedLambdaObject(callee, env); if (!env) return false; } // TODO: Parameter expression-induced extra var environment not // yet handled. if (fun->needsExtraBodyVarEnvironment()) return abort("Extra var environment unsupported"); if (fun->needsCallObject()) { env = createCallObject(callee, env); if (!env) return false; } } } else if (ModuleObject* module = info().module()) { // Modules use a pre-created env object. env = constant(ObjectValue(module->initialEnvironment())); } else { // For global scripts without a non-syntactic global scope, the env // chain is the global lexical env. MOZ_ASSERT(!script()->isForEval()); MOZ_ASSERT(!script()->hasNonSyntacticScope()); env = constant(ObjectValue(script()->global().lexicalEnvironment())); } current->setEnvironmentChain(env); return true; } bool IonBuilder::initArgumentsObject() { JitSpew(JitSpew_IonMIR, "%s:%" PRIuSIZE " - Emitting code to initialize arguments object! block=%p", script()->filename(), script()->lineno(), current); MOZ_ASSERT(info().needsArgsObj()); bool mapped = script()->hasMappedArgsObj(); ArgumentsObject* templateObj = script()->compartment()->maybeArgumentsTemplateObject(mapped); MCreateArgumentsObject* argsObj = MCreateArgumentsObject::New(alloc(), current->environmentChain(), templateObj); current->add(argsObj); current->setArgumentsObject(argsObj); return true; } bool IonBuilder::addOsrValueTypeBarrier(uint32_t slot, MInstruction** def_, MIRType type, TemporaryTypeSet* typeSet) { MInstruction*& def = *def_; MBasicBlock* osrBlock = def->block(); // Clear bogus type information added in newOsrPreheader(). def->setResultType(MIRType::Value); def->setResultTypeSet(nullptr); if (typeSet && !typeSet->unknown()) { MInstruction* barrier = MTypeBarrier::New(alloc(), def, typeSet); osrBlock->insertBefore(osrBlock->lastIns(), barrier); osrBlock->rewriteSlot(slot, barrier); def = barrier; // If the TypeSet is more precise than |type|, adjust |type| for the // code below. if (type == MIRType::Value) type = barrier->type(); } else if (type == MIRType::Null || type == MIRType::Undefined || type == MIRType::MagicOptimizedArguments) { // No unbox instruction will be added below, so check the type by // adding a type barrier for a singleton type set. TypeSet::Type ntype = TypeSet::PrimitiveType(ValueTypeFromMIRType(type)); LifoAlloc* lifoAlloc = alloc().lifoAlloc(); typeSet = lifoAlloc->new_(lifoAlloc, ntype); if (!typeSet) return false; MInstruction* barrier = MTypeBarrier::New(alloc(), def, typeSet); osrBlock->insertBefore(osrBlock->lastIns(), barrier); osrBlock->rewriteSlot(slot, barrier); def = barrier; } switch (type) { case MIRType::Boolean: case MIRType::Int32: case MIRType::Double: case MIRType::String: case MIRType::Symbol: case MIRType::Object: if (type != def->type()) { MUnbox* unbox = MUnbox::New(alloc(), def, type, MUnbox::Fallible); osrBlock->insertBefore(osrBlock->lastIns(), unbox); osrBlock->rewriteSlot(slot, unbox); def = unbox; } break; case MIRType::Null: { MConstant* c = MConstant::New(alloc(), NullValue()); osrBlock->insertBefore(osrBlock->lastIns(), c); osrBlock->rewriteSlot(slot, c); def = c; break; } case MIRType::Undefined: { MConstant* c = MConstant::New(alloc(), UndefinedValue()); osrBlock->insertBefore(osrBlock->lastIns(), c); osrBlock->rewriteSlot(slot, c); def = c; break; } case MIRType::MagicOptimizedArguments: MOZ_ASSERT(lazyArguments_); osrBlock->rewriteSlot(slot, lazyArguments_); def = lazyArguments_; break; default: break; } MOZ_ASSERT(def == osrBlock->getSlot(slot)); return true; } bool IonBuilder::maybeAddOsrTypeBarriers() { if (!info().osrPc()) return true; // The loop has successfully been processed, and the loop header phis // have their final type. Add unboxes and type barriers in the OSR // block to check that the values have the appropriate type, and update // the types in the preheader. MBasicBlock* osrBlock = graph().osrBlock(); if (!osrBlock) { // Because IonBuilder does not compile catch blocks, it's possible to // end up without an OSR block if the OSR pc is only reachable via a // break-statement inside the catch block. For instance: // // for (;;) { // try { // throw 3; // } catch(e) { // break; // } // } // while (..) { } // <= OSR here, only reachable via catch block. // // For now we just abort in this case. MOZ_ASSERT(graph().hasTryBlock()); return abort("OSR block only reachable through catch block"); } MBasicBlock* preheader = osrBlock->getSuccessor(0); MBasicBlock* header = preheader->getSuccessor(0); static const size_t OSR_PHI_POSITION = 1; MOZ_ASSERT(preheader->getPredecessor(OSR_PHI_POSITION) == osrBlock); MResumePoint* headerRp = header->entryResumePoint(); size_t stackDepth = headerRp->stackDepth(); MOZ_ASSERT(stackDepth == osrBlock->stackDepth()); for (uint32_t slot = info().startArgSlot(); slot < stackDepth; slot++) { // Aliased slots are never accessed, since they need to go through // the callobject. The typebarriers are added there and can be // discarded here. if (info().isSlotAliased(slot)) continue; if (!alloc().ensureBallast()) return false; MInstruction* def = osrBlock->getSlot(slot)->toInstruction(); MPhi* preheaderPhi = preheader->getSlot(slot)->toPhi(); MPhi* headerPhi = headerRp->getOperand(slot)->toPhi(); MIRType type = headerPhi->type(); TemporaryTypeSet* typeSet = headerPhi->resultTypeSet(); if (!addOsrValueTypeBarrier(slot, &def, type, typeSet)) return false; preheaderPhi->replaceOperand(OSR_PHI_POSITION, def); preheaderPhi->setResultType(type); preheaderPhi->setResultTypeSet(typeSet); } return true; } // We try to build a control-flow graph in the order that it would be built as // if traversing the AST. This leads to a nice ordering and lets us build SSA // in one pass, since the bytecode is structured. // // We traverse the bytecode iteratively, maintaining a current basic block. // Each basic block has a mapping of local slots to instructions, as well as a // stack depth. As we encounter instructions we mutate this mapping in the // current block. // // Things get interesting when we encounter a control structure. This can be // either an IFEQ, downward GOTO, or a decompiler hint stashed away in source // notes. Once we encounter such an opcode, we recover the structure of the // control flow (its branches and bounds), and push it on a stack. // // As we continue traversing the bytecode, we look for points that would // terminate the topmost control flow path pushed on the stack. These are: // (1) The bounds of the current structure (end of a loop or join/edge of a // branch). // (2) A "return", "break", or "continue" statement. // // For (1), we expect that there is a current block in the progress of being // built, and we complete the necessary edges in the CFG. For (2), we expect // that there is no active block. // // For normal diamond join points, we construct Phi nodes as we add // predecessors. For loops, care must be taken to propagate Phi nodes back // through uses in the loop body. bool IonBuilder::traverseBytecode() { for (;;) { MOZ_ASSERT(pc < info().limitPC()); for (;;) { if (!alloc().ensureBallast()) return false; // Check if we've hit an expected join point or edge in the bytecode. // Leaving one control structure could place us at the edge of another, // thus |while| instead of |if| so we don't skip any opcodes. MOZ_ASSERT_IF(!cfgStack_.empty(), cfgStack_.back().stopAt >= pc); if (!cfgStack_.empty() && cfgStack_.back().stopAt == pc) { ControlStatus status = processCfgStack(); if (status == ControlStatus_Error) return false; if (status == ControlStatus_Abort) return abort("Aborted while processing control flow"); if (!current) return true; continue; } // Some opcodes need to be handled early because they affect control // flow, terminating the current basic block and/or instructing the // traversal algorithm to continue from a new pc. // // (1) If the opcode does not affect control flow, then the opcode // is inspected and transformed to IR. This is the process_opcode // label. // (2) A loop could be detected via a forward GOTO. In this case, // we don't want to process the GOTO, but the following // instruction. // (3) A RETURN, STOP, BREAK, or CONTINUE may require processing the // CFG stack to terminate open branches. // // Similar to above, snooping control flow could land us at another // control flow point, so we iterate until it's time to inspect a real // opcode. ControlStatus status; if ((status = snoopControlFlow(JSOp(*pc))) == ControlStatus_None) break; if (status == ControlStatus_Error) return false; if (status == ControlStatus_Abort) return abort("Aborted while processing control flow"); if (!current) return true; } #ifdef DEBUG // In debug builds, after compiling this op, check that all values // popped by this opcode either: // // (1) Have the ImplicitlyUsed flag set on them. // (2) Have more uses than before compiling this op (the value is // used as operand of a new MIR instruction). // // This is used to catch problems where IonBuilder pops a value without // adding any SSA uses and doesn't call setImplicitlyUsedUnchecked on it. Vector popped(alloc()); Vector poppedUses(alloc()); unsigned nuses = GetUseCount(script_, script_->pcToOffset(pc)); for (unsigned i = 0; i < nuses; i++) { MDefinition* def = current->peek(-int32_t(i + 1)); if (!popped.append(def) || !poppedUses.append(def->defUseCount())) return false; } #endif // Nothing in inspectOpcode() is allowed to advance the pc. JSOp op = JSOp(*pc); if (!inspectOpcode(op)) return false; #ifdef DEBUG for (size_t i = 0; i < popped.length(); i++) { switch (op) { case JSOP_POP: case JSOP_POPN: case JSOP_DUPAT: case JSOP_DUP: case JSOP_DUP2: case JSOP_PICK: case JSOP_UNPICK: case JSOP_SWAP: case JSOP_SETARG: case JSOP_SETLOCAL: case JSOP_INITLEXICAL: case JSOP_SETRVAL: case JSOP_VOID: // Don't require SSA uses for values popped by these ops. break; case JSOP_POS: case JSOP_TOID: case JSOP_TOSTRING: // These ops may leave their input on the stack without setting // the ImplicitlyUsed flag. If this value will be popped immediately, // we may replace it with |undefined|, but the difference is // not observable. MOZ_ASSERT(i == 0); if (current->peek(-1) == popped[0]) break; MOZ_FALLTHROUGH; default: MOZ_ASSERT(popped[i]->isImplicitlyUsed() || // MNewDerivedTypedObject instances are // often dead unless they escape from the // fn. See IonBuilder::loadTypedObjectData() // for more details. popped[i]->isNewDerivedTypedObject() || popped[i]->defUseCount() > poppedUses[i]); break; } } #endif pc += CodeSpec[op].length; current->updateTrackedSite(bytecodeSite(pc)); } return true; } IonBuilder::ControlStatus IonBuilder::snoopControlFlow(JSOp op) { switch (op) { case JSOP_NOP: return maybeLoop(op, info().getNote(gsn, pc)); case JSOP_POP: return maybeLoop(op, info().getNote(gsn, pc)); case JSOP_RETURN: case JSOP_RETRVAL: return processReturn(op); case JSOP_THROW: return processThrow(); case JSOP_GOTO: { jssrcnote* sn = info().getNote(gsn, pc); switch (sn ? SN_TYPE(sn) : SRC_NULL) { case SRC_BREAK: case SRC_BREAK2LABEL: return processBreak(op, sn); case SRC_CONTINUE: return processContinue(op); case SRC_SWITCHBREAK: return processSwitchBreak(op); case SRC_WHILE: case SRC_FOR_IN: case SRC_FOR_OF: // while (cond) { } return whileOrForInLoop(sn); default: // Hard assert for now - make an error later. MOZ_CRASH("unknown goto case"); } break; } case JSOP_TABLESWITCH: return tableSwitch(op, info().getNote(gsn, pc)); case JSOP_IFNE: // We should never reach an IFNE, it's a stopAt point, which will // trigger closing the loop. MOZ_CRASH("we should never reach an ifne!"); default: break; } return ControlStatus_None; } bool IonBuilder::inspectOpcode(JSOp op) { MOZ_ASSERT(analysis_.maybeInfo(pc), "Compiling unreachable op"); switch (op) { case JSOP_NOP: case JSOP_NOP_DESTRUCTURING: case JSOP_TRY_DESTRUCTURING_ITERCLOSE: case JSOP_LINENO: case JSOP_LOOPENTRY: case JSOP_JUMPTARGET: return true; case JSOP_LABEL: return jsop_label(); case JSOP_UNDEFINED: // If this ever changes, change what JSOP_GIMPLICITTHIS does too. pushConstant(UndefinedValue()); return true; case JSOP_IFEQ: return jsop_ifeq(JSOP_IFEQ); case JSOP_TRY: return jsop_try(); case JSOP_CONDSWITCH: return jsop_condswitch(); case JSOP_BITNOT: return jsop_bitnot(); case JSOP_BITAND: case JSOP_BITOR: case JSOP_BITXOR: case JSOP_LSH: case JSOP_RSH: case JSOP_URSH: return jsop_bitop(op); case JSOP_ADD: case JSOP_SUB: case JSOP_MUL: case JSOP_DIV: case JSOP_MOD: return jsop_binary_arith(op); case JSOP_POW: return jsop_pow(); case JSOP_POS: return jsop_pos(); case JSOP_NEG: return jsop_neg(); case JSOP_TOSTRING: return jsop_tostring(); case JSOP_AND: case JSOP_OR: return jsop_andor(op); case JSOP_DEFVAR: return jsop_defvar(GET_UINT32_INDEX(pc)); case JSOP_DEFLET: case JSOP_DEFCONST: return jsop_deflexical(GET_UINT32_INDEX(pc)); case JSOP_DEFFUN: return jsop_deffun(GET_UINT32_INDEX(pc)); case JSOP_EQ: case JSOP_NE: case JSOP_STRICTEQ: case JSOP_STRICTNE: case JSOP_LT: case JSOP_LE: case JSOP_GT: case JSOP_GE: return jsop_compare(op); case JSOP_DOUBLE: pushConstant(info().getConst(pc)); return true; case JSOP_STRING: pushConstant(StringValue(info().getAtom(pc))); return true; case JSOP_SYMBOL: { unsigned which = GET_UINT8(pc); JS::Symbol* sym = compartment->runtime()->wellKnownSymbols().get(which); pushConstant(SymbolValue(sym)); return true; } case JSOP_ZERO: pushConstant(Int32Value(0)); return true; case JSOP_ONE: pushConstant(Int32Value(1)); return true; case JSOP_NULL: pushConstant(NullValue()); return true; case JSOP_VOID: current->pop(); pushConstant(UndefinedValue()); return true; case JSOP_HOLE: pushConstant(MagicValue(JS_ELEMENTS_HOLE)); return true; case JSOP_FALSE: pushConstant(BooleanValue(false)); return true; case JSOP_TRUE: pushConstant(BooleanValue(true)); return true; case JSOP_ARGUMENTS: return jsop_arguments(); case JSOP_RUNONCE: return jsop_runonce(); case JSOP_REST: return jsop_rest(); case JSOP_GETARG: if (info().argsObjAliasesFormals()) { MGetArgumentsObjectArg* getArg = MGetArgumentsObjectArg::New(alloc(), current->argumentsObject(), GET_ARGNO(pc)); current->add(getArg); current->push(getArg); } else { current->pushArg(GET_ARGNO(pc)); } return true; case JSOP_SETARG: return jsop_setarg(GET_ARGNO(pc)); case JSOP_GETLOCAL: current->pushLocal(GET_LOCALNO(pc)); return true; case JSOP_SETLOCAL: current->setLocal(GET_LOCALNO(pc)); return true; case JSOP_THROWSETCONST: case JSOP_THROWSETALIASEDCONST: case JSOP_THROWSETCALLEE: return jsop_throwsetconst(); case JSOP_CHECKLEXICAL: return jsop_checklexical(); case JSOP_INITLEXICAL: current->setLocal(GET_LOCALNO(pc)); return true; case JSOP_INITGLEXICAL: { MOZ_ASSERT(!script()->hasNonSyntacticScope()); MDefinition* value = current->pop(); current->push(constant(ObjectValue(script()->global().lexicalEnvironment()))); current->push(value); return jsop_setprop(info().getAtom(pc)->asPropertyName()); } case JSOP_CHECKALIASEDLEXICAL: return jsop_checkaliasedlexical(EnvironmentCoordinate(pc)); case JSOP_INITALIASEDLEXICAL: return jsop_setaliasedvar(EnvironmentCoordinate(pc)); case JSOP_UNINITIALIZED: pushConstant(MagicValue(JS_UNINITIALIZED_LEXICAL)); return true; case JSOP_POP: { MDefinition* def = current->pop(); // POP opcodes frequently appear where values are killed, e.g. after // SET* opcodes. Place a resume point afterwards to avoid capturing // the dead value in later snapshots, except in places where that // resume point is obviously unnecessary. if (pc[JSOP_POP_LENGTH] == JSOP_POP) return true; if (def->isConstant()) return true; return maybeInsertResume(); } case JSOP_POPN: for (uint32_t i = 0, n = GET_UINT16(pc); i < n; i++) current->pop(); return true; case JSOP_DUPAT: current->pushSlot(current->stackDepth() - 1 - GET_UINT24(pc)); return true; case JSOP_NEWINIT: if (GET_UINT8(pc) == JSProto_Array) return jsop_newarray(0); return jsop_newobject(); case JSOP_NEWARRAY: return jsop_newarray(GET_UINT32(pc)); case JSOP_NEWARRAY_COPYONWRITE: return jsop_newarray_copyonwrite(); case JSOP_NEWOBJECT: return jsop_newobject(); case JSOP_INITELEM: case JSOP_INITHIDDENELEM: return jsop_initelem(); case JSOP_INITELEM_ARRAY: return jsop_initelem_array(); case JSOP_INITPROP: case JSOP_INITLOCKEDPROP: case JSOP_INITHIDDENPROP: { PropertyName* name = info().getAtom(pc)->asPropertyName(); return jsop_initprop(name); } case JSOP_MUTATEPROTO: { return jsop_mutateproto(); } case JSOP_INITPROP_GETTER: case JSOP_INITHIDDENPROP_GETTER: case JSOP_INITPROP_SETTER: case JSOP_INITHIDDENPROP_SETTER: { PropertyName* name = info().getAtom(pc)->asPropertyName(); return jsop_initprop_getter_setter(name); } case JSOP_INITELEM_GETTER: case JSOP_INITHIDDENELEM_GETTER: case JSOP_INITELEM_SETTER: case JSOP_INITHIDDENELEM_SETTER: return jsop_initelem_getter_setter(); case JSOP_FUNCALL: return jsop_funcall(GET_ARGC(pc)); case JSOP_FUNAPPLY: return jsop_funapply(GET_ARGC(pc)); case JSOP_CALL: case JSOP_CALL_IGNORES_RV: case JSOP_CALLITER: case JSOP_NEW: case JSOP_SUPERCALL: if (op == JSOP_CALLITER) { if (!outermostBuilder()->iterators_.append(current->peek(-1))) return false; } return jsop_call(GET_ARGC(pc), (JSOp)*pc == JSOP_NEW || (JSOp)*pc == JSOP_SUPERCALL, (JSOp)*pc == JSOP_CALL_IGNORES_RV); case JSOP_EVAL: case JSOP_STRICTEVAL: return jsop_eval(GET_ARGC(pc)); case JSOP_INT8: pushConstant(Int32Value(GET_INT8(pc))); return true; case JSOP_UINT16: pushConstant(Int32Value(GET_UINT16(pc))); return true; case JSOP_GETGNAME: { PropertyName* name = info().getAtom(pc)->asPropertyName(); if (!script()->hasNonSyntacticScope()) return jsop_getgname(name); return jsop_getname(name); } case JSOP_SETGNAME: case JSOP_STRICTSETGNAME: { PropertyName* name = info().getAtom(pc)->asPropertyName(); JSObject* obj = nullptr; if (!script()->hasNonSyntacticScope()) obj = testGlobalLexicalBinding(name); if (obj) return setStaticName(obj, name); return jsop_setprop(name); } case JSOP_GETNAME: { PropertyName* name = info().getAtom(pc)->asPropertyName(); return jsop_getname(name); } case JSOP_GETINTRINSIC: { PropertyName* name = info().getAtom(pc)->asPropertyName(); return jsop_intrinsic(name); } case JSOP_GETIMPORT: { PropertyName* name = info().getAtom(pc)->asPropertyName(); return jsop_getimport(name); } case JSOP_BINDGNAME: if (!script()->hasNonSyntacticScope()) { if (JSObject* env = testGlobalLexicalBinding(info().getName(pc))) { pushConstant(ObjectValue(*env)); return true; } } // Fall through to JSOP_BINDNAME MOZ_FALLTHROUGH; case JSOP_BINDNAME: return jsop_bindname(info().getName(pc)); case JSOP_BINDVAR: return jsop_bindvar(); case JSOP_DUP: current->pushSlot(current->stackDepth() - 1); return true; case JSOP_DUP2: return jsop_dup2(); case JSOP_SWAP: current->swapAt(-1); return true; case JSOP_PICK: current->pick(-GET_INT8(pc)); return true; case JSOP_UNPICK: current->unpick(-GET_INT8(pc)); return true; case JSOP_GETALIASEDVAR: return jsop_getaliasedvar(EnvironmentCoordinate(pc)); case JSOP_SETALIASEDVAR: return jsop_setaliasedvar(EnvironmentCoordinate(pc)); case JSOP_UINT24: pushConstant(Int32Value(GET_UINT24(pc))); return true; case JSOP_INT32: pushConstant(Int32Value(GET_INT32(pc))); return true; case JSOP_LOOPHEAD: // JSOP_LOOPHEAD is handled when processing the loop header. MOZ_CRASH("JSOP_LOOPHEAD outside loop"); case JSOP_GETELEM: case JSOP_CALLELEM: if (!jsop_getelem()) return false; if (op == JSOP_CALLELEM && !improveThisTypesForCall()) return false; return true; case JSOP_SETELEM: case JSOP_STRICTSETELEM: return jsop_setelem(); case JSOP_LENGTH: return jsop_length(); case JSOP_NOT: return jsop_not(); case JSOP_FUNCTIONTHIS: return jsop_functionthis(); case JSOP_GLOBALTHIS: return jsop_globalthis(); case JSOP_CALLEE: { MDefinition* callee = getCallee(); current->push(callee); return true; } case JSOP_GETPROP: case JSOP_CALLPROP: { PropertyName* name = info().getAtom(pc)->asPropertyName(); if (!jsop_getprop(name)) return false; if (op == JSOP_CALLPROP && !improveThisTypesForCall()) return false; return true; } case JSOP_SETPROP: case JSOP_STRICTSETPROP: case JSOP_SETNAME: case JSOP_STRICTSETNAME: { PropertyName* name = info().getAtom(pc)->asPropertyName(); return jsop_setprop(name); } case JSOP_DELPROP: case JSOP_STRICTDELPROP: { PropertyName* name = info().getAtom(pc)->asPropertyName(); return jsop_delprop(name); } case JSOP_DELELEM: case JSOP_STRICTDELELEM: return jsop_delelem(); case JSOP_REGEXP: return jsop_regexp(info().getRegExp(pc)); case JSOP_CALLSITEOBJ: pushConstant(ObjectValue(*(info().getObject(pc)))); return true; case JSOP_OBJECT: return jsop_object(info().getObject(pc)); case JSOP_TYPEOF: case JSOP_TYPEOFEXPR: return jsop_typeof(); case JSOP_TOASYNC: return jsop_toasync(); case JSOP_TOID: return jsop_toid(); case JSOP_LAMBDA: return jsop_lambda(info().getFunction(pc)); case JSOP_LAMBDA_ARROW: return jsop_lambda_arrow(info().getFunction(pc)); case JSOP_SETFUNNAME: return jsop_setfunname(GET_UINT8(pc)); case JSOP_ITER: return jsop_iter(GET_INT8(pc)); case JSOP_MOREITER: return jsop_itermore(); case JSOP_ISNOITER: return jsop_isnoiter(); case JSOP_ENDITER: return jsop_iterend(); case JSOP_IN: return jsop_in(); case JSOP_SETRVAL: MOZ_ASSERT(!script()->noScriptRval()); current->setSlot(info().returnValueSlot(), current->pop()); return true; case JSOP_INSTANCEOF: return jsop_instanceof(); case JSOP_DEBUGLEAVELEXICALENV: return true; case JSOP_DEBUGGER: return jsop_debugger(); case JSOP_GIMPLICITTHIS: if (!script()->hasNonSyntacticScope()) { pushConstant(UndefinedValue()); return true; } // Just fall through to the unsupported bytecode case. break; case JSOP_NEWTARGET: return jsop_newtarget(); case JSOP_CHECKISOBJ: return jsop_checkisobj(GET_UINT8(pc)); case JSOP_CHECKISCALLABLE: return jsop_checkiscallable(GET_UINT8(pc)); case JSOP_CHECKOBJCOERCIBLE: return jsop_checkobjcoercible(); case JSOP_DEBUGCHECKSELFHOSTED: { #ifdef DEBUG MDebugCheckSelfHosted* check = MDebugCheckSelfHosted::New(alloc(), current->pop()); current->add(check); current->push(check); if (!resumeAfter(check)) return false; #endif return true; } case JSOP_IS_CONSTRUCTING: pushConstant(MagicValue(JS_IS_CONSTRUCTING)); return true; #ifdef DEBUG case JSOP_PUSHLEXICALENV: case JSOP_FRESHENLEXICALENV: case JSOP_RECREATELEXICALENV: case JSOP_POPLEXICALENV: // These opcodes are currently unhandled by Ion, but in principle // there's no reason they couldn't be. Whenever this happens, OSR // will have to consider that JSOP_{FRESHEN,RECREATE}LEXICALENV // mutates the env chain -- right now MBasicBlock::environmentChain() // caches the env chain. JSOP_{FRESHEN,RECREATE}LEXICALENV must // update that stale value. #endif default: // Any unused opcodes and JSOP_LIMIT will end up here without having // to explicitly specify break; } // Track a simpler message, since the actionable abort message is a // static string, and the internal opcode name isn't an actionable // thing anyways. trackActionableAbort("Unsupported bytecode"); #ifdef DEBUG return abort("Unsupported opcode: %s", CodeName[op]); #else return abort("Unsupported opcode: %d", op); #endif } // Given that the current control flow structure has ended forcefully, // via a return, break, or continue (rather than joining), propagate the // termination up. For example, a return nested 5 loops deep may terminate // every outer loop at once, if there are no intervening conditionals: // // for (...) { // for (...) { // return x; // } // } // // If |current| is nullptr when this function returns, then there is no more // control flow to be processed. IonBuilder::ControlStatus IonBuilder::processControlEnd() { MOZ_ASSERT(!current); if (cfgStack_.empty()) { // If there is no more control flow to process, then this is the // last return in the function. return ControlStatus_Ended; } return processCfgStack(); } // Processes the top of the CFG stack. This is used from two places: // (1) processControlEnd(), whereby a break, continue, or return may interrupt // an in-progress CFG structure before reaching its actual termination // point in the bytecode. // (2) traverseBytecode(), whereby we reach the last instruction in a CFG // structure. IonBuilder::ControlStatus IonBuilder::processCfgStack() { ControlStatus status = processCfgEntry(cfgStack_.back()); // If this terminated a CFG structure, act like processControlEnd() and // keep propagating upward. while (status == ControlStatus_Ended) { popCfgStack(); if (cfgStack_.empty()) return status; status = processCfgEntry(cfgStack_.back()); } // If some join took place, the current structure is finished. if (status == ControlStatus_Joined) popCfgStack(); return status; } IonBuilder::ControlStatus IonBuilder::processCfgEntry(CFGState& state) { switch (state.state) { case CFGState::IF_TRUE: case CFGState::IF_TRUE_EMPTY_ELSE: return processIfEnd(state); case CFGState::IF_ELSE_TRUE: return processIfElseTrueEnd(state); case CFGState::IF_ELSE_FALSE: return processIfElseFalseEnd(state); case CFGState::DO_WHILE_LOOP_BODY: return processDoWhileBodyEnd(state); case CFGState::DO_WHILE_LOOP_COND: return processDoWhileCondEnd(state); case CFGState::WHILE_LOOP_COND: return processWhileCondEnd(state); case CFGState::WHILE_LOOP_BODY: return processWhileBodyEnd(state); case CFGState::FOR_LOOP_COND: return processForCondEnd(state); case CFGState::FOR_LOOP_BODY: return processForBodyEnd(state); case CFGState::FOR_LOOP_UPDATE: return processForUpdateEnd(state); case CFGState::TABLE_SWITCH: return processNextTableSwitchCase(state); case CFGState::COND_SWITCH_CASE: return processCondSwitchCase(state); case CFGState::COND_SWITCH_BODY: return processCondSwitchBody(state); case CFGState::AND_OR: return processAndOrEnd(state); case CFGState::LABEL: return processLabelEnd(state); case CFGState::TRY: return processTryEnd(state); default: MOZ_CRASH("unknown cfgstate"); } } IonBuilder::ControlStatus IonBuilder::processIfEnd(CFGState& state) { bool thenBranchTerminated = !current; if (!thenBranchTerminated) { // Here, the false block is the join point. Create an edge from the // current block to the false block. Note that a RETURN opcode // could have already ended the block. current->end(MGoto::New(alloc(), state.branch.ifFalse)); if (!state.branch.ifFalse->addPredecessor(alloc(), current)) return ControlStatus_Error; } if (!setCurrentAndSpecializePhis(state.branch.ifFalse)) return ControlStatus_Error; graph().moveBlockToEnd(current); pc = current->pc(); if (thenBranchTerminated) { // If we can't reach here via the then-branch, we can filter the types // after the if-statement based on the if-condition. MTest* test = state.branch.test; if (!improveTypesAtTest(test->getOperand(0), test->ifTrue() == current, test)) return ControlStatus_Error; } return ControlStatus_Joined; } IonBuilder::ControlStatus IonBuilder::processIfElseTrueEnd(CFGState& state) { // We've reached the end of the true branch of an if-else. Don't // create an edge yet, just transition to parsing the false branch. state.state = CFGState::IF_ELSE_FALSE; state.branch.ifTrue = current; state.stopAt = state.branch.falseEnd; pc = state.branch.ifFalse->pc(); if (!setCurrentAndSpecializePhis(state.branch.ifFalse)) return ControlStatus_Error; graph().moveBlockToEnd(current); MTest* test = state.branch.test; if (!improveTypesAtTest(test->getOperand(0), test->ifTrue() == current, test)) return ControlStatus_Error; return ControlStatus_Jumped; } IonBuilder::ControlStatus IonBuilder::processIfElseFalseEnd(CFGState& state) { // Update the state to have the latest block from the false path. state.branch.ifFalse = current; // To create the join node, we need an incoming edge that has not been // terminated yet. MBasicBlock* pred = state.branch.ifTrue ? state.branch.ifTrue : state.branch.ifFalse; MBasicBlock* other = (pred == state.branch.ifTrue) ? state.branch.ifFalse : state.branch.ifTrue; if (!pred) return ControlStatus_Ended; // Create a new block to represent the join. MBasicBlock* join = newBlock(pred, state.branch.falseEnd); if (!join) return ControlStatus_Error; // Create edges from the true and false blocks as needed. pred->end(MGoto::New(alloc(), join)); if (other) { other->end(MGoto::New(alloc(), join)); if (!join->addPredecessor(alloc(), other)) return ControlStatus_Error; } // Ignore unreachable remainder of false block if existent. if (!setCurrentAndSpecializePhis(join)) return ControlStatus_Error; pc = current->pc(); return ControlStatus_Joined; } IonBuilder::ControlStatus IonBuilder::processBrokenLoop(CFGState& state) { MOZ_ASSERT(!current); MOZ_ASSERT(loopDepth_); loopDepth_--; // A broken loop is not a real loop (it has no header or backedge), so // reset the loop depth. for (MBasicBlockIterator i(graph().begin(state.loop.entry)); i != graph().end(); i++) { if (i->loopDepth() > loopDepth_) i->setLoopDepth(i->loopDepth() - 1); } // If the loop started with a condition (while/for) then even if the // structure never actually loops, the condition itself can still fail and // thus we must resume at the successor, if one exists. if (!setCurrentAndSpecializePhis(state.loop.successor)) return ControlStatus_Error; if (current) { MOZ_ASSERT(current->loopDepth() == loopDepth_); graph().moveBlockToEnd(current); } // Join the breaks together and continue parsing. if (state.loop.breaks) { MBasicBlock* block = createBreakCatchBlock(state.loop.breaks, state.loop.exitpc); if (!block) return ControlStatus_Error; if (current) { current->end(MGoto::New(alloc(), block)); if (!block->addPredecessor(alloc(), current)) return ControlStatus_Error; } if (!setCurrentAndSpecializePhis(block)) return ControlStatus_Error; } // If the loop is not gated on a condition, and has only returns, we'll // reach this case. For example: // do { ... return; } while (); if (!current) return ControlStatus_Ended; // Otherwise, the loop is gated on a condition and/or has breaks so keep // parsing at the successor. pc = current->pc(); return ControlStatus_Joined; } IonBuilder::ControlStatus IonBuilder::finishLoop(CFGState& state, MBasicBlock* successor) { MOZ_ASSERT(current); MOZ_ASSERT(loopDepth_); loopDepth_--; MOZ_ASSERT_IF(successor, successor->loopDepth() == loopDepth_); // Compute phis in the loop header and propagate them throughout the loop, // including the successor. AbortReason r = state.loop.entry->setBackedge(alloc(), current); if (r == AbortReason_Alloc) return ControlStatus_Error; if (r == AbortReason_Disable) { // If there are types for variables on the backedge that were not // present at the original loop header, then uses of the variables' // phis may have generated incorrect nodes. The new types have been // incorporated into the header phis, so remove all blocks for the // loop body and restart with the new types. return restartLoop(state); } if (successor) { graph().moveBlockToEnd(successor); successor->inheritPhis(state.loop.entry); } if (state.loop.breaks) { // Propagate phis placed in the header to individual break exit points. DeferredEdge* edge = state.loop.breaks; while (edge) { edge->block->inheritPhis(state.loop.entry); edge = edge->next; } // Create a catch block to join all break exits. MBasicBlock* block = createBreakCatchBlock(state.loop.breaks, state.loop.exitpc); if (!block) return ControlStatus_Error; if (successor) { // Finally, create an unconditional edge from the successor to the // catch block. successor->end(MGoto::New(alloc(), block)); if (!block->addPredecessor(alloc(), successor)) return ControlStatus_Error; } successor = block; } if (!setCurrentAndSpecializePhis(successor)) return ControlStatus_Error; // An infinite loop (for (;;) { }) will not have a successor. if (!current) return ControlStatus_Ended; pc = current->pc(); return ControlStatus_Joined; } IonBuilder::ControlStatus IonBuilder::restartLoop(const CFGState& state) { AutoTraceLog logCompile(traceLogger(), TraceLogger_IonBuilderRestartLoop); spew("New types at loop header, restarting loop body"); if (JitOptions.limitScriptSize) { if (++numLoopRestarts_ >= MAX_LOOP_RESTARTS) return ControlStatus_Abort; } MBasicBlock* header = state.loop.entry; // Discard unreferenced & pre-allocated resume points. replaceMaybeFallbackFunctionGetter(nullptr); // Remove all blocks in the loop body other than the header, which has phis // of the appropriate type and incoming edges to preserve. graph().removeBlocksAfter(header); // Remove all instructions from the header itself, and all resume points // except the entry resume point. header->discardAllInstructions(); header->discardAllResumePoints(/* discardEntry = */ false); header->setStackDepth(header->getPredecessor(0)->stackDepth()); popCfgStack(); loopDepth_++; // Keep a local copy for these pointers since state will be overwritten in // pushLoop since state is a reference to cfgStack_.back() jsbytecode* condpc = state.loop.condpc; jsbytecode* updatepc = state.loop.updatepc; jsbytecode* updateEnd = state.loop.updateEnd; if (!pushLoop(state.loop.initialState, state.loop.initialStopAt, header, state.loop.osr, state.loop.loopHead, state.loop.initialPc, state.loop.bodyStart, state.loop.bodyEnd, state.loop.exitpc, state.loop.continuepc)) { return ControlStatus_Error; } CFGState& nstate = cfgStack_.back(); nstate.loop.condpc = condpc; nstate.loop.updatepc = updatepc; nstate.loop.updateEnd = updateEnd; // Don't specializePhis(), as the header has been visited before and the // phis have already had their type set. setCurrent(header); if (!jsop_loophead(nstate.loop.loopHead)) return ControlStatus_Error; pc = nstate.loop.initialPc; return ControlStatus_Jumped; } IonBuilder::ControlStatus IonBuilder::processDoWhileBodyEnd(CFGState& state) { if (!processDeferredContinues(state)) return ControlStatus_Error; // No current means control flow cannot reach the condition, so this will // never loop. if (!current) return processBrokenLoop(state); MBasicBlock* header = newBlock(current, state.loop.updatepc); if (!header) return ControlStatus_Error; current->end(MGoto::New(alloc(), header)); state.state = CFGState::DO_WHILE_LOOP_COND; state.stopAt = state.loop.updateEnd; pc = state.loop.updatepc; if (!setCurrentAndSpecializePhis(header)) return ControlStatus_Error; return ControlStatus_Jumped; } IonBuilder::ControlStatus IonBuilder::processDoWhileCondEnd(CFGState& state) { MOZ_ASSERT(JSOp(*pc) == JSOP_IFNE); // We're guaranteed a |current|, it's impossible to break or return from // inside the conditional expression. MOZ_ASSERT(current); // Pop the last value, and create the successor block. MDefinition* vins = current->pop(); MBasicBlock* successor = newBlock(current, GetNextPc(pc), loopDepth_ - 1); if (!successor) return ControlStatus_Error; // Test for do {} while(false) and don't create a loop in that case. if (MConstant* vinsConst = vins->maybeConstantValue()) { bool b; if (vinsConst->valueToBoolean(&b) && !b) { current->end(MGoto::New(alloc(), successor)); current = nullptr; state.loop.successor = successor; return processBrokenLoop(state); } } // Create the test instruction and end the current block. MTest* test = newTest(vins, state.loop.entry, successor); current->end(test); return finishLoop(state, successor); } IonBuilder::ControlStatus IonBuilder::processWhileCondEnd(CFGState& state) { MOZ_ASSERT(JSOp(*pc) == JSOP_IFNE || JSOp(*pc) == JSOP_IFEQ); // Balance the stack past the IFNE. MDefinition* ins = current->pop(); // Create the body and successor blocks. MBasicBlock* body = newBlock(current, state.loop.bodyStart); state.loop.successor = newBlock(current, state.loop.exitpc, loopDepth_ - 1); if (!body || !state.loop.successor) return ControlStatus_Error; MTest* test; if (JSOp(*pc) == JSOP_IFNE) test = newTest(ins, body, state.loop.successor); else test = newTest(ins, state.loop.successor, body); current->end(test); state.state = CFGState::WHILE_LOOP_BODY; state.stopAt = state.loop.bodyEnd; pc = state.loop.bodyStart; if (!setCurrentAndSpecializePhis(body)) return ControlStatus_Error; // Filter the types in the loop body. if (!improveTypesAtTest(test->getOperand(0), test->ifTrue() == current, test)) return ControlStatus_Error; // If this is a for-in loop, unbox the current value as string if possible. if (ins->isIsNoIter()) { MIteratorMore* iterMore = ins->toIsNoIter()->input()->toIteratorMore(); jsbytecode* iterMorePc = iterMore->resumePoint()->pc(); MOZ_ASSERT(*iterMorePc == JSOP_MOREITER); if (!nonStringIteration_ && !inspector->hasSeenNonStringIterMore(iterMorePc)) { MDefinition* val = current->peek(-1); MOZ_ASSERT(val == iterMore); MInstruction* ins = MUnbox::New(alloc(), val, MIRType::String, MUnbox::Fallible, Bailout_NonStringInputInvalidate); current->add(ins); current->rewriteAtDepth(-1, ins); } } return ControlStatus_Jumped; } IonBuilder::ControlStatus IonBuilder::processWhileBodyEnd(CFGState& state) { if (!processDeferredContinues(state)) return ControlStatus_Error; if (!current) return processBrokenLoop(state); current->end(MGoto::New(alloc(), state.loop.entry)); return finishLoop(state, state.loop.successor); } IonBuilder::ControlStatus IonBuilder::processForCondEnd(CFGState& state) { MOZ_ASSERT(JSOp(*pc) == JSOP_IFNE); // Balance the stack past the IFNE. MDefinition* ins = current->pop(); // Create the body and successor blocks. MBasicBlock* body = newBlock(current, state.loop.bodyStart); state.loop.successor = newBlock(current, state.loop.exitpc, loopDepth_ - 1); if (!body || !state.loop.successor) return ControlStatus_Error; MTest* test = newTest(ins, body, state.loop.successor); current->end(test); state.state = CFGState::FOR_LOOP_BODY; state.stopAt = state.loop.bodyEnd; pc = state.loop.bodyStart; if (!setCurrentAndSpecializePhis(body)) return ControlStatus_Error; return ControlStatus_Jumped; } IonBuilder::ControlStatus IonBuilder::processForBodyEnd(CFGState& state) { if (!processDeferredContinues(state)) return ControlStatus_Error; // If there is no updatepc, just go right to processing what would be the // end of the update clause. Otherwise, |current| might be nullptr; if this is // the case, the udpate is unreachable anyway. if (!state.loop.updatepc || !current) return processForUpdateEnd(state); pc = state.loop.updatepc; state.state = CFGState::FOR_LOOP_UPDATE; state.stopAt = state.loop.updateEnd; return ControlStatus_Jumped; } IonBuilder::ControlStatus IonBuilder::processForUpdateEnd(CFGState& state) { // If there is no current, we couldn't reach the loop edge and there was no // update clause. if (!current) return processBrokenLoop(state); current->end(MGoto::New(alloc(), state.loop.entry)); return finishLoop(state, state.loop.successor); } IonBuilder::DeferredEdge* IonBuilder::filterDeadDeferredEdges(DeferredEdge* edge) { DeferredEdge* head = edge; DeferredEdge* prev = nullptr; while (edge) { if (edge->block->isDead()) { if (prev) prev->next = edge->next; else head = edge->next; } else { prev = edge; } edge = edge->next; } // There must be at least one deferred edge from a block that was not // deleted; blocks are deleted when restarting processing of a loop, and // the final version of the loop body will have edges from live blocks. MOZ_ASSERT(head); return head; } bool IonBuilder::processDeferredContinues(CFGState& state) { // If there are any continues for this loop, and there is an update block, // then we need to create a new basic block to house the update. if (state.loop.continues) { DeferredEdge* edge = filterDeadDeferredEdges(state.loop.continues); MBasicBlock* update = newBlock(edge->block, loops_.back().continuepc); if (!update) return false; if (current) { current->end(MGoto::New(alloc(), update)); if (!update->addPredecessor(alloc(), current)) return false; } // No need to use addPredecessor for first edge, // because it is already predecessor. edge->block->end(MGoto::New(alloc(), update)); edge = edge->next; // Remaining edges while (edge) { edge->block->end(MGoto::New(alloc(), update)); if (!update->addPredecessor(alloc(), edge->block)) return false; edge = edge->next; } state.loop.continues = nullptr; if (!setCurrentAndSpecializePhis(update)) return ControlStatus_Error; } return true; } MBasicBlock* IonBuilder::createBreakCatchBlock(DeferredEdge* edge, jsbytecode* pc) { edge = filterDeadDeferredEdges(edge); // Create block, using the first break statement as predecessor MBasicBlock* successor = newBlock(edge->block, pc); if (!successor) return nullptr; // No need to use addPredecessor for first edge, // because it is already predecessor. edge->block->end(MGoto::New(alloc(), successor)); edge = edge->next; // Finish up remaining breaks. while (edge) { MGoto* brk = MGoto::New(alloc().fallible(), successor); if (!brk) return nullptr; edge->block->end(brk); if (!successor->addPredecessor(alloc(), edge->block)) return nullptr; edge = edge->next; } return successor; } IonBuilder::ControlStatus IonBuilder::processNextTableSwitchCase(CFGState& state) { MOZ_ASSERT(state.state == CFGState::TABLE_SWITCH); state.tableswitch.currentBlock++; // Test if there are still unprocessed successors (cases/default) if (state.tableswitch.currentBlock >= state.tableswitch.ins->numBlocks()) return processSwitchEnd(state.tableswitch.breaks, state.tableswitch.exitpc); // Get the next successor MBasicBlock* successor = state.tableswitch.ins->getBlock(state.tableswitch.currentBlock); // Add current block as predecessor if available. // This means the previous case didn't have a break statement. // So flow will continue in this block. if (current) { current->end(MGoto::New(alloc(), successor)); if (!successor->addPredecessor(alloc(), current)) return ControlStatus_Error; } else { // If this is an actual case statement, optimize by replacing the // input to the switch case with the actual number of the case. // This constant has been emitted when creating the case blocks. if (state.tableswitch.ins->getDefault() != successor) { MConstant* constant = successor->begin()->toConstant(); for (uint32_t j = 0; j < successor->stackDepth(); j++) { MDefinition* ins = successor->getSlot(j); if (ins != state.tableswitch.ins->getOperand(0)) continue; constant->setDependency(state.tableswitch.ins); successor->setSlot(j, constant); } } } // Insert successor after the current block, to maintain RPO. graph().moveBlockToEnd(successor); // If this is the last successor the block should stop at the end of the tableswitch // Else it should stop at the start of the next successor if (state.tableswitch.currentBlock+1 < state.tableswitch.ins->numBlocks()) state.stopAt = state.tableswitch.ins->getBlock(state.tableswitch.currentBlock+1)->pc(); else state.stopAt = state.tableswitch.exitpc; if (!setCurrentAndSpecializePhis(successor)) return ControlStatus_Error; pc = current->pc(); return ControlStatus_Jumped; } IonBuilder::ControlStatus IonBuilder::processAndOrEnd(CFGState& state) { MOZ_ASSERT(current); MBasicBlock* lhs = state.branch.ifFalse; // Create a new block to represent the join. MBasicBlock* join = newBlock(current, state.stopAt); if (!join) return ControlStatus_Error; // End the rhs. current->end(MGoto::New(alloc(), join)); // End the lhs. lhs->end(MGoto::New(alloc(), join)); if (!join->addPredecessor(alloc(), state.branch.ifFalse)) return ControlStatus_Error; // Set the join path as current path. if (!setCurrentAndSpecializePhis(join)) return ControlStatus_Error; pc = current->pc(); return ControlStatus_Joined; } IonBuilder::ControlStatus IonBuilder::processLabelEnd(CFGState& state) { MOZ_ASSERT(state.state == CFGState::LABEL); // If there are no breaks and no current, controlflow is terminated. if (!state.label.breaks && !current) return ControlStatus_Ended; // If there are no breaks to this label, there's nothing to do. if (!state.label.breaks) return ControlStatus_Joined; MBasicBlock* successor = createBreakCatchBlock(state.label.breaks, state.stopAt); if (!successor) return ControlStatus_Error; if (current) { current->end(MGoto::New(alloc(), successor)); if (!successor->addPredecessor(alloc(), current)) return ControlStatus_Error; } pc = state.stopAt; if (!setCurrentAndSpecializePhis(successor)) return ControlStatus_Error; return ControlStatus_Joined; } IonBuilder::ControlStatus IonBuilder::processTryEnd(CFGState& state) { MOZ_ASSERT(state.state == CFGState::TRY); if (!state.try_.successor) { MOZ_ASSERT(!current); return ControlStatus_Ended; } if (current) { current->end(MGoto::New(alloc(), state.try_.successor)); if (!state.try_.successor->addPredecessor(alloc(), current)) return ControlStatus_Error; } // Start parsing the code after this try-catch statement. if (!setCurrentAndSpecializePhis(state.try_.successor)) return ControlStatus_Error; graph().moveBlockToEnd(current); pc = current->pc(); return ControlStatus_Joined; } IonBuilder::ControlStatus IonBuilder::processBreak(JSOp op, jssrcnote* sn) { MOZ_ASSERT(op == JSOP_GOTO); MOZ_ASSERT(SN_TYPE(sn) == SRC_BREAK || SN_TYPE(sn) == SRC_BREAK2LABEL); // Find the break target. jsbytecode* target = pc + GetJumpOffset(pc); DebugOnly found = false; if (SN_TYPE(sn) == SRC_BREAK2LABEL) { for (size_t i = labels_.length() - 1; i < labels_.length(); i--) { CFGState& cfg = cfgStack_[labels_[i].cfgEntry]; MOZ_ASSERT(cfg.state == CFGState::LABEL); if (cfg.stopAt == target) { cfg.label.breaks = new(alloc()) DeferredEdge(current, cfg.label.breaks); found = true; break; } } } else { for (size_t i = loops_.length() - 1; i < loops_.length(); i--) { CFGState& cfg = cfgStack_[loops_[i].cfgEntry]; MOZ_ASSERT(cfg.isLoop()); if (cfg.loop.exitpc == target) { cfg.loop.breaks = new(alloc()) DeferredEdge(current, cfg.loop.breaks); found = true; break; } } } MOZ_ASSERT(found); setCurrent(nullptr); pc += CodeSpec[op].length; return processControlEnd(); } static inline jsbytecode* EffectiveContinue(jsbytecode* pc) { if (JSOp(*pc) == JSOP_GOTO) return pc + GetJumpOffset(pc); return pc; } IonBuilder::ControlStatus IonBuilder::processContinue(JSOp op) { MOZ_ASSERT(op == JSOP_GOTO); // Find the target loop. CFGState* found = nullptr; jsbytecode* target = pc + GetJumpOffset(pc); for (size_t i = loops_.length() - 1; i < loops_.length(); i--) { // +1 to skip JSOP_JUMPTARGET. if (loops_[i].continuepc == target + 1 || EffectiveContinue(loops_[i].continuepc) == target) { found = &cfgStack_[loops_[i].cfgEntry]; break; } } // There must always be a valid target loop structure. If not, there's // probably an off-by-something error in which pc we track. MOZ_ASSERT(found); CFGState& state = *found; state.loop.continues = new(alloc()) DeferredEdge(current, state.loop.continues); setCurrent(nullptr); pc += CodeSpec[op].length; return processControlEnd(); } IonBuilder::ControlStatus IonBuilder::processSwitchBreak(JSOp op) { MOZ_ASSERT(op == JSOP_GOTO); // Find the target switch. CFGState* found = nullptr; jsbytecode* target = pc + GetJumpOffset(pc); for (size_t i = switches_.length() - 1; i < switches_.length(); i--) { if (switches_[i].continuepc == target) { found = &cfgStack_[switches_[i].cfgEntry]; break; } } // There must always be a valid target loop structure. If not, there's // probably an off-by-something error in which pc we track. MOZ_ASSERT(found); CFGState& state = *found; DeferredEdge** breaks = nullptr; switch (state.state) { case CFGState::TABLE_SWITCH: breaks = &state.tableswitch.breaks; break; case CFGState::COND_SWITCH_BODY: breaks = &state.condswitch.breaks; break; default: MOZ_CRASH("Unexpected switch state."); } *breaks = new(alloc()) DeferredEdge(current, *breaks); setCurrent(nullptr); pc += CodeSpec[op].length; return processControlEnd(); } IonBuilder::ControlStatus IonBuilder::processSwitchEnd(DeferredEdge* breaks, jsbytecode* exitpc) { // No break statements, no current. // This means that control flow is cut-off from this point // (e.g. all cases have return statements). if (!breaks && !current) return ControlStatus_Ended; // Create successor block. // If there are breaks, create block with breaks as predecessor // Else create a block with current as predecessor MBasicBlock* successor = nullptr; if (breaks) successor = createBreakCatchBlock(breaks, exitpc); else successor = newBlock(current, exitpc); if (!successor) return ControlStatus_Error; // If there is current, the current block flows into this one. // So current is also a predecessor to this block if (current) { current->end(MGoto::New(alloc(), successor)); if (breaks) { if (!successor->addPredecessor(alloc(), current)) return ControlStatus_Error; } } pc = exitpc; if (!setCurrentAndSpecializePhis(successor)) return ControlStatus_Error; return ControlStatus_Joined; } IonBuilder::ControlStatus IonBuilder::maybeLoop(JSOp op, jssrcnote* sn) { // This function looks at the opcode and source note and tries to // determine the structure of the loop. For some opcodes, like // POP/NOP which are not explicitly control flow, this source note is // optional. For opcodes with control flow, like GOTO, an unrecognized // or not-present source note is a compilation failure. switch (op) { case JSOP_POP: // for (init; ; update?) ... if (sn && SN_TYPE(sn) == SRC_FOR) { current->pop(); return forLoop(op, sn); } break; case JSOP_NOP: if (sn) { // do { } while (cond) if (SN_TYPE(sn) == SRC_WHILE) return doWhileLoop(op, sn); // Build a mapping such that given a basic block, whose successor // has a phi // for (; ; update?) if (SN_TYPE(sn) == SRC_FOR) return forLoop(op, sn); } break; default: MOZ_CRASH("unexpected opcode"); } return ControlStatus_None; } void IonBuilder::assertValidLoopHeadOp(jsbytecode* pc) { #ifdef DEBUG MOZ_ASSERT(JSOp(*pc) == JSOP_LOOPHEAD); // Make sure this is the next opcode after the loop header, // unless the for loop is unconditional. CFGState& state = cfgStack_.back(); MOZ_ASSERT_IF((JSOp)*(state.loop.entry->pc()) == JSOP_GOTO, GetNextPc(state.loop.entry->pc()) == pc); // do-while loops have a source note. jssrcnote* sn = info().getNote(gsn, pc); if (sn) { jsbytecode* ifne = pc + GetSrcNoteOffset(sn, 0); jsbytecode* expected_ifne; switch (state.state) { case CFGState::DO_WHILE_LOOP_BODY: expected_ifne = state.loop.updateEnd; break; default: MOZ_CRASH("JSOP_LOOPHEAD unexpected source note"); } // Make sure this loop goes to the same ifne as the loop header's // source notes or GOTO. MOZ_ASSERT(ifne == expected_ifne); } else { MOZ_ASSERT(state.state != CFGState::DO_WHILE_LOOP_BODY); } #endif } IonBuilder::ControlStatus IonBuilder::doWhileLoop(JSOp op, jssrcnote* sn) { // do { } while() loops have the following structure: // NOP ; SRC_WHILE (offset to COND) // LOOPHEAD ; SRC_WHILE (offset to IFNE) // LOOPENTRY // ... ; body // ... // COND ; start of condition // ... // IFNE -> ; goes to LOOPHEAD int condition_offset = GetSrcNoteOffset(sn, 0); jsbytecode* conditionpc = pc + condition_offset; jssrcnote* sn2 = info().getNote(gsn, pc+1); int offset = GetSrcNoteOffset(sn2, 0); jsbytecode* ifne = pc + offset + 1; MOZ_ASSERT(ifne > pc); // Verify that the IFNE goes back to a loophead op. jsbytecode* loopHead = GetNextPc(pc); MOZ_ASSERT(JSOp(*loopHead) == JSOP_LOOPHEAD); MOZ_ASSERT(loopHead == ifne + GetJumpOffset(ifne)); jsbytecode* loopEntry = GetNextPc(loopHead); bool canOsr = LoopEntryCanIonOsr(loopEntry); bool osr = info().hasOsrAt(loopEntry); if (osr) { MBasicBlock* preheader = newOsrPreheader(current, loopEntry, pc); if (!preheader) return ControlStatus_Error; current->end(MGoto::New(alloc(), preheader)); if (!setCurrentAndSpecializePhis(preheader)) return ControlStatus_Error; } unsigned stackPhiCount = 0; MBasicBlock* header = newPendingLoopHeader(current, loopEntry, osr, canOsr, stackPhiCount); if (!header) return ControlStatus_Error; current->end(MGoto::New(alloc(), header)); jsbytecode* loophead = GetNextPc(pc); jsbytecode* bodyStart = GetNextPc(loophead); jsbytecode* bodyEnd = conditionpc; jsbytecode* exitpc = GetNextPc(ifne); if (!analyzeNewLoopTypes(header, bodyStart, exitpc)) return ControlStatus_Error; if (!pushLoop(CFGState::DO_WHILE_LOOP_BODY, conditionpc, header, osr, loopHead, bodyStart, bodyStart, bodyEnd, exitpc, conditionpc)) { return ControlStatus_Error; } CFGState& state = cfgStack_.back(); state.loop.updatepc = conditionpc; state.loop.updateEnd = ifne; if (!setCurrentAndSpecializePhis(header)) return ControlStatus_Error; if (!jsop_loophead(loophead)) return ControlStatus_Error; pc = bodyStart; return ControlStatus_Jumped; } IonBuilder::ControlStatus IonBuilder::whileOrForInLoop(jssrcnote* sn) { // while (cond) { } loops have the following structure: // GOTO cond ; SRC_WHILE (offset to IFNE) // LOOPHEAD // ... // cond: // LOOPENTRY // ... // IFNE ; goes to LOOPHEAD // for (x in y) { } loops are similar; the cond will be a MOREITER. MOZ_ASSERT(SN_TYPE(sn) == SRC_FOR_OF || SN_TYPE(sn) == SRC_FOR_IN || SN_TYPE(sn) == SRC_WHILE); int ifneOffset = GetSrcNoteOffset(sn, 0); jsbytecode* ifne = pc + ifneOffset; MOZ_ASSERT(ifne > pc); // Verify that the IFNE goes back to a loophead op. MOZ_ASSERT(JSOp(*GetNextPc(pc)) == JSOP_LOOPHEAD); MOZ_ASSERT(GetNextPc(pc) == ifne + GetJumpOffset(ifne)); jsbytecode* loopEntry = pc + GetJumpOffset(pc); bool canOsr = LoopEntryCanIonOsr(loopEntry); bool osr = info().hasOsrAt(loopEntry); if (osr) { MBasicBlock* preheader = newOsrPreheader(current, loopEntry, pc); if (!preheader) return ControlStatus_Error; current->end(MGoto::New(alloc(), preheader)); if (!setCurrentAndSpecializePhis(preheader)) return ControlStatus_Error; } unsigned stackPhiCount; if (SN_TYPE(sn) == SRC_FOR_OF) stackPhiCount = 2; else if (SN_TYPE(sn) == SRC_FOR_IN) stackPhiCount = 1; else stackPhiCount = 0; MBasicBlock* header = newPendingLoopHeader(current, loopEntry, osr, canOsr, stackPhiCount); if (!header) return ControlStatus_Error; current->end(MGoto::New(alloc(), header)); // Skip past the JSOP_LOOPHEAD for the body start. jsbytecode* loopHead = GetNextPc(pc); jsbytecode* bodyStart = GetNextPc(loopHead); jsbytecode* bodyEnd = pc + GetJumpOffset(pc); jsbytecode* exitpc = GetNextPc(ifne); jsbytecode* continuepc = pc; if (!analyzeNewLoopTypes(header, bodyStart, exitpc)) return ControlStatus_Error; if (!pushLoop(CFGState::WHILE_LOOP_COND, ifne, header, osr, loopHead, bodyEnd, bodyStart, bodyEnd, exitpc, continuepc)) { return ControlStatus_Error; } // Parse the condition first. if (!setCurrentAndSpecializePhis(header)) return ControlStatus_Error; if (!jsop_loophead(loopHead)) return ControlStatus_Error; pc = bodyEnd; return ControlStatus_Jumped; } IonBuilder::ControlStatus IonBuilder::forLoop(JSOp op, jssrcnote* sn) { // Skip the NOP. MOZ_ASSERT(op == JSOP_NOP); pc = GetNextPc(pc); jsbytecode* condpc = pc + GetSrcNoteOffset(sn, 0); jsbytecode* updatepc = pc + GetSrcNoteOffset(sn, 1); jsbytecode* ifne = pc + GetSrcNoteOffset(sn, 2); jsbytecode* exitpc = GetNextPc(ifne); // for loops have the following structures: // // NOP or POP // [GOTO cond | NOP] // LOOPHEAD // body: // ; [body] // [increment:] // [{FRESHEN,RECREATE}LEXICALENV, if needed by a lexical env] // ; [increment] // [cond:] // LOOPENTRY // GOTO body // // If there is a condition (condpc != ifne), this acts similar to a while // loop otherwise, it acts like a do-while loop. // // Note that currently Ion doesn't compile pushlexicalenv/poplexicalenv, // necessary prerequisites to {freshen,recreate}lexicalenv. So the code // below doesn't and needn't consider either op's implications. jsbytecode* bodyStart = pc; jsbytecode* bodyEnd = updatepc; jsbytecode* loopEntry = condpc; if (condpc != ifne) { MOZ_ASSERT(JSOp(*bodyStart) == JSOP_GOTO); MOZ_ASSERT(bodyStart + GetJumpOffset(bodyStart) == condpc); bodyStart = GetNextPc(bodyStart); } else { // No loop condition, such as for(j = 0; ; j++) if (op != JSOP_NOP) { // If the loop starts with POP, we have to skip a NOP. MOZ_ASSERT(JSOp(*bodyStart) == JSOP_NOP); bodyStart = GetNextPc(bodyStart); } loopEntry = GetNextPc(bodyStart); } jsbytecode* loopHead = bodyStart; MOZ_ASSERT(JSOp(*bodyStart) == JSOP_LOOPHEAD); MOZ_ASSERT(ifne + GetJumpOffset(ifne) == bodyStart); bodyStart = GetNextPc(bodyStart); bool osr = info().hasOsrAt(loopEntry); bool canOsr = LoopEntryCanIonOsr(loopEntry); if (osr) { MBasicBlock* preheader = newOsrPreheader(current, loopEntry, pc); if (!preheader) return ControlStatus_Error; current->end(MGoto::New(alloc(), preheader)); if (!setCurrentAndSpecializePhis(preheader)) return ControlStatus_Error; } unsigned stackPhiCount = 0; MBasicBlock* header = newPendingLoopHeader(current, loopEntry, osr, canOsr, stackPhiCount); if (!header) return ControlStatus_Error; current->end(MGoto::New(alloc(), header)); // If there is no condition, we immediately parse the body. Otherwise, we // parse the condition. jsbytecode* stopAt; CFGState::State initial; if (condpc != ifne) { pc = condpc; stopAt = ifne; initial = CFGState::FOR_LOOP_COND; } else { pc = bodyStart; stopAt = bodyEnd; initial = CFGState::FOR_LOOP_BODY; } if (!analyzeNewLoopTypes(header, bodyStart, exitpc)) return ControlStatus_Error; if (!pushLoop(initial, stopAt, header, osr, loopHead, pc, bodyStart, bodyEnd, exitpc, updatepc)) { return ControlStatus_Error; } CFGState& state = cfgStack_.back(); state.loop.condpc = (condpc != ifne) ? condpc : nullptr; state.loop.updatepc = (updatepc != condpc) ? updatepc : nullptr; if (state.loop.updatepc) state.loop.updateEnd = condpc; if (!setCurrentAndSpecializePhis(header)) return ControlStatus_Error; if (!jsop_loophead(loopHead)) return ControlStatus_Error; return ControlStatus_Jumped; } int IonBuilder::CmpSuccessors(const void* a, const void* b) { const MBasicBlock* a0 = * (MBasicBlock * const*)a; const MBasicBlock* b0 = * (MBasicBlock * const*)b; if (a0->pc() == b0->pc()) return 0; return (a0->pc() > b0->pc()) ? 1 : -1; } IonBuilder::ControlStatus IonBuilder::tableSwitch(JSOp op, jssrcnote* sn) { // TableSwitch op contains the following data // (length between data is JUMP_OFFSET_LEN) // // 0: Offset of default case // 1: Lowest number in tableswitch // 2: Highest number in tableswitch // 3: Offset of case low // 4: Offset of case low+1 // .: ... // .: Offset of case high MOZ_ASSERT(op == JSOP_TABLESWITCH); MOZ_ASSERT(SN_TYPE(sn) == SRC_TABLESWITCH); // Pop input. MDefinition* ins = current->pop(); // Get the default and exit pc jsbytecode* exitpc = pc + GetSrcNoteOffset(sn, 0); jsbytecode* defaultpc = pc + GET_JUMP_OFFSET(pc); MOZ_ASSERT(defaultpc > pc && defaultpc <= exitpc); // Get the low and high from the tableswitch jsbytecode* pc2 = pc; pc2 += JUMP_OFFSET_LEN; int low = GET_JUMP_OFFSET(pc2); pc2 += JUMP_OFFSET_LEN; int high = GET_JUMP_OFFSET(pc2); pc2 += JUMP_OFFSET_LEN; // Create MIR instruction MTableSwitch* tableswitch = MTableSwitch::New(alloc(), ins, low, high); // Create default case MBasicBlock* defaultcase = newBlock(current, defaultpc); if (!defaultcase) return ControlStatus_Error; if (!tableswitch->addDefault(defaultcase)) return ControlStatus_Error; if (!tableswitch->addBlock(defaultcase)) return ControlStatus_Error; // Create cases jsbytecode* casepc = nullptr; for (int i = 0; i < high-low+1; i++) { casepc = pc + GET_JUMP_OFFSET(pc2); MOZ_ASSERT(casepc >= pc && casepc <= exitpc); MBasicBlock* caseblock; if (casepc == pc) { // If the casepc equals the current pc, it is not a written case, // but a filled gap. That way we can use a tableswitch instead of // condswitch, even if not all numbers are consecutive. // In that case this block goes to the default case caseblock = newBlock(current, defaultpc); if (!caseblock) return ControlStatus_Error; caseblock->end(MGoto::New(alloc(), defaultcase)); if (!defaultcase->addPredecessor(alloc(), caseblock)) return ControlStatus_Error; } else { // If this is an actual case (not filled gap), // add this block to the list that still needs to get processed. caseblock = newBlock(current, casepc); if (!caseblock) return ControlStatus_Error; if (!tableswitch->addBlock(caseblock)) return ControlStatus_Error; // Add constant to indicate which case this is for use by // processNextTableSwitchCase. MConstant* constant = MConstant::New(alloc(), Int32Value(i + low)); caseblock->add(constant); } size_t caseIndex; if (!tableswitch->addSuccessor(caseblock, &caseIndex)) return ControlStatus_Error; if (!tableswitch->addCase(caseIndex)) return ControlStatus_Error; pc2 += JUMP_OFFSET_LEN; } // Move defaultcase to the end, to maintain RPO. graph().moveBlockToEnd(defaultcase); MOZ_ASSERT(tableswitch->numCases() == (uint32_t)(high - low + 1)); MOZ_ASSERT(tableswitch->numSuccessors() > 0); // Sort the list of blocks that still needs to get processed by pc qsort(tableswitch->blocks(), tableswitch->numBlocks(), sizeof(MBasicBlock*), CmpSuccessors); // Create info ControlFlowInfo switchinfo(cfgStack_.length(), exitpc); if (!switches_.append(switchinfo)) return ControlStatus_Error; // Use a state to retrieve some information CFGState state = CFGState::TableSwitch(exitpc, tableswitch); // Save the MIR instruction as last instruction of this block. current->end(tableswitch); // If there is only one successor the block should stop at the end of the switch // Else it should stop at the start of the next successor if (tableswitch->numBlocks() > 1) state.stopAt = tableswitch->getBlock(1)->pc(); if (!setCurrentAndSpecializePhis(tableswitch->getBlock(0))) return ControlStatus_Error; if (!cfgStack_.append(state)) return ControlStatus_Error; pc = current->pc(); return ControlStatus_Jumped; } bool IonBuilder::replaceTypeSet(MDefinition* subject, TemporaryTypeSet* type, MTest* test) { if (type->unknown()) return true; // Don't emit MFilterTypeSet if it doesn't improve the typeset. if (subject->resultTypeSet()) { if (subject->resultTypeSet()->equals(type)) return true; } else { TemporaryTypeSet oldTypes(alloc_->lifoAlloc(), subject->type()); if (oldTypes.equals(type)) return true; } MInstruction* replace = nullptr; MDefinition* ins; for (uint32_t i = 0; i < current->stackDepth(); i++) { ins = current->getSlot(i); // Instead of creating a new MFilterTypeSet, try to update the old one. if (ins->isFilterTypeSet() && ins->getOperand(0) == subject && ins->dependency() == test) { TemporaryTypeSet* intersect = TypeSet::intersectSets(ins->resultTypeSet(), type, alloc_->lifoAlloc()); if (!intersect) return false; ins->toFilterTypeSet()->setResultType(intersect->getKnownMIRType()); ins->toFilterTypeSet()->setResultTypeSet(intersect); if (ins->type() == MIRType::Undefined) current->setSlot(i, constant(UndefinedValue())); else if (ins->type() == MIRType::Null) current->setSlot(i, constant(NullValue())); else if (ins->type() == MIRType::MagicOptimizedArguments) current->setSlot(i, constant(MagicValue(JS_OPTIMIZED_ARGUMENTS))); else MOZ_ASSERT(!IsMagicType(ins->type())); continue; } if (ins == subject) { if (!replace) { replace = MFilterTypeSet::New(alloc(), subject, type); if (!replace) return false; current->add(replace); // Make sure we don't hoist it above the MTest, we can use the // 'dependency' of an MInstruction. This is normally used by // Alias Analysis, but won't get overwritten, since this // instruction doesn't have an AliasSet. replace->setDependency(test); if (replace->type() == MIRType::Undefined) replace = constant(UndefinedValue()); else if (replace->type() == MIRType::Null) replace = constant(NullValue()); else if (replace->type() == MIRType::MagicOptimizedArguments) replace = constant(MagicValue(JS_OPTIMIZED_ARGUMENTS)); else MOZ_ASSERT(!IsMagicType(ins->type())); } current->setSlot(i, replace); } } return true; } bool IonBuilder::detectAndOrStructure(MPhi* ins, bool* branchIsAnd) { // Look for a triangle pattern: // // initialBlock // / | // branchBlock | // \ | // testBlock // // Where ins is a phi from testBlock which combines two values // pushed onto the stack by initialBlock and branchBlock. if (ins->numOperands() != 2) return false; MBasicBlock* testBlock = ins->block(); MOZ_ASSERT(testBlock->numPredecessors() == 2); MBasicBlock* initialBlock; MBasicBlock* branchBlock; if (testBlock->getPredecessor(0)->lastIns()->isTest()) { initialBlock = testBlock->getPredecessor(0); branchBlock = testBlock->getPredecessor(1); } else if (testBlock->getPredecessor(1)->lastIns()->isTest()) { initialBlock = testBlock->getPredecessor(1); branchBlock = testBlock->getPredecessor(0); } else { return false; } if (branchBlock->numSuccessors() != 1) return false; if (branchBlock->numPredecessors() != 1 || branchBlock->getPredecessor(0) != initialBlock) return false; if (initialBlock->numSuccessors() != 2) return false; MDefinition* branchResult = ins->getOperand(testBlock->indexForPredecessor(branchBlock)); MDefinition* initialResult = ins->getOperand(testBlock->indexForPredecessor(initialBlock)); if (branchBlock->stackDepth() != initialBlock->stackDepth()) return false; if (branchBlock->stackDepth() != testBlock->stackDepth() + 1) return false; if (branchResult != branchBlock->peek(-1) || initialResult != initialBlock->peek(-1)) return false; MTest* initialTest = initialBlock->lastIns()->toTest(); bool branchIsTrue = branchBlock == initialTest->ifTrue(); if (initialTest->input() == ins->getOperand(0)) *branchIsAnd = branchIsTrue != (testBlock->getPredecessor(0) == branchBlock); else if (initialTest->input() == ins->getOperand(1)) *branchIsAnd = branchIsTrue != (testBlock->getPredecessor(1) == branchBlock); else return false; return true; } bool IonBuilder::improveTypesAtCompare(MCompare* ins, bool trueBranch, MTest* test) { if (ins->compareType() == MCompare::Compare_Undefined || ins->compareType() == MCompare::Compare_Null) { return improveTypesAtNullOrUndefinedCompare(ins, trueBranch, test); } if ((ins->lhs()->isTypeOf() || ins->rhs()->isTypeOf()) && (ins->lhs()->isConstant() || ins->rhs()->isConstant())) { return improveTypesAtTypeOfCompare(ins, trueBranch, test); } return true; } bool IonBuilder::improveTypesAtTypeOfCompare(MCompare* ins, bool trueBranch, MTest* test) { MTypeOf* typeOf = ins->lhs()->isTypeOf() ? ins->lhs()->toTypeOf() : ins->rhs()->toTypeOf(); MConstant* constant = ins->lhs()->isConstant() ? ins->lhs()->toConstant() : ins->rhs()->toConstant(); if (constant->type() != MIRType::String) return true; bool equal = ins->jsop() == JSOP_EQ || ins->jsop() == JSOP_STRICTEQ; bool notEqual = ins->jsop() == JSOP_NE || ins->jsop() == JSOP_STRICTNE; if (notEqual) trueBranch = !trueBranch; // Relational compares not supported. if (!equal && !notEqual) return true; MDefinition* subject = typeOf->input(); TemporaryTypeSet* inputTypes = subject->resultTypeSet(); // Create temporary typeset equal to the type if there is no resultTypeSet. TemporaryTypeSet tmp; if (!inputTypes) { if (subject->type() == MIRType::Value) return true; inputTypes = &tmp; tmp.addType(TypeSet::PrimitiveType(ValueTypeFromMIRType(subject->type())), alloc_->lifoAlloc()); } if (inputTypes->unknown()) return true; // Note: we cannot remove the AnyObject type in the false branch, // since there are multiple ways to get an object. That is the reason // for the 'trueBranch' test. TemporaryTypeSet filter; const JSAtomState& names = GetJitContext()->runtime->names(); if (constant->toString() == TypeName(JSTYPE_VOID, names)) { filter.addType(TypeSet::UndefinedType(), alloc_->lifoAlloc()); if (typeOf->inputMaybeCallableOrEmulatesUndefined() && trueBranch) filter.addType(TypeSet::AnyObjectType(), alloc_->lifoAlloc()); } else if (constant->toString() == TypeName(JSTYPE_BOOLEAN, names)) { filter.addType(TypeSet::BooleanType(), alloc_->lifoAlloc()); } else if (constant->toString() == TypeName(JSTYPE_NUMBER, names)) { filter.addType(TypeSet::Int32Type(), alloc_->lifoAlloc()); filter.addType(TypeSet::DoubleType(), alloc_->lifoAlloc()); } else if (constant->toString() == TypeName(JSTYPE_STRING, names)) { filter.addType(TypeSet::StringType(), alloc_->lifoAlloc()); } else if (constant->toString() == TypeName(JSTYPE_SYMBOL, names)) { filter.addType(TypeSet::SymbolType(), alloc_->lifoAlloc()); } else if (constant->toString() == TypeName(JSTYPE_OBJECT, names)) { filter.addType(TypeSet::NullType(), alloc_->lifoAlloc()); if (trueBranch) filter.addType(TypeSet::AnyObjectType(), alloc_->lifoAlloc()); } else if (constant->toString() == TypeName(JSTYPE_FUNCTION, names)) { if (typeOf->inputMaybeCallableOrEmulatesUndefined() && trueBranch) filter.addType(TypeSet::AnyObjectType(), alloc_->lifoAlloc()); } else { return true; } TemporaryTypeSet* type; if (trueBranch) type = TypeSet::intersectSets(&filter, inputTypes, alloc_->lifoAlloc()); else type = TypeSet::removeSet(inputTypes, &filter, alloc_->lifoAlloc()); if (!type) return false; return replaceTypeSet(subject, type, test); } bool IonBuilder::improveTypesAtNullOrUndefinedCompare(MCompare* ins, bool trueBranch, MTest* test) { MOZ_ASSERT(ins->compareType() == MCompare::Compare_Undefined || ins->compareType() == MCompare::Compare_Null); // altersUndefined/Null represents if we can filter/set Undefined/Null. bool altersUndefined, altersNull; JSOp op = ins->jsop(); switch(op) { case JSOP_STRICTNE: case JSOP_STRICTEQ: altersUndefined = ins->compareType() == MCompare::Compare_Undefined; altersNull = ins->compareType() == MCompare::Compare_Null; break; case JSOP_NE: case JSOP_EQ: altersUndefined = altersNull = true; break; default: MOZ_CRASH("Relational compares not supported"); } MDefinition* subject = ins->lhs(); TemporaryTypeSet* inputTypes = subject->resultTypeSet(); MOZ_ASSERT(IsNullOrUndefined(ins->rhs()->type())); // Create temporary typeset equal to the type if there is no resultTypeSet. TemporaryTypeSet tmp; if (!inputTypes) { if (subject->type() == MIRType::Value) return true; inputTypes = &tmp; tmp.addType(TypeSet::PrimitiveType(ValueTypeFromMIRType(subject->type())), alloc_->lifoAlloc()); } if (inputTypes->unknown()) return true; TemporaryTypeSet* type; // Decide if we need to filter the type or set it. if ((op == JSOP_STRICTEQ || op == JSOP_EQ) ^ trueBranch) { // Remove undefined/null TemporaryTypeSet remove; if (altersUndefined) remove.addType(TypeSet::UndefinedType(), alloc_->lifoAlloc()); if (altersNull) remove.addType(TypeSet::NullType(), alloc_->lifoAlloc()); type = TypeSet::removeSet(inputTypes, &remove, alloc_->lifoAlloc()); } else { // Set undefined/null. TemporaryTypeSet base; if (altersUndefined) { base.addType(TypeSet::UndefinedType(), alloc_->lifoAlloc()); // If TypeSet emulates undefined, then we cannot filter the objects. if (inputTypes->maybeEmulatesUndefined(constraints())) base.addType(TypeSet::AnyObjectType(), alloc_->lifoAlloc()); } if (altersNull) base.addType(TypeSet::NullType(), alloc_->lifoAlloc()); type = TypeSet::intersectSets(&base, inputTypes, alloc_->lifoAlloc()); } if (!type) return false; return replaceTypeSet(subject, type, test); } bool IonBuilder::improveTypesAtTest(MDefinition* ins, bool trueBranch, MTest* test) { // We explore the test condition to try and deduce as much type information // as possible. // All branches of this switch that don't want to fall through to the // default behavior must return. The default behavior assumes that a true // test means the incoming ins is not null or undefined and that a false // tests means it's one of null, undefined, false, 0, "", and objects // emulating undefined switch (ins->op()) { case MDefinition::Op_Not: return improveTypesAtTest(ins->toNot()->getOperand(0), !trueBranch, test); case MDefinition::Op_IsObject: { MDefinition* subject = ins->getOperand(0); TemporaryTypeSet* oldType = subject->resultTypeSet(); // Create temporary typeset equal to the type if there is no resultTypeSet. TemporaryTypeSet tmp; if (!oldType) { if (subject->type() == MIRType::Value) return true; oldType = &tmp; tmp.addType(TypeSet::PrimitiveType(ValueTypeFromMIRType(subject->type())), alloc_->lifoAlloc()); } if (oldType->unknown()) return true; TemporaryTypeSet* type = nullptr; if (trueBranch) type = oldType->cloneObjectsOnly(alloc_->lifoAlloc()); else type = oldType->cloneWithoutObjects(alloc_->lifoAlloc()); if (!type) return false; return replaceTypeSet(subject, type, test); } case MDefinition::Op_Phi: { bool branchIsAnd = true; if (!detectAndOrStructure(ins->toPhi(), &branchIsAnd)) { // Just fall through to the default behavior. break; } // Now we have detected the triangular structure and determined if it // was an AND or an OR. if (branchIsAnd) { if (trueBranch) { if (!improveTypesAtTest(ins->toPhi()->getOperand(0), true, test)) return false; if (!improveTypesAtTest(ins->toPhi()->getOperand(1), true, test)) return false; } } else { /* * if (a || b) { * ... * } else { * ... * } * * If we have a statements like the one described above, * And we are in the else branch of it. It amounts to: * if (!(a || b)) and being in the true branch. * * Simplifying, we have (!a && !b) * In this case we can use the same logic we use for branchIsAnd * */ if (!trueBranch) { if (!improveTypesAtTest(ins->toPhi()->getOperand(0), false, test)) return false; if (!improveTypesAtTest(ins->toPhi()->getOperand(1), false, test)) return false; } } return true; } case MDefinition::Op_Compare: return improveTypesAtCompare(ins->toCompare(), trueBranch, test); default: break; } // By default MTest tests ToBoolean(input). As a result in the true branch we can filter // undefined and null. In false branch we can only encounter undefined, null, false, 0, "" // and objects that emulate undefined. TemporaryTypeSet* oldType = ins->resultTypeSet(); TemporaryTypeSet* type; // Create temporary typeset equal to the type if there is no resultTypeSet. TemporaryTypeSet tmp; if (!oldType) { if (ins->type() == MIRType::Value) return true; oldType = &tmp; tmp.addType(TypeSet::PrimitiveType(ValueTypeFromMIRType(ins->type())), alloc_->lifoAlloc()); } // If ins does not have a typeset we return as we cannot optimize. if (oldType->unknown()) return true; // Decide either to set or remove. if (trueBranch) { TemporaryTypeSet remove; remove.addType(TypeSet::UndefinedType(), alloc_->lifoAlloc()); remove.addType(TypeSet::NullType(), alloc_->lifoAlloc()); type = TypeSet::removeSet(oldType, &remove, alloc_->lifoAlloc()); } else { TemporaryTypeSet base; base.addType(TypeSet::UndefinedType(), alloc_->lifoAlloc()); // ToBoolean(undefined) == false base.addType(TypeSet::NullType(), alloc_->lifoAlloc()); // ToBoolean(null) == false base.addType(TypeSet::BooleanType(), alloc_->lifoAlloc()); // ToBoolean(false) == false base.addType(TypeSet::Int32Type(), alloc_->lifoAlloc()); // ToBoolean(0) == false base.addType(TypeSet::DoubleType(), alloc_->lifoAlloc()); // ToBoolean(0.0) == false base.addType(TypeSet::StringType(), alloc_->lifoAlloc()); // ToBoolean("") == false // If the typeset does emulate undefined, then we cannot filter out // objects. if (oldType->maybeEmulatesUndefined(constraints())) base.addType(TypeSet::AnyObjectType(), alloc_->lifoAlloc()); type = TypeSet::intersectSets(&base, oldType, alloc_->lifoAlloc()); } return type && replaceTypeSet(ins, type, test); } bool IonBuilder::jsop_label() { MOZ_ASSERT(JSOp(*pc) == JSOP_LABEL); jsbytecode* endpc = pc + GET_JUMP_OFFSET(pc); MOZ_ASSERT(endpc > pc); ControlFlowInfo label(cfgStack_.length(), endpc); if (!labels_.append(label)) return false; return cfgStack_.append(CFGState::Label(endpc)); } bool IonBuilder::jsop_condswitch() { // CondSwitch op looks as follows: // condswitch [length +exit_pc; first case offset +next-case ] // { // { // ... any code ... // case (+jump) [pcdelta offset +next-case] // }+ // default (+jump) // ... jump targets ... // } // // The default case is always emitted even if there is no default case in // the source. The last case statement pcdelta source note might have a 0 // offset on the last case (not all the time). // // A conditional evaluate the condition of each case and compare it to the // switch value with a strict equality. Cases conditions are iterated // linearly until one is matching. If one case succeeds, the flow jumps into // the corresponding body block. The body block might alias others and // might continue in the next body block if the body is not terminated with // a break. // // Algorithm: // 1/ Loop over the case chain to reach the default target // & Estimate the number of uniq bodies. // 2/ Generate code for all cases (see processCondSwitchCase). // 3/ Generate code for all bodies (see processCondSwitchBody). MOZ_ASSERT(JSOp(*pc) == JSOP_CONDSWITCH); jssrcnote* sn = info().getNote(gsn, pc); MOZ_ASSERT(SN_TYPE(sn) == SRC_CONDSWITCH); // Get the exit pc jsbytecode* exitpc = pc + GetSrcNoteOffset(sn, 0); jsbytecode* firstCase = pc + GetSrcNoteOffset(sn, 1); // Iterate all cases in the conditional switch. // - Stop at the default case. (always emitted after the last case) // - Estimate the number of uniq bodies. This estimation might be off by 1 // if the default body alias a case body. jsbytecode* curCase = firstCase; jsbytecode* lastTarget = GetJumpOffset(curCase) + curCase; size_t nbBodies = 2; // default target and the first body. MOZ_ASSERT(pc < curCase && curCase <= exitpc); while (JSOp(*curCase) == JSOP_CASE) { // Fetch the next case. jssrcnote* caseSn = info().getNote(gsn, curCase); MOZ_ASSERT(caseSn && SN_TYPE(caseSn) == SRC_NEXTCASE); ptrdiff_t off = GetSrcNoteOffset(caseSn, 0); MOZ_ASSERT_IF(off == 0, JSOp(*GetNextPc(curCase)) == JSOP_JUMPTARGET); curCase = off ? curCase + off : GetNextPc(GetNextPc(curCase)); MOZ_ASSERT(pc < curCase && curCase <= exitpc); // Count non-aliased cases. jsbytecode* curTarget = GetJumpOffset(curCase) + curCase; if (lastTarget < curTarget) nbBodies++; lastTarget = curTarget; } // The current case now be the default case which jump to the body of the // default case, which might be behind the last target. MOZ_ASSERT(JSOp(*curCase) == JSOP_DEFAULT); jsbytecode* defaultTarget = GetJumpOffset(curCase) + curCase; MOZ_ASSERT(curCase < defaultTarget && defaultTarget <= exitpc); // Allocate the current graph state. CFGState state = CFGState::CondSwitch(this, exitpc, defaultTarget); if (!state.condswitch.bodies || !state.condswitch.bodies->init(alloc(), nbBodies)) return ControlStatus_Error; // We loop on case conditions with processCondSwitchCase. MOZ_ASSERT(JSOp(*firstCase) == JSOP_CASE); state.stopAt = firstCase; state.state = CFGState::COND_SWITCH_CASE; return cfgStack_.append(state); } IonBuilder::CFGState IonBuilder::CFGState::CondSwitch(IonBuilder* builder, jsbytecode* exitpc, jsbytecode* defaultTarget) { CFGState state; state.state = COND_SWITCH_CASE; state.stopAt = nullptr; state.condswitch.bodies = (FixedList*)builder->alloc_->allocate( sizeof(FixedList)); state.condswitch.currentIdx = 0; state.condswitch.defaultTarget = defaultTarget; state.condswitch.defaultIdx = uint32_t(-1); state.condswitch.exitpc = exitpc; state.condswitch.breaks = nullptr; return state; } IonBuilder::CFGState IonBuilder::CFGState::Label(jsbytecode* exitpc) { CFGState state; state.state = LABEL; state.stopAt = exitpc; state.label.breaks = nullptr; return state; } IonBuilder::CFGState IonBuilder::CFGState::Try(jsbytecode* exitpc, MBasicBlock* successor) { CFGState state; state.state = TRY; state.stopAt = exitpc; state.try_.successor = successor; return state; } IonBuilder::ControlStatus IonBuilder::processCondSwitchCase(CFGState& state) { MOZ_ASSERT(state.state == CFGState::COND_SWITCH_CASE); MOZ_ASSERT(!state.condswitch.breaks); MOZ_ASSERT(current); MOZ_ASSERT(JSOp(*pc) == JSOP_CASE); FixedList& bodies = *state.condswitch.bodies; jsbytecode* defaultTarget = state.condswitch.defaultTarget; uint32_t& currentIdx = state.condswitch.currentIdx; jsbytecode* lastTarget = currentIdx ? bodies[currentIdx - 1]->pc() : nullptr; // Fetch the following case in which we will continue. jssrcnote* sn = info().getNote(gsn, pc); ptrdiff_t off = GetSrcNoteOffset(sn, 0); MOZ_ASSERT_IF(off == 0, JSOp(*GetNextPc(pc)) == JSOP_JUMPTARGET); jsbytecode* casePc = off ? pc + off : GetNextPc(GetNextPc(pc)); bool caseIsDefault = JSOp(*casePc) == JSOP_DEFAULT; MOZ_ASSERT(JSOp(*casePc) == JSOP_CASE || caseIsDefault); // Allocate the block of the matching case. bool bodyIsNew = false; MBasicBlock* bodyBlock = nullptr; jsbytecode* bodyTarget = pc + GetJumpOffset(pc); if (lastTarget < bodyTarget) { // If the default body is in the middle or aliasing the current target. if (lastTarget < defaultTarget && defaultTarget <= bodyTarget) { MOZ_ASSERT(state.condswitch.defaultIdx == uint32_t(-1)); state.condswitch.defaultIdx = currentIdx; bodies[currentIdx] = nullptr; // If the default body does not alias any and it would be allocated // later and stored in the defaultIdx location. if (defaultTarget < bodyTarget) currentIdx++; } bodyIsNew = true; // Pop switch and case operands. bodyBlock = newBlockPopN(current, bodyTarget, 2); bodies[currentIdx++] = bodyBlock; } else { // This body alias the previous one. MOZ_ASSERT(lastTarget == bodyTarget); MOZ_ASSERT(currentIdx > 0); bodyBlock = bodies[currentIdx - 1]; } if (!bodyBlock) return ControlStatus_Error; lastTarget = bodyTarget; // Allocate the block of the non-matching case. This can either be a normal // case or the default case. bool caseIsNew = false; MBasicBlock* caseBlock = nullptr; if (!caseIsDefault) { caseIsNew = true; // Pop the case operand. caseBlock = newBlockPopN(current, GetNextPc(pc), 1); } else { // The non-matching case is the default case, which jump directly to its // body. Skip the creation of a default case block and directly create // the default body if it does not alias any previous body. if (state.condswitch.defaultIdx == uint32_t(-1)) { // The default target is the last target. MOZ_ASSERT(lastTarget < defaultTarget); state.condswitch.defaultIdx = currentIdx++; caseIsNew = true; } else if (bodies[state.condswitch.defaultIdx] == nullptr) { // The default target is in the middle and it does not alias any // case target. MOZ_ASSERT(defaultTarget < lastTarget); caseIsNew = true; } else { // The default target is in the middle and it alias a case target. MOZ_ASSERT(defaultTarget <= lastTarget); caseBlock = bodies[state.condswitch.defaultIdx]; } // Allocate and register the default body. if (caseIsNew) { // Pop the case & switch operands. caseBlock = newBlockPopN(current, defaultTarget, 2); bodies[state.condswitch.defaultIdx] = caseBlock; } } if (!caseBlock) return ControlStatus_Error; // Terminate the last case condition block by emitting the code // corresponding to JSOP_CASE bytecode. if (bodyBlock != caseBlock) { MDefinition* caseOperand = current->pop(); MDefinition* switchOperand = current->peek(-1); if (!jsop_compare(JSOP_STRICTEQ, switchOperand, caseOperand)) return ControlStatus_Error; MInstruction* cmpResult = current->pop()->toInstruction(); MOZ_ASSERT(!cmpResult->isEffectful()); current->end(newTest(cmpResult, bodyBlock, caseBlock)); // Add last case as predecessor of the body if the body is aliasing // the previous case body. if (!bodyIsNew && !bodyBlock->addPredecessorPopN(alloc(), current, 1)) return ControlStatus_Error; // Add last case as predecessor of the non-matching case if the // non-matching case is an aliased default case. We need to pop the // switch operand as we skip the default case block and use the default // body block directly. MOZ_ASSERT_IF(!caseIsNew, caseIsDefault); if (!caseIsNew && !caseBlock->addPredecessorPopN(alloc(), current, 1)) return ControlStatus_Error; } else { // The default case alias the last case body. MOZ_ASSERT(caseIsDefault); current->pop(); // Case operand current->pop(); // Switch operand current->end(MGoto::New(alloc(), bodyBlock)); if (!bodyIsNew && !bodyBlock->addPredecessor(alloc(), current)) return ControlStatus_Error; } if (caseIsDefault) { // The last case condition is finished. Loop in processCondSwitchBody, // with potential stops in processSwitchBreak. Check that the bodies // fixed list is over-estimate by at most 1, and shrink the size such as // length can be used as an upper bound while iterating bodies. MOZ_ASSERT(currentIdx == bodies.length() || currentIdx + 1 == bodies.length()); bodies.shrink(bodies.length() - currentIdx); // Handle break statements in processSwitchBreak while processing // bodies. ControlFlowInfo breakInfo(cfgStack_.length() - 1, state.condswitch.exitpc); if (!switches_.append(breakInfo)) return ControlStatus_Error; // Jump into the first body. currentIdx = 0; setCurrent(nullptr); state.state = CFGState::COND_SWITCH_BODY; return processCondSwitchBody(state); } // Continue until the case condition. if (!setCurrentAndSpecializePhis(caseBlock)) return ControlStatus_Error; pc = current->pc(); state.stopAt = casePc; return ControlStatus_Jumped; } IonBuilder::ControlStatus IonBuilder::processCondSwitchBody(CFGState& state) { MOZ_ASSERT(state.state == CFGState::COND_SWITCH_BODY); MOZ_ASSERT(pc <= state.condswitch.exitpc); FixedList& bodies = *state.condswitch.bodies; uint32_t& currentIdx = state.condswitch.currentIdx; MOZ_ASSERT(currentIdx <= bodies.length()); if (currentIdx == bodies.length()) { MOZ_ASSERT_IF(current, pc == state.condswitch.exitpc); return processSwitchEnd(state.condswitch.breaks, state.condswitch.exitpc); } // Get the next body MBasicBlock* nextBody = bodies[currentIdx++]; MOZ_ASSERT_IF(current, pc == nextBody->pc()); // Fix the reverse post-order iteration. graph().moveBlockToEnd(nextBody); // The last body continue into the new one. if (current) { current->end(MGoto::New(alloc(), nextBody)); if (!nextBody->addPredecessor(alloc(), current)) return ControlStatus_Error; } // Continue in the next body. if (!setCurrentAndSpecializePhis(nextBody)) return ControlStatus_Error; pc = current->pc(); if (currentIdx < bodies.length()) state.stopAt = bodies[currentIdx]->pc(); else state.stopAt = state.condswitch.exitpc; return ControlStatus_Jumped; } bool IonBuilder::jsop_andor(JSOp op) { MOZ_ASSERT(op == JSOP_AND || op == JSOP_OR); jsbytecode* rhsStart = pc + CodeSpec[op].length; jsbytecode* joinStart = pc + GetJumpOffset(pc); MOZ_ASSERT(joinStart > pc); // We have to leave the LHS on the stack. MDefinition* lhs = current->peek(-1); MBasicBlock* evalLhs = newBlock(current, joinStart); MBasicBlock* evalRhs = newBlock(current, rhsStart); if (!evalLhs || !evalRhs) return false; MTest* test = (op == JSOP_AND) ? newTest(lhs, evalRhs, evalLhs) : newTest(lhs, evalLhs, evalRhs); current->end(test); // Create the lhs block and specialize. if (!setCurrentAndSpecializePhis(evalLhs)) return false; if (!improveTypesAtTest(test->getOperand(0), test->ifTrue() == current, test)) return false; // Create the rhs block. if (!cfgStack_.append(CFGState::AndOr(joinStart, evalLhs))) return false; if (!setCurrentAndSpecializePhis(evalRhs)) return false; if (!improveTypesAtTest(test->getOperand(0), test->ifTrue() == current, test)) return false; return true; } bool IonBuilder::jsop_dup2() { uint32_t lhsSlot = current->stackDepth() - 2; uint32_t rhsSlot = current->stackDepth() - 1; current->pushSlot(lhsSlot); current->pushSlot(rhsSlot); return true; } bool IonBuilder::jsop_loophead(jsbytecode* pc) { assertValidLoopHeadOp(pc); current->add(MInterruptCheck::New(alloc())); insertRecompileCheck(); return true; } bool IonBuilder::jsop_ifeq(JSOp op) { // IFEQ always has a forward offset. jsbytecode* trueStart = pc + CodeSpec[op].length; jsbytecode* falseStart = pc + GetJumpOffset(pc); MOZ_ASSERT(falseStart > pc); // We only handle cases that emit source notes. jssrcnote* sn = info().getNote(gsn, pc); if (!sn) return abort("expected sourcenote"); MDefinition* ins = current->pop(); // Create true and false branches. MBasicBlock* ifTrue = newBlock(current, trueStart); MBasicBlock* ifFalse = newBlock(current, falseStart); if (!ifTrue || !ifFalse) return false; MTest* test = newTest(ins, ifTrue, ifFalse); current->end(test); // The bytecode for if/ternary gets emitted either like this: // // IFEQ X ; src note (IF_ELSE, COND) points to the GOTO // ... // GOTO Z // X: ... ; else/else if // ... // Z: ; join // // Or like this: // // IFEQ X ; src note (IF) has no offset // ... // Z: ... ; join // // We want to parse the bytecode as if we were parsing the AST, so for the // IF_ELSE/COND cases, we use the source note and follow the GOTO. For the // IF case, the IFEQ offset is the join point. switch (SN_TYPE(sn)) { case SRC_IF: if (!cfgStack_.append(CFGState::If(falseStart, test))) return false; break; case SRC_IF_ELSE: case SRC_COND: { // Infer the join point from the JSOP_GOTO[X] sitting here, then // assert as we much we can that this is the right GOTO. jsbytecode* trueEnd = pc + GetSrcNoteOffset(sn, 0); MOZ_ASSERT(trueEnd > pc); MOZ_ASSERT(trueEnd < falseStart); MOZ_ASSERT(JSOp(*trueEnd) == JSOP_GOTO); MOZ_ASSERT(!info().getNote(gsn, trueEnd)); jsbytecode* falseEnd = trueEnd + GetJumpOffset(trueEnd); MOZ_ASSERT(falseEnd > trueEnd); MOZ_ASSERT(falseEnd >= falseStart); if (!cfgStack_.append(CFGState::IfElse(trueEnd, falseEnd, test))) return false; break; } default: MOZ_CRASH("unexpected source note type"); } // Switch to parsing the true branch. Note that no PC update is needed, // it's the next instruction. if (!setCurrentAndSpecializePhis(ifTrue)) return false; // Filter the types in the true branch. if (!improveTypesAtTest(test->getOperand(0), test->ifTrue() == current, test)) return false; return true; } bool IonBuilder::jsop_try() { MOZ_ASSERT(JSOp(*pc) == JSOP_TRY); // Try-finally is not yet supported. if (analysis().hasTryFinally()) return abort("Has try-finally"); // Try-catch within inline frames is not yet supported. MOZ_ASSERT(!isInlineBuilder()); // Try-catch during the arguments usage analysis is not yet supported. Code // accessing the arguments within the 'catch' block is not accounted for. if (info().analysisMode() == Analysis_ArgumentsUsage) return abort("Try-catch during arguments usage analysis"); graph().setHasTryBlock(); jssrcnote* sn = info().getNote(gsn, pc); MOZ_ASSERT(SN_TYPE(sn) == SRC_TRY); // Get the pc of the last instruction in the try block. It's a JSOP_GOTO to // jump over the catch block. jsbytecode* endpc = pc + GetSrcNoteOffset(sn, 0); MOZ_ASSERT(JSOp(*endpc) == JSOP_GOTO); MOZ_ASSERT(GetJumpOffset(endpc) > 0); jsbytecode* afterTry = endpc + GetJumpOffset(endpc); // If controlflow in the try body is terminated (by a return or throw // statement), the code after the try-statement may still be reachable // via the catch block (which we don't compile) and OSR can enter it. // For example: // // try { // throw 3; // } catch(e) { } // // for (var i=0; i<1000; i++) {} // // To handle this, we create two blocks: one for the try block and one // for the code following the try-catch statement. Both blocks are // connected to the graph with an MGotoWithFake instruction that always // jumps to the try block. This ensures the successor block always has a // predecessor. // // If the code after the try block is unreachable (control flow in both the // try and catch blocks is terminated), only create the try block, to avoid // parsing unreachable code. MBasicBlock* tryBlock = newBlock(current, GetNextPc(pc)); if (!tryBlock) return false; MBasicBlock* successor; if (analysis().maybeInfo(afterTry)) { successor = newBlock(current, afterTry); if (!successor) return false; current->end(MGotoWithFake::New(alloc(), tryBlock, successor)); } else { successor = nullptr; current->end(MGoto::New(alloc(), tryBlock)); } if (!cfgStack_.append(CFGState::Try(endpc, successor))) return false; // The baseline compiler should not attempt to enter the catch block // via OSR. MOZ_ASSERT(info().osrPc() < endpc || info().osrPc() >= afterTry); // Start parsing the try block. return setCurrentAndSpecializePhis(tryBlock); } IonBuilder::ControlStatus IonBuilder::processReturn(JSOp op) { MDefinition* def; switch (op) { case JSOP_RETURN: // Return the last instruction. def = current->pop(); break; case JSOP_RETRVAL: // Return undefined eagerly if script doesn't use return value. if (script()->noScriptRval()) { MInstruction* ins = MConstant::New(alloc(), UndefinedValue()); current->add(ins); def = ins; break; } def = current->getSlot(info().returnValueSlot()); break; default: def = nullptr; MOZ_CRASH("unknown return op"); } MReturn* ret = MReturn::New(alloc(), def); current->end(ret); if (!graph().addReturn(current)) return ControlStatus_Error; // Make sure no one tries to use this block now. setCurrent(nullptr); return processControlEnd(); } IonBuilder::ControlStatus IonBuilder::processThrow() { MDefinition* def = current->pop(); // MThrow is not marked as effectful. This means when it throws and we // are inside a try block, we could use an earlier resume point and this // resume point may not be up-to-date, for example: // // (function() { // try { // var x = 1; // foo(); // resume point // x = 2; // throw foo; // } catch(e) { // print(x); // } // ])(); // // If we use the resume point after the call, this will print 1 instead // of 2. To fix this, we create a resume point right before the MThrow. // // Note that this is not a problem for instructions other than MThrow // because they are either marked as effectful (have their own resume // point) or cannot throw a catchable exception. // // We always install this resume point (instead of only when the function // has a try block) in order to handle the Debugger onExceptionUnwind // hook. When we need to handle the hook, we bail out to baseline right // after the throw and propagate the exception when debug mode is on. This // is opposed to the normal behavior of resuming directly in the // associated catch block. MNop* nop = MNop::New(alloc()); current->add(nop); if (!resumeAfter(nop)) return ControlStatus_Error; MThrow* ins = MThrow::New(alloc(), def); current->end(ins); // Make sure no one tries to use this block now. setCurrent(nullptr); return processControlEnd(); } void IonBuilder::pushConstant(const Value& v) { current->push(constant(v)); } bool IonBuilder::bitnotTrySpecialized(bool* emitted, MDefinition* input) { MOZ_ASSERT(*emitted == false); // Try to emit a specialized bitnot instruction based on the input type // of the operand. if (input->mightBeType(MIRType::Object) || input->mightBeType(MIRType::Symbol)) return true; MBitNot* ins = MBitNot::New(alloc(), input); ins->setSpecialization(MIRType::Int32); current->add(ins); current->push(ins); *emitted = true; return true; } bool IonBuilder::jsop_bitnot() { bool emitted = false; MDefinition* input = current->pop(); if (!forceInlineCaches()) { if (!bitnotTrySpecialized(&emitted, input) || emitted) return emitted; } if (!arithTrySharedStub(&emitted, JSOP_BITNOT, nullptr, input) || emitted) return emitted; // Not possible to optimize. Do a slow vm call. MBitNot* ins = MBitNot::New(alloc(), input); current->add(ins); current->push(ins); MOZ_ASSERT(ins->isEffectful()); return resumeAfter(ins); } bool IonBuilder::jsop_bitop(JSOp op) { // Pop inputs. MDefinition* right = current->pop(); MDefinition* left = current->pop(); MBinaryBitwiseInstruction* ins; switch (op) { case JSOP_BITAND: ins = MBitAnd::New(alloc(), left, right); break; case JSOP_BITOR: ins = MBitOr::New(alloc(), left, right); break; case JSOP_BITXOR: ins = MBitXor::New(alloc(), left, right); break; case JSOP_LSH: ins = MLsh::New(alloc(), left, right); break; case JSOP_RSH: ins = MRsh::New(alloc(), left, right); break; case JSOP_URSH: ins = MUrsh::New(alloc(), left, right); break; default: MOZ_CRASH("unexpected bitop"); } current->add(ins); ins->infer(inspector, pc); current->push(ins); if (ins->isEffectful() && !resumeAfter(ins)) return false; return true; } MDefinition::Opcode JSOpToMDefinition(JSOp op) { switch (op) { case JSOP_ADD: return MDefinition::Op_Add; case JSOP_SUB: return MDefinition::Op_Sub; case JSOP_MUL: return MDefinition::Op_Mul; case JSOP_DIV: return MDefinition::Op_Div; case JSOP_MOD: return MDefinition::Op_Mod; default: MOZ_CRASH("unexpected binary opcode"); } } bool IonBuilder::binaryArithTryConcat(bool* emitted, JSOp op, MDefinition* left, MDefinition* right) { MOZ_ASSERT(*emitted == false); // Try to convert an addition into a concat operation if the inputs // indicate this might be a concatenation. // Only try to replace this with concat when we have an addition. if (op != JSOP_ADD) return true; trackOptimizationAttempt(TrackedStrategy::BinaryArith_Concat); // Make sure one of the inputs is a string. if (left->type() != MIRType::String && right->type() != MIRType::String) { trackOptimizationOutcome(TrackedOutcome::OperandNotString); return true; } // The none-string input (if present) should be atleast a numerical type. // Which we can easily coerce to string. if (right->type() != MIRType::String && !IsNumberType(right->type())) { trackOptimizationOutcome(TrackedOutcome::OperandNotStringOrNumber); return true; } if (left->type() != MIRType::String && !IsNumberType(left->type())) { trackOptimizationOutcome(TrackedOutcome::OperandNotStringOrNumber); return true; } MConcat* ins = MConcat::New(alloc(), left, right); current->add(ins); current->push(ins); if (!maybeInsertResume()) return false; trackOptimizationSuccess(); *emitted = true; return true; } bool IonBuilder::powTrySpecialized(bool* emitted, MDefinition* base, MDefinition* power, MIRType outputType) { // Typechecking. MDefinition* output = nullptr; MIRType baseType = base->type(); MIRType powerType = power->type(); if (outputType != MIRType::Int32 && outputType != MIRType::Double) return true; if (!IsNumberType(baseType)) return true; if (!IsNumberType(powerType)) return true; if (powerType == MIRType::Float32) powerType = MIRType::Double; MPow* pow = MPow::New(alloc(), base, power, powerType); current->add(pow); output = pow; // Cast to the right type if (outputType == MIRType::Int32 && output->type() != MIRType::Int32) { MToInt32* toInt = MToInt32::New(alloc(), output); current->add(toInt); output = toInt; } if (outputType == MIRType::Double && output->type() != MIRType::Double) { MToDouble* toDouble = MToDouble::New(alloc(), output); current->add(toDouble); output = toDouble; } current->push(output); *emitted = true; return true; } static inline bool SimpleArithOperand(MDefinition* op) { return !op->mightBeType(MIRType::Object) && !op->mightBeType(MIRType::String) && !op->mightBeType(MIRType::Symbol) && !op->mightBeType(MIRType::MagicOptimizedArguments) && !op->mightBeType(MIRType::MagicHole) && !op->mightBeType(MIRType::MagicIsConstructing); } bool IonBuilder::binaryArithTrySpecialized(bool* emitted, JSOp op, MDefinition* left, MDefinition* right) { MOZ_ASSERT(*emitted == false); // Try to emit a specialized binary instruction based on the input types // of the operands. trackOptimizationAttempt(TrackedStrategy::BinaryArith_SpecializedTypes); // Anything complex - strings, symbols, and objects - are not specialized if (!SimpleArithOperand(left) || !SimpleArithOperand(right)) { trackOptimizationOutcome(TrackedOutcome::OperandNotSimpleArith); return true; } // One of the inputs need to be a number. if (!IsNumberType(left->type()) && !IsNumberType(right->type())) { trackOptimizationOutcome(TrackedOutcome::OperandNotNumber); return true; } MDefinition::Opcode defOp = JSOpToMDefinition(op); MBinaryArithInstruction* ins = MBinaryArithInstruction::New(alloc(), defOp, left, right); ins->setNumberSpecialization(alloc(), inspector, pc); if (op == JSOP_ADD || op == JSOP_MUL) ins->setCommutative(); current->add(ins); current->push(ins); MOZ_ASSERT(!ins->isEffectful()); if (!maybeInsertResume()) return false; trackOptimizationSuccess(); *emitted = true; return true; } bool IonBuilder::binaryArithTrySpecializedOnBaselineInspector(bool* emitted, JSOp op, MDefinition* left, MDefinition* right) { MOZ_ASSERT(*emitted == false); // Try to emit a specialized binary instruction speculating the // type using the baseline caches. trackOptimizationAttempt(TrackedStrategy::BinaryArith_SpecializedOnBaselineTypes); MIRType specialization = inspector->expectedBinaryArithSpecialization(pc); if (specialization == MIRType::None) { trackOptimizationOutcome(TrackedOutcome::NoTypeInfo); return true; } MDefinition::Opcode def_op = JSOpToMDefinition(op); MBinaryArithInstruction* ins = MBinaryArithInstruction::New(alloc(), def_op, left, right); ins->setSpecialization(specialization); current->add(ins); current->push(ins); MOZ_ASSERT(!ins->isEffectful()); if (!maybeInsertResume()) return false; trackOptimizationSuccess(); *emitted = true; return true; } bool IonBuilder::arithTrySharedStub(bool* emitted, JSOp op, MDefinition* left, MDefinition* right) { MOZ_ASSERT(*emitted == false); JSOp actualOp = JSOp(*pc); // Try to emit a shared stub cache. if (JitOptions.disableSharedStubs) return true; // The actual jsop 'jsop_pos' is not supported yet. if (actualOp == JSOP_POS) return true; // FIXME: The JSOP_BITNOT path doesn't track optimizations yet. if (actualOp != JSOP_BITNOT) { trackOptimizationAttempt(TrackedStrategy::BinaryArith_SharedCache); trackOptimizationSuccess(); } MInstruction* stub = nullptr; switch (actualOp) { case JSOP_NEG: case JSOP_BITNOT: MOZ_ASSERT_IF(op == JSOP_MUL, left->maybeConstantValue() && left->maybeConstantValue()->toInt32() == -1); MOZ_ASSERT_IF(op != JSOP_MUL, !left); stub = MUnarySharedStub::New(alloc(), right); break; case JSOP_ADD: case JSOP_SUB: case JSOP_MUL: case JSOP_DIV: case JSOP_MOD: case JSOP_POW: stub = MBinarySharedStub::New(alloc(), left, right); break; default: MOZ_CRASH("unsupported arith"); } current->add(stub); current->push(stub); // Decrease type from 'any type' to 'empty type' when one of the operands // is 'empty typed'. maybeMarkEmpty(stub); if (!resumeAfter(stub)) return false; *emitted = true; return true; } bool IonBuilder::jsop_binary_arith(JSOp op, MDefinition* left, MDefinition* right) { bool emitted = false; startTrackingOptimizations(); trackTypeInfo(TrackedTypeSite::Operand, left->type(), left->resultTypeSet()); trackTypeInfo(TrackedTypeSite::Operand, right->type(), right->resultTypeSet()); if (!forceInlineCaches()) { if (!binaryArithTryConcat(&emitted, op, left, right) || emitted) return emitted; if (!binaryArithTrySpecialized(&emitted, op, left, right) || emitted) return emitted; if (!binaryArithTrySpecializedOnBaselineInspector(&emitted, op, left, right) || emitted) return emitted; } if (!arithTrySharedStub(&emitted, op, left, right) || emitted) return emitted; // Not possible to optimize. Do a slow vm call. trackOptimizationAttempt(TrackedStrategy::BinaryArith_Call); trackOptimizationSuccess(); MDefinition::Opcode def_op = JSOpToMDefinition(op); MBinaryArithInstruction* ins = MBinaryArithInstruction::New(alloc(), def_op, left, right); // Decrease type from 'any type' to 'empty type' when one of the operands // is 'empty typed'. maybeMarkEmpty(ins); current->add(ins); current->push(ins); MOZ_ASSERT(ins->isEffectful()); return resumeAfter(ins); } bool IonBuilder::jsop_binary_arith(JSOp op) { MDefinition* right = current->pop(); MDefinition* left = current->pop(); return jsop_binary_arith(op, left, right); } bool IonBuilder::jsop_pow() { MDefinition* exponent = current->pop(); MDefinition* base = current->pop(); bool emitted = false; if (!forceInlineCaches()) { if (!powTrySpecialized(&emitted, base, exponent, MIRType::Double) || emitted) return emitted; } if (!arithTrySharedStub(&emitted, JSOP_POW, base, exponent) || emitted) return emitted; // For now, use MIRType::Double, as a safe cover-all. See bug 1188079. MPow* pow = MPow::New(alloc(), base, exponent, MIRType::Double); current->add(pow); current->push(pow); return true; } bool IonBuilder::jsop_pos() { if (IsNumberType(current->peek(-1)->type())) { // Already int32 or double. Set the operand as implicitly used so it // doesn't get optimized out if it has no other uses, as we could bail // out. current->peek(-1)->setImplicitlyUsedUnchecked(); return true; } // Compile +x as x * 1. MDefinition* value = current->pop(); MConstant* one = MConstant::New(alloc(), Int32Value(1)); current->add(one); return jsop_binary_arith(JSOP_MUL, value, one); } bool IonBuilder::jsop_neg() { // Since JSOP_NEG does not use a slot, we cannot push the MConstant. // The MConstant is therefore passed to JSOP_MUL without slot traffic. MConstant* negator = MConstant::New(alloc(), Int32Value(-1)); current->add(negator); MDefinition* right = current->pop(); return jsop_binary_arith(JSOP_MUL, negator, right); } bool IonBuilder::jsop_tostring() { if (current->peek(-1)->type() == MIRType::String) return true; MDefinition* value = current->pop(); MToString* ins = MToString::New(alloc(), value); current->add(ins); current->push(ins); MOZ_ASSERT(!ins->isEffectful()); return true; } class AutoAccumulateReturns { MIRGraph& graph_; MIRGraphReturns* prev_; public: AutoAccumulateReturns(MIRGraph& graph, MIRGraphReturns& returns) : graph_(graph) { prev_ = graph_.returnAccumulator(); graph_.setReturnAccumulator(&returns); } ~AutoAccumulateReturns() { graph_.setReturnAccumulator(prev_); } }; IonBuilder::InliningStatus IonBuilder::inlineScriptedCall(CallInfo& callInfo, JSFunction* target) { MOZ_ASSERT(target->hasScript()); MOZ_ASSERT(IsIonInlinablePC(pc)); MBasicBlock::BackupPoint backup(current); if (!backup.init(alloc())) return InliningStatus_Error; callInfo.setImplicitlyUsedUnchecked(); // Ensure sufficient space in the slots: needed for inlining from FUNAPPLY. uint32_t depth = current->stackDepth() + callInfo.numFormals(); if (depth > current->nslots()) { if (!current->increaseSlots(depth - current->nslots())) return InliningStatus_Error; } // Create new |this| on the caller-side for inlined constructors. if (callInfo.constructing()) { MDefinition* thisDefn = createThis(target, callInfo.fun(), callInfo.getNewTarget()); if (!thisDefn) return InliningStatus_Error; callInfo.setThis(thisDefn); } // Capture formals in the outer resume point. callInfo.pushFormals(current); MResumePoint* outerResumePoint = MResumePoint::New(alloc(), current, pc, MResumePoint::Outer); if (!outerResumePoint) return InliningStatus_Error; current->setOuterResumePoint(outerResumePoint); // Pop formals again, except leave |fun| on stack for duration of call. callInfo.popFormals(current); current->push(callInfo.fun()); JSScript* calleeScript = target->nonLazyScript(); BaselineInspector inspector(calleeScript); // Improve type information of |this| when not set. if (callInfo.constructing() && !callInfo.thisArg()->resultTypeSet()) { StackTypeSet* types = TypeScript::ThisTypes(calleeScript); if (types && !types->unknown()) { TemporaryTypeSet* clonedTypes = types->clone(alloc_->lifoAlloc()); if (!clonedTypes) return InliningStatus_Error; MTypeBarrier* barrier = MTypeBarrier::New(alloc(), callInfo.thisArg(), clonedTypes); current->add(barrier); if (barrier->type() == MIRType::Undefined) callInfo.setThis(constant(UndefinedValue())); else if (barrier->type() == MIRType::Null) callInfo.setThis(constant(NullValue())); else callInfo.setThis(barrier); } } // Start inlining. LifoAlloc* lifoAlloc = alloc_->lifoAlloc(); InlineScriptTree* inlineScriptTree = info().inlineScriptTree()->addCallee(alloc_, pc, calleeScript); if (!inlineScriptTree) return InliningStatus_Error; CompileInfo* info = lifoAlloc->new_(calleeScript, target, (jsbytecode*)nullptr, this->info().analysisMode(), /* needsArgsObj = */ false, inlineScriptTree); if (!info) return InliningStatus_Error; MIRGraphReturns returns(alloc()); AutoAccumulateReturns aar(graph(), returns); // Build the graph. IonBuilder inlineBuilder(analysisContext, compartment, options, &alloc(), &graph(), constraints(), &inspector, info, &optimizationInfo(), nullptr, inliningDepth_ + 1, loopDepth_); if (!inlineBuilder.buildInline(this, outerResumePoint, callInfo)) { if (analysisContext && analysisContext->isExceptionPending()) { JitSpew(JitSpew_IonAbort, "Inline builder raised exception."); abortReason_ = AbortReason_Error; return InliningStatus_Error; } // Inlining the callee failed. Mark the callee as uninlineable only if // the inlining was aborted for a non-exception reason. if (inlineBuilder.abortReason_ == AbortReason_Disable) { calleeScript->setUninlineable(); if (!JitOptions.disableInlineBacktracking) { current = backup.restore(); return InliningStatus_NotInlined; } abortReason_ = AbortReason_Inlining; } else if (inlineBuilder.abortReason_ == AbortReason_Inlining) { abortReason_ = AbortReason_Inlining; } else if (inlineBuilder.abortReason_ == AbortReason_Alloc) { abortReason_ = AbortReason_Alloc; } else if (inlineBuilder.abortReason_ == AbortReason_PreliminaryObjects) { const ObjectGroupVector& groups = inlineBuilder.abortedPreliminaryGroups(); MOZ_ASSERT(!groups.empty()); for (size_t i = 0; i < groups.length(); i++) addAbortedPreliminaryGroup(groups[i]); abortReason_ = AbortReason_PreliminaryObjects; } return InliningStatus_Error; } // Create return block. jsbytecode* postCall = GetNextPc(pc); MBasicBlock* returnBlock = newBlock(nullptr, postCall); if (!returnBlock) return InliningStatus_Error; returnBlock->setCallerResumePoint(callerResumePoint_); // Inherit the slots from current and pop |fun|. returnBlock->inheritSlots(current); returnBlock->pop(); // Accumulate return values. if (returns.empty()) { // Inlining of functions that have no exit is not supported. calleeScript->setUninlineable(); if (!JitOptions.disableInlineBacktracking) { current = backup.restore(); return InliningStatus_NotInlined; } abortReason_ = AbortReason_Inlining; return InliningStatus_Error; } MDefinition* retvalDefn = patchInlinedReturns(callInfo, returns, returnBlock); if (!retvalDefn) return InliningStatus_Error; returnBlock->push(retvalDefn); // Initialize entry slots now that the stack has been fixed up. if (!returnBlock->initEntrySlots(alloc())) return InliningStatus_Error; if (!setCurrentAndSpecializePhis(returnBlock)) return InliningStatus_Error; return InliningStatus_Inlined; } MDefinition* IonBuilder::patchInlinedReturn(CallInfo& callInfo, MBasicBlock* exit, MBasicBlock* bottom) { // Replaces the MReturn in the exit block with an MGoto. MDefinition* rdef = exit->lastIns()->toReturn()->input(); exit->discardLastIns(); // Constructors must be patched by the caller to always return an object. if (callInfo.constructing()) { if (rdef->type() == MIRType::Value) { // Unknown return: dynamically detect objects. MReturnFromCtor* filter = MReturnFromCtor::New(alloc(), rdef, callInfo.thisArg()); exit->add(filter); rdef = filter; } else if (rdef->type() != MIRType::Object) { // Known non-object return: force |this|. rdef = callInfo.thisArg(); } } else if (callInfo.isSetter()) { // Setters return their argument, not whatever value is returned. rdef = callInfo.getArg(0); } if (!callInfo.isSetter()) rdef = specializeInlinedReturn(rdef, exit); MGoto* replacement = MGoto::New(alloc(), bottom); exit->end(replacement); if (!bottom->addPredecessorWithoutPhis(exit)) return nullptr; return rdef; } MDefinition* IonBuilder::specializeInlinedReturn(MDefinition* rdef, MBasicBlock* exit) { // Remove types from the return definition that weren't observed. TemporaryTypeSet* types = bytecodeTypes(pc); // The observed typeset doesn't contain extra information. if (types->empty() || types->unknown()) return rdef; // Decide if specializing is needed using the result typeset if available, // else use the result type. if (rdef->resultTypeSet()) { // Don't specialize if return typeset is a subset of the // observed typeset. The return typeset is already more specific. if (rdef->resultTypeSet()->isSubset(types)) return rdef; } else { MIRType observedType = types->getKnownMIRType(); // Don't specialize if type is MIRType::Float32 and TI reports // MIRType::Double. Float is more specific than double. if (observedType == MIRType::Double && rdef->type() == MIRType::Float32) return rdef; // Don't specialize if types are inaccordance, except for MIRType::Value // and MIRType::Object (when not unknown object), since the typeset // contains more specific information. if (observedType == rdef->type() && observedType != MIRType::Value && (observedType != MIRType::Object || types->unknownObject())) { return rdef; } } setCurrent(exit); MTypeBarrier* barrier = nullptr; rdef = addTypeBarrier(rdef, types, BarrierKind::TypeSet, &barrier); if (barrier) barrier->setNotMovable(); return rdef; } MDefinition* IonBuilder::patchInlinedReturns(CallInfo& callInfo, MIRGraphReturns& returns, MBasicBlock* bottom) { // Replaces MReturns with MGotos, returning the MDefinition // representing the return value, or nullptr. MOZ_ASSERT(returns.length() > 0); if (returns.length() == 1) return patchInlinedReturn(callInfo, returns[0], bottom); // Accumulate multiple returns with a phi. MPhi* phi = MPhi::New(alloc()); if (!phi->reserveLength(returns.length())) return nullptr; for (size_t i = 0; i < returns.length(); i++) { MDefinition* rdef = patchInlinedReturn(callInfo, returns[i], bottom); if (!rdef) return nullptr; phi->addInput(rdef); } bottom->addPhi(phi); return phi; } IonBuilder::InliningDecision IonBuilder::makeInliningDecision(JSObject* targetArg, CallInfo& callInfo) { // When there is no target, inlining is impossible. if (targetArg == nullptr) { trackOptimizationOutcome(TrackedOutcome::CantInlineNoTarget); return InliningDecision_DontInline; } // Inlining non-function targets is handled by inlineNonFunctionCall(). if (!targetArg->is()) return InliningDecision_Inline; JSFunction* target = &targetArg->as(); // Never inline during the arguments usage analysis. if (info().analysisMode() == Analysis_ArgumentsUsage) return InliningDecision_DontInline; // Native functions provide their own detection in inlineNativeCall(). if (target->isNative()) return InliningDecision_Inline; // Determine whether inlining is possible at callee site InliningDecision decision = canInlineTarget(target, callInfo); if (decision != InliningDecision_Inline) return decision; // Heuristics! JSScript* targetScript = target->nonLazyScript(); // Callee must not be excessively large. // This heuristic also applies to the callsite as a whole. bool offThread = options.offThreadCompilationAvailable(); if (targetScript->length() > optimizationInfo().inlineMaxBytecodePerCallSite(offThread)) { trackOptimizationOutcome(TrackedOutcome::CantInlineBigCallee); return DontInline(targetScript, "Vetoed: callee excessively large"); } // Callee must have been called a few times to have somewhat stable // type information, except for definite properties analysis, // as the caller has not run yet. if (targetScript->getWarmUpCount() < optimizationInfo().inliningWarmUpThreshold() && !targetScript->baselineScript()->ionCompiledOrInlined() && info().analysisMode() != Analysis_DefiniteProperties) { trackOptimizationOutcome(TrackedOutcome::CantInlineNotHot); JitSpew(JitSpew_Inlining, "Cannot inline %s:%" PRIuSIZE ": callee is insufficiently hot.", targetScript->filename(), targetScript->lineno()); return InliningDecision_WarmUpCountTooLow; } // Don't inline if the callee is known to inline a lot of code, to avoid // huge MIR graphs. uint32_t inlinedBytecodeLength = targetScript->baselineScript()->inlinedBytecodeLength(); if (inlinedBytecodeLength > optimizationInfo().inlineMaxCalleeInlinedBytecodeLength()) { trackOptimizationOutcome(TrackedOutcome::CantInlineBigCalleeInlinedBytecodeLength); return DontInline(targetScript, "Vetoed: callee inlinedBytecodeLength is too big"); } IonBuilder* outerBuilder = outermostBuilder(); // Cap the total bytecode length we inline under a single script, to avoid // excessive inlining in pathological cases. size_t totalBytecodeLength = outerBuilder->inlinedBytecodeLength_ + targetScript->length(); if (totalBytecodeLength > optimizationInfo().inlineMaxTotalBytecodeLength()) { trackOptimizationOutcome(TrackedOutcome::CantInlineExceededTotalBytecodeLength); return DontInline(targetScript, "Vetoed: exceeding max total bytecode length"); } // Cap the inlining depth. uint32_t maxInlineDepth; if (JitOptions.isSmallFunction(targetScript)) { maxInlineDepth = optimizationInfo().smallFunctionMaxInlineDepth(); } else { maxInlineDepth = optimizationInfo().maxInlineDepth(); // Caller must not be excessively large. if (script()->length() >= optimizationInfo().inliningMaxCallerBytecodeLength()) { trackOptimizationOutcome(TrackedOutcome::CantInlineBigCaller); return DontInline(targetScript, "Vetoed: caller excessively large"); } } BaselineScript* outerBaseline = outermostBuilder()->script()->baselineScript(); if (inliningDepth_ >= maxInlineDepth) { // We hit the depth limit and won't inline this function. Give the // outermost script a max inlining depth of 0, so that it won't be // inlined in other scripts. This heuristic is currently only used // when we're inlining scripts with loops, see the comment below. outerBaseline->setMaxInliningDepth(0); trackOptimizationOutcome(TrackedOutcome::CantInlineExceededDepth); return DontInline(targetScript, "Vetoed: exceeding allowed inline depth"); } // Inlining functions with loops can be complicated. For instance, if we're // close to the inlining depth limit and we inline the function f below, we // can no longer inline the call to g: // // function f() { // while (cond) { // g(); // } // } // // If the loop has many iterations, it's more efficient to call f and inline // g in f. // // To avoid this problem, we record a separate max inlining depth for each // script, indicating at which depth we won't be able to inline all functions // we inlined this time. This solves the issue above, because we will only // inline f if it means we can also inline g. if (targetScript->hasLoops() && inliningDepth_ >= targetScript->baselineScript()->maxInliningDepth()) { trackOptimizationOutcome(TrackedOutcome::CantInlineExceededDepth); return DontInline(targetScript, "Vetoed: exceeding allowed script inline depth"); } // Update the max depth at which we can inline the outer script. MOZ_ASSERT(maxInlineDepth > inliningDepth_); uint32_t scriptInlineDepth = maxInlineDepth - inliningDepth_ - 1; if (scriptInlineDepth < outerBaseline->maxInliningDepth()) outerBaseline->setMaxInliningDepth(scriptInlineDepth); // End of heuristics, we will inline this function. // TI calls ObjectStateChange to trigger invalidation of the caller. TypeSet::ObjectKey* targetKey = TypeSet::ObjectKey::get(target); targetKey->watchStateChangeForInlinedCall(constraints()); outerBuilder->inlinedBytecodeLength_ += targetScript->length(); return InliningDecision_Inline; } bool IonBuilder::selectInliningTargets(const ObjectVector& targets, CallInfo& callInfo, BoolVector& choiceSet, uint32_t* numInlineable) { *numInlineable = 0; uint32_t totalSize = 0; // For each target, ask whether it may be inlined. if (!choiceSet.reserve(targets.length())) return false; // Don't inline polymorphic sites during the definite properties analysis. // AddClearDefiniteFunctionUsesInScript depends on this for correctness. if (info().analysisMode() == Analysis_DefiniteProperties && targets.length() > 1) return true; for (size_t i = 0; i < targets.length(); i++) { JSObject* target = targets[i]; trackOptimizationAttempt(TrackedStrategy::Call_Inline); trackTypeInfo(TrackedTypeSite::Call_Target, target); bool inlineable; InliningDecision decision = makeInliningDecision(target, callInfo); switch (decision) { case InliningDecision_Error: return false; case InliningDecision_DontInline: case InliningDecision_WarmUpCountTooLow: inlineable = false; break; case InliningDecision_Inline: inlineable = true; break; default: MOZ_CRASH("Unhandled InliningDecision value!"); } if (target->is()) { // Enforce a maximum inlined bytecode limit at the callsite. if (inlineable && target->as().isInterpreted()) { totalSize += target->as().nonLazyScript()->length(); bool offThread = options.offThreadCompilationAvailable(); if (totalSize > optimizationInfo().inlineMaxBytecodePerCallSite(offThread)) inlineable = false; } } else { // Non-function targets are not supported by polymorphic inlining. inlineable = false; } choiceSet.infallibleAppend(inlineable); if (inlineable) *numInlineable += 1; } // If optimization tracking is turned on and one of the inlineable targets // is a native, track the type info of the call. Most native inlinings // depend on the types of the arguments and the return value. if (isOptimizationTrackingEnabled()) { for (size_t i = 0; i < targets.length(); i++) { if (choiceSet[i] && targets[i]->as().isNative()) { trackTypeInfo(callInfo); break; } } } MOZ_ASSERT(choiceSet.length() == targets.length()); return true; } static bool CanInlineGetPropertyCache(MGetPropertyCache* cache, MDefinition* thisDef) { MOZ_ASSERT(cache->object()->type() == MIRType::Object); if (cache->object() != thisDef) return false; InlinePropertyTable* table = cache->propTable(); if (!table) return false; if (table->numEntries() == 0) return false; return true; } class WrapMGetPropertyCache { MGetPropertyCache* cache_; private: void discardPriorResumePoint() { if (!cache_) return; InlinePropertyTable* propTable = cache_->propTable(); if (!propTable) return; MResumePoint* rp = propTable->takePriorResumePoint(); if (!rp) return; cache_->block()->discardPreAllocatedResumePoint(rp); } public: explicit WrapMGetPropertyCache(MGetPropertyCache* cache) : cache_(cache) { } ~WrapMGetPropertyCache() { discardPriorResumePoint(); } MGetPropertyCache* get() { return cache_; } MGetPropertyCache* operator->() { return get(); } // This function returns the cache given to the constructor if the // GetPropertyCache can be moved into the ObjectGroup fallback path. MGetPropertyCache* moveableCache(bool hasTypeBarrier, MDefinition* thisDef) { // If we have unhandled uses of the MGetPropertyCache, then we cannot // move it to the ObjectGroup fallback path. if (!hasTypeBarrier) { if (cache_->hasUses()) return nullptr; } else { // There is the TypeBarrier consumer, so we check that this is the // only consumer. MOZ_ASSERT(cache_->hasUses()); if (!cache_->hasOneUse()) return nullptr; } // If the this-object is not identical to the object of the // MGetPropertyCache, then we cannot use the InlinePropertyTable, or if // we do not yet have enough information from the ObjectGroup. if (!CanInlineGetPropertyCache(cache_, thisDef)) return nullptr; MGetPropertyCache* ret = cache_; cache_ = nullptr; return ret; } }; MGetPropertyCache* IonBuilder::getInlineableGetPropertyCache(CallInfo& callInfo) { if (callInfo.constructing()) return nullptr; MDefinition* thisDef = callInfo.thisArg(); if (thisDef->type() != MIRType::Object) return nullptr; MDefinition* funcDef = callInfo.fun(); if (funcDef->type() != MIRType::Object) return nullptr; // MGetPropertyCache with no uses may be optimized away. if (funcDef->isGetPropertyCache()) { WrapMGetPropertyCache cache(funcDef->toGetPropertyCache()); return cache.moveableCache(/* hasTypeBarrier = */ false, thisDef); } // Optimize away the following common pattern: // MTypeBarrier[MIRType::Object] <- MGetPropertyCache if (funcDef->isTypeBarrier()) { MTypeBarrier* barrier = funcDef->toTypeBarrier(); if (barrier->hasUses()) return nullptr; if (barrier->type() != MIRType::Object) return nullptr; if (!barrier->input()->isGetPropertyCache()) return nullptr; WrapMGetPropertyCache cache(barrier->input()->toGetPropertyCache()); return cache.moveableCache(/* hasTypeBarrier = */ true, thisDef); } return nullptr; } IonBuilder::InliningStatus IonBuilder::inlineSingleCall(CallInfo& callInfo, JSObject* targetArg) { if (!targetArg->is()) { InliningStatus status = inlineNonFunctionCall(callInfo, targetArg); trackInlineSuccess(status); return status; } JSFunction* target = &targetArg->as(); if (target->isNative()) { InliningStatus status = inlineNativeCall(callInfo, target); trackInlineSuccess(status); return status; } // Track success now, as inlining a scripted call makes a new return block // which has a different pc than the current call pc. trackInlineSuccess(); return inlineScriptedCall(callInfo, target); } IonBuilder::InliningStatus IonBuilder::inlineCallsite(const ObjectVector& targets, CallInfo& callInfo) { if (targets.empty()) { trackOptimizationAttempt(TrackedStrategy::Call_Inline); trackOptimizationOutcome(TrackedOutcome::CantInlineNoTarget); return InliningStatus_NotInlined; } // Is the function provided by an MGetPropertyCache? // If so, the cache may be movable to a fallback path, with a dispatch // instruction guarding on the incoming ObjectGroup. WrapMGetPropertyCache propCache(getInlineableGetPropertyCache(callInfo)); keepFallbackFunctionGetter(propCache.get()); // Inline single targets -- unless they derive from a cache, in which case // avoiding the cache and guarding is still faster. if (!propCache.get() && targets.length() == 1) { JSObject* target = targets[0]; trackOptimizationAttempt(TrackedStrategy::Call_Inline); trackTypeInfo(TrackedTypeSite::Call_Target, target); InliningDecision decision = makeInliningDecision(target, callInfo); switch (decision) { case InliningDecision_Error: return InliningStatus_Error; case InliningDecision_DontInline: return InliningStatus_NotInlined; case InliningDecision_WarmUpCountTooLow: return InliningStatus_WarmUpCountTooLow; case InliningDecision_Inline: break; } // Inlining will elminate uses of the original callee, but it needs to // be preserved in phis if we bail out. Mark the old callee definition as // implicitly used to ensure this happens. callInfo.fun()->setImplicitlyUsedUnchecked(); // If the callee is not going to be a lambda (which may vary across // different invocations), then the callee definition can be replaced by a // constant. if (target->isSingleton()) { // Replace the function with an MConstant. MConstant* constFun = constant(ObjectValue(*target)); if (callInfo.constructing() && callInfo.getNewTarget() == callInfo.fun()) callInfo.setNewTarget(constFun); callInfo.setFun(constFun); } return inlineSingleCall(callInfo, target); } // Choose a subset of the targets for polymorphic inlining. BoolVector choiceSet(alloc()); uint32_t numInlined; if (!selectInliningTargets(targets, callInfo, choiceSet, &numInlined)) return InliningStatus_Error; if (numInlined == 0) return InliningStatus_NotInlined; // Perform a polymorphic dispatch. if (!inlineCalls(callInfo, targets, choiceSet, propCache.get())) return InliningStatus_Error; return InliningStatus_Inlined; } bool IonBuilder::inlineGenericFallback(JSFunction* target, CallInfo& callInfo, MBasicBlock* dispatchBlock) { // Generate a new block with all arguments on-stack. MBasicBlock* fallbackBlock = newBlock(dispatchBlock, pc); if (!fallbackBlock) return false; // Create a new CallInfo to track modified state within this block. CallInfo fallbackInfo(alloc(), callInfo.constructing(), callInfo.ignoresReturnValue()); if (!fallbackInfo.init(callInfo)) return false; fallbackInfo.popFormals(fallbackBlock); // Generate an MCall, which uses stateful |current|. if (!setCurrentAndSpecializePhis(fallbackBlock)) return false; if (!makeCall(target, fallbackInfo)) return false; // Pass return block to caller as |current|. return true; } bool IonBuilder::inlineObjectGroupFallback(CallInfo& callInfo, MBasicBlock* dispatchBlock, MObjectGroupDispatch* dispatch, MGetPropertyCache* cache, MBasicBlock** fallbackTarget) { // Getting here implies the following: // 1. The call function is an MGetPropertyCache, or an MGetPropertyCache // followed by an MTypeBarrier. MOZ_ASSERT(callInfo.fun()->isGetPropertyCache() || callInfo.fun()->isTypeBarrier()); // 2. The MGetPropertyCache has inlineable cases by guarding on the ObjectGroup. MOZ_ASSERT(dispatch->numCases() > 0); // 3. The MGetPropertyCache (and, if applicable, MTypeBarrier) only // have at most a single use. MOZ_ASSERT_IF(callInfo.fun()->isGetPropertyCache(), !cache->hasUses()); MOZ_ASSERT_IF(callInfo.fun()->isTypeBarrier(), cache->hasOneUse()); // This means that no resume points yet capture the MGetPropertyCache, // so everything from the MGetPropertyCache up until the call is movable. // We now move the MGetPropertyCache and friends into a fallback path. MOZ_ASSERT(cache->idempotent()); // Create a new CallInfo to track modified state within the fallback path. CallInfo fallbackInfo(alloc(), callInfo.constructing(), callInfo.ignoresReturnValue()); if (!fallbackInfo.init(callInfo)) return false; // Capture stack prior to the call operation. This captures the function. MResumePoint* preCallResumePoint = MResumePoint::New(alloc(), dispatchBlock, pc, MResumePoint::ResumeAt); if (!preCallResumePoint) return false; DebugOnly preCallFuncIndex = preCallResumePoint->stackDepth() - callInfo.numFormals(); MOZ_ASSERT(preCallResumePoint->getOperand(preCallFuncIndex) == fallbackInfo.fun()); // In the dispatch block, replace the function's slot entry with Undefined. MConstant* undefined = MConstant::New(alloc(), UndefinedValue()); dispatchBlock->add(undefined); dispatchBlock->rewriteAtDepth(-int(callInfo.numFormals()), undefined); // Construct a block that does nothing but remove formals from the stack. // This is effectively changing the entry resume point of the later fallback block. MBasicBlock* prepBlock = newBlock(dispatchBlock, pc); if (!prepBlock) return false; fallbackInfo.popFormals(prepBlock); // Construct a block into which the MGetPropertyCache can be moved. // This is subtle: the pc and resume point are those of the MGetPropertyCache! InlinePropertyTable* propTable = cache->propTable(); MResumePoint* priorResumePoint = propTable->takePriorResumePoint(); MOZ_ASSERT(propTable->pc() != nullptr); MOZ_ASSERT(priorResumePoint != nullptr); MBasicBlock* getPropBlock = newBlock(prepBlock, propTable->pc(), priorResumePoint); if (!getPropBlock) return false; prepBlock->end(MGoto::New(alloc(), getPropBlock)); // Since the getPropBlock inherited the stack from right before the MGetPropertyCache, // the target of the MGetPropertyCache is still on the stack. DebugOnly checkObject = getPropBlock->pop(); MOZ_ASSERT(checkObject == cache->object()); // Move the MGetPropertyCache and friends into the getPropBlock. if (fallbackInfo.fun()->isGetPropertyCache()) { MOZ_ASSERT(fallbackInfo.fun()->toGetPropertyCache() == cache); getPropBlock->addFromElsewhere(cache); getPropBlock->push(cache); } else { MTypeBarrier* barrier = callInfo.fun()->toTypeBarrier(); MOZ_ASSERT(barrier->type() == MIRType::Object); MOZ_ASSERT(barrier->input()->isGetPropertyCache()); MOZ_ASSERT(barrier->input()->toGetPropertyCache() == cache); getPropBlock->addFromElsewhere(cache); getPropBlock->addFromElsewhere(barrier); getPropBlock->push(barrier); } // Construct an end block with the correct resume point. MBasicBlock* preCallBlock = newBlock(getPropBlock, pc, preCallResumePoint); if (!preCallBlock) return false; getPropBlock->end(MGoto::New(alloc(), preCallBlock)); // Now inline the MCallGeneric, using preCallBlock as the dispatch point. if (!inlineGenericFallback(nullptr, fallbackInfo, preCallBlock)) return false; // inlineGenericFallback() set the return block as |current|. preCallBlock->end(MGoto::New(alloc(), current)); *fallbackTarget = prepBlock; return true; } bool IonBuilder::inlineCalls(CallInfo& callInfo, const ObjectVector& targets, BoolVector& choiceSet, MGetPropertyCache* maybeCache) { // Only handle polymorphic inlining. MOZ_ASSERT(IsIonInlinablePC(pc)); MOZ_ASSERT(choiceSet.length() == targets.length()); MOZ_ASSERT_IF(!maybeCache, targets.length() >= 2); MOZ_ASSERT_IF(maybeCache, targets.length() >= 1); MBasicBlock* dispatchBlock = current; callInfo.setImplicitlyUsedUnchecked(); callInfo.pushFormals(dispatchBlock); // Patch any InlinePropertyTable to only contain functions that are // inlineable. The InlinePropertyTable will also be patched at the end to // exclude native functions that vetoed inlining. if (maybeCache) { InlinePropertyTable* propTable = maybeCache->propTable(); propTable->trimToTargets(targets); if (propTable->numEntries() == 0) maybeCache = nullptr; } // Generate a dispatch based on guard kind. MDispatchInstruction* dispatch; if (maybeCache) { dispatch = MObjectGroupDispatch::New(alloc(), maybeCache->object(), maybeCache->propTable()); callInfo.fun()->setImplicitlyUsedUnchecked(); } else { dispatch = MFunctionDispatch::New(alloc(), callInfo.fun()); } // Generate a return block to host the rval-collecting MPhi. jsbytecode* postCall = GetNextPc(pc); MBasicBlock* returnBlock = newBlock(nullptr, postCall); if (!returnBlock) return false; returnBlock->setCallerResumePoint(callerResumePoint_); // Set up stack, used to manually create a post-call resume point. returnBlock->inheritSlots(dispatchBlock); callInfo.popFormals(returnBlock); MPhi* retPhi = MPhi::New(alloc()); returnBlock->addPhi(retPhi); returnBlock->push(retPhi); // Create a resume point from current stack state. if (!returnBlock->initEntrySlots(alloc())) return false; // Reserve the capacity for the phi. // Note: this is an upperbound. Unreachable targets and uninlineable natives are also counted. uint32_t count = 1; // Possible fallback block. for (uint32_t i = 0; i < targets.length(); i++) { if (choiceSet[i]) count++; } if (!retPhi->reserveLength(count)) return false; // Inline each of the inlineable targets. for (uint32_t i = 0; i < targets.length(); i++) { // Target must be inlineable. if (!choiceSet[i]) continue; // Even though we made one round of inline decisions already, we may // be amending them below. amendOptimizationAttempt(i); // Target must be reachable by the MDispatchInstruction. JSFunction* target = &targets[i]->as(); if (maybeCache && !maybeCache->propTable()->hasFunction(target)) { choiceSet[i] = false; trackOptimizationOutcome(TrackedOutcome::CantInlineNotInDispatch); continue; } MBasicBlock* inlineBlock = newBlock(dispatchBlock, pc); if (!inlineBlock) return false; // Create a function MConstant to use in the entry ResumePoint. If we // can't use a constant, add a no-op MPolyInlineGuard, to prevent // hoisting env chain gets above the dispatch instruction. MInstruction* funcDef; if (target->isSingleton()) funcDef = MConstant::New(alloc(), ObjectValue(*target), constraints()); else funcDef = MPolyInlineGuard::New(alloc(), callInfo.fun()); funcDef->setImplicitlyUsedUnchecked(); dispatchBlock->add(funcDef); // Use the inlined callee in the inline resume point and on stack. int funIndex = inlineBlock->entryResumePoint()->stackDepth() - callInfo.numFormals(); inlineBlock->entryResumePoint()->replaceOperand(funIndex, funcDef); inlineBlock->rewriteSlot(funIndex, funcDef); // Create a new CallInfo to track modified state within the inline block. CallInfo inlineInfo(alloc(), callInfo.constructing(), callInfo.ignoresReturnValue()); if (!inlineInfo.init(callInfo)) return false; inlineInfo.popFormals(inlineBlock); inlineInfo.setFun(funcDef); if (maybeCache) { // Assign the 'this' value a TypeSet specialized to the groups that // can generate this inlining target. MOZ_ASSERT(callInfo.thisArg() == maybeCache->object()); TemporaryTypeSet* thisTypes = maybeCache->propTable()->buildTypeSetForFunction(target); if (!thisTypes) return false; MFilterTypeSet* filter = MFilterTypeSet::New(alloc(), inlineInfo.thisArg(), thisTypes); inlineBlock->add(filter); inlineInfo.setThis(filter); } // Inline the call into the inlineBlock. if (!setCurrentAndSpecializePhis(inlineBlock)) return false; InliningStatus status = inlineSingleCall(inlineInfo, target); if (status == InliningStatus_Error) return false; // Natives may veto inlining. if (status == InliningStatus_NotInlined) { MOZ_ASSERT(current == inlineBlock); graph().removeBlock(inlineBlock); choiceSet[i] = false; continue; } // inlineSingleCall() changed |current| to the inline return block. MBasicBlock* inlineReturnBlock = current; setCurrent(dispatchBlock); // Connect the inline path to the returnBlock. ObjectGroup* funcGroup = target->isSingleton() ? nullptr : target->group(); if (!dispatch->addCase(target, funcGroup, inlineBlock)) return false; MDefinition* retVal = inlineReturnBlock->peek(-1); retPhi->addInput(retVal); inlineReturnBlock->end(MGoto::New(alloc(), returnBlock)); if (!returnBlock->addPredecessorWithoutPhis(inlineReturnBlock)) return false; } // Patch the InlinePropertyTable to not dispatch to vetoed paths. bool useFallback; if (maybeCache) { InlinePropertyTable* propTable = maybeCache->propTable(); propTable->trimTo(targets, choiceSet); if (propTable->numEntries() == 0 || !propTable->hasPriorResumePoint()) { // Output a generic fallback path. MOZ_ASSERT_IF(propTable->numEntries() == 0, dispatch->numCases() == 0); maybeCache = nullptr; useFallback = true; } else { // We need a fallback path if the ObjectGroup dispatch does not // handle all incoming objects. useFallback = false; TemporaryTypeSet* objectTypes = maybeCache->object()->resultTypeSet(); for (uint32_t i = 0; i < objectTypes->getObjectCount(); i++) { TypeSet::ObjectKey* obj = objectTypes->getObject(i); if (!obj) continue; if (!obj->isGroup()) { useFallback = true; break; } if (!propTable->hasObjectGroup(obj->group())) { useFallback = true; break; } } if (!useFallback) { // The object group dispatch handles all possible incoming // objects, so the cache and barrier will not be reached and // can be eliminated. if (callInfo.fun()->isGetPropertyCache()) { MOZ_ASSERT(callInfo.fun() == maybeCache); } else { MTypeBarrier* barrier = callInfo.fun()->toTypeBarrier(); MOZ_ASSERT(!barrier->hasUses()); MOZ_ASSERT(barrier->type() == MIRType::Object); MOZ_ASSERT(barrier->input()->isGetPropertyCache()); MOZ_ASSERT(barrier->input()->toGetPropertyCache() == maybeCache); barrier->block()->discard(barrier); } MOZ_ASSERT(!maybeCache->hasUses()); maybeCache->block()->discard(maybeCache); } } } else { useFallback = dispatch->numCases() < targets.length(); } // If necessary, generate a fallback path. if (useFallback) { // Generate fallback blocks, and set |current| to the fallback return block. if (maybeCache) { MBasicBlock* fallbackTarget; if (!inlineObjectGroupFallback(callInfo, dispatchBlock, dispatch->toObjectGroupDispatch(), maybeCache, &fallbackTarget)) { return false; } dispatch->addFallback(fallbackTarget); } else { JSFunction* remaining = nullptr; // If there is only 1 remaining case, we can annotate the fallback call // with the target information. if (dispatch->numCases() + 1 == targets.length()) { for (uint32_t i = 0; i < targets.length(); i++) { if (choiceSet[i]) continue; MOZ_ASSERT(!remaining); if (targets[i]->is() && targets[i]->as().isSingleton()) remaining = &targets[i]->as(); break; } } if (!inlineGenericFallback(remaining, callInfo, dispatchBlock)) return false; dispatch->addFallback(current); } MBasicBlock* fallbackReturnBlock = current; // Connect fallback case to return infrastructure. MDefinition* retVal = fallbackReturnBlock->peek(-1); retPhi->addInput(retVal); fallbackReturnBlock->end(MGoto::New(alloc(), returnBlock)); if (!returnBlock->addPredecessorWithoutPhis(fallbackReturnBlock)) return false; } // Finally add the dispatch instruction. // This must be done at the end so that add() may be called above. dispatchBlock->end(dispatch); // Check the depth change: +1 for retval MOZ_ASSERT(returnBlock->stackDepth() == dispatchBlock->stackDepth() - callInfo.numFormals() + 1); graph().moveBlockToEnd(returnBlock); return setCurrentAndSpecializePhis(returnBlock); } MInstruction* IonBuilder::createNamedLambdaObject(MDefinition* callee, MDefinition* env) { // Get a template CallObject that we'll use to generate inline object // creation. LexicalEnvironmentObject* templateObj = inspector->templateNamedLambdaObject(); // One field is added to the function to handle its name. This cannot be a // dynamic slot because there is still plenty of room on the NamedLambda object. MOZ_ASSERT(!templateObj->hasDynamicSlots()); // Allocate the actual object. It is important that no intervening // instructions could potentially bailout, thus leaking the dynamic slots // pointer. MInstruction* declEnvObj = MNewNamedLambdaObject::New(alloc(), templateObj); current->add(declEnvObj); // Initialize the object's reserved slots. No post barrier is needed here: // the object will be allocated in the nursery if possible, and if the // tenured heap is used instead, a minor collection will have been performed // that moved env/callee to the tenured heap. current->add(MStoreFixedSlot::New(alloc(), declEnvObj, NamedLambdaObject::enclosingEnvironmentSlot(), env)); current->add(MStoreFixedSlot::New(alloc(), declEnvObj, NamedLambdaObject::lambdaSlot(), callee)); return declEnvObj; } MInstruction* IonBuilder::createCallObject(MDefinition* callee, MDefinition* env) { // Get a template CallObject that we'll use to generate inline object // creation. CallObject* templateObj = inspector->templateCallObject(); // Allocate the object. Run-once scripts need a singleton type, so always do // a VM call in such cases. MNewCallObjectBase* callObj; if (script()->treatAsRunOnce() || templateObj->isSingleton()) callObj = MNewSingletonCallObject::New(alloc(), templateObj); else callObj = MNewCallObject::New(alloc(), templateObj); current->add(callObj); // Initialize the object's reserved slots. No post barrier is needed here, // for the same reason as in createNamedLambdaObject. current->add(MStoreFixedSlot::New(alloc(), callObj, CallObject::enclosingEnvironmentSlot(), env)); current->add(MStoreFixedSlot::New(alloc(), callObj, CallObject::calleeSlot(), callee)); //if (!script()->functionHasParameterExprs()) { // Copy closed-over argument slots if there aren't parameter expressions. MSlots* slots = nullptr; for (PositionalFormalParameterIter fi(script()); fi; fi++) { if (!fi.closedOver()) continue; if (!alloc().ensureBallast()) return nullptr; unsigned slot = fi.location().slot(); unsigned formal = fi.argumentSlot(); unsigned numFixedSlots = templateObj->numFixedSlots(); MDefinition* param; if (script()->functionHasParameterExprs()) param = constant(MagicValue(JS_UNINITIALIZED_LEXICAL)); else param = current->getSlot(info().argSlotUnchecked(formal)); if (slot >= numFixedSlots) { if (!slots) { slots = MSlots::New(alloc(), callObj); current->add(slots); } current->add(MStoreSlot::New(alloc(), slots, slot - numFixedSlots, param)); } else { current->add(MStoreFixedSlot::New(alloc(), callObj, slot, param)); } } return callObj; } MDefinition* IonBuilder::createThisScripted(MDefinition* callee, MDefinition* newTarget) { // Get callee.prototype. // // This instruction MUST be idempotent: since it does not correspond to an // explicit operation in the bytecode, we cannot use resumeAfter(). // Getters may not override |prototype| fetching, so this operation is indeed idempotent. // - First try an idempotent property cache. // - Upon failing idempotent property cache, we can't use a non-idempotent cache, // therefore we fallback to CallGetProperty // // Note: both CallGetProperty and GetPropertyCache can trigger a GC, // and thus invalidation. MInstruction* getProto; if (!invalidatedIdempotentCache()) { MConstant* id = constant(StringValue(names().prototype)); MGetPropertyCache* getPropCache = MGetPropertyCache::New(alloc(), newTarget, id, /* monitored = */ false); getPropCache->setIdempotent(); getProto = getPropCache; } else { MCallGetProperty* callGetProp = MCallGetProperty::New(alloc(), newTarget, names().prototype); callGetProp->setIdempotent(); getProto = callGetProp; } current->add(getProto); // Create this from prototype MCreateThisWithProto* createThis = MCreateThisWithProto::New(alloc(), callee, newTarget, getProto); current->add(createThis); return createThis; } JSObject* IonBuilder::getSingletonPrototype(JSFunction* target) { TypeSet::ObjectKey* targetKey = TypeSet::ObjectKey::get(target); if (targetKey->unknownProperties()) return nullptr; jsid protoid = NameToId(names().prototype); HeapTypeSetKey protoProperty = targetKey->property(protoid); return protoProperty.singleton(constraints()); } MDefinition* IonBuilder::createThisScriptedSingleton(JSFunction* target, MDefinition* callee) { if (!target->hasScript()) return nullptr; // Get the singleton prototype (if exists) JSObject* proto = getSingletonPrototype(target); if (!proto) return nullptr; JSObject* templateObject = inspector->getTemplateObject(pc); if (!templateObject) return nullptr; if (!templateObject->is()) return nullptr; if (templateObject->staticPrototype() != proto) return nullptr; TypeSet::ObjectKey* templateObjectKey = TypeSet::ObjectKey::get(templateObject->group()); if (templateObjectKey->hasFlags(constraints(), OBJECT_FLAG_NEW_SCRIPT_CLEARED)) return nullptr; StackTypeSet* thisTypes = TypeScript::ThisTypes(target->nonLazyScript()); if (!thisTypes || !thisTypes->hasType(TypeSet::ObjectType(templateObject))) return nullptr; // Generate an inline path to create a new |this| object with // the given singleton prototype. MConstant* templateConst = MConstant::NewConstraintlessObject(alloc(), templateObject); MCreateThisWithTemplate* createThis = MCreateThisWithTemplate::New(alloc(), constraints(), templateConst, templateObject->group()->initialHeap(constraints())); current->add(templateConst); current->add(createThis); return createThis; } MDefinition* IonBuilder::createThisScriptedBaseline(MDefinition* callee) { // Try to inline |this| creation based on Baseline feedback. JSFunction* target = inspector->getSingleCallee(pc); if (!target || !target->hasScript()) return nullptr; JSObject* templateObject = inspector->getTemplateObject(pc); if (!templateObject) return nullptr; if (!templateObject->is()) return nullptr; Shape* shape = target->lookupPure(compartment->runtime()->names().prototype); if (!shape || !shape->hasDefaultGetter() || !shape->hasSlot()) return nullptr; Value protov = target->getSlot(shape->slot()); if (!protov.isObject()) return nullptr; JSObject* proto = checkNurseryObject(&protov.toObject()); if (proto != templateObject->staticPrototype()) return nullptr; TypeSet::ObjectKey* templateObjectKey = TypeSet::ObjectKey::get(templateObject->group()); if (templateObjectKey->hasFlags(constraints(), OBJECT_FLAG_NEW_SCRIPT_CLEARED)) return nullptr; StackTypeSet* thisTypes = TypeScript::ThisTypes(target->nonLazyScript()); if (!thisTypes || !thisTypes->hasType(TypeSet::ObjectType(templateObject))) return nullptr; // Shape guard. callee = addShapeGuard(callee, target->lastProperty(), Bailout_ShapeGuard); // Guard callee.prototype == proto. MOZ_ASSERT(shape->numFixedSlots() == 0, "Must be a dynamic slot"); MSlots* slots = MSlots::New(alloc(), callee); current->add(slots); MLoadSlot* prototype = MLoadSlot::New(alloc(), slots, shape->slot()); current->add(prototype); MDefinition* protoConst = constant(ObjectValue(*proto)); MGuardObjectIdentity* guard = MGuardObjectIdentity::New(alloc(), prototype, protoConst, /* bailOnEquality = */ false); current->add(guard); // Generate an inline path to create a new |this| object with // the given prototype. MConstant* templateConst = MConstant::NewConstraintlessObject(alloc(), templateObject); MCreateThisWithTemplate* createThis = MCreateThisWithTemplate::New(alloc(), constraints(), templateConst, templateObject->group()->initialHeap(constraints())); current->add(templateConst); current->add(createThis); return createThis; } MDefinition* IonBuilder::createThis(JSFunction* target, MDefinition* callee, MDefinition* newTarget) { // Create |this| for unknown target. if (!target) { if (MDefinition* createThis = createThisScriptedBaseline(callee)) return createThis; MCreateThis* createThis = MCreateThis::New(alloc(), callee, newTarget); current->add(createThis); return createThis; } // Native constructors build the new Object themselves. if (target->isNative()) { if (!target->isConstructor()) return nullptr; MConstant* magic = MConstant::New(alloc(), MagicValue(JS_IS_CONSTRUCTING)); current->add(magic); return magic; } if (target->isBoundFunction()) return constant(MagicValue(JS_UNINITIALIZED_LEXICAL)); if (target->isDerivedClassConstructor()) { MOZ_ASSERT(target->isClassConstructor()); return constant(MagicValue(JS_UNINITIALIZED_LEXICAL)); } // Try baking in the prototype. if (MDefinition* createThis = createThisScriptedSingleton(target, callee)) return createThis; if (MDefinition* createThis = createThisScriptedBaseline(callee)) return createThis; return createThisScripted(callee, newTarget); } bool IonBuilder::jsop_funcall(uint32_t argc) { // Stack for JSOP_FUNCALL: // 1: arg0 // ... // argc: argN // argc+1: JSFunction*, the 'f' in |f.call()|, in |this| position. // argc+2: The native 'call' function. int calleeDepth = -((int)argc + 2); int funcDepth = -((int)argc + 1); // If |Function.prototype.call| may be overridden, don't optimize callsite. TemporaryTypeSet* calleeTypes = current->peek(calleeDepth)->resultTypeSet(); JSFunction* native = getSingleCallTarget(calleeTypes); if (!native || !native->isNative() || native->native() != &fun_call) { CallInfo callInfo(alloc(), /* constructing = */ false, /* ignoresReturnValue = */ BytecodeIsPopped(pc)); if (!callInfo.init(current, argc)) return false; return makeCall(native, callInfo); } current->peek(calleeDepth)->setImplicitlyUsedUnchecked(); // Extract call target. TemporaryTypeSet* funTypes = current->peek(funcDepth)->resultTypeSet(); JSFunction* target = getSingleCallTarget(funTypes); // Shimmy the slots down to remove the native 'call' function. current->shimmySlots(funcDepth - 1); bool zeroArguments = (argc == 0); // If no |this| argument was provided, explicitly pass Undefined. // Pushing is safe here, since one stack slot has been removed. if (zeroArguments) { pushConstant(UndefinedValue()); } else { // |this| becomes implicit in the call. argc -= 1; } CallInfo callInfo(alloc(), /* constructing = */ false, /* ignoresReturnValue = */ BytecodeIsPopped(pc)); if (!callInfo.init(current, argc)) return false; // Try to inline the call. if (!zeroArguments) { InliningDecision decision = makeInliningDecision(target, callInfo); switch (decision) { case InliningDecision_Error: return false; case InliningDecision_DontInline: case InliningDecision_WarmUpCountTooLow: break; case InliningDecision_Inline: if (target->isInterpreted()) { InliningStatus status = inlineScriptedCall(callInfo, target); if (status == InliningStatus_Inlined) return true; if (status == InliningStatus_Error) return false; } break; } } // Call without inlining. return makeCall(target, callInfo); } bool IonBuilder::jsop_funapply(uint32_t argc) { int calleeDepth = -((int)argc + 2); TemporaryTypeSet* calleeTypes = current->peek(calleeDepth)->resultTypeSet(); JSFunction* native = getSingleCallTarget(calleeTypes); if (argc != 2 || info().analysisMode() == Analysis_ArgumentsUsage) { CallInfo callInfo(alloc(), /* constructing = */ false, /* ignoresReturnValue = */ BytecodeIsPopped(pc)); if (!callInfo.init(current, argc)) return false; return makeCall(native, callInfo); } // Disable compilation if the second argument to |apply| cannot be guaranteed // to be either definitely |arguments| or definitely not |arguments|. MDefinition* argument = current->peek(-1); if (script()->argumentsHasVarBinding() && argument->mightBeType(MIRType::MagicOptimizedArguments) && argument->type() != MIRType::MagicOptimizedArguments) { return abort("fun.apply with MaybeArguments"); } // Fallback to regular call if arg 2 is not definitely |arguments|. if (argument->type() != MIRType::MagicOptimizedArguments) { // Optimize fun.apply(self, array) if the length is sane and there are no holes. TemporaryTypeSet* objTypes = argument->resultTypeSet(); if (native && native->isNative() && native->native() == fun_apply && objTypes && objTypes->getKnownClass(constraints()) == &ArrayObject::class_ && !objTypes->hasObjectFlags(constraints(), OBJECT_FLAG_LENGTH_OVERFLOW) && ElementAccessIsPacked(constraints(), argument)) { return jsop_funapplyarray(argc); } CallInfo callInfo(alloc(), /* constructing = */ false, /* ignoresReturnValue = */ BytecodeIsPopped(pc)); if (!callInfo.init(current, argc)) return false; return makeCall(native, callInfo); } if ((!native || !native->isNative() || native->native() != fun_apply) && info().analysisMode() != Analysis_DefiniteProperties) { return abort("fun.apply speculation failed"); } // Use funapply that definitely uses |arguments| return jsop_funapplyarguments(argc); } bool IonBuilder::jsop_funapplyarray(uint32_t argc) { MOZ_ASSERT(argc == 2); int funcDepth = -((int)argc + 1); // Extract call target. TemporaryTypeSet* funTypes = current->peek(funcDepth)->resultTypeSet(); JSFunction* target = getSingleCallTarget(funTypes); // Pop the array agument MDefinition* argObj = current->pop(); MElements* elements = MElements::New(alloc(), argObj); current->add(elements); // Pop the |this| argument. MDefinition* argThis = current->pop(); // Unwrap the (JSFunction *) parameter. MDefinition* argFunc = current->pop(); // Pop apply function. MDefinition* nativeFunc = current->pop(); nativeFunc->setImplicitlyUsedUnchecked(); WrappedFunction* wrappedTarget = target ? new(alloc()) WrappedFunction(target) : nullptr; MApplyArray* apply = MApplyArray::New(alloc(), wrappedTarget, argFunc, elements, argThis); current->add(apply); current->push(apply); if (!resumeAfter(apply)) return false; TemporaryTypeSet* types = bytecodeTypes(pc); return pushTypeBarrier(apply, types, BarrierKind::TypeSet); } bool IonBuilder::jsop_funapplyarguments(uint32_t argc) { // Stack for JSOP_FUNAPPLY: // 1: Vp // 2: This // argc+1: JSFunction*, the 'f' in |f.call()|, in |this| position. // argc+2: The native 'apply' function. int funcDepth = -((int)argc + 1); // Extract call target. TemporaryTypeSet* funTypes = current->peek(funcDepth)->resultTypeSet(); JSFunction* target = getSingleCallTarget(funTypes); // When this script isn't inlined, use MApplyArgs, // to copy the arguments from the stack and call the function if (inliningDepth_ == 0 && info().analysisMode() != Analysis_DefiniteProperties) { // The array argument corresponds to the arguments object. As the JIT // is implicitly reading the arguments object in the next instruction, // we need to prevent the deletion of the arguments object from resume // points, so that Baseline will behave correctly after a bailout. MDefinition* vp = current->pop(); vp->setImplicitlyUsedUnchecked(); MDefinition* argThis = current->pop(); // Unwrap the (JSFunction*) parameter. MDefinition* argFunc = current->pop(); // Pop apply function. MDefinition* nativeFunc = current->pop(); nativeFunc->setImplicitlyUsedUnchecked(); MArgumentsLength* numArgs = MArgumentsLength::New(alloc()); current->add(numArgs); WrappedFunction* wrappedTarget = target ? new(alloc()) WrappedFunction(target) : nullptr; MApplyArgs* apply = MApplyArgs::New(alloc(), wrappedTarget, argFunc, numArgs, argThis); current->add(apply); current->push(apply); if (!resumeAfter(apply)) return false; TemporaryTypeSet* types = bytecodeTypes(pc); return pushTypeBarrier(apply, types, BarrierKind::TypeSet); } // When inlining we have the arguments the function gets called with // and can optimize even more, by just calling the functions with the args. // We also try this path when doing the definite properties analysis, as we // can inline the apply() target and don't care about the actual arguments // that were passed in. CallInfo callInfo(alloc(), /* constructing = */ false, /* ignoresReturnValue = */ BytecodeIsPopped(pc)); // Vp MDefinition* vp = current->pop(); vp->setImplicitlyUsedUnchecked(); // Arguments if (inliningDepth_) { if (!callInfo.setArgs(inlineCallInfo_->argv())) return false; } // This MDefinition* argThis = current->pop(); callInfo.setThis(argThis); // Pop function parameter. MDefinition* argFunc = current->pop(); callInfo.setFun(argFunc); // Pop apply function. MDefinition* nativeFunc = current->pop(); nativeFunc->setImplicitlyUsedUnchecked(); // Try to inline the call. InliningDecision decision = makeInliningDecision(target, callInfo); switch (decision) { case InliningDecision_Error: return false; case InliningDecision_DontInline: case InliningDecision_WarmUpCountTooLow: break; case InliningDecision_Inline: if (target->isInterpreted()) { InliningStatus status = inlineScriptedCall(callInfo, target); if (status == InliningStatus_Inlined) return true; if (status == InliningStatus_Error) return false; } } return makeCall(target, callInfo); } bool IonBuilder::jsop_call(uint32_t argc, bool constructing, bool ignoresReturnValue) { startTrackingOptimizations(); // If this call has never executed, try to seed the observed type set // based on how the call result is used. TemporaryTypeSet* observed = bytecodeTypes(pc); if (observed->empty()) { if (BytecodeFlowsToBitop(pc)) { observed->addType(TypeSet::Int32Type(), alloc_->lifoAlloc()); } else if (*GetNextPc(pc) == JSOP_POS) { // Note: this is lame, overspecialized on the code patterns used // by asm.js and should be replaced by a more general mechanism. // See bug 870847. observed->addType(TypeSet::DoubleType(), alloc_->lifoAlloc()); } } int calleeDepth = -((int)argc + 2 + constructing); // Acquire known call target if existent. ObjectVector targets(alloc()); TemporaryTypeSet* calleeTypes = current->peek(calleeDepth)->resultTypeSet(); if (calleeTypes && !getPolyCallTargets(calleeTypes, constructing, targets, 4)) return false; CallInfo callInfo(alloc(), constructing, ignoresReturnValue); if (!callInfo.init(current, argc)) return false; // Try inlining InliningStatus status = inlineCallsite(targets, callInfo); if (status == InliningStatus_Inlined) return true; if (status == InliningStatus_Error) return false; // Discard unreferenced & pre-allocated resume points. replaceMaybeFallbackFunctionGetter(nullptr); // No inline, just make the call. JSFunction* target = nullptr; if (targets.length() == 1 && targets[0]->is()) target = &targets[0]->as(); if (target && status == InliningStatus_WarmUpCountTooLow) { MRecompileCheck* check = MRecompileCheck::New(alloc(), target->nonLazyScript(), optimizationInfo().inliningRecompileThreshold(), MRecompileCheck::RecompileCheck_Inlining); current->add(check); } return makeCall(target, callInfo); } bool IonBuilder::testShouldDOMCall(TypeSet* inTypes, JSFunction* func, JSJitInfo::OpType opType) { if (!func->isNative() || !func->jitInfo()) return false; // If all the DOM objects flowing through are legal with this // property, we can bake in a call to the bottom half of the DOM // accessor DOMInstanceClassHasProtoAtDepth instanceChecker = compartment->runtime()->DOMcallbacks()->instanceClassMatchesProto; const JSJitInfo* jinfo = func->jitInfo(); if (jinfo->type() != opType) return false; for (unsigned i = 0; i < inTypes->getObjectCount(); i++) { TypeSet::ObjectKey* key = inTypes->getObject(i); if (!key) continue; if (!key->hasStableClassAndProto(constraints())) return false; if (!instanceChecker(key->clasp(), jinfo->protoID, jinfo->depth)) return false; } return true; } static bool ArgumentTypesMatch(MDefinition* def, StackTypeSet* calleeTypes) { if (!calleeTypes) return false; if (def->resultTypeSet()) { MOZ_ASSERT(def->type() == MIRType::Value || def->mightBeType(def->type())); return def->resultTypeSet()->isSubset(calleeTypes); } if (def->type() == MIRType::Value) return false; if (def->type() == MIRType::Object) return calleeTypes->unknownObject(); return calleeTypes->mightBeMIRType(def->type()); } bool IonBuilder::testNeedsArgumentCheck(JSFunction* target, CallInfo& callInfo) { // If we have a known target, check if the caller arg types are a subset of callee. // Since typeset accumulates and can't decrease that means we don't need to check // the arguments anymore. if (!target->hasScript()) return true; JSScript* targetScript = target->nonLazyScript(); if (!ArgumentTypesMatch(callInfo.thisArg(), TypeScript::ThisTypes(targetScript))) return true; uint32_t expected_args = Min(callInfo.argc(), target->nargs()); for (size_t i = 0; i < expected_args; i++) { if (!ArgumentTypesMatch(callInfo.getArg(i), TypeScript::ArgTypes(targetScript, i))) return true; } for (size_t i = callInfo.argc(); i < target->nargs(); i++) { if (!TypeScript::ArgTypes(targetScript, i)->mightBeMIRType(MIRType::Undefined)) return true; } return false; } MCall* IonBuilder::makeCallHelper(JSFunction* target, CallInfo& callInfo) { // This function may be called with mutated stack. // Querying TI for popped types is invalid. uint32_t targetArgs = callInfo.argc(); // Collect number of missing arguments provided that the target is // scripted. Native functions are passed an explicit 'argc' parameter. if (target && !target->isNative()) targetArgs = Max(target->nargs(), callInfo.argc()); bool isDOMCall = false; if (target && !callInfo.constructing()) { // We know we have a single call target. Check whether the "this" types // are DOM types and our function a DOM function, and if so flag the // MCall accordingly. TemporaryTypeSet* thisTypes = callInfo.thisArg()->resultTypeSet(); if (thisTypes && thisTypes->getKnownMIRType() == MIRType::Object && thisTypes->isDOMClass(constraints()) && testShouldDOMCall(thisTypes, target, JSJitInfo::Method)) { isDOMCall = true; } } MCall* call = MCall::New(alloc(), target, targetArgs + 1 + callInfo.constructing(), callInfo.argc(), callInfo.constructing(), callInfo.ignoresReturnValue(), isDOMCall); if (!call) return nullptr; if (callInfo.constructing()) call->addArg(targetArgs + 1, callInfo.getNewTarget()); // Explicitly pad any missing arguments with |undefined|. // This permits skipping the argumentsRectifier. for (int i = targetArgs; i > (int)callInfo.argc(); i--) { MOZ_ASSERT_IF(target, !target->isNative()); MConstant* undef = constant(UndefinedValue()); if (!alloc().ensureBallast()) return nullptr; call->addArg(i, undef); } // Add explicit arguments. // Skip addArg(0) because it is reserved for this for (int32_t i = callInfo.argc() - 1; i >= 0; i--) call->addArg(i + 1, callInfo.getArg(i)); // Now that we've told it about all the args, compute whether it's movable call->computeMovable(); // Inline the constructor on the caller-side. if (callInfo.constructing()) { MDefinition* create = createThis(target, callInfo.fun(), callInfo.getNewTarget()); if (!create) { abort("Failure inlining constructor for call."); return nullptr; } callInfo.thisArg()->setImplicitlyUsedUnchecked(); callInfo.setThis(create); } // Pass |this| and function. MDefinition* thisArg = callInfo.thisArg(); call->addArg(0, thisArg); if (target && !testNeedsArgumentCheck(target, callInfo)) call->disableArgCheck(); call->initFunction(callInfo.fun()); current->add(call); return call; } static bool DOMCallNeedsBarrier(const JSJitInfo* jitinfo, TemporaryTypeSet* types) { MOZ_ASSERT(jitinfo->type() != JSJitInfo::InlinableNative); // If the return type of our DOM native is in "types" already, we don't // actually need a barrier. if (jitinfo->returnType() == JSVAL_TYPE_UNKNOWN) return true; // JSVAL_TYPE_OBJECT doesn't tell us much; we still have to barrier on the // actual type of the object. if (jitinfo->returnType() == JSVAL_TYPE_OBJECT) return true; // No need for a barrier if we're already expecting the type we'll produce. return MIRTypeFromValueType(jitinfo->returnType()) != types->getKnownMIRType(); } bool IonBuilder::makeCall(JSFunction* target, CallInfo& callInfo) { // Constructor calls to non-constructors should throw. We don't want to use // CallKnown in this case. MOZ_ASSERT_IF(callInfo.constructing() && target, target->isConstructor()); MCall* call = makeCallHelper(target, callInfo); if (!call) return false; current->push(call); if (call->isEffectful() && !resumeAfter(call)) return false; TemporaryTypeSet* types = bytecodeTypes(pc); if (call->isCallDOMNative()) return pushDOMTypeBarrier(call, types, call->getSingleTarget()->rawJSFunction()); return pushTypeBarrier(call, types, BarrierKind::TypeSet); } bool IonBuilder::jsop_eval(uint32_t argc) { int calleeDepth = -((int)argc + 2); TemporaryTypeSet* calleeTypes = current->peek(calleeDepth)->resultTypeSet(); // Emit a normal call if the eval has never executed. This keeps us from // disabling compilation for the script when testing with --ion-eager. if (calleeTypes && calleeTypes->empty()) return jsop_call(argc, /* constructing = */ false, false); JSFunction* singleton = getSingleCallTarget(calleeTypes); if (!singleton) return abort("No singleton callee for eval()"); if (script()->global().valueIsEval(ObjectValue(*singleton))) { if (argc != 1) return abort("Direct eval with more than one argument"); if (!info().funMaybeLazy()) return abort("Direct eval in global code"); if (info().funMaybeLazy()->isArrow()) return abort("Direct eval from arrow function"); CallInfo callInfo(alloc(), /* constructing = */ false, /* ignoresReturnValue = */ BytecodeIsPopped(pc)); if (!callInfo.init(current, argc)) return false; callInfo.setImplicitlyUsedUnchecked(); callInfo.fun()->setImplicitlyUsedUnchecked(); MDefinition* envChain = current->environmentChain(); MDefinition* string = callInfo.getArg(0); // Direct eval acts as identity on non-string types according to // ES5 15.1.2.1 step 1. if (!string->mightBeType(MIRType::String)) { current->push(string); TemporaryTypeSet* types = bytecodeTypes(pc); return pushTypeBarrier(string, types, BarrierKind::TypeSet); } if (!jsop_newtarget()) return false; MDefinition* newTargetValue = current->pop(); // Try to pattern match 'eval(v + "()")'. In this case v is likely a // name on the env chain and the eval is performing a call on that // value. Use an env chain lookup rather than a full eval. if (string->isConcat() && string->getOperand(1)->type() == MIRType::String && string->getOperand(1)->maybeConstantValue()) { JSAtom* atom = &string->getOperand(1)->maybeConstantValue()->toString()->asAtom(); if (StringEqualsAscii(atom, "()")) { MDefinition* name = string->getOperand(0); MInstruction* dynamicName = MGetDynamicName::New(alloc(), envChain, name); current->add(dynamicName); current->push(dynamicName); current->push(constant(UndefinedValue())); // thisv CallInfo evalCallInfo(alloc(), /* constructing = */ false, /* ignoresReturnValue = */ BytecodeIsPopped(pc)); if (!evalCallInfo.init(current, /* argc = */ 0)) return false; return makeCall(nullptr, evalCallInfo); } } MInstruction* ins = MCallDirectEval::New(alloc(), envChain, string, newTargetValue, pc); current->add(ins); current->push(ins); TemporaryTypeSet* types = bytecodeTypes(pc); return resumeAfter(ins) && pushTypeBarrier(ins, types, BarrierKind::TypeSet); } return jsop_call(argc, /* constructing = */ false, false); } bool IonBuilder::jsop_compare(JSOp op) { MDefinition* right = current->pop(); MDefinition* left = current->pop(); return jsop_compare(op, left, right); } bool IonBuilder::jsop_compare(JSOp op, MDefinition* left, MDefinition* right) { bool emitted = false; if (!forceInlineCaches()) { if (!compareTrySpecialized(&emitted, op, left, right) || emitted) return emitted; if (!compareTryBitwise(&emitted, op, left, right) || emitted) return emitted; if (!compareTrySpecializedOnBaselineInspector(&emitted, op, left, right) || emitted) return emitted; } if (!compareTrySharedStub(&emitted, op, left, right) || emitted) return emitted; // Not possible to optimize. Do a slow vm call. MCompare* ins = MCompare::New(alloc(), left, right, op); ins->cacheOperandMightEmulateUndefined(constraints()); current->add(ins); current->push(ins); if (ins->isEffectful() && !resumeAfter(ins)) return false; return true; } static bool ObjectOrSimplePrimitive(MDefinition* op) { // Return true if op is either undefined/null/boolean/int32 or an object. return !op->mightBeType(MIRType::String) && !op->mightBeType(MIRType::Symbol) && !op->mightBeType(MIRType::Double) && !op->mightBeType(MIRType::Float32) && !op->mightBeType(MIRType::MagicOptimizedArguments) && !op->mightBeType(MIRType::MagicHole) && !op->mightBeType(MIRType::MagicIsConstructing); } bool IonBuilder::compareTrySpecialized(bool* emitted, JSOp op, MDefinition* left, MDefinition* right) { MOZ_ASSERT(*emitted == false); // Try to emit an compare based on the input types. MCompare::CompareType type = MCompare::determineCompareType(op, left, right); if (type == MCompare::Compare_Unknown) return true; MCompare* ins = MCompare::New(alloc(), left, right, op); ins->setCompareType(type); ins->cacheOperandMightEmulateUndefined(constraints()); // Some compare types need to have the specific type in the rhs. // Swap operands if that is not the case. if (type == MCompare::Compare_StrictString && right->type() != MIRType::String) ins->swapOperands(); else if (type == MCompare::Compare_Null && right->type() != MIRType::Null) ins->swapOperands(); else if (type == MCompare::Compare_Undefined && right->type() != MIRType::Undefined) ins->swapOperands(); else if (type == MCompare::Compare_Boolean && right->type() != MIRType::Boolean) ins->swapOperands(); // Replace inputs with unsigned variants if needed. if (type == MCompare::Compare_UInt32) ins->replaceWithUnsignedOperands(); current->add(ins); current->push(ins); MOZ_ASSERT(!ins->isEffectful()); *emitted = true; return true; } bool IonBuilder::compareTryBitwise(bool* emitted, JSOp op, MDefinition* left, MDefinition* right) { MOZ_ASSERT(*emitted == false); // Try to emit a bitwise compare. Check if a bitwise compare equals the wanted // result for all observed operand types. // Onlye allow loose and strict equality. if (op != JSOP_EQ && op != JSOP_NE && op != JSOP_STRICTEQ && op != JSOP_STRICTNE) return true; // Only primitive (not double/string) or objects are supported. // I.e. Undefined/Null/Boolean/Int32 and Object if (!ObjectOrSimplePrimitive(left) || !ObjectOrSimplePrimitive(right)) return true; // Objects that emulate undefined are not supported. if (left->maybeEmulatesUndefined(constraints()) || right->maybeEmulatesUndefined(constraints())) return true; // In the loose comparison more values could be the same, // but value comparison reporting otherwise. if (op == JSOP_EQ || op == JSOP_NE) { // Undefined compared loosy to Null is not supported, // because tag is different, but value can be the same (undefined == null). if ((left->mightBeType(MIRType::Undefined) && right->mightBeType(MIRType::Null)) || (left->mightBeType(MIRType::Null) && right->mightBeType(MIRType::Undefined))) { return true; } // Int32 compared loosy to Boolean is not supported, // because tag is different, but value can be the same (1 == true). if ((left->mightBeType(MIRType::Int32) && right->mightBeType(MIRType::Boolean)) || (left->mightBeType(MIRType::Boolean) && right->mightBeType(MIRType::Int32))) { return true; } // For loosy comparison of an object with a Boolean/Number/String // the valueOf the object is taken. Therefore not supported. bool simpleLHS = left->mightBeType(MIRType::Boolean) || left->mightBeType(MIRType::Int32); bool simpleRHS = right->mightBeType(MIRType::Boolean) || right->mightBeType(MIRType::Int32); if ((left->mightBeType(MIRType::Object) && simpleRHS) || (right->mightBeType(MIRType::Object) && simpleLHS)) { return true; } } MCompare* ins = MCompare::New(alloc(), left, right, op); ins->setCompareType(MCompare::Compare_Bitwise); ins->cacheOperandMightEmulateUndefined(constraints()); current->add(ins); current->push(ins); MOZ_ASSERT(!ins->isEffectful()); *emitted = true; return true; } bool IonBuilder::compareTrySpecializedOnBaselineInspector(bool* emitted, JSOp op, MDefinition* left, MDefinition* right) { MOZ_ASSERT(*emitted == false); // Try to specialize based on any baseline caches that have been generated // for the opcode. These will cause the instruction's type policy to insert // fallible unboxes to the appropriate input types. // Strict equality isn't supported. if (op == JSOP_STRICTEQ || op == JSOP_STRICTNE) return true; MCompare::CompareType type = inspector->expectedCompareType(pc); if (type == MCompare::Compare_Unknown) return true; MCompare* ins = MCompare::New(alloc(), left, right, op); ins->setCompareType(type); ins->cacheOperandMightEmulateUndefined(constraints()); current->add(ins); current->push(ins); MOZ_ASSERT(!ins->isEffectful()); *emitted = true; return true; } bool IonBuilder::compareTrySharedStub(bool* emitted, JSOp op, MDefinition* left, MDefinition* right) { MOZ_ASSERT(*emitted == false); // Try to emit a shared stub cache. if (JitOptions.disableSharedStubs) return true; if (JSOp(*pc) == JSOP_CASE) return true; MBinarySharedStub* stub = MBinarySharedStub::New(alloc(), left, right); current->add(stub); current->push(stub); if (!resumeAfter(stub)) return false; MUnbox* unbox = MUnbox::New(alloc(), current->pop(), MIRType::Boolean, MUnbox::Infallible); current->add(unbox); current->push(unbox); *emitted = true; return true; } bool IonBuilder::newArrayTryTemplateObject(bool* emitted, JSObject* templateObject, uint32_t length) { MOZ_ASSERT(*emitted == false); if (!templateObject) return true; MOZ_ASSERT(length <= NativeObject::MAX_DENSE_ELEMENTS_COUNT); size_t arraySlots = gc::GetGCKindSlots(templateObject->asTenured().getAllocKind()) - ObjectElements::VALUES_PER_HEADER; if (length > arraySlots) return true; // Emit fastpath. gc::InitialHeap heap = templateObject->group()->initialHeap(constraints()); MConstant* templateConst = MConstant::NewConstraintlessObject(alloc(), templateObject); current->add(templateConst); MNewArray* ins = MNewArray::New(alloc(), constraints(), length, templateConst, heap, pc); current->add(ins); current->push(ins); *emitted = true; return true; } bool IonBuilder::newArrayTrySharedStub(bool* emitted) { MOZ_ASSERT(*emitted == false); // Try to emit a shared stub cache. if (JitOptions.disableSharedStubs) return true; if (*pc != JSOP_NEWINIT && *pc != JSOP_NEWARRAY) return true; MInstruction* stub = MNullarySharedStub::New(alloc()); current->add(stub); current->push(stub); if (!resumeAfter(stub)) return false; *emitted = true; return true; } bool IonBuilder::newArrayTryVM(bool* emitted, JSObject* templateObject, uint32_t length) { MOZ_ASSERT(*emitted == false); // Emit a VM call. gc::InitialHeap heap = gc::DefaultHeap; MConstant* templateConst = MConstant::New(alloc(), NullValue()); if (templateObject) { heap = templateObject->group()->initialHeap(constraints()); templateConst = MConstant::NewConstraintlessObject(alloc(), templateObject); } current->add(templateConst); MNewArray* ins = MNewArray::NewVM(alloc(), constraints(), length, templateConst, heap, pc); current->add(ins); current->push(ins); *emitted = true; return true; } bool IonBuilder::jsop_newarray(uint32_t length) { JSObject* templateObject = inspector->getTemplateObject(pc); if (!jsop_newarray(templateObject, length)) return false; // Improve resulting typeset. ObjectGroup* templateGroup = inspector->getTemplateObjectGroup(pc); if (templateGroup) { TemporaryTypeSet* types = MakeSingletonTypeSet(constraints(), templateGroup); current->peek(-1)->setResultTypeSet(types); } return true; } bool IonBuilder::jsop_newarray(JSObject* templateObject, uint32_t length) { bool emitted = false; if (!forceInlineCaches()) { if (!newArrayTryTemplateObject(&emitted, templateObject, length) || emitted) return emitted; } if (!newArrayTrySharedStub(&emitted) || emitted) return emitted; if (!newArrayTryVM(&emitted, templateObject, length) || emitted) return emitted; MOZ_CRASH("newarray should have been emited"); } bool IonBuilder::jsop_newarray_copyonwrite() { ArrayObject* templateObject = ObjectGroup::getCopyOnWriteObject(script(), pc); // The baseline compiler should have ensured the template object has a type // with the copy on write flag set already. During the arguments usage // analysis the baseline compiler hasn't run yet, however, though in this // case the template object's type doesn't matter. MOZ_ASSERT_IF(info().analysisMode() != Analysis_ArgumentsUsage, templateObject->group()->hasAnyFlags(OBJECT_FLAG_COPY_ON_WRITE)); MNewArrayCopyOnWrite* ins = MNewArrayCopyOnWrite::New(alloc(), constraints(), templateObject, templateObject->group()->initialHeap(constraints())); current->add(ins); current->push(ins); return true; } bool IonBuilder::newObjectTryTemplateObject(bool* emitted, JSObject* templateObject) { MOZ_ASSERT(*emitted == false); if (!templateObject) return true; if (templateObject->is() && templateObject->as().hasDynamicSlots()) return true; // Emit fastpath. MNewObject::Mode mode; if (JSOp(*pc) == JSOP_NEWOBJECT || JSOp(*pc) == JSOP_NEWINIT) mode = MNewObject::ObjectLiteral; else mode = MNewObject::ObjectCreate; gc::InitialHeap heap = templateObject->group()->initialHeap(constraints()); MConstant* templateConst = MConstant::NewConstraintlessObject(alloc(), templateObject); current->add(templateConst); MNewObject* ins = MNewObject::New(alloc(), constraints(), templateConst, heap, mode); current->add(ins); current->push(ins); if (!resumeAfter(ins)) return false; *emitted = true; return true; } bool IonBuilder::newObjectTrySharedStub(bool* emitted) { MOZ_ASSERT(*emitted == false); // Try to emit a shared stub cache. if (JitOptions.disableSharedStubs) return true; MInstruction* stub = MNullarySharedStub::New(alloc()); current->add(stub); current->push(stub); if (!resumeAfter(stub)) return false; *emitted = true; return true; } bool IonBuilder::newObjectTryVM(bool* emitted, JSObject* templateObject) { // Emit a VM call. MOZ_ASSERT(JSOp(*pc) == JSOP_NEWOBJECT || JSOp(*pc) == JSOP_NEWINIT); gc::InitialHeap heap = gc::DefaultHeap; MConstant* templateConst = MConstant::New(alloc(), NullValue()); if (templateObject) { heap = templateObject->group()->initialHeap(constraints()); templateConst = MConstant::NewConstraintlessObject(alloc(), templateObject); } current->add(templateConst); MNewObject* ins = MNewObject::NewVM(alloc(), constraints(), templateConst, heap, MNewObject::ObjectLiteral); current->add(ins); current->push(ins); if (!resumeAfter(ins)) return false; *emitted = true; return true; } bool IonBuilder::jsop_newobject() { bool emitted = false; JSObject* templateObject = inspector->getTemplateObject(pc); if (!forceInlineCaches()) { if (!newObjectTryTemplateObject(&emitted, templateObject) || emitted) return emitted; } if (!newObjectTrySharedStub(&emitted) || emitted) return emitted; if (!newObjectTryVM(&emitted, templateObject) || emitted) return emitted; MOZ_CRASH("newobject should have been emited"); } bool IonBuilder::jsop_initelem() { MDefinition* value = current->pop(); MDefinition* id = current->pop(); MDefinition* obj = current->peek(-1); MInitElem* initElem = MInitElem::New(alloc(), obj, id, value); current->add(initElem); return resumeAfter(initElem); } bool IonBuilder::jsop_initelem_array() { MDefinition* value = current->pop(); MDefinition* obj = current->peek(-1); // Make sure that arrays have the type being written to them by the // intializer, and that arrays are marked as non-packed when writing holes // to them during initialization. bool needStub = false; if (shouldAbortOnPreliminaryGroups(obj)) { needStub = true; } else if (!obj->resultTypeSet() || obj->resultTypeSet()->unknownObject() || obj->resultTypeSet()->getObjectCount() != 1) { needStub = true; } else { MOZ_ASSERT(obj->resultTypeSet()->getObjectCount() == 1); TypeSet::ObjectKey* initializer = obj->resultTypeSet()->getObject(0); if (value->type() == MIRType::MagicHole) { if (!initializer->hasFlags(constraints(), OBJECT_FLAG_NON_PACKED)) needStub = true; } else if (!initializer->unknownProperties()) { HeapTypeSetKey elemTypes = initializer->property(JSID_VOID); if (!TypeSetIncludes(elemTypes.maybeTypes(), value->type(), value->resultTypeSet())) { elemTypes.freeze(constraints()); needStub = true; } } } uint32_t index = GET_UINT32(pc); if (needStub) { MCallInitElementArray* store = MCallInitElementArray::New(alloc(), obj, index, value); current->add(store); return resumeAfter(store); } return initializeArrayElement(obj, index, value, /* addResumePoint = */ true); } bool IonBuilder::initializeArrayElement(MDefinition* obj, size_t index, MDefinition* value, bool addResumePointAndIncrementInitializedLength) { MConstant* id = MConstant::New(alloc(), Int32Value(index)); current->add(id); // Get the elements vector. MElements* elements = MElements::New(alloc(), obj); current->add(elements); if (NeedsPostBarrier(value)) current->add(MPostWriteBarrier::New(alloc(), obj, value)); if ((obj->isNewArray() && obj->toNewArray()->convertDoubleElements()) || (obj->isNullarySharedStub() && obj->resultTypeSet()->convertDoubleElements(constraints()) == TemporaryTypeSet::AlwaysConvertToDoubles)) { MInstruction* valueDouble = MToDouble::New(alloc(), value); current->add(valueDouble); value = valueDouble; } // Store the value. MStoreElement* store = MStoreElement::New(alloc(), elements, id, value, /* needsHoleCheck = */ false); current->add(store); if (addResumePointAndIncrementInitializedLength) { // Update the initialized length. (The template object for this // array has the array's ultimate length, so the length field is // already correct: no updating needed.) MSetInitializedLength* initLength = MSetInitializedLength::New(alloc(), elements, id); current->add(initLength); if (!resumeAfter(initLength)) return false; } return true; } bool IonBuilder::jsop_mutateproto() { MDefinition* value = current->pop(); MDefinition* obj = current->peek(-1); MMutateProto* mutate = MMutateProto::New(alloc(), obj, value); current->add(mutate); return resumeAfter(mutate); } bool IonBuilder::jsop_initprop(PropertyName* name) { bool useSlowPath = false; MDefinition* value = current->peek(-1); MDefinition* obj = current->peek(-2); if (obj->isLambda()) { useSlowPath = true; } else if (obj->isNewObject()) { if (JSObject* templateObject = obj->toNewObject()->templateObject()) { if (templateObject->is()) { if (!templateObject->as().containsPure(name)) useSlowPath = true; } } else { useSlowPath = true; } } else { MOZ_ASSERT(obj->isNullarySharedStub()); useSlowPath = true; } if (useSlowPath) { current->pop(); MInitProp* init = MInitProp::New(alloc(), obj, name, value); current->add(init); return resumeAfter(init); } MInstruction* last = *current->rbegin(); // This is definitely initializing an 'own' property of the object, treat // it as an assignment. if (!jsop_setprop(name)) return false; // SETPROP pushed the value, instead of the object. Fix this on the stack, // and check the most recent resume point to see if it needs updating too. current->pop(); current->push(obj); for (MInstructionReverseIterator riter = current->rbegin(); *riter != last; riter++) { if (MResumePoint* resumePoint = riter->resumePoint()) { MOZ_ASSERT(resumePoint->pc() == pc); if (resumePoint->mode() == MResumePoint::ResumeAfter) { size_t index = resumePoint->numOperands() - 1; resumePoint->replaceOperand(index, obj); } break; } } return true; } bool IonBuilder::jsop_initprop_getter_setter(PropertyName* name) { MDefinition* value = current->pop(); MDefinition* obj = current->peek(-1); MInitPropGetterSetter* init = MInitPropGetterSetter::New(alloc(), obj, name, value); current->add(init); return resumeAfter(init); } bool IonBuilder::jsop_initelem_getter_setter() { MDefinition* value = current->pop(); MDefinition* id = current->pop(); MDefinition* obj = current->peek(-1); MInitElemGetterSetter* init = MInitElemGetterSetter::New(alloc(), obj, id, value); current->add(init); return resumeAfter(init); } MBasicBlock* IonBuilder::addBlock(MBasicBlock* block, uint32_t loopDepth) { if (!block) return nullptr; if (block->pc() && script()->hasScriptCounts()) block->setHitCount(script()->getHitCount(block->pc())); graph().addBlock(block); block->setLoopDepth(loopDepth); return block; } MBasicBlock* IonBuilder::newBlock(MBasicBlock* predecessor, jsbytecode* pc) { MBasicBlock* block = MBasicBlock::New(graph(), &analysis(), info(), predecessor, bytecodeSite(pc), MBasicBlock::NORMAL); return addBlock(block, loopDepth_); } MBasicBlock* IonBuilder::newBlock(MBasicBlock* predecessor, jsbytecode* pc, MResumePoint* priorResumePoint) { MBasicBlock* block = MBasicBlock::NewWithResumePoint(graph(), info(), predecessor, bytecodeSite(pc), priorResumePoint); return addBlock(block, loopDepth_); } MBasicBlock* IonBuilder::newBlockPopN(MBasicBlock* predecessor, jsbytecode* pc, uint32_t popped) { MBasicBlock* block = MBasicBlock::NewPopN(graph(), info(), predecessor, bytecodeSite(pc), MBasicBlock::NORMAL, popped); return addBlock(block, loopDepth_); } MBasicBlock* IonBuilder::newBlockAfter(MBasicBlock* at, MBasicBlock* predecessor, jsbytecode* pc) { MBasicBlock* block = MBasicBlock::New(graph(), &analysis(), info(), predecessor, bytecodeSite(pc), MBasicBlock::NORMAL); if (!block) return nullptr; block->setHitCount(0); // osr block graph().insertBlockAfter(at, block); return block; } MBasicBlock* IonBuilder::newBlock(MBasicBlock* predecessor, jsbytecode* pc, uint32_t loopDepth) { MBasicBlock* block = MBasicBlock::New(graph(), &analysis(), info(), predecessor, bytecodeSite(pc), MBasicBlock::NORMAL); return addBlock(block, loopDepth); } MBasicBlock* IonBuilder::newOsrPreheader(MBasicBlock* predecessor, jsbytecode* loopEntry, jsbytecode* beforeLoopEntry) { MOZ_ASSERT(LoopEntryCanIonOsr(loopEntry)); MOZ_ASSERT(loopEntry == info().osrPc()); // Create two blocks: one for the OSR entry with no predecessors, one for // the preheader, which has the OSR entry block as a predecessor. The // OSR block is always the second block (with id 1). MBasicBlock* osrBlock = newBlockAfter(*graph().begin(), loopEntry); MBasicBlock* preheader = newBlock(predecessor, loopEntry); if (!osrBlock || !preheader) return nullptr; // Give the pre-header the same hit count as the code before the loop. if (script()->hasScriptCounts()) preheader->setHitCount(script()->getHitCount(beforeLoopEntry)); MOsrEntry* entry = MOsrEntry::New(alloc()); osrBlock->add(entry); // Initialize |envChain|. { uint32_t slot = info().environmentChainSlot(); MInstruction* envv; if (analysis().usesEnvironmentChain()) { envv = MOsrEnvironmentChain::New(alloc(), entry); } else { // Use an undefined value if the script does not need its env // chain, to match the type that is already being tracked for the // slot. envv = MConstant::New(alloc(), UndefinedValue()); } osrBlock->add(envv); osrBlock->initSlot(slot, envv); } // Initialize |return value| { MInstruction* returnValue; if (!script()->noScriptRval()) returnValue = MOsrReturnValue::New(alloc(), entry); else returnValue = MConstant::New(alloc(), UndefinedValue()); osrBlock->add(returnValue); osrBlock->initSlot(info().returnValueSlot(), returnValue); } // Initialize arguments object. bool needsArgsObj = info().needsArgsObj(); MInstruction* argsObj = nullptr; if (info().hasArguments()) { if (needsArgsObj) argsObj = MOsrArgumentsObject::New(alloc(), entry); else argsObj = MConstant::New(alloc(), UndefinedValue()); osrBlock->add(argsObj); osrBlock->initSlot(info().argsObjSlot(), argsObj); } if (info().funMaybeLazy()) { // Initialize |this| parameter. MParameter* thisv = MParameter::New(alloc(), MParameter::THIS_SLOT, nullptr); osrBlock->add(thisv); osrBlock->initSlot(info().thisSlot(), thisv); // Initialize arguments. for (uint32_t i = 0; i < info().nargs(); i++) { uint32_t slot = needsArgsObj ? info().argSlotUnchecked(i) : info().argSlot(i); // Only grab arguments from the arguments object if the arguments object // aliases formals. If the argsobj does not alias formals, then the // formals may have been assigned to during interpretation, and that change // will not be reflected in the argsobj. if (needsArgsObj && info().argsObjAliasesFormals()) { MOZ_ASSERT(argsObj && argsObj->isOsrArgumentsObject()); // If this is an aliased formal, then the arguments object // contains a hole at this index. Any references to this // variable in the jitcode will come from JSOP_*ALIASEDVAR // opcodes, so the slot itself can be set to undefined. If // it's not aliased, it must be retrieved from the arguments // object. MInstruction* osrv; if (script()->formalIsAliased(i)) osrv = MConstant::New(alloc(), UndefinedValue()); else osrv = MGetArgumentsObjectArg::New(alloc(), argsObj, i); osrBlock->add(osrv); osrBlock->initSlot(slot, osrv); } else { MParameter* arg = MParameter::New(alloc(), i, nullptr); osrBlock->add(arg); osrBlock->initSlot(slot, arg); } } } // Initialize locals. for (uint32_t i = 0; i < info().nlocals(); i++) { uint32_t slot = info().localSlot(i); ptrdiff_t offset = BaselineFrame::reverseOffsetOfLocal(i); MOsrValue* osrv = MOsrValue::New(alloc().fallible(), entry, offset); if (!osrv) return nullptr; osrBlock->add(osrv); osrBlock->initSlot(slot, osrv); } // Initialize stack. uint32_t numStackSlots = preheader->stackDepth() - info().firstStackSlot(); for (uint32_t i = 0; i < numStackSlots; i++) { uint32_t slot = info().stackSlot(i); ptrdiff_t offset = BaselineFrame::reverseOffsetOfLocal(info().nlocals() + i); MOsrValue* osrv = MOsrValue::New(alloc().fallible(), entry, offset); if (!osrv) return nullptr; osrBlock->add(osrv); osrBlock->initSlot(slot, osrv); } // Create an MStart to hold the first valid MResumePoint. MStart* start = MStart::New(alloc()); osrBlock->add(start); // MOsrValue instructions are infallible, so the first MResumePoint must // occur after they execute, at the point of the MStart. if (!resumeAt(start, loopEntry)) return nullptr; // Link the same MResumePoint from the MStart to each MOsrValue. // This causes logic in ShouldSpecializeInput() to not replace Uses with // Unboxes in the MResumePiont, so that the MStart always sees Values. if (!osrBlock->linkOsrValues(start)) return nullptr; // Clone types of the other predecessor of the pre-header to the osr block, // such as pre-header phi's won't discard specialized type of the // predecessor. MOZ_ASSERT(predecessor->stackDepth() == osrBlock->stackDepth()); MOZ_ASSERT(info().environmentChainSlot() == 0); // Treat the OSR values as having the same type as the existing values // coming in to the loop. These will be fixed up with appropriate // unboxing and type barriers in finishLoop, once the possible types // at the loop header are known. for (uint32_t i = info().startArgSlot(); i < osrBlock->stackDepth(); i++) { MDefinition* existing = current->getSlot(i); MDefinition* def = osrBlock->getSlot(i); MOZ_ASSERT_IF(!needsArgsObj || !info().isSlotAliased(i), def->type() == MIRType::Value); // Aliased slots are never accessed, since they need to go through // the callobject. No need to type them here. if (info().isSlotAliased(i)) continue; def->setResultType(existing->type()); def->setResultTypeSet(existing->resultTypeSet()); } // Finish the osrBlock. osrBlock->end(MGoto::New(alloc(), preheader)); if (!preheader->addPredecessor(alloc(), osrBlock)) return nullptr; graph().setOsrBlock(osrBlock); return preheader; } MBasicBlock* IonBuilder::newPendingLoopHeader(MBasicBlock* predecessor, jsbytecode* pc, bool osr, bool canOsr, unsigned stackPhiCount) { loopDepth_++; // If this site can OSR, all values on the expression stack are part of the loop. if (canOsr) stackPhiCount = predecessor->stackDepth() - info().firstStackSlot(); MBasicBlock* block = MBasicBlock::NewPendingLoopHeader(graph(), info(), predecessor, bytecodeSite(pc), stackPhiCount); if (!addBlock(block, loopDepth_)) return nullptr; if (osr) { // Incorporate type information from the OSR frame into the loop // header. The OSR frame may have unexpected types due to type changes // within the loop body or due to incomplete profiling information, // in which case this may avoid restarts of loop analysis or bailouts // during the OSR itself. MOZ_ASSERT(info().firstLocalSlot() - info().firstArgSlot() == baselineFrame_->argTypes.length()); MOZ_ASSERT(block->stackDepth() - info().firstLocalSlot() == baselineFrame_->varTypes.length()); // Unbox the MOsrValue if it is known to be unboxable. for (uint32_t i = info().startArgSlot(); i < block->stackDepth(); i++) { // The value of aliased args and slots are in the callobject. So we can't // the value from the baseline frame. if (info().isSlotAliased(i)) continue; MPhi* phi = block->getSlot(i)->toPhi(); // Get the type from the baseline frame. TypeSet::Type existingType = TypeSet::UndefinedType(); uint32_t arg = i - info().firstArgSlot(); uint32_t var = i - info().firstLocalSlot(); if (info().funMaybeLazy() && i == info().thisSlot()) existingType = baselineFrame_->thisType; else if (arg < info().nargs()) existingType = baselineFrame_->argTypes[arg]; else existingType = baselineFrame_->varTypes[var]; if (existingType.isSingletonUnchecked()) checkNurseryObject(existingType.singleton()); // Extract typeset from value. LifoAlloc* lifoAlloc = alloc().lifoAlloc(); TemporaryTypeSet* typeSet = lifoAlloc->new_(lifoAlloc, existingType); if (!typeSet) return nullptr; MIRType type = typeSet->getKnownMIRType(); if (!phi->addBackedgeType(alloc(), type, typeSet)) return nullptr; } } return block; } MTest* IonBuilder::newTest(MDefinition* ins, MBasicBlock* ifTrue, MBasicBlock* ifFalse) { MTest* test = MTest::New(alloc(), ins, ifTrue, ifFalse); test->cacheOperandMightEmulateUndefined(constraints()); return test; } // A resume point is a mapping of stack slots to MDefinitions. It is used to // capture the environment such that if a guard fails, and IonMonkey needs // to exit back to the interpreter, the interpreter state can be // reconstructed. // // We capture stack state at critical points: // * (1) At the beginning of every basic block. // * (2) After every effectful operation. // // As long as these two properties are maintained, instructions can // be moved, hoisted, or, eliminated without problems, and ops without side // effects do not need to worry about capturing state at precisely the // right point in time. // // Effectful instructions, of course, need to capture state after completion, // where the interpreter will not attempt to repeat the operation. For this, // ResumeAfter must be used. The state is attached directly to the effectful // instruction to ensure that no intermediate instructions could be injected // in between by a future analysis pass. // // During LIR construction, if an instruction can bail back to the interpreter, // we create an LSnapshot, which uses the last known resume point to request // register/stack assignments for every live value. bool IonBuilder::resume(MInstruction* ins, jsbytecode* pc, MResumePoint::Mode mode) { MOZ_ASSERT(ins->isEffectful() || !ins->isMovable()); MResumePoint* resumePoint = MResumePoint::New(alloc(), ins->block(), pc, mode); if (!resumePoint) { abortReason_ = AbortReason_Alloc; return false; } ins->setResumePoint(resumePoint); return true; } bool IonBuilder::resumeAt(MInstruction* ins, jsbytecode* pc) { return resume(ins, pc, MResumePoint::ResumeAt); } bool IonBuilder::resumeAfter(MInstruction* ins) { return resume(ins, pc, MResumePoint::ResumeAfter); } bool IonBuilder::maybeInsertResume() { // Create a resume point at the current position, without an existing // effectful instruction. This resume point is not necessary for correct // behavior (see above), but is added to avoid holding any values from the // previous resume point which are now dead. This shortens the live ranges // of such values and improves register allocation. // // This optimization is not performed outside of loop bodies, where good // register allocation is not as critical, in order to avoid creating // excessive resume points. if (loopDepth_ == 0) return true; MNop* ins = MNop::New(alloc()); current->add(ins); return resumeAfter(ins); } void IonBuilder::maybeMarkEmpty(MDefinition* ins) { MOZ_ASSERT(ins->type() == MIRType::Value); // When one of the operands has no type information, mark the output // as having no possible types too. This is to avoid degrading // subsequent analysis. for (size_t i = 0; i < ins->numOperands(); i++) { if (!ins->emptyResultTypeSet()) continue; TemporaryTypeSet* types = alloc().lifoAlloc()->new_(); if (types) ins->setResultTypeSet(types); } } // Return whether property lookups can be performed effectlessly on clasp. static bool ClassHasEffectlessLookup(const Class* clasp) { return IsTypedObjectClass(clasp) || (clasp->isNative() && !clasp->getOpsLookupProperty()); } // Return whether an object might have a property for name which is not // accounted for by type information. static bool ObjectHasExtraOwnProperty(CompileCompartment* comp, TypeSet::ObjectKey* object, jsid id) { // Some typed object properties are not reflected in type information. if (object->isGroup() && object->group()->maybeTypeDescr()) return object->group()->typeDescr().hasProperty(comp->runtime()->names(), id); const Class* clasp = object->clasp(); // Array |length| properties are not reflected in type information. if (clasp == &ArrayObject::class_) return JSID_IS_ATOM(id, comp->runtime()->names().length); // Resolve hooks can install new properties on objects on demand. JSObject* singleton = object->isSingleton() ? object->singleton() : nullptr; return ClassMayResolveId(comp->runtime()->names(), clasp, id, singleton); } void IonBuilder::insertRecompileCheck() { // No need for recompile checks if this is the highest optimization level. OptimizationLevel curLevel = optimizationInfo().level(); if (IonOptimizations.isLastLevel(curLevel)) return; // Add recompile check. // Get the topmost builder. The topmost script will get recompiled when // warm-up counter is high enough to justify a higher optimization level. IonBuilder* topBuilder = outermostBuilder(); // Add recompile check to recompile when the warm-up count reaches the // threshold of the next optimization level. OptimizationLevel nextLevel = IonOptimizations.nextLevel(curLevel); const OptimizationInfo* info = IonOptimizations.get(nextLevel); uint32_t warmUpThreshold = info->compilerWarmUpThreshold(topBuilder->script()); MRecompileCheck* check = MRecompileCheck::New(alloc(), topBuilder->script(), warmUpThreshold, MRecompileCheck::RecompileCheck_OptimizationLevel); current->add(check); } JSObject* IonBuilder::testSingletonProperty(JSObject* obj, jsid id) { // We would like to completely no-op property/global accesses which can // produce only a particular JSObject. When indicating the access result is // definitely an object, type inference does not account for the // possibility that the property is entirely missing from the input object // and its prototypes (if this happens, a semantic trigger would be hit and // the pushed types updated, even if there is no type barrier). // // If the access definitely goes through obj, either directly or on the // prototype chain, and the object has singleton type, then the type // information for that property reflects the value that will definitely be // read on accesses to the object. If the property is later deleted or // reconfigured as a getter/setter then the type information for the // property will change and trigger invalidation. while (obj) { if (!ClassHasEffectlessLookup(obj->getClass())) return nullptr; TypeSet::ObjectKey* objKey = TypeSet::ObjectKey::get(obj); if (analysisContext) objKey->ensureTrackedProperty(analysisContext, id); if (objKey->unknownProperties()) return nullptr; HeapTypeSetKey property = objKey->property(id); if (property.isOwnProperty(constraints())) { if (obj->isSingleton()) return property.singleton(constraints()); return nullptr; } if (ObjectHasExtraOwnProperty(compartment, objKey, id)) return nullptr; obj = checkNurseryObject(obj->staticPrototype()); } return nullptr; } JSObject* IonBuilder::testSingletonPropertyTypes(MDefinition* obj, jsid id) { // As for TestSingletonProperty, but the input is any value in a type set // rather than a specific object. TemporaryTypeSet* types = obj->resultTypeSet(); if (types && types->unknownObject()) return nullptr; JSObject* objectSingleton = types ? types->maybeSingleton() : nullptr; if (objectSingleton) return testSingletonProperty(objectSingleton, id); MIRType objType = obj->type(); if (objType == MIRType::Value && types) objType = types->getKnownMIRType(); JSProtoKey key; switch (objType) { case MIRType::String: key = JSProto_String; break; case MIRType::Symbol: key = JSProto_Symbol; break; case MIRType::Int32: case MIRType::Double: key = JSProto_Number; break; case MIRType::Boolean: key = JSProto_Boolean; break; case MIRType::Object: { if (!types) return nullptr; // For property accesses which may be on many objects, we just need to // find a prototype common to all the objects; if that prototype // has the singleton property, the access will not be on a missing property. JSObject* singleton = nullptr; for (unsigned i = 0; i < types->getObjectCount(); i++) { TypeSet::ObjectKey* key = types->getObject(i); if (!key) continue; if (analysisContext) key->ensureTrackedProperty(analysisContext, id); const Class* clasp = key->clasp(); if (!ClassHasEffectlessLookup(clasp) || ObjectHasExtraOwnProperty(compartment, key, id)) return nullptr; if (key->unknownProperties()) return nullptr; HeapTypeSetKey property = key->property(id); if (property.isOwnProperty(constraints())) return nullptr; if (JSObject* proto = checkNurseryObject(key->proto().toObjectOrNull())) { // Test this type. JSObject* thisSingleton = testSingletonProperty(proto, id); if (!thisSingleton) return nullptr; if (singleton) { if (thisSingleton != singleton) return nullptr; } else { singleton = thisSingleton; } } else { // Can't be on the prototype chain with no prototypes... return nullptr; } } return singleton; } default: return nullptr; } JSObject* proto = GetBuiltinPrototypePure(&script()->global(), key); if (proto) return testSingletonProperty(proto, id); return nullptr; } ResultWithOOM IonBuilder::testNotDefinedProperty(MDefinition* obj, jsid id) { TemporaryTypeSet* types = obj->resultTypeSet(); if (!types || types->unknownObject() || types->getKnownMIRType() != MIRType::Object) return ResultWithOOM::ok(false); for (unsigned i = 0, count = types->getObjectCount(); i < count; i++) { TypeSet::ObjectKey* key = types->getObject(i); if (!key) continue; while (true) { if (!alloc().ensureBallast()) return ResultWithOOM::fail(); if (!key->hasStableClassAndProto(constraints()) || key->unknownProperties()) return ResultWithOOM::ok(false); const Class* clasp = key->clasp(); if (!ClassHasEffectlessLookup(clasp) || ObjectHasExtraOwnProperty(compartment, key, id)) return ResultWithOOM::ok(false); // If the object is a singleton, we can do a lookup now to avoid // unnecessary invalidations later on, in case the property types // have not yet been instantiated. if (key->isSingleton() && key->singleton()->is() && key->singleton()->as().lookupPure(id)) { return ResultWithOOM::ok(false); } HeapTypeSetKey property = key->property(id); if (property.isOwnProperty(constraints())) return ResultWithOOM::ok(false); JSObject* proto = checkNurseryObject(key->proto().toObjectOrNull()); if (!proto) break; key = TypeSet::ObjectKey::get(proto); } } return ResultWithOOM::ok(true); } bool IonBuilder::pushTypeBarrier(MDefinition* def, TemporaryTypeSet* observed, BarrierKind kind) { MOZ_ASSERT(def == current->peek(-1)); MDefinition* replace = addTypeBarrier(current->pop(), observed, kind); if (!replace) return false; current->push(replace); return true; } // Given an observed type set, annotates the IR as much as possible: // (1) If no type information is provided, the given value is returned. // (2) If a single type definitely exists, and no type barrier is needed, // then an infallible unbox instruction is returned. // (3) If a type barrier is needed, but has an unknown type set, the given // value is returned. // (4) Lastly, a type barrier instruction is added and returned. MDefinition* IonBuilder::addTypeBarrier(MDefinition* def, TemporaryTypeSet* observed, BarrierKind kind, MTypeBarrier** pbarrier) { // Barriers are never needed for instructions whose result will not be used. if (BytecodeIsPopped(pc)) return def; // If the instruction has no side effects, we'll resume the entire operation. // The actual type barrier will occur in the interpreter. If the // instruction is effectful, even if it has a singleton type, there // must be a resume point capturing the original def, and resuming // to that point will explicitly monitor the new type. if (kind == BarrierKind::NoBarrier) { MDefinition* replace = ensureDefiniteType(def, observed->getKnownMIRType()); replace->setResultTypeSet(observed); return replace; } if (observed->unknown()) return def; MTypeBarrier* barrier = MTypeBarrier::New(alloc(), def, observed, kind); current->add(barrier); if (pbarrier) *pbarrier = barrier; if (barrier->type() == MIRType::Undefined) return constant(UndefinedValue()); if (barrier->type() == MIRType::Null) return constant(NullValue()); return barrier; } bool IonBuilder::pushDOMTypeBarrier(MInstruction* ins, TemporaryTypeSet* observed, JSFunction* func) { MOZ_ASSERT(func && func->isNative() && func->jitInfo()); const JSJitInfo* jitinfo = func->jitInfo(); bool barrier = DOMCallNeedsBarrier(jitinfo, observed); // Need to be a bit careful: if jitinfo->returnType is JSVAL_TYPE_DOUBLE but // types->getKnownMIRType() is MIRType::Int32, then don't unconditionally // unbox as a double. Instead, go ahead and barrier on having an int type, // since we know we need a barrier anyway due to the type mismatch. This is // the only situation in which TI actually has more information about the // JSValueType than codegen can, short of jitinfo->returnType just being // JSVAL_TYPE_UNKNOWN. MDefinition* replace = ins; if (jitinfo->returnType() != JSVAL_TYPE_DOUBLE || observed->getKnownMIRType() != MIRType::Int32) { replace = ensureDefiniteType(ins, MIRTypeFromValueType(jitinfo->returnType())); if (replace != ins) { current->pop(); current->push(replace); } } else { MOZ_ASSERT(barrier); } return pushTypeBarrier(replace, observed, barrier ? BarrierKind::TypeSet : BarrierKind::NoBarrier); } MDefinition* IonBuilder::ensureDefiniteType(MDefinition* def, MIRType definiteType) { MInstruction* replace; switch (definiteType) { case MIRType::Undefined: def->setImplicitlyUsedUnchecked(); replace = MConstant::New(alloc(), UndefinedValue()); break; case MIRType::Null: def->setImplicitlyUsedUnchecked(); replace = MConstant::New(alloc(), NullValue()); break; case MIRType::Value: return def; default: { if (def->type() != MIRType::Value) { if (def->type() == MIRType::Int32 && definiteType == MIRType::Double) { replace = MToDouble::New(alloc(), def); break; } MOZ_ASSERT(def->type() == definiteType); return def; } replace = MUnbox::New(alloc(), def, definiteType, MUnbox::Infallible); break; } } current->add(replace); return replace; } MDefinition* IonBuilder::ensureDefiniteTypeSet(MDefinition* def, TemporaryTypeSet* types) { // We cannot arbitrarily add a typeset to a definition. It can be shared // in another path. So we always need to create a new MIR. // Use ensureDefiniteType to do unboxing. If that happened the type can // be added on the newly created unbox operation. MDefinition* replace = ensureDefiniteType(def, types->getKnownMIRType()); if (replace != def) { replace->setResultTypeSet(types); return replace; } // Don't replace if input type is more accurate than given typeset. if (def->type() != types->getKnownMIRType()) { MOZ_ASSERT(types->getKnownMIRType() == MIRType::Value); return def; } // Create a NOP mir instruction to filter the typeset. MFilterTypeSet* filter = MFilterTypeSet::New(alloc(), def, types); current->add(filter); return filter; } static size_t NumFixedSlots(JSObject* object) { // Note: we can't use object->numFixedSlots() here, as this will read the // shape and can race with the main thread if we are building off thread. // The allocation kind and object class (which goes through the type) can // be read freely, however. gc::AllocKind kind = object->asTenured().getAllocKind(); return gc::GetGCKindSlots(kind, object->getClass()); } static bool IsUninitializedGlobalLexicalSlot(JSObject* obj, PropertyName* name) { LexicalEnvironmentObject &globalLexical = obj->as(); MOZ_ASSERT(globalLexical.isGlobal()); Shape* shape = globalLexical.lookupPure(name); if (!shape) return false; return globalLexical.getSlot(shape->slot()).isMagic(JS_UNINITIALIZED_LEXICAL); } bool IonBuilder::getStaticName(JSObject* staticObject, PropertyName* name, bool* psucceeded, MDefinition* lexicalCheck) { MOZ_ASSERT(*psucceeded == false); jsid id = NameToId(name); bool isGlobalLexical = staticObject->is() && staticObject->as().isGlobal(); MOZ_ASSERT(isGlobalLexical || staticObject->is() || staticObject->is() || staticObject->is()); MOZ_ASSERT(staticObject->isSingleton()); *psucceeded = true; // Always emit the lexical check. This could be optimized, but is // currently not for simplicity's sake. if (lexicalCheck) { *psucceeded = false; return true; } TypeSet::ObjectKey* staticKey = TypeSet::ObjectKey::get(staticObject); if (analysisContext) staticKey->ensureTrackedProperty(analysisContext, NameToId(name)); if (staticKey->unknownProperties()) { *psucceeded = false; return true; } HeapTypeSetKey property = staticKey->property(id); if (!property.maybeTypes() || !property.maybeTypes()->definiteProperty() || property.nonData(constraints())) { // The property has been reconfigured as non-configurable, non-enumerable // or non-writable. *psucceeded = false; return true; } // Don't optimize global lexical bindings if they aren't initialized at // compile time. if (isGlobalLexical && IsUninitializedGlobalLexicalSlot(staticObject, name)) { *psucceeded = false; return true; } TemporaryTypeSet* types = bytecodeTypes(pc); BarrierKind barrier = PropertyReadNeedsTypeBarrier(analysisContext, constraints(), staticKey, name, types, /* updateObserved = */ true); if (barrier == BarrierKind::NoBarrier) { // Try to inline properties holding a known constant object. JSObject* singleton = types->maybeSingleton(); if (singleton) { if (testSingletonProperty(staticObject, id) == singleton) { pushConstant(ObjectValue(*singleton)); return true; } } // Try to inline properties that have never been overwritten. Value constantValue; if (property.constant(constraints(), &constantValue)) { pushConstant(constantValue); return true; } } if (!loadStaticSlot(staticObject, barrier, types, property.maybeTypes()->definiteSlot())) { *psucceeded = false; return false; } return true; } bool IonBuilder::loadStaticSlot(JSObject* staticObject, BarrierKind barrier, TemporaryTypeSet* types, uint32_t slot) { if (barrier == BarrierKind::NoBarrier) { // Try to inline properties that can only have one value. MIRType knownType = types->getKnownMIRType(); if (knownType == MIRType::Undefined) { pushConstant(UndefinedValue()); return true; } if (knownType == MIRType::Null) { pushConstant(NullValue()); return true; } } MInstruction* obj = constant(ObjectValue(*staticObject)); MIRType rvalType = types->getKnownMIRType(); if (barrier != BarrierKind::NoBarrier) rvalType = MIRType::Value; return loadSlot(obj, slot, NumFixedSlots(staticObject), rvalType, barrier, types); } // Whether a write of the given value may need a post-write barrier for GC purposes. bool jit::NeedsPostBarrier(MDefinition* value) { if (!GetJitContext()->runtime->gcNursery().exists()) return false; return value->mightBeType(MIRType::Object); } bool IonBuilder::setStaticName(JSObject* staticObject, PropertyName* name) { jsid id = NameToId(name); bool isGlobalLexical = staticObject->is() && staticObject->as().isGlobal(); MOZ_ASSERT(isGlobalLexical || staticObject->is() || staticObject->is()); MDefinition* value = current->peek(-1); TypeSet::ObjectKey* staticKey = TypeSet::ObjectKey::get(staticObject); if (staticKey->unknownProperties()) return jsop_setprop(name); HeapTypeSetKey property = staticKey->property(id); if (!property.maybeTypes() || !property.maybeTypes()->definiteProperty() || property.nonData(constraints()) || property.nonWritable(constraints())) { // The property has been reconfigured as non-configurable, non-enumerable // or non-writable. return jsop_setprop(name); } if (!CanWriteProperty(alloc(), constraints(), property, value)) return jsop_setprop(name); // Don't optimize global lexical bindings if they aren't initialized at // compile time. if (isGlobalLexical && IsUninitializedGlobalLexicalSlot(staticObject, name)) return jsop_setprop(name); current->pop(); // Pop the bound object on the stack. MDefinition* obj = current->pop(); MOZ_ASSERT(&obj->toConstant()->toObject() == staticObject); if (NeedsPostBarrier(value)) current->add(MPostWriteBarrier::New(alloc(), obj, value)); // If the property has a known type, we may be able to optimize typed stores by not // storing the type tag. MIRType slotType = MIRType::None; MIRType knownType = property.knownMIRType(constraints()); if (knownType != MIRType::Value) slotType = knownType; bool needsBarrier = property.needsBarrier(constraints()); return storeSlot(obj, property.maybeTypes()->definiteSlot(), NumFixedSlots(staticObject), value, needsBarrier, slotType); } JSObject* IonBuilder::testGlobalLexicalBinding(PropertyName* name) { MOZ_ASSERT(JSOp(*pc) == JSOP_BINDGNAME || JSOp(*pc) == JSOP_GETGNAME || JSOp(*pc) == JSOP_SETGNAME || JSOp(*pc) == JSOP_STRICTSETGNAME); // The global isn't the global lexical env's prototype, but its enclosing // env. Test for the existence of |name| manually on the global lexical // env. If it is not found, look for it on the global itself. NativeObject* obj = &script()->global().lexicalEnvironment(); TypeSet::ObjectKey* lexicalKey = TypeSet::ObjectKey::get(obj); jsid id = NameToId(name); if (analysisContext) lexicalKey->ensureTrackedProperty(analysisContext, id); // If the property is not found on the global lexical env but it is found // on the global and is configurable, try to freeze the typeset for its // non-existence. If we don't have type information then fail. // // In the case that it is found on the global but is non-configurable, // the binding cannot be shadowed by a global lexical binding. Maybe lexicalProperty; if (!lexicalKey->unknownProperties()) lexicalProperty.emplace(lexicalKey->property(id)); Shape* shape = obj->lookupPure(name); if (shape) { if ((JSOp(*pc) != JSOP_GETGNAME && !shape->writable()) || obj->getSlot(shape->slot()).isMagic(JS_UNINITIALIZED_LEXICAL)) { return nullptr; } } else { shape = script()->global().lookupPure(name); if (!shape || shape->configurable()) { if (lexicalProperty.isSome()) MOZ_ALWAYS_FALSE(lexicalProperty->isOwnProperty(constraints())); else return nullptr; } obj = &script()->global(); } return obj; } bool IonBuilder::jsop_getgname(PropertyName* name) { // Optimize undefined/NaN/Infinity first. We must ensure we handle these // cases *exactly* like Baseline, because it's invalid to add an Ion IC or // VM call (that might trigger invalidation) if there's no Baseline IC for // this op. if (name == names().undefined) { pushConstant(UndefinedValue()); return true; } if (name == names().NaN) { pushConstant(compartment->runtime()->NaNValue()); return true; } if (name == names().Infinity) { pushConstant(compartment->runtime()->positiveInfinityValue()); return true; } if (JSObject* obj = testGlobalLexicalBinding(name)) { bool emitted = false; if (!getStaticName(obj, name, &emitted) || emitted) return emitted; if (!forceInlineCaches() && obj->is()) { TemporaryTypeSet* types = bytecodeTypes(pc); MDefinition* globalObj = constant(ObjectValue(*obj)); if (!getPropTryCommonGetter(&emitted, globalObj, name, types) || emitted) return emitted; } } return jsop_getname(name); } bool IonBuilder::jsop_getname(PropertyName* name) { MDefinition* object; if (IsGlobalOp(JSOp(*pc)) && !script()->hasNonSyntacticScope()) { MInstruction* global = constant(ObjectValue(script()->global().lexicalEnvironment())); object = global; } else { current->push(current->environmentChain()); object = current->pop(); } MGetNameCache* ins; if (JSOp(*GetNextPc(pc)) == JSOP_TYPEOF) ins = MGetNameCache::New(alloc(), object, name, MGetNameCache::NAMETYPEOF); else ins = MGetNameCache::New(alloc(), object, name, MGetNameCache::NAME); current->add(ins); current->push(ins); if (!resumeAfter(ins)) return false; TemporaryTypeSet* types = bytecodeTypes(pc); return pushTypeBarrier(ins, types, BarrierKind::TypeSet); } bool IonBuilder::jsop_intrinsic(PropertyName* name) { TemporaryTypeSet* types = bytecodeTypes(pc); Value vp = UndefinedValue(); // If the intrinsic value doesn't yet exist, we haven't executed this // opcode yet, so we need to get it and monitor the result. if (!script()->global().maybeExistingIntrinsicValue(name, &vp)) { MCallGetIntrinsicValue* ins = MCallGetIntrinsicValue::New(alloc(), name); current->add(ins); current->push(ins); if (!resumeAfter(ins)) return false; return pushTypeBarrier(ins, types, BarrierKind::TypeSet); } if (types->empty()) types->addType(TypeSet::GetValueType(vp), alloc().lifoAlloc()); // Bake in the intrinsic, guaranteed to exist because a non-empty typeset // means the intrinsic was successfully gotten in the VM call above. // Assert that TI agrees with us on the type. MOZ_ASSERT(types->hasType(TypeSet::GetValueType(vp))); pushConstant(vp); return true; } bool IonBuilder::jsop_getimport(PropertyName* name) { ModuleEnvironmentObject* env = GetModuleEnvironmentForScript(script()); MOZ_ASSERT(env); Shape* shape; ModuleEnvironmentObject* targetEnv; MOZ_ALWAYS_TRUE(env->lookupImport(NameToId(name), &targetEnv, &shape)); PropertyName* localName = JSID_TO_STRING(shape->propid())->asAtom().asPropertyName(); bool emitted = false; if (!getStaticName(targetEnv, localName, &emitted)) return false; if (!emitted) { // This can happen if we don't have type information. TypeSet::ObjectKey* staticKey = TypeSet::ObjectKey::get(targetEnv); TemporaryTypeSet* types = bytecodeTypes(pc); BarrierKind barrier = PropertyReadNeedsTypeBarrier(analysisContext, constraints(), staticKey, name, types, /* updateObserved = */ true); if (!loadStaticSlot(targetEnv, barrier, types, shape->slot())) return false; } // In the rare case where this import hasn't been initialized already (we // have an import cycle where modules reference each other's imports), emit // a check. if (targetEnv->getSlot(shape->slot()).isMagic(JS_UNINITIALIZED_LEXICAL)) { MDefinition* checked = addLexicalCheck(current->pop()); current->push(checked); } return true; } bool IonBuilder::jsop_bindname(PropertyName* name) { MDefinition* envChain; if (analysis().usesEnvironmentChain()) { envChain = current->environmentChain(); } else { // We take the slow path when trying to BINDGNAME a name that resolves // to a 'const' or an uninitialized binding. MOZ_ASSERT(JSOp(*pc) == JSOP_BINDGNAME); envChain = constant(ObjectValue(script()->global().lexicalEnvironment())); } MBindNameCache* ins = MBindNameCache::New(alloc(), envChain, name, script(), pc); current->add(ins); current->push(ins); return resumeAfter(ins); } bool IonBuilder::jsop_bindvar() { MOZ_ASSERT(analysis().usesEnvironmentChain()); MCallBindVar* ins = MCallBindVar::New(alloc(), current->environmentChain()); current->add(ins); current->push(ins); return true; } static MIRType GetElemKnownType(bool needsHoleCheck, TemporaryTypeSet* types) { MIRType knownType = types->getKnownMIRType(); // Null and undefined have no payload so they can't be specialized. // Since folding null/undefined while building SSA is not safe (see the // comment in IsPhiObservable), we just add an untyped load instruction // and rely on pushTypeBarrier and DCE to replace it with a null/undefined // constant. if (knownType == MIRType::Undefined || knownType == MIRType::Null) knownType = MIRType::Value; // Different architectures may want typed element reads which require // hole checks to be done as either value or typed reads. if (needsHoleCheck && !LIRGenerator::allowTypedElementHoleCheck()) knownType = MIRType::Value; return knownType; } bool IonBuilder::jsop_getelem() { startTrackingOptimizations(); MDefinition* index = current->pop(); MDefinition* obj = current->pop(); trackTypeInfo(TrackedTypeSite::Receiver, obj->type(), obj->resultTypeSet()); trackTypeInfo(TrackedTypeSite::Index, index->type(), index->resultTypeSet()); // Always use a call if we are performing analysis and not actually // emitting code, to simplify later analysis. if (info().isAnalysis() || shouldAbortOnPreliminaryGroups(obj)) { MInstruction* ins = MCallGetElement::New(alloc(), obj, index); current->add(ins); current->push(ins); if (!resumeAfter(ins)) return false; TemporaryTypeSet* types = bytecodeTypes(pc); return pushTypeBarrier(ins, types, BarrierKind::TypeSet); } obj = maybeUnboxForPropertyAccess(obj); bool emitted = false; if (!forceInlineCaches()) { trackOptimizationAttempt(TrackedStrategy::GetElem_TypedObject); if (!getElemTryTypedObject(&emitted, obj, index) || emitted) return emitted; // Note: no trackOptimizationAttempt call is needed, getElemTryGetProp // will call it. if (!getElemTryGetProp(&emitted, obj, index) || emitted) return emitted; trackOptimizationAttempt(TrackedStrategy::GetElem_Dense); if (!getElemTryDense(&emitted, obj, index) || emitted) return emitted; trackOptimizationAttempt(TrackedStrategy::GetElem_TypedStatic); if (!getElemTryTypedStatic(&emitted, obj, index) || emitted) return emitted; trackOptimizationAttempt(TrackedStrategy::GetElem_TypedArray); if (!getElemTryTypedArray(&emitted, obj, index) || emitted) return emitted; trackOptimizationAttempt(TrackedStrategy::GetElem_String); if (!getElemTryString(&emitted, obj, index) || emitted) return emitted; trackOptimizationAttempt(TrackedStrategy::GetElem_Arguments); if (!getElemTryArguments(&emitted, obj, index) || emitted) return emitted; trackOptimizationAttempt(TrackedStrategy::GetElem_ArgumentsInlined); if (!getElemTryArgumentsInlined(&emitted, obj, index) || emitted) return emitted; } if (script()->argumentsHasVarBinding() && obj->mightBeType(MIRType::MagicOptimizedArguments)) return abort("Type is not definitely lazy arguments."); trackOptimizationAttempt(TrackedStrategy::GetElem_InlineCache); if (!getElemTryCache(&emitted, obj, index) || emitted) return emitted; // Emit call. MInstruction* ins = MCallGetElement::New(alloc(), obj, index); current->add(ins); current->push(ins); if (!resumeAfter(ins)) return false; if (*pc == JSOP_CALLELEM && IsNullOrUndefined(obj->type())) { // Due to inlining, it's possible the observed TypeSet is non-empty, // even though we know |obj| is null/undefined and the MCallGetElement // will throw. Don't push a TypeBarrier in this case, to avoid // inlining the following (unreachable) JSOP_CALL. return true; } TemporaryTypeSet* types = bytecodeTypes(pc); return pushTypeBarrier(ins, types, BarrierKind::TypeSet); } bool IonBuilder::getElemTryTypedObject(bool* emitted, MDefinition* obj, MDefinition* index) { MOZ_ASSERT(*emitted == false); // The next several failures are all due to types not predicting that we // are definitely doing a getelem access on a typed object. trackOptimizationOutcome(TrackedOutcome::AccessNotTypedObject); TypedObjectPrediction objPrediction = typedObjectPrediction(obj); if (objPrediction.isUseless()) return true; if (!objPrediction.ofArrayKind()) return true; TypedObjectPrediction elemPrediction = objPrediction.arrayElementType(); if (elemPrediction.isUseless()) return true; uint32_t elemSize; if (!elemPrediction.hasKnownSize(&elemSize)) return true; switch (elemPrediction.kind()) { case type::Simd: // FIXME (bug 894105): load into a MIRType::float32x4 etc trackOptimizationOutcome(TrackedOutcome::GenericFailure); return true; case type::Struct: case type::Array: return getElemTryComplexElemOfTypedObject(emitted, obj, index, objPrediction, elemPrediction, elemSize); case type::Scalar: return getElemTryScalarElemOfTypedObject(emitted, obj, index, objPrediction, elemPrediction, elemSize); case type::Reference: return getElemTryReferenceElemOfTypedObject(emitted, obj, index, objPrediction, elemPrediction); } MOZ_CRASH("Bad kind"); } bool IonBuilder::checkTypedObjectIndexInBounds(uint32_t elemSize, MDefinition* obj, MDefinition* index, TypedObjectPrediction objPrediction, LinearSum* indexAsByteOffset) { // Ensure index is an integer. MInstruction* idInt32 = MToInt32::New(alloc(), index); current->add(idInt32); // If we know the length statically from the type, just embed it. // Otherwise, load it from the appropriate reserved slot on the // typed object. We know it's an int32, so we can convert from // Value to int32 using truncation. int32_t lenOfAll; MDefinition* length; if (objPrediction.hasKnownArrayLength(&lenOfAll)) { length = constantInt(lenOfAll); // If we are not loading the length from the object itself, only // optimize if the array buffer can never be a detached array buffer. TypeSet::ObjectKey* globalKey = TypeSet::ObjectKey::get(&script()->global()); if (globalKey->hasFlags(constraints(), OBJECT_FLAG_TYPED_OBJECT_HAS_DETACHED_BUFFER)) { trackOptimizationOutcome(TrackedOutcome::TypedObjectHasDetachedBuffer); return false; } } else { trackOptimizationOutcome(TrackedOutcome::TypedObjectArrayRange); return false; } index = addBoundsCheck(idInt32, length); return indexAsByteOffset->add(index, AssertedCast(elemSize)); } bool IonBuilder::getElemTryScalarElemOfTypedObject(bool* emitted, MDefinition* obj, MDefinition* index, TypedObjectPrediction objPrediction, TypedObjectPrediction elemPrediction, uint32_t elemSize) { MOZ_ASSERT(objPrediction.ofArrayKind()); // Must always be loading the same scalar type ScalarTypeDescr::Type elemType = elemPrediction.scalarType(); MOZ_ASSERT(elemSize == ScalarTypeDescr::alignment(elemType)); LinearSum indexAsByteOffset(alloc()); if (!checkTypedObjectIndexInBounds(elemSize, obj, index, objPrediction, &indexAsByteOffset)) return true; trackOptimizationSuccess(); *emitted = true; return pushScalarLoadFromTypedObject(obj, indexAsByteOffset, elemType); } bool IonBuilder::getElemTryReferenceElemOfTypedObject(bool* emitted, MDefinition* obj, MDefinition* index, TypedObjectPrediction objPrediction, TypedObjectPrediction elemPrediction) { MOZ_ASSERT(objPrediction.ofArrayKind()); ReferenceTypeDescr::Type elemType = elemPrediction.referenceType(); uint32_t elemSize = ReferenceTypeDescr::size(elemType); LinearSum indexAsByteOffset(alloc()); if (!checkTypedObjectIndexInBounds(elemSize, obj, index, objPrediction, &indexAsByteOffset)) return true; trackOptimizationSuccess(); *emitted = true; return pushReferenceLoadFromTypedObject(obj, indexAsByteOffset, elemType, nullptr); } bool IonBuilder::pushScalarLoadFromTypedObject(MDefinition* obj, const LinearSum& byteOffset, ScalarTypeDescr::Type elemType) { uint32_t size = ScalarTypeDescr::size(elemType); MOZ_ASSERT(size == ScalarTypeDescr::alignment(elemType)); // Find location within the owner object. MDefinition* elements; MDefinition* scaledOffset; int32_t adjustment; loadTypedObjectElements(obj, byteOffset, size, &elements, &scaledOffset, &adjustment); // Load the element. MLoadUnboxedScalar* load = MLoadUnboxedScalar::New(alloc(), elements, scaledOffset, elemType, DoesNotRequireMemoryBarrier, adjustment); current->add(load); current->push(load); // If we are reading in-bounds elements, we can use knowledge about // the array type to determine the result type, even if the opcode has // never executed. The known pushed type is only used to distinguish // uint32 reads that may produce either doubles or integers. TemporaryTypeSet* resultTypes = bytecodeTypes(pc); bool allowDouble = resultTypes->hasType(TypeSet::DoubleType()); // Note: knownType is not necessarily in resultTypes; e.g. if we // have only observed integers coming out of float array. MIRType knownType = MIRTypeForTypedArrayRead(elemType, allowDouble); // Note: we can ignore the type barrier here, we know the type must // be valid and unbarriered. Also, need not set resultTypeSet, // because knownType is scalar and a resultTypeSet would provide // no useful additional info. load->setResultType(knownType); return true; } bool IonBuilder::pushReferenceLoadFromTypedObject(MDefinition* typedObj, const LinearSum& byteOffset, ReferenceTypeDescr::Type type, PropertyName* name) { // Find location within the owner object. MDefinition* elements; MDefinition* scaledOffset; int32_t adjustment; uint32_t alignment = ReferenceTypeDescr::alignment(type); loadTypedObjectElements(typedObj, byteOffset, alignment, &elements, &scaledOffset, &adjustment); TemporaryTypeSet* observedTypes = bytecodeTypes(pc); MInstruction* load = nullptr; // initialize to silence GCC warning BarrierKind barrier = PropertyReadNeedsTypeBarrier(analysisContext, constraints(), typedObj, name, observedTypes); switch (type) { case ReferenceTypeDescr::TYPE_ANY: { // Make sure the barrier reflects the possibility of reading undefined. bool bailOnUndefined = barrier == BarrierKind::NoBarrier && !observedTypes->hasType(TypeSet::UndefinedType()); if (bailOnUndefined) barrier = BarrierKind::TypeTagOnly; load = MLoadElement::New(alloc(), elements, scaledOffset, false, false, adjustment); break; } case ReferenceTypeDescr::TYPE_OBJECT: { // Make sure the barrier reflects the possibility of reading null. When // there is no other barrier needed we include the null bailout with // MLoadUnboxedObjectOrNull, which avoids the need to box the result // for a type barrier instruction. MLoadUnboxedObjectOrNull::NullBehavior nullBehavior; if (barrier == BarrierKind::NoBarrier && !observedTypes->hasType(TypeSet::NullType())) nullBehavior = MLoadUnboxedObjectOrNull::BailOnNull; else nullBehavior = MLoadUnboxedObjectOrNull::HandleNull; load = MLoadUnboxedObjectOrNull::New(alloc(), elements, scaledOffset, nullBehavior, adjustment); break; } case ReferenceTypeDescr::TYPE_STRING: { load = MLoadUnboxedString::New(alloc(), elements, scaledOffset, adjustment); observedTypes->addType(TypeSet::StringType(), alloc().lifoAlloc()); break; } } current->add(load); current->push(load); return pushTypeBarrier(load, observedTypes, barrier); } bool IonBuilder::getElemTryComplexElemOfTypedObject(bool* emitted, MDefinition* obj, MDefinition* index, TypedObjectPrediction objPrediction, TypedObjectPrediction elemPrediction, uint32_t elemSize) { MOZ_ASSERT(objPrediction.ofArrayKind()); MDefinition* type = loadTypedObjectType(obj); MDefinition* elemTypeObj = typeObjectForElementFromArrayStructType(type); LinearSum indexAsByteOffset(alloc()); if (!checkTypedObjectIndexInBounds(elemSize, obj, index, objPrediction, &indexAsByteOffset)) return true; return pushDerivedTypedObject(emitted, obj, indexAsByteOffset, elemPrediction, elemTypeObj); } bool IonBuilder::pushDerivedTypedObject(bool* emitted, MDefinition* obj, const LinearSum& baseByteOffset, TypedObjectPrediction derivedPrediction, MDefinition* derivedTypeObj) { // Find location within the owner object. MDefinition* owner; LinearSum ownerByteOffset(alloc()); loadTypedObjectData(obj, &owner, &ownerByteOffset); if (!ownerByteOffset.add(baseByteOffset, 1)) setForceAbort(); MDefinition* offset = ConvertLinearSum(alloc(), current, ownerByteOffset, /* convertConstant = */ true); // Create the derived typed object. MInstruction* derivedTypedObj = MNewDerivedTypedObject::New(alloc(), derivedPrediction, derivedTypeObj, owner, offset); current->add(derivedTypedObj); current->push(derivedTypedObj); // Determine (if possible) the class/proto that `derivedTypedObj` will // have. For derived typed objects, the opacity will be the same as the // incoming object from which the derived typed object is, well, derived. // The prototype will be determined based on the type descriptor (and is // immutable). TemporaryTypeSet* objTypes = obj->resultTypeSet(); const Class* expectedClass = nullptr; if (const Class* objClass = objTypes ? objTypes->getKnownClass(constraints()) : nullptr) { MOZ_ASSERT(IsTypedObjectClass(objClass)); expectedClass = GetOutlineTypedObjectClass(IsOpaqueTypedObjectClass(objClass)); } const TypedProto* expectedProto = derivedPrediction.getKnownPrototype(); MOZ_ASSERT_IF(expectedClass, IsTypedObjectClass(expectedClass)); // Determine (if possible) the class/proto that the observed type set // describes. TemporaryTypeSet* observedTypes = bytecodeTypes(pc); const Class* observedClass = observedTypes->getKnownClass(constraints()); // If expectedClass/expectedProto are both non-null (and hence known), we // can predict precisely what object group derivedTypedObj will have. // Therefore, if we observe that this group is already contained in the set // of observedTypes, we can skip the barrier. // // Barriers still wind up being needed in some relatively // rare cases: // // - if multiple kinds of typed objects flow into this point, // in which case we will not be able to predict expectedClass // nor expectedProto. // // - if the code has never executed, in which case the set of // observed types will be incomplete. // // Barriers are particularly expensive here because they prevent // us from optimizing the MNewDerivedTypedObject away. JSObject* observedProto; if (observedTypes->getCommonPrototype(constraints(), &observedProto) && observedClass && observedProto && observedClass == expectedClass && observedProto == expectedProto) { derivedTypedObj->setResultTypeSet(observedTypes); } else { if (!pushTypeBarrier(derivedTypedObj, observedTypes, BarrierKind::TypeSet)) return false; } trackOptimizationSuccess(); *emitted = true; return true; } bool IonBuilder::getElemTryGetProp(bool* emitted, MDefinition* obj, MDefinition* index) { // If index is a constant string or symbol, try to optimize this GETELEM // as a GETPROP. MOZ_ASSERT(*emitted == false); MConstant* indexConst = index->maybeConstantValue(); jsid id; if (!indexConst || !ValueToIdPure(indexConst->toJSValue(), &id)) return true; if (id != IdToTypeId(id)) return true; TemporaryTypeSet* types = bytecodeTypes(pc); trackOptimizationAttempt(TrackedStrategy::GetProp_Constant); if (!getPropTryConstant(emitted, obj, id, types) || *emitted) { if (*emitted) index->setImplicitlyUsedUnchecked(); return *emitted; } trackOptimizationAttempt(TrackedStrategy::GetProp_NotDefined); if (!getPropTryNotDefined(emitted, obj, id, types) || *emitted) { if (*emitted) index->setImplicitlyUsedUnchecked(); return *emitted; } return true; } bool IonBuilder::getElemTryDense(bool* emitted, MDefinition* obj, MDefinition* index) { MOZ_ASSERT(*emitted == false); if (!ElementAccessIsDenseNative(constraints(), obj, index)) { trackOptimizationOutcome(TrackedOutcome::AccessNotDense); return true; } // Don't generate a fast path if there have been bounds check failures // and this access might be on a sparse property. if (ElementAccessHasExtraIndexedProperty(this, obj) && failedBoundsCheck_) { trackOptimizationOutcome(TrackedOutcome::ProtoIndexedProps); return true; } // Don't generate a fast path if this pc has seen negative indexes accessed, // which will not appear to be extra indexed properties. if (inspector->hasSeenNegativeIndexGetElement(pc)) { trackOptimizationOutcome(TrackedOutcome::ArraySeenNegativeIndex); return true; } if (!jsop_getelem_dense(obj, index)) return false; trackOptimizationSuccess(); *emitted = true; return true; } JSObject* IonBuilder::getStaticTypedArrayObject(MDefinition* obj, MDefinition* index) { Scalar::Type arrayType; if (!ElementAccessIsTypedArray(constraints(), obj, index, &arrayType)) { trackOptimizationOutcome(TrackedOutcome::AccessNotTypedArray); return nullptr; } if (!LIRGenerator::allowStaticTypedArrayAccesses()) { trackOptimizationOutcome(TrackedOutcome::Disabled); return nullptr; } if (ElementAccessHasExtraIndexedProperty(this, obj)) { trackOptimizationOutcome(TrackedOutcome::ProtoIndexedProps); return nullptr; } if (!obj->resultTypeSet()) { trackOptimizationOutcome(TrackedOutcome::NoTypeInfo); return nullptr; } JSObject* tarrObj = obj->resultTypeSet()->maybeSingleton(); if (!tarrObj) { trackOptimizationOutcome(TrackedOutcome::NotSingleton); return nullptr; } TypeSet::ObjectKey* tarrKey = TypeSet::ObjectKey::get(tarrObj); if (tarrKey->unknownProperties()) { trackOptimizationOutcome(TrackedOutcome::UnknownProperties); return nullptr; } return tarrObj; } bool IonBuilder::getElemTryTypedStatic(bool* emitted, MDefinition* obj, MDefinition* index) { MOZ_ASSERT(*emitted == false); JSObject* tarrObj = getStaticTypedArrayObject(obj, index); if (!tarrObj) return true; // LoadTypedArrayElementStatic currently treats uint32 arrays as int32. Scalar::Type viewType = tarrObj->as().type(); if (viewType == Scalar::Uint32) { trackOptimizationOutcome(TrackedOutcome::StaticTypedArrayUint32); return true; } MDefinition* ptr = convertShiftToMaskForStaticTypedArray(index, viewType); if (!ptr) return true; // Emit LoadTypedArrayElementStatic. if (tarrObj->is()) { TypeSet::ObjectKey* tarrKey = TypeSet::ObjectKey::get(tarrObj); tarrKey->watchStateChangeForTypedArrayData(constraints()); } obj->setImplicitlyUsedUnchecked(); index->setImplicitlyUsedUnchecked(); MLoadTypedArrayElementStatic* load = MLoadTypedArrayElementStatic::New(alloc(), tarrObj, ptr); current->add(load); current->push(load); // The load is infallible if an undefined result will be coerced to the // appropriate numeric type if the read is out of bounds. The truncation // analysis picks up some of these cases, but is incomplete with respect // to others. For now, sniff the bytecode for simple patterns following // the load which guarantee a truncation or numeric conversion. if (viewType == Scalar::Float32 || viewType == Scalar::Float64) { jsbytecode* next = pc + JSOP_GETELEM_LENGTH; if (*next == JSOP_POS) load->setInfallible(); } else { jsbytecode* next = pc + JSOP_GETELEM_LENGTH; if (*next == JSOP_ZERO && *(next + JSOP_ZERO_LENGTH) == JSOP_BITOR) load->setInfallible(); } trackOptimizationSuccess(); *emitted = true; return true; } bool IonBuilder::getElemTryTypedArray(bool* emitted, MDefinition* obj, MDefinition* index) { MOZ_ASSERT(*emitted == false); Scalar::Type arrayType; if (!ElementAccessIsTypedArray(constraints(), obj, index, &arrayType)) { trackOptimizationOutcome(TrackedOutcome::AccessNotTypedArray); return true; } // Emit typed getelem variant. if (!jsop_getelem_typed(obj, index, arrayType)) return false; trackOptimizationSuccess(); *emitted = true; return true; } bool IonBuilder::getElemTryString(bool* emitted, MDefinition* obj, MDefinition* index) { MOZ_ASSERT(*emitted == false); if (obj->type() != MIRType::String || !IsNumberType(index->type())) { trackOptimizationOutcome(TrackedOutcome::AccessNotString); return true; } // If the index is expected to be out-of-bounds, don't optimize to avoid // frequent bailouts. if (bytecodeTypes(pc)->hasType(TypeSet::UndefinedType())) { trackOptimizationOutcome(TrackedOutcome::OutOfBounds); return true; } // Emit fast path for string[index]. MInstruction* idInt32 = MToInt32::New(alloc(), index); current->add(idInt32); index = idInt32; MStringLength* length = MStringLength::New(alloc(), obj); current->add(length); index = addBoundsCheck(index, length); MCharCodeAt* charCode = MCharCodeAt::New(alloc(), obj, index); current->add(charCode); MFromCharCode* result = MFromCharCode::New(alloc(), charCode); current->add(result); current->push(result); trackOptimizationSuccess(); *emitted = true; return true; } bool IonBuilder::getElemTryArguments(bool* emitted, MDefinition* obj, MDefinition* index) { MOZ_ASSERT(*emitted == false); if (inliningDepth_ > 0) return true; if (obj->type() != MIRType::MagicOptimizedArguments) return true; // Emit GetFrameArgument. MOZ_ASSERT(!info().argsObjAliasesFormals()); // Type Inference has guaranteed this is an optimized arguments object. obj->setImplicitlyUsedUnchecked(); // To ensure that we are not looking above the number of actual arguments. MArgumentsLength* length = MArgumentsLength::New(alloc()); current->add(length); // Ensure index is an integer. MInstruction* idInt32 = MToInt32::New(alloc(), index); current->add(idInt32); index = idInt32; // Bailouts if we read more than the number of actual arguments. index = addBoundsCheck(index, length); // Load the argument from the actual arguments. MGetFrameArgument* load = MGetFrameArgument::New(alloc(), index, analysis_.hasSetArg()); current->add(load); current->push(load); TemporaryTypeSet* types = bytecodeTypes(pc); if (!pushTypeBarrier(load, types, BarrierKind::TypeSet)) return false; trackOptimizationSuccess(); *emitted = true; return true; } bool IonBuilder::getElemTryArgumentsInlined(bool* emitted, MDefinition* obj, MDefinition* index) { MOZ_ASSERT(*emitted == false); if (inliningDepth_ == 0) return true; if (obj->type() != MIRType::MagicOptimizedArguments) return true; // Emit inlined arguments. obj->setImplicitlyUsedUnchecked(); MOZ_ASSERT(!info().argsObjAliasesFormals()); // When the id is constant, we can just return the corresponding inlined argument MConstant* indexConst = index->maybeConstantValue(); if (indexConst && indexConst->type() == MIRType::Int32) { MOZ_ASSERT(inliningDepth_ > 0); int32_t id = indexConst->toInt32(); index->setImplicitlyUsedUnchecked(); if (id < (int32_t)inlineCallInfo_->argc() && id >= 0) current->push(inlineCallInfo_->getArg(id)); else pushConstant(UndefinedValue()); trackOptimizationSuccess(); *emitted = true; return true; } // inlined not constant not supported, yet. return abort("NYI inlined not constant get argument element"); } bool IonBuilder::getElemTryCache(bool* emitted, MDefinition* obj, MDefinition* index) { MOZ_ASSERT(*emitted == false); // Make sure we have at least an object. if (!obj->mightBeType(MIRType::Object)) { trackOptimizationOutcome(TrackedOutcome::NotObject); return true; } // Don't cache for strings. if (obj->mightBeType(MIRType::String)) { trackOptimizationOutcome(TrackedOutcome::GetElemStringNotCached); return true; } // Index should be integer, string, or symbol if (!index->mightBeType(MIRType::Int32) && !index->mightBeType(MIRType::String) && !index->mightBeType(MIRType::Symbol)) { trackOptimizationOutcome(TrackedOutcome::IndexType); return true; } // Turn off cacheing if the element is int32 and we've seen non-native objects as the target // of this getelem. bool nonNativeGetElement = inspector->hasSeenNonNativeGetElement(pc); if (index->mightBeType(MIRType::Int32) && nonNativeGetElement) { trackOptimizationOutcome(TrackedOutcome::NonNativeReceiver); return true; } // Emit GetElementCache. TemporaryTypeSet* types = bytecodeTypes(pc); BarrierKind barrier = PropertyReadNeedsTypeBarrier(analysisContext, constraints(), obj, nullptr, types); // Always add a barrier if the index might be a string or symbol, so that // the cache can attach stubs for particular properties. if (index->mightBeType(MIRType::String) || index->mightBeType(MIRType::Symbol)) barrier = BarrierKind::TypeSet; MGetPropertyCache* ins = MGetPropertyCache::New(alloc(), obj, index, barrier == BarrierKind::TypeSet); current->add(ins); current->push(ins); if (!resumeAfter(ins)) return false; // Spice up type information. if (index->type() == MIRType::Int32 && barrier == BarrierKind::NoBarrier) { bool needHoleCheck = !ElementAccessIsPacked(constraints(), obj); MIRType knownType = GetElemKnownType(needHoleCheck, types); if (knownType != MIRType::Value && knownType != MIRType::Double) ins->setResultType(knownType); } if (!pushTypeBarrier(ins, types, barrier)) return false; trackOptimizationSuccess(); *emitted = true; return true; } TemporaryTypeSet* IonBuilder::computeHeapType(const TemporaryTypeSet* objTypes, const jsid id) { if (objTypes->unknownObject() || objTypes->getObjectCount() == 0) return nullptr; TemporaryTypeSet empty; TemporaryTypeSet* acc = ∅ LifoAlloc* lifoAlloc = alloc().lifoAlloc(); Vector properties; if (!properties.reserve(objTypes->getObjectCount())) return nullptr; for (unsigned i = 0; i < objTypes->getObjectCount(); i++) { TypeSet::ObjectKey* key = objTypes->getObject(i); if (key->unknownProperties()) return nullptr; HeapTypeSetKey property = key->property(id); HeapTypeSet* currentSet = property.maybeTypes(); if (!currentSet || currentSet->unknown()) return nullptr; properties.infallibleAppend(property); acc = TypeSet::unionSets(acc, currentSet, lifoAlloc); if (!acc) return nullptr; } // Freeze all the properties associated with the refined type set. for (HeapTypeSetKey* i = properties.begin(); i != properties.end(); i++) i->freeze(constraints()); return acc; } bool IonBuilder::jsop_getelem_dense(MDefinition* obj, MDefinition* index) { TemporaryTypeSet* types = bytecodeTypes(pc); MOZ_ASSERT(index->type() == MIRType::Int32 || index->type() == MIRType::Double); if (JSOp(*pc) == JSOP_CALLELEM) { // Indexed call on an element of an array. Populate the observed types // with any objects that could be in the array, to avoid extraneous // type barriers. AddObjectsForPropertyRead(obj, nullptr, types); } BarrierKind barrier = PropertyReadNeedsTypeBarrier(analysisContext, constraints(), obj, nullptr, types); bool needsHoleCheck = !ElementAccessIsPacked(constraints(), obj); // Reads which are on holes in the object do not have to bail out if // undefined values have been observed at this access site and the access // cannot hit another indexed property on the object or its prototypes. bool readOutOfBounds = types->hasType(TypeSet::UndefinedType()) && !ElementAccessHasExtraIndexedProperty(this, obj); MIRType knownType = MIRType::Value; if (barrier == BarrierKind::NoBarrier) knownType = GetElemKnownType(needsHoleCheck, types); // Ensure index is an integer. MInstruction* idInt32 = MToInt32::New(alloc(), index); current->add(idInt32); index = idInt32; // Get the elements vector. MInstruction* elements = MElements::New(alloc(), obj); current->add(elements); // Note: to help GVN, use the original MElements instruction and not // MConvertElementsToDoubles as operand. This is fine because converting // elements to double does not change the initialized length. MInstruction* initLength = initializedLength(obj, elements); // If we can load the element as a definite double, make sure to check that // the array has been converted to homogenous doubles first. TemporaryTypeSet* objTypes = obj->resultTypeSet(); bool inBounds = !readOutOfBounds && !needsHoleCheck; if (inBounds) { TemporaryTypeSet* heapTypes = computeHeapType(objTypes, JSID_VOID); if (heapTypes && heapTypes->isSubset(types)) { knownType = heapTypes->getKnownMIRType(); types = heapTypes; } } bool loadDouble = barrier == BarrierKind::NoBarrier && loopDepth_ && inBounds && knownType == MIRType::Double && objTypes && objTypes->convertDoubleElements(constraints()) == TemporaryTypeSet::AlwaysConvertToDoubles; if (loadDouble) elements = addConvertElementsToDoubles(elements); MInstruction* load; if (!readOutOfBounds) { // This load should not return undefined, so likely we're reading // in-bounds elements, and the array is packed or its holes are not // read. This is the best case: we can separate the bounds check for // hoisting. index = addBoundsCheck(index, initLength); load = MLoadElement::New(alloc(), elements, index, needsHoleCheck, loadDouble); current->add(load); } else { // This load may return undefined, so assume that we *can* read holes, // or that we can read out-of-bounds accesses. In this case, the bounds // check is part of the opcode. load = MLoadElementHole::New(alloc(), elements, index, initLength, needsHoleCheck); current->add(load); // If maybeUndefined was true, the typeset must have undefined, and // then either additional types or a barrier. This means we should // never have a typed version of LoadElementHole. MOZ_ASSERT(knownType == MIRType::Value); } if (knownType != MIRType::Value) { load->setResultType(knownType); load->setResultTypeSet(types); } current->push(load); return pushTypeBarrier(load, types, barrier); } MInstruction* IonBuilder::addArrayBufferByteLength(MDefinition* obj) { MLoadFixedSlot* ins = MLoadFixedSlot::New(alloc(), obj, size_t(ArrayBufferObject::BYTE_LENGTH_SLOT)); current->add(ins); ins->setResultType(MIRType::Int32); return ins; } void IonBuilder::addTypedArrayLengthAndData(MDefinition* obj, BoundsChecking checking, MDefinition** index, MInstruction** length, MInstruction** elements) { MOZ_ASSERT((index != nullptr) == (elements != nullptr)); JSObject* tarr = nullptr; if (MConstant* objConst = obj->maybeConstantValue()) { if (objConst->type() == MIRType::Object) tarr = &objConst->toObject(); } else if (TemporaryTypeSet* types = obj->resultTypeSet()) { tarr = types->maybeSingleton(); } if (tarr) { SharedMem data = tarr->as().viewDataEither(); // Bug 979449 - Optimistically embed the elements and use TI to // invalidate if we move them. bool isTenured = !tarr->runtimeFromMainThread()->gc.nursery.isInside(data); if (isTenured && tarr->isSingleton()) { // The 'data' pointer of TypedArrayObject can change in rare circumstances // (ArrayBufferObject::changeContents). TypeSet::ObjectKey* tarrKey = TypeSet::ObjectKey::get(tarr); if (!tarrKey->unknownProperties()) { if (tarr->is()) tarrKey->watchStateChangeForTypedArrayData(constraints()); obj->setImplicitlyUsedUnchecked(); int32_t len = AssertedCast(tarr->as().length()); *length = MConstant::New(alloc(), Int32Value(len)); current->add(*length); if (index) { if (checking == DoBoundsCheck) *index = addBoundsCheck(*index, *length); *elements = MConstantElements::New(alloc(), data); current->add(*elements); } return; } } } *length = MTypedArrayLength::New(alloc(), obj); current->add(*length); if (index) { if (checking == DoBoundsCheck) *index = addBoundsCheck(*index, *length); *elements = MTypedArrayElements::New(alloc(), obj); current->add(*elements); } } MDefinition* IonBuilder::convertShiftToMaskForStaticTypedArray(MDefinition* id, Scalar::Type viewType) { trackOptimizationOutcome(TrackedOutcome::StaticTypedArrayCantComputeMask); // No shifting is necessary if the typed array has single byte elements. if (TypedArrayShift(viewType) == 0) return id; // If the index is an already shifted constant, undo the shift to get the // absolute offset being accessed. if (MConstant* idConst = id->maybeConstantValue()) { if (idConst->type() == MIRType::Int32) { int32_t index = idConst->toInt32(); MConstant* offset = MConstant::New(alloc(), Int32Value(index << TypedArrayShift(viewType))); current->add(offset); return offset; } } if (!id->isRsh() || id->isEffectful()) return nullptr; MConstant* shiftAmount = id->toRsh()->rhs()->maybeConstantValue(); if (!shiftAmount || shiftAmount->type() != MIRType::Int32) return nullptr; if (uint32_t(shiftAmount->toInt32()) != TypedArrayShift(viewType)) return nullptr; // Instead of shifting, mask off the low bits of the index so that // a non-scaled access on the typed array can be performed. MConstant* mask = MConstant::New(alloc(), Int32Value(~((1 << shiftAmount->toInt32()) - 1))); MBitAnd* ptr = MBitAnd::New(alloc(), id->getOperand(0), mask); ptr->infer(nullptr, nullptr); MOZ_ASSERT(!ptr->isEffectful()); current->add(mask); current->add(ptr); return ptr; } bool IonBuilder::jsop_getelem_typed(MDefinition* obj, MDefinition* index, Scalar::Type arrayType) { TemporaryTypeSet* types = bytecodeTypes(pc); bool maybeUndefined = types->hasType(TypeSet::UndefinedType()); // Reading from an Uint32Array will result in a double for values // that don't fit in an int32. We have to bailout if this happens // and the instruction is not known to return a double. bool allowDouble = types->hasType(TypeSet::DoubleType()); // Ensure id is an integer. MInstruction* idInt32 = MToInt32::New(alloc(), index); current->add(idInt32); index = idInt32; if (!maybeUndefined) { // Assume the index is in range, so that we can hoist the length, // elements vector and bounds check. // If we are reading in-bounds elements, we can use knowledge about // the array type to determine the result type, even if the opcode has // never executed. The known pushed type is only used to distinguish // uint32 reads that may produce either doubles or integers. MIRType knownType = MIRTypeForTypedArrayRead(arrayType, allowDouble); // Get length, bounds-check, then get elements, and add all instructions. MInstruction* length; MInstruction* elements; addTypedArrayLengthAndData(obj, DoBoundsCheck, &index, &length, &elements); // Load the element. MLoadUnboxedScalar* load = MLoadUnboxedScalar::New(alloc(), elements, index, arrayType); current->add(load); current->push(load); // Note: we can ignore the type barrier here, we know the type must // be valid and unbarriered. load->setResultType(knownType); return true; } else { // We need a type barrier if the array's element type has never been // observed (we've only read out-of-bounds values). Note that for // Uint32Array, we only check for int32: if allowDouble is false we // will bailout when we read a double. BarrierKind barrier = BarrierKind::TypeSet; switch (arrayType) { case Scalar::Int8: case Scalar::Uint8: case Scalar::Uint8Clamped: case Scalar::Int16: case Scalar::Uint16: case Scalar::Int32: case Scalar::Uint32: if (types->hasType(TypeSet::Int32Type())) barrier = BarrierKind::NoBarrier; break; case Scalar::Float32: case Scalar::Float64: if (allowDouble) barrier = BarrierKind::NoBarrier; break; default: MOZ_CRASH("Unknown typed array type"); } // Assume we will read out-of-bound values. In this case the // bounds check will be part of the instruction, and the instruction // will always return a Value. MLoadTypedArrayElementHole* load = MLoadTypedArrayElementHole::New(alloc(), obj, index, arrayType, allowDouble); current->add(load); current->push(load); return pushTypeBarrier(load, types, barrier); } } bool IonBuilder::jsop_setelem() { bool emitted = false; startTrackingOptimizations(); MDefinition* value = current->pop(); MDefinition* index = current->pop(); MDefinition* object = current->pop(); trackTypeInfo(TrackedTypeSite::Receiver, object->type(), object->resultTypeSet()); trackTypeInfo(TrackedTypeSite::Index, index->type(), index->resultTypeSet()); trackTypeInfo(TrackedTypeSite::Value, value->type(), value->resultTypeSet()); if (shouldAbortOnPreliminaryGroups(object)) { MInstruction* ins = MCallSetElement::New(alloc(), object, index, value, IsStrictSetPC(pc)); current->add(ins); current->push(value); return resumeAfter(ins); } if (!forceInlineCaches()) { trackOptimizationAttempt(TrackedStrategy::SetElem_TypedObject); if (!setElemTryTypedObject(&emitted, object, index, value) || emitted) return emitted; trackOptimizationAttempt(TrackedStrategy::SetElem_TypedStatic); if (!setElemTryTypedStatic(&emitted, object, index, value) || emitted) return emitted; trackOptimizationAttempt(TrackedStrategy::SetElem_TypedArray); if (!setElemTryTypedArray(&emitted, object, index, value) || emitted) return emitted; trackOptimizationAttempt(TrackedStrategy::SetElem_Dense); SetElemICInspector icInspect(inspector->setElemICInspector(pc)); bool writeHole = icInspect.sawOOBDenseWrite(); if (!setElemTryDense(&emitted, object, index, value, writeHole) || emitted) return emitted; trackOptimizationAttempt(TrackedStrategy::SetElem_Arguments); if (!setElemTryArguments(&emitted, object, index, value) || emitted) return emitted; } if (script()->argumentsHasVarBinding() && object->mightBeType(MIRType::MagicOptimizedArguments) && info().analysisMode() != Analysis_ArgumentsUsage) { return abort("Type is not definitely lazy arguments."); } trackOptimizationAttempt(TrackedStrategy::SetElem_InlineCache); if (!setElemTryCache(&emitted, object, index, value) || emitted) return emitted; // Emit call. MInstruction* ins = MCallSetElement::New(alloc(), object, index, value, IsStrictSetPC(pc)); current->add(ins); current->push(value); return resumeAfter(ins); } bool IonBuilder::setElemTryTypedObject(bool* emitted, MDefinition* obj, MDefinition* index, MDefinition* value) { MOZ_ASSERT(*emitted == false); // The next several failures are all due to types not predicting that we // are definitely doing a getelem access on a typed object. trackOptimizationOutcome(TrackedOutcome::AccessNotTypedObject); TypedObjectPrediction objPrediction = typedObjectPrediction(obj); if (objPrediction.isUseless()) return true; if (!objPrediction.ofArrayKind()) return true; TypedObjectPrediction elemPrediction = objPrediction.arrayElementType(); if (elemPrediction.isUseless()) return true; uint32_t elemSize; if (!elemPrediction.hasKnownSize(&elemSize)) return true; switch (elemPrediction.kind()) { case type::Simd: // FIXME (bug 894105): store a MIRType::float32x4 etc trackOptimizationOutcome(TrackedOutcome::GenericFailure); return true; case type::Reference: return setElemTryReferenceElemOfTypedObject(emitted, obj, index, objPrediction, value, elemPrediction); case type::Scalar: return setElemTryScalarElemOfTypedObject(emitted, obj, index, objPrediction, value, elemPrediction, elemSize); case type::Struct: case type::Array: // Not yet optimized. trackOptimizationOutcome(TrackedOutcome::GenericFailure); return true; } MOZ_CRASH("Bad kind"); } bool IonBuilder::setElemTryReferenceElemOfTypedObject(bool* emitted, MDefinition* obj, MDefinition* index, TypedObjectPrediction objPrediction, MDefinition* value, TypedObjectPrediction elemPrediction) { ReferenceTypeDescr::Type elemType = elemPrediction.referenceType(); uint32_t elemSize = ReferenceTypeDescr::size(elemType); LinearSum indexAsByteOffset(alloc()); if (!checkTypedObjectIndexInBounds(elemSize, obj, index, objPrediction, &indexAsByteOffset)) return true; if (!storeReferenceTypedObjectValue(obj, indexAsByteOffset, elemType, value, nullptr)) return true; current->push(value); trackOptimizationSuccess(); *emitted = true; return true; } bool IonBuilder::setElemTryScalarElemOfTypedObject(bool* emitted, MDefinition* obj, MDefinition* index, TypedObjectPrediction objPrediction, MDefinition* value, TypedObjectPrediction elemPrediction, uint32_t elemSize) { // Must always be loading the same scalar type ScalarTypeDescr::Type elemType = elemPrediction.scalarType(); MOZ_ASSERT(elemSize == ScalarTypeDescr::alignment(elemType)); LinearSum indexAsByteOffset(alloc()); if (!checkTypedObjectIndexInBounds(elemSize, obj, index, objPrediction, &indexAsByteOffset)) return true; // Store the element if (!storeScalarTypedObjectValue(obj, indexAsByteOffset, elemType, value)) return false; current->push(value); trackOptimizationSuccess(); *emitted = true; return true; } bool IonBuilder::setElemTryTypedStatic(bool* emitted, MDefinition* object, MDefinition* index, MDefinition* value) { MOZ_ASSERT(*emitted == false); JSObject* tarrObj = getStaticTypedArrayObject(object, index); if (!tarrObj) return true; SharedMem viewData = tarrObj->as().viewDataEither(); if (tarrObj->runtimeFromMainThread()->gc.nursery.isInside(viewData)) return true; Scalar::Type viewType = tarrObj->as().type(); MDefinition* ptr = convertShiftToMaskForStaticTypedArray(index, viewType); if (!ptr) return true; // Emit StoreTypedArrayElementStatic. if (tarrObj->is()) { TypeSet::ObjectKey* tarrKey = TypeSet::ObjectKey::get(tarrObj); tarrKey->watchStateChangeForTypedArrayData(constraints()); } object->setImplicitlyUsedUnchecked(); index->setImplicitlyUsedUnchecked(); // Clamp value to [0, 255] for Uint8ClampedArray. MDefinition* toWrite = value; if (viewType == Scalar::Uint8Clamped) { toWrite = MClampToUint8::New(alloc(), value); current->add(toWrite->toInstruction()); } MInstruction* store = MStoreTypedArrayElementStatic::New(alloc(), tarrObj, ptr, toWrite); current->add(store); current->push(value); if (!resumeAfter(store)) return false; trackOptimizationSuccess(); *emitted = true; return true; } bool IonBuilder::setElemTryTypedArray(bool* emitted, MDefinition* object, MDefinition* index, MDefinition* value) { MOZ_ASSERT(*emitted == false); Scalar::Type arrayType; if (!ElementAccessIsTypedArray(constraints(), object, index, &arrayType)) { trackOptimizationOutcome(TrackedOutcome::AccessNotTypedArray); return true; } // Emit typed setelem variant. if (!jsop_setelem_typed(arrayType, object, index, value)) return false; trackOptimizationSuccess(); *emitted = true; return true; } bool IonBuilder::setElemTryDense(bool* emitted, MDefinition* object, MDefinition* index, MDefinition* value, bool writeHole) { MOZ_ASSERT(*emitted == false); if (!ElementAccessIsDenseNative(constraints(), object, index)) { trackOptimizationOutcome(TrackedOutcome::AccessNotDense); return true; } if (PropertyWriteNeedsTypeBarrier(alloc(), constraints(), current, &object, nullptr, &value, /* canModify = */ true)) { trackOptimizationOutcome(TrackedOutcome::NeedsTypeBarrier); return true; } if (!object->resultTypeSet()) { trackOptimizationOutcome(TrackedOutcome::NoTypeInfo); return true; } TemporaryTypeSet::DoubleConversion conversion = object->resultTypeSet()->convertDoubleElements(constraints()); // If AmbiguousDoubleConversion, only handle int32 values for now. if (conversion == TemporaryTypeSet::AmbiguousDoubleConversion && value->type() != MIRType::Int32) { trackOptimizationOutcome(TrackedOutcome::ArrayDoubleConversion); return true; } // Don't generate a fast path if there have been bounds check failures // and this access might be on a sparse property. if (ElementAccessHasExtraIndexedProperty(this, object) && failedBoundsCheck_) { trackOptimizationOutcome(TrackedOutcome::ProtoIndexedProps); return true; } // Emit dense setelem variant. if (!jsop_setelem_dense(conversion, object, index, value, writeHole, emitted)) return false; if (!*emitted) { trackOptimizationOutcome(TrackedOutcome::NonWritableProperty); return true; } trackOptimizationSuccess(); return true; } bool IonBuilder::setElemTryArguments(bool* emitted, MDefinition* object, MDefinition* index, MDefinition* value) { MOZ_ASSERT(*emitted == false); if (object->type() != MIRType::MagicOptimizedArguments) return true; // Arguments are not supported yet. return abort("NYI arguments[]="); } bool IonBuilder::setElemTryCache(bool* emitted, MDefinition* object, MDefinition* index, MDefinition* value) { MOZ_ASSERT(*emitted == false); if (!object->mightBeType(MIRType::Object)) { trackOptimizationOutcome(TrackedOutcome::NotObject); return true; } if (!index->mightBeType(MIRType::Int32) && !index->mightBeType(MIRType::String) && !index->mightBeType(MIRType::Symbol)) { trackOptimizationOutcome(TrackedOutcome::IndexType); return true; } bool barrier = true; bool indexIsInt32 = index->type() == MIRType::Int32; if (indexIsInt32 && !PropertyWriteNeedsTypeBarrier(alloc(), constraints(), current, &object, nullptr, &value, /* canModify = */ true)) { barrier = false; } // We can avoid worrying about holes in the IC if we know a priori we are safe // from them. If TI can guard that there are no indexed properties on the prototype // chain, we know that we anen't missing any setters by overwriting the hole with // another value. bool guardHoles = ElementAccessHasExtraIndexedProperty(this, object); // Make sure the object being written to doesn't have copy on write elements. const Class* clasp = object->resultTypeSet() ? object->resultTypeSet()->getKnownClass(constraints()) : nullptr; bool checkNative = !clasp || !clasp->isNative(); object = addMaybeCopyElementsForWrite(object, checkNative); if (NeedsPostBarrier(value)) { if (indexIsInt32) current->add(MPostWriteElementBarrier::New(alloc(), object, value, index)); else current->add(MPostWriteBarrier::New(alloc(), object, value)); } // Emit SetPropertyCache. bool strict = JSOp(*pc) == JSOP_STRICTSETELEM; MSetPropertyCache* ins = MSetPropertyCache::New(alloc(), object, index, value, strict, barrier, guardHoles); current->add(ins); current->push(value); if (!resumeAfter(ins)) return false; trackOptimizationSuccess(); *emitted = true; return true; } bool IonBuilder::jsop_setelem_dense(TemporaryTypeSet::DoubleConversion conversion, MDefinition* obj, MDefinition* id, MDefinition* value, bool writeHole, bool* emitted) { MOZ_ASSERT(*emitted == false); MIRType elementType = DenseNativeElementType(constraints(), obj); bool packed = ElementAccessIsPacked(constraints(), obj); // Writes which are on holes in the object do not have to bail out if they // cannot hit another indexed property on the object or its prototypes. bool hasNoExtraIndexedProperty = !ElementAccessHasExtraIndexedProperty(this, obj); bool mayBeFrozen = ElementAccessMightBeFrozen(constraints(), obj); if (mayBeFrozen && !hasNoExtraIndexedProperty) { // FallibleStoreElement does not know how to deal with extra indexed // properties on the prototype. This case should be rare so we fall back // to an IC. return true; } *emitted = true; // Ensure id is an integer. MInstruction* idInt32 = MToInt32::New(alloc(), id); current->add(idInt32); id = idInt32; if (NeedsPostBarrier(value)) current->add(MPostWriteElementBarrier::New(alloc(), obj, value, id)); // Copy the elements vector if necessary. obj = addMaybeCopyElementsForWrite(obj, /* checkNative = */ false); // Get the elements vector. MElements* elements = MElements::New(alloc(), obj); current->add(elements); // Ensure the value is a double, if double conversion might be needed. MDefinition* newValue = value; switch (conversion) { case TemporaryTypeSet::AlwaysConvertToDoubles: case TemporaryTypeSet::MaybeConvertToDoubles: { MInstruction* valueDouble = MToDouble::New(alloc(), value); current->add(valueDouble); newValue = valueDouble; break; } case TemporaryTypeSet::AmbiguousDoubleConversion: { MOZ_ASSERT(value->type() == MIRType::Int32); MInstruction* maybeDouble = MMaybeToDoubleElement::New(alloc(), elements, value); current->add(maybeDouble); newValue = maybeDouble; break; } case TemporaryTypeSet::DontConvertToDoubles: break; default: MOZ_CRASH("Unknown double conversion"); } // Use MStoreElementHole if this SETELEM has written to out-of-bounds // indexes in the past. Otherwise, use MStoreElement so that we can hoist // the initialized length and bounds check. // If an object may have been frozen, no previous expectation hold and we // fallback to MFallibleStoreElement. MInstruction* store; MStoreElementCommon* common = nullptr; if (writeHole && hasNoExtraIndexedProperty && !mayBeFrozen) { MStoreElementHole* ins = MStoreElementHole::New(alloc(), obj, elements, id, newValue); store = ins; common = ins; current->add(ins); current->push(value); } else if (mayBeFrozen) { MOZ_ASSERT(hasNoExtraIndexedProperty, "FallibleStoreElement codegen assumes no extra indexed properties"); bool strict = IsStrictSetPC(pc); MFallibleStoreElement* ins = MFallibleStoreElement::New(alloc(), obj, elements, id, newValue, strict); store = ins; common = ins; current->add(ins); current->push(value); } else { MInstruction* initLength = initializedLength(obj, elements); id = addBoundsCheck(id, initLength); bool needsHoleCheck = !packed && !hasNoExtraIndexedProperty; MStoreElement* ins = MStoreElement::New(alloc(), elements, id, newValue, needsHoleCheck); store = ins; common = ins; current->add(store); current->push(value); } if (!resumeAfter(store)) return false; if (common) { // Determine whether a write barrier is required. if (obj->resultTypeSet()->propertyNeedsBarrier(constraints(), JSID_VOID)) common->setNeedsBarrier(); if (elementType != MIRType::None && packed) common->setElementType(elementType); } return true; } bool IonBuilder::jsop_setelem_typed(Scalar::Type arrayType, MDefinition* obj, MDefinition* id, MDefinition* value) { SetElemICInspector icInspect(inspector->setElemICInspector(pc)); bool expectOOB = icInspect.sawOOBTypedArrayWrite(); if (expectOOB) spew("Emitting OOB TypedArray SetElem"); // Ensure id is an integer. MInstruction* idInt32 = MToInt32::New(alloc(), id); current->add(idInt32); id = idInt32; // Get length, bounds-check, then get elements, and add all instructions. MInstruction* length; MInstruction* elements; BoundsChecking checking = expectOOB ? SkipBoundsCheck : DoBoundsCheck; addTypedArrayLengthAndData(obj, checking, &id, &length, &elements); // Clamp value to [0, 255] for Uint8ClampedArray. MDefinition* toWrite = value; if (arrayType == Scalar::Uint8Clamped) { toWrite = MClampToUint8::New(alloc(), value); current->add(toWrite->toInstruction()); } // Store the value. MInstruction* ins; if (expectOOB) { ins = MStoreTypedArrayElementHole::New(alloc(), elements, length, id, toWrite, arrayType); } else { MStoreUnboxedScalar* store = MStoreUnboxedScalar::New(alloc(), elements, id, toWrite, arrayType, MStoreUnboxedScalar::TruncateInput); ins = store; } current->add(ins); current->push(value); return resumeAfter(ins); } bool IonBuilder::jsop_length() { if (jsop_length_fastPath()) return true; PropertyName* name = info().getAtom(pc)->asPropertyName(); return jsop_getprop(name); } bool IonBuilder::jsop_length_fastPath() { TemporaryTypeSet* types = bytecodeTypes(pc); if (types->getKnownMIRType() != MIRType::Int32) return false; MDefinition* obj = current->peek(-1); if (shouldAbortOnPreliminaryGroups(obj)) return false; if (obj->mightBeType(MIRType::String)) { if (obj->mightBeType(MIRType::Object)) return false; current->pop(); MStringLength* ins = MStringLength::New(alloc(), obj); current->add(ins); current->push(ins); return true; } if (obj->mightBeType(MIRType::Object)) { TemporaryTypeSet* objTypes = obj->resultTypeSet(); // Compute the length for array objects. if (objTypes && objTypes->getKnownClass(constraints()) == &ArrayObject::class_ && !objTypes->hasObjectFlags(constraints(), OBJECT_FLAG_LENGTH_OVERFLOW)) { current->pop(); MElements* elements = MElements::New(alloc(), obj); current->add(elements); // Read length. MArrayLength* length = MArrayLength::New(alloc(), elements); current->add(length); current->push(length); return true; } // Compute the length for array typed objects. TypedObjectPrediction prediction = typedObjectPrediction(obj); if (!prediction.isUseless()) { TypeSet::ObjectKey* globalKey = TypeSet::ObjectKey::get(&script()->global()); if (globalKey->hasFlags(constraints(), OBJECT_FLAG_TYPED_OBJECT_HAS_DETACHED_BUFFER)) return false; MInstruction* length; int32_t sizedLength; if (prediction.hasKnownArrayLength(&sizedLength)) { obj->setImplicitlyUsedUnchecked(); length = MConstant::New(alloc(), Int32Value(sizedLength)); } else { return false; } current->pop(); current->add(length); current->push(length); return true; } } return false; } bool IonBuilder::jsop_arguments() { if (info().needsArgsObj()) { current->push(current->argumentsObject()); return true; } MOZ_ASSERT(lazyArguments_); current->push(lazyArguments_); return true; } bool IonBuilder::jsop_newtarget() { if (!info().funMaybeLazy()) { MOZ_ASSERT(!info().script()->isForEval()); pushConstant(NullValue()); return true; } MOZ_ASSERT(info().funMaybeLazy()); if (info().funMaybeLazy()->isArrow()) { MArrowNewTarget* arrowNewTarget = MArrowNewTarget::New(alloc(), getCallee()); current->add(arrowNewTarget); current->push(arrowNewTarget); return true; } if (inliningDepth_ == 0) { MNewTarget* newTarget = MNewTarget::New(alloc()); current->add(newTarget); current->push(newTarget); return true; } if (!inlineCallInfo_->constructing()) { pushConstant(UndefinedValue()); return true; } current->push(inlineCallInfo_->getNewTarget()); return true; } bool IonBuilder::jsop_rest() { if (info().analysisMode() == Analysis_ArgumentsUsage) { // There's no BaselineScript with the template object. Just push a // dummy value, it does not affect the arguments analysis. MUnknownValue* unknown = MUnknownValue::New(alloc()); current->add(unknown); current->push(unknown); return true; } ArrayObject* templateObject = &inspector->getTemplateObject(pc)->as(); if (inliningDepth_ == 0) { // We don't know anything about the callee. MArgumentsLength* numActuals = MArgumentsLength::New(alloc()); current->add(numActuals); // Pass in the number of actual arguments, the number of formals (not // including the rest parameter slot itself), and the template object. MRest* rest = MRest::New(alloc(), constraints(), numActuals, info().nargs() - 1, templateObject); current->add(rest); current->push(rest); return true; } // We know the exact number of arguments the callee pushed. unsigned numActuals = inlineCallInfo_->argc(); unsigned numFormals = info().nargs() - 1; unsigned numRest = numActuals > numFormals ? numActuals - numFormals : 0; if (!jsop_newarray(numRest)) return false; if (numRest == 0) { // No more updating to do. (Note that in this one case the length from // the template object is already correct.) return true; } MDefinition *array = current->peek(-1); MElements* elements = MElements::New(alloc(), array); current->add(elements); // Unroll the argument copy loop. We don't need to do any bounds or hole // checking here. MConstant* index = nullptr; for (unsigned i = numFormals; i < numActuals; i++) { index = MConstant::New(alloc(), Int32Value(i - numFormals)); current->add(index); MDefinition* arg = inlineCallInfo_->argv()[i]; MStoreElement* store = MStoreElement::New(alloc(), elements, index, arg, /* needsHoleCheck = */ false); current->add(store); if (NeedsPostBarrier(arg)) current->add(MPostWriteBarrier::New(alloc(), array, arg)); } // The array's length is incorrectly 0 now, from the template object // created by BaselineCompiler::emit_JSOP_REST() before the actual argument // count was known. Set the correct length now that we know that count. MSetArrayLength* length = MSetArrayLength::New(alloc(), elements, index); current->add(length); // Update the initialized length for all the (necessarily non-hole) // elements added. MSetInitializedLength* initLength = MSetInitializedLength::New(alloc(), elements, index); current->add(initLength); return true; } bool IonBuilder::jsop_checkisobj(uint8_t kind) { MDefinition* toCheck = current->peek(-1); if (toCheck->type() == MIRType::Object) { toCheck->setImplicitlyUsedUnchecked(); return true; } MCheckIsObj* check = MCheckIsObj::New(alloc(), current->pop(), kind); current->add(check); current->push(check); return true; } bool IonBuilder::jsop_checkiscallable(uint8_t kind) { MCheckIsCallable* check = MCheckIsCallable::New(alloc(), current->pop(), kind); current->add(check); current->push(check); return true; } bool IonBuilder::jsop_checkobjcoercible() { MDefinition* toCheck = current->peek(-1); if (!toCheck->mightBeType(MIRType::Undefined) && !toCheck->mightBeType(MIRType::Null)) { toCheck->setImplicitlyUsedUnchecked(); return true; } MOZ_ASSERT(toCheck->type() == MIRType::Value || toCheck->type() == MIRType::Null || toCheck->type() == MIRType::Undefined); // If we want to squeeze more perf here, we can throw without checking, // if IsNullOrUndefined(toCheck->type()). Since this is a failure case, // it should be OK. MCheckObjCoercible* check = MCheckObjCoercible::New(alloc(), current->pop()); current->add(check); current->push(check); return resumeAfter(check); } uint32_t IonBuilder::getDefiniteSlot(TemporaryTypeSet* types, PropertyName* name, uint32_t* pnfixed) { if (!types || types->unknownObject() || !types->objectOrSentinel()) { trackOptimizationOutcome(TrackedOutcome::NoTypeInfo); return UINT32_MAX; } uint32_t slot = UINT32_MAX; for (size_t i = 0; i < types->getObjectCount(); i++) { TypeSet::ObjectKey* key = types->getObject(i); if (!key) continue; if (key->unknownProperties()) { trackOptimizationOutcome(TrackedOutcome::UnknownProperties); return UINT32_MAX; } if (key->isSingleton()) { trackOptimizationOutcome(TrackedOutcome::Singleton); return UINT32_MAX; } HeapTypeSetKey property = key->property(NameToId(name)); if (!property.maybeTypes() || !property.maybeTypes()->definiteProperty() || property.nonData(constraints())) { trackOptimizationOutcome(TrackedOutcome::NotFixedSlot); return UINT32_MAX; } // Definite slots will always be fixed slots when they are in the // allowable range for fixed slots. size_t nfixed = NativeObject::MAX_FIXED_SLOTS; uint32_t propertySlot = property.maybeTypes()->definiteSlot(); if (slot == UINT32_MAX) { slot = propertySlot; *pnfixed = nfixed; } else if (slot != propertySlot || nfixed != *pnfixed) { trackOptimizationOutcome(TrackedOutcome::InconsistentFixedSlot); return UINT32_MAX; } } return slot; } bool IonBuilder::jsop_runonce() { MRunOncePrologue* ins = MRunOncePrologue::New(alloc()); current->add(ins); return resumeAfter(ins); } bool IonBuilder::jsop_not() { MDefinition* value = current->pop(); MNot* ins = MNot::New(alloc(), value, constraints()); current->add(ins); current->push(ins); return true; } bool IonBuilder::objectsHaveCommonPrototype(TemporaryTypeSet* types, PropertyName* name, bool isGetter, JSObject* foundProto, bool* guardGlobal) { // With foundProto a prototype with a getter or setter for name, return // whether looking up name on any object in |types| will go through // foundProto, i.e. all the objects have foundProto on their prototype // chain and do not have a property for name before reaching foundProto. // No sense looking if we don't know what's going on. if (!types || types->unknownObject()) return false; *guardGlobal = false; for (unsigned i = 0; i < types->getObjectCount(); i++) { if (types->getSingleton(i) == foundProto) continue; TypeSet::ObjectKey* key = types->getObject(i); if (!key) continue; while (key) { if (key->unknownProperties()) return false; const Class* clasp = key->clasp(); if (!ClassHasEffectlessLookup(clasp)) return false; JSObject* singleton = key->isSingleton() ? key->singleton() : nullptr; if (ObjectHasExtraOwnProperty(compartment, key, NameToId(name))) { if (!singleton || !singleton->is()) return false; *guardGlobal = true; } // Look for a getter/setter on the class itself which may need // to be called. if (isGetter && clasp->getOpsGetProperty()) return false; if (!isGetter && clasp->getOpsSetProperty()) return false; // Test for isOwnProperty() without freezing. If we end up // optimizing, freezePropertiesForCommonPropFunc will freeze the // property type sets later on. HeapTypeSetKey property = key->property(NameToId(name)); if (TypeSet* types = property.maybeTypes()) { if (!types->empty() || types->nonDataProperty()) return false; } if (singleton) { if (CanHaveEmptyPropertyTypesForOwnProperty(singleton)) { MOZ_ASSERT(singleton->is()); *guardGlobal = true; } } JSObject* proto = checkNurseryObject(key->proto().toObjectOrNull()); if (proto == foundProto) break; if (!proto) { // The foundProto being searched for did not show up on the // object's prototype chain. return false; } key = TypeSet::ObjectKey::get(proto); } } return true; } void IonBuilder::freezePropertiesForCommonPrototype(TemporaryTypeSet* types, PropertyName* name, JSObject* foundProto, bool allowEmptyTypesforGlobal/* = false*/) { for (unsigned i = 0; i < types->getObjectCount(); i++) { // If we found a Singleton object's own-property, there's nothing to // freeze. if (types->getSingleton(i) == foundProto) continue; TypeSet::ObjectKey* key = types->getObject(i); if (!key) continue; while (true) { HeapTypeSetKey property = key->property(NameToId(name)); JS_ALWAYS_TRUE(!property.isOwnProperty(constraints(), allowEmptyTypesforGlobal)); // Don't mark the proto. It will be held down by the shape // guard. This allows us to use properties found on prototypes // with properties unknown to TI. if (key->proto() == TaggedProto(foundProto)) break; key = TypeSet::ObjectKey::get(key->proto().toObjectOrNull()); } } } bool IonBuilder::testCommonGetterSetter(TemporaryTypeSet* types, PropertyName* name, bool isGetter, JSObject* foundProto, Shape* lastProperty, JSFunction* getterOrSetter, MDefinition** guard, Shape* globalShape/* = nullptr*/, MDefinition** globalGuard/* = nullptr */) { MOZ_ASSERT_IF(globalShape, globalGuard); bool guardGlobal; // Check if all objects being accessed will lookup the name through foundProto. if (!objectsHaveCommonPrototype(types, name, isGetter, foundProto, &guardGlobal) || (guardGlobal && !globalShape)) { trackOptimizationOutcome(TrackedOutcome::MultiProtoPaths); return false; } // We can optimize the getter/setter, so freeze all involved properties to // ensure there isn't a lower shadowing getter or setter installed in the // future. freezePropertiesForCommonPrototype(types, name, foundProto, guardGlobal); // Add a shape guard on the prototype we found the property on. The rest of // the prototype chain is guarded by TI freezes, except when name is a global // name. In this case, we also have to guard on the globals shape to be able // to optimize, because the way global property sets are handled means // freezing doesn't work for what we want here. Note that a shape guard is // good enough here, even in the proxy case, because we have ensured there // are no lookup hooks for this property. if (guardGlobal) { JSObject* obj = &script()->global(); MDefinition* globalObj = constant(ObjectValue(*obj)); *globalGuard = addShapeGuard(globalObj, globalShape, Bailout_ShapeGuard); } if (foundProto->isNative()) { NativeObject& nativeProto = foundProto->as(); if (nativeProto.lastProperty() == lastProperty) { // The proto shape is the same as it was at the point when we // created the baseline IC, so looking up the prop on the object as // it is now should be safe. Shape* propShape = nativeProto.lookupPure(name); MOZ_ASSERT_IF(isGetter, propShape->getterObject() == getterOrSetter); MOZ_ASSERT_IF(!isGetter, propShape->setterObject() == getterOrSetter); if (propShape && !propShape->configurable()) return true; } } MInstruction* wrapper = constant(ObjectValue(*foundProto)); *guard = addShapeGuard(wrapper, lastProperty, Bailout_ShapeGuard); return true; } void IonBuilder::replaceMaybeFallbackFunctionGetter(MGetPropertyCache* cache) { // Discard the last prior resume point of the previous MGetPropertyCache. WrapMGetPropertyCache rai(maybeFallbackFunctionGetter_); maybeFallbackFunctionGetter_ = cache; } bool IonBuilder::annotateGetPropertyCache(MDefinition* obj, PropertyName* name, MGetPropertyCache* getPropCache, TemporaryTypeSet* objTypes, TemporaryTypeSet* pushedTypes) { // Ensure every pushed value is a singleton. if (pushedTypes->unknownObject() || pushedTypes->baseFlags() != 0) return true; for (unsigned i = 0; i < pushedTypes->getObjectCount(); i++) { if (pushedTypes->getGroup(i) != nullptr) return true; } // Object's typeset should be a proper object if (!objTypes || objTypes->baseFlags() || objTypes->unknownObject()) return true; unsigned int objCount = objTypes->getObjectCount(); if (objCount == 0) return true; InlinePropertyTable* inlinePropTable = getPropCache->initInlinePropertyTable(alloc(), pc); if (!inlinePropTable) return false; // Ensure that the relevant property typeset for each group is // is a single-object typeset containing a JSFunction for (unsigned int i = 0; i < objCount; i++) { ObjectGroup* group = objTypes->getGroup(i); if (!group) continue; TypeSet::ObjectKey* key = TypeSet::ObjectKey::get(group); if (key->unknownProperties() || !key->proto().isObject()) continue; JSObject* proto = checkNurseryObject(key->proto().toObject()); const Class* clasp = key->clasp(); if (!ClassHasEffectlessLookup(clasp) || ObjectHasExtraOwnProperty(compartment, key, NameToId(name))) continue; HeapTypeSetKey ownTypes = key->property(NameToId(name)); if (ownTypes.isOwnProperty(constraints())) continue; JSObject* singleton = testSingletonProperty(proto, NameToId(name)); if (!singleton || !singleton->is()) continue; // Don't add cases corresponding to non-observed pushes if (!pushedTypes->hasType(TypeSet::ObjectType(singleton))) continue; if (!inlinePropTable->addEntry(alloc(), group, &singleton->as())) return false; } if (inlinePropTable->numEntries() == 0) { getPropCache->clearInlinePropertyTable(); return true; } #ifdef JS_JITSPEW if (inlinePropTable->numEntries() > 0) JitSpew(JitSpew_Inlining, "Annotated GetPropertyCache with %d/%d inline cases", (int) inlinePropTable->numEntries(), (int) objCount); #endif // If we successfully annotated the GetPropertyCache and there are inline cases, // then keep a resume point of the state right before this instruction for use // later when we have to bail out to this point in the fallback case of a // PolyInlineDispatch. if (inlinePropTable->numEntries() > 0) { // Push the object back onto the stack temporarily to capture the resume point. current->push(obj); MResumePoint* resumePoint = MResumePoint::New(alloc(), current, pc, MResumePoint::ResumeAt); if (!resumePoint) return false; inlinePropTable->setPriorResumePoint(resumePoint); replaceMaybeFallbackFunctionGetter(getPropCache); current->pop(); } return true; } // Returns true if an idempotent cache has ever invalidated this script // or an outer script. bool IonBuilder::invalidatedIdempotentCache() { IonBuilder* builder = this; do { if (builder->script()->invalidatedIdempotentCache()) return true; builder = builder->callerBuilder_; } while (builder); return false; } bool IonBuilder::loadSlot(MDefinition* obj, size_t slot, size_t nfixed, MIRType rvalType, BarrierKind barrier, TemporaryTypeSet* types) { if (slot < nfixed) { MLoadFixedSlot* load = MLoadFixedSlot::New(alloc(), obj, slot); current->add(load); current->push(load); load->setResultType(rvalType); return pushTypeBarrier(load, types, barrier); } MSlots* slots = MSlots::New(alloc(), obj); current->add(slots); MLoadSlot* load = MLoadSlot::New(alloc(), slots, slot - nfixed); current->add(load); current->push(load); load->setResultType(rvalType); return pushTypeBarrier(load, types, barrier); } bool IonBuilder::loadSlot(MDefinition* obj, Shape* shape, MIRType rvalType, BarrierKind barrier, TemporaryTypeSet* types) { return loadSlot(obj, shape->slot(), shape->numFixedSlots(), rvalType, barrier, types); } bool IonBuilder::storeSlot(MDefinition* obj, size_t slot, size_t nfixed, MDefinition* value, bool needsBarrier, MIRType slotType /* = MIRType::None */) { if (slot < nfixed) { MStoreFixedSlot* store = MStoreFixedSlot::New(alloc(), obj, slot, value); current->add(store); current->push(value); if (needsBarrier) store->setNeedsBarrier(); return resumeAfter(store); } MSlots* slots = MSlots::New(alloc(), obj); current->add(slots); MStoreSlot* store = MStoreSlot::New(alloc(), slots, slot - nfixed, value); current->add(store); current->push(value); if (needsBarrier) store->setNeedsBarrier(); if (slotType != MIRType::None) store->setSlotType(slotType); return resumeAfter(store); } bool IonBuilder::storeSlot(MDefinition* obj, Shape* shape, MDefinition* value, bool needsBarrier, MIRType slotType /* = MIRType::None */) { MOZ_ASSERT(shape->writable()); return storeSlot(obj, shape->slot(), shape->numFixedSlots(), value, needsBarrier, slotType); } bool IonBuilder::shouldAbortOnPreliminaryGroups(MDefinition *obj) { // Watch for groups which still have preliminary object information and // have not had the new script properties or unboxed layout analyses // performed. Normally this is done after a small number of the objects // have been created, but if only a few have been created we can still // perform the analysis with a smaller object population. The analysis can // have side effects so we will end up aborting compilation after building // finishes and retrying later. TemporaryTypeSet *types = obj->resultTypeSet(); if (!types || types->unknownObject()) return false; bool preliminary = false; for (size_t i = 0; i < types->getObjectCount(); i++) { TypeSet::ObjectKey* key = types->getObject(i); if (!key) continue; if (ObjectGroup* group = key->maybeGroup()) { if (group->hasUnanalyzedPreliminaryObjects()) { addAbortedPreliminaryGroup(group); preliminary = true; } } } return preliminary; } MDefinition* IonBuilder::maybeUnboxForPropertyAccess(MDefinition* def) { if (def->type() != MIRType::Value) return def; MIRType type = inspector->expectedPropertyAccessInputType(pc); if (type == MIRType::Value || !def->mightBeType(type)) return def; MUnbox* unbox = MUnbox::New(alloc(), def, type, MUnbox::Fallible); current->add(unbox); // Fixup type information for a common case where a property call // is converted to the following bytecodes // // a.foo() // ================= Compiles to ================ // LOAD "a" // DUP // CALLPROP "foo" // SWAP // CALL 0 // // If we have better type information to unbox the first copy going into // the CALLPROP operation, we can replace the duplicated copy on the // stack as well. if (*pc == JSOP_CALLPROP || *pc == JSOP_CALLELEM) { uint32_t idx = current->stackDepth() - 1; MOZ_ASSERT(current->getSlot(idx) == def); current->setSlot(idx, unbox); } return unbox; } bool IonBuilder::jsop_getprop(PropertyName* name) { bool emitted = false; startTrackingOptimizations(); MDefinition* obj = current->pop(); TemporaryTypeSet* types = bytecodeTypes(pc); trackTypeInfo(TrackedTypeSite::Receiver, obj->type(), obj->resultTypeSet()); if (!info().isAnalysis()) { // The calls below can abort compilation, so we only try this if we're // not analyzing. // Try to optimize arguments.length. trackOptimizationAttempt(TrackedStrategy::GetProp_ArgumentsLength); if (!getPropTryArgumentsLength(&emitted, obj) || emitted) return emitted; // Try to optimize arguments.callee. trackOptimizationAttempt(TrackedStrategy::GetProp_ArgumentsCallee); if (!getPropTryArgumentsCallee(&emitted, obj, name) || emitted) return emitted; } obj = maybeUnboxForPropertyAccess(obj); BarrierKind barrier = PropertyReadNeedsTypeBarrier(analysisContext, constraints(), obj, name, types); // Try to optimize to a specific constant. trackOptimizationAttempt(TrackedStrategy::GetProp_InferredConstant); if (barrier == BarrierKind::NoBarrier) { if (!getPropTryInferredConstant(&emitted, obj, name, types) || emitted) return emitted; } else { trackOptimizationOutcome(TrackedOutcome::NeedsTypeBarrier); } // Always use a call if we are performing analysis and // not actually emitting code, to simplify later analysis. Also skip deeper // analysis if there are no known types for this operation, as it will // always invalidate when executing. if (info().isAnalysis() || types->empty() || shouldAbortOnPreliminaryGroups(obj)) { if (types->empty()) { // Since no further optimizations will be tried, use the IC // strategy, which would have been the last one to be tried, as a // sentinel value for why everything failed. trackOptimizationAttempt(TrackedStrategy::GetProp_InlineCache); trackOptimizationOutcome(TrackedOutcome::NoTypeInfo); } MCallGetProperty* call = MCallGetProperty::New(alloc(), obj, name); current->add(call); // During the definite properties analysis we can still try to bake in // constants read off the prototype chain, to allow inlining later on. // In this case we still need the getprop call so that the later // analysis knows when the |this| value has been read from. if (info().isAnalysis()) { if (!getPropTryConstant(&emitted, obj, NameToId(name), types) || emitted) return emitted; } current->push(call); return resumeAfter(call) && pushTypeBarrier(call, types, BarrierKind::TypeSet); } // Try to optimize accesses on outer window proxies, for example window.foo. // This needs to come before the various strategies getPropTryInnerize tries // internally, since some of those strategies will "succeed" in silly ways // even for an outer object. trackOptimizationAttempt(TrackedStrategy::GetProp_Innerize); if (!getPropTryInnerize(&emitted, obj, name, types) || emitted) return emitted; if (!forceInlineCaches()) { // Try to hardcode known constants. trackOptimizationAttempt(TrackedStrategy::GetProp_Constant); if (!getPropTryConstant(&emitted, obj, NameToId(name), types) || emitted) return emitted; // Try to hardcode known not-defined trackOptimizationAttempt(TrackedStrategy::GetProp_NotDefined); if (!getPropTryNotDefined(&emitted, obj, NameToId(name), types) || emitted) return emitted; // Try to emit loads from known binary data blocks trackOptimizationAttempt(TrackedStrategy::GetProp_TypedObject); if (!getPropTryTypedObject(&emitted, obj, name) || emitted) return emitted; // Try to emit loads from definite slots. trackOptimizationAttempt(TrackedStrategy::GetProp_DefiniteSlot); if (!getPropTryDefiniteSlot(&emitted, obj, name, barrier, types) || emitted) return emitted; // Try to inline a common property getter, or make a call. trackOptimizationAttempt(TrackedStrategy::GetProp_CommonGetter); if (!getPropTryCommonGetter(&emitted, obj, name, types) || emitted) return emitted; // Try to emit a monomorphic/polymorphic access based on baseline caches. trackOptimizationAttempt(TrackedStrategy::GetProp_InlineAccess); if (!getPropTryInlineAccess(&emitted, obj, name, barrier, types) || emitted) return emitted; // Try to emit loads from a module namespace. trackOptimizationAttempt(TrackedStrategy::GetProp_ModuleNamespace); if (!getPropTryModuleNamespace(&emitted, obj, name, barrier, types) || emitted) return emitted; } // Try to emit a polymorphic cache. trackOptimizationAttempt(TrackedStrategy::GetProp_InlineCache); if (!getPropTryCache(&emitted, obj, name, barrier, types) || emitted) return emitted; // Try to emit a shared stub. trackOptimizationAttempt(TrackedStrategy::GetProp_SharedCache); if (!getPropTrySharedStub(&emitted, obj, types) || emitted) return emitted; // Emit a call. MCallGetProperty* call = MCallGetProperty::New(alloc(), obj, name); current->add(call); current->push(call); if (!resumeAfter(call)) return false; if (*pc == JSOP_CALLPROP && IsNullOrUndefined(obj->type())) { // Due to inlining, it's possible the observed TypeSet is non-empty, // even though we know |obj| is null/undefined and the MCallGetProperty // will throw. Don't push a TypeBarrier in this case, to avoid // inlining the following (unreachable) JSOP_CALL. return true; } return pushTypeBarrier(call, types, BarrierKind::TypeSet); } bool IonBuilder::improveThisTypesForCall() { // After a CALLPROP (or CALLELEM) for obj.prop(), the this-value and callee // for the call are on top of the stack: // // ... [this: obj], [callee: obj.prop] // // If obj is null or undefined, obj.prop would have thrown an exception so // at this point we can remove null and undefined from obj's TypeSet, to // improve type information for the call that will follow. MOZ_ASSERT(*pc == JSOP_CALLPROP || *pc == JSOP_CALLELEM); // Ensure |this| has types {object, null/undefined}. MDefinition* thisDef = current->peek(-2); if (thisDef->type() != MIRType::Value || !thisDef->mightBeType(MIRType::Object) || !thisDef->resultTypeSet() || !thisDef->resultTypeSet()->objectOrSentinel()) { return true; } // Remove null/undefined from the TypeSet. TemporaryTypeSet* types = thisDef->resultTypeSet()->cloneObjectsOnly(alloc_->lifoAlloc()); if (!types) return false; MFilterTypeSet* filter = MFilterTypeSet::New(alloc(), thisDef, types); current->add(filter); current->rewriteAtDepth(-2, filter); // FilterTypeSetPolicy::adjustInputs will insert an infallible Unbox(Object) // for the input. Don't hoist this unbox above the getprop or getelem // operation. filter->setDependency(current->peek(-1)->toInstruction()); return true; } bool IonBuilder::checkIsDefinitelyOptimizedArguments(MDefinition* obj, bool* isOptimizedArgs) { if (obj->type() != MIRType::MagicOptimizedArguments) { if (script()->argumentsHasVarBinding() && obj->mightBeType(MIRType::MagicOptimizedArguments)) { return abort("Type is not definitely lazy arguments."); } *isOptimizedArgs = false; return true; } *isOptimizedArgs = true; return true; } bool IonBuilder::getPropTryInferredConstant(bool* emitted, MDefinition* obj, PropertyName* name, TemporaryTypeSet* types) { MOZ_ASSERT(*emitted == false); // Need a result typeset to optimize. TemporaryTypeSet* objTypes = obj->resultTypeSet(); if (!objTypes) { trackOptimizationOutcome(TrackedOutcome::NoTypeInfo); return true; } JSObject* singleton = objTypes->maybeSingleton(); if (!singleton) { trackOptimizationOutcome(TrackedOutcome::NotSingleton); return true; } TypeSet::ObjectKey* key = TypeSet::ObjectKey::get(singleton); if (key->unknownProperties()) { trackOptimizationOutcome(TrackedOutcome::UnknownProperties); return true; } HeapTypeSetKey property = key->property(NameToId(name)); Value constantValue = UndefinedValue(); if (property.constant(constraints(), &constantValue)) { spew("Optimized constant property"); obj->setImplicitlyUsedUnchecked(); pushConstant(constantValue); types->addType(TypeSet::GetValueType(constantValue), alloc_->lifoAlloc()); trackOptimizationSuccess(); *emitted = true; } return true; } bool IonBuilder::getPropTryArgumentsLength(bool* emitted, MDefinition* obj) { MOZ_ASSERT(*emitted == false); if (JSOp(*pc) != JSOP_LENGTH) return true; bool isOptimizedArgs = false; if (!checkIsDefinitelyOptimizedArguments(obj, &isOptimizedArgs)) return false; if (!isOptimizedArgs) return true; trackOptimizationSuccess(); *emitted = true; obj->setImplicitlyUsedUnchecked(); // We don't know anything from the callee if (inliningDepth_ == 0) { MInstruction* ins = MArgumentsLength::New(alloc()); current->add(ins); current->push(ins); return true; } // We are inlining and know the number of arguments the callee pushed pushConstant(Int32Value(inlineCallInfo_->argv().length())); return true; } bool IonBuilder::getPropTryArgumentsCallee(bool* emitted, MDefinition* obj, PropertyName* name) { MOZ_ASSERT(*emitted == false); if (name != names().callee) return true; bool isOptimizedArgs = false; if (!checkIsDefinitelyOptimizedArguments(obj, &isOptimizedArgs)) return false; if (!isOptimizedArgs) return true; MOZ_ASSERT(script()->hasMappedArgsObj()); obj->setImplicitlyUsedUnchecked(); current->push(getCallee()); trackOptimizationSuccess(); *emitted = true; return true; } bool IonBuilder::getPropTryConstant(bool* emitted, MDefinition* obj, jsid id, TemporaryTypeSet* types) { MOZ_ASSERT(*emitted == false); if (!types->mightBeMIRType(MIRType::Object)) { // If we have not observed an object result here, don't look for a // singleton constant. trackOptimizationOutcome(TrackedOutcome::NotObject); return true; } JSObject* singleton = testSingletonPropertyTypes(obj, id); if (!singleton) { trackOptimizationOutcome(TrackedOutcome::NotSingleton); return true; } // Property access is a known constant -- safe to emit. obj->setImplicitlyUsedUnchecked(); pushConstant(ObjectValue(*singleton)); trackOptimizationSuccess(); *emitted = true; return true; } bool IonBuilder::getPropTryNotDefined(bool* emitted, MDefinition* obj, jsid id, TemporaryTypeSet* types) { MOZ_ASSERT(*emitted == false); if (!types->mightBeMIRType(MIRType::Undefined)) { // Only optimize if we expect this property access to return undefined. trackOptimizationOutcome(TrackedOutcome::NotUndefined); return true; } ResultWithOOM res = testNotDefinedProperty(obj, id); if (res.oom) return false; if (!res.value) { trackOptimizationOutcome(TrackedOutcome::GenericFailure); return true; } obj->setImplicitlyUsedUnchecked(); pushConstant(UndefinedValue()); trackOptimizationSuccess(); *emitted = true; return true; } bool IonBuilder::getPropTryTypedObject(bool* emitted, MDefinition* obj, PropertyName* name) { TypedObjectPrediction fieldPrediction; size_t fieldOffset; size_t fieldIndex; if (!typedObjectHasField(obj, name, &fieldOffset, &fieldPrediction, &fieldIndex)) return true; switch (fieldPrediction.kind()) { case type::Simd: // FIXME (bug 894104): load into a MIRType::float32x4 etc return true; case type::Struct: case type::Array: return getPropTryComplexPropOfTypedObject(emitted, obj, fieldOffset, fieldPrediction, fieldIndex); case type::Reference: return getPropTryReferencePropOfTypedObject(emitted, obj, fieldOffset, fieldPrediction, name); case type::Scalar: return getPropTryScalarPropOfTypedObject(emitted, obj, fieldOffset, fieldPrediction); } MOZ_CRASH("Bad kind"); } bool IonBuilder::getPropTryScalarPropOfTypedObject(bool* emitted, MDefinition* typedObj, int32_t fieldOffset, TypedObjectPrediction fieldPrediction) { // Must always be loading the same scalar type Scalar::Type fieldType = fieldPrediction.scalarType(); // Don't optimize if the typed object's underlying buffer may be detached. TypeSet::ObjectKey* globalKey = TypeSet::ObjectKey::get(&script()->global()); if (globalKey->hasFlags(constraints(), OBJECT_FLAG_TYPED_OBJECT_HAS_DETACHED_BUFFER)) return true; trackOptimizationSuccess(); *emitted = true; LinearSum byteOffset(alloc()); if (!byteOffset.add(fieldOffset)) setForceAbort(); return pushScalarLoadFromTypedObject(typedObj, byteOffset, fieldType); } bool IonBuilder::getPropTryReferencePropOfTypedObject(bool* emitted, MDefinition* typedObj, int32_t fieldOffset, TypedObjectPrediction fieldPrediction, PropertyName* name) { ReferenceTypeDescr::Type fieldType = fieldPrediction.referenceType(); TypeSet::ObjectKey* globalKey = TypeSet::ObjectKey::get(&script()->global()); if (globalKey->hasFlags(constraints(), OBJECT_FLAG_TYPED_OBJECT_HAS_DETACHED_BUFFER)) return true; trackOptimizationSuccess(); *emitted = true; LinearSum byteOffset(alloc()); if (!byteOffset.add(fieldOffset)) setForceAbort(); return pushReferenceLoadFromTypedObject(typedObj, byteOffset, fieldType, name); } bool IonBuilder::getPropTryComplexPropOfTypedObject(bool* emitted, MDefinition* typedObj, int32_t fieldOffset, TypedObjectPrediction fieldPrediction, size_t fieldIndex) { // Don't optimize if the typed object's underlying buffer may be detached. TypeSet::ObjectKey* globalKey = TypeSet::ObjectKey::get(&script()->global()); if (globalKey->hasFlags(constraints(), OBJECT_FLAG_TYPED_OBJECT_HAS_DETACHED_BUFFER)) return true; // OK, perform the optimization // Identify the type object for the field. MDefinition* type = loadTypedObjectType(typedObj); MDefinition* fieldTypeObj = typeObjectForFieldFromStructType(type, fieldIndex); LinearSum byteOffset(alloc()); if (!byteOffset.add(fieldOffset)) setForceAbort(); return pushDerivedTypedObject(emitted, typedObj, byteOffset, fieldPrediction, fieldTypeObj); } bool IonBuilder::getPropTryDefiniteSlot(bool* emitted, MDefinition* obj, PropertyName* name, BarrierKind barrier, TemporaryTypeSet* types) { MOZ_ASSERT(*emitted == false); uint32_t nfixed; uint32_t slot = getDefiniteSlot(obj->resultTypeSet(), name, &nfixed); if (slot == UINT32_MAX) return true; if (obj->type() != MIRType::Object) { MGuardObject* guard = MGuardObject::New(alloc(), obj); current->add(guard); obj = guard; } MInstruction* load; if (slot < nfixed) { load = MLoadFixedSlot::New(alloc(), obj, slot); } else { MInstruction* slots = MSlots::New(alloc(), obj); current->add(slots); load = MLoadSlot::New(alloc(), slots, slot - nfixed); } if (barrier == BarrierKind::NoBarrier) load->setResultType(types->getKnownMIRType()); current->add(load); current->push(load); if (!pushTypeBarrier(load, types, barrier)) return false; trackOptimizationSuccess(); *emitted = true; return true; } bool IonBuilder::getPropTryModuleNamespace(bool* emitted, MDefinition* obj, PropertyName* name, BarrierKind barrier, TemporaryTypeSet* types) { MOZ_ASSERT(*emitted == false); TemporaryTypeSet* objTypes = obj->resultTypeSet(); if (!objTypes) { trackOptimizationOutcome(TrackedOutcome::NoTypeInfo); return true; } JSObject* singleton = objTypes->maybeSingleton(); if (!singleton) { trackOptimizationOutcome(TrackedOutcome::NotSingleton); return true; } if (!singleton->is()) { trackOptimizationOutcome(TrackedOutcome::NotModuleNamespace); return true; } ModuleNamespaceObject* ns = &singleton->as(); ModuleEnvironmentObject* env; Shape* shape; if (!ns->bindings().lookup(NameToId(name), &env, &shape)) { trackOptimizationOutcome(TrackedOutcome::UnknownProperty); return true; } obj->setImplicitlyUsedUnchecked(); MConstant* envConst = constant(ObjectValue(*env)); uint32_t slot = shape->slot(); uint32_t nfixed = env->numFixedSlots(); if (!loadSlot(envConst, slot, nfixed, types->getKnownMIRType(), barrier, types)) return false; trackOptimizationSuccess(); *emitted = true; return true; } MDefinition* IonBuilder::addShapeGuardsForGetterSetter(MDefinition* obj, JSObject* holder, Shape* holderShape, const BaselineInspector::ReceiverVector& receivers, bool isOwnProperty) { MOZ_ASSERT(holder); MOZ_ASSERT(holderShape); if (isOwnProperty) { MOZ_ASSERT(receivers.empty()); return addShapeGuard(obj, holderShape, Bailout_ShapeGuard); } MDefinition* holderDef = constant(ObjectValue(*holder)); addShapeGuard(holderDef, holderShape, Bailout_ShapeGuard); return addGuardReceiverPolymorphic(obj, receivers); } bool IonBuilder::getPropTryCommonGetter(bool* emitted, MDefinition* obj, PropertyName* name, TemporaryTypeSet* types) { MOZ_ASSERT(*emitted == false); Shape* lastProperty = nullptr; JSFunction* commonGetter = nullptr; Shape* globalShape = nullptr; JSObject* foundProto = nullptr; bool isOwnProperty = false; BaselineInspector::ReceiverVector receivers(alloc()); if (!inspector->commonGetPropFunction(pc, &foundProto, &lastProperty, &commonGetter, &globalShape, &isOwnProperty, receivers)) { return true; } TemporaryTypeSet* objTypes = obj->resultTypeSet(); MDefinition* guard = nullptr; MDefinition* globalGuard = nullptr; bool canUseTIForGetter = testCommonGetterSetter(objTypes, name, /* isGetter = */ true, foundProto, lastProperty, commonGetter, &guard, globalShape, &globalGuard); if (!canUseTIForGetter) { // If type information is bad, we can still optimize the getter if we // shape guard. obj = addShapeGuardsForGetterSetter(obj, foundProto, lastProperty, receivers, isOwnProperty); if (!obj) return false; } bool isDOM = objTypes && objTypes->isDOMClass(constraints()); if (isDOM && testShouldDOMCall(objTypes, commonGetter, JSJitInfo::Getter)) { const JSJitInfo* jitinfo = commonGetter->jitInfo(); MInstruction* get; if (jitinfo->isAlwaysInSlot) { // If our object is a singleton and we know the property is // constant (which is true if and only if the get doesn't alias // anything), we can just read the slot here and use that constant. JSObject* singleton = objTypes->maybeSingleton(); if (singleton && jitinfo->aliasSet() == JSJitInfo::AliasNone) { size_t slot = jitinfo->slotIndex; *emitted = true; pushConstant(GetReservedSlot(singleton, slot)); return true; } // We can't use MLoadFixedSlot here because it might not have the // right aliasing behavior; we want to alias DOM setters as needed. get = MGetDOMMember::New(alloc(), jitinfo, obj, guard, globalGuard); } else { get = MGetDOMProperty::New(alloc(), jitinfo, obj, guard, globalGuard); } if (!get) { return false; } current->add(get); current->push(get); if (get->isEffectful() && !resumeAfter(get)) return false; if (!pushDOMTypeBarrier(get, types, commonGetter)) return false; trackOptimizationOutcome(TrackedOutcome::DOM); *emitted = true; return true; } // Don't call the getter with a primitive value. if (obj->type() != MIRType::Object) { MGuardObject* guardObj = MGuardObject::New(alloc(), obj); current->add(guardObj); obj = guardObj; } // Spoof stack to expected state for call. // Make sure there's enough room if (!current->ensureHasSlots(2)) return false; current->push(constant(ObjectValue(*commonGetter))); current->push(obj); CallInfo callInfo(alloc(), /* constructing = */ false, /* ignoresReturnValue = */ BytecodeIsPopped(pc)); if (!callInfo.init(current, 0)) return false; if (commonGetter->isNative()) { InliningStatus status = inlineNativeGetter(callInfo, commonGetter); switch (status) { case InliningStatus_Error: return false; case InliningStatus_WarmUpCountTooLow: case InliningStatus_NotInlined: break; case InliningStatus_Inlined: trackOptimizationOutcome(TrackedOutcome::Inlined); *emitted = true; return true; } } // Inline if we can, otherwise, forget it and just generate a call. if (commonGetter->isInterpreted()) { InliningDecision decision = makeInliningDecision(commonGetter, callInfo); switch (decision) { case InliningDecision_Error: return false; case InliningDecision_DontInline: case InliningDecision_WarmUpCountTooLow: break; case InliningDecision_Inline: { InliningStatus status = inlineScriptedCall(callInfo, commonGetter); if (status == InliningStatus_Inlined) { *emitted = true; return true; } if (status == InliningStatus_Error) return false; break; } } } if (!makeCall(commonGetter, callInfo)) return false; // If the getter could have been inlined, don't track success. The call to // makeInliningDecision above would have tracked a specific reason why we // couldn't inline. if (!commonGetter->isInterpreted()) trackOptimizationSuccess(); *emitted = true; return true; } bool IonBuilder::canInlinePropertyOpShapes(const BaselineInspector::ReceiverVector& receivers) { if (receivers.empty()) { trackOptimizationOutcome(TrackedOutcome::NoShapeInfo); return false; } for (size_t i = 0; i < receivers.length(); i++) { // We inline the property access as long as the shape is not in // dictionary mode. We cannot be sure that the shape is still a // lastProperty, and calling Shape::search() on dictionary mode // shapes that aren't lastProperty is invalid. if (receivers[i].shape && receivers[i].shape->inDictionary()) { trackOptimizationOutcome(TrackedOutcome::InDictionaryMode); return false; } } return true; } static Shape* PropertyShapesHaveSameSlot(const BaselineInspector::ReceiverVector& receivers, jsid id) { Shape* firstShape = nullptr; for (size_t i = 0; i < receivers.length(); i++) { if (receivers[i].group) return nullptr; Shape* shape = receivers[i].shape->searchLinear(id); MOZ_ASSERT(shape); if (i == 0) { firstShape = shape; } else if (shape->slot() != firstShape->slot() || shape->numFixedSlots() != firstShape->numFixedSlots()) { return nullptr; } } return firstShape; } bool IonBuilder::getPropTryInlineAccess(bool* emitted, MDefinition* obj, PropertyName* name, BarrierKind barrier, TemporaryTypeSet* types) { MOZ_ASSERT(*emitted == false); BaselineInspector::ReceiverVector receivers(alloc()); if (!inspector->maybeInfoForPropertyOp(pc, receivers)) return false; if (!canInlinePropertyOpShapes(receivers)) return true; MIRType rvalType = types->getKnownMIRType(); if (barrier != BarrierKind::NoBarrier || IsNullOrUndefined(rvalType)) rvalType = MIRType::Value; if (receivers.length() == 1) { if (!receivers[0].group) { // Monomorphic load from a native object. spew("Inlining monomorphic native GETPROP"); obj = addShapeGuard(obj, receivers[0].shape, Bailout_ShapeGuard); Shape* shape = receivers[0].shape->searchLinear(NameToId(name)); MOZ_ASSERT(shape); if (!loadSlot(obj, shape, rvalType, barrier, types)) return false; trackOptimizationOutcome(TrackedOutcome::Monomorphic); *emitted = true; return true; } return true; } MOZ_ASSERT(receivers.length() > 1); spew("Inlining polymorphic GETPROP"); if (Shape* propShape = PropertyShapesHaveSameSlot(receivers, NameToId(name))) { obj = addGuardReceiverPolymorphic(obj, receivers); if (!obj) return false; if (!loadSlot(obj, propShape, rvalType, barrier, types)) return false; trackOptimizationOutcome(TrackedOutcome::Polymorphic); *emitted = true; return true; } MGetPropertyPolymorphic* load = MGetPropertyPolymorphic::New(alloc(), obj, name); current->add(load); current->push(load); for (size_t i = 0; i < receivers.length(); i++) { Shape* propShape = nullptr; if (receivers[i].shape) { propShape = receivers[i].shape->searchLinear(NameToId(name)); MOZ_ASSERT(propShape); } if (!load->addReceiver(receivers[i], propShape)) return false; } if (failedShapeGuard_) load->setNotMovable(); load->setResultType(rvalType); if (!pushTypeBarrier(load, types, barrier)) return false; trackOptimizationOutcome(TrackedOutcome::Polymorphic); *emitted = true; return true; } bool IonBuilder::getPropTryCache(bool* emitted, MDefinition* obj, PropertyName* name, BarrierKind barrier, TemporaryTypeSet* types) { MOZ_ASSERT(*emitted == false); // The input value must either be an object, or we should have strong suspicions // that it can be safely unboxed to an object. if (obj->type() != MIRType::Object) { TemporaryTypeSet* types = obj->resultTypeSet(); if (!types || !types->objectOrSentinel()) { trackOptimizationOutcome(TrackedOutcome::NoTypeInfo); return true; } } // Since getters have no guaranteed return values, we must barrier in order to be // able to attach stubs for them. if (inspector->hasSeenAccessedGetter(pc)) barrier = BarrierKind::TypeSet; // Caches can read values from prototypes, so update the barrier to // reflect such possible values. if (barrier != BarrierKind::TypeSet) { ResultWithOOM protoBarrier = PropertyReadOnPrototypeNeedsTypeBarrier(this, obj, name, types); if (protoBarrier.oom) return false; if (protoBarrier.value != BarrierKind::NoBarrier) { MOZ_ASSERT(barrier <= protoBarrier.value); barrier = protoBarrier.value; } } MConstant* id = constant(StringValue(name)); MGetPropertyCache* load = MGetPropertyCache::New(alloc(), obj, id, barrier == BarrierKind::TypeSet); // Try to mark the cache as idempotent. if (obj->type() == MIRType::Object && !invalidatedIdempotentCache()) { if (PropertyReadIsIdempotent(constraints(), obj, name)) load->setIdempotent(); } // When we are in the context of making a call from the value returned from // a property, we query the typeObject for the given property name to fill // the InlinePropertyTable of the GetPropertyCache. This information is // then used in inlineCallsite and inlineCalls, if the "this" definition is // matching the "object" definition of the GetPropertyCache (see // CanInlineGetPropertyCache). // // If this GetPropertyCache is idempotent, then we can dispatch to the right // function only by checking the typed object, instead of querying the value // of the property. Thus this GetPropertyCache can be moved into the // fallback path (see inlineObjectGroupFallback). Otherwise, we always have // to do the GetPropertyCache, and we can dispatch based on the JSFunction // value. if (JSOp(*pc) == JSOP_CALLPROP && load->idempotent()) { if (!annotateGetPropertyCache(obj, name, load, obj->resultTypeSet(), types)) return false; } current->add(load); current->push(load); if (load->isEffectful() && !resumeAfter(load)) return false; MIRType rvalType = types->getKnownMIRType(); if (barrier != BarrierKind::NoBarrier || IsNullOrUndefined(rvalType)) rvalType = MIRType::Value; load->setResultType(rvalType); if (!pushTypeBarrier(load, types, barrier)) return false; trackOptimizationSuccess(); *emitted = true; return true; } bool IonBuilder::getPropTrySharedStub(bool* emitted, MDefinition* obj, TemporaryTypeSet* types) { MOZ_ASSERT(*emitted == false); // Try to emit a shared stub cache. if (JitOptions.disableSharedStubs) return true; MInstruction* stub = MUnarySharedStub::New(alloc(), obj); current->add(stub); current->push(stub); if (!resumeAfter(stub)) return false; // Due to inlining, it's possible the observed TypeSet is non-empty, // even though we know |obj| is null/undefined and the MCallGetProperty // will throw. Don't push a TypeBarrier in this case, to avoid // inlining the following (unreachable) JSOP_CALL. if (*pc != JSOP_CALLPROP || !IsNullOrUndefined(obj->type())) { if (!pushTypeBarrier(stub, types, BarrierKind::TypeSet)) return false; } *emitted = true; return true; } MDefinition* IonBuilder::tryInnerizeWindow(MDefinition* obj) { // Try to optimize accesses on outer window proxies (window.foo, for // example) to go directly to the inner window, the global. // // Callers should be careful not to pass the inner object to getters or // setters that require outerization. if (obj->type() != MIRType::Object) return obj; TemporaryTypeSet* types = obj->resultTypeSet(); if (!types) return obj; JSObject* singleton = types->maybeSingleton(); if (!singleton) return obj; if (!IsWindowProxy(singleton)) return obj; // This must be a WindowProxy for the current Window/global. Else it'd be // a cross-compartment wrapper and IsWindowProxy returns false for those. MOZ_ASSERT(ToWindowIfWindowProxy(singleton) == &script()->global()); // When we navigate, the WindowProxy is brain transplanted and we'll mark // its ObjectGroup as having unknown properties. The type constraint we add // here will invalidate JIT code when this happens. TypeSet::ObjectKey* key = TypeSet::ObjectKey::get(singleton); if (key->hasFlags(constraints(), OBJECT_FLAG_UNKNOWN_PROPERTIES)) return obj; obj->setImplicitlyUsedUnchecked(); return constant(ObjectValue(script()->global())); } bool IonBuilder::getPropTryInnerize(bool* emitted, MDefinition* obj, PropertyName* name, TemporaryTypeSet* types) { // See the comment in tryInnerizeWindow for how this works. // Note that it's important that we do this _before_ we'd try to // do the optimizations below on obj normally, since some of those // optimizations have fallback paths that are slower than the path // we'd produce here. MOZ_ASSERT(*emitted == false); MDefinition* inner = tryInnerizeWindow(obj); if (inner == obj) return true; if (!forceInlineCaches()) { trackOptimizationAttempt(TrackedStrategy::GetProp_Constant); if (!getPropTryConstant(emitted, inner, NameToId(name), types) || *emitted) return *emitted; trackOptimizationAttempt(TrackedStrategy::GetProp_StaticName); if (!getStaticName(&script()->global(), name, emitted) || *emitted) return *emitted; trackOptimizationAttempt(TrackedStrategy::GetProp_CommonGetter); if (!getPropTryCommonGetter(emitted, inner, name, types) || *emitted) return *emitted; } // Passing the inner object to GetProperty IC is safe, see the // needsOuterizedThisObject check in IsCacheableGetPropCallNative. BarrierKind barrier = PropertyReadNeedsTypeBarrier(analysisContext, constraints(), inner, name, types); trackOptimizationAttempt(TrackedStrategy::GetProp_InlineCache); if (!getPropTryCache(emitted, inner, name, barrier, types) || *emitted) return *emitted; MOZ_ASSERT(*emitted == false); return true; } bool IonBuilder::jsop_setprop(PropertyName* name) { MDefinition* value = current->pop(); MDefinition* obj = current->pop(); bool emitted = false; startTrackingOptimizations(); trackTypeInfo(TrackedTypeSite::Receiver, obj->type(), obj->resultTypeSet()); trackTypeInfo(TrackedTypeSite::Value, value->type(), value->resultTypeSet()); // Always use a call if we are doing the definite properties analysis and // not actually emitting code, to simplify later analysis. if (info().isAnalysis() || shouldAbortOnPreliminaryGroups(obj)) { bool strict = IsStrictSetPC(pc); MInstruction* ins = MCallSetProperty::New(alloc(), obj, value, name, strict); current->add(ins); current->push(value); return resumeAfter(ins); } if (!forceInlineCaches()) { // Try to inline a common property setter, or make a call. trackOptimizationAttempt(TrackedStrategy::SetProp_CommonSetter); if (!setPropTryCommonSetter(&emitted, obj, name, value) || emitted) return emitted; // Try to emit stores to known binary data blocks trackOptimizationAttempt(TrackedStrategy::SetProp_TypedObject); if (!setPropTryTypedObject(&emitted, obj, name, value) || emitted) return emitted; } TemporaryTypeSet* objTypes = obj->resultTypeSet(); bool barrier = PropertyWriteNeedsTypeBarrier(alloc(), constraints(), current, &obj, name, &value, /* canModify = */ true); // Add post barrier if needed. The instructions above manage any post // barriers they need directly. if (NeedsPostBarrier(value)) current->add(MPostWriteBarrier::New(alloc(), obj, value)); if (!forceInlineCaches()) { // Try to emit store from definite slots. trackOptimizationAttempt(TrackedStrategy::SetProp_DefiniteSlot); if (!setPropTryDefiniteSlot(&emitted, obj, name, value, barrier, objTypes) || emitted) return emitted; // Try to emit a monomorphic/polymorphic store based on baseline caches. trackOptimizationAttempt(TrackedStrategy::SetProp_InlineAccess); if (!setPropTryInlineAccess(&emitted, obj, name, value, barrier, objTypes) || emitted) return emitted; } // Emit a polymorphic cache. trackOptimizationAttempt(TrackedStrategy::SetProp_InlineCache); return setPropTryCache(&emitted, obj, name, value, barrier, objTypes); } bool IonBuilder::setPropTryCommonSetter(bool* emitted, MDefinition* obj, PropertyName* name, MDefinition* value) { MOZ_ASSERT(*emitted == false); Shape* lastProperty = nullptr; JSFunction* commonSetter = nullptr; JSObject* foundProto = nullptr; bool isOwnProperty; BaselineInspector::ReceiverVector receivers(alloc()); if (!inspector->commonSetPropFunction(pc, &foundProto, &lastProperty, &commonSetter, &isOwnProperty, receivers)) { trackOptimizationOutcome(TrackedOutcome::NoProtoFound); return true; } TemporaryTypeSet* objTypes = obj->resultTypeSet(); MDefinition* guard = nullptr; bool canUseTIForSetter = testCommonGetterSetter(objTypes, name, /* isGetter = */ false, foundProto, lastProperty, commonSetter, &guard); if (!canUseTIForSetter) { // If type information is bad, we can still optimize the setter if we // shape guard. obj = addShapeGuardsForGetterSetter(obj, foundProto, lastProperty, receivers, isOwnProperty); if (!obj) return false; } // Emit common setter. // Setters can be called even if the property write needs a type // barrier, as calling the setter does not actually write any data // properties. // Try emitting dom call. if (!setPropTryCommonDOMSetter(emitted, obj, value, commonSetter, objTypes)) return false; if (*emitted) { trackOptimizationOutcome(TrackedOutcome::DOM); return true; } // Don't call the setter with a primitive value. if (obj->type() != MIRType::Object) { MGuardObject* guardObj = MGuardObject::New(alloc(), obj); current->add(guardObj); obj = guardObj; } // Dummy up the stack, as in getprop. We are pushing an extra value, so // ensure there is enough space. if (!current->ensureHasSlots(3)) return false; current->push(constant(ObjectValue(*commonSetter))); current->push(obj); current->push(value); // Call the setter. Note that we have to push the original value, not // the setter's return value. CallInfo callInfo(alloc(), /* constructing = */ false, /* ignoresReturnValue = */ BytecodeIsPopped(pc)); if (!callInfo.init(current, 1)) return false; // Ensure that we know we are calling a setter in case we inline it. callInfo.markAsSetter(); // Inline the setter if we can. if (commonSetter->isInterpreted()) { InliningDecision decision = makeInliningDecision(commonSetter, callInfo); switch (decision) { case InliningDecision_Error: return false; case InliningDecision_DontInline: case InliningDecision_WarmUpCountTooLow: break; case InliningDecision_Inline: { InliningStatus status = inlineScriptedCall(callInfo, commonSetter); if (status == InliningStatus_Inlined) { *emitted = true; return true; } if (status == InliningStatus_Error) return false; } } } MCall* call = makeCallHelper(commonSetter, callInfo); if (!call) return false; current->push(value); if (!resumeAfter(call)) return false; // If the setter could have been inlined, don't track success. The call to // makeInliningDecision above would have tracked a specific reason why we // couldn't inline. if (!commonSetter->isInterpreted()) trackOptimizationSuccess(); *emitted = true; return true; } bool IonBuilder::setPropTryCommonDOMSetter(bool* emitted, MDefinition* obj, MDefinition* value, JSFunction* setter, TemporaryTypeSet* objTypes) { MOZ_ASSERT(*emitted == false); if (!objTypes || !objTypes->isDOMClass(constraints())) return true; if (!testShouldDOMCall(objTypes, setter, JSJitInfo::Setter)) return true; // Emit SetDOMProperty. MOZ_ASSERT(setter->jitInfo()->type() == JSJitInfo::Setter); MSetDOMProperty* set = MSetDOMProperty::New(alloc(), setter->jitInfo()->setter, obj, value); current->add(set); current->push(value); if (!resumeAfter(set)) return false; *emitted = true; return true; } bool IonBuilder::setPropTryTypedObject(bool* emitted, MDefinition* obj, PropertyName* name, MDefinition* value) { TypedObjectPrediction fieldPrediction; size_t fieldOffset; size_t fieldIndex; if (!typedObjectHasField(obj, name, &fieldOffset, &fieldPrediction, &fieldIndex)) return true; switch (fieldPrediction.kind()) { case type::Simd: // FIXME (bug 894104): store into a MIRType::float32x4 etc return true; case type::Reference: return setPropTryReferencePropOfTypedObject(emitted, obj, fieldOffset, value, fieldPrediction, name); case type::Scalar: return setPropTryScalarPropOfTypedObject(emitted, obj, fieldOffset, value, fieldPrediction); case type::Struct: case type::Array: return true; } MOZ_CRASH("Unknown kind"); } bool IonBuilder::setPropTryReferencePropOfTypedObject(bool* emitted, MDefinition* obj, int32_t fieldOffset, MDefinition* value, TypedObjectPrediction fieldPrediction, PropertyName* name) { ReferenceTypeDescr::Type fieldType = fieldPrediction.referenceType(); TypeSet::ObjectKey* globalKey = TypeSet::ObjectKey::get(&script()->global()); if (globalKey->hasFlags(constraints(), OBJECT_FLAG_TYPED_OBJECT_HAS_DETACHED_BUFFER)) return true; LinearSum byteOffset(alloc()); if (!byteOffset.add(fieldOffset)) setForceAbort(); if (!storeReferenceTypedObjectValue(obj, byteOffset, fieldType, value, name)) return true; current->push(value); trackOptimizationSuccess(); *emitted = true; return true; } bool IonBuilder::setPropTryScalarPropOfTypedObject(bool* emitted, MDefinition* obj, int32_t fieldOffset, MDefinition* value, TypedObjectPrediction fieldPrediction) { // Must always be loading the same scalar type Scalar::Type fieldType = fieldPrediction.scalarType(); // Don't optimize if the typed object's underlying buffer may be detached. TypeSet::ObjectKey* globalKey = TypeSet::ObjectKey::get(&script()->global()); if (globalKey->hasFlags(constraints(), OBJECT_FLAG_TYPED_OBJECT_HAS_DETACHED_BUFFER)) return true; LinearSum byteOffset(alloc()); if (!byteOffset.add(fieldOffset)) setForceAbort(); if (!storeScalarTypedObjectValue(obj, byteOffset, fieldType, value)) return false; current->push(value); trackOptimizationSuccess(); *emitted = true; return true; } bool IonBuilder::setPropTryDefiniteSlot(bool* emitted, MDefinition* obj, PropertyName* name, MDefinition* value, bool barrier, TemporaryTypeSet* objTypes) { MOZ_ASSERT(*emitted == false); if (barrier) { trackOptimizationOutcome(TrackedOutcome::NeedsTypeBarrier); return true; } uint32_t nfixed; uint32_t slot = getDefiniteSlot(obj->resultTypeSet(), name, &nfixed); if (slot == UINT32_MAX) return true; bool writeBarrier = false; for (size_t i = 0; i < obj->resultTypeSet()->getObjectCount(); i++) { TypeSet::ObjectKey* key = obj->resultTypeSet()->getObject(i); if (!key) continue; HeapTypeSetKey property = key->property(NameToId(name)); if (property.nonWritable(constraints())) { trackOptimizationOutcome(TrackedOutcome::NonWritableProperty); return true; } writeBarrier |= property.needsBarrier(constraints()); } MInstruction* store; if (slot < nfixed) { store = MStoreFixedSlot::New(alloc(), obj, slot, value); if (writeBarrier) store->toStoreFixedSlot()->setNeedsBarrier(); } else { MInstruction* slots = MSlots::New(alloc(), obj); current->add(slots); store = MStoreSlot::New(alloc(), slots, slot - nfixed, value); if (writeBarrier) store->toStoreSlot()->setNeedsBarrier(); } current->add(store); current->push(value); if (!resumeAfter(store)) return false; trackOptimizationSuccess(); *emitted = true; return true; } bool IonBuilder::setPropTryInlineAccess(bool* emitted, MDefinition* obj, PropertyName* name, MDefinition* value, bool barrier, TemporaryTypeSet* objTypes) { MOZ_ASSERT(*emitted == false); if (barrier) { trackOptimizationOutcome(TrackedOutcome::NeedsTypeBarrier); return true; } BaselineInspector::ReceiverVector receivers(alloc()); if (!inspector->maybeInfoForPropertyOp(pc, receivers)) return false; if (!canInlinePropertyOpShapes(receivers)) return true; if (receivers.length() == 1) { if (!receivers[0].group) { // Monomorphic store to a native object. spew("Inlining monomorphic native SETPROP"); obj = addShapeGuard(obj, receivers[0].shape, Bailout_ShapeGuard); Shape* shape = receivers[0].shape->searchLinear(NameToId(name)); MOZ_ASSERT(shape); bool needsBarrier = objTypes->propertyNeedsBarrier(constraints(), NameToId(name)); if (!storeSlot(obj, shape, value, needsBarrier)) return false; trackOptimizationOutcome(TrackedOutcome::Monomorphic); *emitted = true; return true; } return true; } MOZ_ASSERT(receivers.length() > 1); spew("Inlining polymorphic SETPROP"); if (Shape* propShape = PropertyShapesHaveSameSlot(receivers, NameToId(name))) { obj = addGuardReceiverPolymorphic(obj, receivers); if (!obj) return false; bool needsBarrier = objTypes->propertyNeedsBarrier(constraints(), NameToId(name)); if (!storeSlot(obj, propShape, value, needsBarrier)) return false; trackOptimizationOutcome(TrackedOutcome::Polymorphic); *emitted = true; return true; } MSetPropertyPolymorphic* ins = MSetPropertyPolymorphic::New(alloc(), obj, value, name); current->add(ins); current->push(value); for (size_t i = 0; i < receivers.length(); i++) { Shape* propShape = nullptr; if (receivers[i].shape) { propShape = receivers[i].shape->searchLinear(NameToId(name)); MOZ_ASSERT(propShape); } if (!ins->addReceiver(receivers[i], propShape)) return false; } if (objTypes->propertyNeedsBarrier(constraints(), NameToId(name))) ins->setNeedsBarrier(); if (!resumeAfter(ins)) return false; trackOptimizationOutcome(TrackedOutcome::Polymorphic); *emitted = true; return true; } bool IonBuilder::setPropTryCache(bool* emitted, MDefinition* obj, PropertyName* name, MDefinition* value, bool barrier, TemporaryTypeSet* objTypes) { MOZ_ASSERT(*emitted == false); bool strict = IsStrictSetPC(pc); MConstant* id = constant(StringValue(name)); MSetPropertyCache* ins = MSetPropertyCache::New(alloc(), obj, id, value, strict, barrier, /* guardHoles = */ false); current->add(ins); current->push(value); if (!resumeAfter(ins)) return false; trackOptimizationSuccess(); *emitted = true; return true; } bool IonBuilder::jsop_delprop(PropertyName* name) { MDefinition* obj = current->pop(); bool strict = JSOp(*pc) == JSOP_STRICTDELPROP; MInstruction* ins = MDeleteProperty::New(alloc(), obj, name, strict); current->add(ins); current->push(ins); return resumeAfter(ins); } bool IonBuilder::jsop_delelem() { MDefinition* index = current->pop(); MDefinition* obj = current->pop(); bool strict = JSOp(*pc) == JSOP_STRICTDELELEM; MDeleteElement* ins = MDeleteElement::New(alloc(), obj, index, strict); current->add(ins); current->push(ins); return resumeAfter(ins); } bool IonBuilder::jsop_regexp(RegExpObject* reobj) { MRegExp* regexp = MRegExp::New(alloc(), constraints(), reobj); current->add(regexp); current->push(regexp); return true; } bool IonBuilder::jsop_object(JSObject* obj) { if (options.cloneSingletons()) { MCloneLiteral* clone = MCloneLiteral::New(alloc(), constant(ObjectValue(*obj))); current->add(clone); current->push(clone); return resumeAfter(clone); } compartment->setSingletonsAsValues(); pushConstant(ObjectValue(*obj)); return true; } bool IonBuilder::jsop_lambda(JSFunction* fun) { MOZ_ASSERT(analysis().usesEnvironmentChain()); MOZ_ASSERT(!fun->isArrow()); if (IsAsmJSModule(fun)) return abort("asm.js module function"); MConstant* cst = MConstant::NewConstraintlessObject(alloc(), fun); current->add(cst); MLambda* ins = MLambda::New(alloc(), constraints(), current->environmentChain(), cst); current->add(ins); current->push(ins); return resumeAfter(ins); } bool IonBuilder::jsop_lambda_arrow(JSFunction* fun) { MOZ_ASSERT(analysis().usesEnvironmentChain()); MOZ_ASSERT(fun->isArrow()); MOZ_ASSERT(!fun->isNative()); MDefinition* newTargetDef = current->pop(); MLambdaArrow* ins = MLambdaArrow::New(alloc(), constraints(), current->environmentChain(), newTargetDef, fun); current->add(ins); current->push(ins); return resumeAfter(ins); } bool IonBuilder::jsop_setfunname(uint8_t prefixKind) { MDefinition* name = current->pop(); MDefinition* fun = current->pop(); MOZ_ASSERT(fun->type() == MIRType::Object); MSetFunName* ins = MSetFunName::New(alloc(), fun, name, prefixKind); current->add(ins); current->push(fun); return resumeAfter(ins); } bool IonBuilder::jsop_setarg(uint32_t arg) { // To handle this case, we should spill the arguments to the space where // actual arguments are stored. The tricky part is that if we add a MIR // to wrap the spilling action, we don't want the spilling to be // captured by the GETARG and by the resume point, only by // MGetFrameArgument. MOZ_ASSERT(analysis_.hasSetArg()); MDefinition* val = current->peek(-1); // If an arguments object is in use, and it aliases formals, then all SETARGs // must go through the arguments object. if (info().argsObjAliasesFormals()) { if (NeedsPostBarrier(val)) current->add(MPostWriteBarrier::New(alloc(), current->argumentsObject(), val)); current->add(MSetArgumentsObjectArg::New(alloc(), current->argumentsObject(), GET_ARGNO(pc), val)); return true; } // :TODO: if hasArguments() is true, and the script has a JSOP_SETARG, then // convert all arg accesses to go through the arguments object. (see Bug 957475) if (info().hasArguments()) return abort("NYI: arguments & setarg."); // Otherwise, if a magic arguments is in use, and it aliases formals, and there exist // arguments[...] GETELEM expressions in the script, then SetFrameArgument must be used. // If no arguments[...] GETELEM expressions are in the script, and an argsobj is not // required, then it means that any aliased argument set can never be observed, and // the frame does not actually need to be updated with the new arg value. if (info().argumentsAliasesFormals()) { // JSOP_SETARG with magic arguments within inline frames is not yet supported. MOZ_ASSERT(script()->uninlineable() && !isInlineBuilder()); MSetFrameArgument* store = MSetFrameArgument::New(alloc(), arg, val); modifiesFrameArguments_ = true; current->add(store); current->setArg(arg); return true; } // If this assignment is at the start of the function and is coercing // the original value for the argument which was passed in, loosen // the type information for that original argument if it is currently // empty due to originally executing in the interpreter. if (graph().numBlocks() == 1 && (val->isBitOr() || val->isBitAnd() || val->isMul() /* for JSOP_POS */)) { for (size_t i = 0; i < val->numOperands(); i++) { MDefinition* op = val->getOperand(i); if (op->isParameter() && op->toParameter()->index() == (int32_t)arg && op->resultTypeSet() && op->resultTypeSet()->empty()) { bool otherUses = false; for (MUseDefIterator iter(op); iter; iter++) { MDefinition* def = iter.def(); if (def == val) continue; otherUses = true; } if (!otherUses) { MOZ_ASSERT(op->resultTypeSet() == &argTypes[arg]); argTypes[arg].addType(TypeSet::UnknownType(), alloc_->lifoAlloc()); if (val->isMul()) { val->setResultType(MIRType::Double); val->toMul()->setSpecialization(MIRType::Double); } else { MOZ_ASSERT(val->type() == MIRType::Int32); } val->setResultTypeSet(nullptr); } } } } current->setArg(arg); return true; } bool IonBuilder::jsop_defvar(uint32_t index) { MOZ_ASSERT(JSOp(*pc) == JSOP_DEFVAR); PropertyName* name = script()->getName(index); // Bake in attrs. unsigned attrs = JSPROP_ENUMERATE | JSPROP_PERMANENT; MOZ_ASSERT(!script()->isForEval()); // Pass the EnvironmentChain. MOZ_ASSERT(analysis().usesEnvironmentChain()); // Bake the name pointer into the MDefVar. MDefVar* defvar = MDefVar::New(alloc(), name, attrs, current->environmentChain()); current->add(defvar); return resumeAfter(defvar); } bool IonBuilder::jsop_deflexical(uint32_t index) { MOZ_ASSERT(!script()->hasNonSyntacticScope()); MOZ_ASSERT(JSOp(*pc) == JSOP_DEFLET || JSOp(*pc) == JSOP_DEFCONST); PropertyName* name = script()->getName(index); unsigned attrs = JSPROP_ENUMERATE | JSPROP_PERMANENT; if (JSOp(*pc) == JSOP_DEFCONST) attrs |= JSPROP_READONLY; MDefLexical* deflex = MDefLexical::New(alloc(), name, attrs); current->add(deflex); return resumeAfter(deflex); } bool IonBuilder::jsop_deffun(uint32_t index) { MOZ_ASSERT(analysis().usesEnvironmentChain()); MDefFun* deffun = MDefFun::New(alloc(), current->pop(), current->environmentChain()); current->add(deffun); return resumeAfter(deffun); } bool IonBuilder::jsop_throwsetconst() { current->peek(-1)->setImplicitlyUsedUnchecked(); MInstruction* lexicalError = MThrowRuntimeLexicalError::New(alloc(), JSMSG_BAD_CONST_ASSIGN); current->add(lexicalError); return resumeAfter(lexicalError); } bool IonBuilder::jsop_checklexical() { uint32_t slot = info().localSlot(GET_LOCALNO(pc)); MDefinition* lexical = addLexicalCheck(current->getSlot(slot)); if (!lexical) return false; current->setSlot(slot, lexical); return true; } bool IonBuilder::jsop_checkaliasedlexical(EnvironmentCoordinate ec) { MDefinition* let = addLexicalCheck(getAliasedVar(ec)); if (!let) return false; jsbytecode* nextPc = pc + JSOP_CHECKALIASEDLEXICAL_LENGTH; MOZ_ASSERT(JSOp(*nextPc) == JSOP_GETALIASEDVAR || JSOp(*nextPc) == JSOP_SETALIASEDVAR || JSOp(*nextPc) == JSOP_THROWSETALIASEDCONST); MOZ_ASSERT(ec == EnvironmentCoordinate(nextPc)); // If we are checking for a load, push the checked let so that the load // can use it. if (JSOp(*nextPc) == JSOP_GETALIASEDVAR) setLexicalCheck(let); return true; } bool IonBuilder::jsop_functionthis() { MOZ_ASSERT(info().funMaybeLazy()); MOZ_ASSERT(!info().funMaybeLazy()->isArrow()); if (script()->strict() || info().funMaybeLazy()->isSelfHostedBuiltin()) { // No need to wrap primitive |this| in strict mode or self-hosted code. current->pushSlot(info().thisSlot()); return true; } if (thisTypes && (thisTypes->getKnownMIRType() == MIRType::Object || (thisTypes->empty() && baselineFrame_ && baselineFrame_->thisType.isSomeObject()))) { // This is safe, because if the entry type of |this| is an object, it // will necessarily be an object throughout the entire function. OSR // can introduce a phi, but this phi will be specialized. current->pushSlot(info().thisSlot()); return true; } // If we are doing an analysis, we might not yet know the type of |this|. // Instead of bailing out just push the |this| slot, as this code won't // actually execute and it does not matter whether |this| is primitive. if (info().isAnalysis()) { current->pushSlot(info().thisSlot()); return true; } // Hard case: |this| may be a primitive we have to wrap. MDefinition* def = current->getSlot(info().thisSlot()); if (def->type() == MIRType::Object) { current->push(def); return true; } if (IsNullOrUndefined(def->type())) { pushConstant(GetThisValue(&script()->global())); return true; } MComputeThis* thisObj = MComputeThis::New(alloc(), def); current->add(thisObj); current->push(thisObj); return resumeAfter(thisObj); } bool IonBuilder::jsop_globalthis() { if (script()->hasNonSyntacticScope()) { // Ion does not compile global scripts with a non-syntactic scope, but // we can end up here when we're compiling an arrow function. return abort("JSOP_GLOBALTHIS in script with non-syntactic scope"); } LexicalEnvironmentObject* globalLexical = &script()->global().lexicalEnvironment(); pushConstant(globalLexical->thisValue()); return true; } bool IonBuilder::jsop_typeof() { MDefinition* input = current->pop(); MTypeOf* ins = MTypeOf::New(alloc(), input, input->type()); ins->cacheInputMaybeCallableOrEmulatesUndefined(constraints()); current->add(ins); current->push(ins); return true; } bool IonBuilder::jsop_toasync() { MDefinition* unwrapped = current->pop(); MOZ_ASSERT(unwrapped->type() == MIRType::Object); MToAsync* ins = MToAsync::New(alloc(), unwrapped); current->add(ins); current->push(ins); return resumeAfter(ins); } bool IonBuilder::jsop_toid() { // No-op if the index is an integer. if (current->peek(-1)->type() == MIRType::Int32) return true; MDefinition* index = current->pop(); MToId* ins = MToId::New(alloc(), index); current->add(ins); current->push(ins); return resumeAfter(ins); } bool IonBuilder::jsop_iter(uint8_t flags) { if (flags != JSITER_ENUMERATE) nonStringIteration_ = true; MDefinition* obj = current->pop(); MInstruction* ins = MIteratorStart::New(alloc(), obj, flags); if (!outermostBuilder()->iterators_.append(ins)) return false; current->add(ins); current->push(ins); return resumeAfter(ins); } bool IonBuilder::jsop_itermore() { MDefinition* iter = current->peek(-1); MInstruction* ins = MIteratorMore::New(alloc(), iter); current->add(ins); current->push(ins); return resumeAfter(ins); } bool IonBuilder::jsop_isnoiter() { MDefinition* def = current->peek(-1); MOZ_ASSERT(def->isIteratorMore()); MInstruction* ins = MIsNoIter::New(alloc(), def); current->add(ins); current->push(ins); return true; } bool IonBuilder::jsop_iterend() { MDefinition* iter = current->pop(); MInstruction* ins = MIteratorEnd::New(alloc(), iter); current->add(ins); return resumeAfter(ins); } MDefinition* IonBuilder::walkEnvironmentChain(unsigned hops) { MDefinition* env = current->getSlot(info().environmentChainSlot()); for (unsigned i = 0; i < hops; i++) { MInstruction* ins = MEnclosingEnvironment::New(alloc(), env); current->add(ins); env = ins; } return env; } bool IonBuilder::hasStaticEnvironmentObject(EnvironmentCoordinate ec, JSObject** pcall) { JSScript* outerScript = EnvironmentCoordinateFunctionScript(script(), pc); if (!outerScript || !outerScript->treatAsRunOnce()) return false; TypeSet::ObjectKey* funKey = TypeSet::ObjectKey::get(outerScript->functionNonDelazifying()); if (funKey->hasFlags(constraints(), OBJECT_FLAG_RUNONCE_INVALIDATED)) return false; // The script this aliased var operation is accessing will run only once, // so there will be only one call object and the aliased var access can be // compiled in the same manner as a global access. We still need to find // the call object though. // Look for the call object on the current script's function's env chain. // If the current script is inner to the outer script and the function has // singleton type then it should show up here. MDefinition* envDef = current->getSlot(info().environmentChainSlot()); envDef->setImplicitlyUsedUnchecked(); JSObject* environment = script()->functionNonDelazifying()->environment(); while (environment && !environment->is()) { if (environment->is() && environment->as().callee().nonLazyScript() == outerScript) { MOZ_ASSERT(environment->isSingleton()); *pcall = environment; return true; } environment = environment->enclosingEnvironment(); } // Look for the call object on the current frame, if we are compiling the // outer script itself. Don't do this if we are at entry to the outer // script, as the call object we see will not be the real one --- after // entering the Ion code a different call object will be created. if (script() == outerScript && baselineFrame_ && info().osrPc()) { JSObject* singletonScope = baselineFrame_->singletonEnvChain; if (singletonScope && singletonScope->is() && singletonScope->as().callee().nonLazyScript() == outerScript) { MOZ_ASSERT(singletonScope->isSingleton()); *pcall = singletonScope; return true; } } return true; } MDefinition* IonBuilder::getAliasedVar(EnvironmentCoordinate ec) { MDefinition* obj = walkEnvironmentChain(ec.hops()); Shape* shape = EnvironmentCoordinateToEnvironmentShape(script(), pc); MInstruction* load; if (shape->numFixedSlots() <= ec.slot()) { MInstruction* slots = MSlots::New(alloc(), obj); current->add(slots); load = MLoadSlot::New(alloc(), slots, ec.slot() - shape->numFixedSlots()); } else { load = MLoadFixedSlot::New(alloc(), obj, ec.slot()); } current->add(load); return load; } bool IonBuilder::jsop_getaliasedvar(EnvironmentCoordinate ec) { JSObject* call = nullptr; if (hasStaticEnvironmentObject(ec, &call) && call) { PropertyName* name = EnvironmentCoordinateName(envCoordinateNameCache, script(), pc); bool emitted = false; if (!getStaticName(call, name, &emitted, takeLexicalCheck()) || emitted) return emitted; } // See jsop_checkaliasedlexical. MDefinition* load = takeLexicalCheck(); if (!load) load = getAliasedVar(ec); current->push(load); TemporaryTypeSet* types = bytecodeTypes(pc); return pushTypeBarrier(load, types, BarrierKind::TypeSet); } bool IonBuilder::jsop_setaliasedvar(EnvironmentCoordinate ec) { JSObject* call = nullptr; if (hasStaticEnvironmentObject(ec, &call)) { uint32_t depth = current->stackDepth() + 1; if (depth > current->nslots()) { if (!current->increaseSlots(depth - current->nslots())) return false; } MDefinition* value = current->pop(); PropertyName* name = EnvironmentCoordinateName(envCoordinateNameCache, script(), pc); if (call) { // Push the object on the stack to match the bound object expected in // the global and property set cases. pushConstant(ObjectValue(*call)); current->push(value); return setStaticName(call, name); } // The call object has type information we need to respect but we // couldn't find it. Just do a normal property assign. MDefinition* obj = walkEnvironmentChain(ec.hops()); current->push(obj); current->push(value); return jsop_setprop(name); } MDefinition* rval = current->peek(-1); MDefinition* obj = walkEnvironmentChain(ec.hops()); Shape* shape = EnvironmentCoordinateToEnvironmentShape(script(), pc); if (NeedsPostBarrier(rval)) current->add(MPostWriteBarrier::New(alloc(), obj, rval)); MInstruction* store; if (shape->numFixedSlots() <= ec.slot()) { MInstruction* slots = MSlots::New(alloc(), obj); current->add(slots); store = MStoreSlot::NewBarriered(alloc(), slots, ec.slot() - shape->numFixedSlots(), rval); } else { store = MStoreFixedSlot::NewBarriered(alloc(), obj, ec.slot(), rval); } current->add(store); return resumeAfter(store); } bool IonBuilder::jsop_in() { MDefinition* obj = current->pop(); MDefinition* id = current->pop(); bool emitted = false; if (!inTryDense(&emitted, obj, id) || emitted) return emitted; if (!inTryFold(&emitted, obj, id) || emitted) return emitted; MIn* ins = MIn::New(alloc(), id, obj); current->add(ins); current->push(ins); return resumeAfter(ins); } bool IonBuilder::inTryDense(bool* emitted, MDefinition* obj, MDefinition* id) { MOZ_ASSERT(!*emitted); if (shouldAbortOnPreliminaryGroups(obj)) return true; if (!ElementAccessIsDenseNative(constraints(), obj, id)) return true; if (ElementAccessHasExtraIndexedProperty(this, obj)) return true; *emitted = true; bool needsHoleCheck = !ElementAccessIsPacked(constraints(), obj); // Ensure id is an integer. MInstruction* idInt32 = MToInt32::New(alloc(), id); current->add(idInt32); id = idInt32; // Get the elements vector. MElements* elements = MElements::New(alloc(), obj); current->add(elements); MInstruction* initLength = initializedLength(obj, elements); // If there are no holes, speculate the InArray check will not fail. if (!needsHoleCheck && !failedBoundsCheck_) { addBoundsCheck(idInt32, initLength); pushConstant(BooleanValue(true)); return true; } // Check if id < initLength and elem[id] not a hole. MInArray* ins = MInArray::New(alloc(), elements, id, initLength, obj, needsHoleCheck); current->add(ins); current->push(ins); return true; } bool IonBuilder::inTryFold(bool* emitted, MDefinition* obj, MDefinition* id) { // Fold |id in obj| to |false|, if we know the object (or an object on its // prototype chain) does not have this property. MOZ_ASSERT(!*emitted); MConstant* idConst = id->maybeConstantValue(); jsid propId; if (!idConst || !ValueToIdPure(idConst->toJSValue(), &propId)) return true; if (propId != IdToTypeId(propId)) return true; ResultWithOOM res = testNotDefinedProperty(obj, propId); if (res.oom) return false; if (!res.value) return true; *emitted = true; pushConstant(BooleanValue(false)); obj->setImplicitlyUsedUnchecked(); id->setImplicitlyUsedUnchecked(); return true; } bool IonBuilder::hasOnProtoChain(TypeSet::ObjectKey* key, JSObject* protoObject, bool* hasOnProto) { MOZ_ASSERT(protoObject); while (true) { if (!key->hasStableClassAndProto(constraints()) || !key->clasp()->isNative()) return false; JSObject* proto = checkNurseryObject(key->proto().toObjectOrNull()); if (!proto) { *hasOnProto = false; return true; } if (proto == protoObject) { *hasOnProto = true; return true; } key = TypeSet::ObjectKey::get(proto); } MOZ_CRASH("Unreachable"); } bool IonBuilder::tryFoldInstanceOf(MDefinition* lhs, JSObject* protoObject) { // Try to fold the js::IsDelegate part of the instanceof operation. if (!lhs->mightBeType(MIRType::Object)) { // If the lhs is a primitive, the result is false. lhs->setImplicitlyUsedUnchecked(); pushConstant(BooleanValue(false)); return true; } TemporaryTypeSet* lhsTypes = lhs->resultTypeSet(); if (!lhsTypes || lhsTypes->unknownObject()) return false; // We can fold if either all objects have protoObject on their proto chain // or none have. bool isFirst = true; bool knownIsInstance = false; for (unsigned i = 0; i < lhsTypes->getObjectCount(); i++) { TypeSet::ObjectKey* key = lhsTypes->getObject(i); if (!key) continue; bool isInstance; if (!hasOnProtoChain(key, protoObject, &isInstance)) return false; if (isFirst) { knownIsInstance = isInstance; isFirst = false; } else if (knownIsInstance != isInstance) { // Some of the objects have protoObject on their proto chain and // others don't, so we can't optimize this. return false; } } if (knownIsInstance && lhsTypes->getKnownMIRType() != MIRType::Object) { // The result is true for all objects, but the lhs might be a primitive. // We can't fold this completely but we can use a much faster IsObject // test. MIsObject* isObject = MIsObject::New(alloc(), lhs); current->add(isObject); current->push(isObject); return true; } lhs->setImplicitlyUsedUnchecked(); pushConstant(BooleanValue(knownIsInstance)); return true; } bool IonBuilder::jsop_instanceof() { MDefinition* rhs = current->pop(); MDefinition* obj = current->pop(); // If this is an 'x instanceof function' operation and we can determine the // exact function and prototype object being tested for, use a typed path. do { TemporaryTypeSet* rhsTypes = rhs->resultTypeSet(); JSObject* rhsObject = rhsTypes ? rhsTypes->maybeSingleton() : nullptr; if (!rhsObject || !rhsObject->is() || rhsObject->isBoundFunction()) break; // Refuse to optimize anything whose [[Prototype]] isn't Function.prototype // since we can't guarantee that it uses the default @@hasInstance method. if (rhsObject->hasUncacheableProto() || !rhsObject->hasStaticPrototype()) break; Value funProto = script()->global().getPrototype(JSProto_Function); if (!funProto.isObject() || rhsObject->staticPrototype() != &funProto.toObject()) break; // If the user has supplied their own @@hasInstance method we shouldn't // clobber it. JSFunction* fun = &rhsObject->as(); const WellKnownSymbols* symbols = &compartment->runtime()->wellKnownSymbols(); if (!js::FunctionHasDefaultHasInstance(fun, *symbols)) break; // Ensure that we will bail if the @@hasInstance property or [[Prototype]] // change. TypeSet::ObjectKey* rhsKey = TypeSet::ObjectKey::get(rhsObject); if (!rhsKey->hasStableClassAndProto(constraints())) break; if (rhsKey->unknownProperties()) break; HeapTypeSetKey hasInstanceObject = rhsKey->property(SYMBOL_TO_JSID(symbols->hasInstance)); if (hasInstanceObject.isOwnProperty(constraints())) break; HeapTypeSetKey protoProperty = rhsKey->property(NameToId(names().prototype)); JSObject* protoObject = protoProperty.singleton(constraints()); if (!protoObject) break; rhs->setImplicitlyUsedUnchecked(); if (tryFoldInstanceOf(obj, protoObject)) return true; MInstanceOf* ins = MInstanceOf::New(alloc(), obj, protoObject); current->add(ins); current->push(ins); return resumeAfter(ins); } while (false); // Try to inline a fast path based on Baseline ICs. do { Shape* shape; uint32_t slot; JSObject* protoObject; if (!inspector->instanceOfData(pc, &shape, &slot, &protoObject)) break; // Shape guard. rhs = addShapeGuard(rhs, shape, Bailout_ShapeGuard); // Guard .prototype == protoObject. MOZ_ASSERT(shape->numFixedSlots() == 0, "Must be a dynamic slot"); MSlots* slots = MSlots::New(alloc(), rhs); current->add(slots); MLoadSlot* prototype = MLoadSlot::New(alloc(), slots, slot); current->add(prototype); MConstant* protoConst = MConstant::NewConstraintlessObject(alloc(), protoObject); current->add(protoConst); MGuardObjectIdentity* guard = MGuardObjectIdentity::New(alloc(), prototype, protoConst, /* bailOnEquality = */ false); current->add(guard); if (tryFoldInstanceOf(obj, protoObject)) return true; MInstanceOf* ins = MInstanceOf::New(alloc(), obj, protoObject); current->add(ins); current->push(ins); return resumeAfter(ins); } while (false); MCallInstanceOf* ins = MCallInstanceOf::New(alloc(), obj, rhs); current->add(ins); current->push(ins); return resumeAfter(ins); } bool IonBuilder::jsop_debugger() { MDebugger* debugger = MDebugger::New(alloc()); current->add(debugger); // The |debugger;| statement will always bail out to baseline if // cx->compartment()->isDebuggee(). Resume in-place and have baseline // handle the details. return resumeAt(debugger, pc); } MInstruction* IonBuilder::addConvertElementsToDoubles(MDefinition* elements) { MInstruction* convert = MConvertElementsToDoubles::New(alloc(), elements); current->add(convert); return convert; } MDefinition* IonBuilder::addMaybeCopyElementsForWrite(MDefinition* object, bool checkNative) { if (!ElementAccessMightBeCopyOnWrite(constraints(), object)) return object; MInstruction* copy = MMaybeCopyElementsForWrite::New(alloc(), object, checkNative); current->add(copy); return copy; } MInstruction* IonBuilder::addBoundsCheck(MDefinition* index, MDefinition* length) { MInstruction* check = MBoundsCheck::New(alloc(), index, length); current->add(check); // If a bounds check failed in the past, don't optimize bounds checks. if (failedBoundsCheck_) check->setNotMovable(); return check; } MInstruction* IonBuilder::addShapeGuard(MDefinition* obj, Shape* const shape, BailoutKind bailoutKind) { MGuardShape* guard = MGuardShape::New(alloc(), obj, shape, bailoutKind); current->add(guard); // If a shape guard failed in the past, don't optimize shape guard. if (failedShapeGuard_) guard->setNotMovable(); return guard; } MInstruction* IonBuilder::addGroupGuard(MDefinition* obj, ObjectGroup* group, BailoutKind bailoutKind) { MGuardObjectGroup* guard = MGuardObjectGroup::New(alloc(), obj, group, /* bailOnEquality = */ false, bailoutKind); current->add(guard); // If a shape guard failed in the past, don't optimize group guards. if (failedShapeGuard_) guard->setNotMovable(); LifoAlloc* lifoAlloc = alloc().lifoAlloc(); guard->setResultTypeSet(lifoAlloc->new_(lifoAlloc, TypeSet::ObjectType(group))); return guard; } MInstruction* IonBuilder::addGuardReceiverPolymorphic(MDefinition* obj, const BaselineInspector::ReceiverVector& receivers) { if (receivers.length() == 1) { if (!receivers[0].group) { // Monomorphic guard on a native object. return addShapeGuard(obj, receivers[0].shape, Bailout_ShapeGuard); } } MGuardReceiverPolymorphic* guard = MGuardReceiverPolymorphic::New(alloc(), obj); current->add(guard); if (failedShapeGuard_) guard->setNotMovable(); for (size_t i = 0; i < receivers.length(); i++) { if (!guard->addReceiver(receivers[i])) return nullptr; } return guard; } MInstruction* IonBuilder::addSharedTypedArrayGuard(MDefinition* obj) { MGuardSharedTypedArray* guard = MGuardSharedTypedArray::New(alloc(), obj); current->add(guard); return guard; } TemporaryTypeSet* IonBuilder::bytecodeTypes(jsbytecode* pc) { return TypeScript::BytecodeTypes(script(), pc, bytecodeTypeMap, &typeArrayHint, typeArray); } TypedObjectPrediction IonBuilder::typedObjectPrediction(MDefinition* typedObj) { // Extract TypedObjectPrediction directly if we can if (typedObj->isNewDerivedTypedObject()) { return typedObj->toNewDerivedTypedObject()->prediction(); } TemporaryTypeSet* types = typedObj->resultTypeSet(); return typedObjectPrediction(types); } TypedObjectPrediction IonBuilder::typedObjectPrediction(TemporaryTypeSet* types) { // Type set must be known to be an object. if (!types || types->getKnownMIRType() != MIRType::Object) return TypedObjectPrediction(); // And only known objects. if (types->unknownObject()) return TypedObjectPrediction(); TypedObjectPrediction out; for (uint32_t i = 0; i < types->getObjectCount(); i++) { ObjectGroup* group = types->getGroup(i); if (!group || !TypeSet::ObjectKey::get(group)->hasStableClassAndProto(constraints())) return TypedObjectPrediction(); if (!IsTypedObjectClass(group->clasp())) return TypedObjectPrediction(); out.addDescr(group->typeDescr()); } return out; } MDefinition* IonBuilder::loadTypedObjectType(MDefinition* typedObj) { // Shortcircuit derived type objects, meaning the intermediate // objects created to represent `a.b` in an expression like // `a.b.c`. In that case, the type object can be simply pulled // from the operands of that instruction. if (typedObj->isNewDerivedTypedObject()) return typedObj->toNewDerivedTypedObject()->type(); MInstruction* descr = MTypedObjectDescr::New(alloc(), typedObj); current->add(descr); return descr; } // Given a typed object `typedObj` and an offset `offset` into that // object's data, returns another typed object and adusted offset // where the data can be found. Often, these returned values are the // same as the inputs, but in cases where intermediate derived type // objects have been created, the return values will remove // intermediate layers (often rendering those derived type objects // into dead code). void IonBuilder::loadTypedObjectData(MDefinition* typedObj, MDefinition** owner, LinearSum* ownerOffset) { MOZ_ASSERT(typedObj->type() == MIRType::Object); // Shortcircuit derived type objects, meaning the intermediate // objects created to represent `a.b` in an expression like // `a.b.c`. In that case, the owned and a base offset can be // pulled from the operands of the instruction and combined with // `offset`. if (typedObj->isNewDerivedTypedObject()) { MNewDerivedTypedObject* ins = typedObj->toNewDerivedTypedObject(); SimpleLinearSum base = ExtractLinearSum(ins->offset()); if (!ownerOffset->add(base)) setForceAbort(); *owner = ins->owner(); return; } *owner = typedObj; } // Takes as input a typed object, an offset into that typed object's // memory, and the type repr of the data found at that offset. Returns // the elements pointer and a scaled offset. The scaled offset is // expressed in units of `unit`; when working with typed array MIR, // this is typically the alignment. void IonBuilder::loadTypedObjectElements(MDefinition* typedObj, const LinearSum& baseByteOffset, uint32_t scale, MDefinition** ownerElements, MDefinition** ownerScaledOffset, int32_t* ownerByteAdjustment) { MDefinition* owner; LinearSum ownerByteOffset(alloc()); loadTypedObjectData(typedObj, &owner, &ownerByteOffset); if (!ownerByteOffset.add(baseByteOffset)) setForceAbort(); TemporaryTypeSet* ownerTypes = owner->resultTypeSet(); const Class* clasp = ownerTypes ? ownerTypes->getKnownClass(constraints()) : nullptr; if (clasp && IsInlineTypedObjectClass(clasp)) { // Perform the load directly from the owner pointer. if (!ownerByteOffset.add(InlineTypedObject::offsetOfDataStart())) setForceAbort(); *ownerElements = owner; } else { bool definitelyOutline = clasp && IsOutlineTypedObjectClass(clasp); *ownerElements = MTypedObjectElements::New(alloc(), owner, definitelyOutline); current->add((*ownerElements)->toInstruction()); } // Extract the constant adjustment from the byte offset. *ownerByteAdjustment = ownerByteOffset.constant(); int32_t negativeAdjustment; if (!SafeSub(0, *ownerByteAdjustment, &negativeAdjustment)) setForceAbort(); if (!ownerByteOffset.add(negativeAdjustment)) setForceAbort(); // Scale the byte offset if required by the MIR node which will access the // typed object. In principle we should always be able to cleanly divide // the terms in this lienar sum due to alignment restrictions, but due to // limitations of ExtractLinearSum when applied to the terms in derived // typed objects this isn't always be possible. In these cases, fall back // on an explicit division operation. if (ownerByteOffset.divide(scale)) { *ownerScaledOffset = ConvertLinearSum(alloc(), current, ownerByteOffset); } else { MDefinition* unscaledOffset = ConvertLinearSum(alloc(), current, ownerByteOffset); *ownerScaledOffset = MDiv::New(alloc(), unscaledOffset, constantInt(scale), MIRType::Int32, /* unsigned = */ false); current->add((*ownerScaledOffset)->toInstruction()); } } // Looks up the offset/type-repr-set of the field `id`, given the type // set `objTypes` of the field owner. If a field is found, returns true // and sets *fieldOffset, *fieldPrediction, and *fieldIndex. Returns false // otherwise. Infallible. bool IonBuilder::typedObjectHasField(MDefinition* typedObj, PropertyName* name, size_t* fieldOffset, TypedObjectPrediction* fieldPrediction, size_t* fieldIndex) { TypedObjectPrediction objPrediction = typedObjectPrediction(typedObj); if (objPrediction.isUseless()) { trackOptimizationOutcome(TrackedOutcome::AccessNotTypedObject); return false; } // Must be accessing a struct. if (objPrediction.kind() != type::Struct) { trackOptimizationOutcome(TrackedOutcome::NotStruct); return false; } // Determine the type/offset of the field `name`, if any. if (!objPrediction.hasFieldNamed(NameToId(name), fieldOffset, fieldPrediction, fieldIndex)) { trackOptimizationOutcome(TrackedOutcome::StructNoField); return false; } return true; } MDefinition* IonBuilder::typeObjectForElementFromArrayStructType(MDefinition* typeObj) { MInstruction* elemType = MLoadFixedSlot::New(alloc(), typeObj, JS_DESCR_SLOT_ARRAY_ELEM_TYPE); current->add(elemType); MInstruction* unboxElemType = MUnbox::New(alloc(), elemType, MIRType::Object, MUnbox::Infallible); current->add(unboxElemType); return unboxElemType; } MDefinition* IonBuilder::typeObjectForFieldFromStructType(MDefinition* typeObj, size_t fieldIndex) { // Load list of field type objects. MInstruction* fieldTypes = MLoadFixedSlot::New(alloc(), typeObj, JS_DESCR_SLOT_STRUCT_FIELD_TYPES); current->add(fieldTypes); MInstruction* unboxFieldTypes = MUnbox::New(alloc(), fieldTypes, MIRType::Object, MUnbox::Infallible); current->add(unboxFieldTypes); // Index into list with index of field. MInstruction* fieldTypesElements = MElements::New(alloc(), unboxFieldTypes); current->add(fieldTypesElements); MConstant* fieldIndexDef = constantInt(fieldIndex); MInstruction* fieldType = MLoadElement::New(alloc(), fieldTypesElements, fieldIndexDef, false, false); current->add(fieldType); MInstruction* unboxFieldType = MUnbox::New(alloc(), fieldType, MIRType::Object, MUnbox::Infallible); current->add(unboxFieldType); return unboxFieldType; } bool IonBuilder::storeScalarTypedObjectValue(MDefinition* typedObj, const LinearSum& byteOffset, ScalarTypeDescr::Type type, MDefinition* value) { // Find location within the owner object. MDefinition* elements; MDefinition* scaledOffset; int32_t adjustment; uint32_t alignment = ScalarTypeDescr::alignment(type); loadTypedObjectElements(typedObj, byteOffset, alignment, &elements, &scaledOffset, &adjustment); // Clamp value to [0, 255] when type is Uint8Clamped MDefinition* toWrite = value; if (type == Scalar::Uint8Clamped) { toWrite = MClampToUint8::New(alloc(), value); current->add(toWrite->toInstruction()); } MStoreUnboxedScalar* store = MStoreUnboxedScalar::New(alloc(), elements, scaledOffset, toWrite, type, MStoreUnboxedScalar::TruncateInput, DoesNotRequireMemoryBarrier, adjustment); current->add(store); return true; } bool IonBuilder::storeReferenceTypedObjectValue(MDefinition* typedObj, const LinearSum& byteOffset, ReferenceTypeDescr::Type type, MDefinition* value, PropertyName* name) { // Make sure we aren't adding new type information for writes of object and value // references. if (type != ReferenceTypeDescr::TYPE_STRING) { MOZ_ASSERT(type == ReferenceTypeDescr::TYPE_ANY || type == ReferenceTypeDescr::TYPE_OBJECT); MIRType implicitType = (type == ReferenceTypeDescr::TYPE_ANY) ? MIRType::Undefined : MIRType::Null; if (PropertyWriteNeedsTypeBarrier(alloc(), constraints(), current, &typedObj, name, &value, /* canModify = */ true, implicitType)) { trackOptimizationOutcome(TrackedOutcome::NeedsTypeBarrier); return false; } } // Find location within the owner object. MDefinition* elements; MDefinition* scaledOffset; int32_t adjustment; uint32_t alignment = ReferenceTypeDescr::alignment(type); loadTypedObjectElements(typedObj, byteOffset, alignment, &elements, &scaledOffset, &adjustment); MInstruction* store = nullptr; // initialize to silence GCC warning switch (type) { case ReferenceTypeDescr::TYPE_ANY: if (NeedsPostBarrier(value)) current->add(MPostWriteBarrier::New(alloc(), typedObj, value)); store = MStoreElement::New(alloc(), elements, scaledOffset, value, false, adjustment); store->toStoreElement()->setNeedsBarrier(); break; case ReferenceTypeDescr::TYPE_OBJECT: // Note: We cannot necessarily tell at this point whether a post // barrier is needed, because the type policy may insert ToObjectOrNull // instructions later, and those may require a post barrier. Therefore, // defer the insertion of post barriers to the type policy. store = MStoreUnboxedObjectOrNull::New(alloc(), elements, scaledOffset, value, typedObj, adjustment); break; case ReferenceTypeDescr::TYPE_STRING: // Strings are not nursery allocated, so these writes do not need post // barriers. store = MStoreUnboxedString::New(alloc(), elements, scaledOffset, value, adjustment); break; } current->add(store); return true; } JSObject* IonBuilder::checkNurseryObject(JSObject* obj) { // If we try to use any nursery pointers during compilation, make sure that // the main thread will cancel this compilation before performing a minor // GC. All constants used during compilation should either go through this // function or should come from a type set (which has a similar barrier). if (obj && IsInsideNursery(obj)) { compartment->runtime()->setMinorGCShouldCancelIonCompilations(); IonBuilder* builder = this; while (builder) { builder->setNotSafeForMinorGC(); builder = builder->callerBuilder_; } } return obj; } MConstant* IonBuilder::constant(const Value& v) { MOZ_ASSERT(!v.isString() || v.toString()->isAtom(), "Handle non-atomized strings outside IonBuilder."); if (v.isObject()) checkNurseryObject(&v.toObject()); MConstant* c = MConstant::New(alloc(), v, constraints()); current->add(c); return c; } MConstant* IonBuilder::constantInt(int32_t i) { return constant(Int32Value(i)); } MInstruction* IonBuilder::initializedLength(MDefinition* obj, MDefinition* elements) { MInstruction* res = MInitializedLength::New(alloc(), elements); current->add(res); return res; } MInstruction* IonBuilder::setInitializedLength(MDefinition* obj, size_t count) { MOZ_ASSERT(count); // MSetInitializedLength takes the index of the last element, rather // than the count itself. MInstruction* elements = MElements::New(alloc(), obj, /* unboxed = */ false); current->add(elements); MInstruction* res = MSetInitializedLength::New(alloc(), elements, constant(Int32Value(count - 1))); current->add(res); return res; } MDefinition* IonBuilder::getCallee() { if (inliningDepth_ == 0) { MInstruction* callee = MCallee::New(alloc()); current->add(callee); return callee; } return inlineCallInfo_->fun(); } MDefinition* IonBuilder::addLexicalCheck(MDefinition* input) { MOZ_ASSERT(JSOp(*pc) == JSOP_CHECKLEXICAL || JSOp(*pc) == JSOP_CHECKALIASEDLEXICAL || JSOp(*pc) == JSOP_GETIMPORT); MInstruction* lexicalCheck; // If we're guaranteed to not be JS_UNINITIALIZED_LEXICAL, no need to check. if (input->type() == MIRType::MagicUninitializedLexical) { // Mark the input as implicitly used so the JS_UNINITIALIZED_LEXICAL // magic value will be preserved on bailout. input->setImplicitlyUsedUnchecked(); lexicalCheck = MThrowRuntimeLexicalError::New(alloc(), JSMSG_UNINITIALIZED_LEXICAL); current->add(lexicalCheck); if (!resumeAfter(lexicalCheck)) return nullptr; return constant(UndefinedValue()); } if (input->type() == MIRType::Value) { lexicalCheck = MLexicalCheck::New(alloc(), input); current->add(lexicalCheck); if (failedLexicalCheck_) lexicalCheck->setNotMovableUnchecked(); return lexicalCheck; } return input; } MDefinition* IonBuilder::convertToBoolean(MDefinition* input) { // Convert to bool with the '!!' idiom MNot* resultInverted = MNot::New(alloc(), input, constraints()); current->add(resultInverted); MNot* result = MNot::New(alloc(), resultInverted, constraints()); current->add(result); return result; } void IonBuilder::trace(JSTracer* trc) { if (!compartment->runtime()->runtimeMatches(trc->runtime())) return; MOZ_ASSERT(rootList_); rootList_->trace(trc); }