var Debugger = /******/ (function(modules) { // webpackBootstrap /******/ // The module cache /******/ var installedModules = {}; /******/ /******/ // The require function /******/ function __webpack_require__(moduleId) { /******/ /******/ // Check if module is in cache /******/ if(installedModules[moduleId]) /******/ return installedModules[moduleId].exports; /******/ /******/ // Create a new module (and put it into the cache) /******/ var module = installedModules[moduleId] = { /******/ exports: {}, /******/ id: moduleId, /******/ loaded: false /******/ }; /******/ /******/ // Execute the module function /******/ modules[moduleId].call(module.exports, module, module.exports, __webpack_require__); /******/ /******/ // Flag the module as loaded /******/ module.loaded = true; /******/ /******/ // Return the exports of the module /******/ return module.exports; /******/ } /******/ /******/ /******/ // expose the modules object (__webpack_modules__) /******/ __webpack_require__.m = modules; /******/ /******/ // expose the module cache /******/ __webpack_require__.c = installedModules; /******/ /******/ // __webpack_public_path__ /******/ __webpack_require__.p = "/public/build"; /******/ /******/ // Load entry module and return exports /******/ return __webpack_require__(0); /******/ }) /************************************************************************/ /******/ ({ /***/ 0: /***/ function(module, exports, __webpack_require__) { var prettyFast = __webpack_require__(460); var assert = __webpack_require__(247); function prettyPrint(_ref) { var url = _ref.url; var indent = _ref.indent; var source = _ref.source; try { var prettified = prettyFast(source, { url: url, indent: " ".repeat(indent) }); return { code: prettified.code, mappings: prettified.map._mappings }; } catch (e) { return new Error(e.message + "\n" + e.stack); } } function invertMappings(mappings) { return mappings._array.map(m => { var mapping = { generated: { line: m.originalLine, column: m.originalColumn } }; if (m.source) { mapping.source = m.source; mapping.original = { line: m.generatedLine, column: m.generatedColumn }; mapping.name = m.name; } return mapping; }); } self.onmessage = function (msg) { var _msg$data = msg.data; var id = _msg$data.id; var args = _msg$data.args; assert(msg.data.method === "prettyPrint", "Method must be `prettyPrint`"); try { var _prettyPrint = prettyPrint(args[0]); var code = _prettyPrint.code; var mappings = _prettyPrint.mappings; self.postMessage({ id, response: { code, mappings: invertMappings(mappings) } }); } catch (e) { self.postMessage({ id, error: e }); } }; /***/ }, /***/ 247: /***/ function(module, exports) { function assert(condition, message) { if (!condition) { throw new Error("Assertion failure: " + message); } } module.exports = assert; /***/ }, /***/ 460: /***/ function(module, exports, __webpack_require__) { var __WEBPACK_AMD_DEFINE_FACTORY__, __WEBPACK_AMD_DEFINE_RESULT__;/* -*- indent-tabs-mode: nil; js-indent-level: 2; fill-column: 80 -*- */ /* * Copyright 2013 Mozilla Foundation and contributors * Licensed under the New BSD license. See LICENSE.md or: * http://opensource.org/licenses/BSD-2-Clause */ (function (root, factory) { "use strict"; if (true) { !(__WEBPACK_AMD_DEFINE_FACTORY__ = (factory), __WEBPACK_AMD_DEFINE_RESULT__ = (typeof __WEBPACK_AMD_DEFINE_FACTORY__ === 'function' ? (__WEBPACK_AMD_DEFINE_FACTORY__.call(exports, __webpack_require__, exports, module)) : __WEBPACK_AMD_DEFINE_FACTORY__), __WEBPACK_AMD_DEFINE_RESULT__ !== undefined && (module.exports = __WEBPACK_AMD_DEFINE_RESULT__)); } else if (typeof exports === "object") { module.exports = factory(); } else { root.prettyFast = factory(); } }(this, function () { "use strict"; var acorn = this.acorn || __webpack_require__(461); var sourceMap = this.sourceMap || __webpack_require__(462); var SourceNode = sourceMap.SourceNode; // If any of these tokens are seen before a "[" token, we know that "[" token // is the start of an array literal, rather than a property access. // // The only exception is "}", which would need to be disambiguated by // parsing. The majority of the time, an open bracket following a closing // curly is going to be an array literal, so we brush the complication under // the rug, and handle the ambiguity by always assuming that it will be an // array literal. var PRE_ARRAY_LITERAL_TOKENS = { "typeof": true, "void": true, "delete": true, "case": true, "do": true, "=": true, "in": true, "{": true, "*": true, "/": true, "%": true, "else": true, ";": true, "++": true, "--": true, "+": true, "-": true, "~": true, "!": true, ":": true, "?": true, ">>": true, ">>>": true, "<<": true, "||": true, "&&": true, "<": true, ">": true, "<=": true, ">=": true, "instanceof": true, "&": true, "^": true, "|": true, "==": true, "!=": true, "===": true, "!==": true, ",": true, "}": true }; /** * Determines if we think that the given token starts an array literal. * * @param Object token * The token we want to determine if it is an array literal. * @param Object lastToken * The last token we added to the pretty printed results. * * @returns Boolean * True if we believe it is an array literal, false otherwise. */ function isArrayLiteral(token, lastToken) { if (token.type.type != "[") { return false; } if (!lastToken) { return true; } if (lastToken.type.isAssign) { return true; } return !!PRE_ARRAY_LITERAL_TOKENS[ lastToken.type.keyword || lastToken.type.type ]; } // If any of these tokens are followed by a token on a new line, we know that // ASI cannot happen. var PREVENT_ASI_AFTER_TOKENS = { // Binary operators "*": true, "/": true, "%": true, "+": true, "-": true, "<<": true, ">>": true, ">>>": true, "<": true, ">": true, "<=": true, ">=": true, "instanceof": true, "in": true, "==": true, "!=": true, "===": true, "!==": true, "&": true, "^": true, "|": true, "&&": true, "||": true, ",": true, ".": true, "=": true, "*=": true, "/=": true, "%=": true, "+=": true, "-=": true, "<<=": true, ">>=": true, ">>>=": true, "&=": true, "^=": true, "|=": true, // Unary operators "delete": true, "void": true, "typeof": true, "~": true, "!": true, "new": true, // Function calls and grouped expressions "(": true }; // If any of these tokens are on a line after the token before it, we know // that ASI cannot happen. var PREVENT_ASI_BEFORE_TOKENS = { // Binary operators "*": true, "/": true, "%": true, "<<": true, ">>": true, ">>>": true, "<": true, ">": true, "<=": true, ">=": true, "instanceof": true, "in": true, "==": true, "!=": true, "===": true, "!==": true, "&": true, "^": true, "|": true, "&&": true, "||": true, ",": true, ".": true, "=": true, "*=": true, "/=": true, "%=": true, "+=": true, "-=": true, "<<=": true, ">>=": true, ">>>=": true, "&=": true, "^=": true, "|=": true, // Function calls "(": true }; /** * Determines if Automatic Semicolon Insertion (ASI) occurs between these * tokens. * * @param Object token * The current token. * @param Object lastToken * The last token we added to the pretty printed results. * * @returns Boolean * True if we believe ASI occurs. */ function isASI(token, lastToken) { if (!lastToken) { return false; } if (token.startLoc.line === lastToken.startLoc.line) { return false; } if (PREVENT_ASI_AFTER_TOKENS[ lastToken.type.type || lastToken.type.keyword ]) { return false; } if (PREVENT_ASI_BEFORE_TOKENS[token.type.type || token.type.keyword]) { return false; } return true; } /** * Determine if we have encountered a getter or setter. * * @param Object token * The current token. If this is a getter or setter, it would be the * property name. * @param Object lastToken * The last token we added to the pretty printed results. If this is a * getter or setter, it would be the `get` or `set` keyword * respectively. * @param Array stack * The stack of open parens/curlies/brackets/etc. * * @returns Boolean * True if this is a getter or setter. */ function isGetterOrSetter(token, lastToken, stack) { return stack[stack.length - 1] == "{" && lastToken && lastToken.type.type == "name" && (lastToken.value == "get" || lastToken.value == "set") && token.type.type == "name"; } /** * Determine if we should add a newline after the given token. * * @param Object token * The token we are looking at. * @param Array stack * The stack of open parens/curlies/brackets/etc. * * @returns Boolean * True if we should add a newline. */ function isLineDelimiter(token, stack) { if (token.isArrayLiteral) { return true; } var ttt = token.type.type; var top = stack[stack.length - 1]; return ttt == ";" && top != "(" || ttt == "{" || ttt == "," && top != "(" || ttt == ":" && (top == "case" || top == "default"); } /** * Append the necessary whitespace to the result after we have added the given * token. * * @param Object token * The token that was just added to the result. * @param Function write * The function to write to the pretty printed results. * @param Array stack * The stack of open parens/curlies/brackets/etc. * * @returns Boolean * Returns true if we added a newline to result, false in all other * cases. */ function appendNewline(token, write, stack) { if (isLineDelimiter(token, stack)) { write("\n", token.startLoc.line, token.startLoc.column); return true; } return false; } /** * Determines if we need to add a space between the last token we added and * the token we are about to add. * * @param Object token * The token we are about to add to the pretty printed code. * @param Object lastToken * The last token added to the pretty printed code. */ function needsSpaceAfter(token, lastToken) { if (lastToken) { if (lastToken.type.isLoop) { return true; } if (lastToken.type.isAssign) { return true; } if (lastToken.type.binop != null) { return true; } var ltt = lastToken.type.type; if (ltt == "?") { return true; } if (ltt == ":") { return true; } if (ltt == ",") { return true; } if (ltt == ";") { return true; } var ltk = lastToken.type.keyword; if (ltk != null) { if (ltk == "break" || ltk == "continue" || ltk == "return") { return token.type.type != ";"; } if (ltk != "debugger" && ltk != "null" && ltk != "true" && ltk != "false" && ltk != "this" && ltk != "default") { return true; } } if (ltt == ")" && (token.type.type != ")" && token.type.type != "]" && token.type.type != ";" && token.type.type != "," && token.type.type != ".")) { return true; } } if (token.type.isAssign) { return true; } if (token.type.binop != null) { return true; } if (token.type.type == "?") { return true; } return false; } /** * Add the required whitespace before this token, whether that is a single * space, newline, and/or the indent on fresh lines. * * @param Object token * The token we are about to add to the pretty printed code. * @param Object lastToken * The last token we added to the pretty printed code. * @param Boolean addedNewline * Whether we added a newline after adding the last token to the pretty * printed code. * @param Function write * The function to write pretty printed code to the result SourceNode. * @param Object options * The options object. * @param Number indentLevel * The number of indents deep we are. * @param Array stack * The stack of open curlies, brackets, etc. */ function prependWhiteSpace(token, lastToken, addedNewline, write, options, indentLevel, stack) { var ttk = token.type.keyword; var ttt = token.type.type; var newlineAdded = addedNewline; var ltt = lastToken ? lastToken.type.type : null; // Handle whitespace and newlines after "}" here instead of in // `isLineDelimiter` because it is only a line delimiter some of the // time. For example, we don't want to put "else if" on a new line after // the first if's block. if (lastToken && ltt == "}") { if (ttk == "while" && stack[stack.length - 1] == "do") { write(" ", lastToken.startLoc.line, lastToken.startLoc.column); } else if (ttk == "else" || ttk == "catch" || ttk == "finally") { write(" ", lastToken.startLoc.line, lastToken.startLoc.column); } else if (ttt != "(" && ttt != ";" && ttt != "," && ttt != ")" && ttt != ".") { write("\n", lastToken.startLoc.line, lastToken.startLoc.column); newlineAdded = true; } } if (isGetterOrSetter(token, lastToken, stack)) { write(" ", lastToken.startLoc.line, lastToken.startLoc.column); } if (ttt == ":" && stack[stack.length - 1] == "?") { write(" ", lastToken.startLoc.line, lastToken.startLoc.column); } if (lastToken && ltt != "}" && ttk == "else") { write(" ", lastToken.startLoc.line, lastToken.startLoc.column); } function ensureNewline() { if (!newlineAdded) { write("\n", lastToken.startLoc.line, lastToken.startLoc.column); newlineAdded = true; } } if (isASI(token, lastToken)) { ensureNewline(); } if (decrementsIndent(ttt, stack)) { ensureNewline(); } if (newlineAdded) { if (ttk == "case" || ttk == "default") { write(repeat(options.indent, indentLevel - 1), token.startLoc.line, token.startLoc.column); } else { write(repeat(options.indent, indentLevel), token.startLoc.line, token.startLoc.column); } } else if (needsSpaceAfter(token, lastToken)) { write(" ", lastToken.startLoc.line, lastToken.startLoc.column); } } /** * Repeat the `str` string `n` times. * * @param String str * The string to be repeated. * @param Number n * The number of times to repeat the string. * * @returns String * The repeated string. */ function repeat(str, n) { var result = ""; while (n > 0) { if (n & 1) { result += str; } n >>= 1; str += str; } return result; } /** * Make sure that we output the escaped character combination inside string * literals instead of various problematic characters. */ var sanitize = (function () { var escapeCharacters = { // Backslash "\\": "\\\\", // Newlines "\n": "\\n", // Carriage return "\r": "\\r", // Tab "\t": "\\t", // Vertical tab "\v": "\\v", // Form feed "\f": "\\f", // Null character "\0": "\\0", // Single quotes "'": "\\'" }; var regExpString = "(" + Object.keys(escapeCharacters) .map(function (c) { return escapeCharacters[c]; }) .join("|") + ")"; var escapeCharactersRegExp = new RegExp(regExpString, "g"); return function (str) { return str.replace(escapeCharactersRegExp, function (_, c) { return escapeCharacters[c]; }); }; }()); /** * Add the given token to the pretty printed results. * * @param Object token * The token to add. * @param Function write * The function to write pretty printed code to the result SourceNode. */ function addToken(token, write) { if (token.type.type == "string") { write("'" + sanitize(token.value) + "'", token.startLoc.line, token.startLoc.column); } else if (token.type.type == "regexp") { write(String(token.value.value), token.startLoc.line, token.startLoc.column); } else { write(String(token.value != null ? token.value : token.type.type), token.startLoc.line, token.startLoc.column); } } /** * Returns true if the given token type belongs on the stack. */ function belongsOnStack(token) { var ttt = token.type.type; var ttk = token.type.keyword; return ttt == "{" || ttt == "(" || ttt == "[" || ttt == "?" || ttk == "do" || ttk == "switch" || ttk == "case" || ttk == "default"; } /** * Returns true if the given token should cause us to pop the stack. */ function shouldStackPop(token, stack) { var ttt = token.type.type; var ttk = token.type.keyword; var top = stack[stack.length - 1]; return ttt == "]" || ttt == ")" || ttt == "}" || (ttt == ":" && (top == "case" || top == "default" || top == "?")) || (ttk == "while" && top == "do"); } /** * Returns true if the given token type should cause us to decrement the * indent level. */ function decrementsIndent(tokenType, stack) { return tokenType == "}" || (tokenType == "]" && stack[stack.length - 1] == "[\n"); } /** * Returns true if the given token should cause us to increment the indent * level. */ function incrementsIndent(token) { return token.type.type == "{" || token.isArrayLiteral || token.type.keyword == "switch"; } /** * Add a comment to the pretty printed code. * * @param Function write * The function to write pretty printed code to the result SourceNode. * @param Number indentLevel * The number of indents deep we are. * @param Object options * The options object. * @param Boolean block * True if the comment is a multiline block style comment. * @param String text * The text of the comment. * @param Number line * The line number to comment appeared on. * @param Number column * The column number the comment appeared on. */ function addComment(write, indentLevel, options, block, text, line, column) { var indentString = repeat(options.indent, indentLevel); write(indentString, line, column); if (block) { write("/*"); write(text .split(new RegExp("/\n" + indentString + "/", "g")) .join("\n" + indentString)); write("*/"); } else { write("//"); write(text); } write("\n"); } /** * The main function. * * @param String input * The ugly JS code we want to pretty print. * @param Object options * The options object. Provides configurability of the pretty * printing. Properties: * - url: The URL string of the ugly JS code. * - indent: The string to indent code by. * * @returns Object * An object with the following properties: * - code: The pretty printed code string. * - map: A SourceMapGenerator instance. */ return function prettyFast(input, options) { // The level of indents deep we are. var indentLevel = 0; // We will accumulate the pretty printed code in this SourceNode. var result = new SourceNode(); /** * Write a pretty printed string to the result SourceNode. * * We buffer our writes so that we only create one mapping for each line in * the source map. This enhances performance by avoiding extraneous mapping * serialization, and flattening the tree that * `SourceNode#toStringWithSourceMap` will have to recursively walk. When * timing how long it takes to pretty print jQuery, this optimization * brought the time down from ~390 ms to ~190ms! * * @param String str * The string to be added to the result. * @param Number line * The line number the string came from in the ugly source. * @param Number column * The column number the string came from in the ugly source. */ var write = (function () { var buffer = []; var bufferLine = -1; var bufferColumn = -1; return function write(str, line, column) { if (line != null && bufferLine === -1) { bufferLine = line; } if (column != null && bufferColumn === -1) { bufferColumn = column; } buffer.push(str); if (str == "\n") { var lineStr = ""; for (var i = 0, len = buffer.length; i < len; i++) { lineStr += buffer[i]; } result.add(new SourceNode(bufferLine, bufferColumn, options.url, lineStr)); buffer.splice(0, buffer.length); bufferLine = -1; bufferColumn = -1; } }; }()); // Whether or not we added a newline on after we added the last token. var addedNewline = false; // The current token we will be adding to the pretty printed code. var token; // Shorthand for token.type.type, so we don't have to repeatedly access // properties. var ttt; // Shorthand for token.type.keyword, so we don't have to repeatedly access // properties. var ttk; // The last token we added to the pretty printed code. var lastToken; // Stack of token types/keywords that can affect whether we want to add a // newline or a space. We can make that decision based on what token type is // on the top of the stack. For example, a comma in a parameter list should // be followed by a space, while a comma in an object literal should be // followed by a newline. // // Strings that go on the stack: // // - "{" // - "(" // - "[" // - "[\n" // - "do" // - "?" // - "switch" // - "case" // - "default" // // The difference between "[" and "[\n" is that "[\n" is used when we are // treating "[" and "]" tokens as line delimiters and should increment and // decrement the indent level when we find them. var stack = []; // Acorn's tokenizer will always yield comments *before* the token they // follow (unless the very first thing in the source is a comment), so we // have to queue the comments in order to pretty print them in the correct // location. For example, the source file: // // foo // // a // // b // bar // // When tokenized by acorn, gives us the following token stream: // // [ '// a', '// b', foo, bar ] var commentQueue = []; var getToken = acorn.tokenize(input, { locations: true, sourceFile: options.url, onComment: function (block, text, start, end, startLoc, endLoc) { if (lastToken) { commentQueue.push({ block: block, text: text, line: startLoc.line, column: startLoc.column, trailing: lastToken.endLoc.line == startLoc.line }); } else { addComment(write, indentLevel, options, block, text, startLoc.line, startLoc.column); addedNewline = true; } } }); for (;;) { token = getToken(); ttk = token.type.keyword; ttt = token.type.type; if (ttt == "eof") { if (!addedNewline) { write("\n"); } break; } token.isArrayLiteral = isArrayLiteral(token, lastToken); if (belongsOnStack(token)) { if (token.isArrayLiteral) { stack.push("[\n"); } else { stack.push(ttt || ttk); } } if (decrementsIndent(ttt, stack)) { indentLevel--; if (ttt == "}" && stack.length > 1 && stack[stack.length - 2] == "switch") { indentLevel--; } } prependWhiteSpace(token, lastToken, addedNewline, write, options, indentLevel, stack); addToken(token, write); if (commentQueue.length === 0 || !commentQueue[0].trailing) { addedNewline = appendNewline(token, write, stack); } if (shouldStackPop(token, stack)) { stack.pop(); if (token == "}" && stack.length && stack[stack.length - 1] == "switch") { stack.pop(); } } if (incrementsIndent(token)) { indentLevel++; } // Acorn's tokenizer re-uses tokens, so we have to copy the last token on // every iteration. We follow acorn's lead here, and reuse the lastToken // object the same way that acorn reuses the token object. This allows us // to avoid allocations and minimize GC pauses. if (!lastToken) { lastToken = { startLoc: {}, endLoc: {} }; } lastToken.start = token.start; lastToken.end = token.end; lastToken.startLoc.line = token.startLoc.line; lastToken.startLoc.column = token.startLoc.column; lastToken.endLoc.line = token.endLoc.line; lastToken.endLoc.column = token.endLoc.column; lastToken.type = token.type; lastToken.value = token.value; lastToken.isArrayLiteral = token.isArrayLiteral; // Apply all the comments that have been queued up. if (commentQueue.length) { if (!addedNewline && !commentQueue[0].trailing) { write("\n"); } if (commentQueue[0].trailing) { write(" "); } for (var i = 0, n = commentQueue.length; i < n; i++) { var comment = commentQueue[i]; var commentIndentLevel = commentQueue[i].trailing ? 0 : indentLevel; addComment(write, commentIndentLevel, options, comment.block, comment.text, comment.line, comment.column); } addedNewline = true; commentQueue.splice(0, commentQueue.length); } } return result.toStringWithSourceMap({ file: options.url }); }; }.bind(this))); /***/ }, /***/ 461: /***/ function(module, exports, __webpack_require__) { var __WEBPACK_AMD_DEFINE_FACTORY__, __WEBPACK_AMD_DEFINE_ARRAY__, __WEBPACK_AMD_DEFINE_RESULT__;// Acorn is a tiny, fast JavaScript parser written in JavaScript. // // Acorn was written by Marijn Haverbeke and various contributors and // released under an MIT license. The Unicode regexps (for identifiers // and whitespace) were taken from [Esprima](http://esprima.org) by // Ariya Hidayat. // // Git repositories for Acorn are available at // // http://marijnhaverbeke.nl/git/acorn // https://github.com/marijnh/acorn.git // // Please use the [github bug tracker][ghbt] to report issues. // // [ghbt]: https://github.com/marijnh/acorn/issues // // This file defines the main parser interface. The library also comes // with a [error-tolerant parser][dammit] and an // [abstract syntax tree walker][walk], defined in other files. // // [dammit]: acorn_loose.js // [walk]: util/walk.js (function(root, mod) { if (true) return mod(exports); // CommonJS if (true) return !(__WEBPACK_AMD_DEFINE_ARRAY__ = [exports], __WEBPACK_AMD_DEFINE_FACTORY__ = (mod), __WEBPACK_AMD_DEFINE_RESULT__ = (typeof __WEBPACK_AMD_DEFINE_FACTORY__ === 'function' ? (__WEBPACK_AMD_DEFINE_FACTORY__.apply(exports, __WEBPACK_AMD_DEFINE_ARRAY__)) : __WEBPACK_AMD_DEFINE_FACTORY__), __WEBPACK_AMD_DEFINE_RESULT__ !== undefined && (module.exports = __WEBPACK_AMD_DEFINE_RESULT__)); // AMD mod(root.acorn || (root.acorn = {})); // Plain browser env })(this, function(exports) { "use strict"; exports.version = "0.11.0"; // The main exported interface (under `self.acorn` when in the // browser) is a `parse` function that takes a code string and // returns an abstract syntax tree as specified by [Mozilla parser // API][api], with the caveat that inline XML is not recognized. // // [api]: https://developer.mozilla.org/en-US/docs/SpiderMonkey/Parser_API var options, input, inputLen, sourceFile; exports.parse = function(inpt, opts) { input = String(inpt); inputLen = input.length; setOptions(opts); initTokenState(); var startPos = options.locations ? [tokPos, curPosition()] : tokPos; initParserState(); return parseTopLevel(options.program || startNodeAt(startPos)); }; // A second optional argument can be given to further configure // the parser process. These options are recognized: var defaultOptions = exports.defaultOptions = { // `ecmaVersion` indicates the ECMAScript version to parse. Must // be either 3, or 5, or 6. This influences support for strict // mode, the set of reserved words, support for getters and // setters and other features. ecmaVersion: 5, // Turn on `strictSemicolons` to prevent the parser from doing // automatic semicolon insertion. strictSemicolons: false, // When `allowTrailingCommas` is false, the parser will not allow // trailing commas in array and object literals. allowTrailingCommas: true, // By default, reserved words are not enforced. Enable // `forbidReserved` to enforce them. When this option has the // value "everywhere", reserved words and keywords can also not be // used as property names. forbidReserved: false, // When enabled, a return at the top level is not considered an // error. allowReturnOutsideFunction: false, // When enabled, import/export statements are not constrained to // appearing at the top of the program. allowImportExportEverywhere: false, // When `locations` is on, `loc` properties holding objects with // `start` and `end` properties in `{line, column}` form (with // line being 1-based and column 0-based) will be attached to the // nodes. locations: false, // A function can be passed as `onToken` option, which will // cause Acorn to call that function with object in the same // format as tokenize() returns. Note that you are not // allowed to call the parser from the callback—that will // corrupt its internal state. onToken: null, // A function can be passed as `onComment` option, which will // cause Acorn to call that function with `(block, text, start, // end)` parameters whenever a comment is skipped. `block` is a // boolean indicating whether this is a block (`/* */`) comment, // `text` is the content of the comment, and `start` and `end` are // character offsets that denote the start and end of the comment. // When the `locations` option is on, two more parameters are // passed, the full `{line, column}` locations of the start and // end of the comments. Note that you are not allowed to call the // parser from the callback—that will corrupt its internal state. onComment: null, // Nodes have their start and end characters offsets recorded in // `start` and `end` properties (directly on the node, rather than // the `loc` object, which holds line/column data. To also add a // [semi-standardized][range] `range` property holding a `[start, // end]` array with the same numbers, set the `ranges` option to // `true`. // // [range]: https://bugzilla.mozilla.org/show_bug.cgi?id=745678 ranges: false, // It is possible to parse multiple files into a single AST by // passing the tree produced by parsing the first file as // `program` option in subsequent parses. This will add the // toplevel forms of the parsed file to the `Program` (top) node // of an existing parse tree. program: null, // When `locations` is on, you can pass this to record the source // file in every node's `loc` object. sourceFile: null, // This value, if given, is stored in every node, whether // `locations` is on or off. directSourceFile: null, // When enabled, parenthesized expressions are represented by // (non-standard) ParenthesizedExpression nodes preserveParens: false }; // This function tries to parse a single expression at a given // offset in a string. Useful for parsing mixed-language formats // that embed JavaScript expressions. exports.parseExpressionAt = function(inpt, pos, opts) { input = String(inpt); inputLen = input.length; setOptions(opts); initTokenState(pos); initParserState(); return parseExpression(); }; var isArray = function (obj) { return Object.prototype.toString.call(obj) === "[object Array]"; }; function setOptions(opts) { options = {}; for (var opt in defaultOptions) options[opt] = opts && has(opts, opt) ? opts[opt] : defaultOptions[opt]; sourceFile = options.sourceFile || null; if (isArray(options.onToken)) { var tokens = options.onToken; options.onToken = function (token) { tokens.push(token); }; } if (isArray(options.onComment)) { var comments = options.onComment; options.onComment = function (block, text, start, end, startLoc, endLoc) { var comment = { type: block ? 'Block' : 'Line', value: text, start: start, end: end }; if (options.locations) { comment.loc = new SourceLocation(); comment.loc.start = startLoc; comment.loc.end = endLoc; } if (options.ranges) comment.range = [start, end]; comments.push(comment); }; } isKeyword = options.ecmaVersion >= 6 ? isEcma6Keyword : isEcma5AndLessKeyword; } // The `getLineInfo` function is mostly useful when the // `locations` option is off (for performance reasons) and you // want to find the line/column position for a given character // offset. `input` should be the code string that the offset refers // into. var getLineInfo = exports.getLineInfo = function(input, offset) { for (var line = 1, cur = 0;;) { lineBreak.lastIndex = cur; var match = lineBreak.exec(input); if (match && match.index < offset) { ++line; cur = match.index + match[0].length; } else break; } return {line: line, column: offset - cur}; }; function Token() { this.type = tokType; this.value = tokVal; this.start = tokStart; this.end = tokEnd; if (options.locations) { this.loc = new SourceLocation(); this.loc.end = tokEndLoc; // TODO: remove in next major release this.startLoc = tokStartLoc; this.endLoc = tokEndLoc; } if (options.ranges) this.range = [tokStart, tokEnd]; } exports.Token = Token; // Acorn is organized as a tokenizer and a recursive-descent parser. // The `tokenize` export provides an interface to the tokenizer. // Because the tokenizer is optimized for being efficiently used by // the Acorn parser itself, this interface is somewhat crude and not // very modular. Performing another parse or call to `tokenize` will // reset the internal state, and invalidate existing tokenizers. exports.tokenize = function(inpt, opts) { input = String(inpt); inputLen = input.length; setOptions(opts); initTokenState(); skipSpace(); function getToken(forceRegexp) { lastEnd = tokEnd; readToken(forceRegexp); return new Token(); } getToken.jumpTo = function(pos, reAllowed) { tokPos = pos; if (options.locations) { tokCurLine = 1; tokLineStart = lineBreak.lastIndex = 0; var match; while ((match = lineBreak.exec(input)) && match.index < pos) { ++tokCurLine; tokLineStart = match.index + match[0].length; } } tokRegexpAllowed = reAllowed; skipSpace(); }; getToken.noRegexp = function() { tokRegexpAllowed = false; }; getToken.options = options; return getToken; }; // State is kept in (closure-)global variables. We already saw the // `options`, `input`, and `inputLen` variables above. // The current position of the tokenizer in the input. var tokPos; // The start and end offsets of the current token. var tokStart, tokEnd; // When `options.locations` is true, these hold objects // containing the tokens start and end line/column pairs. var tokStartLoc, tokEndLoc; // The type and value of the current token. Token types are objects, // named by variables against which they can be compared, and // holding properties that describe them (indicating, for example, // the precedence of an infix operator, and the original name of a // keyword token). The kind of value that's held in `tokVal` depends // on the type of the token. For literals, it is the literal value, // for operators, the operator name, and so on. var tokType, tokVal; // Internal state for the tokenizer. To distinguish between division // operators and regular expressions, it remembers whether the last // token was one that is allowed to be followed by an expression. // (If it is, a slash is probably a regexp, if it isn't it's a // division operator. See the `parseStatement` function for a // caveat.) var tokRegexpAllowed; // When `options.locations` is true, these are used to keep // track of the current line, and know when a new line has been // entered. var tokCurLine, tokLineStart; // These store the position of the previous token, which is useful // when finishing a node and assigning its `end` position. var lastStart, lastEnd, lastEndLoc; // This is the parser's state. `inFunction` is used to reject // `return` statements outside of functions, `inGenerator` to // reject `yield`s outside of generators, `labels` to verify // that `break` and `continue` have somewhere to jump to, and // `strict` indicates whether strict mode is on. var inFunction, inGenerator, labels, strict; // This counter is used for checking that arrow expressions did // not contain nested parentheses in argument list. var metParenL; // This is used by the tokenizer to track the template strings it is // inside, and count the amount of open braces seen inside them, to // be able to switch back to a template token when the } to match ${ // is encountered. It will hold an array of integers. var templates; function initParserState() { lastStart = lastEnd = tokPos; if (options.locations) lastEndLoc = curPosition(); inFunction = inGenerator = strict = false; labels = []; skipSpace(); readToken(); } // This function is used to raise exceptions on parse errors. It // takes an offset integer (into the current `input`) to indicate // the location of the error, attaches the position to the end // of the error message, and then raises a `SyntaxError` with that // message. function raise(pos, message) { var loc = getLineInfo(input, pos); message += " (" + loc.line + ":" + loc.column + ")"; var err = new SyntaxError(message); err.pos = pos; err.loc = loc; err.raisedAt = tokPos; throw err; } // Reused empty array added for node fields that are always empty. var empty = []; // ## Token types // The assignment of fine-grained, information-carrying type objects // allows the tokenizer to store the information it has about a // token in a way that is very cheap for the parser to look up. // All token type variables start with an underscore, to make them // easy to recognize. // These are the general types. The `type` property is only used to // make them recognizeable when debugging. var _num = {type: "num"}, _regexp = {type: "regexp"}, _string = {type: "string"}; var _name = {type: "name"}, _eof = {type: "eof"}; // Keyword tokens. The `keyword` property (also used in keyword-like // operators) indicates that the token originated from an // identifier-like word, which is used when parsing property names. // // The `beforeExpr` property is used to disambiguate between regular // expressions and divisions. It is set on all token types that can // be followed by an expression (thus, a slash after them would be a // regular expression). // // `isLoop` marks a keyword as starting a loop, which is important // to know when parsing a label, in order to allow or disallow // continue jumps to that label. var _break = {keyword: "break"}, _case = {keyword: "case", beforeExpr: true}, _catch = {keyword: "catch"}; var _continue = {keyword: "continue"}, _debugger = {keyword: "debugger"}, _default = {keyword: "default"}; var _do = {keyword: "do", isLoop: true}, _else = {keyword: "else", beforeExpr: true}; var _finally = {keyword: "finally"}, _for = {keyword: "for", isLoop: true}, _function = {keyword: "function"}; var _if = {keyword: "if"}, _return = {keyword: "return", beforeExpr: true}, _switch = {keyword: "switch"}; var _throw = {keyword: "throw", beforeExpr: true}, _try = {keyword: "try"}, _var = {keyword: "var"}; var _let = {keyword: "let"}, _const = {keyword: "const"}; var _while = {keyword: "while", isLoop: true}, _with = {keyword: "with"}, _new = {keyword: "new", beforeExpr: true}; var _this = {keyword: "this"}; var _class = {keyword: "class"}, _extends = {keyword: "extends", beforeExpr: true}; var _export = {keyword: "export"}, _import = {keyword: "import"}; var _yield = {keyword: "yield", beforeExpr: true}; // The keywords that denote values. var _null = {keyword: "null", atomValue: null}, _true = {keyword: "true", atomValue: true}; var _false = {keyword: "false", atomValue: false}; // Some keywords are treated as regular operators. `in` sometimes // (when parsing `for`) needs to be tested against specifically, so // we assign a variable name to it for quick comparing. var _in = {keyword: "in", binop: 7, beforeExpr: true}; // Map keyword names to token types. var keywordTypes = {"break": _break, "case": _case, "catch": _catch, "continue": _continue, "debugger": _debugger, "default": _default, "do": _do, "else": _else, "finally": _finally, "for": _for, "function": _function, "if": _if, "return": _return, "switch": _switch, "throw": _throw, "try": _try, "var": _var, "let": _let, "const": _const, "while": _while, "with": _with, "null": _null, "true": _true, "false": _false, "new": _new, "in": _in, "instanceof": {keyword: "instanceof", binop: 7, beforeExpr: true}, "this": _this, "typeof": {keyword: "typeof", prefix: true, beforeExpr: true}, "void": {keyword: "void", prefix: true, beforeExpr: true}, "delete": {keyword: "delete", prefix: true, beforeExpr: true}, "class": _class, "extends": _extends, "export": _export, "import": _import, "yield": _yield}; // Punctuation token types. Again, the `type` property is purely for debugging. var _bracketL = {type: "[", beforeExpr: true}, _bracketR = {type: "]"}, _braceL = {type: "{", beforeExpr: true}; var _braceR = {type: "}"}, _parenL = {type: "(", beforeExpr: true}, _parenR = {type: ")"}; var _comma = {type: ",", beforeExpr: true}, _semi = {type: ";", beforeExpr: true}; var _colon = {type: ":", beforeExpr: true}, _dot = {type: "."}, _question = {type: "?", beforeExpr: true}; var _arrow = {type: "=>", beforeExpr: true}, _template = {type: "template"}, _templateContinued = {type: "templateContinued"}; var _ellipsis = {type: "...", prefix: true, beforeExpr: true}; // Operators. These carry several kinds of properties to help the // parser use them properly (the presence of these properties is // what categorizes them as operators). // // `binop`, when present, specifies that this operator is a binary // operator, and will refer to its precedence. // // `prefix` and `postfix` mark the operator as a prefix or postfix // unary operator. `isUpdate` specifies that the node produced by // the operator should be of type UpdateExpression rather than // simply UnaryExpression (`++` and `--`). // // `isAssign` marks all of `=`, `+=`, `-=` etcetera, which act as // binary operators with a very low precedence, that should result // in AssignmentExpression nodes. var _slash = {binop: 10, beforeExpr: true}, _eq = {isAssign: true, beforeExpr: true}; var _assign = {isAssign: true, beforeExpr: true}; var _incDec = {postfix: true, prefix: true, isUpdate: true}, _prefix = {prefix: true, beforeExpr: true}; var _logicalOR = {binop: 1, beforeExpr: true}; var _logicalAND = {binop: 2, beforeExpr: true}; var _bitwiseOR = {binop: 3, beforeExpr: true}; var _bitwiseXOR = {binop: 4, beforeExpr: true}; var _bitwiseAND = {binop: 5, beforeExpr: true}; var _equality = {binop: 6, beforeExpr: true}; var _relational = {binop: 7, beforeExpr: true}; var _bitShift = {binop: 8, beforeExpr: true}; var _plusMin = {binop: 9, prefix: true, beforeExpr: true}; var _modulo = {binop: 10, beforeExpr: true}; // '*' may be multiply or have special meaning in ES6 var _star = {binop: 10, beforeExpr: true}; // Provide access to the token types for external users of the // tokenizer. exports.tokTypes = {bracketL: _bracketL, bracketR: _bracketR, braceL: _braceL, braceR: _braceR, parenL: _parenL, parenR: _parenR, comma: _comma, semi: _semi, colon: _colon, dot: _dot, ellipsis: _ellipsis, question: _question, slash: _slash, eq: _eq, name: _name, eof: _eof, num: _num, regexp: _regexp, string: _string, arrow: _arrow, template: _template, templateContinued: _templateContinued, star: _star, assign: _assign}; for (var kw in keywordTypes) exports.tokTypes["_" + kw] = keywordTypes[kw]; // This is a trick taken from Esprima. It turns out that, on // non-Chrome browsers, to check whether a string is in a set, a // predicate containing a big ugly `switch` statement is faster than // a regular expression, and on Chrome the two are about on par. // This function uses `eval` (non-lexical) to produce such a // predicate from a space-separated string of words. // // It starts by sorting the words by length. function makePredicate(words) { words = words.split(" "); var f = "", cats = []; out: for (var i = 0; i < words.length; ++i) { for (var j = 0; j < cats.length; ++j) if (cats[j][0].length == words[i].length) { cats[j].push(words[i]); continue out; } cats.push([words[i]]); } function compareTo(arr) { if (arr.length == 1) return f += "return str === " + JSON.stringify(arr[0]) + ";"; f += "switch(str){"; for (var i = 0; i < arr.length; ++i) f += "case " + JSON.stringify(arr[i]) + ":"; f += "return true}return false;"; } // When there are more than three length categories, an outer // switch first dispatches on the lengths, to save on comparisons. if (cats.length > 3) { cats.sort(function(a, b) {return b.length - a.length;}); f += "switch(str.length){"; for (var i = 0; i < cats.length; ++i) { var cat = cats[i]; f += "case " + cat[0].length + ":"; compareTo(cat); } f += "}"; // Otherwise, simply generate a flat `switch` statement. } else { compareTo(words); } return new Function("str", f); } // The ECMAScript 3 reserved word list. var isReservedWord3 = makePredicate("abstract boolean byte char class double enum export extends final float goto implements import int interface long native package private protected public short static super synchronized throws transient volatile"); // ECMAScript 5 reserved words. var isReservedWord5 = makePredicate("class enum extends super const export import"); // The additional reserved words in strict mode. var isStrictReservedWord = makePredicate("implements interface let package private protected public static yield"); // The forbidden variable names in strict mode. var isStrictBadIdWord = makePredicate("eval arguments"); // And the keywords. var ecma5AndLessKeywords = "break case catch continue debugger default do else finally for function if return switch throw try var while with null true false instanceof typeof void delete new in this"; var isEcma5AndLessKeyword = makePredicate(ecma5AndLessKeywords); var isEcma6Keyword = makePredicate(ecma5AndLessKeywords + " let const class extends export import yield"); var isKeyword = isEcma5AndLessKeyword; // ## Character categories // Big ugly regular expressions that match characters in the // whitespace, identifier, and identifier-start categories. These // are only applied when a character is found to actually have a // code point above 128. // Generated by `tools/generate-identifier-regex.js`. var nonASCIIwhitespace = /[\u1680\u180e\u2000-\u200a\u202f\u205f\u3000\ufeff]/; var nonASCIIidentifierStartChars = "\xAA\xB5\xBA\xC0-\xD6\xD8-\xF6\xF8-\u02C1\u02C6-\u02D1\u02E0-\u02E4\u02EC\u02EE\u0370-\u0374\u0376\u0377\u037A-\u037D\u037F\u0386\u0388-\u038A\u038C\u038E-\u03A1\u03A3-\u03F5\u03F7-\u0481\u048A-\u052F\u0531-\u0556\u0559\u0561-\u0587\u05D0-\u05EA\u05F0-\u05F2\u0620-\u064A\u066E\u066F\u0671-\u06D3\u06D5\u06E5\u06E6\u06EE\u06EF\u06FA-\u06FC\u06FF\u0710\u0712-\u072F\u074D-\u07A5\u07B1\u07CA-\u07EA\u07F4\u07F5\u07FA\u0800-\u0815\u081A\u0824\u0828\u0840-\u0858\u08A0-\u08B2\u0904-\u0939\u093D\u0950\u0958-\u0961\u0971-\u0980\u0985-\u098C\u098F\u0990\u0993-\u09A8\u09AA-\u09B0\u09B2\u09B6-\u09B9\u09BD\u09CE\u09DC\u09DD\u09DF-\u09E1\u09F0\u09F1\u0A05-\u0A0A\u0A0F\u0A10\u0A13-\u0A28\u0A2A-\u0A30\u0A32\u0A33\u0A35\u0A36\u0A38\u0A39\u0A59-\u0A5C\u0A5E\u0A72-\u0A74\u0A85-\u0A8D\u0A8F-\u0A91\u0A93-\u0AA8\u0AAA-\u0AB0\u0AB2\u0AB3\u0AB5-\u0AB9\u0ABD\u0AD0\u0AE0\u0AE1\u0B05-\u0B0C\u0B0F\u0B10\u0B13-\u0B28\u0B2A-\u0B30\u0B32\u0B33\u0B35-\u0B39\u0B3D\u0B5C\u0B5D\u0B5F-\u0B61\u0B71\u0B83\u0B85-\u0B8A\u0B8E-\u0B90\u0B92-\u0B95\u0B99\u0B9A\u0B9C\u0B9E\u0B9F\u0BA3\u0BA4\u0BA8-\u0BAA\u0BAE-\u0BB9\u0BD0\u0C05-\u0C0C\u0C0E-\u0C10\u0C12-\u0C28\u0C2A-\u0C39\u0C3D\u0C58\u0C59\u0C60\u0C61\u0C85-\u0C8C\u0C8E-\u0C90\u0C92-\u0CA8\u0CAA-\u0CB3\u0CB5-\u0CB9\u0CBD\u0CDE\u0CE0\u0CE1\u0CF1\u0CF2\u0D05-\u0D0C\u0D0E-\u0D10\u0D12-\u0D3A\u0D3D\u0D4E\u0D60\u0D61\u0D7A-\u0D7F\u0D85-\u0D96\u0D9A-\u0DB1\u0DB3-\u0DBB\u0DBD\u0DC0-\u0DC6\u0E01-\u0E30\u0E32\u0E33\u0E40-\u0E46\u0E81\u0E82\u0E84\u0E87\u0E88\u0E8A\u0E8D\u0E94-\u0E97\u0E99-\u0E9F\u0EA1-\u0EA3\u0EA5\u0EA7\u0EAA\u0EAB\u0EAD-\u0EB0\u0EB2\u0EB3\u0EBD\u0EC0-\u0EC4\u0EC6\u0EDC-\u0EDF\u0F00\u0F40-\u0F47\u0F49-\u0F6C\u0F88-\u0F8C\u1000-\u102A\u103F\u1050-\u1055\u105A-\u105D\u1061\u1065\u1066\u106E-\u1070\u1075-\u1081\u108E\u10A0-\u10C5\u10C7\u10CD\u10D0-\u10FA\u10FC-\u1248\u124A-\u124D\u1250-\u1256\u1258\u125A-\u125D\u1260-\u1288\u128A-\u128D\u1290-\u12B0\u12B2-\u12B5\u12B8-\u12BE\u12C0\u12C2-\u12C5\u12C8-\u12D6\u12D8-\u1310\u1312-\u1315\u1318-\u135A\u1380-\u138F\u13A0-\u13F4\u1401-\u166C\u166F-\u167F\u1681-\u169A\u16A0-\u16EA\u16EE-\u16F8\u1700-\u170C\u170E-\u1711\u1720-\u1731\u1740-\u1751\u1760-\u176C\u176E-\u1770\u1780-\u17B3\u17D7\u17DC\u1820-\u1877\u1880-\u18A8\u18AA\u18B0-\u18F5\u1900-\u191E\u1950-\u196D\u1970-\u1974\u1980-\u19AB\u19C1-\u19C7\u1A00-\u1A16\u1A20-\u1A54\u1AA7\u1B05-\u1B33\u1B45-\u1B4B\u1B83-\u1BA0\u1BAE\u1BAF\u1BBA-\u1BE5\u1C00-\u1C23\u1C4D-\u1C4F\u1C5A-\u1C7D\u1CE9-\u1CEC\u1CEE-\u1CF1\u1CF5\u1CF6\u1D00-\u1DBF\u1E00-\u1F15\u1F18-\u1F1D\u1F20-\u1F45\u1F48-\u1F4D\u1F50-\u1F57\u1F59\u1F5B\u1F5D\u1F5F-\u1F7D\u1F80-\u1FB4\u1FB6-\u1FBC\u1FBE\u1FC2-\u1FC4\u1FC6-\u1FCC\u1FD0-\u1FD3\u1FD6-\u1FDB\u1FE0-\u1FEC\u1FF2-\u1FF4\u1FF6-\u1FFC\u2071\u207F\u2090-\u209C\u2102\u2107\u210A-\u2113\u2115\u2119-\u211D\u2124\u2126\u2128\u212A-\u212D\u212F-\u2139\u213C-\u213F\u2145-\u2149\u214E\u2160-\u2188\u2C00-\u2C2E\u2C30-\u2C5E\u2C60-\u2CE4\u2CEB-\u2CEE\u2CF2\u2CF3\u2D00-\u2D25\u2D27\u2D2D\u2D30-\u2D67\u2D6F\u2D80-\u2D96\u2DA0-\u2DA6\u2DA8-\u2DAE\u2DB0-\u2DB6\u2DB8-\u2DBE\u2DC0-\u2DC6\u2DC8-\u2DCE\u2DD0-\u2DD6\u2DD8-\u2DDE\u2E2F\u3005-\u3007\u3021-\u3029\u3031-\u3035\u3038-\u303C\u3041-\u3096\u309D-\u309F\u30A1-\u30FA\u30FC-\u30FF\u3105-\u312D\u3131-\u318E\u31A0-\u31BA\u31F0-\u31FF\u3400-\u4DB5\u4E00-\u9FCC\uA000-\uA48C\uA4D0-\uA4FD\uA500-\uA60C\uA610-\uA61F\uA62A\uA62B\uA640-\uA66E\uA67F-\uA69D\uA6A0-\uA6EF\uA717-\uA71F\uA722-\uA788\uA78B-\uA78E\uA790-\uA7AD\uA7B0\uA7B1\uA7F7-\uA801\uA803-\uA805\uA807-\uA80A\uA80C-\uA822\uA840-\uA873\uA882-\uA8B3\uA8F2-\uA8F7\uA8FB\uA90A-\uA925\uA930-\uA946\uA960-\uA97C\uA984-\uA9B2\uA9CF\uA9E0-\uA9E4\uA9E6-\uA9EF\uA9FA-\uA9FE\uAA00-\uAA28\uAA40-\uAA42\uAA44-\uAA4B\uAA60-\uAA76\uAA7A\uAA7E-\uAAAF\uAAB1\uAAB5\uAAB6\uAAB9-\uAABD\uAAC0\uAAC2\uAADB-\uAADD\uAAE0-\uAAEA\uAAF2-\uAAF4\uAB01-\uAB06\uAB09-\uAB0E\uAB11-\uAB16\uAB20-\uAB26\uAB28-\uAB2E\uAB30-\uAB5A\uAB5C-\uAB5F\uAB64\uAB65\uABC0-\uABE2\uAC00-\uD7A3\uD7B0-\uD7C6\uD7CB-\uD7FB\uF900-\uFA6D\uFA70-\uFAD9\uFB00-\uFB06\uFB13-\uFB17\uFB1D\uFB1F-\uFB28\uFB2A-\uFB36\uFB38-\uFB3C\uFB3E\uFB40\uFB41\uFB43\uFB44\uFB46-\uFBB1\uFBD3-\uFD3D\uFD50-\uFD8F\uFD92-\uFDC7\uFDF0-\uFDFB\uFE70-\uFE74\uFE76-\uFEFC\uFF21-\uFF3A\uFF41-\uFF5A\uFF66-\uFFBE\uFFC2-\uFFC7\uFFCA-\uFFCF\uFFD2-\uFFD7\uFFDA-\uFFDC"; var nonASCIIidentifierChars = "\u0300-\u036F\u0483-\u0487\u0591-\u05BD\u05BF\u05C1\u05C2\u05C4\u05C5\u05C7\u0610-\u061A\u064B-\u0669\u0670\u06D6-\u06DC\u06DF-\u06E4\u06E7\u06E8\u06EA-\u06ED\u06F0-\u06F9\u0711\u0730-\u074A\u07A6-\u07B0\u07C0-\u07C9\u07EB-\u07F3\u0816-\u0819\u081B-\u0823\u0825-\u0827\u0829-\u082D\u0859-\u085B\u08E4-\u0903\u093A-\u093C\u093E-\u094F\u0951-\u0957\u0962\u0963\u0966-\u096F\u0981-\u0983\u09BC\u09BE-\u09C4\u09C7\u09C8\u09CB-\u09CD\u09D7\u09E2\u09E3\u09E6-\u09EF\u0A01-\u0A03\u0A3C\u0A3E-\u0A42\u0A47\u0A48\u0A4B-\u0A4D\u0A51\u0A66-\u0A71\u0A75\u0A81-\u0A83\u0ABC\u0ABE-\u0AC5\u0AC7-\u0AC9\u0ACB-\u0ACD\u0AE2\u0AE3\u0AE6-\u0AEF\u0B01-\u0B03\u0B3C\u0B3E-\u0B44\u0B47\u0B48\u0B4B-\u0B4D\u0B56\u0B57\u0B62\u0B63\u0B66-\u0B6F\u0B82\u0BBE-\u0BC2\u0BC6-\u0BC8\u0BCA-\u0BCD\u0BD7\u0BE6-\u0BEF\u0C00-\u0C03\u0C3E-\u0C44\u0C46-\u0C48\u0C4A-\u0C4D\u0C55\u0C56\u0C62\u0C63\u0C66-\u0C6F\u0C81-\u0C83\u0CBC\u0CBE-\u0CC4\u0CC6-\u0CC8\u0CCA-\u0CCD\u0CD5\u0CD6\u0CE2\u0CE3\u0CE6-\u0CEF\u0D01-\u0D03\u0D3E-\u0D44\u0D46-\u0D48\u0D4A-\u0D4D\u0D57\u0D62\u0D63\u0D66-\u0D6F\u0D82\u0D83\u0DCA\u0DCF-\u0DD4\u0DD6\u0DD8-\u0DDF\u0DE6-\u0DEF\u0DF2\u0DF3\u0E31\u0E34-\u0E3A\u0E47-\u0E4E\u0E50-\u0E59\u0EB1\u0EB4-\u0EB9\u0EBB\u0EBC\u0EC8-\u0ECD\u0ED0-\u0ED9\u0F18\u0F19\u0F20-\u0F29\u0F35\u0F37\u0F39\u0F3E\u0F3F\u0F71-\u0F84\u0F86\u0F87\u0F8D-\u0F97\u0F99-\u0FBC\u0FC6\u102B-\u103E\u1040-\u1049\u1056-\u1059\u105E-\u1060\u1062-\u1064\u1067-\u106D\u1071-\u1074\u1082-\u108D\u108F-\u109D\u135D-\u135F\u1712-\u1714\u1732-\u1734\u1752\u1753\u1772\u1773\u17B4-\u17D3\u17DD\u17E0-\u17E9\u180B-\u180D\u1810-\u1819\u18A9\u1920-\u192B\u1930-\u193B\u1946-\u194F\u19B0-\u19C0\u19C8\u19C9\u19D0-\u19D9\u1A17-\u1A1B\u1A55-\u1A5E\u1A60-\u1A7C\u1A7F-\u1A89\u1A90-\u1A99\u1AB0-\u1ABD\u1B00-\u1B04\u1B34-\u1B44\u1B50-\u1B59\u1B6B-\u1B73\u1B80-\u1B82\u1BA1-\u1BAD\u1BB0-\u1BB9\u1BE6-\u1BF3\u1C24-\u1C37\u1C40-\u1C49\u1C50-\u1C59\u1CD0-\u1CD2\u1CD4-\u1CE8\u1CED\u1CF2-\u1CF4\u1CF8\u1CF9\u1DC0-\u1DF5\u1DFC-\u1DFF\u200C\u200D\u203F\u2040\u2054\u20D0-\u20DC\u20E1\u20E5-\u20F0\u2CEF-\u2CF1\u2D7F\u2DE0-\u2DFF\u302A-\u302F\u3099\u309A\uA620-\uA629\uA66F\uA674-\uA67D\uA69F\uA6F0\uA6F1\uA802\uA806\uA80B\uA823-\uA827\uA880\uA881\uA8B4-\uA8C4\uA8D0-\uA8D9\uA8E0-\uA8F1\uA900-\uA909\uA926-\uA92D\uA947-\uA953\uA980-\uA983\uA9B3-\uA9C0\uA9D0-\uA9D9\uA9E5\uA9F0-\uA9F9\uAA29-\uAA36\uAA43\uAA4C\uAA4D\uAA50-\uAA59\uAA7B-\uAA7D\uAAB0\uAAB2-\uAAB4\uAAB7\uAAB8\uAABE\uAABF\uAAC1\uAAEB-\uAAEF\uAAF5\uAAF6\uABE3-\uABEA\uABEC\uABED\uABF0-\uABF9\uFB1E\uFE00-\uFE0F\uFE20-\uFE2D\uFE33\uFE34\uFE4D-\uFE4F\uFF10-\uFF19\uFF3F"; var nonASCIIidentifierStart = new RegExp("[" + nonASCIIidentifierStartChars + "]"); var nonASCIIidentifier = new RegExp("[" + nonASCIIidentifierStartChars + nonASCIIidentifierChars + "]"); // Whether a single character denotes a newline. var newline = /[\n\r\u2028\u2029]/; function isNewLine(code) { return code === 10 || code === 13 || code === 0x2028 || code == 0x2029; } // Matches a whole line break (where CRLF is considered a single // line break). Used to count lines. var lineBreak = /\r\n|[\n\r\u2028\u2029]/g; // Test whether a given character code starts an identifier. var isIdentifierStart = exports.isIdentifierStart = function(code) { if (code < 65) return code === 36; if (code < 91) return true; if (code < 97) return code === 95; if (code < 123)return true; return code >= 0xaa && nonASCIIidentifierStart.test(String.fromCharCode(code)); }; // Test whether a given character is part of an identifier. var isIdentifierChar = exports.isIdentifierChar = function(code) { if (code < 48) return code === 36; if (code < 58) return true; if (code < 65) return false; if (code < 91) return true; if (code < 97) return code === 95; if (code < 123)return true; return code >= 0xaa && nonASCIIidentifier.test(String.fromCharCode(code)); }; // ## Tokenizer // These are used when `options.locations` is on, for the // `tokStartLoc` and `tokEndLoc` properties. function Position(line, col) { this.line = line; this.column = col; } Position.prototype.offset = function(n) { return new Position(this.line, this.column + n); } function curPosition() { return new Position(tokCurLine, tokPos - tokLineStart); } // Reset the token state. Used at the start of a parse. function initTokenState(pos) { if (pos) { tokPos = pos; tokLineStart = Math.max(0, input.lastIndexOf("\n", pos)); tokCurLine = input.slice(0, tokLineStart).split(newline).length; } else { tokCurLine = 1; tokPos = tokLineStart = 0; } tokRegexpAllowed = true; metParenL = 0; templates = []; } // Called at the end of every token. Sets `tokEnd`, `tokVal`, and // `tokRegexpAllowed`, and skips the space after the token, so that // the next one's `tokStart` will point at the right position. function finishToken(type, val, shouldSkipSpace) { tokEnd = tokPos; if (options.locations) tokEndLoc = curPosition(); tokType = type; if (shouldSkipSpace !== false) skipSpace(); tokVal = val; tokRegexpAllowed = type.beforeExpr; if (options.onToken) { options.onToken(new Token()); } } function skipBlockComment() { var startLoc = options.onComment && options.locations && curPosition(); var start = tokPos, end = input.indexOf("*/", tokPos += 2); if (end === -1) raise(tokPos - 2, "Unterminated comment"); tokPos = end + 2; if (options.locations) { lineBreak.lastIndex = start; var match; while ((match = lineBreak.exec(input)) && match.index < tokPos) { ++tokCurLine; tokLineStart = match.index + match[0].length; } } if (options.onComment) options.onComment(true, input.slice(start + 2, end), start, tokPos, startLoc, options.locations && curPosition()); } function skipLineComment(startSkip) { var start = tokPos; var startLoc = options.onComment && options.locations && curPosition(); var ch = input.charCodeAt(tokPos+=startSkip); while (tokPos < inputLen && ch !== 10 && ch !== 13 && ch !== 8232 && ch !== 8233) { ++tokPos; ch = input.charCodeAt(tokPos); } if (options.onComment) options.onComment(false, input.slice(start + startSkip, tokPos), start, tokPos, startLoc, options.locations && curPosition()); } // Called at the start of the parse and after every token. Skips // whitespace and comments, and. function skipSpace() { while (tokPos < inputLen) { var ch = input.charCodeAt(tokPos); if (ch === 32) { // ' ' ++tokPos; } else if (ch === 13) { ++tokPos; var next = input.charCodeAt(tokPos); if (next === 10) { ++tokPos; } if (options.locations) { ++tokCurLine; tokLineStart = tokPos; } } else if (ch === 10 || ch === 8232 || ch === 8233) { ++tokPos; if (options.locations) { ++tokCurLine; tokLineStart = tokPos; } } else if (ch > 8 && ch < 14) { ++tokPos; } else if (ch === 47) { // '/' var next = input.charCodeAt(tokPos + 1); if (next === 42) { // '*' skipBlockComment(); } else if (next === 47) { // '/' skipLineComment(2); } else break; } else if (ch === 160) { // '\xa0' ++tokPos; } else if (ch >= 5760 && nonASCIIwhitespace.test(String.fromCharCode(ch))) { ++tokPos; } else { break; } } } // ### Token reading // This is the function that is called to fetch the next token. It // is somewhat obscure, because it works in character codes rather // than characters, and because operator parsing has been inlined // into it. // // All in the name of speed. // // The `forceRegexp` parameter is used in the one case where the // `tokRegexpAllowed` trick does not work. See `parseStatement`. function readToken_dot() { var next = input.charCodeAt(tokPos + 1); if (next >= 48 && next <= 57) return readNumber(true); var next2 = input.charCodeAt(tokPos + 2); if (options.ecmaVersion >= 6 && next === 46 && next2 === 46) { // 46 = dot '.' tokPos += 3; return finishToken(_ellipsis); } else { ++tokPos; return finishToken(_dot); } } function readToken_slash() { // '/' var next = input.charCodeAt(tokPos + 1); if (tokRegexpAllowed) {++tokPos; return readRegexp();} if (next === 61) return finishOp(_assign, 2); return finishOp(_slash, 1); } function readToken_mult_modulo(code) { // '%*' var next = input.charCodeAt(tokPos + 1); if (next === 61) return finishOp(_assign, 2); return finishOp(code === 42 ? _star : _modulo, 1); } function readToken_pipe_amp(code) { // '|&' var next = input.charCodeAt(tokPos + 1); if (next === code) return finishOp(code === 124 ? _logicalOR : _logicalAND, 2); if (next === 61) return finishOp(_assign, 2); return finishOp(code === 124 ? _bitwiseOR : _bitwiseAND, 1); } function readToken_caret() { // '^' var next = input.charCodeAt(tokPos + 1); if (next === 61) return finishOp(_assign, 2); return finishOp(_bitwiseXOR, 1); } function readToken_plus_min(code) { // '+-' var next = input.charCodeAt(tokPos + 1); if (next === code) { if (next == 45 && input.charCodeAt(tokPos + 2) == 62 && newline.test(input.slice(lastEnd, tokPos))) { // A `-->` line comment skipLineComment(3); skipSpace(); return readToken(); } return finishOp(_incDec, 2); } if (next === 61) return finishOp(_assign, 2); return finishOp(_plusMin, 1); } function readToken_lt_gt(code) { // '<>' var next = input.charCodeAt(tokPos + 1); var size = 1; if (next === code) { size = code === 62 && input.charCodeAt(tokPos + 2) === 62 ? 3 : 2; if (input.charCodeAt(tokPos + size) === 61) return finishOp(_assign, size + 1); return finishOp(_bitShift, size); } if (next == 33 && code == 60 && input.charCodeAt(tokPos + 2) == 45 && input.charCodeAt(tokPos + 3) == 45) { // `