diff options
author | Matt A. Tobin <mattatobin@localhost.localdomain> | 2018-02-02 04:16:08 -0500 |
---|---|---|
committer | Matt A. Tobin <mattatobin@localhost.localdomain> | 2018-02-02 04:16:08 -0500 |
commit | 5f8de423f190bbb79a62f804151bc24824fa32d8 (patch) | |
tree | 10027f336435511475e392454359edea8e25895d /toolkit/crashreporter/tools | |
parent | 49ee0794b5d912db1f95dce6eb52d781dc210db5 (diff) | |
download | UXP-5f8de423f190bbb79a62f804151bc24824fa32d8.tar UXP-5f8de423f190bbb79a62f804151bc24824fa32d8.tar.gz UXP-5f8de423f190bbb79a62f804151bc24824fa32d8.tar.lz UXP-5f8de423f190bbb79a62f804151bc24824fa32d8.tar.xz UXP-5f8de423f190bbb79a62f804151bc24824fa32d8.zip |
Add m-esr52 at 52.6.0
Diffstat (limited to 'toolkit/crashreporter/tools')
-rwxr-xr-x | toolkit/crashreporter/tools/symbolstore.py | 1078 | ||||
-rw-r--r-- | toolkit/crashreporter/tools/unit-symbolstore.py | 583 | ||||
-rw-r--r-- | toolkit/crashreporter/tools/upload_symbols.py | 102 | ||||
-rwxr-xr-x | toolkit/crashreporter/tools/win32/dump_syms_vc1600.exe | bin | 0 -> 55296 bytes | |||
-rw-r--r-- | toolkit/crashreporter/tools/win32/dump_syms_vc1700.exe | bin | 0 -> 52736 bytes | |||
-rw-r--r-- | toolkit/crashreporter/tools/win32/dump_syms_vc1800.exe | bin | 0 -> 51200 bytes |
6 files changed, 1763 insertions, 0 deletions
diff --git a/toolkit/crashreporter/tools/symbolstore.py b/toolkit/crashreporter/tools/symbolstore.py new file mode 100755 index 000000000..15eb5e112 --- /dev/null +++ b/toolkit/crashreporter/tools/symbolstore.py @@ -0,0 +1,1078 @@ +#!/bin/env python +# This Source Code Form is subject to the terms of the Mozilla Public +# License, v. 2.0. If a copy of the MPL was not distributed with this +# file, You can obtain one at http://mozilla.org/MPL/2.0/. +# +# Usage: symbolstore.py <params> <dump_syms path> <symbol store path> +# <debug info files or dirs> +# Runs dump_syms on each debug info file specified on the command line, +# then places the resulting symbol file in the proper directory +# structure in the symbol store path. Accepts multiple files +# on the command line, so can be called as part of a pipe using +# find <dir> | xargs symbolstore.pl <dump_syms> <storepath> +# But really, you might just want to pass it <dir>. +# +# Parameters accepted: +# -c : Copy debug info files to the same directory structure +# as sym files. On Windows, this will also copy +# binaries into the symbol store. +# -a "<archs>" : Run dump_syms -a <arch> for each space separated +# cpu architecture in <archs> (only on OS X) +# -s <srcdir> : Use <srcdir> as the top source directory to +# generate relative filenames. + +import buildconfig +import errno +import sys +import platform +import os +import re +import shutil +import textwrap +import fnmatch +import subprocess +import time +import ctypes +import urlparse +import concurrent.futures +import multiprocessing + +from optparse import OptionParser +from xml.dom.minidom import parse + +from mozpack.copier import FileRegistry +from mozpack.manifests import ( + InstallManifest, + UnreadableInstallManifest, +) + +# Utility classes + +class VCSFileInfo: + """ A base class for version-controlled file information. Ensures that the + following attributes are generated only once (successfully): + + self.root + self.clean_root + self.revision + self.filename + + The attributes are generated by a single call to the GetRoot, + GetRevision, and GetFilename methods. Those methods are explicitly not + implemented here and must be implemented in derived classes. """ + + def __init__(self, file): + if not file: + raise ValueError + self.file = file + + def __getattr__(self, name): + """ __getattr__ is only called for attributes that are not set on self, + so setting self.[attr] will prevent future calls to the GetRoot, + GetRevision, and GetFilename methods. We don't set the values on + failure on the off chance that a future call might succeed. """ + + if name == "root": + root = self.GetRoot() + if root: + self.root = root + return root + + elif name == "clean_root": + clean_root = self.GetCleanRoot() + if clean_root: + self.clean_root = clean_root + return clean_root + + elif name == "revision": + revision = self.GetRevision() + if revision: + self.revision = revision + return revision + + elif name == "filename": + filename = self.GetFilename() + if filename: + self.filename = filename + return filename + + raise AttributeError + + def GetRoot(self): + """ This method should return the unmodified root for the file or 'None' + on failure. """ + raise NotImplementedError + + def GetCleanRoot(self): + """ This method should return the repository root for the file or 'None' + on failure. """ + raise NotImplementedError + + def GetRevision(self): + """ This method should return the revision number for the file or 'None' + on failure. """ + raise NotImplementedError + + def GetFilename(self): + """ This method should return the repository-specific filename for the + file or 'None' on failure. """ + raise NotImplementedError + + +# This regex separates protocol and optional username/password from a url. +# For instance, all the following urls will be transformed into +# 'foo.com/bar': +# +# http://foo.com/bar +# svn+ssh://user@foo.com/bar +# svn+ssh://user:pass@foo.com/bar +# +rootRegex = re.compile(r'^\S+?:/+(?:[^\s/]*@)?(\S+)$') + +def read_output(*args): + (stdout, _) = subprocess.Popen(args=args, stdout=subprocess.PIPE).communicate() + return stdout.rstrip() + +class HGRepoInfo: + def __init__(self, path): + self.path = path + + rev = os.environ.get('MOZ_SOURCE_CHANGESET') + if not rev: + rev = read_output('hg', '-R', path, + 'parent', '--template={node|short}') + + # Look for the default hg path. If MOZ_SOURCE_REPO is set, we + # don't bother asking hg. + hg_root = os.environ.get('MOZ_SOURCE_REPO') + if hg_root: + root = hg_root + else: + root = read_output('hg', '-R', path, + 'showconfig', 'paths.default') + if not root: + print >> sys.stderr, "Failed to get HG Repo for %s" % path + cleanroot = None + if root: + match = rootRegex.match(root) + if match: + cleanroot = match.group(1) + if cleanroot.endswith('/'): + cleanroot = cleanroot[:-1] + if cleanroot is None: + print >> sys.stderr, textwrap.dedent("""\ + Could not determine repo info for %s. This is either not a clone of the web-based + repository, or you have not specified MOZ_SOURCE_REPO, or the clone is corrupt.""") % path + sys.exit(1) + self.rev = rev + self.root = root + self.cleanroot = cleanroot + + def GetFileInfo(self, file): + return HGFileInfo(file, self) + +class HGFileInfo(VCSFileInfo): + def __init__(self, file, repo): + VCSFileInfo.__init__(self, file) + self.repo = repo + self.file = os.path.relpath(file, repo.path) + + def GetRoot(self): + return self.repo.root + + def GetCleanRoot(self): + return self.repo.cleanroot + + def GetRevision(self): + return self.repo.rev + + def GetFilename(self): + if self.revision and self.clean_root: + return "hg:%s:%s:%s" % (self.clean_root, self.file, self.revision) + return self.file + +class GitRepoInfo: + """ + Info about a local git repository. Does not currently + support discovering info about a git clone, the info must be + provided out-of-band. + """ + def __init__(self, path, rev, root): + self.path = path + cleanroot = None + if root: + match = rootRegex.match(root) + if match: + cleanroot = match.group(1) + if cleanroot.endswith('/'): + cleanroot = cleanroot[:-1] + if cleanroot is None: + print >> sys.stderr, textwrap.dedent("""\ + Could not determine repo info for %s (%s). This is either not a clone of a web-based + repository, or you have not specified MOZ_SOURCE_REPO, or the clone is corrupt.""") % (path, root) + sys.exit(1) + self.rev = rev + self.cleanroot = cleanroot + + def GetFileInfo(self, file): + return GitFileInfo(file, self) + +class GitFileInfo(VCSFileInfo): + def __init__(self, file, repo): + VCSFileInfo.__init__(self, file) + self.repo = repo + self.file = os.path.relpath(file, repo.path) + + def GetRoot(self): + return self.repo.path + + def GetCleanRoot(self): + return self.repo.cleanroot + + def GetRevision(self): + return self.repo.rev + + def GetFilename(self): + if self.revision and self.clean_root: + return "git:%s:%s:%s" % (self.clean_root, self.file, self.revision) + return self.file + +# Utility functions + +# A cache of files for which VCS info has already been determined. Used to +# prevent extra filesystem activity or process launching. +vcsFileInfoCache = {} + +def IsInDir(file, dir): + # the lower() is to handle win32+vc8, where + # the source filenames come out all lowercase, + # but the srcdir can be mixed case + return os.path.abspath(file).lower().startswith(os.path.abspath(dir).lower()) + +def GetVCSFilenameFromSrcdir(file, srcdir): + if srcdir not in Dumper.srcdirRepoInfo: + # Not in cache, so find it adnd cache it + if os.path.isdir(os.path.join(srcdir, '.hg')): + Dumper.srcdirRepoInfo[srcdir] = HGRepoInfo(srcdir) + else: + # Unknown VCS or file is not in a repo. + return None + return Dumper.srcdirRepoInfo[srcdir].GetFileInfo(file) + +def GetVCSFilename(file, srcdirs): + """Given a full path to a file, and the top source directory, + look for version control information about this file, and return + a tuple containing + 1) a specially formatted filename that contains the VCS type, + VCS location, relative filename, and revision number, formatted like: + vcs:vcs location:filename:revision + For example: + cvs:cvs.mozilla.org/cvsroot:mozilla/browser/app/nsBrowserApp.cpp:1.36 + 2) the unmodified root information if it exists""" + (path, filename) = os.path.split(file) + if path == '' or filename == '': + return (file, None) + + fileInfo = None + root = '' + if file in vcsFileInfoCache: + # Already cached this info, use it. + fileInfo = vcsFileInfoCache[file] + else: + for srcdir in srcdirs: + if not IsInDir(file, srcdir): + continue + fileInfo = GetVCSFilenameFromSrcdir(file, srcdir) + if fileInfo: + vcsFileInfoCache[file] = fileInfo + break + + if fileInfo: + file = fileInfo.filename + root = fileInfo.root + + # we want forward slashes on win32 paths + return (file.replace("\\", "/"), root) + +def validate_install_manifests(install_manifest_args): + args = [] + for arg in install_manifest_args: + bits = arg.split(',') + if len(bits) != 2: + raise ValueError('Invalid format for --install-manifest: ' + 'specify manifest,target_dir') + manifest_file, destination = map(os.path.abspath, bits) + if not os.path.isfile(manifest_file): + raise IOError(errno.ENOENT, 'Manifest file not found', + manifest_file) + if not os.path.isdir(destination): + raise IOError(errno.ENOENT, 'Install directory not found', + destination) + try: + manifest = InstallManifest(manifest_file) + except UnreadableInstallManifest: + raise IOError(errno.EINVAL, 'Error parsing manifest file', + manifest_file) + args.append((manifest, destination)) + return args + +def make_file_mapping(install_manifests): + file_mapping = {} + for manifest, destination in install_manifests: + destination = os.path.abspath(destination) + reg = FileRegistry() + manifest.populate_registry(reg) + for dst, src in reg: + if hasattr(src, 'path'): + abs_dest = os.path.normpath(os.path.join(destination, dst)) + file_mapping[abs_dest] = src.path + return file_mapping + +def GetPlatformSpecificDumper(**kwargs): + """This function simply returns a instance of a subclass of Dumper + that is appropriate for the current platform.""" + return {'WINNT': Dumper_Win32, + 'Linux': Dumper_Linux, + 'Darwin': Dumper_Mac}[buildconfig.substs['OS_ARCH']](**kwargs) + +def SourceIndex(fileStream, outputPath, vcs_root): + """Takes a list of files, writes info to a data block in a .stream file""" + # Creates a .pdb.stream file in the mozilla\objdir to be used for source indexing + # Create the srcsrv data block that indexes the pdb file + result = True + pdbStreamFile = open(outputPath, "w") + pdbStreamFile.write('''SRCSRV: ini ------------------------------------------------\r\nVERSION=2\r\nINDEXVERSION=2\r\nVERCTRL=http\r\nSRCSRV: variables ------------------------------------------\r\nHGSERVER=''') + pdbStreamFile.write(vcs_root) + pdbStreamFile.write('''\r\nSRCSRVVERCTRL=http\r\nHTTP_EXTRACT_TARGET=%hgserver%/raw-file/%var3%/%var2%\r\nSRCSRVTRG=%http_extract_target%\r\nSRCSRV: source files ---------------------------------------\r\n''') + pdbStreamFile.write(fileStream) # can't do string interpolation because the source server also uses this and so there are % in the above + pdbStreamFile.write("SRCSRV: end ------------------------------------------------\r\n\n") + pdbStreamFile.close() + return result + +def StartJob(dumper, lock, srcdirRepoInfo, func_name, args): + # Windows worker processes won't have run GlobalInit, + # and due to a lack of fork(), won't inherit the class + # variables from the parent, so set them here. + Dumper.lock = lock + Dumper.srcdirRepoInfo = srcdirRepoInfo + return getattr(dumper, func_name)(*args) + +class JobPool(object): + jobs = {} + executor = None + + @classmethod + def init(cls, executor): + cls.executor = executor + + @classmethod + def shutdown(cls): + cls.executor.shutdown() + + @classmethod + def submit(cls, args, callback): + cls.jobs[cls.executor.submit(StartJob, *args)] = callback + + @classmethod + def as_completed(cls): + '''Like concurrent.futures.as_completed, but allows adding new futures + between generator steps. Iteration will end when the generator has + yielded all completed futures and JobQueue.jobs is empty. + Yields (future, callback) pairs. + ''' + while cls.jobs: + completed, _ = concurrent.futures.wait(cls.jobs.keys(), return_when=concurrent.futures.FIRST_COMPLETED) + for f in completed: + callback = cls.jobs[f] + del cls.jobs[f] + yield f, callback + +class Dumper: + """This class can dump symbols from a file with debug info, and + store the output in a directory structure that is valid for use as + a Breakpad symbol server. Requires a path to a dump_syms binary-- + |dump_syms| and a directory to store symbols in--|symbol_path|. + Optionally takes a list of processor architectures to process from + each debug file--|archs|, the full path to the top source + directory--|srcdir|, for generating relative source file names, + and an option to copy debug info files alongside the dumped + symbol files--|copy_debug|, mostly useful for creating a + Microsoft Symbol Server from the resulting output. + + You don't want to use this directly if you intend to process files. + Instead, call GetPlatformSpecificDumper to get an instance of a + subclass. + + Processing is performed asynchronously via worker processes; in + order to wait for processing to finish and cleanup correctly, you + must call Finish after all ProcessFiles calls have been made. + You must also call Dumper.GlobalInit before creating or using any + instances.""" + def __init__(self, dump_syms, symbol_path, + archs=None, + srcdirs=[], + copy_debug=False, + vcsinfo=False, + srcsrv=False, + exclude=[], + repo_manifest=None, + file_mapping=None): + # popen likes absolute paths, at least on windows + self.dump_syms = os.path.abspath(dump_syms) + self.symbol_path = symbol_path + if archs is None: + # makes the loop logic simpler + self.archs = [''] + else: + self.archs = ['-a %s' % a for a in archs.split()] + self.srcdirs = [os.path.normpath(a) for a in srcdirs] + self.copy_debug = copy_debug + self.vcsinfo = vcsinfo + self.srcsrv = srcsrv + self.exclude = exclude[:] + if repo_manifest: + self.parse_repo_manifest(repo_manifest) + self.file_mapping = file_mapping or {} + + # book-keeping to keep track of the cleanup work per file tuple + self.files_record = {} + + @classmethod + def GlobalInit(cls, executor=concurrent.futures.ProcessPoolExecutor): + """Initialize the class globals for the multiprocessing setup; must + be called before any Dumper instances are created and used. Test cases + may pass in a different executor to use, usually + concurrent.futures.ThreadPoolExecutor.""" + num_cpus = multiprocessing.cpu_count() + if num_cpus is None: + # assume a dual core machine if we can't find out for some reason + # probably better on single core anyway due to I/O constraints + num_cpus = 2 + + # have to create any locks etc before the pool + manager = multiprocessing.Manager() + cls.lock = manager.RLock() + cls.srcdirRepoInfo = manager.dict() + JobPool.init(executor(max_workers=num_cpus)) + + def output(self, dest, output_str): + """Writes |output_str| to |dest|, holding |lock|; + terminates with a newline.""" + with Dumper.lock: + dest.write(output_str + "\n") + dest.flush() + + def output_pid(self, dest, output_str): + """Debugging output; prepends the pid to the string.""" + self.output(dest, "%d: %s" % (os.getpid(), output_str)) + + def parse_repo_manifest(self, repo_manifest): + """ + Parse an XML manifest of repository info as produced + by the `repo manifest -r` command. + """ + doc = parse(repo_manifest) + if doc.firstChild.tagName != "manifest": + return + # First, get remotes. + def ensure_slash(u): + if not u.endswith("/"): + return u + "/" + return u + remotes = dict([(r.getAttribute("name"), ensure_slash(r.getAttribute("fetch"))) for r in doc.getElementsByTagName("remote")]) + # And default remote. + default_remote = None + if doc.getElementsByTagName("default"): + default_remote = doc.getElementsByTagName("default")[0].getAttribute("remote") + # Now get projects. Assume they're relative to repo_manifest. + base_dir = os.path.abspath(os.path.dirname(repo_manifest)) + for proj in doc.getElementsByTagName("project"): + # name is the repository URL relative to the remote path. + name = proj.getAttribute("name") + # path is the path on-disk, relative to the manifest file. + path = proj.getAttribute("path") + # revision is the changeset ID. + rev = proj.getAttribute("revision") + # remote is the base URL to use. + remote = proj.getAttribute("remote") + # remote defaults to the <default remote>. + if not remote: + remote = default_remote + # path defaults to name. + if not path: + path = name + if not (name and path and rev and remote): + print "Skipping project %s" % proj.toxml() + continue + remote = remotes[remote] + # Turn git URLs into http URLs so that urljoin works. + if remote.startswith("git:"): + remote = "http" + remote[3:] + # Add this project to srcdirs. + srcdir = os.path.join(base_dir, path) + self.srcdirs.append(srcdir) + # And cache its VCS file info. Currently all repos mentioned + # in a repo manifest are assumed to be git. + root = urlparse.urljoin(remote, name) + Dumper.srcdirRepoInfo[srcdir] = GitRepoInfo(srcdir, rev, root) + + # subclasses override this + def ShouldProcess(self, file): + return not any(fnmatch.fnmatch(os.path.basename(file), exclude) for exclude in self.exclude) + + # and can override this + def ShouldSkipDir(self, dir): + return False + + def RunFileCommand(self, file): + """Utility function, returns the output of file(1)""" + try: + # we use -L to read the targets of symlinks, + # and -b to print just the content, not the filename + return os.popen("file -Lb " + file).read() + except: + return "" + + # This is a no-op except on Win32 + def FixFilenameCase(self, file): + return file + + # This is a no-op except on Win32 + def SourceServerIndexing(self, debug_file, guid, sourceFileStream, vcs_root): + return "" + + # subclasses override this if they want to support this + def CopyDebug(self, file, debug_file, guid, code_file, code_id): + pass + + def Finish(self, stop_pool=True): + '''Process all pending jobs and any jobs their callbacks submit. + By default, will shutdown the executor, but for testcases that + need multiple runs, pass stop_pool = False.''' + for job, callback in JobPool.as_completed(): + try: + res = job.result() + except Exception as e: + self.output(sys.stderr, 'Job raised exception: %s' % e) + continue + callback(res) + if stop_pool: + JobPool.shutdown() + + def Process(self, *args): + """Process files recursively in args.""" + # We collect all files to process first then sort by size to schedule + # larger files first because larger files tend to take longer and we + # don't like long pole stragglers. + files = set() + for arg in args: + for f in self.get_files_to_process(arg): + files.add(f) + + for f in sorted(files, key=os.path.getsize, reverse=True): + self.ProcessFiles((f,)) + + def get_files_to_process(self, file_or_dir): + """Generate the files to process from an input.""" + if os.path.isdir(file_or_dir) and not self.ShouldSkipDir(file_or_dir): + for f in self.get_files_to_process_in_dir(file_or_dir): + yield f + elif os.path.isfile(file_or_dir): + yield file_or_dir + + def get_files_to_process_in_dir(self, path): + """Generate the files to process in a directory. + + Valid files are are determined by calling ShouldProcess. + """ + for root, dirs, files in os.walk(path): + for d in dirs[:]: + if self.ShouldSkipDir(d): + dirs.remove(d) + for f in files: + fullpath = os.path.join(root, f) + if self.ShouldProcess(fullpath): + yield fullpath + + def SubmitJob(self, file_key, func_name, args, callback): + """Submits a job to the pool of workers""" + JobPool.submit((self, Dumper.lock, Dumper.srcdirRepoInfo, func_name, args), callback) + + def ProcessFilesFinished(self, res): + """Callback from multiprocesing when ProcessFilesWork finishes; + run the cleanup work, if any""" + # only run the cleanup function once per tuple of files + self.files_record[res['files']] += 1 + if self.files_record[res['files']] == len(self.archs): + del self.files_record[res['files']] + if res['after']: + res['after'](res['status'], res['after_arg']) + + def ProcessFiles(self, files, after=None, after_arg=None): + """Dump symbols from these files into a symbol file, stored + in the proper directory structure in |symbol_path|; processing is performed + asynchronously, and Finish must be called to wait for it complete and cleanup. + All files after the first are fallbacks in case the first file does not process + successfully; if it does, no other files will be touched.""" + self.output_pid(sys.stderr, "Submitting jobs for files: %s" % str(files)) + + # tries to get the vcs root from the .mozconfig first - if it's not set + # the tinderbox vcs path will be assigned further down + vcs_root = os.environ.get('MOZ_SOURCE_REPO') + for arch_num, arch in enumerate(self.archs): + self.files_record[files] = 0 # record that we submitted jobs for this tuple of files + self.SubmitJob(files[-1], 'ProcessFilesWork', args=(files, arch_num, arch, vcs_root, after, after_arg), callback=self.ProcessFilesFinished) + + def dump_syms_cmdline(self, file, arch, files): + ''' + Get the commandline used to invoke dump_syms. + ''' + # The Mac dumper overrides this. + return [self.dump_syms, file] + + def ProcessFilesWork(self, files, arch_num, arch, vcs_root, after, after_arg): + t_start = time.time() + self.output_pid(sys.stderr, "Worker processing files: %s" % (files,)) + + # our result is a status, a cleanup function, an argument to that function, and the tuple of files we were called on + result = { 'status' : False, 'after' : after, 'after_arg' : after_arg, 'files' : files } + + sourceFileStream = '' + code_id, code_file = None, None + for file in files: + # files is a tuple of files, containing fallbacks in case the first file doesn't process successfully + try: + cmd = self.dump_syms_cmdline(file, arch, files) + self.output_pid(sys.stderr, ' '.join(cmd)) + proc = subprocess.Popen(cmd, stdout=subprocess.PIPE, + stderr=open(os.devnull, 'wb')) + module_line = proc.stdout.next() + if module_line.startswith("MODULE"): + # MODULE os cpu guid debug_file + (guid, debug_file) = (module_line.split())[3:5] + # strip off .pdb extensions, and append .sym + sym_file = re.sub("\.pdb$", "", debug_file) + ".sym" + # we do want forward slashes here + rel_path = os.path.join(debug_file, + guid, + sym_file).replace("\\", "/") + full_path = os.path.normpath(os.path.join(self.symbol_path, + rel_path)) + try: + os.makedirs(os.path.dirname(full_path)) + except OSError: # already exists + pass + f = open(full_path, "w") + f.write(module_line) + # now process the rest of the output + for line in proc.stdout: + if line.startswith("FILE"): + # FILE index filename + (x, index, filename) = line.rstrip().split(None, 2) + filename = os.path.normpath(self.FixFilenameCase(filename)) + # We want original file paths for the source server. + sourcepath = filename + if filename in self.file_mapping: + filename = self.file_mapping[filename] + if self.vcsinfo: + (filename, rootname) = GetVCSFilename(filename, self.srcdirs) + # sets vcs_root in case the loop through files were to end on an empty rootname + if vcs_root is None: + if rootname: + vcs_root = rootname + # gather up files with hg for indexing + if filename.startswith("hg"): + (ver, checkout, source_file, revision) = filename.split(":", 3) + sourceFileStream += sourcepath + "*" + source_file + '*' + revision + "\r\n" + f.write("FILE %s %s\n" % (index, filename)) + elif line.startswith("INFO CODE_ID "): + # INFO CODE_ID code_id code_file + # This gives some info we can use to + # store binaries in the symbol store. + bits = line.rstrip().split(None, 3) + if len(bits) == 4: + code_id, code_file = bits[2:] + f.write(line) + else: + # pass through all other lines unchanged + f.write(line) + # we want to return true only if at least one line is not a MODULE or FILE line + result['status'] = True + f.close() + proc.wait() + # we output relative paths so callers can get a list of what + # was generated + self.output(sys.stdout, rel_path) + if self.srcsrv and vcs_root: + # add source server indexing to the pdb file + self.SourceServerIndexing(file, guid, sourceFileStream, vcs_root) + # only copy debug the first time if we have multiple architectures + if self.copy_debug and arch_num == 0: + self.CopyDebug(file, debug_file, guid, + code_file, code_id) + except StopIteration: + pass + except Exception as e: + self.output(sys.stderr, "Unexpected error: %s" % (str(e),)) + raise + if result['status']: + # we only need 1 file to work + break + + elapsed = time.time() - t_start + self.output_pid(sys.stderr, 'Worker finished processing %s in %.2fs' % + (files, elapsed)) + return result + +# Platform-specific subclasses. For the most part, these just have +# logic to determine what files to extract symbols from. + +class Dumper_Win32(Dumper): + fixedFilenameCaseCache = {} + + def ShouldProcess(self, file): + """This function will allow processing of pdb files that have dll + or exe files with the same base name next to them.""" + if not Dumper.ShouldProcess(self, file): + return False + if file.endswith(".pdb"): + (path,ext) = os.path.splitext(file) + if os.path.isfile(path + ".exe") or os.path.isfile(path + ".dll"): + return True + return False + + def FixFilenameCase(self, file): + """Recent versions of Visual C++ put filenames into + PDB files as all lowercase. If the file exists + on the local filesystem, fix it.""" + + # Use a cached version if we have one. + if file in self.fixedFilenameCaseCache: + return self.fixedFilenameCaseCache[file] + + result = file + + ctypes.windll.kernel32.SetErrorMode(ctypes.c_uint(1)) + if not isinstance(file, unicode): + file = unicode(file, sys.getfilesystemencoding()) + handle = ctypes.windll.kernel32.CreateFileW(file, + # GENERIC_READ + 0x80000000, + # FILE_SHARE_READ + 1, + None, + # OPEN_EXISTING + 3, + 0, + None) + if handle != -1: + size = ctypes.windll.kernel32.GetFinalPathNameByHandleW(handle, + None, + 0, + 0) + buf = ctypes.create_unicode_buffer(size) + if ctypes.windll.kernel32.GetFinalPathNameByHandleW(handle, + buf, + size, + 0) > 0: + # The return value of GetFinalPathNameByHandleW uses the + # '\\?\' prefix. + result = buf.value.encode(sys.getfilesystemencoding())[4:] + ctypes.windll.kernel32.CloseHandle(handle) + + # Cache the corrected version to avoid future filesystem hits. + self.fixedFilenameCaseCache[file] = result + return result + + def CopyDebug(self, file, debug_file, guid, code_file, code_id): + def compress(path): + compressed_file = path[:-1] + '_' + # ignore makecab's output + success = subprocess.call(["makecab.exe", "/D", + "CompressionType=MSZIP", + path, compressed_file], + stdout=open(os.devnull, 'w'), + stderr=subprocess.STDOUT) + if success == 0 and os.path.exists(compressed_file): + os.unlink(path) + return True + return False + + rel_path = os.path.join(debug_file, + guid, + debug_file).replace("\\", "/") + full_path = os.path.normpath(os.path.join(self.symbol_path, + rel_path)) + shutil.copyfile(file, full_path) + if compress(full_path): + self.output(sys.stdout, rel_path[:-1] + '_') + else: + self.output(sys.stdout, rel_path) + + # Copy the binary file as well + if code_file and code_id: + full_code_path = os.path.join(os.path.dirname(file), + code_file) + if os.path.exists(full_code_path): + rel_path = os.path.join(code_file, + code_id, + code_file).replace("\\", "/") + full_path = os.path.normpath(os.path.join(self.symbol_path, + rel_path)) + try: + os.makedirs(os.path.dirname(full_path)) + except OSError as e: + if e.errno != errno.EEXIST: + raise + shutil.copyfile(full_code_path, full_path) + if compress(full_path): + self.output(sys.stdout, rel_path[:-1] + '_') + else: + self.output(sys.stdout, rel_path) + + def SourceServerIndexing(self, debug_file, guid, sourceFileStream, vcs_root): + # Creates a .pdb.stream file in the mozilla\objdir to be used for source indexing + debug_file = os.path.abspath(debug_file) + streamFilename = debug_file + ".stream" + stream_output_path = os.path.abspath(streamFilename) + # Call SourceIndex to create the .stream file + result = SourceIndex(sourceFileStream, stream_output_path, vcs_root) + if self.copy_debug: + pdbstr_path = os.environ.get("PDBSTR_PATH") + pdbstr = os.path.normpath(pdbstr_path) + subprocess.call([pdbstr, "-w", "-p:" + os.path.basename(debug_file), + "-i:" + os.path.basename(streamFilename), "-s:srcsrv"], + cwd=os.path.dirname(stream_output_path)) + # clean up all the .stream files when done + os.remove(stream_output_path) + return result + +class Dumper_Linux(Dumper): + objcopy = os.environ['OBJCOPY'] if 'OBJCOPY' in os.environ else 'objcopy' + def ShouldProcess(self, file): + """This function will allow processing of files that are + executable, or end with the .so extension, and additionally + file(1) reports as being ELF files. It expects to find the file + command in PATH.""" + if not Dumper.ShouldProcess(self, file): + return False + if file.endswith(".so") or os.access(file, os.X_OK): + return self.RunFileCommand(file).startswith("ELF") + return False + + def CopyDebug(self, file, debug_file, guid, code_file, code_id): + # We want to strip out the debug info, and add a + # .gnu_debuglink section to the object, so the debugger can + # actually load our debug info later. + file_dbg = file + ".dbg" + if subprocess.call([self.objcopy, '--only-keep-debug', file, file_dbg]) == 0 and \ + subprocess.call([self.objcopy, '--add-gnu-debuglink=%s' % file_dbg, file]) == 0: + rel_path = os.path.join(debug_file, + guid, + debug_file + ".dbg") + full_path = os.path.normpath(os.path.join(self.symbol_path, + rel_path)) + shutil.move(file_dbg, full_path) + # gzip the shipped debug files + os.system("gzip -4 -f %s" % full_path) + self.output(sys.stdout, rel_path + ".gz") + else: + if os.path.isfile(file_dbg): + os.unlink(file_dbg) + +class Dumper_Solaris(Dumper): + def RunFileCommand(self, file): + """Utility function, returns the output of file(1)""" + try: + output = os.popen("file " + file).read() + return output.split('\t')[1]; + except: + return "" + + def ShouldProcess(self, file): + """This function will allow processing of files that are + executable, or end with the .so extension, and additionally + file(1) reports as being ELF files. It expects to find the file + command in PATH.""" + if not Dumper.ShouldProcess(self, file): + return False + if file.endswith(".so") or os.access(file, os.X_OK): + return self.RunFileCommand(file).startswith("ELF") + return False + +def AfterMac(status, dsymbundle): + """Cleanup function to run on Macs after we process the file(s).""" + # CopyDebug will already have been run from Dumper.ProcessFiles + shutil.rmtree(dsymbundle) + +class Dumper_Mac(Dumper): + def ShouldProcess(self, file): + """This function will allow processing of files that are + executable, or end with the .dylib extension, and additionally + file(1) reports as being Mach-O files. It expects to find the file + command in PATH.""" + if not Dumper.ShouldProcess(self, file): + return False + if file.endswith(".dylib") or os.access(file, os.X_OK): + return self.RunFileCommand(file).startswith("Mach-O") + return False + + def ShouldSkipDir(self, dir): + """We create .dSYM bundles on the fly, but if someone runs + buildsymbols twice, we should skip any bundles we created + previously, otherwise we'll recurse into them and try to + dump the inner bits again.""" + if dir.endswith(".dSYM"): + return True + return False + + def ProcessFiles(self, files, after=None, after_arg=None): + # also note, files must be len 1 here, since we're the only ones + # that ever add more than one file to the list + self.output_pid(sys.stderr, "Submitting job for Mac pre-processing on file: %s" % (files[0])) + self.SubmitJob(files[0], 'ProcessFilesWorkMac', args=(files[0],), callback=self.ProcessFilesMacFinished) + + def ProcessFilesMacFinished(self, result): + if result['status']: + # kick off new jobs per-arch with our new list of files + Dumper.ProcessFiles(self, result['files'], after=AfterMac, after_arg=result['files'][0]) + + def dump_syms_cmdline(self, file, arch, files): + ''' + Get the commandline used to invoke dump_syms. + ''' + # dump_syms wants the path to the original binary and the .dSYM + # in order to dump all the symbols. + if len(files) == 2 and file == files[0] and file.endswith('.dSYM'): + # This is the .dSYM bundle. + return [self.dump_syms] + arch.split() + ['-g', file, files[1]] + return Dumper.dump_syms_cmdline(self, file, arch, files) + + def ProcessFilesWorkMac(self, file): + """dump_syms on Mac needs to be run on a dSYM bundle produced + by dsymutil(1), so run dsymutil here and pass the bundle name + down to the superclass method instead.""" + t_start = time.time() + self.output_pid(sys.stderr, "Worker running Mac pre-processing on file: %s" % (file,)) + + # our return is a status and a tuple of files to dump symbols for + # the extra files are fallbacks; as soon as one is dumped successfully, we stop + result = { 'status' : False, 'files' : None, 'file_key' : file } + dsymbundle = file + ".dSYM" + if os.path.exists(dsymbundle): + shutil.rmtree(dsymbundle) + dsymutil = buildconfig.substs['DSYMUTIL'] + # dsymutil takes --arch=foo instead of -a foo like everything else + try: + cmd = ([dsymutil] + + [a.replace('-a ', '--arch=') for a in self.archs if a] + + [file]) + self.output_pid(sys.stderr, ' '.join(cmd)) + subprocess.check_call(cmd, stdout=open(os.devnull, 'w')) + except subprocess.CalledProcessError as e: + self.output_pid(sys.stderr, 'Error running dsymutil: %s' % str(e)) + + if not os.path.exists(dsymbundle): + # dsymutil won't produce a .dSYM for files without symbols + self.output_pid(sys.stderr, "No symbols found in file: %s" % (file,)) + result['status'] = False + result['files'] = (file, ) + return result + + result['status'] = True + result['files'] = (dsymbundle, file) + elapsed = time.time() - t_start + self.output_pid(sys.stderr, 'Worker finished processing %s in %.2fs' % + (file, elapsed)) + return result + + def CopyDebug(self, file, debug_file, guid, code_file, code_id): + """ProcessFiles has already produced a dSYM bundle, so we should just + copy that to the destination directory. However, we'll package it + into a .tar.bz2 because the debug symbols are pretty huge, and + also because it's a bundle, so it's a directory. |file| here is the + dSYM bundle, and |debug_file| is the original filename.""" + rel_path = os.path.join(debug_file, + guid, + os.path.basename(file) + ".tar.bz2") + full_path = os.path.abspath(os.path.join(self.symbol_path, + rel_path)) + success = subprocess.call(["tar", "cjf", full_path, os.path.basename(file)], + cwd=os.path.dirname(file), + stdout=open(os.devnull, 'w'), stderr=subprocess.STDOUT) + if success == 0 and os.path.exists(full_path): + self.output(sys.stdout, rel_path) + +# Entry point if called as a standalone program +def main(): + parser = OptionParser(usage="usage: %prog [options] <dump_syms binary> <symbol store path> <debug info files>") + parser.add_option("-c", "--copy", + action="store_true", dest="copy_debug", default=False, + help="Copy debug info files into the same directory structure as symbol files") + parser.add_option("-a", "--archs", + action="store", dest="archs", + help="Run dump_syms -a <arch> for each space separated cpu architecture in ARCHS (only on OS X)") + parser.add_option("-s", "--srcdir", + action="append", dest="srcdir", default=[], + help="Use SRCDIR to determine relative paths to source files") + parser.add_option("-v", "--vcs-info", + action="store_true", dest="vcsinfo", + help="Try to retrieve VCS info for each FILE listed in the output") + parser.add_option("-i", "--source-index", + action="store_true", dest="srcsrv", default=False, + help="Add source index information to debug files, making them suitable for use in a source server.") + parser.add_option("-x", "--exclude", + action="append", dest="exclude", default=[], metavar="PATTERN", + help="Skip processing files matching PATTERN.") + parser.add_option("--repo-manifest", + action="store", dest="repo_manifest", + help="""Get source information from this XML manifest +produced by the `repo manifest -r` command. +""") + parser.add_option("--install-manifest", + action="append", dest="install_manifests", + default=[], + help="""Use this install manifest to map filenames back +to canonical locations in the source repository. Specify +<install manifest filename>,<install destination> as a comma-separated pair. +""") + (options, args) = parser.parse_args() + + #check to see if the pdbstr.exe exists + if options.srcsrv: + pdbstr = os.environ.get("PDBSTR_PATH") + if not os.path.exists(pdbstr): + print >> sys.stderr, "Invalid path to pdbstr.exe - please set/check PDBSTR_PATH.\n" + sys.exit(1) + + if len(args) < 3: + parser.error("not enough arguments") + exit(1) + + try: + manifests = validate_install_manifests(options.install_manifests) + except (IOError, ValueError) as e: + parser.error(str(e)) + exit(1) + file_mapping = make_file_mapping(manifests) + dumper = GetPlatformSpecificDumper(dump_syms=args[0], + symbol_path=args[1], + copy_debug=options.copy_debug, + archs=options.archs, + srcdirs=options.srcdir, + vcsinfo=options.vcsinfo, + srcsrv=options.srcsrv, + exclude=options.exclude, + repo_manifest=options.repo_manifest, + file_mapping=file_mapping) + + dumper.Process(*args[2:]) + dumper.Finish() + +# run main if run directly +if __name__ == "__main__": + # set up the multiprocessing infrastructure before we start; + # note that this needs to be in the __main__ guard, or else Windows will choke + Dumper.GlobalInit() + + main() diff --git a/toolkit/crashreporter/tools/unit-symbolstore.py b/toolkit/crashreporter/tools/unit-symbolstore.py new file mode 100644 index 000000000..021efbeeb --- /dev/null +++ b/toolkit/crashreporter/tools/unit-symbolstore.py @@ -0,0 +1,583 @@ +#!/usr/bin/env python +# This Source Code Form is subject to the terms of the Mozilla Public +# License, v. 2.0. If a copy of the MPL was not distributed with this +# file, You can obtain one at http://mozilla.org/MPL/2.0/. + +import concurrent.futures +import mock +import mozunit +import os +import platform +import shutil +import struct +import subprocess +import sys +import tempfile +import unittest + +from mock import patch +from mozpack.manifests import InstallManifest + +import symbolstore + +# Some simple functions to mock out files that the platform-specific dumpers will accept. +# dump_syms itself will not be run (we mock that call out), but we can't override +# the ShouldProcessFile method since we actually want to test that. +def write_elf(filename): + open(filename, "wb").write(struct.pack("<7B45x", 0x7f, ord("E"), ord("L"), ord("F"), 1, 1, 1)) + +def write_macho(filename): + open(filename, "wb").write(struct.pack("<I28x", 0xfeedface)) + +def write_pdb(filename): + open(filename, "w").write("aaa") + # write out a fake DLL too + open(os.path.splitext(filename)[0] + ".dll", "w").write("aaa") + +writer = {'Windows': write_pdb, + 'Microsoft': write_pdb, + 'Linux': write_elf, + 'Sunos5': write_elf, + 'Darwin': write_macho}[platform.system()] +extension = {'Windows': ".pdb", + 'Microsoft': ".pdb", + 'Linux': ".so", + 'Sunos5': ".so", + 'Darwin': ".dylib"}[platform.system()] + +def add_extension(files): + return [f + extension for f in files] + +class HelperMixin(object): + """ + Test that passing filenames to exclude from processing works. + """ + def setUp(self): + self.test_dir = tempfile.mkdtemp() + if not self.test_dir.endswith(os.sep): + self.test_dir += os.sep + symbolstore.srcdirRepoInfo = {} + symbolstore.vcsFileInfoCache = {} + + # Remove environment variables that can influence tests. + for e in ('MOZ_SOURCE_CHANGESET', 'MOZ_SOURCE_REPO'): + try: + del os.environ[e] + except KeyError: + pass + + def tearDown(self): + shutil.rmtree(self.test_dir) + symbolstore.srcdirRepoInfo = {} + symbolstore.vcsFileInfoCache = {} + + def make_dirs(self, f): + d = os.path.dirname(f) + if d and not os.path.exists(d): + os.makedirs(d) + + def add_test_files(self, files): + for f in files: + f = os.path.join(self.test_dir, f) + self.make_dirs(f) + writer(f) + +class TestSizeOrder(HelperMixin, unittest.TestCase): + def test_size_order(self): + """ + Test that files are processed ordered by size on disk. + """ + processed = [] + def mock_process_file(filenames): + for filename in filenames: + processed.append((filename[len(self.test_dir):] if filename.startswith(self.test_dir) else filename).replace('\\', '/')) + return True + for f, size in (('a/one', 10), ('b/c/two', 30), ('c/three', 20)): + f = os.path.join(self.test_dir, f) + d = os.path.dirname(f) + if d and not os.path.exists(d): + os.makedirs(d) + open(f, 'wb').write('x' * size) + d = symbolstore.GetPlatformSpecificDumper(dump_syms="dump_syms", + symbol_path="symbol_path") + d.ShouldProcess = lambda f: True + d.ProcessFiles = mock_process_file + d.Process(self.test_dir) + d.Finish(stop_pool=False) + self.assertEqual(processed, ['b/c/two', 'c/three', 'a/one']) + + +class TestExclude(HelperMixin, unittest.TestCase): + def test_exclude_wildcard(self): + """ + Test that using an exclude list with a wildcard pattern works. + """ + processed = [] + def mock_process_file(filenames): + for filename in filenames: + processed.append((filename[len(self.test_dir):] if filename.startswith(self.test_dir) else filename).replace('\\', '/')) + return True + self.add_test_files(add_extension(["foo", "bar", "abc/xyz", "abc/fooxyz", "def/asdf", "def/xyzfoo"])) + d = symbolstore.GetPlatformSpecificDumper(dump_syms="dump_syms", + symbol_path="symbol_path", + exclude=["*foo*"]) + d.ProcessFiles = mock_process_file + d.Process(self.test_dir) + d.Finish(stop_pool=False) + processed.sort() + expected = add_extension(["bar", "abc/xyz", "def/asdf"]) + expected.sort() + self.assertEqual(processed, expected) + + + def test_exclude_filenames(self): + """ + Test that excluding a filename without a wildcard works. + """ + processed = [] + def mock_process_file(filenames): + for filename in filenames: + processed.append((filename[len(self.test_dir):] if filename.startswith(self.test_dir) else filename).replace('\\', '/')) + return True + self.add_test_files(add_extension(["foo", "bar", "abc/foo", "abc/bar", "def/foo", "def/bar"])) + d = symbolstore.GetPlatformSpecificDumper(dump_syms="dump_syms", + symbol_path="symbol_path", + exclude=add_extension(["foo"])) + d.ProcessFiles = mock_process_file + d.Process(self.test_dir) + d.Finish(stop_pool=False) + processed.sort() + expected = add_extension(["bar", "abc/bar", "def/bar"]) + expected.sort() + self.assertEqual(processed, expected) + + +def mock_dump_syms(module_id, filename, extra=[]): + return ["MODULE os x86 %s %s" % (module_id, filename) + ] + extra + [ + "FILE 0 foo.c", + "PUBLIC xyz 123"] + + +class TestCopyDebug(HelperMixin, unittest.TestCase): + def setUp(self): + HelperMixin.setUp(self) + self.symbol_dir = tempfile.mkdtemp() + self.mock_call = patch("subprocess.call").start() + self.stdouts = [] + self.mock_popen = patch("subprocess.Popen").start() + stdout_iter = self.next_mock_stdout() + def next_popen(*args, **kwargs): + m = mock.MagicMock() + m.stdout = stdout_iter.next() + m.wait.return_value = 0 + return m + self.mock_popen.side_effect = next_popen + shutil.rmtree = patch("shutil.rmtree").start() + + def tearDown(self): + HelperMixin.tearDown(self) + patch.stopall() + shutil.rmtree(self.symbol_dir) + + def next_mock_stdout(self): + if not self.stdouts: + yield iter([]) + for s in self.stdouts: + yield iter(s) + + def test_copy_debug_universal(self): + """ + Test that dumping symbols for multiple architectures only copies debug symbols once + per file. + """ + copied = [] + def mock_copy_debug(filename, debug_file, guid, code_file, code_id): + copied.append(filename[len(self.symbol_dir):] if filename.startswith(self.symbol_dir) else filename) + self.add_test_files(add_extension(["foo"])) + self.stdouts.append(mock_dump_syms("X" * 33, add_extension(["foo"])[0])) + self.stdouts.append(mock_dump_syms("Y" * 33, add_extension(["foo"])[0])) + def mock_dsymutil(args, **kwargs): + filename = args[-1] + os.makedirs(filename + ".dSYM") + return 0 + self.mock_call.side_effect = mock_dsymutil + d = symbolstore.GetPlatformSpecificDumper(dump_syms="dump_syms", + symbol_path=self.symbol_dir, + copy_debug=True, + archs="abc xyz") + d.CopyDebug = mock_copy_debug + d.Process(self.test_dir) + d.Finish(stop_pool=False) + self.assertEqual(1, len(copied)) + + def test_copy_debug_copies_binaries(self): + """ + Test that CopyDebug copies binaries as well on Windows. + """ + test_file = os.path.join(self.test_dir, 'foo.pdb') + write_pdb(test_file) + code_file = 'foo.dll' + code_id = 'abc123' + self.stdouts.append(mock_dump_syms('X' * 33, 'foo.pdb', + ['INFO CODE_ID %s %s' % (code_id, code_file)])) + def mock_compress(args, **kwargs): + filename = args[-1] + open(filename, 'w').write('stuff') + return 0 + self.mock_call.side_effect = mock_compress + d = symbolstore.Dumper_Win32(dump_syms='dump_syms', + symbol_path=self.symbol_dir, + copy_debug=True) + d.FixFilenameCase = lambda f: f + d.Process(self.test_dir) + d.Finish(stop_pool=False) + self.assertTrue(os.path.isfile(os.path.join(self.symbol_dir, code_file, code_id, code_file[:-1] + '_'))) + +class TestGetVCSFilename(HelperMixin, unittest.TestCase): + def setUp(self): + HelperMixin.setUp(self) + + def tearDown(self): + HelperMixin.tearDown(self) + + @patch("subprocess.Popen") + def testVCSFilenameHg(self, mock_Popen): + # mock calls to `hg parent` and `hg showconfig paths.default` + mock_communicate = mock_Popen.return_value.communicate + mock_communicate.side_effect = [("abcd1234", ""), + ("http://example.com/repo", "")] + os.mkdir(os.path.join(self.test_dir, ".hg")) + filename = os.path.join(self.test_dir, "foo.c") + self.assertEqual("hg:example.com/repo:foo.c:abcd1234", + symbolstore.GetVCSFilename(filename, [self.test_dir])[0]) + + @patch("subprocess.Popen") + def testVCSFilenameHgMultiple(self, mock_Popen): + # mock calls to `hg parent` and `hg showconfig paths.default` + mock_communicate = mock_Popen.return_value.communicate + mock_communicate.side_effect = [("abcd1234", ""), + ("http://example.com/repo", ""), + ("0987ffff", ""), + ("http://example.com/other", "")] + srcdir1 = os.path.join(self.test_dir, "one") + srcdir2 = os.path.join(self.test_dir, "two") + os.makedirs(os.path.join(srcdir1, ".hg")) + os.makedirs(os.path.join(srcdir2, ".hg")) + filename1 = os.path.join(srcdir1, "foo.c") + filename2 = os.path.join(srcdir2, "bar.c") + self.assertEqual("hg:example.com/repo:foo.c:abcd1234", + symbolstore.GetVCSFilename(filename1, [srcdir1, srcdir2])[0]) + self.assertEqual("hg:example.com/other:bar.c:0987ffff", + symbolstore.GetVCSFilename(filename2, [srcdir1, srcdir2])[0]) + + def testVCSFilenameEnv(self): + # repo URL and changeset read from environment variables if defined. + os.environ['MOZ_SOURCE_REPO'] = 'https://somewhere.com/repo' + os.environ['MOZ_SOURCE_CHANGESET'] = 'abcdef0123456' + os.mkdir(os.path.join(self.test_dir, '.hg')) + filename = os.path.join(self.test_dir, 'foo.c') + self.assertEqual('hg:somewhere.com/repo:foo.c:abcdef0123456', + symbolstore.GetVCSFilename(filename, [self.test_dir])[0]) + + +class TestRepoManifest(HelperMixin, unittest.TestCase): + def testRepoManifest(self): + manifest = os.path.join(self.test_dir, "sources.xml") + open(manifest, "w").write("""<?xml version="1.0" encoding="UTF-8"?> +<manifest> +<remote fetch="http://example.com/foo/" name="foo"/> +<remote fetch="git://example.com/bar/" name="bar"/> +<default remote="bar"/> +<project name="projects/one" revision="abcd1234"/> +<project name="projects/two" path="projects/another" revision="ffffffff" remote="foo"/> +<project name="something_else" revision="00000000" remote="bar"/> +</manifest> +""") + # Use a source file from each of the three projects + file1 = os.path.join(self.test_dir, "projects", "one", "src1.c") + file2 = os.path.join(self.test_dir, "projects", "another", "src2.c") + file3 = os.path.join(self.test_dir, "something_else", "src3.c") + d = symbolstore.Dumper("dump_syms", "symbol_path", + repo_manifest=manifest) + self.assertEqual("git:example.com/bar/projects/one:src1.c:abcd1234", + symbolstore.GetVCSFilename(file1, d.srcdirs)[0]) + self.assertEqual("git:example.com/foo/projects/two:src2.c:ffffffff", + symbolstore.GetVCSFilename(file2, d.srcdirs)[0]) + self.assertEqual("git:example.com/bar/something_else:src3.c:00000000", + symbolstore.GetVCSFilename(file3, d.srcdirs)[0]) + +if platform.system() in ("Windows", "Microsoft"): + class TestFixFilenameCase(HelperMixin, unittest.TestCase): + def test_fix_filename_case(self): + # self.test_dir is going to be 8.3 paths... + junk = os.path.join(self.test_dir, 'x') + with open(junk, 'wb') as o: + o.write('x') + d = symbolstore.Dumper_Win32(dump_syms='dump_syms', + symbol_path=self.test_dir) + fixed_dir = os.path.dirname(d.FixFilenameCase(junk)) + files = [ + 'one\\two.c', + 'three\\Four.d', + 'Five\\Six.e', + 'seven\\Eight\\nine.F', + ] + for rel_path in files: + full_path = os.path.normpath(os.path.join(self.test_dir, + rel_path)) + self.make_dirs(full_path) + with open(full_path, 'wb') as o: + o.write('x') + fixed_path = d.FixFilenameCase(full_path.lower()) + fixed_path = os.path.relpath(fixed_path, fixed_dir) + self.assertEqual(rel_path, fixed_path) + + class TestSourceServer(HelperMixin, unittest.TestCase): + @patch("subprocess.call") + @patch("subprocess.Popen") + def test_HGSERVER(self, mock_Popen, mock_call): + """ + Test that HGSERVER gets set correctly in the source server index. + """ + symbolpath = os.path.join(self.test_dir, "symbols") + os.makedirs(symbolpath) + srcdir = os.path.join(self.test_dir, "srcdir") + os.makedirs(os.path.join(srcdir, ".hg")) + sourcefile = os.path.join(srcdir, "foo.c") + test_files = add_extension(["foo"]) + self.add_test_files(test_files) + # srcsrv needs PDBSTR_PATH set + os.environ["PDBSTR_PATH"] = "pdbstr" + # mock calls to `dump_syms`, `hg parent` and + # `hg showconfig paths.default` + mock_Popen.return_value.stdout = iter([ + "MODULE os x86 %s %s" % ("X" * 33, test_files[0]), + "FILE 0 %s" % sourcefile, + "PUBLIC xyz 123" + ]) + mock_communicate = mock_Popen.return_value.communicate + mock_communicate.side_effect = [("abcd1234", ""), + ("http://example.com/repo", ""), + ] + # And mock the call to pdbstr to capture the srcsrv stream data. + global srcsrv_stream + srcsrv_stream = None + def mock_pdbstr(args, cwd="", **kwargs): + for arg in args: + if arg.startswith("-i:"): + global srcsrv_stream + srcsrv_stream = open(os.path.join(cwd, arg[3:]), 'r').read() + return 0 + mock_call.side_effect = mock_pdbstr + d = symbolstore.GetPlatformSpecificDumper(dump_syms="dump_syms", + symbol_path=symbolpath, + srcdirs=[srcdir], + vcsinfo=True, + srcsrv=True, + copy_debug=True) + # stub out CopyDebug + d.CopyDebug = lambda *args: True + d.Process(self.test_dir) + d.Finish(stop_pool=False) + self.assertNotEqual(srcsrv_stream, None) + hgserver = [x.rstrip() for x in srcsrv_stream.splitlines() if x.startswith("HGSERVER=")] + self.assertEqual(len(hgserver), 1) + self.assertEqual(hgserver[0].split("=")[1], "http://example.com/repo") + +class TestInstallManifest(HelperMixin, unittest.TestCase): + def setUp(self): + HelperMixin.setUp(self) + self.srcdir = os.path.join(self.test_dir, 'src') + os.mkdir(self.srcdir) + self.objdir = os.path.join(self.test_dir, 'obj') + os.mkdir(self.objdir) + self.manifest = InstallManifest() + self.canonical_mapping = {} + for s in ['src1', 'src2']: + srcfile = os.path.join(self.srcdir, s) + objfile = os.path.join(self.objdir, s) + self.canonical_mapping[objfile] = srcfile + self.manifest.add_copy(srcfile, s) + self.manifest_file = os.path.join(self.test_dir, 'install-manifest') + self.manifest.write(self.manifest_file) + + def testMakeFileMapping(self): + ''' + Test that valid arguments are validated. + ''' + arg = '%s,%s' % (self.manifest_file, self.objdir) + ret = symbolstore.validate_install_manifests([arg]) + self.assertEqual(len(ret), 1) + manifest, dest = ret[0] + self.assertTrue(isinstance(manifest, InstallManifest)) + self.assertEqual(dest, self.objdir) + + file_mapping = symbolstore.make_file_mapping(ret) + for obj, src in self.canonical_mapping.iteritems(): + self.assertTrue(obj in file_mapping) + self.assertEqual(file_mapping[obj], src) + + def testMissingFiles(self): + ''' + Test that missing manifest files or install directories give errors. + ''' + missing_manifest = os.path.join(self.test_dir, 'missing-manifest') + arg = '%s,%s' % (missing_manifest, self.objdir) + with self.assertRaises(IOError) as e: + symbolstore.validate_install_manifests([arg]) + self.assertEqual(e.filename, missing_manifest) + + missing_install_dir = os.path.join(self.test_dir, 'missing-dir') + arg = '%s,%s' % (self.manifest_file, missing_install_dir) + with self.assertRaises(IOError) as e: + symbolstore.validate_install_manifests([arg]) + self.assertEqual(e.filename, missing_install_dir) + + def testBadManifest(self): + ''' + Test that a bad manifest file give errors. + ''' + bad_manifest = os.path.join(self.test_dir, 'bad-manifest') + with open(bad_manifest, 'wb') as f: + f.write('junk\n') + arg = '%s,%s' % (bad_manifest, self.objdir) + with self.assertRaises(IOError) as e: + symbolstore.validate_install_manifests([arg]) + self.assertEqual(e.filename, bad_manifest) + + def testBadArgument(self): + ''' + Test that a bad manifest argument gives an error. + ''' + with self.assertRaises(ValueError) as e: + symbolstore.validate_install_manifests(['foo']) + +class TestFileMapping(HelperMixin, unittest.TestCase): + def setUp(self): + HelperMixin.setUp(self) + self.srcdir = os.path.join(self.test_dir, 'src') + os.mkdir(self.srcdir) + self.objdir = os.path.join(self.test_dir, 'obj') + os.mkdir(self.objdir) + self.symboldir = os.path.join(self.test_dir, 'symbols') + os.mkdir(self.symboldir) + + @patch("subprocess.Popen") + def testFileMapping(self, mock_Popen): + files = [('a/b', 'mozilla/b'), + ('c/d', 'foo/d')] + if os.sep != '/': + files = [[f.replace('/', os.sep) for f in x] for x in files] + file_mapping = {} + dumped_files = [] + expected_files = [] + for s, o in files: + srcfile = os.path.join(self.srcdir, s) + expected_files.append(srcfile) + file_mapping[os.path.join(self.objdir, o)] = srcfile + dumped_files.append(os.path.join(self.objdir, 'x', 'y', + '..', '..', o)) + # mock the dump_syms output + file_id = ("X" * 33, 'somefile') + def mk_output(files): + return iter( + [ + 'MODULE os x86 %s %s\n' % file_id + ] + + [ + 'FILE %d %s\n' % (i,s) for i, s in enumerate(files) + ] + + [ + 'PUBLIC xyz 123\n' + ] + ) + mock_Popen.return_value.stdout = mk_output(dumped_files) + + d = symbolstore.Dumper('dump_syms', self.symboldir, + file_mapping=file_mapping) + f = os.path.join(self.objdir, 'somefile') + open(f, 'wb').write('blah') + d.Process(f) + d.Finish(stop_pool=False) + expected_output = ''.join(mk_output(expected_files)) + symbol_file = os.path.join(self.symboldir, + file_id[1], file_id[0], file_id[1] + '.sym') + self.assertEqual(open(symbol_file, 'r').read(), expected_output) + +class TestFunctional(HelperMixin, unittest.TestCase): + '''Functional tests of symbolstore.py, calling it with a real + dump_syms binary and passing in a real binary to dump symbols from. + + Since the rest of the tests in this file mock almost everything and + don't use the actual process pool like buildsymbols does, this tests + that the way symbolstore.py gets called in buildsymbols works. + ''' + def setUp(self): + HelperMixin.setUp(self) + import buildconfig + self.skip_test = False + if buildconfig.substs['MOZ_BUILD_APP'] != 'browser': + self.skip_test = True + self.topsrcdir = buildconfig.topsrcdir + self.script_path = os.path.join(self.topsrcdir, 'toolkit', + 'crashreporter', 'tools', + 'symbolstore.py') + if platform.system() in ("Windows", "Microsoft"): + if buildconfig.substs['MSVC_HAS_DIA_SDK']: + self.dump_syms = os.path.join(buildconfig.topobjdir, + 'dist', 'host', 'bin', + 'dump_syms.exe') + else: + self.dump_syms = os.path.join(self.topsrcdir, + 'toolkit', + 'crashreporter', + 'tools', + 'win32', + 'dump_syms_vc{_MSC_VER}.exe'.format(**buildconfig.substs)) + self.target_bin = os.path.join(buildconfig.topobjdir, + 'browser', + 'app', + 'firefox.pdb') + else: + self.dump_syms = os.path.join(buildconfig.topobjdir, + 'dist', 'host', 'bin', + 'dump_syms') + self.target_bin = os.path.join(buildconfig.topobjdir, + 'dist', 'bin', 'firefox') + + + def tearDown(self): + HelperMixin.tearDown(self) + + def testSymbolstore(self): + if self.skip_test: + raise unittest.SkipTest('Skipping test in non-Firefox product') + output = subprocess.check_output([sys.executable, + self.script_path, + '--vcs-info', + '-s', self.topsrcdir, + self.dump_syms, + self.test_dir, + self.target_bin], + stderr=open(os.devnull, 'w')) + lines = filter(lambda x: x.strip(), output.splitlines()) + self.assertEqual(1, len(lines), + 'should have one filename in the output') + symbol_file = os.path.join(self.test_dir, lines[0]) + self.assertTrue(os.path.isfile(symbol_file)) + symlines = open(symbol_file, 'r').readlines() + file_lines = filter(lambda x: x.startswith('FILE') and 'nsBrowserApp.cpp' in x, symlines) + self.assertEqual(len(file_lines), 1, + 'should have nsBrowserApp.cpp FILE line') + filename = file_lines[0].split(None, 2)[2] + self.assertEqual('hg:', filename[:3]) + + +if __name__ == '__main__': + # use ThreadPoolExecutor to use threading instead of processes so + # that our mocking/module-patching works. + symbolstore.Dumper.GlobalInit(concurrent.futures.ThreadPoolExecutor) + + mozunit.main() + diff --git a/toolkit/crashreporter/tools/upload_symbols.py b/toolkit/crashreporter/tools/upload_symbols.py new file mode 100644 index 000000000..7e94b238d --- /dev/null +++ b/toolkit/crashreporter/tools/upload_symbols.py @@ -0,0 +1,102 @@ +#!/usr/bin/env python +# +# This Source Code Form is subject to the terms of the Mozilla Public +# License, v. 2.0. If a copy of the MPL was not distributed with this +# file, You can obtain one at http://mozilla.org/MPL/2.0/. +# +# This script uploads a symbol zip file passed on the commandline +# to the Socorro symbol upload API hosted on crash-stats.mozilla.org. +# +# Using this script requires you to have generated an authentication +# token in the crash-stats web interface. You must put the token in a file +# and set SOCORRO_SYMBOL_UPLOAD_TOKEN_FILE to the path to the file in +# the mozconfig you're using. + +from __future__ import print_function + +import os +import redo +import requests +import sys + +try: + from buildconfig import substs +except ImportError: + # Allow standalone use of this script, for use in TaskCluster + from os import environ as substs + +url = 'https://symbols.mozilla.org/upload/' +# Allow overwriting of the upload url with an environmental variable +if 'SOCORRO_SYMBOL_UPLOAD_URL' in os.environ: + url = os.environ['SOCORRO_SYMBOL_UPLOAD_URL'] +MAX_RETRIES = 5 + +def print_error(r): + if r.status_code < 400: + print('Error: bad auth token? ({0}: {1})'.format(r.status_code, + r.reason), + file=sys.stderr) + else: + print('Error: got HTTP response {0}: {1}'.format(r.status_code, + r.reason), + file=sys.stderr) + + print('Response body:\n{sep}\n{body}\n{sep}\n'.format( + sep='=' * 20, + body=r.text + )) + +def main(): + if len(sys.argv) != 2: + print('Usage: uploadsymbols.py <zip file>', file=sys.stderr) + return 1 + + if not os.path.isfile(sys.argv[1]): + print('Error: zip file "{0}" does not exist!'.format(sys.argv[1]), + file=sys.stderr) + return 1 + symbols_zip = sys.argv[1] + + if 'SOCORRO_SYMBOL_UPLOAD_TOKEN_FILE' not in substs: + print('Error: you must set SOCORRO_SYMBOL_UPLOAD_TOKEN_FILE in your mozconfig!', file=sys.stderr) + return 1 + token_file = substs['SOCORRO_SYMBOL_UPLOAD_TOKEN_FILE'] + + if not os.path.isfile(token_file): + print('Error: SOCORRO_SYMBOL_UPLOAD_TOKEN_FILE "{0}" does not exist!'.format(token_file), file=sys.stderr) + return 1 + auth_token = open(token_file, 'r').read().strip() + + print('Uploading symbol file "{0}" to "{1}"'.format(sys.argv[1], url)) + + for i, _ in enumerate(redo.retrier(attempts=MAX_RETRIES), start=1): + print('Attempt %d of %d...' % (i, MAX_RETRIES)) + try: + r = requests.post( + url, + files={'symbols.zip': open(sys.argv[1], 'rb')}, + headers={'Auth-Token': auth_token}, + allow_redirects=False, + timeout=120) + # 500 is likely to be a transient failure. + # Break out for success or other error codes. + if r.status_code < 500: + break + print_error(r) + except requests.exceptions.RequestException as e: + print('Error: {0}'.format(e)) + print('Retrying...') + else: + print('Maximum retries hit, giving up!') + return 1 + + if r.status_code >= 200 and r.status_code < 300: + print('Uploaded successfully!') + return 0 + + print_error(r) + return 1 + +if __name__ == '__main__': + sys.exit(main()) + diff --git a/toolkit/crashreporter/tools/win32/dump_syms_vc1600.exe b/toolkit/crashreporter/tools/win32/dump_syms_vc1600.exe Binary files differnew file mode 100755 index 000000000..489cc4c4a --- /dev/null +++ b/toolkit/crashreporter/tools/win32/dump_syms_vc1600.exe diff --git a/toolkit/crashreporter/tools/win32/dump_syms_vc1700.exe b/toolkit/crashreporter/tools/win32/dump_syms_vc1700.exe Binary files differnew file mode 100644 index 000000000..0ead28957 --- /dev/null +++ b/toolkit/crashreporter/tools/win32/dump_syms_vc1700.exe diff --git a/toolkit/crashreporter/tools/win32/dump_syms_vc1800.exe b/toolkit/crashreporter/tools/win32/dump_syms_vc1800.exe Binary files differnew file mode 100644 index 000000000..22e7398ab --- /dev/null +++ b/toolkit/crashreporter/tools/win32/dump_syms_vc1800.exe |