diff options
author | wolfbeast <mcwerewolf@wolfbeast.com> | 2019-04-01 13:05:24 +0200 |
---|---|---|
committer | wolfbeast <mcwerewolf@wolfbeast.com> | 2019-04-01 13:05:24 +0200 |
commit | ff2f287f82630ab3887d7d5c1e64e5b888ea0beb (patch) | |
tree | 4e96cb32aa2320a327024942d247c6b56ef8c199 /toolkit/crashreporter/tools | |
parent | bfc97728065cbbc7f6bbc281b654a2d1e079b48d (diff) | |
download | UXP-ff2f287f82630ab3887d7d5c1e64e5b888ea0beb.tar UXP-ff2f287f82630ab3887d7d5c1e64e5b888ea0beb.tar.gz UXP-ff2f287f82630ab3887d7d5c1e64e5b888ea0beb.tar.lz UXP-ff2f287f82630ab3887d7d5c1e64e5b888ea0beb.tar.xz UXP-ff2f287f82630ab3887d7d5c1e64e5b888ea0beb.zip |
Remove crashreporter toolkit files.
Resolves #20
Diffstat (limited to 'toolkit/crashreporter/tools')
-rwxr-xr-x | toolkit/crashreporter/tools/symbolstore.py | 1078 | ||||
-rw-r--r-- | toolkit/crashreporter/tools/unit-symbolstore.py | 583 | ||||
-rw-r--r-- | toolkit/crashreporter/tools/upload_symbols.py | 102 | ||||
-rwxr-xr-x | toolkit/crashreporter/tools/win32/dump_syms_vc1600.exe | bin | 55296 -> 0 bytes | |||
-rw-r--r-- | toolkit/crashreporter/tools/win32/dump_syms_vc1700.exe | bin | 52736 -> 0 bytes | |||
-rw-r--r-- | toolkit/crashreporter/tools/win32/dump_syms_vc1800.exe | bin | 51200 -> 0 bytes |
6 files changed, 0 insertions, 1763 deletions
diff --git a/toolkit/crashreporter/tools/symbolstore.py b/toolkit/crashreporter/tools/symbolstore.py deleted file mode 100755 index 15eb5e112..000000000 --- a/toolkit/crashreporter/tools/symbolstore.py +++ /dev/null @@ -1,1078 +0,0 @@ -#!/bin/env python -# This Source Code Form is subject to the terms of the Mozilla Public -# License, v. 2.0. If a copy of the MPL was not distributed with this -# file, You can obtain one at http://mozilla.org/MPL/2.0/. -# -# Usage: symbolstore.py <params> <dump_syms path> <symbol store path> -# <debug info files or dirs> -# Runs dump_syms on each debug info file specified on the command line, -# then places the resulting symbol file in the proper directory -# structure in the symbol store path. Accepts multiple files -# on the command line, so can be called as part of a pipe using -# find <dir> | xargs symbolstore.pl <dump_syms> <storepath> -# But really, you might just want to pass it <dir>. -# -# Parameters accepted: -# -c : Copy debug info files to the same directory structure -# as sym files. On Windows, this will also copy -# binaries into the symbol store. -# -a "<archs>" : Run dump_syms -a <arch> for each space separated -# cpu architecture in <archs> (only on OS X) -# -s <srcdir> : Use <srcdir> as the top source directory to -# generate relative filenames. - -import buildconfig -import errno -import sys -import platform -import os -import re -import shutil -import textwrap -import fnmatch -import subprocess -import time -import ctypes -import urlparse -import concurrent.futures -import multiprocessing - -from optparse import OptionParser -from xml.dom.minidom import parse - -from mozpack.copier import FileRegistry -from mozpack.manifests import ( - InstallManifest, - UnreadableInstallManifest, -) - -# Utility classes - -class VCSFileInfo: - """ A base class for version-controlled file information. Ensures that the - following attributes are generated only once (successfully): - - self.root - self.clean_root - self.revision - self.filename - - The attributes are generated by a single call to the GetRoot, - GetRevision, and GetFilename methods. Those methods are explicitly not - implemented here and must be implemented in derived classes. """ - - def __init__(self, file): - if not file: - raise ValueError - self.file = file - - def __getattr__(self, name): - """ __getattr__ is only called for attributes that are not set on self, - so setting self.[attr] will prevent future calls to the GetRoot, - GetRevision, and GetFilename methods. We don't set the values on - failure on the off chance that a future call might succeed. """ - - if name == "root": - root = self.GetRoot() - if root: - self.root = root - return root - - elif name == "clean_root": - clean_root = self.GetCleanRoot() - if clean_root: - self.clean_root = clean_root - return clean_root - - elif name == "revision": - revision = self.GetRevision() - if revision: - self.revision = revision - return revision - - elif name == "filename": - filename = self.GetFilename() - if filename: - self.filename = filename - return filename - - raise AttributeError - - def GetRoot(self): - """ This method should return the unmodified root for the file or 'None' - on failure. """ - raise NotImplementedError - - def GetCleanRoot(self): - """ This method should return the repository root for the file or 'None' - on failure. """ - raise NotImplementedError - - def GetRevision(self): - """ This method should return the revision number for the file or 'None' - on failure. """ - raise NotImplementedError - - def GetFilename(self): - """ This method should return the repository-specific filename for the - file or 'None' on failure. """ - raise NotImplementedError - - -# This regex separates protocol and optional username/password from a url. -# For instance, all the following urls will be transformed into -# 'foo.com/bar': -# -# http://foo.com/bar -# svn+ssh://user@foo.com/bar -# svn+ssh://user:pass@foo.com/bar -# -rootRegex = re.compile(r'^\S+?:/+(?:[^\s/]*@)?(\S+)$') - -def read_output(*args): - (stdout, _) = subprocess.Popen(args=args, stdout=subprocess.PIPE).communicate() - return stdout.rstrip() - -class HGRepoInfo: - def __init__(self, path): - self.path = path - - rev = os.environ.get('MOZ_SOURCE_CHANGESET') - if not rev: - rev = read_output('hg', '-R', path, - 'parent', '--template={node|short}') - - # Look for the default hg path. If MOZ_SOURCE_REPO is set, we - # don't bother asking hg. - hg_root = os.environ.get('MOZ_SOURCE_REPO') - if hg_root: - root = hg_root - else: - root = read_output('hg', '-R', path, - 'showconfig', 'paths.default') - if not root: - print >> sys.stderr, "Failed to get HG Repo for %s" % path - cleanroot = None - if root: - match = rootRegex.match(root) - if match: - cleanroot = match.group(1) - if cleanroot.endswith('/'): - cleanroot = cleanroot[:-1] - if cleanroot is None: - print >> sys.stderr, textwrap.dedent("""\ - Could not determine repo info for %s. This is either not a clone of the web-based - repository, or you have not specified MOZ_SOURCE_REPO, or the clone is corrupt.""") % path - sys.exit(1) - self.rev = rev - self.root = root - self.cleanroot = cleanroot - - def GetFileInfo(self, file): - return HGFileInfo(file, self) - -class HGFileInfo(VCSFileInfo): - def __init__(self, file, repo): - VCSFileInfo.__init__(self, file) - self.repo = repo - self.file = os.path.relpath(file, repo.path) - - def GetRoot(self): - return self.repo.root - - def GetCleanRoot(self): - return self.repo.cleanroot - - def GetRevision(self): - return self.repo.rev - - def GetFilename(self): - if self.revision and self.clean_root: - return "hg:%s:%s:%s" % (self.clean_root, self.file, self.revision) - return self.file - -class GitRepoInfo: - """ - Info about a local git repository. Does not currently - support discovering info about a git clone, the info must be - provided out-of-band. - """ - def __init__(self, path, rev, root): - self.path = path - cleanroot = None - if root: - match = rootRegex.match(root) - if match: - cleanroot = match.group(1) - if cleanroot.endswith('/'): - cleanroot = cleanroot[:-1] - if cleanroot is None: - print >> sys.stderr, textwrap.dedent("""\ - Could not determine repo info for %s (%s). This is either not a clone of a web-based - repository, or you have not specified MOZ_SOURCE_REPO, or the clone is corrupt.""") % (path, root) - sys.exit(1) - self.rev = rev - self.cleanroot = cleanroot - - def GetFileInfo(self, file): - return GitFileInfo(file, self) - -class GitFileInfo(VCSFileInfo): - def __init__(self, file, repo): - VCSFileInfo.__init__(self, file) - self.repo = repo - self.file = os.path.relpath(file, repo.path) - - def GetRoot(self): - return self.repo.path - - def GetCleanRoot(self): - return self.repo.cleanroot - - def GetRevision(self): - return self.repo.rev - - def GetFilename(self): - if self.revision and self.clean_root: - return "git:%s:%s:%s" % (self.clean_root, self.file, self.revision) - return self.file - -# Utility functions - -# A cache of files for which VCS info has already been determined. Used to -# prevent extra filesystem activity or process launching. -vcsFileInfoCache = {} - -def IsInDir(file, dir): - # the lower() is to handle win32+vc8, where - # the source filenames come out all lowercase, - # but the srcdir can be mixed case - return os.path.abspath(file).lower().startswith(os.path.abspath(dir).lower()) - -def GetVCSFilenameFromSrcdir(file, srcdir): - if srcdir not in Dumper.srcdirRepoInfo: - # Not in cache, so find it adnd cache it - if os.path.isdir(os.path.join(srcdir, '.hg')): - Dumper.srcdirRepoInfo[srcdir] = HGRepoInfo(srcdir) - else: - # Unknown VCS or file is not in a repo. - return None - return Dumper.srcdirRepoInfo[srcdir].GetFileInfo(file) - -def GetVCSFilename(file, srcdirs): - """Given a full path to a file, and the top source directory, - look for version control information about this file, and return - a tuple containing - 1) a specially formatted filename that contains the VCS type, - VCS location, relative filename, and revision number, formatted like: - vcs:vcs location:filename:revision - For example: - cvs:cvs.mozilla.org/cvsroot:mozilla/browser/app/nsBrowserApp.cpp:1.36 - 2) the unmodified root information if it exists""" - (path, filename) = os.path.split(file) - if path == '' or filename == '': - return (file, None) - - fileInfo = None - root = '' - if file in vcsFileInfoCache: - # Already cached this info, use it. - fileInfo = vcsFileInfoCache[file] - else: - for srcdir in srcdirs: - if not IsInDir(file, srcdir): - continue - fileInfo = GetVCSFilenameFromSrcdir(file, srcdir) - if fileInfo: - vcsFileInfoCache[file] = fileInfo - break - - if fileInfo: - file = fileInfo.filename - root = fileInfo.root - - # we want forward slashes on win32 paths - return (file.replace("\\", "/"), root) - -def validate_install_manifests(install_manifest_args): - args = [] - for arg in install_manifest_args: - bits = arg.split(',') - if len(bits) != 2: - raise ValueError('Invalid format for --install-manifest: ' - 'specify manifest,target_dir') - manifest_file, destination = map(os.path.abspath, bits) - if not os.path.isfile(manifest_file): - raise IOError(errno.ENOENT, 'Manifest file not found', - manifest_file) - if not os.path.isdir(destination): - raise IOError(errno.ENOENT, 'Install directory not found', - destination) - try: - manifest = InstallManifest(manifest_file) - except UnreadableInstallManifest: - raise IOError(errno.EINVAL, 'Error parsing manifest file', - manifest_file) - args.append((manifest, destination)) - return args - -def make_file_mapping(install_manifests): - file_mapping = {} - for manifest, destination in install_manifests: - destination = os.path.abspath(destination) - reg = FileRegistry() - manifest.populate_registry(reg) - for dst, src in reg: - if hasattr(src, 'path'): - abs_dest = os.path.normpath(os.path.join(destination, dst)) - file_mapping[abs_dest] = src.path - return file_mapping - -def GetPlatformSpecificDumper(**kwargs): - """This function simply returns a instance of a subclass of Dumper - that is appropriate for the current platform.""" - return {'WINNT': Dumper_Win32, - 'Linux': Dumper_Linux, - 'Darwin': Dumper_Mac}[buildconfig.substs['OS_ARCH']](**kwargs) - -def SourceIndex(fileStream, outputPath, vcs_root): - """Takes a list of files, writes info to a data block in a .stream file""" - # Creates a .pdb.stream file in the mozilla\objdir to be used for source indexing - # Create the srcsrv data block that indexes the pdb file - result = True - pdbStreamFile = open(outputPath, "w") - pdbStreamFile.write('''SRCSRV: ini ------------------------------------------------\r\nVERSION=2\r\nINDEXVERSION=2\r\nVERCTRL=http\r\nSRCSRV: variables ------------------------------------------\r\nHGSERVER=''') - pdbStreamFile.write(vcs_root) - pdbStreamFile.write('''\r\nSRCSRVVERCTRL=http\r\nHTTP_EXTRACT_TARGET=%hgserver%/raw-file/%var3%/%var2%\r\nSRCSRVTRG=%http_extract_target%\r\nSRCSRV: source files ---------------------------------------\r\n''') - pdbStreamFile.write(fileStream) # can't do string interpolation because the source server also uses this and so there are % in the above - pdbStreamFile.write("SRCSRV: end ------------------------------------------------\r\n\n") - pdbStreamFile.close() - return result - -def StartJob(dumper, lock, srcdirRepoInfo, func_name, args): - # Windows worker processes won't have run GlobalInit, - # and due to a lack of fork(), won't inherit the class - # variables from the parent, so set them here. - Dumper.lock = lock - Dumper.srcdirRepoInfo = srcdirRepoInfo - return getattr(dumper, func_name)(*args) - -class JobPool(object): - jobs = {} - executor = None - - @classmethod - def init(cls, executor): - cls.executor = executor - - @classmethod - def shutdown(cls): - cls.executor.shutdown() - - @classmethod - def submit(cls, args, callback): - cls.jobs[cls.executor.submit(StartJob, *args)] = callback - - @classmethod - def as_completed(cls): - '''Like concurrent.futures.as_completed, but allows adding new futures - between generator steps. Iteration will end when the generator has - yielded all completed futures and JobQueue.jobs is empty. - Yields (future, callback) pairs. - ''' - while cls.jobs: - completed, _ = concurrent.futures.wait(cls.jobs.keys(), return_when=concurrent.futures.FIRST_COMPLETED) - for f in completed: - callback = cls.jobs[f] - del cls.jobs[f] - yield f, callback - -class Dumper: - """This class can dump symbols from a file with debug info, and - store the output in a directory structure that is valid for use as - a Breakpad symbol server. Requires a path to a dump_syms binary-- - |dump_syms| and a directory to store symbols in--|symbol_path|. - Optionally takes a list of processor architectures to process from - each debug file--|archs|, the full path to the top source - directory--|srcdir|, for generating relative source file names, - and an option to copy debug info files alongside the dumped - symbol files--|copy_debug|, mostly useful for creating a - Microsoft Symbol Server from the resulting output. - - You don't want to use this directly if you intend to process files. - Instead, call GetPlatformSpecificDumper to get an instance of a - subclass. - - Processing is performed asynchronously via worker processes; in - order to wait for processing to finish and cleanup correctly, you - must call Finish after all ProcessFiles calls have been made. - You must also call Dumper.GlobalInit before creating or using any - instances.""" - def __init__(self, dump_syms, symbol_path, - archs=None, - srcdirs=[], - copy_debug=False, - vcsinfo=False, - srcsrv=False, - exclude=[], - repo_manifest=None, - file_mapping=None): - # popen likes absolute paths, at least on windows - self.dump_syms = os.path.abspath(dump_syms) - self.symbol_path = symbol_path - if archs is None: - # makes the loop logic simpler - self.archs = [''] - else: - self.archs = ['-a %s' % a for a in archs.split()] - self.srcdirs = [os.path.normpath(a) for a in srcdirs] - self.copy_debug = copy_debug - self.vcsinfo = vcsinfo - self.srcsrv = srcsrv - self.exclude = exclude[:] - if repo_manifest: - self.parse_repo_manifest(repo_manifest) - self.file_mapping = file_mapping or {} - - # book-keeping to keep track of the cleanup work per file tuple - self.files_record = {} - - @classmethod - def GlobalInit(cls, executor=concurrent.futures.ProcessPoolExecutor): - """Initialize the class globals for the multiprocessing setup; must - be called before any Dumper instances are created and used. Test cases - may pass in a different executor to use, usually - concurrent.futures.ThreadPoolExecutor.""" - num_cpus = multiprocessing.cpu_count() - if num_cpus is None: - # assume a dual core machine if we can't find out for some reason - # probably better on single core anyway due to I/O constraints - num_cpus = 2 - - # have to create any locks etc before the pool - manager = multiprocessing.Manager() - cls.lock = manager.RLock() - cls.srcdirRepoInfo = manager.dict() - JobPool.init(executor(max_workers=num_cpus)) - - def output(self, dest, output_str): - """Writes |output_str| to |dest|, holding |lock|; - terminates with a newline.""" - with Dumper.lock: - dest.write(output_str + "\n") - dest.flush() - - def output_pid(self, dest, output_str): - """Debugging output; prepends the pid to the string.""" - self.output(dest, "%d: %s" % (os.getpid(), output_str)) - - def parse_repo_manifest(self, repo_manifest): - """ - Parse an XML manifest of repository info as produced - by the `repo manifest -r` command. - """ - doc = parse(repo_manifest) - if doc.firstChild.tagName != "manifest": - return - # First, get remotes. - def ensure_slash(u): - if not u.endswith("/"): - return u + "/" - return u - remotes = dict([(r.getAttribute("name"), ensure_slash(r.getAttribute("fetch"))) for r in doc.getElementsByTagName("remote")]) - # And default remote. - default_remote = None - if doc.getElementsByTagName("default"): - default_remote = doc.getElementsByTagName("default")[0].getAttribute("remote") - # Now get projects. Assume they're relative to repo_manifest. - base_dir = os.path.abspath(os.path.dirname(repo_manifest)) - for proj in doc.getElementsByTagName("project"): - # name is the repository URL relative to the remote path. - name = proj.getAttribute("name") - # path is the path on-disk, relative to the manifest file. - path = proj.getAttribute("path") - # revision is the changeset ID. - rev = proj.getAttribute("revision") - # remote is the base URL to use. - remote = proj.getAttribute("remote") - # remote defaults to the <default remote>. - if not remote: - remote = default_remote - # path defaults to name. - if not path: - path = name - if not (name and path and rev and remote): - print "Skipping project %s" % proj.toxml() - continue - remote = remotes[remote] - # Turn git URLs into http URLs so that urljoin works. - if remote.startswith("git:"): - remote = "http" + remote[3:] - # Add this project to srcdirs. - srcdir = os.path.join(base_dir, path) - self.srcdirs.append(srcdir) - # And cache its VCS file info. Currently all repos mentioned - # in a repo manifest are assumed to be git. - root = urlparse.urljoin(remote, name) - Dumper.srcdirRepoInfo[srcdir] = GitRepoInfo(srcdir, rev, root) - - # subclasses override this - def ShouldProcess(self, file): - return not any(fnmatch.fnmatch(os.path.basename(file), exclude) for exclude in self.exclude) - - # and can override this - def ShouldSkipDir(self, dir): - return False - - def RunFileCommand(self, file): - """Utility function, returns the output of file(1)""" - try: - # we use -L to read the targets of symlinks, - # and -b to print just the content, not the filename - return os.popen("file -Lb " + file).read() - except: - return "" - - # This is a no-op except on Win32 - def FixFilenameCase(self, file): - return file - - # This is a no-op except on Win32 - def SourceServerIndexing(self, debug_file, guid, sourceFileStream, vcs_root): - return "" - - # subclasses override this if they want to support this - def CopyDebug(self, file, debug_file, guid, code_file, code_id): - pass - - def Finish(self, stop_pool=True): - '''Process all pending jobs and any jobs their callbacks submit. - By default, will shutdown the executor, but for testcases that - need multiple runs, pass stop_pool = False.''' - for job, callback in JobPool.as_completed(): - try: - res = job.result() - except Exception as e: - self.output(sys.stderr, 'Job raised exception: %s' % e) - continue - callback(res) - if stop_pool: - JobPool.shutdown() - - def Process(self, *args): - """Process files recursively in args.""" - # We collect all files to process first then sort by size to schedule - # larger files first because larger files tend to take longer and we - # don't like long pole stragglers. - files = set() - for arg in args: - for f in self.get_files_to_process(arg): - files.add(f) - - for f in sorted(files, key=os.path.getsize, reverse=True): - self.ProcessFiles((f,)) - - def get_files_to_process(self, file_or_dir): - """Generate the files to process from an input.""" - if os.path.isdir(file_or_dir) and not self.ShouldSkipDir(file_or_dir): - for f in self.get_files_to_process_in_dir(file_or_dir): - yield f - elif os.path.isfile(file_or_dir): - yield file_or_dir - - def get_files_to_process_in_dir(self, path): - """Generate the files to process in a directory. - - Valid files are are determined by calling ShouldProcess. - """ - for root, dirs, files in os.walk(path): - for d in dirs[:]: - if self.ShouldSkipDir(d): - dirs.remove(d) - for f in files: - fullpath = os.path.join(root, f) - if self.ShouldProcess(fullpath): - yield fullpath - - def SubmitJob(self, file_key, func_name, args, callback): - """Submits a job to the pool of workers""" - JobPool.submit((self, Dumper.lock, Dumper.srcdirRepoInfo, func_name, args), callback) - - def ProcessFilesFinished(self, res): - """Callback from multiprocesing when ProcessFilesWork finishes; - run the cleanup work, if any""" - # only run the cleanup function once per tuple of files - self.files_record[res['files']] += 1 - if self.files_record[res['files']] == len(self.archs): - del self.files_record[res['files']] - if res['after']: - res['after'](res['status'], res['after_arg']) - - def ProcessFiles(self, files, after=None, after_arg=None): - """Dump symbols from these files into a symbol file, stored - in the proper directory structure in |symbol_path|; processing is performed - asynchronously, and Finish must be called to wait for it complete and cleanup. - All files after the first are fallbacks in case the first file does not process - successfully; if it does, no other files will be touched.""" - self.output_pid(sys.stderr, "Submitting jobs for files: %s" % str(files)) - - # tries to get the vcs root from the .mozconfig first - if it's not set - # the tinderbox vcs path will be assigned further down - vcs_root = os.environ.get('MOZ_SOURCE_REPO') - for arch_num, arch in enumerate(self.archs): - self.files_record[files] = 0 # record that we submitted jobs for this tuple of files - self.SubmitJob(files[-1], 'ProcessFilesWork', args=(files, arch_num, arch, vcs_root, after, after_arg), callback=self.ProcessFilesFinished) - - def dump_syms_cmdline(self, file, arch, files): - ''' - Get the commandline used to invoke dump_syms. - ''' - # The Mac dumper overrides this. - return [self.dump_syms, file] - - def ProcessFilesWork(self, files, arch_num, arch, vcs_root, after, after_arg): - t_start = time.time() - self.output_pid(sys.stderr, "Worker processing files: %s" % (files,)) - - # our result is a status, a cleanup function, an argument to that function, and the tuple of files we were called on - result = { 'status' : False, 'after' : after, 'after_arg' : after_arg, 'files' : files } - - sourceFileStream = '' - code_id, code_file = None, None - for file in files: - # files is a tuple of files, containing fallbacks in case the first file doesn't process successfully - try: - cmd = self.dump_syms_cmdline(file, arch, files) - self.output_pid(sys.stderr, ' '.join(cmd)) - proc = subprocess.Popen(cmd, stdout=subprocess.PIPE, - stderr=open(os.devnull, 'wb')) - module_line = proc.stdout.next() - if module_line.startswith("MODULE"): - # MODULE os cpu guid debug_file - (guid, debug_file) = (module_line.split())[3:5] - # strip off .pdb extensions, and append .sym - sym_file = re.sub("\.pdb$", "", debug_file) + ".sym" - # we do want forward slashes here - rel_path = os.path.join(debug_file, - guid, - sym_file).replace("\\", "/") - full_path = os.path.normpath(os.path.join(self.symbol_path, - rel_path)) - try: - os.makedirs(os.path.dirname(full_path)) - except OSError: # already exists - pass - f = open(full_path, "w") - f.write(module_line) - # now process the rest of the output - for line in proc.stdout: - if line.startswith("FILE"): - # FILE index filename - (x, index, filename) = line.rstrip().split(None, 2) - filename = os.path.normpath(self.FixFilenameCase(filename)) - # We want original file paths for the source server. - sourcepath = filename - if filename in self.file_mapping: - filename = self.file_mapping[filename] - if self.vcsinfo: - (filename, rootname) = GetVCSFilename(filename, self.srcdirs) - # sets vcs_root in case the loop through files were to end on an empty rootname - if vcs_root is None: - if rootname: - vcs_root = rootname - # gather up files with hg for indexing - if filename.startswith("hg"): - (ver, checkout, source_file, revision) = filename.split(":", 3) - sourceFileStream += sourcepath + "*" + source_file + '*' + revision + "\r\n" - f.write("FILE %s %s\n" % (index, filename)) - elif line.startswith("INFO CODE_ID "): - # INFO CODE_ID code_id code_file - # This gives some info we can use to - # store binaries in the symbol store. - bits = line.rstrip().split(None, 3) - if len(bits) == 4: - code_id, code_file = bits[2:] - f.write(line) - else: - # pass through all other lines unchanged - f.write(line) - # we want to return true only if at least one line is not a MODULE or FILE line - result['status'] = True - f.close() - proc.wait() - # we output relative paths so callers can get a list of what - # was generated - self.output(sys.stdout, rel_path) - if self.srcsrv and vcs_root: - # add source server indexing to the pdb file - self.SourceServerIndexing(file, guid, sourceFileStream, vcs_root) - # only copy debug the first time if we have multiple architectures - if self.copy_debug and arch_num == 0: - self.CopyDebug(file, debug_file, guid, - code_file, code_id) - except StopIteration: - pass - except Exception as e: - self.output(sys.stderr, "Unexpected error: %s" % (str(e),)) - raise - if result['status']: - # we only need 1 file to work - break - - elapsed = time.time() - t_start - self.output_pid(sys.stderr, 'Worker finished processing %s in %.2fs' % - (files, elapsed)) - return result - -# Platform-specific subclasses. For the most part, these just have -# logic to determine what files to extract symbols from. - -class Dumper_Win32(Dumper): - fixedFilenameCaseCache = {} - - def ShouldProcess(self, file): - """This function will allow processing of pdb files that have dll - or exe files with the same base name next to them.""" - if not Dumper.ShouldProcess(self, file): - return False - if file.endswith(".pdb"): - (path,ext) = os.path.splitext(file) - if os.path.isfile(path + ".exe") or os.path.isfile(path + ".dll"): - return True - return False - - def FixFilenameCase(self, file): - """Recent versions of Visual C++ put filenames into - PDB files as all lowercase. If the file exists - on the local filesystem, fix it.""" - - # Use a cached version if we have one. - if file in self.fixedFilenameCaseCache: - return self.fixedFilenameCaseCache[file] - - result = file - - ctypes.windll.kernel32.SetErrorMode(ctypes.c_uint(1)) - if not isinstance(file, unicode): - file = unicode(file, sys.getfilesystemencoding()) - handle = ctypes.windll.kernel32.CreateFileW(file, - # GENERIC_READ - 0x80000000, - # FILE_SHARE_READ - 1, - None, - # OPEN_EXISTING - 3, - 0, - None) - if handle != -1: - size = ctypes.windll.kernel32.GetFinalPathNameByHandleW(handle, - None, - 0, - 0) - buf = ctypes.create_unicode_buffer(size) - if ctypes.windll.kernel32.GetFinalPathNameByHandleW(handle, - buf, - size, - 0) > 0: - # The return value of GetFinalPathNameByHandleW uses the - # '\\?\' prefix. - result = buf.value.encode(sys.getfilesystemencoding())[4:] - ctypes.windll.kernel32.CloseHandle(handle) - - # Cache the corrected version to avoid future filesystem hits. - self.fixedFilenameCaseCache[file] = result - return result - - def CopyDebug(self, file, debug_file, guid, code_file, code_id): - def compress(path): - compressed_file = path[:-1] + '_' - # ignore makecab's output - success = subprocess.call(["makecab.exe", "/D", - "CompressionType=MSZIP", - path, compressed_file], - stdout=open(os.devnull, 'w'), - stderr=subprocess.STDOUT) - if success == 0 and os.path.exists(compressed_file): - os.unlink(path) - return True - return False - - rel_path = os.path.join(debug_file, - guid, - debug_file).replace("\\", "/") - full_path = os.path.normpath(os.path.join(self.symbol_path, - rel_path)) - shutil.copyfile(file, full_path) - if compress(full_path): - self.output(sys.stdout, rel_path[:-1] + '_') - else: - self.output(sys.stdout, rel_path) - - # Copy the binary file as well - if code_file and code_id: - full_code_path = os.path.join(os.path.dirname(file), - code_file) - if os.path.exists(full_code_path): - rel_path = os.path.join(code_file, - code_id, - code_file).replace("\\", "/") - full_path = os.path.normpath(os.path.join(self.symbol_path, - rel_path)) - try: - os.makedirs(os.path.dirname(full_path)) - except OSError as e: - if e.errno != errno.EEXIST: - raise - shutil.copyfile(full_code_path, full_path) - if compress(full_path): - self.output(sys.stdout, rel_path[:-1] + '_') - else: - self.output(sys.stdout, rel_path) - - def SourceServerIndexing(self, debug_file, guid, sourceFileStream, vcs_root): - # Creates a .pdb.stream file in the mozilla\objdir to be used for source indexing - debug_file = os.path.abspath(debug_file) - streamFilename = debug_file + ".stream" - stream_output_path = os.path.abspath(streamFilename) - # Call SourceIndex to create the .stream file - result = SourceIndex(sourceFileStream, stream_output_path, vcs_root) - if self.copy_debug: - pdbstr_path = os.environ.get("PDBSTR_PATH") - pdbstr = os.path.normpath(pdbstr_path) - subprocess.call([pdbstr, "-w", "-p:" + os.path.basename(debug_file), - "-i:" + os.path.basename(streamFilename), "-s:srcsrv"], - cwd=os.path.dirname(stream_output_path)) - # clean up all the .stream files when done - os.remove(stream_output_path) - return result - -class Dumper_Linux(Dumper): - objcopy = os.environ['OBJCOPY'] if 'OBJCOPY' in os.environ else 'objcopy' - def ShouldProcess(self, file): - """This function will allow processing of files that are - executable, or end with the .so extension, and additionally - file(1) reports as being ELF files. It expects to find the file - command in PATH.""" - if not Dumper.ShouldProcess(self, file): - return False - if file.endswith(".so") or os.access(file, os.X_OK): - return self.RunFileCommand(file).startswith("ELF") - return False - - def CopyDebug(self, file, debug_file, guid, code_file, code_id): - # We want to strip out the debug info, and add a - # .gnu_debuglink section to the object, so the debugger can - # actually load our debug info later. - file_dbg = file + ".dbg" - if subprocess.call([self.objcopy, '--only-keep-debug', file, file_dbg]) == 0 and \ - subprocess.call([self.objcopy, '--add-gnu-debuglink=%s' % file_dbg, file]) == 0: - rel_path = os.path.join(debug_file, - guid, - debug_file + ".dbg") - full_path = os.path.normpath(os.path.join(self.symbol_path, - rel_path)) - shutil.move(file_dbg, full_path) - # gzip the shipped debug files - os.system("gzip -4 -f %s" % full_path) - self.output(sys.stdout, rel_path + ".gz") - else: - if os.path.isfile(file_dbg): - os.unlink(file_dbg) - -class Dumper_Solaris(Dumper): - def RunFileCommand(self, file): - """Utility function, returns the output of file(1)""" - try: - output = os.popen("file " + file).read() - return output.split('\t')[1]; - except: - return "" - - def ShouldProcess(self, file): - """This function will allow processing of files that are - executable, or end with the .so extension, and additionally - file(1) reports as being ELF files. It expects to find the file - command in PATH.""" - if not Dumper.ShouldProcess(self, file): - return False - if file.endswith(".so") or os.access(file, os.X_OK): - return self.RunFileCommand(file).startswith("ELF") - return False - -def AfterMac(status, dsymbundle): - """Cleanup function to run on Macs after we process the file(s).""" - # CopyDebug will already have been run from Dumper.ProcessFiles - shutil.rmtree(dsymbundle) - -class Dumper_Mac(Dumper): - def ShouldProcess(self, file): - """This function will allow processing of files that are - executable, or end with the .dylib extension, and additionally - file(1) reports as being Mach-O files. It expects to find the file - command in PATH.""" - if not Dumper.ShouldProcess(self, file): - return False - if file.endswith(".dylib") or os.access(file, os.X_OK): - return self.RunFileCommand(file).startswith("Mach-O") - return False - - def ShouldSkipDir(self, dir): - """We create .dSYM bundles on the fly, but if someone runs - buildsymbols twice, we should skip any bundles we created - previously, otherwise we'll recurse into them and try to - dump the inner bits again.""" - if dir.endswith(".dSYM"): - return True - return False - - def ProcessFiles(self, files, after=None, after_arg=None): - # also note, files must be len 1 here, since we're the only ones - # that ever add more than one file to the list - self.output_pid(sys.stderr, "Submitting job for Mac pre-processing on file: %s" % (files[0])) - self.SubmitJob(files[0], 'ProcessFilesWorkMac', args=(files[0],), callback=self.ProcessFilesMacFinished) - - def ProcessFilesMacFinished(self, result): - if result['status']: - # kick off new jobs per-arch with our new list of files - Dumper.ProcessFiles(self, result['files'], after=AfterMac, after_arg=result['files'][0]) - - def dump_syms_cmdline(self, file, arch, files): - ''' - Get the commandline used to invoke dump_syms. - ''' - # dump_syms wants the path to the original binary and the .dSYM - # in order to dump all the symbols. - if len(files) == 2 and file == files[0] and file.endswith('.dSYM'): - # This is the .dSYM bundle. - return [self.dump_syms] + arch.split() + ['-g', file, files[1]] - return Dumper.dump_syms_cmdline(self, file, arch, files) - - def ProcessFilesWorkMac(self, file): - """dump_syms on Mac needs to be run on a dSYM bundle produced - by dsymutil(1), so run dsymutil here and pass the bundle name - down to the superclass method instead.""" - t_start = time.time() - self.output_pid(sys.stderr, "Worker running Mac pre-processing on file: %s" % (file,)) - - # our return is a status and a tuple of files to dump symbols for - # the extra files are fallbacks; as soon as one is dumped successfully, we stop - result = { 'status' : False, 'files' : None, 'file_key' : file } - dsymbundle = file + ".dSYM" - if os.path.exists(dsymbundle): - shutil.rmtree(dsymbundle) - dsymutil = buildconfig.substs['DSYMUTIL'] - # dsymutil takes --arch=foo instead of -a foo like everything else - try: - cmd = ([dsymutil] + - [a.replace('-a ', '--arch=') for a in self.archs if a] + - [file]) - self.output_pid(sys.stderr, ' '.join(cmd)) - subprocess.check_call(cmd, stdout=open(os.devnull, 'w')) - except subprocess.CalledProcessError as e: - self.output_pid(sys.stderr, 'Error running dsymutil: %s' % str(e)) - - if not os.path.exists(dsymbundle): - # dsymutil won't produce a .dSYM for files without symbols - self.output_pid(sys.stderr, "No symbols found in file: %s" % (file,)) - result['status'] = False - result['files'] = (file, ) - return result - - result['status'] = True - result['files'] = (dsymbundle, file) - elapsed = time.time() - t_start - self.output_pid(sys.stderr, 'Worker finished processing %s in %.2fs' % - (file, elapsed)) - return result - - def CopyDebug(self, file, debug_file, guid, code_file, code_id): - """ProcessFiles has already produced a dSYM bundle, so we should just - copy that to the destination directory. However, we'll package it - into a .tar.bz2 because the debug symbols are pretty huge, and - also because it's a bundle, so it's a directory. |file| here is the - dSYM bundle, and |debug_file| is the original filename.""" - rel_path = os.path.join(debug_file, - guid, - os.path.basename(file) + ".tar.bz2") - full_path = os.path.abspath(os.path.join(self.symbol_path, - rel_path)) - success = subprocess.call(["tar", "cjf", full_path, os.path.basename(file)], - cwd=os.path.dirname(file), - stdout=open(os.devnull, 'w'), stderr=subprocess.STDOUT) - if success == 0 and os.path.exists(full_path): - self.output(sys.stdout, rel_path) - -# Entry point if called as a standalone program -def main(): - parser = OptionParser(usage="usage: %prog [options] <dump_syms binary> <symbol store path> <debug info files>") - parser.add_option("-c", "--copy", - action="store_true", dest="copy_debug", default=False, - help="Copy debug info files into the same directory structure as symbol files") - parser.add_option("-a", "--archs", - action="store", dest="archs", - help="Run dump_syms -a <arch> for each space separated cpu architecture in ARCHS (only on OS X)") - parser.add_option("-s", "--srcdir", - action="append", dest="srcdir", default=[], - help="Use SRCDIR to determine relative paths to source files") - parser.add_option("-v", "--vcs-info", - action="store_true", dest="vcsinfo", - help="Try to retrieve VCS info for each FILE listed in the output") - parser.add_option("-i", "--source-index", - action="store_true", dest="srcsrv", default=False, - help="Add source index information to debug files, making them suitable for use in a source server.") - parser.add_option("-x", "--exclude", - action="append", dest="exclude", default=[], metavar="PATTERN", - help="Skip processing files matching PATTERN.") - parser.add_option("--repo-manifest", - action="store", dest="repo_manifest", - help="""Get source information from this XML manifest -produced by the `repo manifest -r` command. -""") - parser.add_option("--install-manifest", - action="append", dest="install_manifests", - default=[], - help="""Use this install manifest to map filenames back -to canonical locations in the source repository. Specify -<install manifest filename>,<install destination> as a comma-separated pair. -""") - (options, args) = parser.parse_args() - - #check to see if the pdbstr.exe exists - if options.srcsrv: - pdbstr = os.environ.get("PDBSTR_PATH") - if not os.path.exists(pdbstr): - print >> sys.stderr, "Invalid path to pdbstr.exe - please set/check PDBSTR_PATH.\n" - sys.exit(1) - - if len(args) < 3: - parser.error("not enough arguments") - exit(1) - - try: - manifests = validate_install_manifests(options.install_manifests) - except (IOError, ValueError) as e: - parser.error(str(e)) - exit(1) - file_mapping = make_file_mapping(manifests) - dumper = GetPlatformSpecificDumper(dump_syms=args[0], - symbol_path=args[1], - copy_debug=options.copy_debug, - archs=options.archs, - srcdirs=options.srcdir, - vcsinfo=options.vcsinfo, - srcsrv=options.srcsrv, - exclude=options.exclude, - repo_manifest=options.repo_manifest, - file_mapping=file_mapping) - - dumper.Process(*args[2:]) - dumper.Finish() - -# run main if run directly -if __name__ == "__main__": - # set up the multiprocessing infrastructure before we start; - # note that this needs to be in the __main__ guard, or else Windows will choke - Dumper.GlobalInit() - - main() diff --git a/toolkit/crashreporter/tools/unit-symbolstore.py b/toolkit/crashreporter/tools/unit-symbolstore.py deleted file mode 100644 index 021efbeeb..000000000 --- a/toolkit/crashreporter/tools/unit-symbolstore.py +++ /dev/null @@ -1,583 +0,0 @@ -#!/usr/bin/env python -# This Source Code Form is subject to the terms of the Mozilla Public -# License, v. 2.0. If a copy of the MPL was not distributed with this -# file, You can obtain one at http://mozilla.org/MPL/2.0/. - -import concurrent.futures -import mock -import mozunit -import os -import platform -import shutil -import struct -import subprocess -import sys -import tempfile -import unittest - -from mock import patch -from mozpack.manifests import InstallManifest - -import symbolstore - -# Some simple functions to mock out files that the platform-specific dumpers will accept. -# dump_syms itself will not be run (we mock that call out), but we can't override -# the ShouldProcessFile method since we actually want to test that. -def write_elf(filename): - open(filename, "wb").write(struct.pack("<7B45x", 0x7f, ord("E"), ord("L"), ord("F"), 1, 1, 1)) - -def write_macho(filename): - open(filename, "wb").write(struct.pack("<I28x", 0xfeedface)) - -def write_pdb(filename): - open(filename, "w").write("aaa") - # write out a fake DLL too - open(os.path.splitext(filename)[0] + ".dll", "w").write("aaa") - -writer = {'Windows': write_pdb, - 'Microsoft': write_pdb, - 'Linux': write_elf, - 'Sunos5': write_elf, - 'Darwin': write_macho}[platform.system()] -extension = {'Windows': ".pdb", - 'Microsoft': ".pdb", - 'Linux': ".so", - 'Sunos5': ".so", - 'Darwin': ".dylib"}[platform.system()] - -def add_extension(files): - return [f + extension for f in files] - -class HelperMixin(object): - """ - Test that passing filenames to exclude from processing works. - """ - def setUp(self): - self.test_dir = tempfile.mkdtemp() - if not self.test_dir.endswith(os.sep): - self.test_dir += os.sep - symbolstore.srcdirRepoInfo = {} - symbolstore.vcsFileInfoCache = {} - - # Remove environment variables that can influence tests. - for e in ('MOZ_SOURCE_CHANGESET', 'MOZ_SOURCE_REPO'): - try: - del os.environ[e] - except KeyError: - pass - - def tearDown(self): - shutil.rmtree(self.test_dir) - symbolstore.srcdirRepoInfo = {} - symbolstore.vcsFileInfoCache = {} - - def make_dirs(self, f): - d = os.path.dirname(f) - if d and not os.path.exists(d): - os.makedirs(d) - - def add_test_files(self, files): - for f in files: - f = os.path.join(self.test_dir, f) - self.make_dirs(f) - writer(f) - -class TestSizeOrder(HelperMixin, unittest.TestCase): - def test_size_order(self): - """ - Test that files are processed ordered by size on disk. - """ - processed = [] - def mock_process_file(filenames): - for filename in filenames: - processed.append((filename[len(self.test_dir):] if filename.startswith(self.test_dir) else filename).replace('\\', '/')) - return True - for f, size in (('a/one', 10), ('b/c/two', 30), ('c/three', 20)): - f = os.path.join(self.test_dir, f) - d = os.path.dirname(f) - if d and not os.path.exists(d): - os.makedirs(d) - open(f, 'wb').write('x' * size) - d = symbolstore.GetPlatformSpecificDumper(dump_syms="dump_syms", - symbol_path="symbol_path") - d.ShouldProcess = lambda f: True - d.ProcessFiles = mock_process_file - d.Process(self.test_dir) - d.Finish(stop_pool=False) - self.assertEqual(processed, ['b/c/two', 'c/three', 'a/one']) - - -class TestExclude(HelperMixin, unittest.TestCase): - def test_exclude_wildcard(self): - """ - Test that using an exclude list with a wildcard pattern works. - """ - processed = [] - def mock_process_file(filenames): - for filename in filenames: - processed.append((filename[len(self.test_dir):] if filename.startswith(self.test_dir) else filename).replace('\\', '/')) - return True - self.add_test_files(add_extension(["foo", "bar", "abc/xyz", "abc/fooxyz", "def/asdf", "def/xyzfoo"])) - d = symbolstore.GetPlatformSpecificDumper(dump_syms="dump_syms", - symbol_path="symbol_path", - exclude=["*foo*"]) - d.ProcessFiles = mock_process_file - d.Process(self.test_dir) - d.Finish(stop_pool=False) - processed.sort() - expected = add_extension(["bar", "abc/xyz", "def/asdf"]) - expected.sort() - self.assertEqual(processed, expected) - - - def test_exclude_filenames(self): - """ - Test that excluding a filename without a wildcard works. - """ - processed = [] - def mock_process_file(filenames): - for filename in filenames: - processed.append((filename[len(self.test_dir):] if filename.startswith(self.test_dir) else filename).replace('\\', '/')) - return True - self.add_test_files(add_extension(["foo", "bar", "abc/foo", "abc/bar", "def/foo", "def/bar"])) - d = symbolstore.GetPlatformSpecificDumper(dump_syms="dump_syms", - symbol_path="symbol_path", - exclude=add_extension(["foo"])) - d.ProcessFiles = mock_process_file - d.Process(self.test_dir) - d.Finish(stop_pool=False) - processed.sort() - expected = add_extension(["bar", "abc/bar", "def/bar"]) - expected.sort() - self.assertEqual(processed, expected) - - -def mock_dump_syms(module_id, filename, extra=[]): - return ["MODULE os x86 %s %s" % (module_id, filename) - ] + extra + [ - "FILE 0 foo.c", - "PUBLIC xyz 123"] - - -class TestCopyDebug(HelperMixin, unittest.TestCase): - def setUp(self): - HelperMixin.setUp(self) - self.symbol_dir = tempfile.mkdtemp() - self.mock_call = patch("subprocess.call").start() - self.stdouts = [] - self.mock_popen = patch("subprocess.Popen").start() - stdout_iter = self.next_mock_stdout() - def next_popen(*args, **kwargs): - m = mock.MagicMock() - m.stdout = stdout_iter.next() - m.wait.return_value = 0 - return m - self.mock_popen.side_effect = next_popen - shutil.rmtree = patch("shutil.rmtree").start() - - def tearDown(self): - HelperMixin.tearDown(self) - patch.stopall() - shutil.rmtree(self.symbol_dir) - - def next_mock_stdout(self): - if not self.stdouts: - yield iter([]) - for s in self.stdouts: - yield iter(s) - - def test_copy_debug_universal(self): - """ - Test that dumping symbols for multiple architectures only copies debug symbols once - per file. - """ - copied = [] - def mock_copy_debug(filename, debug_file, guid, code_file, code_id): - copied.append(filename[len(self.symbol_dir):] if filename.startswith(self.symbol_dir) else filename) - self.add_test_files(add_extension(["foo"])) - self.stdouts.append(mock_dump_syms("X" * 33, add_extension(["foo"])[0])) - self.stdouts.append(mock_dump_syms("Y" * 33, add_extension(["foo"])[0])) - def mock_dsymutil(args, **kwargs): - filename = args[-1] - os.makedirs(filename + ".dSYM") - return 0 - self.mock_call.side_effect = mock_dsymutil - d = symbolstore.GetPlatformSpecificDumper(dump_syms="dump_syms", - symbol_path=self.symbol_dir, - copy_debug=True, - archs="abc xyz") - d.CopyDebug = mock_copy_debug - d.Process(self.test_dir) - d.Finish(stop_pool=False) - self.assertEqual(1, len(copied)) - - def test_copy_debug_copies_binaries(self): - """ - Test that CopyDebug copies binaries as well on Windows. - """ - test_file = os.path.join(self.test_dir, 'foo.pdb') - write_pdb(test_file) - code_file = 'foo.dll' - code_id = 'abc123' - self.stdouts.append(mock_dump_syms('X' * 33, 'foo.pdb', - ['INFO CODE_ID %s %s' % (code_id, code_file)])) - def mock_compress(args, **kwargs): - filename = args[-1] - open(filename, 'w').write('stuff') - return 0 - self.mock_call.side_effect = mock_compress - d = symbolstore.Dumper_Win32(dump_syms='dump_syms', - symbol_path=self.symbol_dir, - copy_debug=True) - d.FixFilenameCase = lambda f: f - d.Process(self.test_dir) - d.Finish(stop_pool=False) - self.assertTrue(os.path.isfile(os.path.join(self.symbol_dir, code_file, code_id, code_file[:-1] + '_'))) - -class TestGetVCSFilename(HelperMixin, unittest.TestCase): - def setUp(self): - HelperMixin.setUp(self) - - def tearDown(self): - HelperMixin.tearDown(self) - - @patch("subprocess.Popen") - def testVCSFilenameHg(self, mock_Popen): - # mock calls to `hg parent` and `hg showconfig paths.default` - mock_communicate = mock_Popen.return_value.communicate - mock_communicate.side_effect = [("abcd1234", ""), - ("http://example.com/repo", "")] - os.mkdir(os.path.join(self.test_dir, ".hg")) - filename = os.path.join(self.test_dir, "foo.c") - self.assertEqual("hg:example.com/repo:foo.c:abcd1234", - symbolstore.GetVCSFilename(filename, [self.test_dir])[0]) - - @patch("subprocess.Popen") - def testVCSFilenameHgMultiple(self, mock_Popen): - # mock calls to `hg parent` and `hg showconfig paths.default` - mock_communicate = mock_Popen.return_value.communicate - mock_communicate.side_effect = [("abcd1234", ""), - ("http://example.com/repo", ""), - ("0987ffff", ""), - ("http://example.com/other", "")] - srcdir1 = os.path.join(self.test_dir, "one") - srcdir2 = os.path.join(self.test_dir, "two") - os.makedirs(os.path.join(srcdir1, ".hg")) - os.makedirs(os.path.join(srcdir2, ".hg")) - filename1 = os.path.join(srcdir1, "foo.c") - filename2 = os.path.join(srcdir2, "bar.c") - self.assertEqual("hg:example.com/repo:foo.c:abcd1234", - symbolstore.GetVCSFilename(filename1, [srcdir1, srcdir2])[0]) - self.assertEqual("hg:example.com/other:bar.c:0987ffff", - symbolstore.GetVCSFilename(filename2, [srcdir1, srcdir2])[0]) - - def testVCSFilenameEnv(self): - # repo URL and changeset read from environment variables if defined. - os.environ['MOZ_SOURCE_REPO'] = 'https://somewhere.com/repo' - os.environ['MOZ_SOURCE_CHANGESET'] = 'abcdef0123456' - os.mkdir(os.path.join(self.test_dir, '.hg')) - filename = os.path.join(self.test_dir, 'foo.c') - self.assertEqual('hg:somewhere.com/repo:foo.c:abcdef0123456', - symbolstore.GetVCSFilename(filename, [self.test_dir])[0]) - - -class TestRepoManifest(HelperMixin, unittest.TestCase): - def testRepoManifest(self): - manifest = os.path.join(self.test_dir, "sources.xml") - open(manifest, "w").write("""<?xml version="1.0" encoding="UTF-8"?> -<manifest> -<remote fetch="http://example.com/foo/" name="foo"/> -<remote fetch="git://example.com/bar/" name="bar"/> -<default remote="bar"/> -<project name="projects/one" revision="abcd1234"/> -<project name="projects/two" path="projects/another" revision="ffffffff" remote="foo"/> -<project name="something_else" revision="00000000" remote="bar"/> -</manifest> -""") - # Use a source file from each of the three projects - file1 = os.path.join(self.test_dir, "projects", "one", "src1.c") - file2 = os.path.join(self.test_dir, "projects", "another", "src2.c") - file3 = os.path.join(self.test_dir, "something_else", "src3.c") - d = symbolstore.Dumper("dump_syms", "symbol_path", - repo_manifest=manifest) - self.assertEqual("git:example.com/bar/projects/one:src1.c:abcd1234", - symbolstore.GetVCSFilename(file1, d.srcdirs)[0]) - self.assertEqual("git:example.com/foo/projects/two:src2.c:ffffffff", - symbolstore.GetVCSFilename(file2, d.srcdirs)[0]) - self.assertEqual("git:example.com/bar/something_else:src3.c:00000000", - symbolstore.GetVCSFilename(file3, d.srcdirs)[0]) - -if platform.system() in ("Windows", "Microsoft"): - class TestFixFilenameCase(HelperMixin, unittest.TestCase): - def test_fix_filename_case(self): - # self.test_dir is going to be 8.3 paths... - junk = os.path.join(self.test_dir, 'x') - with open(junk, 'wb') as o: - o.write('x') - d = symbolstore.Dumper_Win32(dump_syms='dump_syms', - symbol_path=self.test_dir) - fixed_dir = os.path.dirname(d.FixFilenameCase(junk)) - files = [ - 'one\\two.c', - 'three\\Four.d', - 'Five\\Six.e', - 'seven\\Eight\\nine.F', - ] - for rel_path in files: - full_path = os.path.normpath(os.path.join(self.test_dir, - rel_path)) - self.make_dirs(full_path) - with open(full_path, 'wb') as o: - o.write('x') - fixed_path = d.FixFilenameCase(full_path.lower()) - fixed_path = os.path.relpath(fixed_path, fixed_dir) - self.assertEqual(rel_path, fixed_path) - - class TestSourceServer(HelperMixin, unittest.TestCase): - @patch("subprocess.call") - @patch("subprocess.Popen") - def test_HGSERVER(self, mock_Popen, mock_call): - """ - Test that HGSERVER gets set correctly in the source server index. - """ - symbolpath = os.path.join(self.test_dir, "symbols") - os.makedirs(symbolpath) - srcdir = os.path.join(self.test_dir, "srcdir") - os.makedirs(os.path.join(srcdir, ".hg")) - sourcefile = os.path.join(srcdir, "foo.c") - test_files = add_extension(["foo"]) - self.add_test_files(test_files) - # srcsrv needs PDBSTR_PATH set - os.environ["PDBSTR_PATH"] = "pdbstr" - # mock calls to `dump_syms`, `hg parent` and - # `hg showconfig paths.default` - mock_Popen.return_value.stdout = iter([ - "MODULE os x86 %s %s" % ("X" * 33, test_files[0]), - "FILE 0 %s" % sourcefile, - "PUBLIC xyz 123" - ]) - mock_communicate = mock_Popen.return_value.communicate - mock_communicate.side_effect = [("abcd1234", ""), - ("http://example.com/repo", ""), - ] - # And mock the call to pdbstr to capture the srcsrv stream data. - global srcsrv_stream - srcsrv_stream = None - def mock_pdbstr(args, cwd="", **kwargs): - for arg in args: - if arg.startswith("-i:"): - global srcsrv_stream - srcsrv_stream = open(os.path.join(cwd, arg[3:]), 'r').read() - return 0 - mock_call.side_effect = mock_pdbstr - d = symbolstore.GetPlatformSpecificDumper(dump_syms="dump_syms", - symbol_path=symbolpath, - srcdirs=[srcdir], - vcsinfo=True, - srcsrv=True, - copy_debug=True) - # stub out CopyDebug - d.CopyDebug = lambda *args: True - d.Process(self.test_dir) - d.Finish(stop_pool=False) - self.assertNotEqual(srcsrv_stream, None) - hgserver = [x.rstrip() for x in srcsrv_stream.splitlines() if x.startswith("HGSERVER=")] - self.assertEqual(len(hgserver), 1) - self.assertEqual(hgserver[0].split("=")[1], "http://example.com/repo") - -class TestInstallManifest(HelperMixin, unittest.TestCase): - def setUp(self): - HelperMixin.setUp(self) - self.srcdir = os.path.join(self.test_dir, 'src') - os.mkdir(self.srcdir) - self.objdir = os.path.join(self.test_dir, 'obj') - os.mkdir(self.objdir) - self.manifest = InstallManifest() - self.canonical_mapping = {} - for s in ['src1', 'src2']: - srcfile = os.path.join(self.srcdir, s) - objfile = os.path.join(self.objdir, s) - self.canonical_mapping[objfile] = srcfile - self.manifest.add_copy(srcfile, s) - self.manifest_file = os.path.join(self.test_dir, 'install-manifest') - self.manifest.write(self.manifest_file) - - def testMakeFileMapping(self): - ''' - Test that valid arguments are validated. - ''' - arg = '%s,%s' % (self.manifest_file, self.objdir) - ret = symbolstore.validate_install_manifests([arg]) - self.assertEqual(len(ret), 1) - manifest, dest = ret[0] - self.assertTrue(isinstance(manifest, InstallManifest)) - self.assertEqual(dest, self.objdir) - - file_mapping = symbolstore.make_file_mapping(ret) - for obj, src in self.canonical_mapping.iteritems(): - self.assertTrue(obj in file_mapping) - self.assertEqual(file_mapping[obj], src) - - def testMissingFiles(self): - ''' - Test that missing manifest files or install directories give errors. - ''' - missing_manifest = os.path.join(self.test_dir, 'missing-manifest') - arg = '%s,%s' % (missing_manifest, self.objdir) - with self.assertRaises(IOError) as e: - symbolstore.validate_install_manifests([arg]) - self.assertEqual(e.filename, missing_manifest) - - missing_install_dir = os.path.join(self.test_dir, 'missing-dir') - arg = '%s,%s' % (self.manifest_file, missing_install_dir) - with self.assertRaises(IOError) as e: - symbolstore.validate_install_manifests([arg]) - self.assertEqual(e.filename, missing_install_dir) - - def testBadManifest(self): - ''' - Test that a bad manifest file give errors. - ''' - bad_manifest = os.path.join(self.test_dir, 'bad-manifest') - with open(bad_manifest, 'wb') as f: - f.write('junk\n') - arg = '%s,%s' % (bad_manifest, self.objdir) - with self.assertRaises(IOError) as e: - symbolstore.validate_install_manifests([arg]) - self.assertEqual(e.filename, bad_manifest) - - def testBadArgument(self): - ''' - Test that a bad manifest argument gives an error. - ''' - with self.assertRaises(ValueError) as e: - symbolstore.validate_install_manifests(['foo']) - -class TestFileMapping(HelperMixin, unittest.TestCase): - def setUp(self): - HelperMixin.setUp(self) - self.srcdir = os.path.join(self.test_dir, 'src') - os.mkdir(self.srcdir) - self.objdir = os.path.join(self.test_dir, 'obj') - os.mkdir(self.objdir) - self.symboldir = os.path.join(self.test_dir, 'symbols') - os.mkdir(self.symboldir) - - @patch("subprocess.Popen") - def testFileMapping(self, mock_Popen): - files = [('a/b', 'mozilla/b'), - ('c/d', 'foo/d')] - if os.sep != '/': - files = [[f.replace('/', os.sep) for f in x] for x in files] - file_mapping = {} - dumped_files = [] - expected_files = [] - for s, o in files: - srcfile = os.path.join(self.srcdir, s) - expected_files.append(srcfile) - file_mapping[os.path.join(self.objdir, o)] = srcfile - dumped_files.append(os.path.join(self.objdir, 'x', 'y', - '..', '..', o)) - # mock the dump_syms output - file_id = ("X" * 33, 'somefile') - def mk_output(files): - return iter( - [ - 'MODULE os x86 %s %s\n' % file_id - ] + - [ - 'FILE %d %s\n' % (i,s) for i, s in enumerate(files) - ] + - [ - 'PUBLIC xyz 123\n' - ] - ) - mock_Popen.return_value.stdout = mk_output(dumped_files) - - d = symbolstore.Dumper('dump_syms', self.symboldir, - file_mapping=file_mapping) - f = os.path.join(self.objdir, 'somefile') - open(f, 'wb').write('blah') - d.Process(f) - d.Finish(stop_pool=False) - expected_output = ''.join(mk_output(expected_files)) - symbol_file = os.path.join(self.symboldir, - file_id[1], file_id[0], file_id[1] + '.sym') - self.assertEqual(open(symbol_file, 'r').read(), expected_output) - -class TestFunctional(HelperMixin, unittest.TestCase): - '''Functional tests of symbolstore.py, calling it with a real - dump_syms binary and passing in a real binary to dump symbols from. - - Since the rest of the tests in this file mock almost everything and - don't use the actual process pool like buildsymbols does, this tests - that the way symbolstore.py gets called in buildsymbols works. - ''' - def setUp(self): - HelperMixin.setUp(self) - import buildconfig - self.skip_test = False - if buildconfig.substs['MOZ_BUILD_APP'] != 'browser': - self.skip_test = True - self.topsrcdir = buildconfig.topsrcdir - self.script_path = os.path.join(self.topsrcdir, 'toolkit', - 'crashreporter', 'tools', - 'symbolstore.py') - if platform.system() in ("Windows", "Microsoft"): - if buildconfig.substs['MSVC_HAS_DIA_SDK']: - self.dump_syms = os.path.join(buildconfig.topobjdir, - 'dist', 'host', 'bin', - 'dump_syms.exe') - else: - self.dump_syms = os.path.join(self.topsrcdir, - 'toolkit', - 'crashreporter', - 'tools', - 'win32', - 'dump_syms_vc{_MSC_VER}.exe'.format(**buildconfig.substs)) - self.target_bin = os.path.join(buildconfig.topobjdir, - 'browser', - 'app', - 'firefox.pdb') - else: - self.dump_syms = os.path.join(buildconfig.topobjdir, - 'dist', 'host', 'bin', - 'dump_syms') - self.target_bin = os.path.join(buildconfig.topobjdir, - 'dist', 'bin', 'firefox') - - - def tearDown(self): - HelperMixin.tearDown(self) - - def testSymbolstore(self): - if self.skip_test: - raise unittest.SkipTest('Skipping test in non-Firefox product') - output = subprocess.check_output([sys.executable, - self.script_path, - '--vcs-info', - '-s', self.topsrcdir, - self.dump_syms, - self.test_dir, - self.target_bin], - stderr=open(os.devnull, 'w')) - lines = filter(lambda x: x.strip(), output.splitlines()) - self.assertEqual(1, len(lines), - 'should have one filename in the output') - symbol_file = os.path.join(self.test_dir, lines[0]) - self.assertTrue(os.path.isfile(symbol_file)) - symlines = open(symbol_file, 'r').readlines() - file_lines = filter(lambda x: x.startswith('FILE') and 'nsBrowserApp.cpp' in x, symlines) - self.assertEqual(len(file_lines), 1, - 'should have nsBrowserApp.cpp FILE line') - filename = file_lines[0].split(None, 2)[2] - self.assertEqual('hg:', filename[:3]) - - -if __name__ == '__main__': - # use ThreadPoolExecutor to use threading instead of processes so - # that our mocking/module-patching works. - symbolstore.Dumper.GlobalInit(concurrent.futures.ThreadPoolExecutor) - - mozunit.main() - diff --git a/toolkit/crashreporter/tools/upload_symbols.py b/toolkit/crashreporter/tools/upload_symbols.py deleted file mode 100644 index 7e94b238d..000000000 --- a/toolkit/crashreporter/tools/upload_symbols.py +++ /dev/null @@ -1,102 +0,0 @@ -#!/usr/bin/env python -# -# This Source Code Form is subject to the terms of the Mozilla Public -# License, v. 2.0. If a copy of the MPL was not distributed with this -# file, You can obtain one at http://mozilla.org/MPL/2.0/. -# -# This script uploads a symbol zip file passed on the commandline -# to the Socorro symbol upload API hosted on crash-stats.mozilla.org. -# -# Using this script requires you to have generated an authentication -# token in the crash-stats web interface. You must put the token in a file -# and set SOCORRO_SYMBOL_UPLOAD_TOKEN_FILE to the path to the file in -# the mozconfig you're using. - -from __future__ import print_function - -import os -import redo -import requests -import sys - -try: - from buildconfig import substs -except ImportError: - # Allow standalone use of this script, for use in TaskCluster - from os import environ as substs - -url = 'https://symbols.mozilla.org/upload/' -# Allow overwriting of the upload url with an environmental variable -if 'SOCORRO_SYMBOL_UPLOAD_URL' in os.environ: - url = os.environ['SOCORRO_SYMBOL_UPLOAD_URL'] -MAX_RETRIES = 5 - -def print_error(r): - if r.status_code < 400: - print('Error: bad auth token? ({0}: {1})'.format(r.status_code, - r.reason), - file=sys.stderr) - else: - print('Error: got HTTP response {0}: {1}'.format(r.status_code, - r.reason), - file=sys.stderr) - - print('Response body:\n{sep}\n{body}\n{sep}\n'.format( - sep='=' * 20, - body=r.text - )) - -def main(): - if len(sys.argv) != 2: - print('Usage: uploadsymbols.py <zip file>', file=sys.stderr) - return 1 - - if not os.path.isfile(sys.argv[1]): - print('Error: zip file "{0}" does not exist!'.format(sys.argv[1]), - file=sys.stderr) - return 1 - symbols_zip = sys.argv[1] - - if 'SOCORRO_SYMBOL_UPLOAD_TOKEN_FILE' not in substs: - print('Error: you must set SOCORRO_SYMBOL_UPLOAD_TOKEN_FILE in your mozconfig!', file=sys.stderr) - return 1 - token_file = substs['SOCORRO_SYMBOL_UPLOAD_TOKEN_FILE'] - - if not os.path.isfile(token_file): - print('Error: SOCORRO_SYMBOL_UPLOAD_TOKEN_FILE "{0}" does not exist!'.format(token_file), file=sys.stderr) - return 1 - auth_token = open(token_file, 'r').read().strip() - - print('Uploading symbol file "{0}" to "{1}"'.format(sys.argv[1], url)) - - for i, _ in enumerate(redo.retrier(attempts=MAX_RETRIES), start=1): - print('Attempt %d of %d...' % (i, MAX_RETRIES)) - try: - r = requests.post( - url, - files={'symbols.zip': open(sys.argv[1], 'rb')}, - headers={'Auth-Token': auth_token}, - allow_redirects=False, - timeout=120) - # 500 is likely to be a transient failure. - # Break out for success or other error codes. - if r.status_code < 500: - break - print_error(r) - except requests.exceptions.RequestException as e: - print('Error: {0}'.format(e)) - print('Retrying...') - else: - print('Maximum retries hit, giving up!') - return 1 - - if r.status_code >= 200 and r.status_code < 300: - print('Uploaded successfully!') - return 0 - - print_error(r) - return 1 - -if __name__ == '__main__': - sys.exit(main()) - diff --git a/toolkit/crashreporter/tools/win32/dump_syms_vc1600.exe b/toolkit/crashreporter/tools/win32/dump_syms_vc1600.exe Binary files differdeleted file mode 100755 index 489cc4c4a..000000000 --- a/toolkit/crashreporter/tools/win32/dump_syms_vc1600.exe +++ /dev/null diff --git a/toolkit/crashreporter/tools/win32/dump_syms_vc1700.exe b/toolkit/crashreporter/tools/win32/dump_syms_vc1700.exe Binary files differdeleted file mode 100644 index 0ead28957..000000000 --- a/toolkit/crashreporter/tools/win32/dump_syms_vc1700.exe +++ /dev/null diff --git a/toolkit/crashreporter/tools/win32/dump_syms_vc1800.exe b/toolkit/crashreporter/tools/win32/dump_syms_vc1800.exe Binary files differdeleted file mode 100644 index 22e7398ab..000000000 --- a/toolkit/crashreporter/tools/win32/dump_syms_vc1800.exe +++ /dev/null |