2009-09-05 18:08:49 -07:00
|
|
|
#
|
2012-05-21 04:12:37 -07:00
|
|
|
# This Source Code Form is subject to the terms of the Mozilla Public
|
|
|
|
# License, v. 2.0. If a copy of the MPL was not distributed with this
|
|
|
|
# file, You can obtain one at http://mozilla.org/MPL/2.0/.
|
2009-09-05 18:08:49 -07:00
|
|
|
|
2011-11-14 20:33:21 -08:00
|
|
|
from __future__ import with_statement
|
|
|
|
import glob, logging, os, platform, shutil, subprocess, sys, tempfile, urllib2, zipfile
|
2009-09-05 18:08:49 -07:00
|
|
|
import re
|
2010-05-06 05:06:09 -07:00
|
|
|
from urlparse import urlparse
|
2009-07-22 12:24:54 -07:00
|
|
|
|
|
|
|
__all__ = [
|
2011-11-14 20:33:21 -08:00
|
|
|
"ZipFileReader",
|
2009-09-05 18:08:49 -07:00
|
|
|
"addCommonOptions",
|
|
|
|
"dumpLeakLog",
|
2010-05-06 05:06:09 -07:00
|
|
|
"isURL",
|
2009-09-05 18:08:49 -07:00
|
|
|
"processLeakLog",
|
2009-10-19 16:12:09 -07:00
|
|
|
"getDebuggerInfo",
|
|
|
|
"DEBUGGER_INFO",
|
2010-03-12 14:57:29 -08:00
|
|
|
"replaceBackSlashes",
|
2010-09-23 09:19:31 -07:00
|
|
|
"wrapCommand",
|
2009-09-05 18:08:49 -07:00
|
|
|
]
|
2009-07-22 12:24:54 -07:00
|
|
|
|
2009-10-19 16:12:09 -07:00
|
|
|
# Map of debugging programs to information about them, like default arguments
|
|
|
|
# and whether or not they are interactive.
|
|
|
|
DEBUGGER_INFO = {
|
|
|
|
# gdb requires that you supply the '--args' flag in order to pass arguments
|
|
|
|
# after the executable name to the executable.
|
|
|
|
"gdb": {
|
|
|
|
"interactive": True,
|
|
|
|
"args": "-q --args"
|
|
|
|
},
|
|
|
|
|
|
|
|
# valgrind doesn't explain much about leaks unless you set the
|
|
|
|
# '--leak-check=full' flag.
|
|
|
|
"valgrind": {
|
|
|
|
"interactive": False,
|
|
|
|
"args": "--leak-check=full"
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2011-11-14 20:33:21 -08:00
|
|
|
class ZipFileReader(object):
|
|
|
|
"""
|
|
|
|
Class to read zip files in Python 2.5 and later. Limited to only what we
|
|
|
|
actually use.
|
|
|
|
"""
|
|
|
|
|
|
|
|
def __init__(self, filename):
|
|
|
|
self._zipfile = zipfile.ZipFile(filename, "r")
|
|
|
|
|
|
|
|
def __del__(self):
|
|
|
|
self._zipfile.close()
|
|
|
|
|
|
|
|
def _getnormalizedpath(self, path):
|
|
|
|
"""
|
|
|
|
Gets a normalized path from 'path' (or the current working directory if
|
|
|
|
'path' is None). Also asserts that the path exists.
|
|
|
|
"""
|
|
|
|
if path is None:
|
|
|
|
path = os.curdir
|
|
|
|
path = os.path.normpath(os.path.expanduser(path))
|
|
|
|
assert os.path.isdir(path)
|
|
|
|
return path
|
|
|
|
|
|
|
|
def _extractname(self, name, path):
|
|
|
|
"""
|
|
|
|
Extracts a file with the given name from the zip file to the given path.
|
|
|
|
Also creates any directories needed along the way.
|
|
|
|
"""
|
|
|
|
filename = os.path.normpath(os.path.join(path, name))
|
|
|
|
if name.endswith("/"):
|
|
|
|
os.makedirs(filename)
|
|
|
|
else:
|
|
|
|
path = os.path.split(filename)[0]
|
|
|
|
if not os.path.isdir(path):
|
|
|
|
os.makedirs(path)
|
|
|
|
with open(filename, "wb") as dest:
|
|
|
|
dest.write(self._zipfile.read(name))
|
|
|
|
|
|
|
|
def namelist(self):
|
|
|
|
return self._zipfile.namelist()
|
|
|
|
|
|
|
|
def read(self, name):
|
|
|
|
return self._zipfile.read(name)
|
|
|
|
|
|
|
|
def extract(self, name, path = None):
|
|
|
|
if hasattr(self._zipfile, "extract"):
|
|
|
|
return self._zipfile.extract(name, path)
|
|
|
|
|
|
|
|
# This will throw if name is not part of the zip file.
|
|
|
|
self._zipfile.getinfo(name)
|
|
|
|
|
|
|
|
self._extractname(name, self._getnormalizedpath(path))
|
|
|
|
|
|
|
|
def extractall(self, path = None):
|
|
|
|
if hasattr(self._zipfile, "extractall"):
|
|
|
|
return self._zipfile.extractall(path)
|
|
|
|
|
|
|
|
path = self._getnormalizedpath(path)
|
|
|
|
|
|
|
|
for name in self._zipfile.namelist():
|
|
|
|
self._extractname(name, path)
|
|
|
|
|
2009-07-22 12:24:54 -07:00
|
|
|
log = logging.getLogger()
|
|
|
|
|
2010-05-06 05:06:09 -07:00
|
|
|
def isURL(thing):
|
|
|
|
"""Return True if |thing| looks like a URL."""
|
2011-11-29 14:13:08 -08:00
|
|
|
# We want to download URLs like http://... but not Windows paths like c:\...
|
|
|
|
return len(urlparse(thing).scheme) >= 2
|
2010-05-06 05:06:09 -07:00
|
|
|
|
2009-07-22 12:24:54 -07:00
|
|
|
def addCommonOptions(parser, defaults={}):
|
2009-09-05 18:08:49 -07:00
|
|
|
parser.add_option("--xre-path",
|
|
|
|
action = "store", type = "string", dest = "xrePath",
|
|
|
|
# individual scripts will set a sane default
|
|
|
|
default = None,
|
|
|
|
help = "absolute path to directory containing XRE (probably xulrunner)")
|
|
|
|
if 'SYMBOLS_PATH' not in defaults:
|
|
|
|
defaults['SYMBOLS_PATH'] = None
|
|
|
|
parser.add_option("--symbols-path",
|
|
|
|
action = "store", type = "string", dest = "symbolsPath",
|
|
|
|
default = defaults['SYMBOLS_PATH'],
|
2010-05-06 05:06:09 -07:00
|
|
|
help = "absolute path to directory containing breakpad symbols, or the URL of a zip file containing symbols")
|
2009-10-19 16:12:09 -07:00
|
|
|
parser.add_option("--debugger",
|
|
|
|
action = "store", dest = "debugger",
|
|
|
|
help = "use the given debugger to launch the application")
|
|
|
|
parser.add_option("--debugger-args",
|
|
|
|
action = "store", dest = "debuggerArgs",
|
|
|
|
help = "pass the given args to the debugger _before_ "
|
|
|
|
"the application on the command line")
|
|
|
|
parser.add_option("--debugger-interactive",
|
|
|
|
action = "store_true", dest = "debuggerInteractive",
|
|
|
|
help = "prevents the test harness from redirecting "
|
|
|
|
"stdout and stderr for interactive debuggers")
|
2009-07-22 12:24:54 -07:00
|
|
|
|
2009-10-19 16:12:09 -07:00
|
|
|
def getFullPath(directory, path):
|
|
|
|
"Get an absolute path relative to 'directory'."
|
|
|
|
return os.path.normpath(os.path.join(directory, os.path.expanduser(path)))
|
|
|
|
|
|
|
|
def searchPath(directory, path):
|
|
|
|
"Go one step beyond getFullPath and try the various folders in PATH"
|
|
|
|
# Try looking in the current working directory first.
|
|
|
|
newpath = getFullPath(directory, path)
|
2010-01-21 12:02:51 -08:00
|
|
|
if os.path.isfile(newpath):
|
2009-10-19 16:12:09 -07:00
|
|
|
return newpath
|
|
|
|
|
|
|
|
# At this point we have to fail if a directory was given (to prevent cases
|
|
|
|
# like './gdb' from matching '/usr/bin/./gdb').
|
|
|
|
if not os.path.dirname(path):
|
|
|
|
for dir in os.environ['PATH'].split(os.pathsep):
|
|
|
|
newpath = os.path.join(dir, path)
|
2010-01-21 12:02:51 -08:00
|
|
|
if os.path.isfile(newpath):
|
2009-10-19 16:12:09 -07:00
|
|
|
return newpath
|
|
|
|
return None
|
|
|
|
|
|
|
|
def getDebuggerInfo(directory, debugger, debuggerArgs, debuggerInteractive = False):
|
|
|
|
|
|
|
|
debuggerInfo = None
|
|
|
|
|
|
|
|
if debugger:
|
|
|
|
debuggerPath = searchPath(directory, debugger)
|
|
|
|
if not debuggerPath:
|
|
|
|
print "Error: Path %s doesn't exist." % debugger
|
|
|
|
sys.exit(1)
|
|
|
|
|
|
|
|
debuggerName = os.path.basename(debuggerPath).lower()
|
|
|
|
|
|
|
|
def getDebuggerInfo(type, default):
|
|
|
|
if debuggerName in DEBUGGER_INFO and type in DEBUGGER_INFO[debuggerName]:
|
|
|
|
return DEBUGGER_INFO[debuggerName][type]
|
|
|
|
return default
|
|
|
|
|
|
|
|
debuggerInfo = {
|
|
|
|
"path": debuggerPath,
|
|
|
|
"interactive" : getDebuggerInfo("interactive", False),
|
|
|
|
"args": getDebuggerInfo("args", "").split()
|
|
|
|
}
|
|
|
|
|
|
|
|
if debuggerArgs:
|
|
|
|
debuggerInfo["args"] = debuggerArgs.split()
|
|
|
|
if debuggerInteractive:
|
|
|
|
debuggerInfo["interactive"] = debuggerInteractive
|
2011-11-29 14:13:08 -08:00
|
|
|
|
2009-10-19 16:12:09 -07:00
|
|
|
return debuggerInfo
|
|
|
|
|
2009-09-05 18:08:49 -07:00
|
|
|
|
|
|
|
def dumpLeakLog(leakLogFile, filter = False):
|
|
|
|
"""Process the leak log, without parsing it.
|
|
|
|
|
|
|
|
Use this function if you want the raw log only.
|
|
|
|
Use it preferably with the |XPCOM_MEM_LEAK_LOG| environment variable.
|
|
|
|
"""
|
|
|
|
|
|
|
|
# Don't warn (nor "info") if the log file is not there.
|
|
|
|
if not os.path.exists(leakLogFile):
|
|
|
|
return
|
|
|
|
|
2013-03-18 10:18:06 -07:00
|
|
|
with open(leakLogFile, "r") as leaks:
|
|
|
|
leakReport = leaks.read()
|
2009-09-05 18:08:49 -07:00
|
|
|
|
|
|
|
# Only |XPCOM_MEM_LEAK_LOG| reports can be actually filtered out.
|
|
|
|
# Only check whether an actual leak was reported.
|
|
|
|
if filter and not "0 TOTAL " in leakReport:
|
|
|
|
return
|
|
|
|
|
|
|
|
# Simply copy the log.
|
|
|
|
log.info(leakReport.rstrip("\n"))
|
|
|
|
|
2013-03-18 10:18:08 -07:00
|
|
|
def processSingleLeakFile(leakLogFileName, processType, leakThreshold):
|
|
|
|
"""Process a single leak log.
|
2009-09-05 18:08:49 -07:00
|
|
|
"""
|
|
|
|
|
|
|
|
# Per-Inst Leaked Total Rem ...
|
|
|
|
# 0 TOTAL 17 192 419115886 2 ...
|
|
|
|
# 833 nsTimerImpl 60 120 24726 2 ...
|
|
|
|
lineRe = re.compile(r"^\s*\d+\s+(?P<name>\S+)\s+"
|
|
|
|
r"(?P<size>-?\d+)\s+(?P<bytesLeaked>-?\d+)\s+"
|
|
|
|
r"-?\d+\s+(?P<numLeaked>-?\d+)")
|
|
|
|
|
2009-11-10 11:16:12 -08:00
|
|
|
processString = ""
|
2013-03-18 10:18:08 -07:00
|
|
|
if processType:
|
|
|
|
# eg 'plugin'
|
|
|
|
processString = " %s process:" % processType
|
2013-03-18 10:18:06 -07:00
|
|
|
|
2013-03-18 10:18:07 -07:00
|
|
|
crashedOnPurpose = False
|
|
|
|
totalBytesLeaked = None
|
|
|
|
leakAnalysis = []
|
|
|
|
leakedObjectNames = []
|
2013-03-18 10:18:06 -07:00
|
|
|
with open(leakLogFileName, "r") as leaks:
|
|
|
|
for line in leaks:
|
|
|
|
if line.find("purposefully crash") > -1:
|
|
|
|
crashedOnPurpose = True
|
|
|
|
matches = lineRe.match(line)
|
|
|
|
if not matches:
|
2013-03-18 10:18:07 -07:00
|
|
|
# eg: the leak table header row
|
|
|
|
log.info(line.rstrip())
|
2013-03-18 10:18:06 -07:00
|
|
|
continue
|
|
|
|
name = matches.group("name")
|
|
|
|
size = int(matches.group("size"))
|
|
|
|
bytesLeaked = int(matches.group("bytesLeaked"))
|
|
|
|
numLeaked = int(matches.group("numLeaked"))
|
2013-03-18 10:18:07 -07:00
|
|
|
# Output the raw line from the leak log table if it is the TOTAL row,
|
|
|
|
# or is for an object row that has been leaked.
|
|
|
|
if numLeaked != 0 or name == "TOTAL":
|
|
|
|
log.info(line.rstrip())
|
|
|
|
# Analyse the leak log, but output later or it will interrupt the leak table
|
2013-03-18 10:18:06 -07:00
|
|
|
if name == "TOTAL":
|
|
|
|
totalBytesLeaked = bytesLeaked
|
|
|
|
if size < 0 or bytesLeaked < 0 or numLeaked < 0:
|
2013-03-18 10:18:08 -07:00
|
|
|
leakAnalysis.append("TEST-UNEXPECTED-FAIL | leakcheck |%s negative leaks caught!"
|
2013-03-18 10:18:07 -07:00
|
|
|
% processString)
|
2013-03-18 10:18:06 -07:00
|
|
|
continue
|
|
|
|
if name != "TOTAL" and numLeaked != 0:
|
|
|
|
leakedObjectNames.append(name)
|
2013-03-18 10:18:08 -07:00
|
|
|
leakAnalysis.append("TEST-INFO | leakcheck |%s leaked %d %s (%s bytes)"
|
2013-03-18 10:18:07 -07:00
|
|
|
% (processString, numLeaked, name, bytesLeaked))
|
|
|
|
log.info('\n'.join(leakAnalysis))
|
2013-03-18 10:18:05 -07:00
|
|
|
|
2013-03-18 10:18:04 -07:00
|
|
|
if totalBytesLeaked is None:
|
|
|
|
# We didn't see a line with name 'TOTAL'
|
2009-12-09 10:51:38 -08:00
|
|
|
if crashedOnPurpose:
|
2013-03-18 10:18:08 -07:00
|
|
|
log.info("TEST-INFO | leakcheck |%s deliberate crash and thus no leak log"
|
|
|
|
% processString)
|
2009-12-09 10:51:38 -08:00
|
|
|
else:
|
2013-03-18 10:18:05 -07:00
|
|
|
# TODO: This should be a TEST-UNEXPECTED-FAIL, but was changed to a warning
|
|
|
|
# due to too many intermittent failures (see bug 831223).
|
2013-03-18 10:18:08 -07:00
|
|
|
log.info("WARNING | leakcheck |%s missing output line for total leaks!"
|
|
|
|
% processString)
|
2013-03-18 10:18:07 -07:00
|
|
|
return
|
|
|
|
|
|
|
|
if totalBytesLeaked == 0:
|
2013-03-18 10:18:08 -07:00
|
|
|
log.info("TEST-PASS | leakcheck |%s no leaks detected!" % processString)
|
2013-03-18 10:18:07 -07:00
|
|
|
return
|
|
|
|
|
2013-03-18 10:18:08 -07:00
|
|
|
# totalBytesLeaked was seen and is non-zero.
|
2013-03-18 10:18:07 -07:00
|
|
|
if totalBytesLeaked > leakThreshold:
|
2013-03-18 10:18:08 -07:00
|
|
|
# Fail the run if we're over the threshold (which defaults to 0)
|
2013-03-18 10:18:07 -07:00
|
|
|
prefix = "TEST-UNEXPECTED-FAIL"
|
2013-03-18 10:18:04 -07:00
|
|
|
else:
|
2013-03-18 10:18:07 -07:00
|
|
|
prefix = "WARNING"
|
|
|
|
# Create a comma delimited string of the first N leaked objects found,
|
2013-03-18 10:18:08 -07:00
|
|
|
# to aid with bug summary matching in TBPL. Note: The order of the objects
|
|
|
|
# had no significance (they're sorted alphabetically).
|
2013-03-18 10:18:07 -07:00
|
|
|
maxSummaryObjects = 5
|
|
|
|
leakedObjectSummary = ', '.join(leakedObjectNames[:maxSummaryObjects])
|
|
|
|
if len(leakedObjectNames) > maxSummaryObjects:
|
|
|
|
leakedObjectSummary += ', ...'
|
2013-03-18 10:18:08 -07:00
|
|
|
log.info("%s | leakcheck |%s %d bytes leaked (%s)"
|
2013-03-18 10:18:07 -07:00
|
|
|
% (prefix, processString, totalBytesLeaked, leakedObjectSummary))
|
2009-11-10 11:16:12 -08:00
|
|
|
|
|
|
|
def processLeakLog(leakLogFile, leakThreshold = 0):
|
|
|
|
"""Process the leak log, including separate leak logs created
|
|
|
|
by child processes.
|
|
|
|
|
|
|
|
Use this function if you want an additional PASS/FAIL summary.
|
|
|
|
It must be used with the |XPCOM_MEM_BLOAT_LOG| environment variable.
|
|
|
|
"""
|
|
|
|
|
|
|
|
if not os.path.exists(leakLogFile):
|
2013-03-18 10:18:03 -07:00
|
|
|
log.info("WARNING | leakcheck | refcount logging is off, so leaks can't be detected!")
|
2009-11-10 11:16:12 -08:00
|
|
|
return
|
|
|
|
|
2013-03-18 10:18:06 -07:00
|
|
|
if leakThreshold != 0:
|
|
|
|
log.info("TEST-INFO | leakcheck | threshold set at %d bytes" % leakThreshold)
|
|
|
|
|
2009-11-10 11:16:12 -08:00
|
|
|
(leakLogFileDir, leakFileBase) = os.path.split(leakLogFile)
|
2013-03-18 10:18:08 -07:00
|
|
|
fileNameRegExp = re.compile(r".*?_([a-z]*)_pid\d*$")
|
2009-11-10 11:16:12 -08:00
|
|
|
if leakFileBase[-4:] == ".log":
|
|
|
|
leakFileBase = leakFileBase[:-4]
|
2013-03-18 10:18:08 -07:00
|
|
|
fileNameRegExp = re.compile(r".*?_([a-z]*)_pid\d*.log$")
|
2009-11-10 11:16:12 -08:00
|
|
|
|
|
|
|
for fileName in os.listdir(leakLogFileDir):
|
|
|
|
if fileName.find(leakFileBase) != -1:
|
|
|
|
thisFile = os.path.join(leakLogFileDir, fileName)
|
|
|
|
processType = None
|
2013-03-18 10:18:08 -07:00
|
|
|
m = fileNameRegExp.search(fileName)
|
2009-11-10 11:16:12 -08:00
|
|
|
if m:
|
|
|
|
processType = m.group(1)
|
2013-03-18 10:18:08 -07:00
|
|
|
processSingleLeakFile(thisFile, processType, leakThreshold)
|
2010-03-12 14:57:29 -08:00
|
|
|
|
|
|
|
def replaceBackSlashes(input):
|
|
|
|
return input.replace('\\', '/')
|
2010-09-23 09:19:31 -07:00
|
|
|
|
|
|
|
def wrapCommand(cmd):
|
|
|
|
"""
|
|
|
|
If running on OS X 10.5 or older, wrap |cmd| so that it will
|
|
|
|
be executed as an i386 binary, in case it's a 32-bit/64-bit universal
|
|
|
|
binary.
|
|
|
|
"""
|
|
|
|
if platform.system() == "Darwin" and \
|
|
|
|
hasattr(platform, 'mac_ver') and \
|
|
|
|
platform.mac_ver()[0][:4] < '10.6':
|
|
|
|
return ["arch", "-arch", "i386"] + cmd
|
|
|
|
# otherwise just execute the command normally
|
|
|
|
return cmd
|