2009-09-05 18:08:49 -07:00
|
|
|
#
|
2012-05-21 04:12:37 -07:00
|
|
|
# This Source Code Form is subject to the terms of the Mozilla Public
|
|
|
|
# License, v. 2.0. If a copy of the MPL was not distributed with this
|
|
|
|
# file, You can obtain one at http://mozilla.org/MPL/2.0/.
|
2009-09-05 18:08:49 -07:00
|
|
|
|
2011-11-14 20:33:21 -08:00
|
|
|
from __future__ import with_statement
|
|
|
|
import glob, logging, os, platform, shutil, subprocess, sys, tempfile, urllib2, zipfile
|
2009-09-05 18:08:49 -07:00
|
|
|
import re
|
2010-05-06 05:06:09 -07:00
|
|
|
from urlparse import urlparse
|
2009-07-22 12:24:54 -07:00
|
|
|
|
|
|
|
__all__ = [
|
2011-11-14 20:33:21 -08:00
|
|
|
"ZipFileReader",
|
2009-09-05 18:08:49 -07:00
|
|
|
"addCommonOptions",
|
|
|
|
"checkForCrashes",
|
|
|
|
"dumpLeakLog",
|
2010-05-06 05:06:09 -07:00
|
|
|
"isURL",
|
2009-09-05 18:08:49 -07:00
|
|
|
"processLeakLog",
|
2009-10-19 16:12:09 -07:00
|
|
|
"getDebuggerInfo",
|
|
|
|
"DEBUGGER_INFO",
|
2010-03-12 14:57:29 -08:00
|
|
|
"replaceBackSlashes",
|
2010-09-23 09:19:31 -07:00
|
|
|
"wrapCommand",
|
2009-09-05 18:08:49 -07:00
|
|
|
]
|
2009-07-22 12:24:54 -07:00
|
|
|
|
2009-10-19 16:12:09 -07:00
|
|
|
# Map of debugging programs to information about them, like default arguments
|
|
|
|
# and whether or not they are interactive.
|
|
|
|
DEBUGGER_INFO = {
|
|
|
|
# gdb requires that you supply the '--args' flag in order to pass arguments
|
|
|
|
# after the executable name to the executable.
|
|
|
|
"gdb": {
|
|
|
|
"interactive": True,
|
|
|
|
"args": "-q --args"
|
|
|
|
},
|
|
|
|
|
|
|
|
# valgrind doesn't explain much about leaks unless you set the
|
|
|
|
# '--leak-check=full' flag.
|
|
|
|
"valgrind": {
|
|
|
|
"interactive": False,
|
|
|
|
"args": "--leak-check=full"
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2011-11-14 20:33:21 -08:00
|
|
|
class ZipFileReader(object):
|
|
|
|
"""
|
|
|
|
Class to read zip files in Python 2.5 and later. Limited to only what we
|
|
|
|
actually use.
|
|
|
|
"""
|
|
|
|
|
|
|
|
def __init__(self, filename):
|
|
|
|
self._zipfile = zipfile.ZipFile(filename, "r")
|
|
|
|
|
|
|
|
def __del__(self):
|
|
|
|
self._zipfile.close()
|
|
|
|
|
|
|
|
def _getnormalizedpath(self, path):
|
|
|
|
"""
|
|
|
|
Gets a normalized path from 'path' (or the current working directory if
|
|
|
|
'path' is None). Also asserts that the path exists.
|
|
|
|
"""
|
|
|
|
if path is None:
|
|
|
|
path = os.curdir
|
|
|
|
path = os.path.normpath(os.path.expanduser(path))
|
|
|
|
assert os.path.isdir(path)
|
|
|
|
return path
|
|
|
|
|
|
|
|
def _extractname(self, name, path):
|
|
|
|
"""
|
|
|
|
Extracts a file with the given name from the zip file to the given path.
|
|
|
|
Also creates any directories needed along the way.
|
|
|
|
"""
|
|
|
|
filename = os.path.normpath(os.path.join(path, name))
|
|
|
|
if name.endswith("/"):
|
|
|
|
os.makedirs(filename)
|
|
|
|
else:
|
|
|
|
path = os.path.split(filename)[0]
|
|
|
|
if not os.path.isdir(path):
|
|
|
|
os.makedirs(path)
|
|
|
|
with open(filename, "wb") as dest:
|
|
|
|
dest.write(self._zipfile.read(name))
|
|
|
|
|
|
|
|
def namelist(self):
|
|
|
|
return self._zipfile.namelist()
|
|
|
|
|
|
|
|
def read(self, name):
|
|
|
|
return self._zipfile.read(name)
|
|
|
|
|
|
|
|
def extract(self, name, path = None):
|
|
|
|
if hasattr(self._zipfile, "extract"):
|
|
|
|
return self._zipfile.extract(name, path)
|
|
|
|
|
|
|
|
# This will throw if name is not part of the zip file.
|
|
|
|
self._zipfile.getinfo(name)
|
|
|
|
|
|
|
|
self._extractname(name, self._getnormalizedpath(path))
|
|
|
|
|
|
|
|
def extractall(self, path = None):
|
|
|
|
if hasattr(self._zipfile, "extractall"):
|
|
|
|
return self._zipfile.extractall(path)
|
|
|
|
|
|
|
|
path = self._getnormalizedpath(path)
|
|
|
|
|
|
|
|
for name in self._zipfile.namelist():
|
|
|
|
self._extractname(name, path)
|
|
|
|
|
2009-07-22 12:24:54 -07:00
|
|
|
log = logging.getLogger()
|
|
|
|
|
2010-05-06 05:06:09 -07:00
|
|
|
def isURL(thing):
|
|
|
|
"""Return True if |thing| looks like a URL."""
|
2011-11-29 14:13:08 -08:00
|
|
|
# We want to download URLs like http://... but not Windows paths like c:\...
|
|
|
|
return len(urlparse(thing).scheme) >= 2
|
2010-05-06 05:06:09 -07:00
|
|
|
|
2009-07-22 12:24:54 -07:00
|
|
|
def addCommonOptions(parser, defaults={}):
|
2009-09-05 18:08:49 -07:00
|
|
|
parser.add_option("--xre-path",
|
|
|
|
action = "store", type = "string", dest = "xrePath",
|
|
|
|
# individual scripts will set a sane default
|
|
|
|
default = None,
|
|
|
|
help = "absolute path to directory containing XRE (probably xulrunner)")
|
|
|
|
if 'SYMBOLS_PATH' not in defaults:
|
|
|
|
defaults['SYMBOLS_PATH'] = None
|
|
|
|
parser.add_option("--symbols-path",
|
|
|
|
action = "store", type = "string", dest = "symbolsPath",
|
|
|
|
default = defaults['SYMBOLS_PATH'],
|
2010-05-06 05:06:09 -07:00
|
|
|
help = "absolute path to directory containing breakpad symbols, or the URL of a zip file containing symbols")
|
2009-10-19 16:12:09 -07:00
|
|
|
parser.add_option("--debugger",
|
|
|
|
action = "store", dest = "debugger",
|
|
|
|
help = "use the given debugger to launch the application")
|
|
|
|
parser.add_option("--debugger-args",
|
|
|
|
action = "store", dest = "debuggerArgs",
|
|
|
|
help = "pass the given args to the debugger _before_ "
|
|
|
|
"the application on the command line")
|
|
|
|
parser.add_option("--debugger-interactive",
|
|
|
|
action = "store_true", dest = "debuggerInteractive",
|
|
|
|
help = "prevents the test harness from redirecting "
|
|
|
|
"stdout and stderr for interactive debuggers")
|
2009-07-22 12:24:54 -07:00
|
|
|
|
|
|
|
def checkForCrashes(dumpDir, symbolsPath, testName=None):
|
|
|
|
stackwalkPath = os.environ.get('MINIDUMP_STACKWALK', None)
|
|
|
|
# try to get the caller's filename if no test name is given
|
|
|
|
if testName is None:
|
2009-09-05 18:08:49 -07:00
|
|
|
try:
|
|
|
|
testName = os.path.basename(sys._getframe(1).f_code.co_filename)
|
|
|
|
except:
|
|
|
|
testName = "unknown"
|
2009-07-22 12:24:54 -07:00
|
|
|
|
2011-11-14 20:33:21 -08:00
|
|
|
# Check preconditions
|
2009-07-22 12:24:54 -07:00
|
|
|
dumps = glob.glob(os.path.join(dumpDir, '*.dmp'))
|
2011-11-14 20:33:21 -08:00
|
|
|
if len(dumps) == 0:
|
|
|
|
return False
|
|
|
|
|
|
|
|
removeSymbolsPath = False
|
|
|
|
|
|
|
|
# If our symbols are at a remote URL, download them now
|
2012-09-15 11:33:02 -07:00
|
|
|
if symbolsPath and isURL(symbolsPath):
|
2011-11-14 20:33:21 -08:00
|
|
|
print "Downloading symbols from: " + symbolsPath
|
|
|
|
removeSymbolsPath = True
|
|
|
|
# Get the symbols and write them to a temporary zipfile
|
|
|
|
data = urllib2.urlopen(symbolsPath)
|
|
|
|
symbolsFile = tempfile.TemporaryFile()
|
|
|
|
symbolsFile.write(data.read())
|
|
|
|
# extract symbols to a temporary directory (which we'll delete after
|
|
|
|
# processing all crashes)
|
|
|
|
symbolsPath = tempfile.mkdtemp()
|
|
|
|
zfile = ZipFileReader(symbolsFile)
|
|
|
|
zfile.extractall(symbolsPath)
|
|
|
|
|
|
|
|
try:
|
|
|
|
for d in dumps:
|
|
|
|
log.info("PROCESS-CRASH | %s | application crashed (minidump found)", testName)
|
|
|
|
print "Crash dump filename: " + d
|
|
|
|
if symbolsPath and stackwalkPath and os.path.exists(stackwalkPath):
|
|
|
|
# run minidump stackwalk
|
|
|
|
p = subprocess.Popen([stackwalkPath, d, symbolsPath],
|
|
|
|
stdout=subprocess.PIPE,
|
|
|
|
stderr=subprocess.PIPE)
|
|
|
|
(out, err) = p.communicate()
|
|
|
|
if len(out) > 3:
|
|
|
|
# minidump_stackwalk is chatty, so ignore stderr when it succeeds.
|
|
|
|
print out
|
2010-11-06 13:30:32 -07:00
|
|
|
else:
|
2011-11-14 20:33:21 -08:00
|
|
|
print "stderr from minidump_stackwalk:"
|
|
|
|
print err
|
|
|
|
if p.returncode != 0:
|
|
|
|
print "minidump_stackwalk exited with return code %d" % p.returncode
|
2009-10-20 10:27:02 -07:00
|
|
|
else:
|
2011-11-14 20:33:21 -08:00
|
|
|
if not symbolsPath:
|
|
|
|
print "No symbols path given, can't process dump."
|
|
|
|
if not stackwalkPath:
|
|
|
|
print "MINIDUMP_STACKWALK not set, can't process dump."
|
|
|
|
elif stackwalkPath and not os.path.exists(stackwalkPath):
|
2009-10-20 10:27:02 -07:00
|
|
|
print "MINIDUMP_STACKWALK binary not found: %s" % stackwalkPath
|
2011-11-14 20:33:21 -08:00
|
|
|
dumpSavePath = os.environ.get('MINIDUMP_SAVE_PATH', None)
|
|
|
|
if dumpSavePath:
|
|
|
|
shutil.move(d, dumpSavePath)
|
|
|
|
print "Saved dump as %s" % os.path.join(dumpSavePath,
|
|
|
|
os.path.basename(d))
|
|
|
|
else:
|
|
|
|
os.remove(d)
|
|
|
|
extra = os.path.splitext(d)[0] + ".extra"
|
|
|
|
if os.path.exists(extra):
|
|
|
|
os.remove(extra)
|
|
|
|
finally:
|
|
|
|
if removeSymbolsPath:
|
|
|
|
shutil.rmtree(symbolsPath)
|
2009-09-05 18:08:49 -07:00
|
|
|
|
2012-11-20 07:24:28 -08:00
|
|
|
return True
|
2011-11-14 20:33:21 -08:00
|
|
|
|
2009-10-19 16:12:09 -07:00
|
|
|
def getFullPath(directory, path):
|
|
|
|
"Get an absolute path relative to 'directory'."
|
|
|
|
return os.path.normpath(os.path.join(directory, os.path.expanduser(path)))
|
|
|
|
|
|
|
|
def searchPath(directory, path):
|
|
|
|
"Go one step beyond getFullPath and try the various folders in PATH"
|
|
|
|
# Try looking in the current working directory first.
|
|
|
|
newpath = getFullPath(directory, path)
|
2010-01-21 12:02:51 -08:00
|
|
|
if os.path.isfile(newpath):
|
2009-10-19 16:12:09 -07:00
|
|
|
return newpath
|
|
|
|
|
|
|
|
# At this point we have to fail if a directory was given (to prevent cases
|
|
|
|
# like './gdb' from matching '/usr/bin/./gdb').
|
|
|
|
if not os.path.dirname(path):
|
|
|
|
for dir in os.environ['PATH'].split(os.pathsep):
|
|
|
|
newpath = os.path.join(dir, path)
|
2010-01-21 12:02:51 -08:00
|
|
|
if os.path.isfile(newpath):
|
2009-10-19 16:12:09 -07:00
|
|
|
return newpath
|
|
|
|
return None
|
|
|
|
|
|
|
|
def getDebuggerInfo(directory, debugger, debuggerArgs, debuggerInteractive = False):
|
|
|
|
|
|
|
|
debuggerInfo = None
|
|
|
|
|
|
|
|
if debugger:
|
|
|
|
debuggerPath = searchPath(directory, debugger)
|
|
|
|
if not debuggerPath:
|
|
|
|
print "Error: Path %s doesn't exist." % debugger
|
|
|
|
sys.exit(1)
|
|
|
|
|
|
|
|
debuggerName = os.path.basename(debuggerPath).lower()
|
|
|
|
|
|
|
|
def getDebuggerInfo(type, default):
|
|
|
|
if debuggerName in DEBUGGER_INFO and type in DEBUGGER_INFO[debuggerName]:
|
|
|
|
return DEBUGGER_INFO[debuggerName][type]
|
|
|
|
return default
|
|
|
|
|
|
|
|
debuggerInfo = {
|
|
|
|
"path": debuggerPath,
|
|
|
|
"interactive" : getDebuggerInfo("interactive", False),
|
|
|
|
"args": getDebuggerInfo("args", "").split()
|
|
|
|
}
|
|
|
|
|
|
|
|
if debuggerArgs:
|
|
|
|
debuggerInfo["args"] = debuggerArgs.split()
|
|
|
|
if debuggerInteractive:
|
|
|
|
debuggerInfo["interactive"] = debuggerInteractive
|
2011-11-29 14:13:08 -08:00
|
|
|
|
2009-10-19 16:12:09 -07:00
|
|
|
return debuggerInfo
|
|
|
|
|
2009-09-05 18:08:49 -07:00
|
|
|
|
|
|
|
def dumpLeakLog(leakLogFile, filter = False):
|
|
|
|
"""Process the leak log, without parsing it.
|
|
|
|
|
|
|
|
Use this function if you want the raw log only.
|
|
|
|
Use it preferably with the |XPCOM_MEM_LEAK_LOG| environment variable.
|
|
|
|
"""
|
|
|
|
|
|
|
|
# Don't warn (nor "info") if the log file is not there.
|
|
|
|
if not os.path.exists(leakLogFile):
|
|
|
|
return
|
|
|
|
|
|
|
|
leaks = open(leakLogFile, "r")
|
|
|
|
leakReport = leaks.read()
|
|
|
|
leaks.close()
|
|
|
|
|
|
|
|
# Only |XPCOM_MEM_LEAK_LOG| reports can be actually filtered out.
|
|
|
|
# Only check whether an actual leak was reported.
|
|
|
|
if filter and not "0 TOTAL " in leakReport:
|
|
|
|
return
|
|
|
|
|
|
|
|
# Simply copy the log.
|
|
|
|
log.info(leakReport.rstrip("\n"))
|
|
|
|
|
2009-11-10 11:16:12 -08:00
|
|
|
def processSingleLeakFile(leakLogFileName, PID, processType, leakThreshold):
|
|
|
|
"""Process a single leak log, corresponding to the specified
|
|
|
|
process PID and type.
|
2009-09-05 18:08:49 -07:00
|
|
|
"""
|
|
|
|
|
|
|
|
# Per-Inst Leaked Total Rem ...
|
|
|
|
# 0 TOTAL 17 192 419115886 2 ...
|
|
|
|
# 833 nsTimerImpl 60 120 24726 2 ...
|
|
|
|
lineRe = re.compile(r"^\s*\d+\s+(?P<name>\S+)\s+"
|
|
|
|
r"(?P<size>-?\d+)\s+(?P<bytesLeaked>-?\d+)\s+"
|
|
|
|
r"-?\d+\s+(?P<numLeaked>-?\d+)")
|
|
|
|
|
2009-11-10 11:16:12 -08:00
|
|
|
processString = ""
|
|
|
|
if PID and processType:
|
|
|
|
processString = "| %s process %s " % (processType, PID)
|
|
|
|
leaks = open(leakLogFileName, "r")
|
2009-09-05 18:08:49 -07:00
|
|
|
for line in leaks:
|
|
|
|
matches = lineRe.match(line)
|
|
|
|
if (matches and
|
|
|
|
int(matches.group("numLeaked")) == 0 and
|
|
|
|
matches.group("name") != "TOTAL"):
|
|
|
|
continue
|
|
|
|
log.info(line.rstrip())
|
|
|
|
leaks.close()
|
|
|
|
|
2009-11-10 11:16:12 -08:00
|
|
|
leaks = open(leakLogFileName, "r")
|
2009-09-05 18:08:49 -07:00
|
|
|
seenTotal = False
|
2009-12-09 10:51:38 -08:00
|
|
|
crashedOnPurpose = False
|
2009-09-05 18:08:49 -07:00
|
|
|
prefix = "TEST-PASS"
|
2010-06-10 05:27:47 -07:00
|
|
|
numObjects = 0
|
2009-09-05 18:08:49 -07:00
|
|
|
for line in leaks:
|
2009-12-09 10:51:38 -08:00
|
|
|
if line.find("purposefully crash") > -1:
|
|
|
|
crashedOnPurpose = True
|
2009-09-05 18:08:49 -07:00
|
|
|
matches = lineRe.match(line)
|
|
|
|
if not matches:
|
|
|
|
continue
|
|
|
|
name = matches.group("name")
|
|
|
|
size = int(matches.group("size"))
|
|
|
|
bytesLeaked = int(matches.group("bytesLeaked"))
|
|
|
|
numLeaked = int(matches.group("numLeaked"))
|
|
|
|
if size < 0 or bytesLeaked < 0 or numLeaked < 0:
|
2009-11-10 11:16:12 -08:00
|
|
|
log.info("TEST-UNEXPECTED-FAIL %s| automationutils.processLeakLog() | negative leaks caught!" %
|
|
|
|
processString)
|
2009-09-05 18:08:49 -07:00
|
|
|
if name == "TOTAL":
|
|
|
|
seenTotal = True
|
|
|
|
elif name == "TOTAL":
|
|
|
|
seenTotal = True
|
|
|
|
# Check for leaks.
|
|
|
|
if bytesLeaked < 0 or bytesLeaked > leakThreshold:
|
|
|
|
prefix = "TEST-UNEXPECTED-FAIL"
|
2009-11-10 11:16:12 -08:00
|
|
|
leakLog = "TEST-UNEXPECTED-FAIL %s| automationutils.processLeakLog() | leaked" \
|
|
|
|
" %d bytes during test execution" % (processString, bytesLeaked)
|
2009-09-05 18:08:49 -07:00
|
|
|
elif bytesLeaked > 0:
|
2009-11-10 11:16:12 -08:00
|
|
|
leakLog = "TEST-PASS %s| automationutils.processLeakLog() | WARNING leaked" \
|
|
|
|
" %d bytes during test execution" % (processString, bytesLeaked)
|
2009-09-05 18:08:49 -07:00
|
|
|
else:
|
2009-11-10 11:16:12 -08:00
|
|
|
leakLog = "TEST-PASS %s| automationutils.processLeakLog() | no leaks detected!" \
|
|
|
|
% processString
|
2009-09-05 18:08:49 -07:00
|
|
|
# Remind the threshold if it is not 0, which is the default/goal.
|
|
|
|
if leakThreshold != 0:
|
|
|
|
leakLog += " (threshold set at %d bytes)" % leakThreshold
|
|
|
|
# Log the information.
|
|
|
|
log.info(leakLog)
|
|
|
|
else:
|
|
|
|
if numLeaked != 0:
|
|
|
|
if numLeaked > 1:
|
|
|
|
instance = "instances"
|
|
|
|
rest = " each (%s bytes total)" % matches.group("bytesLeaked")
|
|
|
|
else:
|
|
|
|
instance = "instance"
|
|
|
|
rest = ""
|
2010-06-10 05:27:47 -07:00
|
|
|
numObjects += 1
|
|
|
|
if numObjects > 5:
|
|
|
|
# don't spam brief tinderbox logs with tons of leak output
|
|
|
|
prefix = "TEST-INFO"
|
2009-11-10 11:16:12 -08:00
|
|
|
log.info("%(prefix)s %(process)s| automationutils.processLeakLog() | leaked %(numLeaked)d %(instance)s of %(name)s "
|
2009-09-05 18:08:49 -07:00
|
|
|
"with size %(size)s bytes%(rest)s" %
|
|
|
|
{ "prefix": prefix,
|
2009-11-10 11:16:12 -08:00
|
|
|
"process": processString,
|
2009-09-05 18:08:49 -07:00
|
|
|
"numLeaked": numLeaked,
|
|
|
|
"instance": instance,
|
|
|
|
"name": name,
|
|
|
|
"size": matches.group("size"),
|
|
|
|
"rest": rest })
|
|
|
|
if not seenTotal:
|
2009-12-09 10:51:38 -08:00
|
|
|
if crashedOnPurpose:
|
|
|
|
log.info("INFO | automationutils.processLeakLog() | process %s was " \
|
|
|
|
"deliberately crashed and thus has no leak log" % PID)
|
|
|
|
else:
|
|
|
|
log.info("TEST-UNEXPECTED-FAIL %s| automationutils.processLeakLog() | missing output line for total leaks!" %
|
2009-11-10 11:16:12 -08:00
|
|
|
processString)
|
2009-09-05 18:08:49 -07:00
|
|
|
leaks.close()
|
2009-11-10 11:16:12 -08:00
|
|
|
|
|
|
|
|
|
|
|
def processLeakLog(leakLogFile, leakThreshold = 0):
|
|
|
|
"""Process the leak log, including separate leak logs created
|
|
|
|
by child processes.
|
|
|
|
|
|
|
|
Use this function if you want an additional PASS/FAIL summary.
|
|
|
|
It must be used with the |XPCOM_MEM_BLOAT_LOG| environment variable.
|
|
|
|
"""
|
|
|
|
|
|
|
|
if not os.path.exists(leakLogFile):
|
|
|
|
log.info("WARNING | automationutils.processLeakLog() | refcount logging is off, so leaks can't be detected!")
|
|
|
|
return
|
|
|
|
|
|
|
|
(leakLogFileDir, leakFileBase) = os.path.split(leakLogFile)
|
|
|
|
pidRegExp = re.compile(r".*?_([a-z]*)_pid(\d*)$")
|
|
|
|
if leakFileBase[-4:] == ".log":
|
|
|
|
leakFileBase = leakFileBase[:-4]
|
|
|
|
pidRegExp = re.compile(r".*?_([a-z]*)_pid(\d*).log$")
|
|
|
|
|
|
|
|
for fileName in os.listdir(leakLogFileDir):
|
|
|
|
if fileName.find(leakFileBase) != -1:
|
|
|
|
thisFile = os.path.join(leakLogFileDir, fileName)
|
|
|
|
processPID = 0
|
|
|
|
processType = None
|
|
|
|
m = pidRegExp.search(fileName)
|
|
|
|
if m:
|
|
|
|
processType = m.group(1)
|
|
|
|
processPID = m.group(2)
|
|
|
|
processSingleLeakFile(thisFile, processPID, processType, leakThreshold)
|
2010-03-12 14:57:29 -08:00
|
|
|
|
|
|
|
def replaceBackSlashes(input):
|
|
|
|
return input.replace('\\', '/')
|
2010-09-23 09:19:31 -07:00
|
|
|
|
|
|
|
def wrapCommand(cmd):
|
|
|
|
"""
|
|
|
|
If running on OS X 10.5 or older, wrap |cmd| so that it will
|
|
|
|
be executed as an i386 binary, in case it's a 32-bit/64-bit universal
|
|
|
|
binary.
|
|
|
|
"""
|
|
|
|
if platform.system() == "Darwin" and \
|
|
|
|
hasattr(platform, 'mac_ver') and \
|
|
|
|
platform.mac_ver()[0][:4] < '10.6':
|
|
|
|
return ["arch", "-arch", "i386"] + cmd
|
|
|
|
# otherwise just execute the command normally
|
|
|
|
return cmd
|