Permalink
Cannot retrieve contributors at this time
Name already in use
A tag already exists with the provided branch name. Many Git commands accept both tag and branch names, so creating this branch may cause unexpected behavior. Are you sure you want to create this branch?
Phoenix/build.py
Go to fileThis commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
executable file
2372 lines (1914 sloc)
84.3 KB
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
#!/usr/bin/python | |
#---------------------------------------------------------------------- | |
# Name: build.py | |
# Purpose: Master build controller script. | |
# This script is used to run through the commands used for the | |
# various stages of building Phoenix, and can also be a front-end | |
# for building wxWidgets and the wxPython distribution files. | |
# | |
# Author: Robin Dunn | |
# | |
# Created: 3-Dec-2010 | |
# Copyright: (c) 2010-2020 by Total Control Software | |
# License: wxWindows License | |
#---------------------------------------------------------------------- | |
from __future__ import absolute_import | |
import sys | |
import glob | |
import hashlib | |
import optparse | |
import os | |
import re | |
import shutil | |
import subprocess | |
import tarfile | |
import tempfile | |
import datetime | |
import shlex | |
import textwrap | |
try: | |
import pathlib | |
except ImportError: | |
import buildtools.backports.pathlib2 as pathlib | |
try: | |
from shutil import which | |
except ImportError: | |
from buildtools.backports.shutil_which import which | |
from distutils.dep_util import newer, newer_group | |
from buildtools.config import Config, msg, opj, posixjoin, loadETG, etg2sip, findCmd, \ | |
phoenixDir, wxDir, copyIfNewer, copyFile, \ | |
macSetLoaderNames, \ | |
getVcsRev, runcmd, textfile_open, getSipFiles, \ | |
getVisCVersion, getToolsPlatformName, updateLicenseFiles, \ | |
TemporaryDirectory, getMSVCInfo | |
import buildtools.version as version | |
# which version of Python is running this script | |
PY2 = sys.version_info[0] == 2 | |
PY3 = sys.version_info[0] == 3 | |
# defaults | |
PYVER = '2.7' | |
PYSHORTVER = '27' | |
PYTHON = None # it will be set later | |
PYTHON_ARCH = 'UNKNOWN' | |
# convenience access to the wxPython version digits | |
version2 = "%d.%d" % (version.VER_MAJOR, version.VER_MINOR) | |
version3 = "%d.%d.%d" % (version.VER_MAJOR, version.VER_MINOR, version.VER_RELEASE) | |
version2_nodot = version2.replace(".", "") | |
version3_nodot = version3.replace(".", "") | |
# same for the wxWidgets version | |
wxversion2 = "%d.%d" % (version.wxVER_MAJOR, version.wxVER_MINOR) | |
wxversion3 = "%d.%d.%d" % (version.wxVER_MAJOR, version.wxVER_MINOR, version.wxVER_RELEASE) | |
wxversion2_nodot = wxversion2.replace(".", "") | |
wxversion3_nodot = wxversion3.replace(".", "") | |
unstable_series = (version.wxVER_MINOR % 2) == 1 # is the minor version odd or even? | |
isWindows = sys.platform.startswith('win') | |
isDarwin = sys.platform == "darwin" | |
devMode = False | |
baseName = version.PROJECT_NAME | |
eggInfoName = baseName + '.egg-info' | |
defaultMask='%s-%s*' % (baseName, version.VER_MAJOR) | |
pyICON = 'packaging/docset/Vippi-blocks-icon-32.png' | |
wxICON = 'packaging/docset/mondrian.png' | |
# Some tools will be downloaded for the builds. These are the versions and | |
# MD5s of the tool binaries currently in use. | |
wafCurrentVersion = '2.0.24' | |
wafMD5 = '698f382cca34a08323670f34830325c4' | |
doxygenCurrentVersion = '1.8.8' | |
doxygenMD5 = { | |
'darwin' : '71c590e6cab47100f23919a2696cc7fd', | |
'win32' : 'a3dcff227458e423c132f16f57e26510', | |
'linux' : '083b3d8f614b538696041c7364e0f334', | |
} | |
# And the location where they can be downloaded from | |
toolsURL = 'https://wxpython.org/Phoenix/tools' | |
# MS Edge code and DLLs needed for the wxWEBVIEW_BACKEND_EDGE backend | |
MS_edge_version = '1.0.1185.39' | |
MS_edge_url = 'https://www.nuget.org/api/v2/package/Microsoft.Web.WebView2/{}'.format(MS_edge_version) | |
#--------------------------------------------------------------------------- | |
def usage(): | |
print ("""\ | |
Usage: ./build.py [command(s)] [options] | |
Commands: | |
N.N NN Major.Minor version number of the Python to use to run | |
the other commands. Default is 2.7. Or you can use | |
--python to specify the actual Python executable to use. | |
dox Run Doxygen to produce the XML file used by ETG scripts | |
doxhtml Run Doxygen to create the HTML documentation for wx | |
touch 'touch' the etg files so they will all get run the next | |
time the etg command is run. | |
etg Run the ETG scripts that are out of date to update their | |
SIP files and their Sphinx input files | |
sip Run sip to generate the C++ wrapper source | |
wxlib Build the Sphinx input files for wx.lib | |
sphinx Run the documentation building process using Sphinx | |
docset Build Dash or Zeal compatible docsets | |
build Build both wxWidgets and wxPython | |
build_wx Do only the wxWidgets part of the build | |
build_py Build wxPython only | |
install Install both wxWidgets and wxPython | |
install_wx Install wxWidgets (but only if this tool was used to | |
build it) | |
install_py Install wxPython only | |
sdist Build a tarball containing all source files | |
sdist_demo Build a tarball containing just the demo and samples folders | |
bdist Create a binary tarball release of wxPython Phoenix | |
bdist_docs Build a tarball containing the documentation | |
bdist_egg Build a Python egg. Requires magic. | |
bdist_wheel Build a Python wheel. Requires magic. | |
test Run the unit test suite | |
test_* Run just the one named test module | |
clean_wx Clean the wx parts of the build | |
clean_py Clean the wxPython parts of the build | |
clean_sphinx Clean the sphinx files | |
clean Clean both wx and wxPython | |
cleanall Clean all and do a little extra scrubbing too | |
""") | |
parser = makeOptionParser() | |
parser.print_help() | |
def main(args): | |
setPythonVersion(args) | |
setDevModeOptions(args) | |
os.environ['PYTHONPATH'] = os.environ.get('PYTHONPATH', '') + os.pathsep + phoenixDir() | |
os.environ['PYTHONUNBUFFERED'] = 'yes' | |
os.environ['WXWIN'] = wxDir() | |
# Ensure that PYTHON's containing folder is on the PATH. Normally it will | |
# already be there, and probably at the beginning. But if this script was | |
# invoked with a venv Python without activating the venv (like we do on the | |
# BuildBot) then it may not be in the PATH at all. | |
os.environ['PATH'] += os.pathsep + os.path.dirname(PYTHON) | |
wxpydir = os.path.join(phoenixDir(), "wx") | |
if not os.path.exists(wxpydir): | |
os.makedirs(wxpydir) | |
if not args or 'help' in args or '--help' in args or '-h' in args: | |
usage() | |
sys.exit(1) | |
options, commands = parseArgs(args) | |
cfg = Config(noWxConfig=True) | |
msg('cfg.VERSION: %s' % cfg.VERSION) | |
msg('') | |
# For collecting test names or files, to be run after all have been pulled | |
# off the command line | |
test_names = [] | |
while commands: | |
# ensure that each command starts with the CWD being the phoenix dir. | |
os.chdir(phoenixDir()) | |
cmd = commands.pop(0) | |
if not cmd: | |
continue # ignore empty command-line args (possible with the buildbot) | |
elif cmd.startswith('test_'): | |
test_names.append('unittests/%s.py' % cmd) | |
elif cmd.startswith('unittests/test_'): | |
test_names.append(cmd) | |
elif 'cmd_'+cmd in globals(): | |
function = globals()['cmd_'+cmd] | |
function(options, args) | |
else: | |
print('*** Unknown command: ' + cmd) | |
usage() | |
sys.exit(1) | |
# Now run the collected tests names, if any | |
if test_names: | |
cmd_test(options, args, test_names) | |
msg("Done!") | |
#--------------------------------------------------------------------------- | |
# Helper functions (see also buildtools.config for more) | |
#--------------------------------------------------------------------------- | |
def setPythonVersion(args): | |
global PYVER | |
global PYSHORTVER | |
global PYTHON | |
global PYTHON_ARCH | |
havePyVer = False | |
havePyPath = False | |
for idx, arg in enumerate(args): | |
if re.match(r'^[0-9]\.[0-9][0-9]?$', arg): | |
havePyVer = True | |
PYVER = arg | |
PYSHORTVER = arg[0] + arg[2:] | |
del args[idx] | |
break | |
if re.match(r'^[0-9][0-9][0-9]?$', arg): | |
havePyVer = True | |
PYVER = '%s.%s' % (arg[0], arg[1:]) | |
PYSHORTVER = arg | |
del args[idx] | |
break | |
if arg.startswith('--python'): | |
havePyPath = True | |
if '=' in arg: | |
PYTHON = arg.split('=')[1] | |
del args[idx] | |
else: | |
PYTHON = args[idx+1] | |
del args[idx:idx+2] | |
PYVER = runcmd([PYTHON, '-c', 'import sys; print(sys.version[:3])'], | |
getOutput=True, echoCmd=False) | |
PYSHORTVER = PYVER[0] + PYVER[2:] | |
break | |
if havePyVer: | |
if isWindows and os.environ.get('TOOLS'): | |
# Use $TOOLS to find the correct Python. If set then it should be | |
# the install root of all Python's on the system, with the 64-bit | |
# ones in an amd64 subfolder, like this: | |
# | |
# $TOOLS\Python27\python.exe | |
# $TOOLS\Python33\python.exe | |
# $TOOLS\amd64\Python27\python.exe | |
# $TOOLS\amd64\Python33\python.exe | |
# | |
TOOLS = os.environ.get('TOOLS') | |
if 'cygdrive' in TOOLS: | |
TOOLS = bash2dosPath(TOOLS) | |
use64flag = '--x64' in args | |
if use64flag: | |
args.remove('--x64') | |
CPU = os.environ.get('CPU') | |
if use64flag or CPU in ['AMD64', 'X64', 'amd64', 'x64']: | |
TOOLS = posixjoin(TOOLS, 'amd64') | |
PYTHON = posixjoin(TOOLS, | |
'python%s' % PYSHORTVER, | |
'python.exe') | |
elif isWindows: | |
# Otherwise check if the invoking Python is the right version | |
if sys.version[:3] != PYVER: | |
msg('ERROR: The invoking Python is not the requested version. Perhaps you should use --python') | |
sys.exit(1) | |
PYTHON = sys.executable | |
PYVER = sys.version[:3] | |
PYSHORTVER = PYVER[0] + PYVER[2] | |
elif not isWindows: | |
# find a pythonX.Y on the PATH | |
PYTHON = runcmd("which python%s" % PYVER, True, False) | |
if not PYTHON: | |
# If no version or path were specified then default to the python | |
# that invoked this script | |
PYTHON = sys.executable | |
PYVER = '{}.{}'.format(sys.version_info.major, sys.version_info.minor) | |
PYSHORTVER = '{}{}'.format(sys.version_info.major, sys.version_info.minor) | |
PYTHON = os.path.abspath(PYTHON) | |
msg('Will build using: "%s"' % PYTHON) | |
msg(runcmd([PYTHON, '-c', 'import sys; print(sys.version)'], True, False)) | |
PYTHON_ARCH = runcmd( | |
[PYTHON, '-c', 'import platform; print(platform.architecture()[0])'], | |
True, False) | |
msg('Python\'s architecture is %s' % PYTHON_ARCH) | |
os.environ['PYTHON'] = PYTHON | |
if PYTHON_ARCH == '64bit': | |
# Make sure this is set in case it wasn't above. | |
os.environ['CPU'] = 'X64' | |
def setDevModeOptions(args): | |
# Using --dev is a shortcut for setting several build options that I use | |
# while working on the code in my local workspaces. Most people will | |
# probably not use this so it is not part for the documented options and | |
# is explicitly handled here before the options parser is created. If | |
# anybody besides Robin is using this option do not depend on the options | |
# it inserts into the args list being consistent. They could change at any | |
# update from the repository. | |
myDevModeOptions = [ | |
#'--build_dir=../bld', | |
#'--prefix=/opt/wx/2.9', | |
'--jobs={}'.format(max(2, int(numCPUs()/2))), | |
# These will be ignored on the other platforms so it is okay to | |
# include them unconditionally | |
'--osx_cocoa', | |
'--mac_arch=arm64,x86_64', | |
'--no_allmo', | |
] | |
if not isWindows: | |
myDevModeOptions.append('--debug') | |
if isWindows: | |
myDevModeOptions.append('--cairo') | |
if '--dev' in args: | |
global devMode | |
devMode = True | |
idx = args.index('--dev') | |
# replace the --dev item with the items from the list | |
args[idx:idx+1] = myDevModeOptions | |
def numCPUs(): | |
""" | |
Detects the number of CPUs on a system. | |
This approach is from detectCPUs here: | |
http://www.artima.com/weblogs/viewpost.jsp?thread=230001 | |
""" | |
# Linux, Unix and MacOS: | |
if hasattr(os, "sysconf"): | |
if "SC_NPROCESSORS_ONLN" in os.sysconf_names: | |
# Linux & Unix: | |
ncpus = os.sysconf("SC_NPROCESSORS_ONLN") | |
if isinstance(ncpus, int) and ncpus > 0: | |
return ncpus | |
else: # OSX: | |
p = subprocess.Popen("sysctl -n hw.ncpu", shell=True, stdout=subprocess.PIPE) | |
return int(p.stdout.read()) | |
# Windows: | |
if "NUMBER_OF_PROCESSORS" in os.environ: | |
ncpus = int(os.environ["NUMBER_OF_PROCESSORS"]) | |
if ncpus > 0: | |
return ncpus | |
return 1 # Default | |
def getMSWSettings(options): | |
checkCompiler(quiet=True) | |
class MSWsettings(object): | |
pass | |
msw = MSWsettings() | |
msw.CPU = os.environ.get('CPU') | |
if msw.CPU in ['AMD64', 'X64'] or PYTHON_ARCH == '64bit': | |
msw.dllDir = posixjoin(wxDir(), "lib", "vc%s_x64_dll" % getVisCVersion()) | |
else: | |
msw.dllDir = posixjoin(wxDir(), "lib", "vc%s_dll" % getVisCVersion()) | |
msw.buildDir = posixjoin(wxDir(), "build", "msw") | |
msw.dll_type = "u" | |
if options.debug: | |
msw.dll_type = "ud" | |
return msw | |
def makeOptionParser(): | |
OPTS = [ | |
("python", ("", "The python executable to build for.")), | |
("debug", (False, "Build wxPython with debug symbols")), | |
("relwithdebug", (False, "Turn on the generation of debug info for release builds on MSW.")), | |
("release", (False, "Turn off some development options for a release build.")), | |
("keep_hash_lines",(False, "Don't remove the '#line N' lines from the SIP generated code")), | |
("gtk2", (False, "On Linux build for gtk2 (default gtk3)")), | |
("gtk3", (True, "On Linux build for gtk3")), | |
("osx_cocoa", (True, "Build the OSX Cocoa port on Mac (default)")), | |
("osx_carbon", (False, "Build the OSX Carbon port on Mac (unsupported)")), | |
("mac_framework", (False, "Build wxWidgets as a Mac framework.")), | |
("mac_arch", ("", "Comma separated list of architectures to build on Mac")), | |
("use_syswx", (False, "Try to use an installed wx rather than building the " | |
"one in this source tree. The wx-config in {prefix}/bin " | |
"or the first found on the PATH determines which wx is " | |
"used. Implies --no_magic.")), | |
("force_config", (False, "Run configure when building even if the script " | |
"determines it's not necessary.")), | |
("no_config", (False, "Turn off configure step on autoconf builds")), | |
("no_magic", (False, "Do NOT use the magic that will enable the wxWidgets " | |
"libraries to be bundled with wxPython. (Such as when " | |
"using an uninstalled wx/wxPython from the build dir, " | |
"or when distributing wxPython as an egg.) When using " | |
"this flag you should either build with an already " | |
"installed wxWidgets, or allow this script to build and " | |
"install wxWidgets independently of wxPython.")), | |
("build_dir", ("", "Directory to store wx build files. (Not used on Windows)")), | |
("prefix", ("", "Prefix value to pass to the wx build.")), | |
("destdir", ("", "Installation root for wxWidgets, files will go to {destdir}/{prefix}")), | |
("extra_setup", ("", "Extra args to pass on setup.py's command line.")), | |
("extra_make", ("", "Extra args to pass on [n]make's command line.")), | |
("extra_waf", ("", "Extra args to pass on waf's command line.")), | |
("extra_pytest", ("", "Extra args to pass on py.test's command line.")), | |
(("j","jobs"), ("", "Number of parallel compile jobs to do, if supported.")), | |
("both", (False, "Build both a debug and release version. (Only used on Windows)")), | |
("unicode", (True, "Build wxPython with unicode support (always on for wx2.9+)")), | |
(("v", "verbose"), (False, "Print out more information during the build.")), | |
("nodoc", (False, "Do not run the default docs generator")), | |
("upload", (False, "Upload bdist and/or sdist packages to snapshot server.")), | |
("cairo", (False, "Allow Cairo use with wxGraphicsContext (Windows only)")), | |
("x64", (False, "Use and build for the 64bit version of Python on Windows")), | |
("jom", (False, "Use jom instead of nmake for the wxMSW build")), | |
("pytest_timeout", ("0", "Timeout, in seconds, for stopping stuck test cases. (Currently not working as expected, so disabled by default.)")), | |
("pytest_jobs", ("", "Number of parallel processes py.test should run")), | |
("docker_img", ("all", "Comma separated list of image tags to use for the build_docker command. Defaults to \"all\"")), | |
("dump_waf_log", (False, "If the waf build tool fails then using this option will cause waf's configure log to be printed")), | |
("regenerate_sysconfig", (False, "Waf uses Python's sysconfig and related tools to configure the build. In some cases that info can be incorrect, so this option regenerates it. Must have write access to Python's lib folder.")), | |
("no_allmo", (False, "Skip regenerating the wxWidgets message catalogs")), | |
("no_msedge", (False, "Do not include the MS Edge backend for wx.html2.WebView. (Windows only)")), | |
] | |
parser = optparse.OptionParser("build options:") | |
for opt, info in OPTS: | |
default, txt = info | |
action = 'store' | |
if type(default) == bool: | |
action = 'store_true' | |
if isinstance(opt, str): | |
opts = ('--'+opt, ) | |
dest = opt | |
else: | |
opts = ('-'+opt[0], '--'+opt[1]) | |
dest = opt[1] | |
parser.add_option(*opts, default=default, action=action, | |
dest=dest, help=txt) | |
return parser | |
def parseArgs(args): | |
# If WXPYTHON_BUILD_ARGS is set in the environment, split it and add to args | |
if os.environ.get('WXPYTHON_BUILD_ARGS', None): | |
args += shlex.split(os.environ.get('WXPYTHON_BUILD_ARGS')) | |
# Parse the args into options | |
parser = makeOptionParser() | |
options, args = parser.parse_args(args) | |
if isWindows: | |
# We always use magic on Windows | |
options.no_magic = False | |
options.use_syswx = False | |
elif options.use_syswx: | |
# Turn off magic if using the system wx | |
options.no_magic = True | |
# Some options don't make sense for release builds | |
if options.release: | |
options.debug = False | |
options.both = False | |
if os.path.exists('REV.txt'): | |
os.unlink('REV.txt') | |
if options.gtk2: | |
options.gtk3 = False | |
return options, args | |
class pushDir(object): | |
def __init__(self, newDir): | |
self.cwd = os.getcwd() | |
os.chdir(newDir) | |
def __del__(self): | |
# pop back to the original dir | |
os.chdir(self.cwd) | |
def getBuildDir(options): | |
if not isDarwin and not isWindows: | |
BUILD_DIR = opj(phoenixDir(), 'build', 'wxbld', 'gtk3' if options.gtk3 else 'gtk2') | |
else: | |
BUILD_DIR = opj(phoenixDir(), 'build', 'wxbld') | |
if options.build_dir: | |
BUILD_DIR = os.path.abspath(options.build_dir) | |
return BUILD_DIR | |
def deleteIfExists(deldir, verbose=True): | |
if os.path.exists(deldir) and os.path.isdir(deldir): | |
try: | |
if verbose: | |
msg("Removing folder: %s" % deldir) | |
shutil.rmtree(deldir) | |
except Exception: | |
if verbose: | |
import traceback | |
msg("Error: %s" % traceback.format_exc(1)) | |
else: | |
if verbose: | |
msg("Unable to delete: '%s' (it doesn't exist or is not a folder)" % deldir) | |
def delFiles(fileList, verbose=True): | |
for afile in fileList: | |
if verbose: | |
print("Removing file: %s" % afile) | |
os.remove(afile) | |
def getTool(cmdName, version, MD5, envVar, platformBinary, linuxBits=False): | |
# Check in the bin dir for the specified version of the tool command. If | |
# it's not there then attempt to download it. Validity of the binary is | |
# checked with an MD5 hash. | |
if os.environ.get(envVar): | |
# Setting a value in the environment overrides other options | |
return os.environ.get(envVar) | |
else: | |
# setup | |
if platformBinary: | |
platform = getToolsPlatformName(linuxBits) | |
ext = '' | |
if platform == 'win32': | |
ext = '.exe' | |
cmd = opj(phoenixDir(), 'bin', '%s-%s-%s%s' % (cmdName, version, platform, ext)) | |
md5 = MD5[platform] | |
else: | |
cmd = opj(phoenixDir(), 'bin', '%s-%s' % (cmdName, version)) | |
md5 = MD5 | |
def _error_msg(txt): | |
msg('ERROR: ' + txt) | |
msg(' Set %s in the environment to use a local build of %s instead' % (envVar, cmdName)) | |
msg('Checking for %s...' % cmd) | |
if os.path.exists(cmd): | |
# if the file exists run some verification checks on it | |
# first make sure it is a normal file | |
if not os.path.isfile(cmd): | |
_error_msg('%s exists but is not a regular file.' % cmd) | |
sys.exit(1) | |
# now check the MD5 if not in dev mode and it's set to None | |
if not (devMode and md5 is None): | |
m = hashlib.md5() | |
with open(cmd, 'rb') as fid: | |
m.update(fid.read()) | |
if m.hexdigest() != md5: | |
_error_msg('MD5 mismatch, got "%s"\n ' | |
'expected "%s"' % (m.hexdigest(), md5)) | |
sys.exit(1) | |
# If the cmd is a script run by some interpreter, or similar, | |
# then we don't need to check anything else | |
if not platformBinary: | |
return cmd | |
# Ensure that commands that are platform binaries are executable | |
if not os.access(cmd, os.R_OK | os.X_OK): | |
_error_msg('Cannot execute %s due to permissions error' % cmd) | |
sys.exit(1) | |
try: | |
p = subprocess.Popen([cmd, '--help'], stdout=subprocess.PIPE, stderr=subprocess.PIPE, env=os.environ) | |
p.communicate() | |
except OSError as e: | |
_error_msg('Could not execute %s, got "%s"' % (cmd, e)) | |
sys.exit(1) | |
# if we get this far then all is well, the cmd is good to go | |
return cmd | |
msg('Not found. Attempting to download...') | |
url = '%s/%s.bz2' % (toolsURL, os.path.basename(cmd)) | |
try: | |
import requests | |
resp = requests.get(url) | |
resp.raise_for_status() | |
msg('Connection successful...') | |
data = resp.content | |
msg('Data downloaded...') | |
except Exception: | |
_error_msg('Unable to download ' + url) | |
import traceback | |
traceback.print_exc() | |
sys.exit(1) | |
import bz2 | |
data = bz2.decompress(data) | |
with open(cmd, 'wb') as f: | |
f.write(data) | |
os.chmod(cmd, 0o755) | |
# Recursive call so the MD5 value will be double-checked on what was | |
# just downloaded | |
return getTool(cmdName, version, MD5, envVar, platformBinary, linuxBits) | |
# The download and MD5 check only needs to happen once per run, cache the sip | |
# cmd value here the first time through. | |
_wafCmd = None | |
def getWafCmd(): | |
global _wafCmd | |
if _wafCmd is None: | |
_wafCmd = getTool('waf', wafCurrentVersion, wafMD5, 'WAF', False) | |
return _wafCmd | |
# and Doxygen | |
_doxCmd = None | |
def getDoxCmd(): | |
global _doxCmd | |
if _doxCmd is None: | |
_doxCmd = getTool('doxygen', doxygenCurrentVersion, doxygenMD5, 'DOXYGEN', True) | |
return _doxCmd | |
def getMSWebView2(): | |
fname = '{}.zip'.format(MS_edge_version) | |
dest = opj(wxDir(), '3rdparty', 'webview2') | |
if not os.path.exists(dest) or not os.path.exists(opj(dest, fname)): | |
if os.path.exists(dest): | |
shutil.rmtree(dest) | |
os.makedirs(dest) | |
msg('Downloading microsoft.web.webview2 {}...'.format(MS_edge_version)) | |
try: | |
import requests | |
resp = requests.get(MS_edge_url) | |
resp.raise_for_status() | |
msg('Connection successful...') | |
data = resp.content | |
msg('Data downloaded...') | |
except Exception: | |
import traceback | |
traceback.print_exc() | |
sys.exit(1) | |
# Write the downloaded data to a local file | |
with open(opj(dest, fname), 'wb') as f: | |
f.write(data) | |
# Unzip it | |
from zipfile import ZipFile | |
with ZipFile(opj(dest, fname)) as zip: | |
zip.extractall(dest) | |
class CommandTimer(object): | |
def __init__(self, name): | |
self.name = name | |
self.startTime = datetime.datetime.now() | |
msg('Running command: %s' % self.name) | |
def __del__(self): | |
delta = datetime.datetime.now() - self.startTime | |
time = "" | |
if delta.seconds / 60 > 0: | |
time = "%dm" % (delta.seconds / 60) | |
time += "%d.%ds" % (delta.seconds % 60, delta.microseconds / 1000) | |
msg('Finished command: %s (%s)' % (self.name, time)) | |
def uploadPackage(fileName, options, mask=defaultMask, keep=75): | |
""" | |
Upload the given filename to the configured package server location. Only | |
the `keep` most recent files matching `mask` will be kept so the server | |
space is not overly consumed. It is assumed that if the files are in | |
sorted order then the end of the list will be the newest files. | |
""" | |
fileName = os.path.relpath(fileName) | |
fileName = fileName.replace('\\', '/') | |
msg("Uploading %s..." % fileName) | |
# NOTE: It is expected that there will be a host entry defined in | |
# ~/.ssh/config named wxpython-rbot, with the proper host, user, identity | |
# file, etc. needed for making an SSH connection to the snapshots server. | |
# Release builds work similarly with their own host configuration defined | |
# in the ~/.ssh/config file. | |
if options.release: | |
host = 'wxpython-release' | |
uploadDir = 'release-builds' | |
else: | |
host = 'wxpython-rbot' | |
uploadDir = 'snapshot-builds' | |
# copy the new file to the server | |
cmd = 'scp {} {}:{}'.format(fileName, host, uploadDir) | |
runcmd(cmd) | |
# Make sure it is readable by all, and writable by rbot | |
cmd = 'ssh {} "cd {}; chmod 644 {}"'.format(host, uploadDir, os.path.basename(fileName)) | |
runcmd(cmd) | |
if not options.release: | |
# get the list of all snapshot files on the server | |
cmd = 'ssh {} "cd {}; ls {}"'.format(host, uploadDir, mask) | |
allFiles = runcmd(cmd, getOutput=True) | |
allFiles = allFiles.strip().split('\n') | |
allFiles.sort() | |
# Leave the last keep number of builds, including this new one, on the server. | |
# Delete the rest. | |
rmFiles = allFiles[:-keep] | |
if rmFiles: | |
msg("Deleting %s" % ", ".join(rmFiles)) | |
cmd = 'ssh {} "cd {}; rm {}"'.format(host, uploadDir, " ".join(rmFiles)) | |
runcmd(cmd) | |
msg("Upload complete!") | |
def uploadTree(srcPath, destPath, options, days=30): | |
""" | |
Similar to the above but uploads a tree of files. | |
""" | |
msg("Uploading tree at {}...".format(srcPath)) | |
if options.release: | |
host = 'wxpython-release' | |
uploadDir = opj('release-builds', destPath) | |
else: | |
host = 'wxpython-rbot' | |
uploadDir = opj('snapshot-builds', destPath) | |
# Ensure the destination exists | |
cmd = 'ssh {0} "if [ ! -d {1} ]; then mkdir -p {1}; fi"'.format(host, uploadDir) | |
runcmd(cmd) | |
# Upload the tree | |
cmd = 'scp -r {} {}:{}'.format(srcPath, host, uploadDir) | |
runcmd(cmd) | |
# Make sure it is readable by all | |
cmd = 'ssh {} "chmod -R a+r {}"'.format(host, uploadDir) | |
runcmd(cmd) | |
if not options.release: | |
# Remove files that were last modified more than `days` days ago | |
msg("Cleaning up old builds.") | |
cmd = 'ssh {} "find {} -type f -mtime +{} -delete"'.format(host, uploadDir, days) | |
runcmd(cmd) | |
msg("Tree upload and cleanup complete!") | |
def checkCompiler(quiet=False): | |
if isWindows: | |
# Set up the PATH and other environment variables so the proper version | |
# of MSVC will be used. The setuptools package is used to find the | |
# needed info, so the target python shoudl have a recent version of | |
# setuptools installed. | |
arch = 'x64' if PYTHON_ARCH == '64bit' else 'x86' | |
info = getMSVCInfo(PYTHON, arch, set_env=True) | |
# Make sure there is now a cl.exe on the PATH | |
CL = 'NOT FOUND' | |
for d in os.environ['PATH'].split(os.pathsep): | |
p = pathlib.Path(d, 'cl.exe') | |
if p.exists(): | |
CL = p | |
break | |
if not quiet: | |
msg(f"CL.exe: {CL}") | |
# Just needed for debugging | |
# msg('include: ' + info.include) | |
# msg('lib: ' + info.lib) | |
# msg('libpath: ' + info.libpath) | |
# for d in info.include.split(os.pathsep): | |
# p = pathlib.Path(d, 'tchar.h') | |
# if p.exists(): | |
# msg('tchar.h: ' + str(p)) | |
# break | |
# NOTE: SIP is now generating code with scoped-enums. Older linux | |
# platforms like what we're using for builds, and also TravisCI for | |
# example, are using GCC versions that are still defaulting to C++98, | |
# so this flag is needed to turn on the C++11 mode. If this flag | |
# causes problems with other non-Windows, non-Darwin compilers then | |
# we'll need to make this a little smarter about what flag (if any) | |
# needs to be used. | |
# | |
# NOTE 2: SIP changed its output such that this doesn't appear to be | |
# needed anymore, but we'll leave the code in place to make it easy to | |
# turn it back on again if/when needed. | |
if False and not isWindows and not isDarwin: | |
stdflag = '-std=c++11' | |
curflags = os.environ.get('CXXFLAGS', '') | |
if stdflag not in curflags: | |
os.environ['CXXFLAGS'] = '{} {}'.format(stdflag, curflags) | |
#print('**** Using CXXFLAGS:', os.environ.get('CXXFLAGS', '')) | |
def getWafBuildBase(): | |
base = posixjoin('build', 'waf', PYVER) | |
if isWindows: | |
if PYTHON_ARCH == '64bit': | |
base = posixjoin(base, 'x64') | |
else: | |
base = posixjoin(base, 'x86') | |
return base | |
def getBashPath(): | |
"""Check if there is a bash.exe on the PATH""" | |
bash = which('bash.exe') | |
return bash | |
def dos2bashPath(path): | |
""" | |
Convert an absolute dos-style path to one bash.exe can understand. | |
""" | |
path = path.replace('\\', '/') | |
cygpath = which('cygpath') | |
wsl = which('wsl') | |
# If we have cygwin then we can use cygpath to convert the path. | |
# Note that MSYS2 (and Git Bash) now also have cygpath so this should | |
# work there too. | |
if cygpath: | |
path = runcmd('"{}" -u "{}"'.format(cygpath, path), getOutput=True, echoCmd=False) | |
return path | |
elif wsl: | |
# Are we using Windows System for Linux? (untested) | |
path = runcmd('"{}" wslpath -a -u "{}"'.format(wsl, path), getOutput=True, echoCmd=False) | |
return path | |
else: | |
# Otherwise, do a simple translate and hope for the best? | |
# c:/foo --> /c/foo | |
# TODO: Check this!! | |
drive, rest = os.path.splitdrive(path) | |
path = '/{}/{}'.format(drive[0], rest) | |
return path | |
def bash2dosPath(path): | |
""" | |
Convert an absolute unix-style path to one Windows can understand. | |
""" | |
cygpath = which('cygpath') | |
wsl = which('wsl') | |
# If we have cygwin then we can use cygpath to convert the path. | |
# Note that MSYS2 (and Git Bash) now also have cygpath so this should | |
# work there too. | |
if cygpath: | |
path = runcmd('"{}" -w "{}"'.format(cygpath, path), getOutput=True, echoCmd=False) | |
return path | |
elif wsl: | |
# Are we using Windows System for Linux? (untested) | |
path = runcmd('"{}" wslpath -a -w "{}"'.format(wsl, path), getOutput=True, echoCmd=False) | |
return path | |
else: | |
# Otherwise, do a simple translate and hope for the best? | |
# /c/foo --> c:/foo | |
# There's also paths like /cygdrive/c/foo or /mnt/c/foo, but in those | |
# cases cygpath or wsl should be available. | |
components = path.split('/') | |
assert components[0] == '' and len(components[1]) == 1, "Expecting a path like /c/foo" | |
path = components[1] + ':/' + '/'.join(components[2:]) | |
return path | |
def do_regenerate_sysconfig(): | |
""" | |
If a Python environment has been relocated to a new folder then it's | |
possible that the sysconfig can still be using paths for the original | |
location. Since wxPython's build uses WAF, which uses the sysconfig (via | |
python-config, distutils.sysconfig, etc.) then we need to ensure that these | |
paths match the current environment. | |
TODO: Can this be done in a way that doesn't require overwriting a file in | |
the environment? | |
""" | |
with TemporaryDirectory() as td: | |
pwd = pushDir(td) | |
# generate a new sysconfig data file | |
cmd = [PYTHON, '-m', 'sysconfig', '--generate-posix-vars'] | |
runcmd(cmd) | |
# On success the new data module will have been written to a subfolder | |
# of the current folder, which is recorded in ./pybuilddir.txt | |
with open('pybuilddir.txt', 'r') as fp: | |
pybd = fp.read() | |
# grab the file in that folder and copy it into the Python lib | |
p = opj(td, pybd, '*') | |
datafile = glob.glob(opj(td, pybd, '*'))[0] | |
cmd = [PYTHON, '-c', 'import sysconfig; print(sysconfig.get_path("stdlib"))'] | |
stdlib = runcmd(cmd, getOutput=True) | |
shutil.copy(datafile, stdlib) | |
del pwd | |
def _setTarItemPerms(tarinfo): | |
""" | |
Used to set permissions of the files and fodlers in the source tarball | |
""" | |
if tarinfo.isdir(): | |
tarinfo.mode = 0o755 | |
else: | |
tarinfo.mode |= 0o644 | |
return tarinfo | |
#--------------------------------------------------------------------------- | |
# Command functions and helpers | |
#--------------------------------------------------------------------------- | |
def _doDox(arg): | |
doxCmd = getDoxCmd() | |
doxCmd = os.path.abspath(doxCmd) | |
if isWindows: | |
bash = getBashPath() | |
if not bash: | |
raise RuntimeError("ERROR: Unable to find bash.exe, needed for running regen.sh") | |
doxCmd = dos2bashPath(doxCmd) | |
print(doxCmd) | |
os.environ['DOXYGEN'] = doxCmd | |
os.environ['WX_SKIP_DOXYGEN_VERSION_CHECK'] = '1' | |
d = posixjoin(wxDir(), 'docs/doxygen') | |
d = d.replace('\\', '/') | |
cmd = '"{}" -l -c "cd {} && ./regen.sh {}"'.format(bash, d, arg) | |
else: | |
os.environ['DOXYGEN'] = doxCmd | |
os.environ['WX_SKIP_DOXYGEN_VERSION_CHECK'] = '1' | |
pwd = pushDir(posixjoin(wxDir(), 'docs/doxygen')) | |
cmd = 'bash ./regen.sh %s' % arg | |
runcmd(cmd) | |
def cmd_dox(options, args): | |
cmdTimer = CommandTimer('dox') | |
_doDox('xml') | |
def cmd_doxhtml(options, args): | |
cmdTimer = CommandTimer('doxhtml') | |
_doDox('html') | |
_doDox('chm') | |
def cmd_docset_wx(options, args): | |
cmdTimer = CommandTimer('docset_wx') | |
cfg = Config() | |
# Use Doxygen to generate the docset | |
_doDox('docset') | |
# Remove any existing docset in the dist dir and move the new docset in | |
srcname = posixjoin(wxDir(), 'docs/doxygen/out/docset', | |
'wxWidgets-%s.docset' % wxversion2) | |
destname = 'dist/wxWidgets-%s.docset' % wxversion3 | |
if not os.path.isdir(srcname): | |
msg('ERROR: %s not found' % srcname) | |
sys.exit(1) | |
if os.path.isdir(destname): | |
shutil.rmtree(destname) | |
shutil.move(srcname, destname) | |
shutil.copyfile(wxICON, posixjoin(destname, 'icon.png')) | |
def cmd_docset_py(options, args): | |
cmdTimer = CommandTimer('docset_py') | |
cfg = Config(noWxConfig=True) | |
if not os.path.isdir('docs/html'): | |
msg('ERROR: No docs/html dir found, has the sphinx build command been run?') | |
sys.exit(1) | |
# clear out any old docset build | |
name = 'wxPython-{}'.format(cfg.VERSION) | |
docset = posixjoin('dist', '{}.docset'.format(name)) | |
if os.path.isdir(docset): | |
shutil.rmtree(docset) | |
# run the docset generator | |
VERBOSE = '--verbose' if options.verbose else '' | |
URL = 'https://docs.wxpython.org/' if options.release else 'https://wxpython.org/Phoenix/docs/html/' | |
cmd = [PYTHON, '-m doc2dash', VERBOSE, | |
'--name', name, '--icon', pyICON, | |
'--index-page index.html', '--enable-js', | |
'--online-redirect-url', URL, | |
'--destination dist docs/html'] | |
runcmd(' '.join(cmd)) | |
# Remove the sidebar from the pages in the docset | |
msg('Removing sidebar from docset pages...') | |
_removeSidebar(opj('dist', name+'.docset', 'Contents', 'Resources', 'Documents')) | |
# build the tarball | |
msg('Archiving Phoenix docset...') | |
rootname = "wxPython-docset-{}".format(cfg.VERSION) | |
tarfilename = posixjoin(phoenixDir(), 'dist', '{}.tar.gz'.format(rootname)) | |
if os.path.exists(tarfilename): | |
os.remove(tarfilename) | |
with tarfile.open(name=tarfilename, mode="w:gz") as tarball: | |
tarball.add(opj('dist', name+'.docset'), name+'.docset', filter=_setTarItemPerms) | |
if options.upload: | |
uploadPackage(tarfilename, options, keep=5, | |
mask='%s-docset-%s*' % (baseName, cfg.VER_MAJOR)) | |
msg("Docset file built at %s" % tarfilename) | |
def _removeSidebar(path): | |
""" | |
Remove the sidebar <div> from the pages going into the docset | |
""" | |
from bs4 import BeautifulSoup | |
for filename in glob.glob(opj(path, '*.html')): | |
with textfile_open(filename, 'rt') as f: | |
text = f.read() | |
text = text.replace('<script src="_static/javascript/sidebar.js" type="text/javascript"></script>', '') | |
soup = BeautifulSoup(text, 'html.parser') | |
tag = soup.find('div', 'sphinxsidebar') | |
if tag: | |
tag.extract() | |
tag = soup.find('div', 'document') | |
if tag: | |
tag.attrs['class'] = ['document-no-sidebar'] | |
text = unicode(soup) if PY2 else str(soup) | |
with textfile_open(filename, 'wt') as f: | |
f.write(text) | |
def cmd_docset(options, args): | |
cmd_docset_wx(options, args) | |
cmd_docset_py(options, args) | |
def cmd_etg(options, args): | |
cmdTimer = CommandTimer('etg') | |
cfg = Config() | |
assert os.path.exists(cfg.DOXY_XML_DIR), "Doxygen XML folder not found: " + cfg.DOXY_XML_DIR | |
pwd = pushDir(cfg.ROOT_DIR) | |
# TODO: Better support for selecting etg cmd-line flags... | |
flags = '--sip' | |
if options.nodoc: | |
flags += ' --nodoc' | |
# get the files to run, moving _core the to the front of the list | |
etgfiles = glob.glob(opj('etg', '_*.py')) | |
core_file = opj('etg', '_core.py') | |
if core_file in etgfiles: | |
etgfiles.remove(core_file) | |
etgfiles.insert(0, core_file) | |
for script in etgfiles: | |
sipfile = etg2sip(script) | |
deps = [script] | |
ns = loadETG(script) | |
if hasattr(ns, 'ETGFILES'): | |
etgfiles += ns.ETGFILES[1:] # all but itself | |
if hasattr(ns, 'DEPENDS'): | |
deps += ns.DEPENDS | |
if hasattr(ns, 'OTHERDEPS'): | |
deps += ns.OTHERDEPS | |
# run the script only if any dependencies are newer | |
if newer_group(deps, sipfile): | |
runcmd('"%s" %s %s' % (PYTHON, script, flags)) | |
def cmd_sphinx(options, args): | |
from sphinxtools.postprocess import genIndexes, makeHeadings, postProcess, genGallery | |
cmdTimer = CommandTimer('sphinx') | |
pwd = pushDir(phoenixDir()) | |
sphinxDir = os.path.join(phoenixDir(), 'docs', 'sphinx') | |
if not os.path.isdir(sphinxDir): | |
raise Exception('Missing sphinx folder in the distribution') | |
textFiles = glob.glob(sphinxDir + '/*.txt') | |
if not textFiles: | |
raise Exception('No documentation files found. Please run "build.py touch etg" first') | |
# Copy the rst files into txt files | |
restDir = os.path.join(sphinxDir, 'rest_substitutions', 'overviews') | |
rstFiles = glob.glob(restDir + '/*.rst') | |
for rst in rstFiles: | |
rstName = os.path.split(rst)[1] | |
txt = os.path.join(sphinxDir, os.path.splitext(rstName)[0] + '.txt') | |
copyIfNewer(rst, txt) | |
genIndexes(sphinxDir) | |
genGallery() | |
# Copy the hand-edited top level doc files too | |
rstFiles = glob.glob(os.path.join(phoenixDir(), 'docs', '*.rst')) | |
for rst in rstFiles: | |
txt = os.path.join(sphinxDir, os.path.splitext(os.path.basename(rst))[0] + '.txt') | |
copyIfNewer(rst, txt) | |
makeHeadings() | |
pwd2 = pushDir(sphinxDir) | |
buildDir = os.path.join(sphinxDir, 'build') | |
htmlDir = os.path.join(phoenixDir(), 'docs', 'html') | |
runcmd('{} -m sphinx -b html -d {}/doctrees . {}'.format(PYTHON, buildDir, htmlDir)) | |
del pwd2 | |
msg('Postprocessing sphinx output...') | |
postProcess(htmlDir, options) | |
def cmd_wxlib(options, args): | |
from sphinxtools.modulehunter import ModuleHunter | |
cmdTimer = CommandTimer('wxlib') | |
pwd = pushDir(phoenixDir()) | |
for wx_pkg in ['lib', 'py', 'svg', 'tools']: | |
libDir = os.path.join(phoenixDir(), 'wx', wx_pkg) | |
if not os.path.isdir(libDir): | |
raise Exception('Missing wx.{} folder in the distribution'.format(wx_pkg)) | |
init_name = os.path.join(libDir, '__init__.py') | |
import_name = 'wx.{}'.format(wx_pkg) | |
ModuleHunter(init_name, import_name, version3) | |
def cmd_wxpy(options, args): | |
msg('Command wxpy has been folded into command wxlib.') | |
def cmd_wxtools(options, args): | |
msg('Command wxtools has been folded into command wxlib.') | |
def cmd_bdist_docs(options, args): | |
cmdTimer = CommandTimer('bdist_docs') | |
pwd = pushDir(phoenixDir()) | |
cfg = Config() | |
msg("Archiving wxPython Phoenix documentation...") | |
rootname = "%s-docs-%s" % (baseName, cfg.VERSION) | |
tarfilename = posixjoin(phoenixDir(), 'dist', '%s.tar.gz' % rootname) | |
if not os.path.exists('dist'): | |
os.makedirs('dist') | |
if os.path.exists(tarfilename): | |
os.remove(tarfilename) | |
with tarfile.open(name=tarfilename, mode="w:gz") as tarball: | |
tarball.add('docs/html', os.path.join(rootname, 'docs/html'), | |
filter=_setTarItemPerms) | |
if options.upload: | |
uploadPackage(tarfilename, options, keep=5, | |
mask='%s-docs-%s*' % (baseName, cfg.VER_MAJOR)) | |
msg('Documentation tarball built at %s' % tarfilename) | |
# # pythonhosted.org can host the wxPython documentation for us, so let's | |
# # use it for the docs associated with the latest release of wxPython. It | |
# # requires that the docs be in a .zip file with an index.html file at the | |
# # top level. To build this we'll just need to do like the above tarball | |
# # code, except add the files from within the docs/html folder so they will | |
# # all be at the top level of the archive. shutil.make_archive can be used | |
# # in this case because we don't need to rewrite the pathnames in the | |
# # archive. | |
# if options.release: | |
# msg("Archiving wxPython Phoenix documentation for pythonhosted.org...") | |
# rootname = "%s-docs-pythonhosted-%s" % (baseName, cfg.VERSION) | |
# zipfilename = "dist/%s.zip" % rootname | |
# | |
# if os.path.exists(zipfilename): | |
# os.remove(zipfilename) | |
# | |
# # with zipfile.ZipFile(zipfilename, 'w', zipfile.ZIP_DEFLATED) as zip: | |
# # pwd2 = pushDir('docs/html') | |
# | |
# zipfilename = shutil.make_archive(base_name=os.path.splitext(zipfilename)[0], | |
# format="zip", | |
# root_dir="docs/html") | |
# zipfilename = os.path.relpath(zipfilename) | |
# zipfilename = zipfilename.replace('\\', '/') | |
# | |
# if options.upload: | |
# uploadPackage(zipfilename, options, keep=5, | |
# mask='%s-docs-pythonhosted-%s*' % (baseName, cfg.VER_MAJOR)) | |
# | |
# msg('Pythonhosted zip file built at %s' % zipfilename) | |
def cmd_sip(options, args): | |
cmdTimer = CommandTimer('sip') | |
cfg = Config() | |
pwd = pushDir(cfg.ROOT_DIR) | |
modules = glob.glob(opj(cfg.SIPGEN, '_*.sip')) | |
# move _core the to the front of the list | |
core_file = opj(cfg.SIPGEN, '_core.sip') | |
if core_file in modules: | |
modules.remove(core_file) | |
modules.insert(0, core_file) | |
for src_name in modules: | |
tmpdir = tempfile.mkdtemp() | |
tmpdir = tmpdir.replace('\\', '/') | |
src_name = src_name.replace('\\', '/') | |
base = os.path.basename(os.path.splitext(src_name)[0]) | |
sbf = posixjoin(cfg.ROOT_DIR, cfg.SIPOUT, base) + '.sbf' | |
pycode = base[1:] # remove the leading _ | |
pycode = opj(cfg.ROOT_DIR, cfg.PKGDIR, pycode) + '.py' | |
# Check if any of the included files are newer than the .sbf file | |
# produced by the previous run of sip. If not then we don't need to | |
# run sip again. | |
etg = loadETG(posixjoin('etg', base + '.py')) | |
sipFiles = getSipFiles(etg.INCLUDES) + [opj(cfg.SIPGEN, base+'.sip')] | |
if not newer_group(sipFiles, sbf) and os.path.exists(pycode): | |
continue | |
# Leave it turned off for now. TODO: Experiment with this... | |
# pyi_extract = posixjoin(cfg.PKGDIR, base[1:]) + '.pyi' | |
pyi_extract = None | |
# SIP extracts are used to pull python snippets and put them into the | |
# module's .py file | |
pycode = 'pycode'+base+':'+pycode | |
# Write out a pyproject.toml to configure sip | |
pyproject_toml = textwrap.dedent("""\ | |
[build-system] | |
requires = ["sip >=6.6.2, <7"] | |
build-backend = "sipbuild.api" | |
[tool.sip.metadata] | |
name = "{base}" | |
[tool.sip.bindings.{base}] | |
docstrings = true | |
release-gil = true | |
exceptions = false | |
tracing = {tracing} | |
protected-is-public = false | |
generate-extracts = [\'{extracts}\'] | |
pep484-pyi = false | |
[tool.sip.project] | |
abi-version = "{abi_version}" | |
sip-files-dir = '{sip_gen_dir}' | |
sip-include-dirs = ['{src_dir}'] | |
sip-module = "wx.siplib" | |
""".format( | |
base=base, | |
abi_version=cfg.SIP_ABI, | |
tracing=str(cfg.SIP_TRACE).lower(), | |
extracts=pycode, | |
src_dir=opj(phoenixDir(), 'src'), | |
sip_gen_dir=opj(phoenixDir(), 'sip', 'gen'), | |
) | |
) | |
with open(opj(tmpdir, 'pyproject.toml'), 'w') as f: | |
f.write(pyproject_toml) | |
sip_pwd = pushDir(tmpdir) | |
cmd = 'sip-build --no-compile' | |
runcmd(cmd) | |
del sip_pwd | |
# Write out a sip build file (no longer done by sip itself) | |
sip_tmp_out_dir = opj(tmpdir, 'build', base) | |
sip_pwd = pushDir(sip_tmp_out_dir) | |
header = glob.glob('*.h')[0] | |
sources = glob.glob('*.cpp') | |
del sip_pwd | |
with open(sbf, 'w') as f: | |
f.write("sources = {}\n".format(' '.join(sources))) | |
f.write("headers = {}\n".format(header)) | |
classesNeedingClassInfo = { 'sip_corewxTreeCtrl.cpp' : 'wxTreeCtrl', } | |
def processSrc(src, keepHashLines=False): | |
with textfile_open(src, 'rt') as f: | |
srcTxt = f.read() | |
if keepHashLines: | |
# Either just fix the pathnames in the #line lines... | |
srcTxt = srcTxt.replace(sip_tmp_out_dir, cfg.SIPOUT) | |
else: | |
# ...or totally remove them by replacing those lines with '' | |
import re | |
srcTxt = re.sub(r'^\s*#line.*\n', '', srcTxt, flags=re.MULTILINE) | |
className = classesNeedingClassInfo.get(os.path.basename(src)) | |
if className: | |
srcTxt = injectClassInfo(className, srcTxt) | |
return srcTxt | |
def injectClassInfo(className, srcTxt): | |
# inject wxClassInfo macros into the sip generated wrapper class | |
lines = srcTxt.splitlines() | |
# find the beginning of the class declaration | |
for idx, line in enumerate(lines): | |
if line.startswith('class sip{}'.format(className)): | |
# next line is '{', insert after that | |
lines[idx+2:idx+2] = ['wxDECLARE_ABSTRACT_CLASS(sip{});'.format(className)] | |
break | |
# find the '};' that terminates the class | |
for idx, line in enumerate(lines[idx+2:], idx+2): | |
if line.startswith('};'): | |
lines[idx+1:idx+1] = ['\nwxIMPLEMENT_ABSTRACT_CLASS(sip{0}, {0});'.format(className)] | |
break | |
# join it back up and return | |
return '\n'.join(lines) | |
# Check each file in tmpdir to see if it is different than the same file | |
# in cfg.SIPOUT. If so then copy the new one to cfg.SIPOUT, otherwise | |
# ignore it. | |
for src in glob.glob(sip_tmp_out_dir + '/*'): | |
dest = opj(cfg.SIPOUT, os.path.basename(src)) | |
if not os.path.exists(dest): | |
msg('%s is a new file, copying...' % os.path.basename(src)) | |
srcTxt = processSrc(src, options.keep_hash_lines) | |
with textfile_open(dest, 'wt') as f: | |
f.write(srcTxt) | |
continue | |
srcTxt = processSrc(src, options.keep_hash_lines) | |
with textfile_open(dest, 'rt') as f: | |
destTxt = f.read() | |
if srcTxt == destTxt: | |
pass | |
else: | |
msg('%s is changed, copying...' % os.path.basename(src)) | |
with textfile_open(dest, 'wt') as f: | |
f.write(srcTxt) | |
# Remove tmpdir and its contents | |
shutil.rmtree(tmpdir) | |
# Generate sip module code | |
deleteIfExists(cfg.SIPINC) | |
with tempfile.TemporaryDirectory() as tmpdir: | |
cmd = 'sip-module --sdist --abi-version {} --target-dir {} wx.siplib'.format(cfg.SIP_ABI, tmpdir) | |
runcmd(cmd) | |
tf_name = glob.glob(tmpdir + '/*.tar*')[0] | |
tf_dir = os.path.splitext(os.path.splitext(tf_name)[0])[0] | |
with tarfile.open(tf_name) as tf: | |
tf.extractall(tmpdir) | |
shutil.move(tf_dir, cfg.SIPINC) | |
def cmd_touch(options, args): | |
cmdTimer = CommandTimer('touch') | |
pwd = pushDir(phoenixDir()) | |
etg = pathlib.Path('etg') | |
for item in etg.glob('*.py'): | |
item.touch() | |
cmd_touch_others(options, args) | |
def cmd_test(options, args, tests=None): | |
# Run all tests | |
cmdTimer = CommandTimer('test') | |
pwd = pushDir(phoenixDir()) | |
# --boxed runs each test in a new process (only for posix *&##$#@$^!!) | |
# -n is the number of processes to run in parallel | |
# --timeout will kill the test process if it gets stuck | |
jobs = '-n{}'.format(options.pytest_jobs) if options.pytest_jobs else '' | |
boxed = '--forked' if not isWindows else '' | |
sec = options.pytest_timeout | |
timeout = '--timeout={}'.format(sec) if sec and sec != "0" else '' | |
cmd = '"{}" -m pytest {} {} {} {} {} '.format( | |
PYTHON, | |
'-v' if options.verbose else '', | |
boxed, | |
jobs, | |
timeout, | |
options.extra_pytest) | |
if not tests: | |
# let pytest find all tests in the unittest folder | |
cmd += 'unittests' | |
else: | |
# otherwise, run only the test modules given | |
cmd += ' '.join(tests) | |
runcmd(cmd, fatal=False) | |
def cmd_build(options, args): | |
cmdTimer = CommandTimer('build') | |
cmd_build_wx(options, args) | |
cmd_build_py(options, args) | |
def cmd_build_wx(options, args): | |
cmdTimer = CommandTimer('build_wx') | |
if not isWindows and options.use_syswx: | |
msg("use_syswx option specified, skipping wxWidgets build") | |
return | |
checkCompiler() | |
build_options = ['--wxpython', '--unicode'] | |
if options.jobs: | |
build_options.append('--jobs=%s' % options.jobs) | |
if isWindows: | |
# Windows-specific pre build stuff | |
if not options.no_msedge: | |
getMSWebView2() | |
else: | |
build_options.append('--no_msedge') | |
if options.cairo: | |
build_options.append('--cairo') | |
cairo_root = os.path.join(phoenixDir(), 'packaging', 'msw-cairo') | |
os.environ['CAIRO_ROOT'] = cairo_root | |
if options.jom: | |
build_options.append('--jom') | |
if PY2: | |
build_options.append('--no_dpi_aware') | |
else: | |
# Platform is something other than MSW | |
if options.osx_carbon: | |
options.osx_cocoa = False | |
BUILD_DIR = getBuildDir(options) | |
DESTDIR = options.destdir | |
PREFIX = options.prefix | |
if options.mac_framework and isDarwin: | |
# TODO: Don't hard-code this path | |
PREFIX = "/Library/Frameworks/wx.framework/Versions/%s" % wxversion2 | |
if PREFIX: | |
build_options.append('--prefix=%s' % PREFIX) | |
if not os.path.exists(BUILD_DIR): | |
os.makedirs(BUILD_DIR) | |
if isDarwin: | |
if options.osx_cocoa: | |
build_options.append("--osx_cocoa") | |
if options.mac_arch: | |
build_options.append("--mac_universal_binary=%s" % options.mac_arch) | |
else: | |
build_options.append("--mac_universal_binary=default") | |
if options.no_config: | |
build_options.append('--no_config') | |
elif not options.force_config: | |
dependencies = [ os.path.join(wxDir(), 'Makefile.in'), | |
os.path.join(wxDir(), 'configure'), | |
os.path.join(wxDir(), 'setup.h.in'), | |
os.path.join(wxDir(), 'version-script.in'), | |
os.path.join(wxDir(), 'wx-config.in'), | |
] | |
for dep in dependencies: | |
if newer(dep, os.path.join(BUILD_DIR, "Makefile")): | |
break | |
else: | |
build_options.append("--no_config") | |
#if options.install: | |
# build_options.append('--installdir=%s' % DESTDIR) | |
# build_options.append("--install") | |
if options.mac_framework and isDarwin: | |
build_options.append("--mac_framework") | |
if not isDarwin: | |
if options.gtk2: | |
build_options.append('--gtk2') | |
if options.gtk3: | |
build_options.append('--gtk3') | |
# Change to what will be the wxWidgets build folder | |
# (Note, this needs to be after any testing for file/path existence, etc. | |
# because they may be specified as relative paths.) | |
pwd = pushDir(BUILD_DIR) | |
if options.debug or (isWindows and options.both): | |
build_options.append('--debug') | |
if options.extra_make: | |
build_options.append('--extra_make="%s"' % options.extra_make) | |
if not isWindows and not isDarwin and not options.no_magic and not options.use_syswx: | |
# Using $ORIGIN in the rpath will cause the dynamic linker to look | |
# for shared libraries in a folder relative to the loading binary's | |
# location. Here we'll use just $ORIGIN so it should look in the same | |
# folder as the wxPython extension modules. | |
os.environ['LD_RUN_PATH'] = '$ORIGIN' | |
try: | |
# Import and run the wxWidgets build script | |
from buildtools import build_wxwidgets as wxbuild | |
print('wxWidgets build options: ' + str(build_options)) | |
wxbuild.main(wxDir(), build_options) | |
# build again without the --debug flag? | |
if isWindows and options.both: | |
build_options.remove('--debug') | |
print('wxWidgets build options: ' + str(build_options)) | |
wxbuild.main(wxDir(), build_options) | |
except Exception: | |
print("ERROR: failed building wxWidgets") | |
import traceback | |
traceback.print_exc() | |
sys.exit(1) | |
if not options.no_allmo: | |
# Build the wx message catalogs, but first check that there is a msgfmt | |
# command available | |
if findCmd('msgfmt') and findCmd('make'): | |
locale_pwd = pushDir(posixjoin(wxDir(), 'locale')) | |
print('Building message catalogs in ' + os.getcwd()) | |
runcmd('make allmo') | |
del locale_pwd | |
else: | |
print("WARNING: msgfmt and/or make commands not found, message catalogs not \n" | |
" rebuilt. Please install gettext and associated tools.") | |
def copyWxDlls(options): | |
if options.no_magic or options.use_syswx: | |
return | |
if isWindows: | |
# Copy the wxWidgets DLLs to the wxPython package folder | |
msw = getMSWSettings(options) | |
cfg = Config() | |
ver = wxversion3_nodot if unstable_series else wxversion2_nodot | |
arch = 'x64' if PYTHON_ARCH == '64bit' else 'x86' | |
dlls = list() | |
if not options.debug or options.both: | |
dlls += glob.glob(os.path.join(msw.dllDir, "wx*%su_*.dll" % ver)) | |
if options.relwithdebug: | |
dlls += glob.glob(os.path.join(msw.dllDir, "wx*%su_*.pdb" % ver)) | |
if options.debug or options.both: | |
dlls += glob.glob(os.path.join(msw.dllDir, "wx*%sud_*.dll" % ver)) | |
dlls += glob.glob(os.path.join(msw.dllDir, "wx*%sud_*.pdb" % ver)) | |
# Also copy the cairo DLLs if needed | |
if options.cairo: | |
cairo_root = os.path.join(phoenixDir(), 'packaging', 'msw-cairo') | |
dlls += glob.glob(os.path.join(cairo_root, arch, 'bin', '*.dll')) | |
# And the webview2 (MS EDGE) DLL | |
wv2_root = os.path.join(phoenixDir(), 'packaging', 'msw-webview2') | |
dlls += glob.glob(os.path.join(wv2_root, arch, '*.dll')) | |
# For Python 3.5 and 3.6 builds we also need to copy some VC14 redist DLLs. | |
# NOTE: Do it for 3.7+ too for now. But when we fully switch over to VS 2017 | |
# this may need to change. See notes in wscript about it. | |
if PYVER in ['3.5', '3.6', '3.7', '3.8', '3.9', '3.10']: | |
redist_dir = os.path.join( | |
phoenixDir(), 'packaging', 'msw-vcredist', | |
arch, 'Microsoft.VC140.CRT', '*.dll') | |
dlls += glob.glob(redist_dir) | |
for dll in dlls: | |
copyIfNewer(dll, posixjoin(phoenixDir(), cfg.PKGDIR, os.path.basename(dll)), verbose=True) | |
elif isDarwin: | |
# Copy the wxWidgets dylibs | |
cfg = Config() | |
wxlibdir = os.path.join(getBuildDir(options), "lib") | |
dlls = glob.glob(wxlibdir + '/*.dylib') | |
for dll in dlls: | |
copyIfNewer(dll, posixjoin(phoenixDir(), cfg.PKGDIR, os.path.basename(dll)), verbose=True) | |
# Now use install_name_tool to change the extension modules to look | |
# in the same folder for the wx libs, instead of the build dir. Also | |
# change the wx libs the same way. | |
macSetLoaderNames(glob.glob(opj(phoenixDir(), cfg.PKGDIR, '*.so')) + | |
glob.glob(opj(phoenixDir(), cfg.PKGDIR, '*.dylib'))) | |
else: | |
# Not Windows and not OSX. For now that means that we'll assume it's wxGTK. | |
cfg = Config() | |
wxlibdir = os.path.join(getBuildDir(options), "lib") | |
dlls = glob.glob(wxlibdir + '/libwx_*.so') | |
dlls += glob.glob(wxlibdir + '/libwx_*.so.[0-9]*') | |
for dll in dlls: | |
copyIfNewer(dll, posixjoin(phoenixDir(), cfg.PKGDIR, os.path.basename(dll)), verbose=True) | |
# If all went well the wxPython extensions should already be looking | |
# for the wxlib's in $ORIGIN, so there is nothing else to do here | |
def cmd_build_py(options, args): | |
cmdTimer = CommandTimer('build_py') | |
waf = getWafCmd() | |
checkCompiler() | |
BUILD_DIR = getBuildDir(options) | |
if options.release: | |
os.environ['WXPYTHON_RELEASE'] = 'yes' | |
if not isWindows: | |
WX_CONFIG = os.environ.get('WX_CONFIG', None) | |
if WX_CONFIG is None: | |
WX_CONFIG = posixjoin(BUILD_DIR, 'wx-config') | |
if options.use_syswx: | |
wxcfg = posixjoin(options.prefix, 'bin', 'wx-config') | |
if options.prefix and os.path.exists(wxcfg): | |
WX_CONFIG = wxcfg | |
else: | |
WX_CONFIG = 'wx-config' # hope it is on the PATH | |
if isDarwin: | |
# WAF does a test build as part of the configuration phase, but the | |
# default compiler and flags it uses are not even close to how we'll | |
# configure it later in the configuration process. At a minimum we need | |
# to add the -isysroot for builds on Darwin, and wxWidgets configure is | |
# adding this to the compiler command so we can fetch it there. | |
def _getWxCompiler(flag, compName, flagName): | |
cmd = "%s %s" % (WX_CONFIG, flag) | |
value = os.popen(cmd, 'r').read()[:-1] | |
cmd = shlex.split(value) | |
compiler = cmd[0] | |
flags = cmd[1:] | |
#os.environ[compName] = compiler # don't reset the compiler here, it will be done later | |
os.environ[flagName] = ' '.join(flags) | |
_getWxCompiler('--cc', 'CC', 'CFLAGS') | |
_getWxCompiler('--cxx', 'CXX', 'CXXFLAGS') | |
wafBuildBase = wafBuildDir = getWafBuildBase() | |
if isWindows: | |
wafBuildDir = posixjoin(wafBuildBase, 'release') | |
build_options = list() | |
if options.verbose: | |
build_options.append('--verbose') | |
if options.debug or (isWindows and options.both): | |
build_options.append("--debug") | |
if isWindows: | |
wafBuildDir = posixjoin(wafBuildBase, 'debug') | |
if isDarwin and options.mac_arch: | |
build_options.append("--mac_arch=%s" % options.mac_arch) | |
if isWindows: | |
if PYTHON_ARCH == '64bit': | |
build_options.append('--msvc_arch=x64') | |
else: | |
build_options.append('--msvc_arch=x86') | |
if not isWindows: | |
build_options.append('--wx_config=%s' % WX_CONFIG) | |
if options.jobs: | |
build_options.append('--jobs=%s' % options.jobs) | |
if options.relwithdebug: | |
build_options.append('--msvc_relwithdebug') | |
if not isDarwin and not isWindows: | |
if options.gtk2: | |
build_options.append('--gtk2') | |
wafBuildDir = posixjoin(wafBuildBase, 'gtk2') | |
if options.gtk3: | |
build_options.append('--gtk3') | |
wafBuildDir = posixjoin(wafBuildBase, 'gtk3') | |
build_options.append('--python="%s"' % PYTHON) | |
build_options.append('--out=%s' % wafBuildDir) # this needs to be the last option | |
if not isWindows and not isDarwin and not options.no_magic and not options.use_syswx: | |
# Using $ORIGIN in the rpath will cause the dynamic linker to look | |
# for shared libraries in a folder relative to the loading binary's | |
# location. Here we'll use just $ORIGIN so it should look in the same | |
# folder as the wxPython extension modules. | |
os.environ['LD_RUN_PATH'] = '$ORIGIN' | |
# Regenerate the _sysconfigdata module? | |
if options.regenerate_sysconfig: | |
do_regenerate_sysconfig() | |
# Run waf to perform the builds | |
pwd = pushDir(phoenixDir()) | |
cmd = '"%s" %s %s configure build %s' % (PYTHON, waf, ' '.join(build_options), options.extra_waf) | |
def _onWafError(): | |
if options.dump_waf_log: | |
logfilename = opj(wafBuildDir, 'config.log') | |
if not os.path.exists(logfilename): | |
msg('WARNING: waf log "{}" not found!'.format(logfilename)) | |
return | |
msg('*-'*40) | |
msg('WAF config log "{}":'.format(logfilename)) | |
with open(logfilename, 'r') as log: | |
msg(log.read()) | |
msg('*-'*40) | |
runcmd(cmd, onError=_onWafError) | |
if isWindows and options.both: | |
build_options.remove('--debug') | |
del build_options[-1] | |
wafBuildDir = posixjoin(wafBuildBase, 'release') | |
build_options.append('--out=%s' % wafBuildDir) | |
cmd = '"%s" %s %s configure build %s' % (PYTHON, waf, ' '.join(build_options), options.extra_waf) | |
runcmd(cmd, onError=_onWafError) | |
copyWxDlls(options) | |
cmd_build_others(options, args) | |
cfg = Config() | |
cfg.build_locale_dir(opj(cfg.PKGDIR, 'locale')) | |
print("\n------------ BUILD FINISHED ------------") | |
print("To use wxPython from the build folder (without installing):") | |
print(" - Set your PYTHONPATH variable to %s." % phoenixDir()) | |
if not isWindows: | |
print(" - You may also need to set your (DY)LD_LIBRARY_PATH to %s/lib,\n or wherever the wxWidgets libs have been installed." % BUILD_DIR) | |
print("") | |
def cmd_build_docker(options, args): | |
# Uses docker images defined by the Dockerfiles under ./docker/build for | |
# building wxPython wheels for a few versions of Linux. | |
# | |
# Requirements: Docker | |
# Internet connection for downloading docker images, if needed | |
# One wxPython source archive in ./dist (created with cmd_sdist) | |
cmdTimer = CommandTimer('build_docker') | |
cfg = Config(noWxConfig=True) | |
cmd = ['inv', 'build-wxpython'] | |
if options.docker_img != 'all': | |
for img in options.docker_img.split(','): | |
cmd.append('-i') | |
cmd.append(img) | |
# Do just the gtk2 builds? | |
if options.gtk2: | |
cmd.extend(['--port', 'gtk2']) | |
# TODO: Instead of the simple options.gtk2 test above, do something like the | |
# following to select both. But currently if gtk2 is selected then | |
# options.gtk3 is explicitly set to False... That needs to be made a little | |
# smarter. | |
# if options.gtk2 and options.gtk3: | |
# cmd.extend(['--port', 'all']) | |
# elif options.gtk2: | |
# cmd.extend(['--port', 'gtk2']) | |
# else: | |
# cmd.extend(['--port', 'gtk3']) | |
# 'none' can be used to skip building and go straight to uploading | |
if options.docker_img != 'none': | |
pwd = pushDir('docker') | |
runcmd(cmd, echoCmd=True) | |
if options.upload: | |
for tag in ['gtk2', 'gtk3']: | |
src = opj(phoenixDir(), 'dist', 'linux', tag) | |
if os.path.isdir(src): | |
uploadTree(src, 'linux', options) | |
def cmd_build_others(options, args): | |
# Build other stuff that may have their own separate build commands instead | |
# of the (ab)normal etg/tweak/generate/sip/compile sequence that the rest of | |
# wxPython uses. So far, it's just the wx.svg package | |
cmdTimer = CommandTimer('build_others') | |
cmd = [PYTHON, 'setup-wxsvg.py', 'build_ext', '--inplace'] | |
if options.verbose: | |
cmd.append('--verbose') | |
runcmd(cmd) | |
def cmd_touch_others(options, args): | |
cmdTimer = CommandTimer('touch_others') | |
pwd = pushDir(phoenixDir()) | |
cfg = Config(noWxConfig=True) | |
pth = pathlib.Path(opj(cfg.PKGDIR, 'svg')) | |
for item in pth.glob('*.pyx'): | |
item.touch() | |
def cmd_clean_others(options, args): | |
cmdTimer = CommandTimer('clean_others') | |
pwd = pushDir(phoenixDir()) | |
cfg = Config(noWxConfig=True) | |
files = [] | |
for wc in ['*.pyd', '*.so']: | |
files += glob.glob(opj(cfg.PKGDIR, 'svg', wc)) | |
delFiles(files) | |
def cmd_install(options, args): | |
cmdTimer = CommandTimer('install') | |
cmd_install_wx(options, args) | |
cmd_install_py(options, args) | |
def cmd_install_wx(options, args): | |
cmdTimer = CommandTimer('install_wx') | |
if isWindows: | |
msg('No wxWidgets installation required on Windows') | |
return | |
if options.use_syswx: | |
msg("use_syswx option specified, skipping wxWidgets install") | |
return | |
if not options.no_magic: | |
msg('Magic is in use, wxWidgets will be installed with wxPython') | |
return | |
# Otherwise, run 'make install' in the wx build folder | |
BUILD_DIR = getBuildDir(options) | |
DESTDIR = '' if not options.destdir else 'DESTDIR=' + options.destdir | |
pwd = pushDir(BUILD_DIR) | |
runcmd("make install %s %s" % (DESTDIR, options.extra_make)) | |
def cmd_install_py(options, args): | |
cmdTimer = CommandTimer('install_py') | |
DESTDIR = '' if not options.destdir else '--root=' + options.destdir | |
VERBOSE = '--verbose' if options.verbose else '' | |
cmd = '"%s" setup.py install --skip-build %s %s %s' % ( | |
PYTHON, DESTDIR, VERBOSE, options.extra_setup) | |
runcmd(cmd) | |
def cmd_build_pdbzip(options, args): | |
if isWindows and options.relwithdebug: | |
cmdTimer = CommandTimer('build_pdbzip') | |
if not os.path.exists('dist'): | |
os.mkdir('dist') | |
cfg = Config() | |
filenames = glob.glob('./wx/*.pdb') | |
if not filenames: | |
msg('No PDB files found in ./wx!') | |
return | |
arch = 'win_amd64' if PYTHON_ARCH == '64bit' else 'win32' | |
pyver = 'py{}'.format(PYSHORTVER) | |
zipname = 'dist/{}-pdb-{}-{}-{}.zip'.format(baseName, cfg.VERSION, pyver, arch) | |
from zipfile import ZipFile, ZIP_DEFLATED | |
with ZipFile(zipname, 'w', ZIP_DEFLATED) as zip: | |
for name in filenames: | |
zip.write(name) | |
msg('PDB zip file created at: {}'.format(zipname)) | |
for name in filenames: | |
os.unlink(name) | |
msg('PDB files removed from ./wx') | |
return zipname | |
def _doSimpleSetupCmd(options, args, setupCmd): | |
cmdTimer = CommandTimer(setupCmd) | |
VERBOSE = '--verbose' if options.verbose else '' | |
cmd = '"%s" setup.py %s --skip-build %s %s' % (PYTHON, setupCmd, VERBOSE, options.extra_setup) | |
runcmd(cmd) | |
def cmd_bdist_egg(options, args): | |
pdbzip = cmd_build_pdbzip(options, args) | |
_doSimpleSetupCmd(options, args, 'bdist_egg') | |
cfg = Config() | |
if options.upload: | |
filemask = "dist/%s-%s-*.egg" % (baseName, cfg.VERSION) | |
filenames = glob.glob(filemask) | |
assert len(filenames) == 1, "Unknown files found:"+repr(filenames) | |
uploadPackage(filenames[0], options) | |
if pdbzip: | |
uploadPackage(pdbzip, options, keep=24, | |
mask='%s-pdb-%s*' % (baseName, cfg.VER_MAJOR)) | |
def cmd_bdist_wheel(options, args): | |
pdbzip = cmd_build_pdbzip(options, args) | |
_doSimpleSetupCmd(options, args, 'bdist_wheel') | |
cfg = Config() | |
if options.upload: | |
filemask = "dist/%s-%s-*.whl" % (baseName, cfg.VERSION) | |
filenames = glob.glob(filemask) | |
print(f'**** filemask: {filemask}') | |
print(f'**** matched: {filenames}') | |
print(f'**** all dist: {glob.glob("dist/*")}') | |
assert len(filenames) == 1, "Unknown files found:"+repr(filenames) | |
uploadPackage(filenames[0], options) | |
if pdbzip: | |
uploadPackage(pdbzip, options, keep=24, | |
mask='%s-pdb-%s*' % (baseName, cfg.VER_MAJOR)) | |
def cmd_bdist_wininst(options, args): | |
pdbzip = cmd_build_pdbzip(options, args) | |
_doSimpleSetupCmd(options, args, 'bdist_wininst') | |
cfg = Config() | |
if options.upload: | |
filemask = "dist/%s-%s-*.exe" % (baseName, cfg.VERSION) | |
filenames = glob.glob(filemask) | |
assert len(filenames) == 1, "Unknown files found:"+repr(filenames) | |
uploadPackage(filenames[0], options) | |
if pdbzip: | |
uploadPackage(pdbzip, options, keep=24, | |
mask='%s-pdb-%s*' % (baseName, cfg.VER_MAJOR)) | |
def cmd_bdist_msi(options, args): | |
cmd_build_pdbzip(options, args) | |
_doSimpleSetupCmd(options, args, 'bdist_msi') | |
def cmd_egg_info(options, args, egg_base=None): | |
cmdTimer = CommandTimer('egg_info') | |
VERBOSE = '--verbose' if options.verbose else '' | |
BASE = '--egg-base '+egg_base if egg_base is not None else '' | |
cmd = '"%s" setup.py egg_info %s %s' % (PYTHON, VERBOSE, BASE) | |
runcmd(cmd) | |
def cmd_clean_wx(options, args): | |
cmdTimer = CommandTimer('clean_wx') | |
if isWindows: | |
if options.both: | |
options.debug = True | |
msw = getMSWSettings(options) | |
deleteIfExists(opj(msw.dllDir, 'msw'+msw.dll_type)) | |
delFiles(glob.glob(opj(msw.dllDir, 'wx*%s%s*' % (wxversion2_nodot, msw.dll_type)))) | |
delFiles(glob.glob(opj(msw.dllDir, 'wx*%s%s*' % (wxversion3_nodot, msw.dll_type)))) | |
if PYTHON_ARCH == '64bit': | |
deleteIfExists(opj(msw.buildDir, 'vc%s_x64_msw%sdll' % (getVisCVersion(), msw.dll_type))) | |
else: | |
deleteIfExists(opj(msw.buildDir, 'vc%s_msw%sdll' % (getVisCVersion(), msw.dll_type))) | |
if options.both: | |
options.debug = False | |
options.both = False | |
cmd_clean_wx(options, args) | |
options.both = True | |
else: | |
BUILD_DIR = getBuildDir(options) | |
deleteIfExists(BUILD_DIR) | |
def cmd_clean_py(options, args): | |
cmdTimer = CommandTimer('clean_py') | |
assert os.getcwd() == phoenixDir() | |
if isWindows and options.both: | |
options.debug = True | |
cfg = Config() | |
deleteIfExists(getWafBuildBase()) | |
files = list() | |
for wc in ['*.py', '*.pyc', '*.so', '*.dylib', '*.pyd', '*.pdb', '*.pi', '*.pyi']: | |
files += glob.glob(opj(cfg.PKGDIR, wc)) | |
if isWindows: | |
msw = getMSWSettings(options) | |
for wc in [ 'wx*' + wxversion2_nodot + msw.dll_type + '*.dll', | |
'wx*' + wxversion3_nodot + msw.dll_type + '*.dll']: | |
files += glob.glob(opj(cfg.PKGDIR, wc)) | |
delFiles(files) | |
# Also remove any remaining DLLs just to make sure. This includes the C++ | |
# runtime DLLs, Cairo, etc. | |
# TODO: Check for specific files, not just *.dll | |
if isWindows: | |
files = glob.glob(opj(cfg.PKGDIR, '*.dll')) | |
delFiles(files) | |
if options.both: | |
options.debug = False | |
options.both = False | |
cmd_clean_py(options, args) | |
options.both = True | |
cmd_clean_others(options, args) | |
def cmd_clean_sphinx(options, args): | |
cmdTimer = CommandTimer('clean_sphinx') | |
assert os.getcwd() == phoenixDir() | |
sphinxDir = opj(phoenixDir(), 'docs', 'sphinx') | |
globs = [ opj(sphinxDir, '*.txt'), | |
opj(sphinxDir, '*.rst'), | |
opj(sphinxDir, '*.inc'), | |
opj(sphinxDir, '*.pkl'), | |
opj(sphinxDir, '*.lst'), | |
opj(sphinxDir, '_templates/gallery.html'), | |
opj(sphinxDir, 'rest_substitutions/snippets/python/*.py'), | |
opj(sphinxDir, 'rest_substitutions/snippets/cpp/*.cpp'), | |
opj(sphinxDir, '_static/images/inheritance/*.*'), | |
] | |
for wc in globs: | |
for f in glob.glob(wc): | |
os.remove(f) | |
dirs = [opj(sphinxDir, 'build'), | |
opj(phoenixDir(), 'docs/html'), | |
] | |
for d in dirs: | |
if os.path.exists(d): | |
shutil.rmtree(d) | |
def cmd_clean(options, args): | |
cmd_clean_wx(options, args) | |
cmd_clean_py(options, args) | |
def cmd_clean_docker(options, args): | |
cmdTimer = CommandTimer('clean_docker') | |
assert os.getcwd() == phoenixDir() | |
d = opj(phoenixDir(), 'dist', 'linux') | |
if os.path.exists(d): | |
shutil.rmtree(d) | |
def cmd_clean_all(options, args): | |
cmd_cleanall(options, args) | |
def cmd_cleanall(options, args): | |
# These take care of all the object, lib, shared lib files created by the | |
# compilation part of build | |
cmd_clean_wx(options, args) | |
cmd_clean_py(options, args) | |
# Clean all the intermediate and generated files from the sphinx command | |
cmd_clean_sphinx(options, args) | |
# Now also scrub out all of the SIP and C++ source files that are | |
# generated by the Phoenix ETG system. | |
cmdTimer = CommandTimer('cleanall') | |
assert os.getcwd() == phoenixDir() | |
files = list() | |
for wc in ['sip/cpp/*.h', 'sip/cpp/*.cpp', 'sip/cpp/*.sbf', 'sip/gen/*.sip']: | |
files += glob.glob(wc) | |
delFiles(files) | |
cmd_clean_docker(options, args) | |
def cmd_buildall(options, args): | |
# (re)build everything | |
cmd_build_wx(options, args) | |
cmd_dox(options, args) | |
cmd_touch(options, args) | |
cmd_etg(options, args) | |
cmd_sip(options, args) | |
cmd_build_py(options, args) | |
cmd_test(options, args) | |
def cmd_sdist(options, args): | |
# Build a source tarball that includes the generated SIP and CPP files. | |
cmdTimer = CommandTimer('sdist') | |
assert os.getcwd() == phoenixDir() | |
cfg = Config() | |
isGit = os.path.exists('.git') | |
if not isGit: | |
msg("Sorry, I don't know what to do in this source tree, no git workspace found.") | |
return | |
# Make a place to export everything to | |
PDEST = 'build/sdist' | |
if not os.path.exists(PDEST): | |
os.makedirs(PDEST) | |
# and a place to put the final tarball | |
if not os.path.exists('dist'): | |
os.mkdir('dist') | |
# recursively export a git archive of this repo and submodules | |
def _archive_submodules(root, dest): | |
msg('Exporting {}...'.format(root)) | |
if not os.path.exists(dest): | |
os.path.makedirs(dest) | |
pwd = pushDir(root) | |
runcmd('git archive HEAD | tar -x -C %s' % dest, echoCmd=False) | |
if os.path.exists('.gitmodules'): | |
with open('.gitmodules', 'rt') as fid: | |
for line in fid: | |
line = line.strip() | |
if line.startswith('path = '): | |
sub = line[7:] | |
_archive_submodules(sub, opj(dest, sub)) | |
_archive_submodules('.', os.path.abspath(PDEST)) | |
# copy Phoenix's generated code into the archive tree | |
msg('Copying generated files...') | |
os.mkdir(posixjoin(PDEST, 'sip', 'siplib')) | |
for srcdir in ['cpp', 'gen', 'siplib']: | |
destdir = posixjoin(PDEST, 'sip', srcdir) | |
for name in glob.glob(posixjoin('sip', srcdir, '*')): | |
try: | |
copyFile(name, destdir) | |
except IsADirectoryError: | |
pass | |
sip_h_dir = posixjoin(cfg.PKGDIR, 'include', 'wxPython') | |
copyFile(posixjoin(sip_h_dir, 'sip.h'), posixjoin(PDEST, sip_h_dir)) | |
for wc in ['*.py', '*.pi', '*.pyi']: | |
destdir = posixjoin(PDEST, cfg.PKGDIR) | |
for name in glob.glob(posixjoin(cfg.PKGDIR, wc)): | |
copyFile(name, destdir) | |
copyFile('demo/version.py', posixjoin(PDEST, 'demo')) | |
# Copy the license files from wxWidgets | |
msg('Copying license files...') | |
updateLicenseFiles(cfg) | |
shutil.copytree('license', opj(PDEST, 'license')) | |
copyFile('LICENSE.txt', PDEST) | |
# Copy the locale message catalogs | |
msg('Copying message catalog files...') | |
cfg.build_locale_dir(opj(cfg.PKGDIR, 'locale')) | |
shutil.copytree(opj(cfg.PKGDIR, 'locale'), opj(PDEST, cfg.PKGDIR, 'locale')) | |
# Also add the waf executable, fetching it first if we don't already have it | |
getWafCmd() | |
copyFile('bin/waf-%s' % wafCurrentVersion, os.path.join(PDEST, 'bin')) | |
# And the REV.txt if there is one | |
if os.path.exists('REV.txt'): | |
copyFile('REV.txt', PDEST) | |
# Copy the Sphinx source files in the docs tree, excluding the html and | |
# sphinx/build folders, if present. | |
shutil.rmtree(opj(PDEST, 'docs'), ignore_errors=True) | |
shutil.copytree('docs', opj(PDEST, 'docs'), | |
ignore=shutil.ignore_patterns('html', 'build', '__pycache__', 'cpp')) | |
# Add some extra stuff to the root folder | |
cmd_egg_info(options, args, egg_base=PDEST) | |
copyFile(opj(PDEST, '{}.egg-info/PKG-INFO'.format(baseName)), | |
opj(PDEST, 'PKG-INFO')) | |
# build the tarball | |
msg('Archiving Phoenix source...') | |
rootname = "%s-%s" % (baseName, cfg.VERSION) | |
tarfilename = posixjoin(phoenixDir(), 'dist', '%s.tar.gz' % rootname) | |
if os.path.exists(tarfilename): | |
os.remove(tarfilename) | |
pwd = pushDir(PDEST) | |
with tarfile.open(name=tarfilename, mode="w:gz") as tarball: | |
for name in glob.glob('*'): | |
tarball.add(name, os.path.join(rootname, name), filter=_setTarItemPerms) | |
msg('Cleaning up...') | |
del pwd | |
shutil.rmtree(PDEST) | |
if options.upload: | |
uploadPackage(tarfilename, options) | |
msg("Source release built at %s" % tarfilename) | |
def cmd_sdist_demo(options, args): | |
# Build a tarball containing the demo and samples | |
cmdTimer = CommandTimer('sdist_demo') | |
assert os.getcwd() == phoenixDir() | |
cfg = Config() | |
isGit = os.path.exists('.git') | |
if not isGit: | |
msg("Sorry, I don't know what to do in this source tree, no git workspace found.") | |
return | |
# make a tree for building up the archive files | |
PDEST = 'build/sdist_demo' | |
if not os.path.exists(PDEST): | |
os.makedirs(PDEST) | |
# and a place to put the final tarball | |
if not os.path.exists('dist'): | |
os.mkdir('dist') | |
# pull out an archive copy of the repo files | |
msg('Exporting Phoenix/demo...') | |
runcmd('git archive HEAD demo | tar -x -C %s' % PDEST, echoCmd=False) | |
msg('Exporting Phoenix/demo/samples...') | |
runcmd('git archive HEAD samples | tar -x -C %s' % PDEST, echoCmd=False) | |
# Add in the README file | |
copyFile('packaging/README-sdist_demo.txt', posixjoin(PDEST, 'README.txt')) | |
copyFile('demo/version.py', posixjoin(PDEST, 'demo')) | |
# build the tarball | |
msg('Archiving Phoenix demo and samples...') | |
rootname = "%s-demo-%s" % (baseName, cfg.VERSION) | |
tarfilename = posixjoin(phoenixDir(), 'dist', '%s.tar.gz' % rootname) | |
if os.path.exists(tarfilename): | |
os.remove(tarfilename) | |
pwd = pushDir(PDEST) | |
with tarfile.open(name=tarfilename, mode="w:gz") as tarball: | |
for name in glob.glob('*'): | |
tarball.add(name, os.path.join(rootname, name), filter=_setTarItemPerms) | |
msg('Cleaning up...') | |
del pwd | |
shutil.rmtree(PDEST) | |
if options.upload: | |
uploadPackage(tarfilename, options, keep=5, | |
mask='%s-demo-%s*' % (baseName, cfg.VER_MAJOR)) | |
msg("demo and samples tarball built at %s" % tarfilename) | |
def cmd_bdist(options, args): | |
# Build a tarball and/or installer that includes all the files needed at | |
# runtime for the current platform and the current version of Python. | |
cmdTimer = CommandTimer('bdist') | |
assert os.getcwd() == phoenixDir() | |
cmd_egg_info(options, args) | |
cfg = Config() | |
dllext = ".so" | |
wxlibdir = os.path.join(getBuildDir(options), "lib") | |
if sys.platform.startswith('darwin'): | |
dllext = ".dylib" | |
platform = sys.platform | |
if isWindows and PYTHON_ARCH == '64bit': | |
platform = 'win64' | |
rootname = "%s-%s-%s-py%s" % (baseName, cfg.VERSION, platform, PYVER) | |
tarfilename = posixjoin(phoenixDir(), 'dist', '%s.tar.gz' % rootname) | |
if not os.path.exists('dist'): | |
os.makedirs('dist') | |
if os.path.exists(tarfilename): | |
os.remove(tarfilename) | |
msg("Archiving Phoenix binaries to %s..." % tarfilename) | |
with tarfile.open(name=tarfilename, mode="w:gz") as tarball: | |
tarball.add('wx', opj(rootname, 'wx'), | |
filter=lambda info: None if '.svn' in info.name \ | |
or info.name.endswith('.pyc') \ | |
or '__pycache__' in info.name else info) | |
tarball.add(eggInfoName, opj(rootname, eggInfoName)) | |
if not isDarwin and not isWindows and not options.no_magic and not options.use_syswx: | |
# If the DLLs are not already in the wx package folder then go fetch | |
# them now. | |
msg("Archiving wxWidgets shared libraries...") | |
dlls = glob.glob(os.path.join(wxlibdir, "*%s" % dllext)) | |
for dll in dlls: | |
tarball.add(dll, os.path.join(rootname, 'wx', os.path.basename(dll))) | |
tarball.add('packaging/README-bdist.txt', os.path.join(rootname, 'README.txt')) | |
if options.upload: | |
uploadPackage(tarfilename, options) | |
msg("Binary release built at %s" % tarfilename) | |
def cmd_setrev(options, args): | |
# Grab the current revision number from the version control system | |
# (if possible) and write it to a file we'll use later for | |
# building the package version number | |
cmdTimer = CommandTimer('setrev') | |
assert os.getcwd() == phoenixDir() | |
if options.release: | |
# Ignore this command for release builds, they are not supposed to | |
# include current VCS revision info in the version number. | |
msg('This is a release build, setting REV.txt skipped') | |
else: | |
svnrev = getVcsRev() | |
with open('REV.txt', 'w') as f: | |
svnrev = '.dev'+svnrev | |
f.write(svnrev) | |
msg('REV.txt set to "%s"' % svnrev) | |
cfg = Config() | |
cfg.resetVersion() | |
msg('cfg.VERSION: %s' % cfg.VERSION) | |
def cmd_setpythonpath(options, args): | |
cmdTimer = CommandTimer('setpythonpath') | |
assert os.getcwd() == phoenixDir() | |
sys.path.insert(0, phoenixDir()) | |
os.environ['PYTHONPATH'] = os.environ.get('PYTHONPATH', '') + os.pathsep + phoenixDir() | |
#--------------------------------------------------------------------------- | |
if __name__ == '__main__': | |
main(sys.argv[1:]) |