| #!/usr/bin/env python3 |
| # |
| # Copyright (C) 2018, 2020 Igalia S.L. |
| # |
| # Redistribution and use in source and binary forms, with or without |
| # modification, are permitted provided that the following conditions are met: |
| # |
| # 1. Redistributions of source code must retain the above copyright notice, this |
| # list of conditions and the following disclaimer. |
| # 2. Redistributions in binary form must reproduce the above copyright notice, |
| # this list of conditions and the following disclaimer in the documentation |
| # and/or other materials provided with the distribution. |
| # |
| # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND |
| # ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED |
| # WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE |
| # DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE LIABLE FOR |
| # ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES |
| # (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; |
| # LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND |
| # ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT |
| # (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS |
| # SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. |
| |
| import argparse |
| import base64 |
| import datetime |
| import hashlib |
| import json |
| import logging |
| import os |
| from pathlib import PurePath |
| import shutil |
| import subprocess |
| import sys |
| import tarfile |
| import tempfile |
| import zipfile |
| |
| top_level_directory = os.path.normpath(os.path.join(os.path.dirname(__file__), '..', '..')) |
| sys.path.insert(0, os.path.join(top_level_directory, 'Tools', 'flatpak')) |
| sys.path.insert(0, os.path.join(top_level_directory, 'Tools', 'jhbuild')) |
| sys.path.insert(0, os.path.join(top_level_directory, 'Tools', 'Scripts', 'webkitpy')) |
| import jhbuildutils |
| import flatpakutils |
| from binary_bundling.ldd import SharedObjectResolver |
| from binary_bundling.bundle import BinaryBundler |
| |
| INSTALL_DEPS_SCRIPT_TEMPLATE = """\ |
| #!/bin/bash |
| set -eu -o pipefail |
| |
| REQUIREDPACKAGES="%(packages_needed)s" |
| |
| if ! which apt-get >/dev/null; then |
| echo "This script only supports apt-get based distributions like Debian or Ubuntu." |
| exit 1 |
| fi |
| |
| # Calling dpkg-query is slow, so call it only once and cache the results |
| TMPCHECKPACKAGES="$(mktemp)" |
| dpkg-query --show --showformat='${binary:Package} ${db:Status-Status}\\n' > "${TMPCHECKPACKAGES}" |
| TOINSTALL="" |
| for PACKAGE in ${REQUIREDPACKAGES}; do |
| if ! grep -qxF "${PACKAGE} installed" "${TMPCHECKPACKAGES}"; then |
| TOINSTALL="${TOINSTALL} ${PACKAGE}" |
| fi |
| done |
| rm -f "${TMPCHECKPACKAGES}" |
| |
| if [[ -z "${TOINSTALL}" ]]; then |
| echo "All required dependencies are already installed" |
| else |
| echo "Need to install the following extra packages: ${TOINSTALL}" |
| [[ ${#} -gt 0 ]] && [[ "${1}" == "--printonly" ]] && exit 0 |
| SUDO="" |
| [[ ${UID} -ne 0 ]] && SUDO="sudo" |
| AUTOINSTALL="" |
| if [[ ${#} -gt 0 ]] && [[ "${1}" == "--autoinstall" ]]; then |
| AUTOINSTALL="-y" |
| export DEBIAN_FRONTEND="noninteractive" |
| [[ ${UID} -ne 0 ]] && SUDO="sudo --preserve-env=DEBIAN_FRONTEND" |
| ${SUDO} apt-get update |
| fi |
| set -x |
| ${SUDO} apt-get install --no-install-recommends ${AUTOINSTALL} ${TOINSTALL} |
| fi |
| """ |
| |
| _log = logging.getLogger(__name__) |
| LOG_MESSAGE = 25 |
| |
| class Archiver(object): |
| |
| def __enter__(self): |
| return self |
| |
| def __exit__(self, type, v, tb): |
| return self._archive.close() |
| |
| class TarArchiver(Archiver): |
| |
| def __init__(self, path): |
| self._archive = tarfile.open(path, 'w:xz') |
| |
| def add_file(self, system_path, zip_path): |
| return self._archive.add(system_path, zip_path) |
| |
| class ZipArchiver(Archiver): |
| |
| def __init__(self, path): |
| self._archive = zipfile.ZipFile(path, 'w', compression=zipfile.ZIP_DEFLATED) |
| |
| def add_file(self, system_path, zip_path): |
| if os.path.islink(system_path): |
| symlink_zip_info = zipfile.ZipInfo(zip_path) |
| symlink_zip_info.create_system = 3 # Unix (for symlink support) |
| symlink_zip_info.external_attr = 0xA1ED0000 # Zip softlink magic number |
| return self._archive.writestr(symlink_zip_info, os.readlink(system_path)) |
| return self._archive.write(system_path, zip_path) |
| |
| class BundleCreator(object): |
| |
| def __init__(self, configuration, platform, bundle_type, syslibs, ldd, should_strip_objects, compression_type, destination = None, revision = None, builder_name = None): |
| self._configuration = configuration |
| self._platform = platform.lower() |
| self._revision = revision |
| self._bundle_binaries = ['jsc', 'MiniBrowser'] if bundle_type == 'all' else [ bundle_type ] |
| self._bundle_type = bundle_type |
| self._buildername = builder_name |
| self._syslibs = syslibs |
| self._shared_object_resolver = SharedObjectResolver(ldd) |
| self._should_strip_objects = should_strip_objects |
| self._compression_type = compression_type |
| self._tmpdir = None |
| self._wrapper_scripts = [] |
| self._port_binary_preffix = 'WebKit' if self._platform == 'gtk' else 'WPE' |
| wk_build_path = os.environ['WEBKIT_OUTPUTDIR'] if 'WEBKIT_OUTPUTDIR' in os.environ else \ |
| os.path.abspath(os.path.join(os.path.dirname(__file__), '..', '..', 'WebKitBuild')) |
| self._buildpath = os.path.join(wk_build_path, self._configuration.capitalize()) |
| |
| default_bundle_name = bundle_type + '_' + self._platform + '_' + self._configuration + '.' + self._compression_type |
| if destination and os.path.isdir(destination): |
| self._bundle_file_path = os.path.join(destination, default_bundle_name) |
| else: |
| self._bundle_file_path = os.path.join(wk_build_path, default_bundle_name) |
| |
| |
| def _create_tempdir(self, basedir = None): |
| if basedir is not None: |
| if not os.path.isdir(basedir): |
| raise ValueError('%s is not a directory' % basedir) |
| return tempfile.mkdtemp(prefix=os.path.join(os.path.abspath(basedir), 'tmp')) |
| return tempfile.mkdtemp() |
| |
| |
| def _run_cmd_and_get_output(self, command): |
| _log.debug("EXEC %s" % command) |
| command_process = subprocess.Popen(command, |
| stdout=subprocess.PIPE, |
| stderr=subprocess.STDOUT, |
| encoding='utf-8') |
| stdout, stderr = command_process.communicate() |
| return command_process.returncode, stdout, stderr |
| |
| def _get_osprettyname(self): |
| with open('/etc/os-release', 'r') as osrelease_handle: |
| for line in osrelease_handle.readlines(): |
| if line.startswith('PRETTY_NAME='): |
| return line.split('=')[1].strip().strip('"') |
| return None |
| |
| |
| def _generate_readme(self): |
| _log.info('Generate README.txt file') |
| readme_file = os.path.join(self._tmpdir, 'README.txt') |
| with open(readme_file, 'w') as readme_handle: |
| readme_handle.write('Bundle details:\n') |
| readme_handle.write(' - WebKit Platform: %s\n' % self._platform.upper()) |
| readme_handle.write(' - Configuration: %s\n' % self._configuration.capitalize()) |
| if self._revision: |
| readme_handle.write(' - WebKit Revision: %s\n' % self._revision) |
| readme_handle.write(' - Bundle type: %s\n' % self._bundle_type) |
| if self._buildername: |
| readme_handle.write(' - Builder name: %s\n' % self._buildername) |
| readme_handle.write(' - Builder date: %s\n' % datetime.datetime.now().isoformat()) |
| readme_handle.write(' - Builder OS: %s\n' % self._get_osprettyname()) |
| if self._syslibs == 'generate-install-script': |
| readme_handle.write('\nRequired dependencies:\n') |
| readme_handle.write(' - This bundle depends on several system libraries that are assumed to be installed.\n') |
| readme_handle.write(' - To ensure all the required libraries are installed execute the script: install-dependencies.sh\n') |
| readme_handle.write(' - You can pass the flag "--autoinstall" to this script to automatically install the dependencies if needed.\n') |
| readme_handle.write('\nRun instructions:\n') |
| scripts = "script" if len(self._wrapper_scripts) == 1 else "scripts" |
| readme_handle.write(' - Execute the wrapper %s in this directory:\n' % scripts) |
| for wrapper_script in self._wrapper_scripts: |
| readme_handle.write(' * %s\n' %wrapper_script) |
| return True |
| |
| |
| def _generate_wrapper_script(self, interpreter, binary_to_wrap): |
| variables = dict() |
| mydir = self._bundler.VAR_MYDIR |
| if os.path.isdir(os.path.join(self._bundler.destination_dir(), 'gio')): |
| gio_var = 'GIO_MODULE_DIR' if self._syslibs == 'bundle-all' else 'GIO_EXTRA_MODULES' |
| variables[gio_var] = "${%s}/gio" % mydir |
| |
| if os.path.isdir(os.path.join(self._bundler.destination_dir(), 'gst')): |
| gst_var = 'GST_PLUGIN_SYSTEM_PATH_1_0' if self._syslibs == 'bundle-all' else 'GST_PLUGIN_PATH_1_0' |
| variables[gst_var] = "${%s}/gst" % mydir |
| variables['GST_REGISTRY_1_0'] = "${%s}/gst/gstreamer-1.0.registry" % mydir |
| if binary_to_wrap != "jsc": |
| variables['WEBKIT_EXEC_PATH'] = "${%s}/bin" % mydir |
| variables['WEBKIT_INJECTED_BUNDLE_PATH'] = "${%s}/lib" % mydir |
| if self._syslibs == 'bundle-all': |
| if binary_to_wrap != "jsc": |
| for var in ['WEB_PROCESS_CMD_PREFIX', |
| 'PLUGIN_PROCESS_CMD_PREFIX', |
| 'NETWORK_PROCESS_CMD_PREFIX', |
| 'GPU_PROCESS_CMD_PREFIX']: |
| variables[var] = "${%s}" % self._bundler.VAR_INTERPRETER |
| else: |
| interpreter = None |
| self._bundler.generate_wrapper_script(interpreter, binary_to_wrap, variables) |
| self._wrapper_scripts.append(binary_to_wrap) |
| |
| def _generate_install_deps_script(self, system_packages_needed): |
| if not system_packages_needed: |
| return |
| if 'MiniBrowser' in self._bundle_binaries: |
| # Add some extra packages that are needed but the script can't automatically detect |
| for extra_needed_pkg in ['ca-certificates', 'shared-mime-info']: |
| system_packages_needed.add(extra_needed_pkg) |
| # And remove some packages that may be detected due to indirect deps (gstreamer/gio) but are really not needed |
| for not_needed_pkg in ['dconf-gsettings-backend', 'gvfs', 'pitivi', 'gstreamer1.0-convolver-pulseeffects', 'gstreamer1.0-x', |
| 'gstreamer1.0-adapter-pulseeffects', 'gstreamer1.0-autogain-pulseeffects', 'gstreamer1.0-alsa', |
| 'gstreamer1.0-clutter-3.0', 'gstreamer1.0-crystalizer-pulseeffects', 'gstreamer1.0-gtk3', 'gstreamer1.0-nice']: |
| if not_needed_pkg in system_packages_needed: |
| system_packages_needed.remove(not_needed_pkg) |
| # Sometimes the package is identified with an arch suffix, but not always |
| not_needed_pkg_arch = not_needed_pkg + ':amd64' |
| if not_needed_pkg_arch in system_packages_needed: |
| system_packages_needed.remove(not_needed_pkg_arch) |
| installdeps_file = os.path.join(self._tmpdir, 'install-dependencies.sh') |
| with open(installdeps_file, 'w') as installdeps_handle: |
| installdeps_handle.write(INSTALL_DEPS_SCRIPT_TEMPLATE % {'packages_needed' : ' '.join(system_packages_needed)} ) |
| os.chmod(installdeps_file, 0o755) |
| |
| def _remove_tempdir(self): |
| if not self._tmpdir: |
| return |
| if os.path.isdir(self._tmpdir): |
| shutil.rmtree(self._tmpdir) |
| |
| def create(self): |
| self._tmpdir = self._create_tempdir(self._buildpath) |
| self._bundler = BinaryBundler(self._tmpdir, self._should_strip_objects) |
| |
| if os.path.isfile(self._bundle_file_path): |
| _log.info('Removing previous bundle %s' % self._bundle_file_path) |
| os.remove(self._bundle_file_path) |
| |
| for bundle_binary in self._bundle_binaries: |
| self._create_bundle(bundle_binary) |
| self._generate_readme() |
| |
| if self._compression_type == 'zip': |
| archiver = ZipArchiver(self._bundle_file_path) |
| elif self._compression_type == 'tar.xz': |
| archiver = TarArchiver(self._bundle_file_path) |
| else: |
| raise NotImplementedError('Support for compression type %s not implemented' % self._compression_type) |
| self._create_archive(archiver) |
| self._remove_tempdir() |
| if not os.path.isfile(self._bundle_file_path): |
| raise RuntimeError('Unable to create the file %s' % self._bundle_file_path) |
| _log.log(LOG_MESSAGE, 'Bundle file created at: %s' % self._bundle_file_path) |
| return self._bundle_file_path |
| |
| |
| def _get_webkit_binaries(self): |
| webkit_binaries = [] |
| bin_dir = os.path.join(self._buildpath, 'bin') |
| for entry in os.listdir(bin_dir): |
| if entry.startswith(self._port_binary_preffix) and (entry.endswith('Process') or entry.endswith('Driver')): |
| binary = os.path.join(bin_dir, entry) |
| if os.path.isfile(binary) and os.access(binary, os.X_OK): |
| webkit_binaries.append(binary) |
| if len(webkit_binaries) < 2: |
| raise RuntimeError('Could not find required WebKit Process binaries. Check if you are passing the right platform value.') |
| return webkit_binaries |
| |
| |
| def _get_webkit_bundlelib(self): |
| lib_dir = os.path.join(self._buildpath, 'lib') |
| bundle_lib = None |
| for entry in os.listdir(lib_dir): |
| if entry.endswith('.so') and 'injectedbundle' in entry.lower() and 'test' not in entry.lower(): |
| assert(bundle_lib == None) |
| bundle_lib = os.path.join(lib_dir, entry) |
| break |
| assert(bundle_lib) |
| return bundle_lib |
| |
| |
| def _create_archive(self, archiver): |
| _log.info('Create archive') |
| with archiver: |
| for dirname, subdirs, files in os.walk(self._tmpdir): |
| for filename in files: |
| system_file_path = os.path.join(dirname, filename) |
| zip_file_path = system_file_path.replace(self._tmpdir, '', 1).lstrip('/') |
| archiver.add_file(system_file_path, zip_file_path) |
| |
| |
| def _get_system_package_name(self, object): |
| if not shutil.which('dpkg'): |
| raise RuntimeError('Adding system dependencies only supported for dpkg-based distros. Try passing --syslibs=bundle-all') |
| retcode, stdout, stderr = self._run_cmd_and_get_output(['dpkg', '-S', object]) |
| if retcode != 0: |
| # Give a second-try with the realpath of the object. |
| # This fixes issue on Ubuntu-20.04 that has a /lib symlink to /usr/lib |
| # and objects point to /lib, but dpkg only recognizes the files on /usr/lib |
| object_realpath = os.path.realpath(object) |
| if object_realpath != object: |
| retcode, stdout, stderr = self._run_cmd_and_get_output(['dpkg', '-S', object_realpath]) |
| if retcode != 0: |
| # Package not found |
| return None |
| package = stdout.split(' ')[0].rstrip(':') |
| _log.info('Add dependency on system package [%s]: %s' %(package, object)) |
| return package |
| |
| |
| def _get_gio_modules(self): |
| gio_modules = [] |
| retcode, stdout, stderr = self._run_cmd_and_get_output(['pkg-config', '--variable=giomoduledir', 'gio-2.0']) |
| if retcode != 0: |
| raise RuntimeError('The pkg-config command returned status %d' % retcode) |
| gio_module_dir = stdout.strip() |
| if not os.path.isdir(gio_module_dir): |
| raise RuntimeError('The pkg-config entry for giomoduledir is not a directory: %s' % gio_module_dir) |
| for entry in os.listdir(gio_module_dir): |
| if entry.endswith('.so'): |
| gio_modules.append(os.path.join(gio_module_dir, entry)) |
| return gio_modules |
| |
| def _get_gstreamer_modules(self): |
| gstreamer_plugins = [] |
| retcode, stdout, stderr = self._run_cmd_and_get_output(['pkg-config', '--variable=pluginsdir', 'gstreamer-1.0']) |
| if retcode != 0: |
| raise RuntimeError('The pkg-config command returned status %d' % retcode) |
| gstramer_plugins_dir = stdout.strip() |
| if not os.path.isdir(gstramer_plugins_dir): |
| raise RuntimeError('The pkg-config entry for pluginsdir is not a directory: %s' % gstramer_plugins_dir) |
| for entry in os.listdir(gstramer_plugins_dir): |
| if entry.endswith('.so'): |
| gstreamer_plugins.append(os.path.join(gstramer_plugins_dir, entry)) |
| return gstreamer_plugins |
| |
| |
| def _add_object_or_get_sysdep(self, object, object_type): |
| provided_by_system_package = None |
| if self._syslibs == 'bundle-all': |
| self._bundler.copy_and_remove_rpath(object, type=object_type) |
| else: |
| provided_by_system_package = self._get_system_package_name(object) |
| if not provided_by_system_package: |
| self._bundler.copy_and_remove_rpath(object, type=object_type) |
| return provided_by_system_package |
| |
| def _ensure_wpe_backend_symlink(self): |
| # WPE/WPERenderer dlopens this library without a version suffix, |
| # so we need to ensure there is a proper symlink |
| bundle_lib_dir = os.path.join(self._tmpdir, 'lib') |
| wpe_backend_soname = 'libWPEBackend-fdo-1.0.so' |
| previous_dir = os.getcwd() |
| for entry in os.listdir(bundle_lib_dir): |
| if entry.startswith(wpe_backend_soname + '.'): |
| os.chdir(bundle_lib_dir) |
| if not os.path.exists(wpe_backend_soname): |
| os.symlink(entry, wpe_backend_soname) |
| os.chdir(previous_dir) |
| break |
| |
| def _create_bundle(self, bundle_binary): |
| main_binary_path = os.path.join(self._buildpath, 'bin', bundle_binary) |
| if not os.path.isfile(main_binary_path) or not os.access(main_binary_path, os.X_OK): |
| raise ValueError('Cannot find binary for %s at %s' % (bundle_binary, main_binary_path) ) |
| |
| copied_interpreter = None |
| gio_modules = [] |
| gstreamer_modules = [] |
| libraries_checked = set() |
| system_packages_needed = set() |
| objects_to_copy = [ main_binary_path ] |
| if bundle_binary == 'MiniBrowser': |
| gio_modules = self._get_gio_modules() |
| gstreamer_modules = self._get_gstreamer_modules() |
| objects_to_copy.extend(self._get_webkit_binaries()) |
| objects_to_copy.append(self._get_webkit_bundlelib()) |
| objects_to_copy.extend(gio_modules) |
| objects_to_copy.extend(gstreamer_modules) |
| for object in objects_to_copy: |
| system_package = None |
| if object in gio_modules: |
| system_package = self._add_object_or_get_sysdep(object, 'gio') |
| if system_package: |
| system_packages_needed.add(system_package) |
| elif object in gstreamer_modules: |
| system_package = self._add_object_or_get_sysdep(object, 'gst') |
| if system_package: |
| system_packages_needed.add(system_package) |
| elif object.endswith('.so'): |
| self._bundler.copy_and_remove_rpath(object, type='lib') |
| else: |
| self._bundler.copy_and_remove_rpath(object, type='bin') |
| # There is no need to examine the libraries linked with objects coming from a system package, |
| # because system packages already declare dependencies between them. |
| # However, if we are running with self._syslibs == 'bundle-all' then system_package will be None, |
| # and everything will be examined and bundled as we don't account for system packages in that case. |
| if not system_package: |
| libraries, interpreter = self._shared_object_resolver.get_libs_and_interpreter(object) |
| if interpreter is None: |
| raise RuntimeError("Could not determine interpreter for binary %s" % object) |
| if copied_interpreter is None: |
| if self._syslibs == 'bundle-all': |
| self._bundler.copy_and_remove_rpath(interpreter, type='interpreter') |
| copied_interpreter = interpreter |
| elif copied_interpreter != interpreter: |
| raise RuntimeError('Detected binaries with different interpreters: %s != %s' %(copied_interpreter, interpreter)) |
| # FIXME: for --syslibs=bundle-all we would have to copy the libnss_*so* libraries which are dlopen'ed <https://bugs.debian.org/203014> |
| # Also we should include config files like fontconfig stuff or support files like icons. |
| for library in libraries: |
| if library in libraries_checked: |
| _log.debug('Skip already checked [lib]: %s' % library) |
| continue |
| libraries_checked.add(library) |
| system_package = self._add_object_or_get_sysdep(library, 'lib') |
| if system_package: |
| system_packages_needed.add(system_package) |
| |
| self._ensure_wpe_backend_symlink() |
| self._generate_wrapper_script(interpreter, bundle_binary) |
| if bundle_binary == "MiniBrowser": |
| self._generate_wrapper_script(interpreter, self._port_binary_preffix + 'WebDriver') |
| self._generate_install_deps_script(system_packages_needed) |
| |
| |
| class BundleUploader(object): |
| |
| def __init__(self, bundle_file_path, remote_config_file, bundle_type, platform, configuration, compression_type, revision, log_level): |
| self._bundle_file_path = bundle_file_path |
| self._remote_config_file = remote_config_file |
| self._configuration = configuration |
| self._revision = revision |
| self._bundle_type = bundle_type |
| self._platform = platform |
| self._compression_type = compression_type |
| self._sftp_quiet = log_level == 'quiet' or log_level == 'minimal' |
| if not os.path.isfile(self._remote_config_file): |
| raise ValueError('Can not find remote config file for upload at path %s' % self._remote_config_file) |
| |
| def _sha256sum(self, file): |
| hash = hashlib.sha256() |
| with open(file, 'rb') as f: |
| for chunk in iter(lambda: f.read(4096), b''): |
| hash.update(chunk) |
| return hash.hexdigest() |
| |
| def _get_osidversion(self): |
| with open('/etc/os-release', 'r') as osrelease_handle: |
| for line in osrelease_handle.readlines(): |
| if line.startswith('ID='): |
| os_id = line.split('=')[1].strip().strip('"') |
| if line.startswith('VERSION_ID='): |
| version_id = line.split('=')[1].strip().strip('"') |
| assert(os_id) |
| assert(version_id) |
| osidversion = os_id + '-' + version_id |
| assert(' ' not in osidversion) |
| assert(len(osidversion) > 3) |
| return osidversion |
| |
| # The expected format for --remote-config-file is something like: |
| # { |
| # "servername": "webkitgtk.org", |
| # "serveraddress": "webkitgtk.intranet-address.local", |
| # "serverport": "23", |
| # "username": "upload-bot-64", |
| # "baseurl": "https://webkitgtk.org/built-products", |
| # "remotepath" : "x86_64/nightly/%(bundletype)s/%(distro_id_ver)s/%(bundletype)s_%(platform)s_%(configuration)s_r%(version)s.%(compression_type)s", |
| # "sshkey": "output of the priv key in base64. E.g. cat ~/.ssh/id_rsa|base64 -w0" |
| # } |
| def upload(self): |
| remote_data = json.load(open(self._remote_config_file)) |
| remote_file_bundle_path = remote_data['remotepath'] % { 'bundletype' : self._bundle_type, |
| 'configuration' : self._configuration, |
| 'compression_type' : self._compression_type, |
| 'distro_id_ver' : self._get_osidversion().capitalize(), |
| 'platform' : self._platform, |
| 'version' : self._revision } |
| with tempfile.NamedTemporaryFile(mode='w+b') as sshkeyfile, tempfile.NamedTemporaryFile(mode='w+') as hashcheckfile, \ |
| tempfile.NamedTemporaryFile(mode='w+') as lastisfile, tempfile.NamedTemporaryFile(mode='w+') as uploadinstructionsfile: |
| |
| # In theory NamedTemporaryFile() is already created 0600. But it don't hurts ensuring this again here. |
| os.chmod(sshkeyfile.name, 0o600) |
| sshkeyfile.write(base64.b64decode(remote_data['sshkey'])) |
| sshkeyfile.flush() |
| # Generate and upload also a sha256 hash |
| hashforbundle = self._sha256sum(self._bundle_file_path) |
| os.chmod(hashcheckfile.name, 0o644) |
| hashcheckfile.write('%s %s\n' % (hashforbundle, os.path.basename(remote_file_bundle_path))) |
| hashcheckfile.flush() |
| # A LAST-IS file for convenience |
| os.chmod(lastisfile.name, 0o644) |
| lastisfile.write('%s\n' % os.path.basename(remote_file_bundle_path)) |
| lastisfile.flush() |
| # SFTP upload instructions file |
| uploadinstructionsfile.write('progress\n') |
| uploadinstructionsfile.write('put %s %s\n' % (self._bundle_file_path, remote_file_bundle_path)) |
| remote_file_bundle_path_no_ext, _ = os.path.splitext(remote_file_bundle_path) |
| uploadinstructionsfile.write('put %s %s\n' % (hashcheckfile.name, remote_file_bundle_path_no_ext + '.sha256sum')) |
| uploadinstructionsfile.write('put %s %s\n' % (lastisfile.name, os.path.join(os.path.dirname(remote_file_bundle_path), 'LAST-IS'))) |
| uploadinstructionsfile.write('quit\n') |
| uploadinstructionsfile.flush() |
| # The idea of this is to ensure scp doesn't ask any question (not even on the first run). |
| # This should be secure enough according to https://www.gremwell.com/ssh-mitm-public-key-authentication |
| sftpCommand = ['sftp', |
| '-o', 'StrictHostKeyChecking=no', |
| '-o', 'UserKnownHostsFile=/dev/null', |
| '-o', 'LogLevel=ERROR', |
| '-P', remote_data['serverport'], |
| '-i', sshkeyfile.name, |
| '-b', uploadinstructionsfile.name, |
| '%s@%s' % (remote_data['username'], remote_data['serveraddress'])] |
| _log.info('Uploading bundle to %s as %s with sha256 hash %s' % (remote_data['servername'], remote_file_bundle_path, hashforbundle)) |
| sftp_out = subprocess.DEVNULL if self._sftp_quiet else sys.stdout |
| if subprocess.call(sftpCommand, stdout=sftp_out, stderr=sftp_out) != 0: |
| raise RuntimeError('The sftp command returned non-zero status') |
| |
| _log.log(LOG_MESSAGE, 'Done: archive sucesfully uploaded to %s/%s' % (remote_data['baseurl'], remote_file_bundle_path)) |
| return 0 |
| |
| |
| def configure_logging(selected_log_level='info'): |
| |
| class LogHandler(logging.StreamHandler): |
| def __init__(self, stream): |
| super().__init__(stream) |
| |
| def format(self, record): |
| if record.levelno > LOG_MESSAGE: |
| return '%s: %s' % (record.levelname, record.getMessage()) |
| return record.getMessage() |
| |
| logging.addLevelName(LOG_MESSAGE, 'MESSAGE') |
| if selected_log_level == 'debug': |
| log_level = logging.DEBUG |
| elif selected_log_level == 'info': |
| log_level = logging.INFO |
| elif selected_log_level == 'quiet': |
| log_level = logging.NOTSET |
| elif selected_log_level == 'minimal': |
| log_level = logging.getLevelName(LOG_MESSAGE) |
| |
| handler = LogHandler(sys.stdout) |
| logger = logging.getLogger() |
| logger.addHandler(handler) |
| logger.setLevel(log_level) |
| return handler |
| |
| |
| def main(): |
| parser = argparse.ArgumentParser('usage: %prog [options]') |
| configuration = parser.add_mutually_exclusive_group(required=True) |
| configuration.add_argument('--debug', action='store_const', const='debug', dest='configuration') |
| configuration.add_argument('--release', action='store_const', const='release', dest='configuration') |
| parser.add_argument('--platform', dest='platform', choices=['gtk', 'wpe'], required=True, |
| help='The WebKit port to generate the bundle') |
| parser.add_argument('--bundle', dest='bundle_binary', choices=['jsc', 'MiniBrowser', 'all'], required=True, |
| help='Select what main binary should be included in the bundle') |
| parser.add_argument('--syslibs', dest='syslibs', choices=['bundle-all', 'generate-install-script'], default='generate-install-script', |
| help='If value is "bundle-all", the bundle will include _all_ the system libraries instead of a install-dependencies script.\n' |
| 'If value is "generate-install-script", the system libraries will not be bundled and a install-dependencies script will be generated for this distribution.') |
| parser.add_argument('--ldd', dest='ldd', default='ldd', help='Use alternative ldd (useful for non-native binaries') |
| parser.add_argument('--compression', dest='compression', choices=['zip', 'tar.xz'], default='zip') |
| parser.add_argument('--destination', action='store', dest='destination', |
| help='Optional path were to store the bundle') |
| parser.add_argument('--no-strip', action='store_true', dest='no_strip', |
| help='Do not strip the binaries and libraries inside the bundle') |
| parser.add_argument('--log-level', dest='log_level', choices=['quiet', 'minimal', 'info', 'debug'], default='info') |
| parser.add_argument('--revision', action='store', dest='webkit_version') |
| parser.add_argument('--builder-name', action='store', dest='builder_name') |
| parser.add_argument('--remote-config-file', action='store', dest='remote_config_file', |
| help='Optional configuration file with the configuration needed to upload the generated the bundle to a remote server via sftp/ssh.') |
| options = parser.parse_args() |
| |
| flatpakutils.run_in_sandbox_if_available([sys.argv[0], '--flatpak-' + options.platform] + sys.argv[1:]) |
| if not flatpakutils.is_sandboxed(): |
| jhbuildutils.enter_jhbuild_environment_if_available(options.platform) |
| |
| configure_logging(options.log_level) |
| bundle_creator = BundleCreator(options.configuration, options.platform, options.bundle_binary, options.syslibs, options.ldd, |
| not options.no_strip, options.compression, options.destination, options.webkit_version, options.builder_name) |
| bundle_file_path = bundle_creator.create() |
| |
| if options.remote_config_file is not None: |
| bundle_uploader = BundleUploader(bundle_file_path, options.remote_config_file, options.bundle_binary, options.platform, |
| options.configuration, options.compression, options.webkit_version, options.log_level) |
| return bundle_uploader.upload() |
| return 0 |
| |
| |
| if __name__ == '__main__': |
| sys.exit(main()) |