2017-08-13 03:24:00 +02:00
|
|
|
#!/usr/bin/env python
|
|
|
|
"""
|
|
|
|
Tool for packaging Python apps for Android
|
|
|
|
==========================================
|
|
|
|
|
|
|
|
This module defines the entry point for command line and programmatic use.
|
|
|
|
"""
|
|
|
|
|
|
|
|
from __future__ import print_function
|
2017-12-21 08:24:31 +01:00
|
|
|
from pythonforandroid import __version__
|
2017-08-13 03:24:00 +02:00
|
|
|
|
|
|
|
def check_python_dependencies():
|
|
|
|
# Check if the Python requirements are installed. This appears
|
|
|
|
# before the imports because otherwise they're imported elsewhere.
|
|
|
|
|
|
|
|
# Using the ok check instead of failing immediately so that all
|
|
|
|
# errors are printed at once
|
|
|
|
|
|
|
|
from distutils.version import LooseVersion
|
|
|
|
from importlib import import_module
|
|
|
|
import sys
|
|
|
|
|
|
|
|
ok = True
|
|
|
|
|
|
|
|
modules = [('colorama', '0.3.3'), 'appdirs', ('sh', '1.10'), 'jinja2',
|
|
|
|
'six']
|
|
|
|
|
|
|
|
for module in modules:
|
|
|
|
if isinstance(module, tuple):
|
|
|
|
module, version = module
|
|
|
|
else:
|
|
|
|
version = None
|
|
|
|
|
|
|
|
try:
|
|
|
|
import_module(module)
|
|
|
|
except ImportError:
|
|
|
|
if version is None:
|
|
|
|
print('ERROR: The {} Python module could not be found, please '
|
|
|
|
'install it.'.format(module))
|
|
|
|
ok = False
|
|
|
|
else:
|
|
|
|
print('ERROR: The {} Python module could not be found, '
|
|
|
|
'please install version {} or higher'.format(
|
|
|
|
module, version))
|
|
|
|
ok = False
|
|
|
|
else:
|
|
|
|
if version is None:
|
|
|
|
continue
|
|
|
|
try:
|
|
|
|
cur_ver = sys.modules[module].__version__
|
|
|
|
except AttributeError: # this is sometimes not available
|
|
|
|
continue
|
|
|
|
if LooseVersion(cur_ver) < LooseVersion(version):
|
|
|
|
print('ERROR: {} version is {}, but python-for-android needs '
|
|
|
|
'at least {}.'.format(module, cur_ver, version))
|
|
|
|
ok = False
|
|
|
|
|
|
|
|
if not ok:
|
|
|
|
print('python-for-android is exiting due to the errors above.')
|
|
|
|
exit(1)
|
|
|
|
|
|
|
|
|
|
|
|
check_python_dependencies()
|
|
|
|
|
|
|
|
|
|
|
|
import sys
|
|
|
|
from sys import platform
|
2017-12-21 08:24:31 +01:00
|
|
|
from os.path import (join, dirname, realpath, exists, expanduser, basename)
|
2017-08-13 03:24:00 +02:00
|
|
|
import os
|
|
|
|
import glob
|
|
|
|
import shutil
|
|
|
|
import re
|
|
|
|
import imp
|
|
|
|
import shlex
|
|
|
|
from functools import wraps
|
|
|
|
|
|
|
|
import argparse
|
|
|
|
import sh
|
|
|
|
import imp
|
|
|
|
from appdirs import user_data_dir
|
|
|
|
import logging
|
2017-12-21 08:24:31 +01:00
|
|
|
from distutils.version import LooseVersion
|
2017-08-13 03:24:00 +02:00
|
|
|
|
|
|
|
from pythonforandroid.recipe import (Recipe, PythonRecipe, CythonRecipe,
|
|
|
|
CompiledComponentsPythonRecipe,
|
|
|
|
BootstrapNDKRecipe, NDKRecipe)
|
|
|
|
from pythonforandroid.archs import (ArchARM, ArchARMv7_a, Archx86)
|
|
|
|
from pythonforandroid.logger import (logger, info, warning, setup_color,
|
|
|
|
Out_Style, Out_Fore, Err_Style, Err_Fore,
|
|
|
|
info_notify, info_main, shprint, error)
|
|
|
|
from pythonforandroid.util import current_directory, ensure_dir
|
|
|
|
from pythonforandroid.bootstrap import Bootstrap
|
|
|
|
from pythonforandroid.distribution import Distribution, pretty_log_dists
|
|
|
|
from pythonforandroid.graph import get_recipe_order_and_bootstrap
|
|
|
|
from pythonforandroid.build import Context, build_recipes
|
|
|
|
|
|
|
|
user_dir = dirname(realpath(os.path.curdir))
|
|
|
|
toolchain_dir = dirname(__file__)
|
|
|
|
sys.path.insert(0, join(toolchain_dir, "tools", "external"))
|
|
|
|
|
|
|
|
|
|
|
|
def add_boolean_option(parser, names, no_names=None,
|
|
|
|
default=True, dest=None, description=None):
|
|
|
|
group = parser.add_argument_group(description=description)
|
|
|
|
if not isinstance(names, (list, tuple)):
|
|
|
|
names = [names]
|
|
|
|
if dest is None:
|
|
|
|
dest = names[0].strip("-").replace("-", "_")
|
|
|
|
|
|
|
|
def add_dashes(x):
|
|
|
|
return x if x.startswith("-") else "--"+x
|
|
|
|
|
|
|
|
opts = [add_dashes(x) for x in names]
|
|
|
|
group.add_argument(
|
|
|
|
*opts, help=("(this is the default)" if default else None),
|
|
|
|
dest=dest, action='store_true')
|
|
|
|
if no_names is None:
|
|
|
|
def add_no(x):
|
|
|
|
x = x.lstrip("-")
|
|
|
|
return ("no_"+x) if "_" in x else ("no-"+x)
|
|
|
|
no_names = [add_no(x) for x in names]
|
|
|
|
opts = [add_dashes(x) for x in no_names]
|
|
|
|
group.add_argument(
|
|
|
|
*opts, help=(None if default else "(this is the default)"),
|
|
|
|
dest=dest, action='store_false')
|
|
|
|
parser.set_defaults(**{dest: default})
|
|
|
|
|
|
|
|
|
|
|
|
def require_prebuilt_dist(func):
|
|
|
|
'''Decorator for ToolchainCL methods. If present, the method will
|
|
|
|
automatically make sure a dist has been built before continuing
|
|
|
|
or, if no dists are present or can be obtained, will raise an
|
|
|
|
error.
|
|
|
|
|
|
|
|
'''
|
|
|
|
|
|
|
|
@wraps(func)
|
|
|
|
def wrapper_func(self, args):
|
|
|
|
ctx = self.ctx
|
|
|
|
ctx.set_archs(self._archs)
|
|
|
|
ctx.prepare_build_environment(user_sdk_dir=self.sdk_dir,
|
|
|
|
user_ndk_dir=self.ndk_dir,
|
|
|
|
user_android_api=self.android_api,
|
2018-04-11 17:30:16 +02:00
|
|
|
user_android_min_api=self.android_min_api,
|
2017-08-13 03:24:00 +02:00
|
|
|
user_ndk_ver=self.ndk_version)
|
|
|
|
dist = self._dist
|
|
|
|
if dist.needs_build:
|
|
|
|
info_notify('No dist exists that meets your requirements, '
|
|
|
|
'so one will be built.')
|
|
|
|
build_dist_from_args(ctx, dist, args)
|
|
|
|
func(self, args)
|
|
|
|
return wrapper_func
|
|
|
|
|
|
|
|
|
|
|
|
def dist_from_args(ctx, args):
|
|
|
|
'''Parses out any distribution-related arguments, and uses them to
|
|
|
|
obtain a Distribution class instance for the build.
|
|
|
|
'''
|
|
|
|
return Distribution.get_distribution(
|
|
|
|
ctx,
|
|
|
|
name=args.dist_name,
|
|
|
|
recipes=split_argument_list(args.requirements),
|
|
|
|
require_perfect_match=args.require_perfect_match)
|
|
|
|
|
|
|
|
|
|
|
|
def build_dist_from_args(ctx, dist, args):
|
|
|
|
'''Parses out any bootstrap related arguments, and uses them to build
|
|
|
|
a dist.'''
|
|
|
|
bs = Bootstrap.get_bootstrap(args.bootstrap, ctx)
|
|
|
|
build_order, python_modules, bs \
|
|
|
|
= get_recipe_order_and_bootstrap(ctx, dist.recipes, bs)
|
|
|
|
ctx.recipe_build_order = build_order
|
|
|
|
ctx.python_modules = python_modules
|
|
|
|
|
|
|
|
if python_modules and hasattr(sys, 'real_prefix'):
|
|
|
|
error('virtualenv is needed to install pure-Python modules, but')
|
|
|
|
error('virtualenv does not support nesting, and you are running')
|
|
|
|
error('python-for-android in one. Please run p4a outside of a')
|
|
|
|
error('virtualenv instead.')
|
|
|
|
exit(1)
|
|
|
|
|
|
|
|
info('The selected bootstrap is {}'.format(bs.name))
|
|
|
|
info_main('# Creating dist with {} bootstrap'.format(bs.name))
|
|
|
|
bs.distribution = dist
|
|
|
|
info_notify('Dist will have name {} and recipes ({})'.format(
|
|
|
|
dist.name, ', '.join(dist.recipes)))
|
2017-12-21 08:24:31 +01:00
|
|
|
info('Dist will also contain modules ({}) installed from pip'.format(
|
|
|
|
', '.join(ctx.python_modules)))
|
2017-08-13 03:24:00 +02:00
|
|
|
|
|
|
|
ctx.dist_name = bs.distribution.name
|
|
|
|
ctx.prepare_bootstrap(bs)
|
|
|
|
ctx.prepare_dist(ctx.dist_name)
|
|
|
|
|
|
|
|
build_recipes(build_order, python_modules, ctx)
|
|
|
|
|
|
|
|
ctx.bootstrap.run_distribute()
|
|
|
|
|
|
|
|
info_main('# Your distribution was created successfully, exiting.')
|
|
|
|
info('Dist can be found at (for now) {}'
|
|
|
|
.format(join(ctx.dist_dir, ctx.dist_name)))
|
|
|
|
|
|
|
|
|
|
|
|
def split_argument_list(l):
|
|
|
|
if not len(l):
|
|
|
|
return []
|
|
|
|
return re.split(r'[ ,]+', l)
|
|
|
|
|
|
|
|
class NoAbbrevParser(argparse.ArgumentParser):
|
|
|
|
'''We want to disable argument abbreviation so as not to interfere
|
|
|
|
with passing through arguments to build.py, but in python2 argparse
|
|
|
|
doesn't have this option.
|
|
|
|
|
|
|
|
This subclass alternative is follows the suggestion at
|
|
|
|
https://bugs.python.org/issue14910.
|
|
|
|
'''
|
|
|
|
def _get_option_tuples(self, option_string):
|
|
|
|
return []
|
|
|
|
|
|
|
|
class ToolchainCL(object):
|
|
|
|
|
|
|
|
def __init__(self):
|
|
|
|
|
|
|
|
argv = sys.argv
|
|
|
|
# Buildozer used to pass these arguments in a now-invalid order
|
|
|
|
# If that happens, apply this fix
|
|
|
|
# This fix will be removed once a fixed buildozer is released
|
|
|
|
if (len(argv) > 2 and
|
|
|
|
argv[1].startswith('--color') and
|
|
|
|
argv[2].startswith('--storage-dir')):
|
|
|
|
argv.append(argv.pop(1)) # the --color arg
|
|
|
|
argv.append(argv.pop(1)) # the --storage-dir arg
|
|
|
|
|
|
|
|
parser = NoAbbrevParser(
|
|
|
|
description=('A packaging tool for turning Python scripts and apps '
|
|
|
|
'into Android APKs'))
|
|
|
|
|
|
|
|
generic_parser = argparse.ArgumentParser(
|
|
|
|
add_help=False,
|
|
|
|
description=('Generic arguments applied to all commands'))
|
|
|
|
dist_parser = argparse.ArgumentParser(
|
|
|
|
add_help=False,
|
|
|
|
description=('Arguments for dist building'))
|
|
|
|
|
|
|
|
generic_parser.add_argument(
|
|
|
|
'--debug', dest='debug', action='store_true',
|
|
|
|
default=False,
|
|
|
|
help='Display debug output and all build info')
|
|
|
|
generic_parser.add_argument(
|
|
|
|
'--color', dest='color', choices=['always', 'never', 'auto'],
|
|
|
|
help='Enable or disable color output (default enabled on tty)')
|
|
|
|
generic_parser.add_argument(
|
|
|
|
'--sdk-dir', '--sdk_dir', dest='sdk_dir', default='',
|
|
|
|
help='The filepath where the Android SDK is installed')
|
|
|
|
generic_parser.add_argument(
|
|
|
|
'--ndk-dir', '--ndk_dir', dest='ndk_dir', default='',
|
|
|
|
help='The filepath where the Android NDK is installed')
|
|
|
|
generic_parser.add_argument(
|
|
|
|
'--android-api', '--android_api', dest='android_api', default=0, type=int,
|
|
|
|
help='The Android API level to build against.')
|
2018-04-11 17:30:16 +02:00
|
|
|
generic_parser.add_argument(
|
|
|
|
'--android-minapi', '--android_minapi', dest='android_min_api', default=0, type=int,
|
|
|
|
help='The Minimum Android API level to build against (will be used for all C compilation).')
|
2017-08-13 03:24:00 +02:00
|
|
|
generic_parser.add_argument(
|
|
|
|
'--ndk-version', '--ndk_version', dest='ndk_version', default='',
|
|
|
|
help=('The version of the Android NDK. This is optional, '
|
|
|
|
'we try to work it out automatically from the ndk_dir.'))
|
|
|
|
generic_parser.add_argument(
|
|
|
|
'--symlink-java-src', '--symlink_java_src',
|
|
|
|
action='store_true',
|
|
|
|
dest='symlink_java_src',
|
|
|
|
default=False,
|
|
|
|
help=('If True, symlinks the java src folder during build and dist '
|
|
|
|
'creation. This is useful for development only, it could also '
|
|
|
|
'cause weird problems.'))
|
|
|
|
|
|
|
|
default_storage_dir = user_data_dir('python-for-android')
|
|
|
|
if ' ' in default_storage_dir:
|
|
|
|
default_storage_dir = '~/.python-for-android'
|
|
|
|
generic_parser.add_argument(
|
|
|
|
'--storage-dir', dest='storage_dir',
|
|
|
|
default=default_storage_dir,
|
|
|
|
help=('Primary storage directory for downloads and builds '
|
|
|
|
'(default: {})'.format(default_storage_dir)))
|
|
|
|
|
|
|
|
generic_parser.add_argument(
|
|
|
|
'--arch',
|
|
|
|
help='The archs to build for, separated by commas.',
|
|
|
|
default='armeabi')
|
|
|
|
|
|
|
|
# Options for specifying the Distribution
|
|
|
|
generic_parser.add_argument(
|
|
|
|
'--dist-name', '--dist_name',
|
|
|
|
help='The name of the distribution to use or create',
|
|
|
|
default='')
|
|
|
|
|
|
|
|
generic_parser.add_argument(
|
|
|
|
'--requirements',
|
|
|
|
help=('Dependencies of your app, should be recipe names or '
|
|
|
|
'Python modules'),
|
|
|
|
default='')
|
|
|
|
|
|
|
|
generic_parser.add_argument(
|
|
|
|
'--bootstrap',
|
|
|
|
help='The bootstrap to build with. Leave unset to choose automatically.',
|
|
|
|
default=None)
|
|
|
|
|
|
|
|
generic_parser.add_argument(
|
|
|
|
'--hook',
|
|
|
|
help='Filename to a module that contain python-for-android hooks',
|
|
|
|
default=None)
|
|
|
|
|
|
|
|
add_boolean_option(
|
|
|
|
generic_parser, ["force-build"],
|
|
|
|
default=False,
|
|
|
|
description='Whether to force compilation of a new distribution:')
|
|
|
|
|
|
|
|
add_boolean_option(
|
|
|
|
generic_parser, ["require-perfect-match"],
|
|
|
|
default=False,
|
|
|
|
description=('Whether the dist recipes must perfectly match '
|
|
|
|
'those requested'))
|
|
|
|
|
|
|
|
generic_parser.add_argument(
|
|
|
|
'--local-recipes', '--local_recipes',
|
|
|
|
dest='local_recipes', default='./p4a-recipes',
|
|
|
|
help='Directory to look for local recipes')
|
|
|
|
|
2017-12-21 08:24:31 +01:00
|
|
|
generic_parser.add_argument(
|
|
|
|
'--java-build-tool',
|
|
|
|
dest='java_build_tool', default='auto',
|
|
|
|
choices=['auto', 'ant', 'gradle'],
|
|
|
|
help=('The java build tool to use when packaging the APK, defaults '
|
|
|
|
'to automatically selecting an appropriate tool.'))
|
|
|
|
|
2017-08-13 03:24:00 +02:00
|
|
|
add_boolean_option(
|
|
|
|
generic_parser, ['copy-libs'],
|
|
|
|
default=False,
|
|
|
|
description='Copy libraries instead of using biglink (Android 4.3+)')
|
|
|
|
|
|
|
|
self._read_configuration()
|
|
|
|
|
|
|
|
subparsers = parser.add_subparsers(dest='subparser_name',
|
|
|
|
help='The command to run')
|
|
|
|
|
|
|
|
def add_parser(subparsers, *args, **kwargs):
|
|
|
|
'''
|
|
|
|
argparse in python2 doesn't support the aliases option,
|
|
|
|
so we just don't provide the aliases there.
|
|
|
|
'''
|
|
|
|
if 'aliases' in kwargs and sys.version_info.major < 3:
|
|
|
|
kwargs.pop('aliases')
|
|
|
|
return subparsers.add_parser(*args, **kwargs)
|
|
|
|
|
|
|
|
parser_recipes = add_parser(subparsers,
|
|
|
|
'recipes',
|
|
|
|
parents=[generic_parser],
|
|
|
|
help='List the available recipes')
|
|
|
|
parser_recipes.add_argument(
|
|
|
|
"--compact", action="store_true", default=False,
|
|
|
|
help="Produce a compact list suitable for scripting")
|
|
|
|
|
|
|
|
parser_bootstraps = add_parser(
|
|
|
|
subparsers, 'bootstraps',
|
|
|
|
help='List the available bootstraps',
|
|
|
|
parents=[generic_parser])
|
|
|
|
parser_clean_all = add_parser(
|
|
|
|
subparsers, 'clean_all',
|
|
|
|
aliases=['clean-all'],
|
|
|
|
help='Delete all builds, dists and caches',
|
|
|
|
parents=[generic_parser])
|
|
|
|
parser_clean_dists = add_parser(
|
|
|
|
subparsers,
|
|
|
|
'clean_dists', aliases=['clean-dists'],
|
|
|
|
help='Delete all dists',
|
|
|
|
parents=[generic_parser])
|
|
|
|
parser_clean_bootstrap_builds = add_parser(
|
|
|
|
subparsers,
|
|
|
|
'clean_bootstrap_builds', aliases=['clean-bootstrap-builds'],
|
|
|
|
help='Delete all bootstrap builds',
|
|
|
|
parents=[generic_parser])
|
|
|
|
parser_clean_builds = add_parser(
|
|
|
|
subparsers,
|
|
|
|
'clean_builds', aliases=['clean-builds'],
|
|
|
|
help='Delete all builds',
|
|
|
|
parents=[generic_parser])
|
|
|
|
|
|
|
|
parser_clean = add_parser(subparsers, 'clean',
|
|
|
|
help='Delete build components.',
|
|
|
|
parents=[generic_parser])
|
|
|
|
parser_clean.add_argument(
|
|
|
|
'component', nargs='+',
|
|
|
|
help=('The build component(s) to delete. You can pass any '
|
|
|
|
'number of arguments from "all", "builds", "dists", '
|
|
|
|
'"distributions", "bootstrap_builds", "downloads".'))
|
|
|
|
|
|
|
|
parser_clean_recipe_build = add_parser(subparsers,
|
|
|
|
'clean_recipe_build', aliases=['clean-recipe-build'],
|
|
|
|
help=('Delete the build components of the given recipe. '
|
|
|
|
'By default this will also delete built dists'),
|
|
|
|
parents=[generic_parser])
|
|
|
|
parser_clean_recipe_build.add_argument('recipe', help='The recipe name')
|
|
|
|
parser_clean_recipe_build.add_argument('--no-clean-dists', default=False,
|
|
|
|
dest='no_clean_dists',
|
|
|
|
action='store_true',
|
|
|
|
help='If passed, do not delete existing dists')
|
|
|
|
|
|
|
|
parser_clean_download_cache= add_parser(subparsers,
|
|
|
|
'clean_download_cache', aliases=['clean-download-cache'],
|
|
|
|
help='Delete cached downloads for requirement builds',
|
|
|
|
parents=[generic_parser])
|
|
|
|
parser_clean_download_cache.add_argument(
|
|
|
|
'recipes', nargs='*',
|
|
|
|
help=('The recipes to clean (space-separated). If no recipe name is '
|
|
|
|
'provided, the entire cache is cleared.'))
|
|
|
|
|
|
|
|
parser_export_dist = add_parser(subparsers,
|
|
|
|
'export_dist', aliases=['export-dist'],
|
|
|
|
help='Copy the named dist to the given path',
|
|
|
|
parents=[generic_parser])
|
|
|
|
parser_export_dist.add_argument('output_dir', help=('The output dir to copy to'))
|
|
|
|
parser_export_dist.add_argument('--symlink', action='store_true',
|
|
|
|
help=('Symlink the dist instead of copying'))
|
|
|
|
|
|
|
|
parser_apk = add_parser(subparsers,
|
|
|
|
'apk', help='Build an APK',
|
|
|
|
parents=[generic_parser])
|
|
|
|
parser_apk.add_argument('--release', dest='build_mode', action='store_const',
|
|
|
|
const='release', default='debug',
|
|
|
|
help='Build the PARSER_APK. in Release mode')
|
|
|
|
parser_apk.add_argument('--keystore', dest='keystore', action='store', default=None,
|
|
|
|
help=('Keystore for JAR signing key, will use jarsigner '
|
|
|
|
'default if not specified (release build only)'))
|
|
|
|
parser_apk.add_argument('--signkey', dest='signkey', action='store', default=None,
|
|
|
|
help='Key alias to sign PARSER_APK. with (release build only)')
|
|
|
|
parser_apk.add_argument('--keystorepw', dest='keystorepw', action='store', default=None,
|
|
|
|
help='Password for keystore')
|
|
|
|
parser_apk.add_argument('--signkeypw', dest='signkeypw', action='store', default=None,
|
|
|
|
help='Password for key alias')
|
|
|
|
|
|
|
|
parser_create = add_parser(subparsers,
|
|
|
|
'create', help='Compile a set of requirements into a dist',
|
|
|
|
parents=[generic_parser])
|
|
|
|
parser_archs = add_parser(subparsers,
|
|
|
|
'archs', help='List the available target architectures',
|
|
|
|
parents=[generic_parser])
|
|
|
|
parser_distributions = add_parser(subparsers,
|
|
|
|
'distributions', aliases=['dists'],
|
|
|
|
help='List the currently available (compiled) dists',
|
|
|
|
parents=[generic_parser])
|
|
|
|
parser_delete_dist = add_parser(subparsers,
|
|
|
|
'delete_dist', aliases=['delete-dist'], help='Delete a compiled dist',
|
|
|
|
parents=[generic_parser])
|
|
|
|
|
|
|
|
parser_sdk_tools = add_parser(subparsers,
|
|
|
|
'sdk_tools', aliases=['sdk-tools'],
|
|
|
|
help='Run the given binary from the SDK tools dis',
|
|
|
|
parents=[generic_parser])
|
|
|
|
parser_sdk_tools.add_argument(
|
|
|
|
'tool', help=('The tool binary name to run'))
|
|
|
|
|
|
|
|
parser_adb = add_parser(subparsers,
|
|
|
|
'adb', help='Run adb from the given SDK',
|
|
|
|
parents=[generic_parser])
|
|
|
|
parser_logcat = add_parser(subparsers,
|
|
|
|
'logcat', help='Run logcat from the given SDK',
|
|
|
|
parents=[generic_parser])
|
|
|
|
parser_build_status = add_parser(subparsers,
|
|
|
|
'build_status', aliases=['build-status'],
|
|
|
|
help='Print some debug information about current built components',
|
|
|
|
parents=[generic_parser])
|
|
|
|
|
2017-12-21 08:24:31 +01:00
|
|
|
parser.add_argument('-v', '--version', action='version', version=__version__)
|
|
|
|
|
2017-08-13 03:24:00 +02:00
|
|
|
args, unknown = parser.parse_known_args(sys.argv[1:])
|
|
|
|
args.unknown_args = unknown
|
|
|
|
|
|
|
|
self.args = args
|
|
|
|
|
|
|
|
if args.subparser_name is None:
|
|
|
|
parser.print_help()
|
|
|
|
exit(1)
|
|
|
|
|
|
|
|
setup_color(args.color)
|
|
|
|
|
|
|
|
if args.debug:
|
|
|
|
logger.setLevel(logging.DEBUG)
|
|
|
|
|
|
|
|
# strip version from requirements, and put them in environ
|
|
|
|
if hasattr(args, 'requirements'):
|
|
|
|
requirements = []
|
|
|
|
for requirement in split_argument_list(args.requirements):
|
|
|
|
if "==" in requirement:
|
|
|
|
requirement, version = requirement.split(u"==", 1)
|
|
|
|
os.environ["VERSION_{}".format(requirement)] = version
|
|
|
|
info('Recipe {}: version "{}" requested'.format(
|
|
|
|
requirement, version))
|
|
|
|
requirements.append(requirement)
|
|
|
|
args.requirements = u",".join(requirements)
|
|
|
|
|
|
|
|
self.ctx = Context()
|
|
|
|
self.storage_dir = args.storage_dir
|
|
|
|
self.ctx.setup_dirs(self.storage_dir)
|
|
|
|
self.sdk_dir = args.sdk_dir
|
|
|
|
self.ndk_dir = args.ndk_dir
|
|
|
|
self.android_api = args.android_api
|
2018-04-11 17:30:16 +02:00
|
|
|
self.android_min_api = args.android_min_api
|
2017-08-13 03:24:00 +02:00
|
|
|
self.ndk_version = args.ndk_version
|
|
|
|
self.ctx.symlink_java_src = args.symlink_java_src
|
2017-12-21 08:24:31 +01:00
|
|
|
self.ctx.java_build_tool = args.java_build_tool
|
2017-08-13 03:24:00 +02:00
|
|
|
|
|
|
|
self._archs = split_argument_list(args.arch)
|
|
|
|
|
|
|
|
self.ctx.local_recipes = args.local_recipes
|
|
|
|
self.ctx.copy_libs = args.copy_libs
|
|
|
|
|
|
|
|
# Each subparser corresponds to a method
|
|
|
|
getattr(self, args.subparser_name.replace('-', '_'))(args)
|
|
|
|
|
|
|
|
def hook(self, name):
|
|
|
|
if not self.args.hook:
|
|
|
|
return
|
|
|
|
if not hasattr(self, "hook_module"):
|
|
|
|
# first time, try to load the hook module
|
|
|
|
self.hook_module = imp.load_source("pythonforandroid.hook", self.args.hook)
|
|
|
|
if hasattr(self.hook_module, name):
|
|
|
|
info("Hook: execute {}".format(name))
|
|
|
|
getattr(self.hook_module, name)(self)
|
|
|
|
else:
|
|
|
|
info("Hook: ignore {}".format(name))
|
|
|
|
|
|
|
|
@property
|
|
|
|
def default_storage_dir(self):
|
|
|
|
udd = user_data_dir('python-for-android')
|
|
|
|
if ' ' in udd:
|
|
|
|
udd = '~/.python-for-android'
|
|
|
|
return udd
|
|
|
|
|
|
|
|
def _read_configuration(self):
|
|
|
|
# search for a .p4a configuration file in the current directory
|
|
|
|
if not exists(".p4a"):
|
|
|
|
return
|
|
|
|
info("Reading .p4a configuration")
|
|
|
|
with open(".p4a") as fd:
|
|
|
|
lines = fd.readlines()
|
|
|
|
lines = [shlex.split(line)
|
|
|
|
for line in lines if not line.startswith("#")]
|
|
|
|
for line in lines:
|
|
|
|
for arg in line:
|
|
|
|
sys.argv.append(arg)
|
|
|
|
|
|
|
|
def recipes(self, args):
|
|
|
|
ctx = self.ctx
|
|
|
|
if args.compact:
|
|
|
|
print(" ".join(set(Recipe.list_recipes(ctx))))
|
|
|
|
else:
|
|
|
|
for name in sorted(Recipe.list_recipes(ctx)):
|
|
|
|
try:
|
|
|
|
recipe = Recipe.get_recipe(name, ctx)
|
|
|
|
except IOError:
|
|
|
|
warning('Recipe "{}" could not be loaded'.format(name))
|
2017-12-21 08:24:31 +01:00
|
|
|
except SyntaxError:
|
|
|
|
import traceback
|
|
|
|
traceback.print_exc()
|
|
|
|
warning(('Recipe "{}" could not be loaded due to a '
|
|
|
|
'syntax error').format(name))
|
2017-08-13 03:24:00 +02:00
|
|
|
version = str(recipe.version)
|
|
|
|
print('{Fore.BLUE}{Style.BRIGHT}{recipe.name:<12} '
|
|
|
|
'{Style.RESET_ALL}{Fore.LIGHTBLUE_EX}'
|
|
|
|
'{version:<8}{Style.RESET_ALL}'.format(
|
|
|
|
recipe=recipe, Fore=Out_Fore, Style=Out_Style,
|
|
|
|
version=version))
|
|
|
|
print(' {Fore.GREEN}depends: {recipe.depends}'
|
|
|
|
'{Fore.RESET}'.format(recipe=recipe, Fore=Out_Fore))
|
|
|
|
if recipe.conflicts:
|
|
|
|
print(' {Fore.RED}conflicts: {recipe.conflicts}'
|
|
|
|
'{Fore.RESET}'
|
|
|
|
.format(recipe=recipe, Fore=Out_Fore))
|
|
|
|
if recipe.opt_depends:
|
|
|
|
print(' {Fore.YELLOW}optional depends: '
|
|
|
|
'{recipe.opt_depends}{Fore.RESET}'
|
|
|
|
.format(recipe=recipe, Fore=Out_Fore))
|
|
|
|
|
|
|
|
def bootstraps(self, args):
|
|
|
|
'''List all the bootstraps available to build with.'''
|
|
|
|
for bs in Bootstrap.list_bootstraps():
|
|
|
|
bs = Bootstrap.get_bootstrap(bs, self.ctx)
|
|
|
|
print('{Fore.BLUE}{Style.BRIGHT}{bs.name}{Style.RESET_ALL}'
|
|
|
|
.format(bs=bs, Fore=Out_Fore, Style=Out_Style))
|
|
|
|
print(' {Fore.GREEN}depends: {bs.recipe_depends}{Fore.RESET}'
|
|
|
|
.format(bs=bs, Fore=Out_Fore))
|
|
|
|
|
|
|
|
def clean(self, args):
|
|
|
|
components = args.component
|
|
|
|
|
|
|
|
component_clean_methods = {'all': self.clean_all,
|
|
|
|
'dists': self.clean_dists,
|
|
|
|
'distributions': self.clean_dists,
|
|
|
|
'builds': self.clean_builds,
|
|
|
|
'bootstrap_builds': self.clean_bootstrap_builds,
|
|
|
|
'downloads': self.clean_download_cache}
|
|
|
|
|
|
|
|
for component in components:
|
|
|
|
if component not in component_clean_methods:
|
|
|
|
raise ValueError((
|
|
|
|
'Asked to clean "{}" but this argument is not '
|
|
|
|
'recognised'.format(component)))
|
|
|
|
component_clean_methods[component](args)
|
2017-12-21 08:24:31 +01:00
|
|
|
|
2017-08-13 03:24:00 +02:00
|
|
|
|
|
|
|
def clean_all(self, args):
|
|
|
|
'''Delete all build components; the package cache, package builds,
|
|
|
|
bootstrap builds and distributions.'''
|
|
|
|
self.clean_dists(args)
|
|
|
|
self.clean_builds(args)
|
|
|
|
self.clean_download_cache(args)
|
|
|
|
|
|
|
|
def clean_dists(self, args):
|
|
|
|
'''Delete all compiled distributions in the internal distribution
|
|
|
|
directory.'''
|
|
|
|
ctx = self.ctx
|
|
|
|
if exists(ctx.dist_dir):
|
|
|
|
shutil.rmtree(ctx.dist_dir)
|
|
|
|
|
|
|
|
def clean_bootstrap_builds(self, args):
|
|
|
|
'''Delete all the bootstrap builds.'''
|
|
|
|
if exists(join(self.ctx.build_dir, 'bootstrap_builds')):
|
|
|
|
shutil.rmtree(join(self.ctx.build_dir, 'bootstrap_builds'))
|
|
|
|
# for bs in Bootstrap.list_bootstraps():
|
|
|
|
# bs = Bootstrap.get_bootstrap(bs, self.ctx)
|
|
|
|
# if bs.build_dir and exists(bs.build_dir):
|
|
|
|
# info('Cleaning build for {} bootstrap.'.format(bs.name))
|
|
|
|
# shutil.rmtree(bs.build_dir)
|
|
|
|
|
|
|
|
def clean_builds(self, args):
|
|
|
|
'''Delete all build caches for each recipe, python-install, java code
|
|
|
|
and compiled libs collection.
|
|
|
|
|
|
|
|
This does *not* delete the package download cache or the final
|
|
|
|
distributions. You can also use clean_recipe_build to delete the build
|
|
|
|
of a specific recipe.
|
|
|
|
'''
|
|
|
|
ctx = self.ctx
|
|
|
|
# if exists(ctx.dist_dir):
|
|
|
|
# shutil.rmtree(ctx.dist_dir)
|
|
|
|
if exists(ctx.build_dir):
|
|
|
|
shutil.rmtree(ctx.build_dir)
|
|
|
|
if exists(ctx.python_installs_dir):
|
|
|
|
shutil.rmtree(ctx.python_installs_dir)
|
|
|
|
libs_dir = join(self.ctx.build_dir, 'libs_collections')
|
|
|
|
if exists(libs_dir):
|
|
|
|
shutil.rmtree(libs_dir)
|
|
|
|
|
|
|
|
def clean_recipe_build(self, args):
|
|
|
|
'''Deletes the build files of the given recipe.
|
|
|
|
|
|
|
|
This is intended for debug purposes, you may experience
|
|
|
|
strange behaviour or problems with some recipes (if their
|
|
|
|
build has done unexpected state changes). If this happens, run
|
|
|
|
clean_builds, or attempt to clean other recipes until things
|
|
|
|
work again.
|
|
|
|
'''
|
|
|
|
recipe = Recipe.get_recipe(args.recipe, self.ctx)
|
|
|
|
info('Cleaning build for {} recipe.'.format(recipe.name))
|
|
|
|
recipe.clean_build()
|
|
|
|
if not args.no_clean_dists:
|
|
|
|
self.clean_dists(args)
|
|
|
|
|
|
|
|
def clean_download_cache(self, args):
|
|
|
|
'''
|
|
|
|
Deletes a download cache for recipes stated as arguments. If no
|
|
|
|
argument is passed, it'll delete *all* downloaded cache. ::
|
|
|
|
|
|
|
|
p4a clean_download_cache kivy,pyjnius
|
|
|
|
|
|
|
|
This does *not* delete the build caches or final distributions.
|
|
|
|
'''
|
|
|
|
ctx = self.ctx
|
|
|
|
if hasattr(args, 'recipes') and args.recipes:
|
|
|
|
for package in args.recipes:
|
|
|
|
remove_path = join(ctx.packages_path, package)
|
|
|
|
if exists(remove_path):
|
|
|
|
shutil.rmtree(remove_path)
|
|
|
|
info('Download cache removed for: "{}"'.format(package))
|
|
|
|
else:
|
|
|
|
warning('No download cache found for "{}", skipping'.format(package))
|
|
|
|
else:
|
|
|
|
if exists(ctx.packages_path):
|
|
|
|
shutil.rmtree(ctx.packages_path)
|
|
|
|
info('Download cache removed.')
|
|
|
|
else:
|
|
|
|
print('No cache found at "{}"'.format(ctx.packages_path))
|
|
|
|
|
|
|
|
@require_prebuilt_dist
|
|
|
|
def export_dist(self, args):
|
|
|
|
'''Copies a created dist to an output dir.
|
|
|
|
|
|
|
|
This makes it easy to navigate to the dist to investigate it
|
|
|
|
or call build.py, though you do not in general need to do this
|
|
|
|
and can use the apk command instead.
|
|
|
|
'''
|
|
|
|
ctx = self.ctx
|
|
|
|
dist = dist_from_args(ctx, args)
|
|
|
|
if dist.needs_build:
|
|
|
|
info('You asked to export a dist, but there is no dist '
|
|
|
|
'with suitable recipes available. For now, you must '
|
|
|
|
' create one first with the create argument.')
|
|
|
|
exit(1)
|
|
|
|
if args.symlink:
|
|
|
|
shprint(sh.ln, '-s', dist.dist_dir, args.output_dir)
|
|
|
|
else:
|
|
|
|
shprint(sh.cp, '-r', dist.dist_dir, args.output_dir)
|
|
|
|
|
|
|
|
@property
|
|
|
|
def _dist(self):
|
|
|
|
ctx = self.ctx
|
|
|
|
dist = dist_from_args(ctx, self.args)
|
|
|
|
return dist
|
|
|
|
|
|
|
|
@require_prebuilt_dist
|
|
|
|
def apk(self, args):
|
|
|
|
'''Create an APK using the given distribution.'''
|
|
|
|
|
|
|
|
ctx = self.ctx
|
|
|
|
dist = self._dist
|
|
|
|
|
|
|
|
# Manually fixing these arguments at the string stage is
|
|
|
|
# unsatisfactory and should probably be changed somehow, but
|
|
|
|
# we can't leave it until later as the build.py scripts assume
|
|
|
|
# they are in the current directory.
|
|
|
|
fix_args = ('--dir', '--private', '--add-jar', '--add-source',
|
|
|
|
'--whitelist', '--blacklist', '--presplash', '--icon')
|
|
|
|
unknown_args = args.unknown_args
|
|
|
|
for i, arg in enumerate(unknown_args[:-1]):
|
|
|
|
argx = arg.split('=')
|
|
|
|
if argx[0] in fix_args:
|
|
|
|
if len(argx) > 1:
|
|
|
|
unknown_args[i] = '='.join((argx[0],
|
|
|
|
realpath(expanduser(argx[1]))))
|
|
|
|
else:
|
|
|
|
unknown_args[i+1] = realpath(expanduser(unknown_args[i+1]))
|
|
|
|
|
|
|
|
env = os.environ.copy()
|
|
|
|
if args.build_mode == 'release':
|
|
|
|
if args.keystore:
|
|
|
|
env['P4A_RELEASE_KEYSTORE'] = realpath(expanduser(args.keystore))
|
|
|
|
if args.signkey:
|
|
|
|
env['P4A_RELEASE_KEYALIAS'] = args.signkey
|
|
|
|
if args.keystorepw:
|
|
|
|
env['P4A_RELEASE_KEYSTORE_PASSWD'] = args.keystorepw
|
|
|
|
if args.signkeypw:
|
|
|
|
env['P4A_RELEASE_KEYALIAS_PASSWD'] = args.signkeypw
|
|
|
|
elif args.keystorepw and 'P4A_RELEASE_KEYALIAS_PASSWD' not in env:
|
|
|
|
env['P4A_RELEASE_KEYALIAS_PASSWD'] = args.keystorepw
|
|
|
|
|
|
|
|
build = imp.load_source('build', join(dist.dist_dir, 'build.py'))
|
|
|
|
with current_directory(dist.dist_dir):
|
|
|
|
self.hook("before_apk_build")
|
2017-12-21 08:24:31 +01:00
|
|
|
os.environ["ANDROID_API"] = str(self.ctx.android_api)
|
2017-08-13 03:24:00 +02:00
|
|
|
build_args = build.parse_args(args.unknown_args)
|
|
|
|
self.hook("after_apk_build")
|
|
|
|
self.hook("before_apk_assemble")
|
|
|
|
|
2017-12-21 08:24:31 +01:00
|
|
|
build_type = ctx.java_build_tool
|
|
|
|
if build_type == 'auto':
|
|
|
|
info('Selecting java build tool:')
|
|
|
|
|
|
|
|
build_tools_versions = os.listdir(join(ctx.sdk_dir, 'build-tools'))
|
|
|
|
build_tools_versions = sorted(build_tools_versions,
|
|
|
|
key=LooseVersion)
|
|
|
|
build_tools_version = build_tools_versions[-1]
|
|
|
|
info(('Detected highest available build tools '
|
|
|
|
'version to be {}').format(build_tools_version))
|
|
|
|
|
|
|
|
if build_tools_version >= '25.0' and exists('gradlew'):
|
|
|
|
build_type = 'gradle'
|
|
|
|
info(' Building with gradle, as gradle executable is present')
|
|
|
|
else:
|
|
|
|
build_type = 'ant'
|
|
|
|
if build_tools_version < '25.0':
|
|
|
|
info((' Building with ant, as the highest '
|
|
|
|
'build-tools-version is only {}').format(build_tools_version))
|
|
|
|
else:
|
|
|
|
info(' Building with ant, as no gradle executable detected')
|
|
|
|
|
|
|
|
if build_type == 'gradle':
|
|
|
|
# gradle-based build
|
|
|
|
env["ANDROID_NDK_HOME"] = self.ctx.ndk_dir
|
|
|
|
env["ANDROID_HOME"] = self.ctx.sdk_dir
|
|
|
|
|
|
|
|
gradlew = sh.Command('./gradlew')
|
|
|
|
if args.build_mode == "debug":
|
|
|
|
gradle_task = "assembleDebug"
|
|
|
|
elif args.build_mode == "release":
|
|
|
|
gradle_task = "assembleRelease"
|
|
|
|
else:
|
|
|
|
error("Unknown build mode {} for apk()".format(
|
|
|
|
args.build_mode))
|
|
|
|
exit(1)
|
|
|
|
output = shprint(gradlew, gradle_task, _tail=20,
|
|
|
|
_critical=True, _env=env)
|
|
|
|
|
|
|
|
# gradle output apks somewhere else
|
|
|
|
# and don't have version in file
|
|
|
|
apk_dir = join(dist.dist_dir, "build", "outputs", "apk")
|
|
|
|
apk_glob = "*-{}.apk"
|
|
|
|
apk_add_version = True
|
|
|
|
|
|
|
|
else:
|
|
|
|
# ant-based build
|
|
|
|
try:
|
|
|
|
ant = sh.Command('ant')
|
|
|
|
except sh.CommandNotFound:
|
|
|
|
error('Could not find ant binary, please install it '
|
|
|
|
'and make sure it is in your $PATH.')
|
|
|
|
exit(1)
|
|
|
|
output = shprint(ant, args.build_mode, _tail=20,
|
|
|
|
_critical=True, _env=env)
|
|
|
|
apk_dir = join(dist.dist_dir, "bin")
|
|
|
|
apk_glob = "*-*-{}.apk"
|
|
|
|
apk_add_version = False
|
2017-08-13 03:24:00 +02:00
|
|
|
|
|
|
|
self.hook("after_apk_assemble")
|
|
|
|
|
|
|
|
info_main('# Copying APK to current directory')
|
|
|
|
|
|
|
|
apk_re = re.compile(r'.*Package: (.*\.apk)$')
|
|
|
|
apk_file = None
|
|
|
|
for line in reversed(output.splitlines()):
|
|
|
|
m = apk_re.match(line)
|
|
|
|
if m:
|
|
|
|
apk_file = m.groups()[0]
|
|
|
|
break
|
|
|
|
|
|
|
|
if not apk_file:
|
|
|
|
info_main('# APK filename not found in build output, trying to guess')
|
2017-12-21 08:24:31 +01:00
|
|
|
if args.build_mode == "release":
|
|
|
|
suffixes = ("release", "release-unsigned")
|
|
|
|
else:
|
|
|
|
suffixes = ("debug", )
|
|
|
|
for suffix in suffixes:
|
|
|
|
apks = glob.glob(join(apk_dir, apk_glob.format(suffix)))
|
|
|
|
if apks:
|
|
|
|
if len(apks) > 1:
|
|
|
|
info('More than one built APK found... guessing you '
|
|
|
|
'just built {}'.format(apks[-1]))
|
|
|
|
apk_file = apks[-1]
|
|
|
|
break
|
|
|
|
else:
|
2017-08-13 03:24:00 +02:00
|
|
|
raise ValueError('Couldn\'t find the built APK')
|
|
|
|
|
|
|
|
info_main('# Found APK file: {}'.format(apk_file))
|
2017-12-21 08:24:31 +01:00
|
|
|
if apk_add_version:
|
|
|
|
info('# Add version number to APK')
|
|
|
|
apk_name, apk_suffix = basename(apk_file).split("-", 1)
|
|
|
|
apk_file_dest = "{}-{}-{}".format(
|
|
|
|
apk_name, build_args.version, apk_suffix)
|
|
|
|
info('# APK renamed to {}'.format(apk_file_dest))
|
|
|
|
shprint(sh.cp, apk_file, apk_file_dest)
|
|
|
|
else:
|
|
|
|
shprint(sh.cp, apk_file, './')
|
2017-08-13 03:24:00 +02:00
|
|
|
|
|
|
|
@require_prebuilt_dist
|
|
|
|
def create(self, args):
|
|
|
|
'''Create a distribution directory if it doesn't already exist, run
|
|
|
|
any recipes if necessary, and build the apk.
|
|
|
|
'''
|
|
|
|
pass # The decorator does everything
|
|
|
|
|
|
|
|
def archs(self, args):
|
|
|
|
'''List the target architectures available to be built for.'''
|
|
|
|
print('{Style.BRIGHT}Available target architectures are:'
|
|
|
|
'{Style.RESET_ALL}'.format(Style=Out_Style))
|
|
|
|
for arch in self.ctx.archs:
|
|
|
|
print(' {}'.format(arch.arch))
|
|
|
|
|
|
|
|
def dists(self, args):
|
|
|
|
'''The same as :meth:`distributions`.'''
|
|
|
|
self.distributions(args)
|
|
|
|
|
|
|
|
def distributions(self, args):
|
|
|
|
'''Lists all distributions currently available (i.e. that have already
|
|
|
|
been built).'''
|
|
|
|
ctx = self.ctx
|
|
|
|
dists = Distribution.get_distributions(ctx)
|
|
|
|
|
|
|
|
if dists:
|
|
|
|
print('{Style.BRIGHT}Distributions currently installed are:'
|
|
|
|
'{Style.RESET_ALL}'.format(Style=Out_Style, Fore=Out_Fore))
|
|
|
|
pretty_log_dists(dists, print)
|
|
|
|
else:
|
|
|
|
print('{Style.BRIGHT}There are no dists currently built.'
|
|
|
|
'{Style.RESET_ALL}'.format(Style=Out_Style))
|
|
|
|
|
|
|
|
def delete_dist(self, args):
|
|
|
|
dist = self._dist
|
|
|
|
if dist.needs_build:
|
|
|
|
info('No dist exists that matches your specifications, '
|
|
|
|
'exiting without deleting.')
|
|
|
|
shutil.rmtree(dist.dist_dir)
|
|
|
|
|
|
|
|
def sdk_tools(self, args):
|
|
|
|
'''Runs the android binary from the detected SDK directory, passing
|
|
|
|
all arguments straight to it. This binary is used to install
|
|
|
|
e.g. platform-tools for different API level targets. This is
|
|
|
|
intended as a convenience function if android is not in your
|
|
|
|
$PATH.
|
|
|
|
'''
|
|
|
|
ctx = self.ctx
|
|
|
|
ctx.prepare_build_environment(user_sdk_dir=self.sdk_dir,
|
|
|
|
user_ndk_dir=self.ndk_dir,
|
|
|
|
user_android_api=self.android_api,
|
2018-04-11 17:30:16 +02:00
|
|
|
user_android_min_api=self.android_min_api,
|
2017-08-13 03:24:00 +02:00
|
|
|
user_ndk_ver=self.ndk_version)
|
|
|
|
android = sh.Command(join(ctx.sdk_dir, 'tools', args.tool))
|
|
|
|
output = android(
|
|
|
|
*args.unknown_args, _iter=True, _out_bufsize=1, _err_to_out=True)
|
|
|
|
for line in output:
|
|
|
|
sys.stdout.write(line)
|
|
|
|
sys.stdout.flush()
|
|
|
|
|
|
|
|
def adb(self, args):
|
|
|
|
'''Runs the adb binary from the detected SDK directory, passing all
|
|
|
|
arguments straight to it. This is intended as a convenience
|
|
|
|
function if adb is not in your $PATH.
|
|
|
|
'''
|
|
|
|
self._adb(args.unknown_args)
|
|
|
|
|
|
|
|
def logcat(self, args):
|
|
|
|
'''Runs ``adb logcat`` using the adb binary from the detected SDK
|
|
|
|
directory. All extra args are passed as arguments to logcat.'''
|
|
|
|
self._adb(['logcat'] + args.unknown_args)
|
|
|
|
|
|
|
|
def _adb(self, commands):
|
|
|
|
'''Call the adb executable from the SDK, passing the given commands as
|
|
|
|
arguments.'''
|
|
|
|
ctx = self.ctx
|
|
|
|
ctx.prepare_build_environment(user_sdk_dir=self.sdk_dir,
|
|
|
|
user_ndk_dir=self.ndk_dir,
|
|
|
|
user_android_api=self.android_api,
|
2018-04-11 17:30:16 +02:00
|
|
|
user_android_min_api=self.android_min_api,
|
2017-08-13 03:24:00 +02:00
|
|
|
user_ndk_ver=self.ndk_version)
|
|
|
|
if platform in ('win32', 'cygwin'):
|
|
|
|
adb = sh.Command(join(ctx.sdk_dir, 'platform-tools', 'adb.exe'))
|
|
|
|
else:
|
|
|
|
adb = sh.Command(join(ctx.sdk_dir, 'platform-tools', 'adb'))
|
|
|
|
info_notify('Starting adb...')
|
|
|
|
output = adb(*commands, _iter=True, _out_bufsize=1, _err_to_out=True)
|
|
|
|
for line in output:
|
|
|
|
sys.stdout.write(line)
|
|
|
|
sys.stdout.flush()
|
|
|
|
|
|
|
|
|
|
|
|
def build_status(self, args):
|
|
|
|
print('{Style.BRIGHT}Bootstraps whose core components are probably '
|
|
|
|
'already built:{Style.RESET_ALL}'.format(Style=Out_Style))
|
|
|
|
|
|
|
|
bootstrap_dir = join(self.ctx.build_dir, 'bootstrap_builds')
|
|
|
|
if exists(bootstrap_dir):
|
|
|
|
for filen in os.listdir(bootstrap_dir):
|
|
|
|
print(' {Fore.GREEN}{Style.BRIGHT}{filen}{Style.RESET_ALL}'
|
|
|
|
.format(filen=filen, Fore=Out_Fore, Style=Out_Style))
|
|
|
|
|
|
|
|
print('{Style.BRIGHT}Recipes that are probably already built:'
|
|
|
|
'{Style.RESET_ALL}'.format(Style=Out_Style))
|
|
|
|
other_builds_dir = join(self.ctx.build_dir, 'other_builds')
|
|
|
|
if exists(other_builds_dir):
|
|
|
|
for filen in sorted(os.listdir(other_builds_dir)):
|
|
|
|
name = filen.split('-')[0]
|
|
|
|
dependencies = filen.split('-')[1:]
|
|
|
|
recipe_str = (' {Style.BRIGHT}{Fore.GREEN}{name}'
|
|
|
|
'{Style.RESET_ALL}'.format(
|
|
|
|
Style=Out_Style, name=name, Fore=Out_Fore))
|
|
|
|
if dependencies:
|
|
|
|
recipe_str += (
|
|
|
|
' ({Fore.BLUE}with ' + ', '.join(dependencies) +
|
|
|
|
'{Fore.RESET})').format(Fore=Out_Fore)
|
|
|
|
recipe_str += '{Style.RESET_ALL}'.format(Style=Out_Style)
|
|
|
|
print(recipe_str)
|
|
|
|
|
|
|
|
|
|
|
|
def main():
|
|
|
|
ToolchainCL()
|
|
|
|
|
|
|
|
if __name__ == "__main__":
|
|
|
|
main()
|